diff --git "a/1414.jsonl" "b/1414.jsonl" new file mode 100644--- /dev/null +++ "b/1414.jsonl" @@ -0,0 +1,445 @@ +{"seq_id": "433515869", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Sat Jul 7 20:23:22 2018\r\n\r\n@author: Debashree_Debalaxmi\r\n\"\"\"\r\n\r\nimport pandas as pd\r\nimport matplotlib.pyplot as plt\r\ndf=pd.read_excel(\"C:\\\\Users\\\\acer\\\\Desktop\\\\youtube.xlsx\")\r\nprint(df)\r\n#to find out top 5 categories with max no of videos uploaded\r\npp = df.groupby('category').size().sort_values(ascending=False)[:5]\r\nprint(pp)\r\nplt.plot(df[\"category\"])\r\nplt.title(\"Graph\")\r\nplt.xlabel(\"X-axis\")\r\nplt.ylabel(\"Y-axis\")\r\nplt.grid(True)\r\nplt.show()\r\n", "sub_path": "youtube.py", "file_name": "youtube.py", "file_ext": "py", "file_size_in_byte": 491, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pandas.read_excel", "line_number": 10, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}]} +{"seq_id": "23973338", "text": "import sys\n\nimport pytest\nimport xarray as xr\nimport xgcm\n\nimport pop_tools\nfrom pop_tools import DATASETS\n\n\n@pytest.mark.parametrize(\n 'file', ['tend_zint_100m_Fe.nc', 'g.e20.G.TL319_t13.control.001_hfreq-coarsen.nc']\n)\ndef test_to_xgcm_grid_dataset(file):\n filepath = DATASETS.fetch(file)\n ds = xr.open_dataset(filepath)\n grid, ds_new = pop_tools.to_xgcm_grid_dataset(ds, metrics=None)\n assert isinstance(grid, xgcm.Grid)\n assert set(['X', 'Y', 'Z']) == set(grid.axes.keys())\n new_spatial_coords = ['nlon_u', 'nlat_u', 'nlon_t', 'nlat_t']\n for coord in new_spatial_coords:\n assert coord in ds_new.coords\n assert coord not in ds.coords\n old_spatial_coords = ['nlat', 'nlon', 'z_w']\n for coord in old_spatial_coords:\n assert coord not in ds_new.coords\n\n\ndef test_to_xgcm_grid_dataset_missing_xgcm():\n from unittest import mock\n\n with pytest.raises(ImportError):\n with mock.patch.dict(sys.modules, {'xgcm': None}):\n filepath = DATASETS.fetch('tend_zint_100m_Fe.nc')\n ds = xr.open_dataset(filepath)\n _, _ = pop_tools.to_xgcm_grid_dataset(ds, metrics=None)\n", "sub_path": "tests/test_xgcm_util.py", "file_name": "test_xgcm_util.py", "file_ext": "py", "file_size_in_byte": 1151, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pop_tools.DATASETS.fetch", "line_number": 15, "usage_type": "call"}, {"api_name": "pop_tools.DATASETS", "line_number": 15, "usage_type": "name"}, {"api_name": "xarray.open_dataset", "line_number": 16, "usage_type": "call"}, {"api_name": "pop_tools.to_xgcm_grid_dataset", "line_number": 17, "usage_type": "call"}, {"api_name": "xgcm.Grid", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 11, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pytest.raises", "line_number": 32, "usage_type": "call"}, {"api_name": "unittest.mock.patch.dict", "line_number": 33, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 33, "usage_type": "attribute"}, {"api_name": "unittest.mock", "line_number": 33, "usage_type": "name"}, {"api_name": "sys.modules", "line_number": 33, "usage_type": "attribute"}, {"api_name": "pop_tools.DATASETS.fetch", "line_number": 34, "usage_type": "call"}, {"api_name": "pop_tools.DATASETS", "line_number": 34, "usage_type": "name"}, {"api_name": "xarray.open_dataset", "line_number": 35, "usage_type": "call"}, {"api_name": "pop_tools.to_xgcm_grid_dataset", "line_number": 36, "usage_type": "call"}]} +{"seq_id": "571502130", "text": "import os\n\nimport environ\nfrom django.core.exceptions import ImproperlyConfigured, ValidationError\nfrom django.core.validators import EmailValidator\n\nBASE_DIR = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))\nENV_PATH = os.path.join(BASE_DIR, \".env\")\n\n\ndef raise_env_error(varname, msg):\n raise ImproperlyConfigured(f\"Configuration error for {ENV_PATH}:{varname} : {msg}\")\n\n\ndef validate_env_email(email, varname):\n try:\n EmailValidator()(email)\n except ValidationError:\n raise_env_error(varname, f\"{email} is an invalid email\")\n\n\ndef read_env(*args, **kwargs):\n env = environ.Env(*args, **kwargs)\n env.read_env(ENV_PATH)\n return env\n\n\nenv = read_env()\n\n\nDEBUG = env.bool(\"DEBUG\")\n\nINSTALLED_APPS = [\n \"permapp.admin.CustomAdminConfig\",\n \"django.contrib.auth\",\n \"django.contrib.contenttypes\",\n \"django.contrib.sessions\",\n \"django.contrib.messages\",\n \"django.contrib.staticfiles\",\n \"corsheaders\",\n \"rest_framework\",\n \"admin_auto_filters\",\n \"django_filters\",\n \"categories.editor\",\n \"inline_actions\",\n \"django_json_widget\",\n \"jsonschemaform\",\n \"polymorphic\",\n \"solo\",\n \"tagging\",\n \"colorfield\",\n \"leaflet\",\n \"djgeojson\",\n \"custom_auth\",\n \"designs\",\n]\n\nMIDDLEWARE = [\n \"corsheaders.middleware.CorsMiddleware\",\n \"django.middleware.security.SecurityMiddleware\",\n \"django.contrib.sessions.middleware.SessionMiddleware\",\n \"django.middleware.common.CommonMiddleware\",\n \"django.middleware.csrf.CsrfViewMiddleware\",\n \"django.contrib.auth.middleware.AuthenticationMiddleware\",\n \"django.contrib.messages.middleware.MessageMiddleware\",\n \"django.middleware.clickjacking.XFrameOptionsMiddleware\",\n]\n\nROOT_URLCONF = \"permapp.urls\"\n\nTEMPLATES = [\n {\n \"BACKEND\": \"django.template.backends.django.DjangoTemplates\",\n \"DIRS\": [os.path.join(BASE_DIR, \"templates\")],\n \"APP_DIRS\": True,\n \"OPTIONS\": {\n \"context_processors\": [\n \"django.template.context_processors.debug\",\n \"django.template.context_processors.request\",\n \"django.contrib.auth.context_processors.auth\",\n \"django.contrib.messages.context_processors.messages\",\n ]\n },\n }\n]\n\nWSGI_APPLICATION = \"permapp.wsgi.application\"\n", "sub_path": "permapp/settings/base.py", "file_name": "base.py", "file_ext": "py", "file_size_in_byte": 2330, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "os.path.dirname", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.core.exceptions.ImproperlyConfigured", "line_number": 12, "usage_type": "call"}, {"api_name": "django.core.validators.EmailValidator", "line_number": 17, "usage_type": "call"}, {"api_name": "django.core.exceptions.ValidationError", "line_number": 18, "usage_type": "name"}, {"api_name": "environ.Env", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path", "line_number": 74, "usage_type": "attribute"}]} +{"seq_id": "533583234", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nimport datetime\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('storage', '__first__'),\n ]\n\n operations = [\n migrations.CreateModel(\n name='FundRaising',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('begin_date', models.DateField(default=datetime.datetime(2014, 9, 9, 9, 53, 51, 620440), verbose_name=b'Date de d\\xc3\\xa9but')),\n ('end_date', models.DateField(default=datetime.datetime(2014, 9, 9, 9, 53, 51, 620533), verbose_name=b'Date de fin')),\n ('status', models.CharField(default=b'current', max_length=10, verbose_name=b'Statut', choices=[(b'current', b'En cours'), (b'closing', b'En cl\\xc3\\xb4ture'), (b'closed', b'Cl\\xc3\\xb4tur\\xc3\\xa9e')])),\n ('is_closed', models.BooleanField(default=False, verbose_name=b'Cl\\xc3\\xb4tur\\xc3\\xa9e')),\n ],\n options={\n },\n bases=(models.Model,),\n ),\n migrations.CreateModel(\n name='InvestBook',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('cap_table', models.ForeignKey(related_name=b'cap_table', verbose_name=b'Table de capitalisation avant Lev\\xc3\\xa9e de Fonds', to='storage.KYCFile')),\n ('executive_summary', models.ForeignKey(related_name=b'executive_summary', verbose_name=b'Executive summary', to='storage.KYCFile')),\n ('operation_note', models.ForeignKey(related_name=b'operation_note', verbose_name=b'Note d\\xe2\\x80\\x99op\\xc3\\xa9ration', to='storage.KYCFile')),\n ('pitch_deck', models.ForeignKey(related_name=b'pitch_deck', verbose_name=b'Pitch Deck', to='storage.KYCFile')),\n ('statuts', models.ForeignKey(related_name=b'statuts', verbose_name=b'Status', to='storage.KYCFile')),\n ],\n options={\n },\n bases=(models.Model,),\n ),\n migrations.CreateModel(\n name='RaisingTerm',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('lead_part', models.FloatField(verbose_name=b'Apport du Lead (en \\xe2\\x82\\xac)')),\n ('round_size', models.FloatField(verbose_name=b'Montant de lev\\xc3\\xa9e (en \\xe2\\x82\\xac)')),\n ('minimum_investment', models.FloatField(verbose_name=b'Ticket minimum (en \\xe2\\x82\\xac) pour un utilisateur particulier')),\n ('invest_type', models.CharField(default=b'HOL', max_length=20, verbose_name=b'Type d\\xe2\\x80\\x99investissement', choices=[(b'HOL', b'Holding'), (b'DIR', b'Direct')])),\n ('product', models.CharField(default=b'SHARES', max_length=20, verbose_name=b'Instrument', choices=[(b'SHARES', b'Actions')])),\n ('valuation', models.FloatField(verbose_name=b'Valorisation (en \\xe2\\x82\\xac)')),\n ('new_investors_part', models.IntegerField(verbose_name=b'Pourcentage c\\xc3\\xa9d\\xc3\\xa9 aux nouveaux investisseurs (en %)')),\n ('subscription_date', models.DateField(verbose_name=b'Date limite de souscription')),\n ],\n options={\n },\n bases=(models.Model,),\n ),\n migrations.AddField(\n model_name='fundraising',\n name='invest_book',\n field=models.OneToOneField(to='fundraising.InvestBook'),\n preserve_default=True,\n ),\n migrations.AddField(\n model_name='fundraising',\n name='raising_term',\n field=models.OneToOneField(to='fundraising.RaisingTerm'),\n preserve_default=True,\n ),\n ]\n", "sub_path": "api/fundraising/migrations/0001_initial.py", "file_name": "0001_initial.py", "file_ext": "py", "file_size_in_byte": 3930, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 15, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models.DateField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 26, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 28, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 28, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 31, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 31, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 32, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 32, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 33, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 34, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 34, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 35, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 35, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 36, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 36, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 40, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 40, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 42, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 42, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 45, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 45, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 46, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 46, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 47, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 47, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 48, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 48, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 49, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 49, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 50, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 50, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 51, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 51, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 52, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 52, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 53, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 53, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 57, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 57, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 59, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 59, "usage_type": "name"}, {"api_name": "django.db.models.OneToOneField", "line_number": 62, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 62, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 65, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 65, "usage_type": "name"}, {"api_name": "django.db.models.OneToOneField", "line_number": 68, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 68, "usage_type": "name"}]} +{"seq_id": "34905618", "text": "import telebot\nimport requests\nfrom bs4 import BeautifulSoup\n\nbot = telebot.TeleBot('278758102:AAENnsnUdlaLTklqCaoj0F8hSckxoR-_t10')\n\n\n\n@bot.message_handler(commands=['start'])\ndef default_test(message):\n bot.send_message(message.from_user.id, 'Please input audio track title or performer\\'s name.')\n\n\n@bot.message_handler(content_types=[\"text\"])\ndef _query(message):\n url = 'https://downloadmusicvk.ru/audio/search?q=' + parser(message.text)\n source_code = requests.get(url)\n plain_text = source_code.text\n soup = BeautifulSoup(plain_text, \"html.parser\")\n counter = 1\n for link in soup.findAll('div', {'class' : 'row audio'}, limit=5):\n name_of_song = link.find('div', {'class' : 'col-lg-9 col-md-8 col-sm-7 col-xs-5'}).text\n # go to 3rd block\n link_to_download = link.find('div', {'class' : 'col-lg-2 col-md-3 col-sm-4 col-xs-5'})\n # go to download button\n link_to_download = link_to_download.find('a', {'class' : 'btn btn-primary btn-xs download'})\n \n # link to page to download\n main_link = 'https://downloadmusicvk.ru' + link_to_download.get('href')\n # link to download\n answer = handle_song(main_link)\n \n code = '' + str(counter) + '. ' + name_of_song.strip() + ''\n bot.send_message(message.from_user.id, parse_mode='HTML', text = code)\n counter += 1\n\ndef handle_song(main_link):\n source_code = requests.get(main_link)\n plain_text = source_code.text\n soup = BeautifulSoup(plain_text, \"html.parser\")\n button = soup.find('a', {'class' : 'btn btn-success btn-lg btn-block download'})\n return 'https://downloadmusicvk.ru' + button.get('href')\n\n\ndef parser(message):\n message = message.strip()\n answer = \"\"\n for c in message:\n if c != \" \":\n answer += c\n else:\n answer += '+'\n return answer\n\n\nif __name__ == '__main__':\n bot.polling(none_stop=True)\n\n\n", "sub_path": "sample.py", "file_name": "sample.py", "file_ext": "py", "file_size_in_byte": 1967, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "telebot.TeleBot", "line_number": 5, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 17, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 19, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 38, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 40, "usage_type": "call"}]} +{"seq_id": "314820446", "text": "from functools import cmp_to_key\nclass Player:\n def __init__(self, name, score):\n self.name = name\n self.score = score\n \n def __repr__(self):\n return \"{}: {}\".format(self.name, self.score)\n \n def comparator(a, b):\n if a.score != b.score:\n return b.score - a.score\n else:\n if a.name == b.name:\n return 0\n else: \n return -1 if a.name < b.name else 1\n\ninputs = [('allen1', 100), ('allen2', 90), ('ben1', 100), ('ben2', 50)]\ndata = []\nfor i in inputs:\n name, score = i\n data.append(Player(name, score))\n\nprint(data) \ndata = sorted(data, key=cmp_to_key(Player.comparator))\n# for i in data:\n# print(i.name, i.score)\n", "sub_path": "algo_comparator.py", "file_name": "algo_comparator.py", "file_ext": "py", "file_size_in_byte": 717, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "functools.cmp_to_key", "line_number": 26, "usage_type": "call"}]} +{"seq_id": "93728991", "text": "import pymysql\nfrom common.settings import configData\n\ndata = configData()['mysql']\n\nclass Mysql():\n '''数据库操作'''\n\n def __init__(self,databaseName='test_tcljx_rpm'):\n self.db = pymysql.connect(host= data['host'],\n user = data['user'],\n password = data['password'],\n port = data['port'],\n charset = data['charset'],\n database= databaseName)\n self.c = self.db.cursor()\n\n\n def select(self,sql,fetch=True):\n '''\n :param sql: 执行语句\n :param fetch: 返回one/all\n :return:\n '''\n try:\n self.c.execute(sql)\n except:\n self.db.rollback()\n else:\n if fetch == True:\n return self.c.fetchone()[0]\n else:\n return self.c.fetchall()[0]\n finally:\n self.db.close()\n self.c.close()\n\n def delect(self):\n print()\n\nif __name__ == '__main__':\n\n print(True is None)\n\n\n\n", "sub_path": "common/mysql.py", "file_name": "mysql.py", "file_ext": "py", "file_size_in_byte": 1114, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "common.settings.configData", "line_number": 4, "usage_type": "call"}, {"api_name": "pymysql.connect", "line_number": 10, "usage_type": "call"}]} +{"seq_id": "254015859", "text": "import os\n\nfrom setuptools import setup, find_packages\n\ntry:\n if os.environ.get('CI_COMMIT_TAG'):\n if os.environ['CI_COMMIT_TAG'].startswith('v'):\n version = os.environ['CI_COMMIT_TAG'][1:]\n else:\n version = os.environ['CI_COMMIT_TAG']\n else:\n version = '0.' + os.environ['CI_JOB_ID'] # Use job ID if no commmit tag provided\nexcept KeyError:\n import datetime\n\n version = '0.' + str(datetime.datetime.now())[0:23].replace(' ', '-').replace(':', '')\n\nsetup(name='gym_grasshoppers',\n description='OpenAI Gym environment for Grasshoppers project',\n version=version,\n url='https://gitlab.com/kdg-ti/integratieproject-2/dekwo-kybons-fanclub/environment-ai',\n author='Dekwo Kybons Fanclub',\n author_email='mees.vankaam@student.kdg.be',\n packages=find_packages(),\n zip_safe=True,\n install_requires=['gym>=0.16', 'numpy', 'shapely', 'matplotlib', 'pyproj<=2.4.1']\n )\n", "sub_path": "pypi_install_script/gym_grasshoppers-0.484241930.tar/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 959, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "os.environ.get", "line_number": 6, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 12, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 16, "usage_type": "attribute"}, {"api_name": "setuptools.setup", "line_number": 18, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 24, "usage_type": "call"}]} +{"seq_id": "588522889", "text": "import calendar\nimport datetime\nfrom collections import defaultdict\n\nfrom dateutil.relativedelta import relativedelta\nfrom django.db.models import Sum, Q\nfrom django.db.models.functions import Coalesce\nfrom rest_framework import viewsets, status\nfrom rest_framework.decorators import action\nfrom rest_framework.response import Response\n\nfrom balance.models import Category, Balance\nfrom balance.serializers import (\n CategorySerializer,\n SimplyBalanceSerializer,\n CategorySimplySerializer,\n CategoryBalanceSerializer,\n BalanceSerializer,\n)\n\n\nclass CategoryView(viewsets.ModelViewSet):\n serializer_class = CategorySerializer\n\n def get_queryset(self):\n return Category.objects.filter(user=self.request.user)\n\n def list(self, request, *args, **kwargs):\n categories = Category.objects.filter(\n user=request.user,\n is_income=(request.query_params.get(\"isIncome\") == \"true\"),\n )\n return Response(CategorySimplySerializer(categories, many=True).data)\n\n def create(self, request, *args, **kwargs):\n serializer = self.get_serializer(data={**request.data, \"user\": request.user.id})\n serializer.is_valid(raise_exception=True)\n\n category = Category.objects.create(**serializer.validated_data)\n\n return Response(\n CategoryBalanceSerializer(category).data, status=status.HTTP_201_CREATED\n )\n\n @action(methods=[\"get\"], detail=False)\n def list_with_balance(self, request, *args, **kwargs):\n categories = Category.objects.filter(\n user=request.user,\n is_income=(request.query_params.get(\"isIncome\") == \"true\"),\n )\n return Response(CategoryBalanceSerializer(categories, many=True).data)\n\n\nclass BalanceView(viewsets.ModelViewSet):\n serializer_class = SimplyBalanceSerializer\n\n def get_queryset(self):\n return Balance.objects.filter(category__user=self.request.user)\n\n def list(self, request, *args, **kwargs):\n self.serializer_class = BalanceSerializer\n return super().list(request, *args, **kwargs)\n\n @action(methods=[\"get\"], detail=False)\n def balance_summary(self, request, *args, **kwargs):\n user_balances = Balance.objects.filter(category__user=request.user)\n today = datetime.date.today()\n\n q_incomes_total = Q(category__user=request.user, category__is_income=True)\n q_incomes_monthly = Q(\n category__user=request.user,\n category__is_income=True,\n date__month=today.month,\n )\n q_incomes_today = Q(\n category__user=request.user, category__is_income=True, date=today\n )\n\n q_expenses_total = Q(category__user=request.user, category__is_income=False)\n q_expenses_monthly = Q(\n category__user=request.user,\n category__is_income=False,\n date__month=today.month,\n )\n q_expenses_today = Q(\n category__user=request.user, category__is_income=False, date=today\n )\n\n aggregated_balance = user_balances.aggregate(\n income_total=Coalesce(Sum(\"amount\", filter=q_incomes_total), 0),\n income_monthly=Coalesce(Sum(\"amount\", filter=q_incomes_monthly), 0),\n income_today=Coalesce(Sum(\"amount\", filter=q_incomes_today), 0),\n expenses_total=Coalesce(Sum(\"amount\", filter=q_expenses_total), 0),\n expenses_monthly=Coalesce(Sum(\"amount\", filter=q_expenses_monthly), 0),\n expenses_today=Coalesce(Sum(\"amount\", filter=q_expenses_today), 0),\n )\n\n return Response(status=status.HTTP_200_OK, data=aggregated_balance)\n\n @action(methods=[\"get\"], detail=False)\n def annual_balance(self, request, *args, **kwargs):\n today = datetime.date.today()\n balance_list = Balance.objects.filter(category__user=request.user)\n incomes_sum_qs = (\n balance_list.filter(category__is_income=True)\n .order_by(\"date__month\")\n .values(\"date__month\")\n .annotate(Sum(\"amount\"))\n )\n\n expenses_sum_qs = (\n balance_list.filter(category__is_income=False)\n .order_by(\"date__month\")\n .values(\"date__month\")\n .annotate(Sum(\"amount\"))\n )\n\n incomes_dict = {\n balance[\"date__month\"]: balance[\"amount__sum\"] for balance in incomes_sum_qs\n }\n expenses_dict = {\n balance[\"date__month\"]: balance[\"amount__sum\"]\n for balance in expenses_sum_qs\n }\n\n annual_balance = defaultdict(list)\n for i in range(11, -1, -1):\n current_date = today - relativedelta(months=i)\n annual_balance[\"months\"].append(calendar.month_abbr[current_date.month])\n annual_balance[\"incomes\"].append(incomes_dict.get(current_date.month, 0))\n annual_balance[\"expenses\"].append(expenses_dict.get(current_date.month, 0))\n\n return Response(status=status.HTTP_200_OK, data=annual_balance)\n", "sub_path": "backend/src/balance/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 5000, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "rest_framework.viewsets.ModelViewSet", "line_number": 22, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 22, "usage_type": "name"}, {"api_name": "balance.serializers.CategorySerializer", "line_number": 23, "usage_type": "name"}, {"api_name": "balance.models.Category.objects.filter", "line_number": 26, "usage_type": "call"}, {"api_name": "balance.models.Category.objects", "line_number": 26, "usage_type": "attribute"}, {"api_name": "balance.models.Category", "line_number": 26, "usage_type": "name"}, {"api_name": "balance.models.Category.objects.filter", "line_number": 29, "usage_type": "call"}, {"api_name": "balance.models.Category.objects", "line_number": 29, "usage_type": "attribute"}, {"api_name": "balance.models.Category", "line_number": 29, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 33, "usage_type": "call"}, {"api_name": "balance.serializers.CategorySimplySerializer", "line_number": 33, "usage_type": "call"}, {"api_name": "balance.models.Category.objects.create", "line_number": 39, "usage_type": "call"}, {"api_name": "balance.models.Category.objects", "line_number": 39, "usage_type": "attribute"}, {"api_name": "balance.models.Category", "line_number": 39, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 41, "usage_type": "call"}, {"api_name": "balance.serializers.CategoryBalanceSerializer", "line_number": 42, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 42, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 42, "usage_type": "name"}, {"api_name": "balance.models.Category.objects.filter", "line_number": 47, "usage_type": "call"}, {"api_name": "balance.models.Category.objects", "line_number": 47, "usage_type": "attribute"}, {"api_name": "balance.models.Category", "line_number": 47, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 51, "usage_type": "call"}, {"api_name": "balance.serializers.CategoryBalanceSerializer", "line_number": 51, "usage_type": "call"}, {"api_name": "rest_framework.decorators.action", "line_number": 45, "usage_type": "call"}, {"api_name": "rest_framework.viewsets.ModelViewSet", "line_number": 54, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 54, "usage_type": "name"}, {"api_name": "balance.serializers.SimplyBalanceSerializer", "line_number": 55, "usage_type": "name"}, {"api_name": "balance.models.Balance.objects.filter", "line_number": 58, "usage_type": "call"}, {"api_name": "balance.models.Balance.objects", "line_number": 58, "usage_type": "attribute"}, {"api_name": "balance.models.Balance", "line_number": 58, "usage_type": "name"}, {"api_name": "balance.serializers.BalanceSerializer", "line_number": 61, "usage_type": "name"}, {"api_name": "balance.models.Balance.objects.filter", "line_number": 66, "usage_type": "call"}, {"api_name": "balance.models.Balance.objects", "line_number": 66, "usage_type": "attribute"}, {"api_name": "balance.models.Balance", "line_number": 66, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 67, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 67, "usage_type": "attribute"}, {"api_name": "django.db.models.Q", "line_number": 69, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 70, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 75, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 79, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 80, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 85, "usage_type": "call"}, {"api_name": "django.db.models.functions.Coalesce", "line_number": 90, "usage_type": "call"}, {"api_name": "django.db.models.Sum", "line_number": 90, "usage_type": "call"}, {"api_name": "django.db.models.functions.Coalesce", "line_number": 91, "usage_type": "call"}, {"api_name": "django.db.models.Sum", "line_number": 91, "usage_type": "call"}, {"api_name": "django.db.models.functions.Coalesce", "line_number": 92, "usage_type": "call"}, {"api_name": "django.db.models.Sum", "line_number": 92, "usage_type": "call"}, {"api_name": "django.db.models.functions.Coalesce", "line_number": 93, "usage_type": "call"}, {"api_name": "django.db.models.Sum", "line_number": 93, "usage_type": "call"}, {"api_name": "django.db.models.functions.Coalesce", "line_number": 94, "usage_type": "call"}, {"api_name": "django.db.models.Sum", "line_number": 94, "usage_type": "call"}, {"api_name": "django.db.models.functions.Coalesce", "line_number": 95, "usage_type": "call"}, {"api_name": "django.db.models.Sum", "line_number": 95, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 98, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 98, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 98, "usage_type": "name"}, {"api_name": "rest_framework.decorators.action", "line_number": 64, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 102, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 102, "usage_type": "attribute"}, {"api_name": "balance.models.Balance.objects.filter", "line_number": 103, "usage_type": "call"}, {"api_name": "balance.models.Balance.objects", "line_number": 103, "usage_type": "attribute"}, {"api_name": "balance.models.Balance", "line_number": 103, "usage_type": "name"}, {"api_name": "django.db.models.Sum", "line_number": 108, "usage_type": "call"}, {"api_name": "django.db.models.Sum", "line_number": 115, "usage_type": "call"}, {"api_name": "balance.models", "line_number": 119, "usage_type": "name"}, {"api_name": "balance.models", "line_number": 122, "usage_type": "name"}, {"api_name": "balance.models", "line_number": 123, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 126, "usage_type": "call"}, {"api_name": "dateutil.relativedelta.relativedelta", "line_number": 128, "usage_type": "call"}, {"api_name": "calendar.month_abbr", "line_number": 129, "usage_type": "attribute"}, {"api_name": "rest_framework.response.Response", "line_number": 133, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 133, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 133, "usage_type": "name"}, {"api_name": "rest_framework.decorators.action", "line_number": 100, "usage_type": "call"}]} +{"seq_id": "649382009", "text": "\"\"\"\nFor debug & testing\n\"\"\"\nimport pandas as pd\n\nfrom utils import grammar_processing as gp\nfrom utils.first_follow import compute_firsts, compute_follows\nfrom utils.grammar_cleaner import (GrammarPipeline, remove_ambiguity,\n remove_epsilon, remove_left_recursion,\n remove_unit, remove_unreachable,\n remove_vars_nothing)\nfrom utils.tokenizer import tokenize\n\nG = gp.load_grammar()[1]\n\n# print(tokenize(G, 'int + int * ( int * int + int )'))\n\n# print('Before:')\n# print(G)\n\n# GrammarPipeline(G, [\n# remove_epsilon,\n# remove_unit,\n# remove_vars_nothing,\n# remove_unreachable,\n# remove_left_recursion,\n# remove_ambiguity,\n# ]).run()\n\n# print('After:')\n# print(G)\n\n# Testing automatons\n\nimport numpy as np\nimport pydot\n\nfrom utils.DFA import DFA\nfrom utils.NFA import NFA\n\n# automaton = NFA(states=3, finals=[1], transitions={\n# (0, 'a'): [1],\n# (1, 'a'): [2],\n# (2, 'b'): [1],\n# (1, 'b'): [0]\n# })\nautomaton = NFA(states=6, finals=[3, 5], transitions={\n (0, ''): [ 1, 2 ],\n (1, ''): [ 3 ],\n (1,'b'): [ 4 ],\n (2,'a'): [ 4 ],\n (3,'c'): [ 3 ],\n (4, ''): [ 5 ],\n (5,'d'): [ 5 ]\n})\n\nautomaton.graph().write_png('automaton.png')\n\ndfa = automaton.to_dfa()\n\ndfa.graph().write_png('to_dfa.png')\nprint(automaton.finals)\n\nnew = NFA.extend_automaton(automaton)\n\nnew.graph().write_png('extended.png')\n\nadj_list = {}\nr_adj_list = {}\nnew._build_adj_list(adj_list, r_adj_list)\n\nprint('adj_list:', adj_list)\nprint('r_adj_list:', r_adj_list)\n\nregex = new.to_regex()\nprint(regex)\n\n# testing the regex obtained\nimport re\n\ntests = [\n 'bdddddd',\n 'addddddddd',\n 'a',\n 'b',\n 'c',\n 'ccccc',\n '',\n]\nfor t in tests:\n print(re.fullmatch(regex, t))\n assert re.fullmatch(regex, t).end() == len(t)\n", "sub_path": "debug.py", "file_name": "debug.py", "file_ext": "py", "file_size_in_byte": 1855, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "utils.grammar_processing.load_grammar", "line_number": 14, "usage_type": "call"}, {"api_name": "utils.grammar_processing", "line_number": 14, "usage_type": "name"}, {"api_name": "utils.NFA.NFA", "line_number": 47, "usage_type": "call"}, {"api_name": "utils.NFA.NFA.extend_automaton", "line_number": 64, "usage_type": "call"}, {"api_name": "utils.NFA.NFA", "line_number": 64, "usage_type": "name"}, {"api_name": "re.fullmatch", "line_number": 91, "usage_type": "call"}, {"api_name": "re.fullmatch", "line_number": 92, "usage_type": "call"}]} +{"seq_id": "289062458", "text": "\"\"\"\nNon-purely generic tree data structure and functions manipulating it\n\"\"\"\nimport functools\nimport collections\nimport operator\n\n\nif False: # pragma: no cover\n from typing import ( # noqa: F401\n Any,\n Dict,\n List,\n Optional,\n Iterable,\n Tuple,\n )\n\n\nNode = collections.namedtuple('Node', ['key', 'value', 'children'])\n\n\ndef make_node(key, value=None, children=None):\n # type: (str, Any, Optional[Iterable[Tuple[str, Node]]]) -> Node\n return Node(\n key=key,\n value=value,\n children=collections.OrderedDict(children or []),\n )\n\n\ndef _list_children(node):\n # type: (Node) -> List[Node]\n return list(node.children.values())\n\n\ndef insert_by_path(node, path, value=None):\n # type: (Node, List[str], Any) -> Node\n \"\"\"Add descendants to a node\n\n :param node:\n descendants are added to this node\n :param path:\n path to a leaf node. This must not include above `node`.\n :param value:\n a value to be set as leaf node's value\n \"\"\"\n p, ps = path[0], path[1:]\n # An intermediate node of path\n if ps:\n if p in node.children:\n insert_by_path(\n node.children[p],\n ps,\n value,\n )\n return node\n else:\n node.children[p] = insert_by_path(\n make_node(p),\n ps,\n value,\n )\n return node\n # A leaf node of path\n else:\n if p in node.children:\n # recreate a node to replace a value\n node.children[p] = make_node(\n p,\n value=value,\n children=node.children[p].children.items(),\n )\n return node\n else:\n node.children[p] = make_node(p, value=value)\n return node\n\n\ndef to_paths(node):\n # type: (Node) -> List[Tuple[Any, List[str]]]\n \"\"\"Convert node into paths list with their values\n\n example input::\n\n root\n `- path\n +- to1 (has `value` `'v1'`)\n | `- leaf1\n `- to2\n\n example output::\n\n [\n (None, ['path']),\n ('v1', ['path', 'to1']),\n (None, ['path', 'to1', 'leaf1']),\n (None, ['path', 'to2']),\n ]\n\n :param node:\n Child nodes of this node are converted\n :return:\n list of pairs of node's value and path\n \"\"\"\n return functools.reduce(\n operator.add,\n (_to_paths(c) for c in _list_children(node)),\n [],\n )\n\n\ndef _to_paths(node):\n # type: (Node) -> List[Tuple[Any, List[str]]]\n \"\"\"Convert node into paths list\n\n :param node:\n a node which is coverted\n :return:\n list of pairs of node's value and path\n \"\"\"\n if node.children:\n return [(node.value, [node.key])] + [\n (v, [node.key] + path)\n for c in _list_children(node)\n for v, path in _to_paths(c)\n ]\n else:\n return [(node.value, [node.key])]\n\n\ndef draw(\n node,\n tail='`- ',\n middle='+- ',\n vertical='| ',\n whitespace=' ',\n):\n # type: (Node, str, str, str, str) -> List[Tuple[Any, str]]\n \"\"\"Draw children\n\n example::\n\n +- path1\n | +- to1\n | | `- leaf1\n | `- to2\n `- path2\n `- leaf2\n\n :param node:\n Child nodes of this node are drawn\n :param tail:\n shown as ``-` in example\n :param middle:\n shown as `+-` in example\n :param vertical:\n shown as `|` in example\n :param whitespace:\n leading whitespace for nodes in 2nd or deeper layers\n :return:\n list of pairs of node's value and drawing\n \"\"\"\n return _draw(\n _list_children(node),\n tail, middle, vertical, whitespace,\n )\n\n\ndef _draw(children, tail, middle, vertical, whitespace):\n # type: (List[Node], str, str, str, str) -> List[Tuple[Any, str]]\n \"\"\"Draw children\n\n :param children:\n Child nodes which is drawn\n :param tail:\n see: :func:`draw`\n :param middle:\n see: :func:`draw`\n :param vertical:\n see: :func:`draw`\n :param whitespace:\n see: :func:`draw`\n :return:\n see: :func:`draw`\n \"\"\"\n # A leaf's children:\n # `- path\n # `- to\n # `- leaf <- `children` = this node's children (= empty)\n if not children:\n return []\n\n # An intermediate node has one or more children\n else:\n c, cs = children[0], children[1:]\n # An intermediate node of nodes in the same layer:\n # `- path\n # +- to1 <- `c`\n # +- to2 <- one of `cs`\n # `- to3 <- one of `cs`\n if cs:\n return [(c.value, middle + c.key)] + [\n (v, vertical + l)\n for v, l in _draw( # type: ignore\n _list_children(c),\n tail, middle, vertical, whitespace,\n )\n ] + _draw(\n cs,\n tail, middle, vertical, whitespace,\n )\n # A tail node of nodes in the same layer:\n # `- path\n # +- to1 <- already accumulated\n # `- to2 <- `c`\n else:\n return [(c.value, tail + c.key)] + [\n (v, whitespace + l)\n for v, l in _draw( # type: ignore\n _list_children(c),\n tail, middle, vertical, whitespace,\n )\n ]\n", "sub_path": "me/hirai/snippets/data/tree.py", "file_name": "tree.py", "file_ext": "py", "file_size_in_byte": 5473, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "collections.namedtuple", "line_number": 20, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 28, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 106, "usage_type": "call"}, {"api_name": "operator.add", "line_number": 107, "usage_type": "attribute"}]} +{"seq_id": "623026911", "text": "#! coding: utf-8\n\nfrom django.conf import settings\nfrom django_datajsonar.models import \\\n Catalog, Dataset, Distribution, Field\n\nfrom series_tiempo_ar_api.apps.management import meta_keys\n\n\ndef setup_database():\n catalog = Catalog.objects.create(title='test_catalog', metadata='{}')\n dataset = Dataset.objects.create(identifier=\"132\",\n metadata='{}',\n catalog=catalog)\n init_year_series(dataset)\n init_semester_series(dataset)\n init_month_series(dataset)\n init_daily_series(dataset)\n\n\ndef init_year_series(dataset):\n distrib = Distribution.objects.create(identifier='132.1',\n metadata='{}',\n download_url=\"invalid_url\",\n dataset=dataset)\n distrib.save()\n distrib.enhanced_meta.create(key=meta_keys.PERIODICITY, value='R/P1Y')\n field = Field.objects.create(\n identifier=settings.TEST_SERIES_NAME.format('year'),\n metadata='{}',\n distribution=distrib,\n title='random_year_0_title'\n )\n field.save()\n field.enhanced_meta.create(key=meta_keys.AVAILABLE, value='True')\n\n\ndef init_semester_series(dataset):\n distrib = Distribution.objects.create(identifier='132.2',\n metadata='{}',\n download_url=\"invalid_url\",\n dataset=dataset)\n distrib.save()\n distrib.enhanced_meta.create(key=meta_keys.PERIODICITY, value='R/P6M')\n\n field = Field.objects.create(\n identifier=settings.TEST_SERIES_NAME.format('semester'),\n metadata='{}',\n distribution=distrib,\n title='random_semester_0_title'\n )\n field.save()\n field.enhanced_meta.create(key=meta_keys.AVAILABLE, value='True')\n\n\ndef init_month_series(dataset):\n\n distrib = Distribution.objects.create(identifier='132.3',\n metadata='{}',\n download_url=\"invalid_url\",\n dataset=dataset)\n distrib.save()\n distrib.enhanced_meta.create(key=meta_keys.PERIODICITY, value='R/P1M')\n field = Field.objects.create(\n identifier=settings.TEST_SERIES_NAME.format('month'),\n metadata='{\"description\": \"test_series_description\", \"units\": \"percentage\"}',\n distribution=distrib,\n title='random_month_0_title'\n )\n field.save()\n field.enhanced_meta.create(key=meta_keys.AVAILABLE, value='True')\n field.enhanced_meta.create(key=meta_keys.PERIODICITY, value='R/P1M')\n field.enhanced_meta.create(key=meta_keys.INDEX_START, value='1910-01-01')\n\n return dataset\n\n\ndef init_daily_series(dataset):\n distrib = Distribution.objects.create(identifier='132.4',\n metadata='{}',\n download_url=\"invalid_url\",\n dataset=dataset)\n distrib.save()\n distrib.enhanced_meta.create(key=meta_keys.PERIODICITY, value='R/P1D')\n\n field = Field.objects.create(\n identifier=settings.TEST_SERIES_NAME.format('day'),\n metadata='{\"description\": \"test_series_description\"}',\n distribution=distrib,\n title='random_day_0_title'\n )\n field.save()\n field.enhanced_meta.create(key=meta_keys.AVAILABLE, value='True')\n\n\ndef get_series_id(periodicity):\n return settings.TEST_SERIES_NAME.format(periodicity)\n", "sub_path": "series_tiempo_ar_api/apps/api/tests/helpers.py", "file_name": "helpers.py", "file_ext": "py", "file_size_in_byte": 3567, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "django_datajsonar.models.Catalog.objects.create", "line_number": 11, "usage_type": "call"}, {"api_name": "django_datajsonar.models.Catalog.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "django_datajsonar.models.Catalog", "line_number": 11, "usage_type": "name"}, {"api_name": "django_datajsonar.models.Dataset.objects.create", "line_number": 12, "usage_type": "call"}, {"api_name": "django_datajsonar.models.Dataset.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "django_datajsonar.models.Dataset", "line_number": 12, "usage_type": "name"}, {"api_name": "django_datajsonar.models.Distribution.objects.create", "line_number": 22, "usage_type": "call"}, {"api_name": "django_datajsonar.models.Distribution.objects", "line_number": 22, "usage_type": "attribute"}, {"api_name": "django_datajsonar.models.Distribution", "line_number": 22, "usage_type": "name"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys.PERIODICITY", "line_number": 27, "usage_type": "attribute"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys", "line_number": 27, "usage_type": "name"}, {"api_name": "django_datajsonar.models.Field.objects.create", "line_number": 28, "usage_type": "call"}, {"api_name": "django_datajsonar.models.Field.objects", "line_number": 28, "usage_type": "attribute"}, {"api_name": "django_datajsonar.models.Field", "line_number": 28, "usage_type": "name"}, {"api_name": "django.conf.settings.TEST_SERIES_NAME.format", "line_number": 29, "usage_type": "call"}, {"api_name": "django.conf.settings.TEST_SERIES_NAME", "line_number": 29, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 29, "usage_type": "name"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys.AVAILABLE", "line_number": 35, "usage_type": "attribute"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys", "line_number": 35, "usage_type": "name"}, {"api_name": "django_datajsonar.models.Distribution.objects.create", "line_number": 39, "usage_type": "call"}, {"api_name": "django_datajsonar.models.Distribution.objects", "line_number": 39, "usage_type": "attribute"}, {"api_name": "django_datajsonar.models.Distribution", "line_number": 39, "usage_type": "name"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys.PERIODICITY", "line_number": 44, "usage_type": "attribute"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys", "line_number": 44, "usage_type": "name"}, {"api_name": "django_datajsonar.models.Field.objects.create", "line_number": 46, "usage_type": "call"}, {"api_name": "django_datajsonar.models.Field.objects", "line_number": 46, "usage_type": "attribute"}, {"api_name": "django_datajsonar.models.Field", "line_number": 46, "usage_type": "name"}, {"api_name": "django.conf.settings.TEST_SERIES_NAME.format", "line_number": 47, "usage_type": "call"}, {"api_name": "django.conf.settings.TEST_SERIES_NAME", "line_number": 47, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 47, "usage_type": "name"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys.AVAILABLE", "line_number": 53, "usage_type": "attribute"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys", "line_number": 53, "usage_type": "name"}, {"api_name": "django_datajsonar.models.Distribution.objects.create", "line_number": 58, "usage_type": "call"}, {"api_name": "django_datajsonar.models.Distribution.objects", "line_number": 58, "usage_type": "attribute"}, {"api_name": "django_datajsonar.models.Distribution", "line_number": 58, "usage_type": "name"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys.PERIODICITY", "line_number": 63, "usage_type": "attribute"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys", "line_number": 63, "usage_type": "name"}, {"api_name": "django_datajsonar.models.Field.objects.create", "line_number": 64, "usage_type": "call"}, {"api_name": "django_datajsonar.models.Field.objects", "line_number": 64, "usage_type": "attribute"}, {"api_name": "django_datajsonar.models.Field", "line_number": 64, "usage_type": "name"}, {"api_name": "django.conf.settings.TEST_SERIES_NAME.format", "line_number": 65, "usage_type": "call"}, {"api_name": "django.conf.settings.TEST_SERIES_NAME", "line_number": 65, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 65, "usage_type": "name"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys.AVAILABLE", "line_number": 71, "usage_type": "attribute"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys", "line_number": 71, "usage_type": "name"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys.PERIODICITY", "line_number": 72, "usage_type": "attribute"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys", "line_number": 72, "usage_type": "name"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys.INDEX_START", "line_number": 73, "usage_type": "attribute"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys", "line_number": 73, "usage_type": "name"}, {"api_name": "django_datajsonar.models.Distribution.objects.create", "line_number": 79, "usage_type": "call"}, {"api_name": "django_datajsonar.models.Distribution.objects", "line_number": 79, "usage_type": "attribute"}, {"api_name": "django_datajsonar.models.Distribution", "line_number": 79, "usage_type": "name"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys.PERIODICITY", "line_number": 84, "usage_type": "attribute"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys", "line_number": 84, "usage_type": "name"}, {"api_name": "django_datajsonar.models.Field.objects.create", "line_number": 86, "usage_type": "call"}, {"api_name": "django_datajsonar.models.Field.objects", "line_number": 86, "usage_type": "attribute"}, {"api_name": "django_datajsonar.models.Field", "line_number": 86, "usage_type": "name"}, {"api_name": "django.conf.settings.TEST_SERIES_NAME.format", "line_number": 87, "usage_type": "call"}, {"api_name": "django.conf.settings.TEST_SERIES_NAME", "line_number": 87, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 87, "usage_type": "name"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys.AVAILABLE", "line_number": 93, "usage_type": "attribute"}, {"api_name": "series_tiempo_ar_api.apps.management.meta_keys", "line_number": 93, "usage_type": "name"}, {"api_name": "django.conf.settings.TEST_SERIES_NAME.format", "line_number": 97, "usage_type": "call"}, {"api_name": "django.conf.settings.TEST_SERIES_NAME", "line_number": 97, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 97, "usage_type": "name"}]} +{"seq_id": "494244289", "text": "#!/usr/bin/python3\n\"\"\"this is a test string\"\"\"\n\nfrom flask import request, jsonify, abort\nfrom api.v1.views import app_views\nfrom models import storage\nfrom models.place import Place\nfrom models.review import Review\nfrom models.user import User\n\n\n@app_views.route(\"/places//reviews\",\n strict_slashes=False,\n methods=[\"GET\", \"POST\"])\ndef reviews_base(p_id):\n \"\"\"this is a test string\"\"\"\n if request.method == \"GET\":\n out = []\n place = storage.get(Place, p_id)\n if place:\n for review in place.reviews:\n out.append(review.to_dict())\n return jsonify(out)\n abort(404)\n if request.method == \"POST\":\n if not request.is_json:\n return \"Not a JSON\", 400\n place = storage.get(Place, p_id)\n if place:\n kwargs = {\"place_id\": p_id}\n kwargs.update(request.get_json())\n out = Review(**kwargs)\n info = out.to_dict()\n if \"user_id\" not in info.keys():\n return \"Missing user_id\", 400\n if not storage.get(User, info.get(\"user_id\")):\n abort(404)\n if \"text\" not in info.keys():\n return \"Missing text\", 400\n out.save()\n return out.to_dict(), 201\n abort(404)\n\n\n@app_views.route(\"/reviews/\",\n strict_slashes=False,\n methods=[\"GET\", \"DELETE\", \"PUT\"])\ndef reviews_id(r_id):\n \"\"\"this is a test string\"\"\"\n if request.method == \"GET\":\n review = storage.get(Review, r_id)\n if review:\n return review.to_dict()\n abort(404)\n if request.method == \"DELETE\":\n review = storage.get(Review, r_id)\n if review:\n review.delete()\n storage.save()\n return {}, 200\n abort(404)\n if request.method == \"PUT\":\n review = storage.get(Review, r_id)\n if review:\n if not request.is_json:\n return \"Not a JSON\", 400\n for k, v in request.get_json().items():\n if k not in [\"id\", \"user_id\", \"place_id\",\n \"created_at\", \"updated_at\"]:\n setattr(review, k, v)\n storage.save()\n return review.to_dict(), 200\n abort(404)\n", "sub_path": "api/v1/views/places_reviews.py", "file_name": "places_reviews.py", "file_ext": "py", "file_size_in_byte": 2326, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "flask.request.method", "line_number": 17, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 17, "usage_type": "name"}, {"api_name": "models.storage.get", "line_number": 19, "usage_type": "call"}, {"api_name": "models.place.Place", "line_number": 19, "usage_type": "argument"}, {"api_name": "models.storage", "line_number": 19, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 23, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 24, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 25, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 25, "usage_type": "name"}, {"api_name": "flask.request.is_json", "line_number": 26, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 26, "usage_type": "name"}, {"api_name": "models.storage.get", "line_number": 28, "usage_type": "call"}, {"api_name": "models.place.Place", "line_number": 28, "usage_type": "argument"}, {"api_name": "models.storage", "line_number": 28, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 31, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 31, "usage_type": "name"}, {"api_name": "models.review.Review", "line_number": 32, "usage_type": "call"}, {"api_name": "models.storage.get", "line_number": 36, "usage_type": "call"}, {"api_name": "models.user.User", "line_number": 36, "usage_type": "argument"}, {"api_name": "models.storage", "line_number": 36, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 37, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 42, "usage_type": "call"}, {"api_name": "api.v1.views.app_views.route", "line_number": 12, "usage_type": "call"}, {"api_name": "api.v1.views.app_views", "line_number": 12, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 50, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 50, "usage_type": "name"}, {"api_name": "models.storage.get", "line_number": 51, "usage_type": "call"}, {"api_name": "models.review.Review", "line_number": 51, "usage_type": "argument"}, {"api_name": "models.storage", "line_number": 51, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 54, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 55, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 55, "usage_type": "name"}, {"api_name": "models.storage.get", "line_number": 56, "usage_type": "call"}, {"api_name": "models.review.Review", "line_number": 56, "usage_type": "argument"}, {"api_name": "models.storage", "line_number": 56, "usage_type": "name"}, {"api_name": "models.storage.save", "line_number": 59, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 59, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 61, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 62, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 62, "usage_type": "name"}, {"api_name": "models.storage.get", "line_number": 63, "usage_type": "call"}, {"api_name": "models.review.Review", "line_number": 63, "usage_type": "argument"}, {"api_name": "models.storage", "line_number": 63, "usage_type": "name"}, {"api_name": "flask.request.is_json", "line_number": 65, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 65, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 67, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 67, "usage_type": "name"}, {"api_name": "models.storage.save", "line_number": 71, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 71, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 73, "usage_type": "call"}, {"api_name": "api.v1.views.app_views.route", "line_number": 45, "usage_type": "call"}, {"api_name": "api.v1.views.app_views", "line_number": 45, "usage_type": "name"}]} +{"seq_id": "416091327", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Jul 12 15:55:22 2017\n\n@author: Boulot\n\"\"\"\nfrom SPARQLWrapper import SPARQLWrapper, JSON\n\n\ndef get_wikidata(value):\n sparql = SPARQLWrapper(\"https://query.wikidata.org/sparql\")\n sparql.setQuery(\"\"\"\n PREFIX skos: \n SELECT ?item ?label ?description ?score1 WHERE {\n ?item wdt:P17 wd:Q31.\n ?item rdfs:label ?label.\n ?item schema:description ?description.\n FILTER(EXISTS { ?item wdt:P625 ?x. })\n FILTER(CONTAINS(LCASE(STR(?label)), \"%s\"))\n FILTER(LANG(?label) = \"fr\").\n FILTER(LANG(?description) = \"fr\").\n}\nLIMIT 1 \"\"\" % value)\n\n sparql.setReturnFormat(JSON)\n results = sparql.query().convert()\n\n for result in results[\"results\"][\"bindings\"]:\n return result[\"label\"][\"value\"] + \"||\" + \\\n result[\"item\"][\"value\"] + \"||\" + result[\"description\"][\"value\"]\n\n\nprint(get_wikidata(\"cathédrale de tournai\"))\n", "sub_path": "functions/get_wikidata.py", "file_name": "get_wikidata.py", "file_ext": "py", "file_size_in_byte": 902, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "SPARQLWrapper.SPARQLWrapper", "line_number": 11, "usage_type": "call"}, {"api_name": "SPARQLWrapper.JSON", "line_number": 25, "usage_type": "argument"}]} +{"seq_id": "133378566", "text": "import xml.etree.ElementTree as ET\nfrom os import getcwd\n\nsets=['allData']\nclasses = [\"plum\"]\n\ndataPath = '/home/jasper/Datasets/Day'\ndataName = 'Day'\n# dataPath = '/home/jasper/Datasets/Night'\n# dataName = 'Night'\n\ndef count_annotation(dataPath, image_id):\n in_file = open(dataPath + '/Annotations/%s.xml'%(image_id))\n tree=ET.parse(in_file)\n root = tree.getroot()\n\n boxes = 0\n\n for obj in root.iter('object'):\n difficult = obj.find('difficult').text\n cls = obj.find('name').text\n if cls not in classes or int(difficult)==1:\n continue\n cls_id = classes.index(cls)\n xmlbox = obj.find('bndbox')\n b = (int(xmlbox.find('xmin').text), int(xmlbox.find('ymin').text), int(xmlbox.find('xmax').text), int(xmlbox.find('ymax').text))\n boxes = boxes + 1\n\n return boxes\n\nfor image_set in sets:\n count = 0\n image_ids = open(dataPath + '/ImageSets/Main/%s.txt'%(image_set)).read().strip().split()\n for image_id in image_ids:\n count = count + count_annotation(dataPath, image_id)\n\n print(\"For image set {} a total of {} bounding boxes were found from {} files\".format(image_set, count, len(image_ids)))\n", "sub_path": "count_voc_annotation.py", "file_name": "count_voc_annotation.py", "file_ext": "py", "file_size_in_byte": 1185, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "xml.etree.ElementTree.parse", "line_number": 14, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 14, "usage_type": "name"}]} +{"seq_id": "515150096", "text": "from flask import Blueprint, request, flash, redirect, render_template, send_from_directory, abort, send_file\nfrom werkzeug.utils import secure_filename\nfrom pathlib import Path\nimport openpyxl as oxl\nimport os\n\nbp = Blueprint('colclear', __name__)\n\nallowed_extensions = set(['txt', 'xlsx'])\n\n\n# local paths\npath = os.path.abspath(\"app/filesuploaded\")\npath1 = os.path.abspath(\"app/col_rm_final\")\n# for my pythonanywhere paths\n# path = \"/home/Flippy9004/excel_app/app/static/filesuploaded\"\n# path1 = \"/home/Flippy9004/excel_app/app/static/col_rm_final\"\n\n\ndef allowed_files(filename):\n return '.' in filename and filename.rsplit('.', 1)[1].lower() in allowed_extensions\n\n\n@bp.route('/clupload', methods=[\"GET\", \"POST\"])\ndef clupload():\n # Check if there is any \"final\" file in directory for final files\n # If there is, remove it\n dirs = os.listdir(path1)\n for file in dirs:\n os.remove(path1 + '/' + file)\n\n if request.method == \"POST\":\n\n if 'file' not in request.files:\n flash('No file part')\n return redirect(request.url)\n\n if 'file1' not in request.files:\n flash('No file part')\n return redirect(request.url)\n\n filse = request.files['file']\n filse1 = request.files['file1']\n\n files = [filse, filse1]\n for file in files:\n if file.filename == '':\n flash('No file selected')\n return redirect(request.url)\n\n for file in files:\n if file.filename.rsplit('.', 1)[1].lower() not in allowed_extensions:\n files.remove(file)\n\n if len(files) == 1:\n fext = files[0].filename.split('.', 1)[1]\n if fext == 'txt':\n return render_template(\"col_rm/upload.html\", numfiles=len(files), fext=True, wrong=True)\n else:\n return render_template(\"col_rm/upload.html\", numfiles=len(files), fext=False, wrong=True)\n elif len(files) == 2:\n fext = files[0].filename.split('.', 1)[1]\n fext1 = files[1].filename.split('.', 1)[1]\n if fext == fext1 and fext == 'txt':\n return render_template(\"col_rm/upload.html\", fext=True, wrong=True)\n elif fext == fext1 and fext == 'xlsx':\n return render_template(\"col_rm/upload.html\", fext=False, wrong=True)\n\n filenames = []\n for file in files:\n if file and allowed_files(file.filename):\n filename = secure_filename(file.filename)\n filenames.append(filename)\n file.save(os.path.join(Path(path), filename))\n\n return render_template(\"col_rm/upload.html\", filse=filenames, lfse=len(filenames), wrong=False)\n return render_template(\"col_rm/upload.html\")\n\n\n@bp.route('/removed')\ndef removed():\n dirs = os.listdir(path)\n # fileX = ['name', 'extension']\n file0 = [dirs[0].split('.', 1)[0], dirs[0].split('.', 1)[1]]\n file1 = [dirs[1].split('.', 1)[0], dirs[1].split('.', 1)[1]]\n if file0[1] == 'txt':\n txtfile = file0[0] + '.' + file0[1]\n xlsxfile = file1[0] + '.' + file1[1]\n else:\n txtfile = file1[0] + '.' + file1[1]\n xlsxfile = file0[0] + '.' + file0[1]\n filename = remove_columns(txtfile, xlsxfile)\n return render_template(\"col_rm/removed.html\", filse=filename)\n\n\n@bp.route(\"/download/\")\ndef download_file(filename):\n try:\n return send_from_directory(path1, filename=filename, as_attachment=True)\n except FileNotFoundError:\n abort(404)\n\n\ndef get_all_good_headers(txtfile):\n file = open(txtfile, 'r')\n dobre = file.readlines()\n good = []\n for i in dobre:\n good.append(i.strip())\n return good\n\n\ndef remove_columns(txtfile, xlsxfile):\n excel = path + '/' + xlsxfile\n txtfile = path + '/' + txtfile\n good = get_all_good_headers(txtfile)\n wbf = oxl.load_workbook(excel)\n sh0 = wbf[wbf.sheetnames[0]]\n\n number_of_columns = sh0.max_column\n\n i = 1\n while True:\n if i > number_of_columns:\n break\n cell = sh0.cell(row=1, column=i).value\n great = False\n for j in range(len(good)):\n header = good[j]\n great = True\n if header == cell:\n great = False\n i += 1\n break\n if great:\n sh0.delete_cols(i)\n number_of_columns -= 1\n finalname = xlsxfile.split(\".\", 1)\n finale = finalname[0] + \"_final.xlsx\"\n wbf.save(path1 + '/' + finale)\n os.remove(txtfile)\n os.remove(excel)\n return finale\n", "sub_path": "app/col_rm.py", "file_name": "col_rm.py", "file_ext": "py", "file_size_in_byte": 4582, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "flask.Blueprint", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 28, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 30, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 32, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 32, "usage_type": "name"}, {"api_name": "flask.request.files", "line_number": 34, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 34, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 35, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 36, "usage_type": "call"}, {"api_name": "flask.request.url", "line_number": 36, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 36, "usage_type": "name"}, {"api_name": "flask.request.files", "line_number": 38, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 38, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 39, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 40, "usage_type": "call"}, {"api_name": "flask.request.url", "line_number": 40, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 40, "usage_type": "name"}, {"api_name": "flask.request.files", "line_number": 42, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 42, "usage_type": "name"}, {"api_name": "flask.request.files", "line_number": 43, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 43, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 48, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 49, "usage_type": "call"}, {"api_name": "flask.request.url", "line_number": 49, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 49, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 58, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 60, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 65, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 67, "usage_type": "call"}, {"api_name": "werkzeug.utils.secure_filename", "line_number": 72, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path", "line_number": 74, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 74, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 76, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 77, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 82, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 93, "usage_type": "call"}, {"api_name": "flask.send_from_directory", "line_number": 99, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 101, "usage_type": "call"}, {"api_name": "openpyxl.load_workbook", "line_number": 117, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 141, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 142, "usage_type": "call"}]} +{"seq_id": "73700501", "text": "import json\n\n\ndef grow_cash(json_data):\n starting_amount = int(json_data[\"st_amt\"])\n growth_percentage = float(json_data[\"gw_pct\"])\n num_of_rotations = int(json_data[\"num_rota\"])\n return starting_amount * growth_percentage ** num_of_rotations\n\n\ndef growth_predictions(args):\n all_results = []\n for x in args:\n all_results.append(grow_cash(x))\n return all_results\n\n\ndef load_file(json_file):\n with open(json_file) as f:\n data = json.loads(f.read())\n f.close()\n return data\n", "sub_path": "cash_growth/cash_money.py", "file_name": "cash_money.py", "file_ext": "py", "file_size_in_byte": 524, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "json.loads", "line_number": 20, "usage_type": "call"}]} +{"seq_id": "144100380", "text": "import time\nfrom selenium import webdriver\nfrom selenium.webdriver.common.keys import Keys\nfrom selenium.webdriver.support import expected_conditions as EC\nfrom selenium.webdriver.support.wait import WebDriverWait\n\nwaiting_sec = 5\n\n\ndef getDriver(Debug=False):\n DRIVER_PATH = './chromedriver.exe'\n USER_AGENT = 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/60.0.3112.50 ' \\\n 'Safari/537.36 '\n options = webdriver.ChromeOptions()\n if not Debug:\n options.add_argument('headless')\n options.add_argument('window-size=1920x1080')\n options.add_argument('disable-gpu')\n options.add_argument('--start-maximized')\n options.add_argument('user-agent={0}'.format(USER_AGENT)) # for pretend not to be a bot\n options.add_argument('incognito') # for secret window\n chrome_driver = webdriver.Chrome(DRIVER_PATH, options=options)\n return chrome_driver\n\n\ndef click_element(element, driver, ctrl=False):\n if ctrl:\n element.send_keys(Keys.CONTROL + Keys.ENTER)\n else:\n driver.execute_script(\"arguments[0].click();\", element)\n\n\ndef wait(by, div, driver):\n time.sleep(.1)\n WebDriverWait(driver, waiting_sec).until(EC.presence_of_element_located((by, div)))\n\n\ndef click(by, div, driver, isWait=True, ctrl=False):\n if isWait:\n wait(by, div, driver)\n if ctrl:\n driver.find_element(by, div).send_keys(Keys.CONTROL + Keys.ENTER)\n else:\n driver.execute_script(\"arguments[0].click();\", driver.find_element(by, div))\n", "sub_path": "저학년 대회 결과물/러버덕/Rubber_Duck-master/chaeminuim.py", "file_name": "chaeminuim.py", "file_ext": "py", "file_size_in_byte": 1542, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "selenium.webdriver.ChromeOptions", "line_number": 14, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 14, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 22, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 22, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.keys.Keys.CONTROL", "line_number": 28, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.keys.Keys", "line_number": 28, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.keys.Keys.ENTER", "line_number": 28, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 34, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.wait.WebDriverWait", "line_number": 35, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions.presence_of_element_located", "line_number": 35, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.expected_conditions", "line_number": 35, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.keys.Keys.CONTROL", "line_number": 42, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.keys.Keys", "line_number": 42, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.keys.Keys.ENTER", "line_number": 42, "usage_type": "attribute"}]} +{"seq_id": "212337132", "text": "import sys, time, math, os, random\nfrom pyglet.gl import *\n\nwindow = pyglet.window.Window()\nkeyboard = pyglet.window.key.KeyStateHandler()\nwindow.push_handlers(keyboard)\n\n\nclass ball:\n def __init__(self, x, y, Xsize, Ysize):\n self.x=x\n self.y=y\n self.Xsize = Xsize\n self.Ysize = Ysize\n self.dy = 2\n self.dx = 2\n\n def draw(self):\n pyglet.graphics.draw(4, pyglet.gl.GL_QUADS, ('v2f', \n [self.x, self.y, self.x-self.Xsize, self.y, self.x-self.Xsize, self.y-self.Ysize, self.x, self.y-self.Ysize]))\n\n def update(self):\n global window\n if(self.x==window.width):\n self.dx=-2\n if(self.x==self.Xsize):\n self.dx=2\n if(self.y==window.height):\n self.dy=-2\n if(self.y==self.Ysize):\n self.dy=2\n self.y+=self.dy\n self.x+=self.dx\n\n\n@window.event\ndef on_draw():\n global ball\n glClearColor(0, 0.2, 0.2, 0)\n glClear(GL_COLOR_BUFFER_BIT)\n ball.draw()\n\ndef update(tm):\n global ball\n ball.update()\n \n@window.event\ndef on_key_press(symbol, modifiers):\n global ball\n ball.update()\nball=ball(100,200,50,50)\n\n\npyglet.clock.schedule_interval(update, 0.005)\npyglet.app.run()", "sub_path": "ball.py", "file_name": "ball.py", "file_ext": "py", "file_size_in_byte": 1236, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "pyglet.gl.window.Window", "line_number": 4, "usage_type": "call"}, {"api_name": "pyglet.gl.window", "line_number": 4, "usage_type": "attribute"}, {"api_name": "pyglet.gl", "line_number": 4, "usage_type": "name"}, {"api_name": "pyglet.gl.window.key.KeyStateHandler", "line_number": 5, "usage_type": "call"}, {"api_name": "pyglet.gl.window", "line_number": 5, "usage_type": "attribute"}, {"api_name": "pyglet.gl", "line_number": 5, "usage_type": "name"}, {"api_name": "pyglet.gl.graphics.draw", "line_number": 19, "usage_type": "call"}, {"api_name": "pyglet.gl.graphics", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pyglet.gl", "line_number": 19, "usage_type": "name"}, {"api_name": "pyglet.gl.gl", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pyglet.gl.clock.schedule_interval", "line_number": 54, "usage_type": "call"}, {"api_name": "pyglet.gl.clock", "line_number": 54, "usage_type": "attribute"}, {"api_name": "pyglet.gl", "line_number": 54, "usage_type": "name"}, {"api_name": "pyglet.gl.app.run", "line_number": 55, "usage_type": "call"}, {"api_name": "pyglet.gl.app", "line_number": 55, "usage_type": "attribute"}, {"api_name": "pyglet.gl", "line_number": 55, "usage_type": "name"}]} +{"seq_id": "423262527", "text": "from pathlib import Path\nimport dateutil.parser\nimport datetime\n\nimport xarray as xr\nimport numpy as np\n\nimport lagtraj.produce.lagrangian_trajectory\nimport lagtraj.produce.forcing_profiles\n\n\ndef test_create_dummy_forcing_from_stationary_trajectory():\n ds_traj = lagtraj.produce.lagrangian_trajectory.stationary.extract_trajectory(\n lat0=-10,\n lon0=40,\n t0=dateutil.parser.parse(\"2020-01-22T12:00\"),\n t_max=dateutil.parser.parse(\"2020-01-24T12:00\"),\n dt=datetime.timedelta(hours=4),\n )\n\n da_levels = xr.DataArray(\n np.array([100e3, 90e3, 80e3, 70e3]),\n attrs=dict(long_name=\"pressure levels\", units=\"Pa\"),\n )\n\n ds_forcing_profiles = lagtraj.produce.forcing_profiles.dummy.extract_forcing_profiles(\n ds_traj=ds_traj, required_variables=[\"ddt__qv\",], da_levels=da_levels\n )\n\n\n# testing command line interface\ndef test_main():\n fn_trajectory = \"trajectory_test.nc\"\n lagtraj.produce.lagrangian_trajectory.main(\n lat0=-10,\n lon0=40,\n t0=dateutil.parser.parse(\"2020-01-22T12:00\"),\n t_max=dateutil.parser.parse(\"2020-01-24T12:00\"),\n dt=datetime.timedelta(hours=4),\n trajectory_type=\"stationary\",\n out_filename=fn_trajectory,\n )\n\n fn_forcing_profiles = \"forcing_profiles.nc\"\n lagtraj.produce.forcing_profiles.main(\n fn_trajectory=fn_trajectory,\n source_data=\"dummy\",\n out_filename=fn_forcing_profiles,\n )\n\n p_traj = Path(fn_trajectory)\n p_forcing = Path(fn_forcing_profiles)\n\n assert p_forcing.exists()\n p_traj.unlink()\n p_forcing.unlink()\n", "sub_path": "tests/test_forcing_profiles_extraction.py", "file_name": "test_forcing_profiles_extraction.py", "file_ext": "py", "file_size_in_byte": 1609, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "lagtraj.produce.lagrangian_trajectory.produce.lagrangian_trajectory.stationary.extract_trajectory", "line_number": 13, "usage_type": "call"}, {"api_name": "lagtraj.produce.lagrangian_trajectory.produce", "line_number": 13, "usage_type": "attribute"}, {"api_name": "lagtraj.produce.lagrangian_trajectory", "line_number": 13, "usage_type": "name"}, {"api_name": "dateutil.parser.parser.parse", "line_number": 16, "usage_type": "call"}, {"api_name": "dateutil.parser.parser", "line_number": 16, "usage_type": "attribute"}, {"api_name": "dateutil.parser", "line_number": 16, "usage_type": "name"}, {"api_name": "dateutil.parser.parser.parse", "line_number": 17, "usage_type": "call"}, {"api_name": "dateutil.parser.parser", "line_number": 17, "usage_type": "attribute"}, {"api_name": "dateutil.parser", "line_number": 17, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 18, "usage_type": "call"}, {"api_name": "xarray.DataArray", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 22, "usage_type": "call"}, {"api_name": "lagtraj.produce.lagrangian_trajectory.produce.forcing_profiles.dummy.extract_forcing_profiles", "line_number": 26, "usage_type": "call"}, {"api_name": "lagtraj.produce.lagrangian_trajectory.produce", "line_number": 26, "usage_type": "attribute"}, {"api_name": "lagtraj.produce.lagrangian_trajectory", "line_number": 26, "usage_type": "name"}, {"api_name": "lagtraj.produce.lagrangian_trajectory.produce.lagrangian_trajectory.main", "line_number": 34, "usage_type": "call"}, {"api_name": "lagtraj.produce.lagrangian_trajectory.produce", "line_number": 34, "usage_type": "attribute"}, {"api_name": "lagtraj.produce.lagrangian_trajectory", "line_number": 34, "usage_type": "name"}, {"api_name": "dateutil.parser.parser.parse", "line_number": 37, "usage_type": "call"}, {"api_name": "dateutil.parser.parser", "line_number": 37, "usage_type": "attribute"}, {"api_name": "dateutil.parser", "line_number": 37, "usage_type": "name"}, {"api_name": "dateutil.parser.parser.parse", "line_number": 38, "usage_type": "call"}, {"api_name": "dateutil.parser.parser", "line_number": 38, "usage_type": "attribute"}, {"api_name": "dateutil.parser", "line_number": 38, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 39, "usage_type": "call"}, {"api_name": "lagtraj.produce.lagrangian_trajectory.produce.forcing_profiles.main", "line_number": 45, "usage_type": "call"}, {"api_name": "lagtraj.produce.lagrangian_trajectory.produce", "line_number": 45, "usage_type": "attribute"}, {"api_name": "lagtraj.produce.lagrangian_trajectory", "line_number": 45, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 51, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 52, "usage_type": "call"}]} +{"seq_id": "568851536", "text": "import re\r\nimport datetime\r\n\r\ndef printTimeStamp(name):\r\n print('Автор програми: ' + name)\r\n print('Час компіля��ії: ' + str(datetime.datetime.now()))\r\nprintTimeStamp('Пікула. Погорілий')\r\n\r\nwhile True:\r\n try:\r\n numlet = input('Позиція: ')\r\n list1 = re.findall(r'\\D+', numlet)\r\n list2 = re.findall(r'\\d+', numlet)\r\n let = ''.join(list1)\r\n num = ''.join(list2)\r\n num = int(num)\r\n except ValueError:\r\n print('Некоректне значення. Спробуйте ще раз.')\r\n continue\r\n if num < 1 and num > 8:\r\n print('Некоректне значення. Спробуйте ще раз.')\r\n continue\r\n elif let == \"a\" or let == \"b\" or let == \"c\" or let == \"d\" or let == \"e\" or let == \"f\" or let == \"g\" or let == \"h\":\r\n break\r\n else:\r\n print('Некоректне значення. Спробуйте ще раз.')\r\n continue\r\nif num % 2 == 0:\r\n if let == \"a\" or let == \"c\" or let == \"e\" or let == \"g\":\r\n print('Біла')\r\n elif let == \"b\" or let == \"d\" or let == \"f\" or let == \"h\":\r\n print('Чорна')\r\nelse:\r\n if let == \"a\" or let == \"c\" or let == \"e\" or let == \"g\":\r\n print('Чорна')\r\n elif let == \"b\" or let == \"d\" or let == \"f\" or let == \"h\":\r\n print('Біла')\r\n", "sub_path": "моя практика/1 день/B/20б.py", "file_name": "20б.py", "file_ext": "py", "file_size_in_byte": 1385, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "datetime.datetime.now", "line_number": 6, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 6, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 12, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 13, "usage_type": "call"}]} +{"seq_id": "152925761", "text": "from konlpy.tag import Okt\nokt = Okt()\n\nfile = open('cover_Letter.txt', 'r', encoding='UTF-8')\n\ntexts = file.readlines()\nfile.close()\n\nfile = open('noun_Letter.txt', 'w', encoding='UTF-8')\n\ndoc_list = []\nall_text = \"\"\ntext_list = []\ni = 1\n\nfor text in texts:\n\twords = \"\"\n\tdoc_list = okt.nouns(text)\n\tfor word in doc_list:\n\t\tif (len(word) > 1):\n\t\t\ttext_list.append(word)\n\t\t\twords += word+ ' '\n\tfile.write(words+\" \\n\")\nfile.close()\n\n", "sub_path": "make_model/letter_to_noun.py", "file_name": "letter_to_noun.py", "file_ext": "py", "file_size_in_byte": 431, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "konlpy.tag.Okt", "line_number": 2, "usage_type": "call"}]} +{"seq_id": "456207398", "text": "\"\"\"SQLAlchemy column generation.\"\"\"\n\nimport typing\n\nimport sqlalchemy\n\nfrom ... import exceptions\nfrom ... import helpers\nfrom ... import types\n\n# Remapping SQLAlchemy classes\nColumn: sqlalchemy.Column = sqlalchemy.Column\nType: sqlalchemy.sql.type_api.TypeEngine = sqlalchemy.sql.type_api.TypeEngine\nForeignKey: sqlalchemy.ForeignKey = sqlalchemy.ForeignKey\nInteger: sqlalchemy.Integer = sqlalchemy.Integer\nBigInteger: sqlalchemy.BigInteger = sqlalchemy.BigInteger\nNumber: sqlalchemy.Float = sqlalchemy.Float\nString: sqlalchemy.String = sqlalchemy.String\nBinary: sqlalchemy.LargeBinary = sqlalchemy.LargeBinary\nDate: sqlalchemy.Date = sqlalchemy.Date\nDateTime: sqlalchemy.DateTime = sqlalchemy.DateTime\nBoolean: sqlalchemy.Boolean = sqlalchemy.Boolean\n\n\ndef construct(*, artifacts: types.ColumnArtifacts) -> Column:\n \"\"\"\n Construct column from artifacts.\n\n Args:\n artifacts: The artifacts of the column.\n\n Returns:\n The SQLAlchemy column.\n\n \"\"\"\n type_ = _determine_type(artifacts=artifacts)\n foreign_key: typing.Optional[ForeignKey] = None\n if artifacts.extension.foreign_key is not None:\n foreign_key_kwargs: types.TKwargs = {}\n if artifacts.extension.foreign_key_kwargs is not None:\n foreign_key_kwargs = artifacts.extension.foreign_key_kwargs\n foreign_key = ForeignKey(artifacts.extension.foreign_key, **foreign_key_kwargs)\n # Map default value\n default = helpers.oa_to_py_type.convert(\n value=artifacts.open_api.default,\n type_=artifacts.open_api.type,\n format_=artifacts.open_api.format,\n )\n # Generate kwargs\n kwargs: types.TKwargs = {}\n if artifacts.extension.kwargs is not None:\n kwargs = artifacts.extension.kwargs\n return Column(\n type_,\n foreign_key,\n nullable=artifacts.open_api.nullable,\n default=default,\n primary_key=artifacts.extension.primary_key,\n autoincrement=artifacts.extension.autoincrement,\n index=artifacts.extension.index,\n unique=artifacts.extension.unique,\n **kwargs,\n )\n\n\ndef _determine_type(*, artifacts: types.ColumnArtifacts) -> Type:\n \"\"\"\n Determine the type for a specification.\n\n Raise FeatureNotImplementedError for unsupported types.\n\n Args:\n artifacts: The artifacts for the column.\n\n Returns:\n The type for the column.\n\n \"\"\"\n # Determining the type\n type_: typing.Optional[Type] = None\n if artifacts.open_api.type == \"integer\":\n type_ = _handle_integer(artifacts=artifacts)\n elif artifacts.open_api.type == \"number\":\n type_ = _handle_number(artifacts=artifacts)\n elif artifacts.open_api.type == \"string\":\n type_ = _handle_string(artifacts=artifacts)\n elif artifacts.open_api.type == \"boolean\":\n type_ = _handle_boolean(artifacts=artifacts)\n\n if type_ is None:\n raise exceptions.FeatureNotImplementedError(\n f\"{artifacts.open_api.type} has not been implemented\"\n )\n\n return type_\n\n\ndef _handle_integer(\n *, artifacts: types.ColumnArtifacts\n) -> typing.Union[Integer, BigInteger]:\n \"\"\"\n Handle artifacts for an integer type.\n\n Raise FeatureNotImplementedError if a format that is not supported is defined.\n\n Args:\n artifacts: The artifacts for the column.\n\n Returns:\n The SQLAlchemy integer type of the column.\n\n \"\"\"\n if artifacts.open_api.format is None or artifacts.open_api.format == \"int32\":\n return Integer\n if artifacts.open_api.format == \"int64\":\n return BigInteger\n raise exceptions.FeatureNotImplementedError(\n f\"{artifacts.open_api.format} format for integer is not supported.\"\n )\n\n\ndef _handle_number(*, artifacts: types.ColumnArtifacts) -> Number:\n \"\"\"\n Handle artifacts for an number type.\n\n Raise FeatureNotImplementedError if a format that is not supported is defined.\n\n Args:\n artifacts: The artifacts for the column.\n\n Returns:\n The SQLAlchemy number type of the column.\n\n \"\"\"\n if artifacts.open_api.format is None or artifacts.open_api.format == \"float\":\n return Number\n raise exceptions.FeatureNotImplementedError(\n f\"{artifacts.open_api.format} format for number is not supported.\"\n )\n\n\ndef _handle_string(\n *, artifacts: types.ColumnArtifacts\n) -> typing.Union[String, Binary, Date, DateTime]:\n \"\"\"\n Handle artifacts for an string type.\n\n Raise FeatureNotImplementedError if a format that is not supported is defined.\n\n Args:\n artifacts: The artifacts for the column.\n\n Returns:\n The SQLAlchemy string type of the column.\n\n \"\"\"\n if artifacts.open_api.format in {None, \"byte\", \"password\"}:\n if artifacts.open_api.max_length is None:\n return String\n return String(length=artifacts.open_api.max_length)\n if artifacts.open_api.format == \"binary\":\n if artifacts.open_api.max_length is None:\n return Binary\n return Binary(length=artifacts.open_api.max_length)\n if artifacts.open_api.format == \"date\":\n return Date\n if artifacts.open_api.format == \"date-time\":\n return DateTime\n raise exceptions.FeatureNotImplementedError(\n f\"{artifacts.open_api.format} format for string is not supported.\"\n )\n\n\ndef _handle_boolean(\n *, artifacts: types.ColumnArtifacts # pylint: disable=unused-argument\n) -> Boolean:\n \"\"\"\n Handle artifacts for an boolean type.\n\n Args:\n artifacts: The artifacts for the column.\n\n Returns:\n The SQLAlchemy boolean type of the column.\n\n \"\"\"\n return Boolean\n", "sub_path": "open_alchemy/facades/sqlalchemy/column.py", "file_name": "column.py", "file_ext": "py", "file_size_in_byte": 5602, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "sqlalchemy.Column", "line_number": 12, "usage_type": "attribute"}, {"api_name": "sqlalchemy.sql", "line_number": 13, "usage_type": "attribute"}, {"api_name": "sqlalchemy.ForeignKey", "line_number": 14, "usage_type": "attribute"}, {"api_name": "sqlalchemy.Integer", "line_number": 15, "usage_type": "attribute"}, {"api_name": "sqlalchemy.BigInteger", "line_number": 16, "usage_type": "attribute"}, {"api_name": "sqlalchemy.Float", "line_number": 17, "usage_type": "attribute"}, {"api_name": "sqlalchemy.String", "line_number": 18, "usage_type": "attribute"}, {"api_name": "sqlalchemy.LargeBinary", "line_number": 19, "usage_type": "attribute"}, {"api_name": "sqlalchemy.Date", "line_number": 20, "usage_type": "attribute"}, {"api_name": "sqlalchemy.DateTime", "line_number": 21, "usage_type": "attribute"}, {"api_name": "sqlalchemy.Boolean", "line_number": 22, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 37, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 80, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 100, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 144, "usage_type": "attribute"}]} +{"seq_id": "314392964", "text": "'''class defination'''\nimport torch\nimport torch.nn as nn\nfrom torch.utils.data import Dataset, DataLoader\nimport numpy as np\n\nimport math\nimport json\n#for preprocessing\nimport re\nfrom nltk.tokenize import word_tokenize\nfrom nltk.tokenize import RegexpTokenizer\n\n\nclass char_lstm(nn.Module):\n '''forward output: output sequence of the last bilstm layer\n size = batch*longest_seq*(2hidden_size)'''\n def __init__(self,char_vocab_size,char_embed=300,hidden = 150,char_layer = 2):\n super().__init__()\n self.input_size = char_embed\n self.hidden_size = hidden\n self.layers = char_layer\n self.char_embeddings = nn.Embedding(char_vocab_size,self.input_size,padding_idx=0)\n self.bilstm = nn.LSTM(self.input_size,self.hidden_size,self.layers,bidirectional=True,dropout=0)\n\n def forward(self,input_word,char2index,init_hidden):\n '''input word of the form of string'''\n char_ids,batch_lens = self.word2char_id(input_word,char2index)\n #(batch_size,seq_len,1) -> (batch_size,(longest)seq_len,embedding_dim)\n char_ids = torch.from_numpy(char_ids).long()\n char_embeds = self.char_embeddings(char_ids)\n '''print(input_word)\n print(\"padded input\")\n print(char_ids)\n print(char_ids.size())\n print(\"embedding\")\n print(char_embeds)\n print(char_embeds.size())'''\n #packed_padded_sequence,input of size BxTx*\n X = torch.nn.utils.rnn.pack_padded_sequence(char_embeds,batch_lens,batch_first = True)\n '''print(\"packed input\")\n print(X.data)\n print(type(X.data))\n print(X.data.size())'''\n hidden,cell = self.bilstm(X,init_hidden)\n '''print(\"hidden output\")\n print(hidden.data)\n print(hidden.data.size())'''\n #unpack output of lstm\n hidden, _ = torch.nn.utils.rnn.pad_packed_sequence(hidden, batch_first=True)\n print(\"unpacked hidden output\")\n print(hidden.data) \n print(hidden.data.size())\n return hidden,cell\n \n def init_hidden(self,batch_size):\n start_hidden = (torch.randn(self.layers*2, batch_size, self.hidden_size),\n torch.randn(self.layers*2, batch_size, self.hidden_size)) #h0,c0\n return start_hidden\n \n def word2char_id(self,word,char2index):\n '''convert a string of word to a collection of chars;\n and map to char indices; remove unknow chars in a word\n use white space as the start/end marker of word'''\n batch_chars = [list(i) for i in word] # assume word=[seq_len*batch_size]\n char_ids = [[char2index[i] for i in chars if i in char2index] for chars in batch_chars] \n padded_ids, batch_lens = self.padding(char_ids)\n return padded_ids,batch_lens\n \n def padding(self,indices):\n '''input: int lists of variable length sequence minibatch\n e.g. x= [[0,1,2,3,4,5],[7,7],[6,8]]\n =batch_size*seq_len\n Also, assume the padding token always has an index of 0'''\n #sort the batch in the descending order of length\n indices = sorted(indices,key=lambda seq:len(seq),reverse=True)\n batch_size = len(indices)\n lengths = [len(seq) for seq in indices]\n longest_seq = max(lengths)\n pad_token = int(0)\n padded_indices = np.ones((batch_size,longest_seq),dtype = int)*pad_token\n \n for i,len_i in enumerate(lengths):\n seq = indices[i]\n padded_indices[i,0:len_i] = seq[0:len_i]\n \n return padded_indices,lengths\n \n \nclass json_data(Dataset):\n '''data generator'''\n def __init__(self,file,num_sent): \n self.num_sent = num_sent\n self.file = file\n return\n \n def __len__(self):\n return self.num_sent\n \n def __getitem__(self,index):\n '''generate one sample of data'''\n with open(self.file,'r') as f:\n sent = json.load(f,object_hook =lambda d:\n self.json_hooker(d,index))\n \n sent,target = self.preprocess(sent) #tokenization + start/end marker \n return sent,target\n \n def json_hooker(self,dict,i):\n '''helper func for __get__item: return the ith sentence in the json record'''\n if 'text' in dict:\n return dict['text'][i]\n \n def preprocess(self,sent):\n '''remove symbols,and underscores(non-alpha)\n and tokenize\n keep digits'''\n reg = re.compile('([^\\s\\w]|_)+')\n sent = word_tokenize(re.sub(reg,'',sent.lower()))\n target = sent[1:]\n target.append('>')\n return [sent,target]\n \n#params = {'batch_size':64,'shuffle':True,'num_worker':6}\n\nmax_epoch = 100 \n\ntotal = 310 #built_in_intents\npartition = [0.75,0.15,0.1]\nnum_train = math.ceil(total*partition[0])\nnum_val = math.ceil(total*partition[1])\nnum_test = math.ceil(total*partition[2])\n\n\nfile = 'built_in_intents.json'\ntrain_set = json_data(file,total)\ntrain_loader = DataLoader(train_set, batch_size=1, num_workers=4, shuffle=True)\ntorch.manual_seed(1)\n\nif __name__ == '__main__':\n #run()\n for batch_idx, data in enumerate(train_loader):\n if batch_idx == 1:\n print('batch: {}\\tdata: {}'.format(batch_idx, data))\n #padded_batch = pad_sent(data)\n data5 = data\n print(data5[0])\n print(data5[1])\n print(type(data5[0]))\n\n \ndef pad_sent():\n pass\n", "sub_path": "Embeddings/class_def.py", "file_name": "class_def.py", "file_ext": "py", "file_size_in_byte": 5483, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "torch.nn.Module", "line_number": 15, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.nn.Embedding", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 23, "usage_type": "name"}, {"api_name": "torch.nn.LSTM", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 24, "usage_type": "name"}, {"api_name": "torch.from_numpy", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.nn.utils.rnn.pack_padded_sequence", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 40, "usage_type": "attribute"}, {"api_name": "torch.nn.utils.rnn.pad_packed_sequence", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 50, "usage_type": "attribute"}, {"api_name": "torch.randn", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 81, "usage_type": "call"}, {"api_name": "torch.utils.data.Dataset", "line_number": 90, "usage_type": "name"}, {"api_name": "json.load", "line_number": 103, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 118, "usage_type": "call"}, {"api_name": "nltk.tokenize.word_tokenize", "line_number": 119, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 119, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 130, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 131, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 132, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 137, "usage_type": "call"}, {"api_name": "torch.manual_seed", "line_number": 138, "usage_type": "call"}]} +{"seq_id": "151945035", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Tue Oct 18 14:47:58 2016\r\n\r\n@author: HP\r\n\"\"\"\r\n\r\nimport pandas as pd\r\nimport numpy as np\r\nimport csv\r\nfrom decimal import *\r\nimport seaborn as sns\r\nimport matplotlib.pyplot as plt\r\nfrom sklearn.preprocessing import LabelEncoder\r\nfrom sklearn.ensemble import RandomForestRegressor\r\n\r\n\r\ntrain_df = pd.read_csv('Housing1.csv')\r\ntest_df = pd.read_csv('Housing2.csv')\r\ntrain_df.head()\r\nprint(train_df.describe())\r\nprint(train_df.dtypes)\r\n\r\n#Drop columns deemed irrelevant\r\ntrain_df = train_df.drop(['Id','LotFrontage','Street', 'Alley','Utilities', 'LotConfig','LandSlope','Condition2','RoofMatl',\r\n 'ExterCond','BsmtFinSF2','Heating','LowQualFinSF','3SsnPorch',\r\n 'ScreenPorch','PoolArea','PoolQC','MiscFeature','MiscVal',\r\n 'MoSold','YrSold','SaleType',], axis=1)\r\n\r\n\r\ntest_df = test_df.drop(['LotFrontage', 'Street', 'Alley','Utilities','LotConfig','LandSlope','Condition2','RoofMatl',\r\n 'ExterCond','BsmtFinSF2','Heating','LowQualFinSF','3SsnPorch',\r\n 'ScreenPorch','PoolArea','PoolQC','MiscFeature','MiscVal',\r\n 'MoSold','YrSold','SaleType'], axis=1)\r\n\r\n\r\n#Get a sense of the median price of a house\r\nprint(\"The median housing price:\", np.median(train_df['SalePrice']))\r\n\r\n#\r\ncategorical_fields=['MSZoning','LotShape','LandContour','Neighborhood','Condition1','BldgType','HouseStyle',\r\n 'RoofStyle','Exterior1st','Exterior2nd','MasVnrType','ExterQual','Foundation','BsmtQual','BsmtCond','BsmtExposure','BsmtFinType1',\r\n 'BsmtFinType2','HeatingQC','CentralAir','Electrical','KitchenQual','Functional','FireplaceQu','GarageType','GarageFinish','GarageQual',\r\n 'GarageCond','PavedDrive','Fence','SaleCondition']\r\n \r\nnumerical_fields=['MSSubClass','LotArea','OverallQual','OverallCond','YearBuilt','YearRemodAdd','RoofStyle','Exterior1st','Exterior2nd','MasVnrType','MasVnrArea','ExterQual',\t\r\n 'Foundation','BsmtQual','BsmtCond','BsmtExposure','BsmtFinType1','BsmtFinSF1',\t'BsmtFinType2','BsmtUnfSF','TotalBsmtSF','HeatingQC','CentralAir','Electrical',\r\n '1stFlrSF','2ndFlrSF','GrLivArea','BsmtFullBath','BsmtHalfBath','FullBath','HalfBath','BedroomAbvGr','KitchenAbvGr','KitchenQual',\t'TotRmsAbvGrd','Functional','Fireplaces','FireplaceQu',\r\n 'GarageType','GarageYrBlt',\t'GarageFinish','GarageCars',\t'GarageArea','GarageQual','GarageCond','PavedDrive','WoodDeckSF','OpenPorchSF','EnclosedPorch','Fence']\r\n #find the null variables\r\nprint(train_df.apply(lambda x: sum(x.isnull())))\r\nprint(test_df.apply(lambda x: sum(x.isnull()))) \r\n\r\n#fill the null variables \r\ndef pre_process_data():\r\n print('Preprocessing the Data...')\r\nfor col in categorical_fields:\r\n train_df[col].fillna('default',inplace=True)\r\n test_df[col].fillna('default', inplace=True) \r\n \r\nfor col in numerical_fields:\r\n train_df[col].fillna(0,inplace=True)\r\n test_df[col].fillna(0, inplace=True)\r\n \r\n#check to make sure null variables are filled in\r\nprint(train_df.apply(lambda x: sum(x.isnull())))\r\nprint(test_df.apply(lambda x: sum(x.isnull()))) \r\n\r\n#Label encode\r\nencode=LabelEncoder()\r\n \r\nfor col in categorical_fields:\r\n train_df[col]=encode.fit_transform(train_df[col])\r\n test_df[col]=encode.fit_transform(test_df[col])\r\n\r\n#drop SalesPrice and Id\r\nX_train = train_df.drop('SalePrice', axis=1)\r\nY_train = train_df['SalePrice']\r\nX_test = test_df.drop('Id',axis=1).copy()\r\n\r\n#fit to run RandomForestReressor\r\nrandom_forest = RandomForestRegressor(n_estimators=100)\r\nrandom_forest.fit(X_train, Y_train)\r\nY_pred = random_forest.predict(X_test)\r\nprint(random_forest.score(X_train, Y_train))\r\n\r\nrf = RandomForestRegressor()\r\nrf.fit(X_train, Y_train)", "sub_path": "Housing prediction.py", "file_name": "Housing prediction.py", "file_ext": "py", "file_size_in_byte": 3905, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pandas.read_csv", "line_number": 18, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 38, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 70, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestRegressor", "line_number": 82, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestRegressor", "line_number": 87, "usage_type": "call"}]} +{"seq_id": "581921958", "text": "from pygame import Rect\nfrom m_engine.scene import *\nfrom m_engine.control import *\nfrom m_engine.gameobject import *\n\nfrom m_engine import eng_math\n\n\n\"\"\"\nPhysics section :DDD\n\nPhysx is a class that handles calculations that are physics related\n-May grow in the future\n\"\"\"\nclass Collision():\n def __init__(self,A,B,nature):\n self.A = A\n self.B = B\n self.nature = nature\n\nclass Physx():\n def mapRect(x):\n return [x[i]-1 if i < 2 else x[i]+2 for i in range(len(x))]\n \n def rectCollideNature(rectA,rectB):\n aLeft,aTop,aRight,aBottom = rectA\n aBottom+=aTop\n aRight+=aLeft\n bLeft,bTop,bRight,bBottom = rectB\n bBottom+=bTop\n bRight+=bLeft\n possible = [[abs(aTop-bBottom),\"TOP\"],[abs(aBottom-bTop),\"BOTTOM\"],[abs(aLeft-bRight),\"LEFT\"],[abs(aRight-bLeft),\"RIGHT\"]]\n possible.sort()\n if possible[0][0] == possible[1][0]:\n return \"CORNER\"\n return possible[0][1]\n \n def __init__(self,space):\n self.space = space\n space.bindPhysx(self)\n self.rbcToResolve = []\n\n def checkCollisions(self,rigidBodyControl,static=False):\n rect = rigidBodyControl.getRect()\n if static:\n rect = Physx.mapRect(rect)\n collisions = []\n for obj in self.space.query(rect):\n c = obj.getControl(RigidBodyControl)\n if not c == None and c.getRect().colliderect(rect) and not c.id == rigidBodyControl.id:\n collisions.append(c)\n\n return collisions\n\n #objects must be registered into the physics space before getting\n #affected by it; objects must have a rigidbodycontrol to do this\n def registerObject(self,gameObject):\n if not isinstance(gameObject,GameObject):\n print (\"Fatal: Trying to add \"+repr(gameObject)+\". Required: GameObject!\")\n return\n rbc = gameObject.getControl(RigidBodyControl)\n if rbc == None:\n print (\"Fatal: GameObject must have a rigidBodyControl!\")\n return\n rbc.bindSpace(self)\n \n def updatePhysx(self):\n '''\n after all of the move operations are done,\n we must resolve what happened to determine order obj's,etc.\n '''\n\n for rbc in self.rbcToResolve:\n collisions = self.checkCollisions(rbc,True)\n friction = eng_math.Vect2D()\n\n x,xx,y,yy = True,True,True,True\n for c in collisions:\n if c.solid and rbc.solid:\n nature = Physx.rectCollideNature(Physx.mapRect(rbc.getRect()),c.getRect())\n if nature == \"LEFT\":\n x = False\n friction.components[\"y\"]+=c.friction\n if nature == \"RIGHT\":\n xx = False\n friction.components[\"y\"]+=c.friction\n if nature == \"TOP\":\n y = False\n friction.components[\"x\"]+=c.friction\n if nature == \"BOTTOM\":\n yy = False\n friction.components[\"x\"]+=c.friction\n\n for force in rbc.forces.values():\n dx,dy = True,True\n if not x and force.vect[\"x\"] < 0 or not xx and force.vect[\"x\"] > 0:\n dx = False\n rbc.normal.components[\"y\"]+= abs(force.vect[\"x\"])\n elif not y and force.vect[\"y\"] < 0 or not yy and force.vect[\"y\"] > 0:\n dy = False\n rbc.normal.components[\"x\"]+= abs(force.vect[\"y\"])\n\n force.apply(dx,dy)\n\n rbc.normal.components[\"x\"]*= friction[\"x\"] #friction in the x and y axis\n rbc.normal.components[\"y\"]*= friction[\"y\"]\n\n V = rbc.velocity\n normal = rbc.normal\n\n if V.components[\"x\"] < 0:\n V.components[\"x\"] = min(V[\"x\"]+normal[\"x\"],0)\n elif V.components[\"x\"] > 0:\n V.components[\"x\"] = max(V[\"x\"]-normal[\"x\"],0)\n if V.components[\"y\"] < 0:\n V.components[\"y\"] = min(V[\"y\"]+normal[\"y\"],0)\n if V.components[\"y\"] > 0:\n V.components[\"y\"] = max(V[\"y\"]-normal[\"y\"],0)\n\n priority = eng_math.PriorityQueue()\n \n while self.rbcToResolve:\n priority.push(self.rbcToResolve.pop())\n\n order = []\n while priority.hasNext():\n order.append(priority.pop())\n\n speedThreshold = abs(order[0].velocity)\n\n for i in range(int(speedThreshold)):\n for j in range(len(order)-1,-1,-1):\n old = order[j].bound.location.copy()\n V = order[j].velocity\n delta = V/speedThreshold\n order[j].bound.location+=delta\n collisions = self.checkCollisions(order[j])\n for c in collisions:\n nature = Physx.rectCollideNature(order[j].getRect(),c.getRect())\n order[j].alertCollision(Collision(order[j],c,nature))\n if c.solid and order[j].solid:\n if nature == \"LEFT\" or nature == \"RIGHT\":\n order[j].bound.location.components[\"x\"] = old[\"x\"]\n V.components[\"x\"] *= c.elasticity\n if abs(V.components[\"x\"]) < 1:\n V.components[\"x\"] = 0\n elif nature == \"TOP\" or nature == \"BOTTOM\":\n order[j].bound.location.components[\"y\"] = old[\"y\"]\n V.components[\"y\"] *= c.elasticity\n if abs(V.components[\"y\"]) < 1:\n V.components[\"y\"] = 0\n else:\n order[j].bound.location.components[\"x\"] = old[\"x\"]-delta[\"x\"]\n order[j].bound.location.components[\"y\"] = old[\"y\"]-delta[\"y\"]\n #clear normals\n for x in order:\n x.normal-=x.normal\n\nclass Force():\n \"\"\"\n I do call this forces, but its a bit of a misnomer.\n It's more like something that causes acceleration per loop.\n Ex: gravity\n \"\"\"\n ids = 0\n def __init__(self,vect2D):\n self.id = Force.ids\n Force.ids+=1\n self.vect = vect2D\n self.bound = None\n\n def bind(self,rigidBodyControl):\n self.bound = rigidBodyControl\n\n def unbind(self):\n self.bound = None\n\n def apply(self,x,y):\n vect = self.vect.copy()\n if not x:\n vect.components[\"x\"] = 0\n if not y:\n vect.components[\"y\"] = 0\n self.bound.velocity+=vect\n\nclass RigidBodyControl(Control):\n\n def __init__(self,width,height,gravity=0,solid=True,elasticity=0):\n #if solid is false, this object will pass through other objects\n #however, listeners will still function\n #Elasticity is a measure of the percent of rebound that occurs\n Control.__init__(self)\n self.velocity = eng_math.Vect2D()\n self.normal = eng_math.Vect2D()\n self.forces = {}\n self.listeners = {}\n self.dimensions = [width,height]\n self.solid = solid\n self.elasticity = -elasticity\n self.friction = 0\n self.mass = width*height/100/2000\n if gravity > 0:\n self.addForce(Gravity(gravity))\n #this is the resultant velocity that will be parsed\n #by physx to determine wtf happened to the objects\n self.space = None\n\n #should be called when added into the game\n def bindSpace(self,space):\n self.space = space\n \n def __repr__(self):\n return \"RigidBodyControl id:\"+str(self.id);\n def __str__(self):\n return \"RigidBodyControl id:\"+str(self.id);\n def setFriction(self,value):\n self.friction = value\n\n #comparator operators\n def __eq__(self,rbc):\n if isinstance(rbc,RigidBodyControl):\n return abs(self.velocity)==abs(rbc.velocity)\n def __lt__(self,rbc):\n if isinstance(rbc,RigidBodyControl):\n return abs(self.velocity)>abs(rbc.velocity)\n def __gt__(self,rbc):\n if isinstance(rbc,RigidBodyControl):\n return abs(self.velocity) 90:\n conclusion = \"SUCCESS\"\n else:\n conclusion = \"FAILED\"\n try:\n copy(test_picture, os.path.join(self.app.report_dir, \"%s_%s.bmp\" % (self.app.testscript, testcase)))\n cv2.imwrite(os.path.join(self.app.report_dir, \"%s_%s_dif.jpg\" % (self.app.testscript, testcase)), dif)\n except:\n return\n self.app.write_log(\"[GRAB] Check the comparison result: %s\" % conclusion) # logger - check the comparison result: SUCCESS/FAILED\n print(\"[GRAB] Check the comparison result: %s\" % conclusion) # print - check the comparison result: SUCCESS/FAILED\n self.report_to_xlsx(testcase=testcase,\n col=self.app.testintex,\n report=self.app.report_file,\n result=conclusion)\n self.app.write_log(\"[EXCEL] Fill the xlsx report\") # logger - fill the xlsx report\n print(\"[EXCEL] Fill the xlsx report\") # print - fill the xlsx report\n\n\nclass Capture:\n \"\"\"\n Class for creating Capture obj:\n Capturing frames from Video Capture card and transmit them to the handlers\n \"\"\"\n def __init__(self, app):\n self.app = app\n self._thread = KThread(target=self._loop)\n self._handlers = []\n self.cap = cv2.VideoCapture(0)\n self.cap.set(3, 720)\n self.cap.set(4, 576)\n\n def start(self):\n self._thread.start()\n self.app.write_log(\"[GRAB] Launch Grabber. Start capturing\") # logger - Launch Grabber. Start capturing\n print(\"[GRAB] Launch Grabber. Start capturing\") # print - Launch Grabber. Start capturing\n\n def stop(self):\n self._thread.kill()\n self.app.write_log(\"[GRAB] Turn off Grabber. Stop capturing\") # logger - Turn off Grabber. Stop capturing\n print(\"[GRAB] Turn off Grabber. Stop capturing\") # print - Turn off Grabber. Stop capturing\n\n def _loop(self):\n \"\"\"\n Reading frame and transmit it to all handlers\n \"\"\"\n while 1:\n ret, frame = self.cap.read()\n for handler in self._handlers:\n handler(ret, frame)\n\n def register_handler(self, handler):\n \"\"\"\n Add handlers\n \"\"\"\n self._handlers.append(handler)\n\n def close_session(self):\n \"\"\"\n Finalizer of Capture obj\n \"\"\"\n self._handlers = []\n self.cap.release()\n\n\n", "sub_path": "Devices/grabber.py", "file_name": "grabber.py", "file_ext": "py", "file_size_in_byte": 6535, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "threading.Thread", "line_number": 16, "usage_type": "attribute"}, {"api_name": "threading.Thread.__init__", "line_number": 20, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 20, "usage_type": "attribute"}, {"api_name": "threading.Thread.start", "line_number": 27, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 27, "usage_type": "attribute"}, {"api_name": "sys.settrace", "line_number": 32, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 56, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 57, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 66, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 67, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 80, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 81, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path", "line_number": 90, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path", "line_number": 92, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 102, "usage_type": "call"}, {"api_name": "openpyxl.load_workbook", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 122, "usage_type": "call"}, {"api_name": "os.path", "line_number": 122, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path", "line_number": 123, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 124, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path", "line_number": 140, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 141, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 141, "usage_type": "call"}, {"api_name": "os.path", "line_number": 141, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 163, "usage_type": "call"}]} +{"seq_id": "299565628", "text": "import discord\nimport time\nfrom replit import db\nfrom datetime import datetime\nfrom pytz import timezone\n\nasync def utilityCheck(message, command, args, client, p, defaultColor, fullArgs):\n\n\tif(command == 'unmute'):\n\t\tif(message.author.guild_permissions.manage_roles == False): return await message.channel.send(\"You need the `MANAGE_ROLES` permission in order to unmute other users in this guild.\")\n\t\tif(len(message.mentions) < 1): return await message.channel.send('Please mention someone to unmute them.')\n\t\trole = discord.utils.get(message.guild.roles, name=\"muted-donut\")\n\t\tperson = message.mentions[0]\n\t\tif role == None: return await message.channel.send(f\"The muted role does not exist in the server and **{person.name}** is not Muted.\")\n\t\tif role not in person.roles: return await message.channel.send(f\"**{person.name}** is not Muted.\")\n\t\tawait person.remove_roles(role)\n\t\tawait message.channel.send(f\"You have unmuted **{person.name}**.\")\n\tif(command == 'mute'):\n\t\tif(message.author.guild_permissions.manage_roles == False): return await message.channel.send(\"You need the `MANAGE_ROLES` permission in order to mute other users in this guild.\")\n\t\tif(len(message.mentions) < 1): return await message.channel.send('Please mention someone to mute them.')\n\t\ttoMute = message.mentions[0]\n\t\ttoMuteMember = message.guild.get_member(toMute.id)\n\t\tif(toMuteMember.guild_permissions.administrator == True): return await message.channel.send(\"This person has the `ADMINISTRATOR` permission meaning they cannot be muted.\")\n\t\trole = discord.utils.get(message.guild.roles, name=\"muted-donut\")\n\t\tif(role == None):\n\t\t\tawait message.channel.send(\"The `muted-donut` role wasn't found. Generating a new one...\")\n\t\t\trole = await message.guild.create_role(name='muted-donut')\n\t\t\tfor channel in message.guild.channels:\n\t\t\t\tawait channel.set_permissions(role, speak=False, send_messages=False)\n\n\t\tawait toMute.add_roles(role)\n\t\tawait message.channel.send(f\"You have muted **{toMute.name}**.\")\n\n\tif(command == 'clear'):\n if message.author.guild_permissions.manage_messages == False: return await message.channel.send(f\"You need the `MANAGE_MESSAGES` permission in order to clear messages.\")\n if len(fullArgs) < 1:\n await message.channel.send('Please provide a number of messages to delete after the command!')\n else: \n amount = str(args[0])\n digit = amount.isdigit()\n print(fullArgs)\n\t\t \n if(digit == True):\n if int(amount) < 100 and int(amount) > 0:\n deleted = await message.channel.purge(limit = int(amount) + 1)\n done = await message.channel.send(f'Sucessfully deleted **{len(deleted)}** messages in {message.channel}!')\n time.sleep(1)\n await done.delete()\n else: \n await message.channel.send('Please provide a number between 1 and 100!')\n else:\n await message.channel.send('Please provide a valid number!')\n\tif command == 'poll' or command == 'vote':\n\t\tif(len(fullArgs) < 1): return await message.channel.send(f'Please either provide a yes or no poll or a multiple choice poll. Examples: \\n Yes or No: `{p}poll How are you?`\\n Multiple Choice: `{p}poll How are you?: Good - Excited - Bored`')\n\t\tsubDash = ''\n\t\tsubColon = ':'\n\t\tif(subDash in fullArgs and subColon in fullArgs):\n\t\t\tquestion = fullArgs.split(':')[0]\n\t\t\tmsgArgs = fullArgs.split(':')\n\t\t\tdel msgArgs[:1]\n\t\t\tstrOptions = ' '.join(msgArgs)\n\t\t\toptions = strOptions.split('-')\n\t\t\toptionsDesc = \"\"\n\t\t\ti = -1\n\t\t\tif(len(options) > 20): return await message.channel.send('The max amount of reactions on a Discord message is 20! I cannot put more than that. Please provide 20 options max.')\n\t\t\tif(len(options) < 2): return await message.channel.send('Please provide at least 2 options.')\n\t\t\treactions = ['🇦', '🇧', '🇨', '🇩', '🇪', '🇫', '🇬', '🇭', '🇮', '🇯', '🇰', '🇱', '🇲', '🇳', '🇴', '🇵', '🇶', '🇷', '🇸', '🇹', '🇺', '🇻', '🇼', '🇽', '🇾', '🇿']\n\t\t\tfor element in options: i+=1;optionsDesc += f\"{reactions[i]} = {element}\\n\\n\"\n\t\t\te = discord.Embed(color=defaultColor, title=question, description=optionsDesc)\n\t\t\te.set_author(name=message.author, icon_url=message.author.avatar_url)\n\t\t\tmsg = await message.channel.send(embed=e)\n\t\t\ti = -1\n\t\t\tfor element in options: i+=1;await msg.add_reaction(f\"{reactions[i]}\")\n\t\telse:\n\t\t\tquestion = fullArgs\n\t\t\te = discord.Embed(color=defaultColor, title=question)\n\t\t\te.set_author(name=message.author, icon_url=message.author.avatar_url)\n\t\t\tmsg = await message.channel.send(embed=e)\n\t\t\tawait msg.add_reaction('✅')\n\t\t\tawait msg.add_reaction('❌')\n\n\tif command == 'clear-warns':\n\t\tif(message.author.guild_permissions.manage_roles == False): return await message.channel.send(\"You need the `MANAGE_ROLES` permission in order to manage people's warns.\")\n\t\tif(len(message.mentions) < 1): return await message.channel.send(\"Please provide a user to remove their warns.\")\n\t\tperson = message.mentions[0]\n\t\tdb[f\"warns_{message.guild.id}_{person.id}\"] = []\n\t\tawait message.channel.send(f\"You have cleared **{person.name}**'s warns.\")\n\tif command == 'warn':\n\t\tif(message.author.guild_permissions.manage_roles == False): return await message.channel.send(\"You need the `MANAGE_ROLES` permission in order to warn people.\")\n\t\tif(len(message.mentions) < 1): return await message.channel.send(\"Please provide a user to warn after the command.\")\n\t\tperson = message.mentions[0]\n\t\tif(len(args) < 2):\n\t\t\twarnReason = 'No Reason'\n\t\telse:\n\t\t\tmsgArgs = args\n\t\t\tdel msgArgs[:1]\n\t\t\twarnReason = \" \".join(msgArgs)\n\t\ttry:\n\t\t\twarns = db[f\"warns_{message.guild.id}_{person.id}\"]\n\t\texcept:\n\t\t\twarns = []\n\t\test = timezone('EST')\n\t\tdate = datetime.now(est).strftime('%l:%M%p %Z on %b %d, %Y')\n\t\tfinalDate = f\"{date}\"\n\t\twarns.append({\"date\": date, \"mod\": str(message.author), \"reason\": warnReason})\n\t\tdb[f\"warns_{message.guild.id}_{person.id}\"] = warns\n\t\tawait message.channel.send(f\"You have warned **{person.name}** for **{warnReason}** on **{date}**.\")\n\tif command == 'warns':\n\t\tif(len(message.mentions) < 1): \n\t\t\tperson = message.author\n\t\telse:\n\t\t\tperson = message.mentions[0]\n\n\t\ttry:\n\t\t\twarns = db[f\"warns_{message.guild.id}_{person.id}\"]\n\t\texcept:\n\t\t\twarns = []\n\t\te = discord.Embed(color=defaultColor, title=f'__{person.name}\\'s Warns in {message.guild.name}__')\n\t\tfor element in warns:\n\t\t\te.add_field(name=f\"{element['date']}\", value=f\"Warned By: {element['mod']}\\nReason for Warn: {element['reason']}\", inline=False)\n\t\tif(len(warns) < 1): e.add_field(name='None', value=f\"{person.name} has no warns in {message.guild.name}!\")\n\t\tawait message.channel.send(embed=e)\n\tif command == 'kick':\n if message.author.guild_permissions.kick_members == False: \n await message.channel.send('You do not have the kick members permission!')\n else:\n if len(message.mentions) < 1:\n await message.channel.send('Who do you want to kick?')\n else:\n user = message.mentions[0]\n member = message.guild.get_member(user.id)\n if len(args) < 2:\n kickReason = 'None'\n else:\n msgArgs = args\n del msgArgs[:1]\n space = ' '\n kickReason = space.join(msgArgs)\n try:\n await member.kick(reason=f'{message.author}: {kickReason}')\n embed=discord.Embed(title=f'Action: Kicked {member}', color=defaultColor)\n embed.add_field(name='Kicked User ID:', value=member.id)\n embed.add_field(name='Reason:', value=kickReason)\n embed.add_field(name='Kicked By:', value=message.author)\n await message.channel.send(embed=embed)\n except:\n await message.channel.send(\"I don't have enough permissions to do that!\")\n\n\tif command == 'ban':\n if message.author.guild_permissions.ban_members == False: \n await message.channel.send('You do not have the ban members permission!')\n else:\n if len(message.mentions) < 1:\n await message.channel.send('Who do you want to ban?')\n else:\n user = message.mentions[0]\n member = message.guild.get_member(user.id)\n if len(args) < 2:\n banReason = 'None'\n else:\n msgArgs = args\n del msgArgs[:1]\n space = ' '\n banReason = space.join(msgArgs)\n try:\n await member.ban(reason=f'{message.author}: {banReason}')\n embed=discord.Embed(title=f'Action: Banned {member}', color=defaultColor)\n embed.add_field(name='Banned User ID:', value=member.id)\n embed.add_field(name='Reason:', value=banReason)\n embed.add_field(name='Banned By:', value=message.author)\n await message.channel.send(embed=embed)\n except:\n await message.channel.send(\"I don't have enough permissions to do that!\")", "sub_path": "utility.py", "file_name": "utility.py", "file_ext": "py", "file_size_in_byte": 10304, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "discord.utils.get", "line_number": 12, "usage_type": "call"}, {"api_name": "discord.utils", "line_number": 12, "usage_type": "attribute"}, {"api_name": "discord.utils.get", "line_number": 24, "usage_type": "call"}, {"api_name": "discord.utils", "line_number": 24, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 47, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 69, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 76, "usage_type": "call"}, {"api_name": "replit.db", "line_number": 86, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 99, "usage_type": "name"}, {"api_name": "pytz.timezone", "line_number": 102, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 103, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 103, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 106, "usage_type": "name"}, {"api_name": "replit.db", "line_number": 115, "usage_type": "name"}, {"api_name": "discord.Embed", "line_number": 118, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 141, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 167, "usage_type": "call"}]} +{"seq_id": "495631434", "text": "import cv2\nimport numpy as np\nfrom scipy import stats\ndef image_convert(filename):\n file = cv2.imread(filename)\n file = np.array(file[:,:,0]).flatten()\n file = (255-file)/100\n return file\n\n\ndef cos(a,b):\n norm1 = np.sqrt(np.sum(a**2))\n norm2 = np.sqrt(np.sum(b**2))\n return np.dot(a,b)/(norm1*norm2)\n\n\ndef strnum_make(i):\n if i < 10:\n strnum = \"00\"+str(i)\n elif i < 100:\n strnum = \"0\"+str(i)\n else:\n strnum = str(i)\n return strnum\n\n\nsamples_length = 482\npluses = np.empty((0,784))\nsubs = np.empty((0,784))\nmultis = np.empty((0,784))\ndivs = np.empty((0,784))\n\nfor i in range(0,samples_length):\n print(i)\n img_plus = image_convert(\"plus/plus_\"+strnum_make(i)+\".bmp\")\n img_sub = image_convert(\"sub/sub_\"+strnum_make(i)+\".bmp\")\n img_multi = image_convert(\"multi/multi_\"+strnum_make(i)+\".bmp\")\n img_div = image_convert(\"div/div_\"+strnum_make(i)+\".bmp\")\n\n pluses = np.append(pluses,np.array([img_plus]),axis=0)\n subs = np.append(subs,np.array([img_sub]),axis=0)\n multis = np.append(multis,np.array([img_multi]),axis=0)\n divs = np.append(divs,np.array([img_div]),axis=0)\n\nall_X = np.append(pluses,subs,axis=0)\nall_X = np.append(all_X,multis,axis=0)\nall_X = np.append(all_X,divs,axis=0)\nall_y = np.append(np.ones(samples_length)*0,np.ones(samples_length)*1)\nall_y = np.append(all_y,np.ones(samples_length)*2)\nall_y = np.append(all_y,np.ones(samples_length)*3)\n\n\ntest_x = cv2.imread(\"test_009.bmp\")\ntest_x = np.array(test_x[:,:,0]).flatten()\ntest_x = np.abs(test_x-255)/255.0\n\n\ndistance = [cos(test_x,all_X[i]) for i in range(0,samples_length*4)]\nprint(distance)\nprint(all_y[np.argsort(distance)][-10::])\nmode = stats.mode(all_y[np.argsort(distance)][-3:])\nmode = int(mode[0])\n\nif mode == 0:\n print(\"これは +\")\nelif mode == 1:\n print(\"これは -\")\nelif mode == 2:\n print(\"これは *\")\nelif mode == 3:\n print(\"これは /\")\n\n\n\n", "sub_path": "test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 1923, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "cv2.imread", "line_number": 5, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 50, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 60, "usage_type": "call"}, {"api_name": "scipy.stats.mode", "line_number": 61, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 61, "usage_type": "name"}, {"api_name": "numpy.argsort", "line_number": 61, "usage_type": "call"}]} +{"seq_id": "289491733", "text": "import threading\nimport logging\nlogging.basicConfig(format='%(asctime)s - %(message)s', datefmt='%d-%b-%y %H:%M:%S')\n\n\nclass Node:\n def __init__(self, key, val):\n # Doubly link list.\n self.key = key\n self.val = val\n self.next = None\n self.prev = None\n\n\nclass LRUCache:\n \"\"\"\n 1. Uses DL to remove node in O(1)\n 2. get for accessing key\n 3. put for inserting a key\n\n \"\"\"\n def __init__(self, capacity):\n if capacity < 1:\n print('LRUCache capacity must be > 0')\n return None\n self.capacity = capacity\n self.size = 0\n self.node_map = {}\n self.lock = threading.Lock()\n self.head = None\n self.tail = None\n\n def use_node(self, node):\n \"\"\"\n Util function for updating node order\n \"\"\"\n if node is self.head:\n return\n if node.next:\n node.next.prev = node.prev\n if node.prev:\n node.prev.next = node.next\n if node is self.tail:\n self.tail = self.tail.prev\n self.head.prev = node\n node.next = self.head\n node.prev = None\n self.head = node\n\n def get(self, key):\n with self.lock:\n if key in self.node_map:\n self.use_node(self.node_map[key])\n logging.info(\"Key {} found with value {}\".format(key, self.node_map[key].val))\n return self.node_map[key].val\n else:\n logging.error(\"Key not found\")\n raise KeyError(\"Key not found\")\n\n def put(self, key, value=1):\n with self.lock:\n if key in self.node_map:\n self.use_node(self.node_map[key])\n self.node_map[key].val = value\n else:\n node = Node(key, value)\n self.node_map[key] = node\n if self.size == 0:\n self.head = node\n self.tail = node\n if self.size < self.capacity:\n self.size += 1\n elif self.size == self.capacity:\n k = self.tail.key\n if self.size == 1:\n self.head = node\n self.tail = node\n else:\n self.tail = self.tail.prev\n self.tail.next = None\n del self.node_map[k]\n self.use_node(node)\n\n def items(self):\n print(self.head.key)\n print(self.tail.key)\n print(self.size)\n print(self.node_map.keys())\n\n\nif __name__ == '__main__':\n print(\"Please test\")\n lru = LRUCache(5)\n lru.put(1, 1)\n lru.put(2, 1)\n lru.put(3, 1)\n lru.put(4, 1)\n lru.put(5, 1)\n lru.put(6, 1)\n lru.put(7, 1)\n print(lru.get(9))\n print(lru.get(3))\n lru.items()\n\n", "sub_path": "rivetlabs/lru.py", "file_name": "lru.py", "file_ext": "py", "file_size_in_byte": 2860, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "logging.basicConfig", "line_number": 3, "usage_type": "call"}, {"api_name": "threading.Lock", "line_number": 29, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 54, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 57, "usage_type": "call"}]} +{"seq_id": "559546535", "text": "import sys\nsys.path.append(\"../\")\n\nimport schedule\nimport time\nimport argparse\nimport json\nimport logging\n\nfrom datetime import datetime\nfrom app_package.models import Player, InternalState\nfrom app_package import db\n\nIS_DEBUG = False\nSTATE_TABLE=\"state_table.json\"\nLOG_FILENAME = \"scheduler.log\"\n\n\n# Logging configuration\n# -- Create logger\nlogger = logging.getLogger(\"scheduler\")\nlogger.setLevel(logging.DEBUG)\n# create console handler and set level to debug\nch = logging.StreamHandler()\nfh = logging.FileHandler(LOG_FILENAME)\nch.setLevel(logging.DEBUG)\nfh.setLevel(logging.DEBUG)\nformatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')\nch.setFormatter(formatter)\nfh.setFormatter(formatter)\nlogger.addHandler(ch)\nlogger.addHandler(fh)\n\n# # -- Create handler and set level:\n# file_handler = logging.FileHandler(LOG_FILENAME)\n# file_handler.setLevel(logging.DEBUG)\n# # -- Create formatter:\n# formatter = logging.Formatter('[%(levelnale)s]-%(asctime)s-%(name)s-%(message)s')\n# # -- Add formatter to handler\n# file_handler.setFormatter(file_handler)\n# # -- Add handler to logger\n# logger.addHandler(file_handler)\n\n\ndef get_params():\n parser = argparse.ArgumentParser()\n parser.add_argument(\"--params\", help=\"Scheduling parameters\")\n args = parser.parse_args()\n params_file = args.params\n with open(params_file, 'r') as f:\n params = json.load(f)\n logger.debug(\"Successfully opened {}\".format(params_file))\n\n return params\n\n\ndef get_internal_state():\n \"\"\"\n Query the DB to retrieve the latest internal state. Output is a dict.\n \"\"\"\n state = (\n InternalState.query\n .with_entities(InternalState.is_challenge_open,\n InternalState.is_final_vote_open,\n InternalState.is_leaderboard_open)\n .order_by(InternalState.id.desc())\n .first()\n )\n logger.debug(\"Successfully QUERIED InternalState\")\n return state._asdict()\n\n\ndef set_new_internal_state(state_d):\n \"\"\"\n Insert record into the InternalState view. Input is a dict.\n \"\"\"\n\n state = InternalState(is_challenge_open=state_d[\"is_challenge_open\"],\n is_final_vote_open=state_d[\"is_final_vote_open\"],\n is_leaderboard_open=state_d[\"is_leaderboard_open\"])\n db.session.add(state)\n db.session.commit()\n logger.debug(\"Successfully WROTE into InternalState\")\n print(state)\n\n\ndef refill_credits(params):\n \"\"\"\n Execute the refill operation with penalty on score if not all previous credits\n were used.\n \"\"\"\n logger.info(\"Starting refill operation\")\n refill_qty = params[\"refill\"][\"qty\"]\n score_penalty = params[\"refill\"][\"score_penalty\"]\n logger.debug(\"qty={}, penalty={}\".format(refill_qty, score_penalty))\n all_players = Player.query.all()\n for p in all_players:\n # Apply penalty if player didnt spend all his credits\n current_credit = p.credit\n logger.info(\"{}: {} credits remaining\".format(p.playername,\n current_credit))\n if current_credit > 0:\n if p.score > 0:\n p.score -= score_penalty\n logger.info(\"UNUSED CREDITS! -{}pts\".format(score_penalty))\n p.credit = refill_qty\n db.session.commit()\n logger.info(\"Refill operation is DONE.\")\n\n\ndef switch_state(params):\n logger.info(\"Starting switch_state\")\n now = datetime.now()\n current_state = get_internal_state()\n logger.debug(\"Current state: %s\", current_state)\n print(current_state)\n # Get state from file\n with open(STATE_TABLE, 'r') as f:\n state_table = json.load(f)\n logger.debug(\"Successfully opened %s\", STATE_TABLE)\n\n # Fetching all interval data:\n pattern = \"%Y-%m-%d-%H:%M\"\n date_start_challenge = datetime.strptime(params[\"open_challenge\"][\"start\"], pattern)\n date_end_challenge = datetime.strptime(params[\"open_challenge\"][\"stop\"], pattern)\n date_start_fvote = datetime.strptime(params[\"fvote\"][\"start\"], pattern)\n date_stop_fvote = datetime.strptime(params[\"fvote\"][\"stop\"], pattern)\n # Pauses\n is_pause = False\n pause_list = params[\"pauses\"]\n pause_bounds = []\n for pause in pause_list:\n date_start_pause = datetime.strptime(pause[\"start\"], pattern)\n date_stop_pause = datetime.strptime(pause[\"stop\"], pattern)\n pause_bounds.append((date_start_pause, date_stop_pause))\n\n logger.debug(\"Successfully loaded interval data\")\n\n\n if current_state == state_table[0] and now < date_start_challenge:\n logger.info(\"CASE: challenge NOT STARTED\")\n new_state = current_state\n\n elif current_state == state_table[0] and now >= date_start_challenge:\n logger.info(\"CASE: challenge JUST STARTED\")\n new_state = state_table[1]\n\n elif current_state == state_table[1] and now < date_end_challenge:\n # Check if now is inside a pause interval\n for (pause_start, pause_stop) in pause_bounds:\n if pause_start < now < pause_stop:\n logger.info(\"CASE: challenge ONGOING + PAUSE\")\n is_pause = True\n if not is_pause:\n logger.info(\"CASE: challenge ONGOING\")\n refill_credits(params)\n new_state = current_state\n\n elif current_state == state_table[1] and date_end_challenge < now < date_start_fvote:\n logger.info(\"CASE: challenge JUST FINISHED\")\n new_state = state_table[0]\n\n elif current_state == state_table[0] and date_start_fvote <= now < date_stop_fvote:\n logger.info(\"CASE: final vote STARTED\")\n all_players = Player.query.all()\n for p in all_players:\n p.credit = params[\"refill\"][\"qty_fvote\"]\n new_state = state_table[2]\n\n elif current_state == state_table[2] and now >= date_stop_fvote:\n # Check that everybody did vote:\n is_fvote_complete = False\n all_players = Player.query.all()\n for p in all_players:\n if p.credit > 0:\n is_fvote_complete = True\n logger.info(\"CASE: final vote FINISHED\")\n new_state = state_table[3]\n else:\n logger.info(\"CASE: final vote NEED EVERYONE TO VOTE\")\n new_state = current_state\n\n else:\n logger.warning(\"UNKNOWN STATE !\")\n new_state = current_state\n\n logger.info(\"Setting new state: %s\", new_state)\n set_new_internal_state(new_state)\n \n\ndef job():\n \"\"\"\n Function to be executed by the scheduler.\n \"\"\"\n\n params = get_params()\n switch_state(params)\n\nif IS_DEBUG:\n # Manual execution of the job\n job()\nelse:\n params = get_params()\n schedule.every(params[\"refill\"][\"frequency_min\"]).minutes.do(job)\n while True:\n schedule.run_pending()\n time.sleep(1)\n\n\n", "sub_path": "scheduler/scheduler_v2.py", "file_name": "scheduler_v2.py", "file_ext": "py", "file_size_in_byte": 6822, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "sys.path.append", "line_number": 2, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 2, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 21, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 22, "usage_type": "attribute"}, {"api_name": "logging.StreamHandler", "line_number": 24, "usage_type": "call"}, {"api_name": "logging.FileHandler", "line_number": 25, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 26, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 27, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 28, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 46, "usage_type": "call"}, {"api_name": "json.load", "line_number": 51, "usage_type": "call"}, {"api_name": "app_package.models.InternalState.query.with_entities", "line_number": 62, "usage_type": "call"}, {"api_name": "app_package.models.InternalState.query", "line_number": 62, "usage_type": "attribute"}, {"api_name": "app_package.models.InternalState", "line_number": 62, "usage_type": "name"}, {"api_name": "app_package.models.InternalState.is_challenge_open", "line_number": 63, "usage_type": "attribute"}, {"api_name": "app_package.models.InternalState", "line_number": 63, "usage_type": "name"}, {"api_name": "app_package.models.InternalState.is_final_vote_open", "line_number": 64, "usage_type": "attribute"}, {"api_name": "app_package.models.InternalState", "line_number": 64, "usage_type": "name"}, {"api_name": "app_package.models.InternalState.is_leaderboard_open", "line_number": 65, "usage_type": "attribute"}, {"api_name": "app_package.models.InternalState", "line_number": 65, "usage_type": "name"}, {"api_name": "app_package.models.InternalState.id.desc", "line_number": 66, "usage_type": "call"}, {"api_name": "app_package.models.InternalState.id", "line_number": 66, "usage_type": "attribute"}, {"api_name": "app_package.models.InternalState", "line_number": 66, "usage_type": "name"}, {"api_name": "app_package.models.InternalState", "line_number": 78, "usage_type": "call"}, {"api_name": "app_package.db.session.add", "line_number": 81, "usage_type": "call"}, {"api_name": "app_package.db.session", "line_number": 81, "usage_type": "attribute"}, {"api_name": "app_package.db", "line_number": 81, "usage_type": "name"}, {"api_name": "app_package.db.session.commit", "line_number": 82, "usage_type": "call"}, {"api_name": "app_package.db.session", "line_number": 82, "usage_type": "attribute"}, {"api_name": "app_package.db", "line_number": 82, "usage_type": "name"}, {"api_name": "app_package.models.Player.query.all", "line_number": 96, "usage_type": "call"}, {"api_name": "app_package.models.Player.query", "line_number": 96, "usage_type": "attribute"}, {"api_name": "app_package.models.Player", "line_number": 96, "usage_type": "name"}, {"api_name": "app_package.db.session.commit", "line_number": 107, "usage_type": "call"}, {"api_name": "app_package.db.session", "line_number": 107, "usage_type": "attribute"}, {"api_name": "app_package.db", "line_number": 107, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 113, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 113, "usage_type": "name"}, {"api_name": "json.load", "line_number": 119, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 124, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 124, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 125, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 125, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 126, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 126, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 127, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 127, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 133, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 133, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 134, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 134, "usage_type": "name"}, {"api_name": "app_package.models.Player.query.all", "line_number": 165, "usage_type": "call"}, {"api_name": "app_package.models.Player.query", "line_number": 165, "usage_type": "attribute"}, {"api_name": "app_package.models.Player", "line_number": 165, "usage_type": "name"}, {"api_name": "app_package.models.Player.query.all", "line_number": 173, "usage_type": "call"}, {"api_name": "app_package.models.Player.query", "line_number": 173, "usage_type": "attribute"}, {"api_name": "app_package.models.Player", "line_number": 173, "usage_type": "name"}, {"api_name": "schedule.every", "line_number": 204, "usage_type": "call"}, {"api_name": "schedule.run_pending", "line_number": 206, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 207, "usage_type": "call"}]} +{"seq_id": "95754539", "text": "# -*- encoding: utf-8 -*-\n\"\"\"\nLicense: MIT\nCopyright (c) 2019 - present AppSeed.us\n\"\"\"\n\n# Python modules\nimport os, logging \n\n# Flask modules\nfrom flask import render_template, request, url_for, redirect, send_from_directory, session, make_response\nfrom flask_login import login_user, logout_user, current_user, login_required\nfrom werkzeug.exceptions import HTTPException, NotFound, abort, BadRequestKeyError\n\n# App modules\nfrom app import app\nfrom app.models import User\nfrom app.constants import *\nfrom app.forms import LoginForm, RegisterForm\nimport os\n# Firestore\nfrom firebase_admin import firestore\n\nuser = User()\ndb = firestore.client()\n#Index Route\n@app.route('/')\ndef index():\n return render_template('pages/index.html')\n\n@app.route('/about')\ndef about():\n return render_template('pages/aboutus.html')\n\n@app.route('/contact-us')\ndef contactus():\n return render_template('pages/contactus.html')\n\n\n@app.route('/coursehub')\ndef courses():\n if 'uname' not in session:\n return redirect(url_for('login'))\n return render_template('courses/courses.html')\n\n@app.route('/components')\ndef components():\n return render_template('pages/components.html')\n\n@app.route('/courses/python', methods=['GET', 'POST'])\ndef py():\n if 'uname' not in session:\n return redirect(url_for('index'))\n hidden_vals = py_course['courses']\n doc = user.getCourseDoc(session['uname'], u'python', db)\n\n if request.method == 'POST':\n user.assignmentCompleted(hidden_vals, session['uname'], u'python', db)\n return redirect(url_for('py'))\n\n session['py'] = user.getProgress(doc, py_course['courses'])\n return render_template('courses/python.html', doc=doc, py_course=py_course['courses'])\n\n@app.route('/courses/swift', methods=['GET', 'POST'])\ndef swift():\n if 'uname' not in session:\n return redirect(url_for('index'))\n hidden_vals = sw_course['courses']\n doc = user.getCourseDoc(session['uname'], u'swift', db)\n\n if request.method == 'POST':\n user.assignmentCompleted(hidden_vals, session['uname'], u'swift', db)\n return redirect(url_for('swift'))\n\n session['sw'] = user.getProgress(doc, sw_course['courses'])\n return render_template('courses/swift.html', doc=doc, sw_course=sw_course['courses'])\n\n@app.route('/courses/electrical-engineering', methods=['GET', 'POST'])\ndef ee():\n if 'uname' not in session:\n return redirect(url_for('index'))\n hidden_vals = ee_course['courses']\n doc = user.getCourseDoc(session['uname'], u'electrical-engineering', db)\n\n if request.method == 'POST':\n user.assignmentCompleted(hidden_vals, session['uname'], u'electrical-engineering', db)\n return redirect(url_for('ee'))\n\n session['ee'] = user.getProgress(doc, ee_course['courses'])\n return render_template('courses/electroeg.html', doc=doc, ee_course=ee_course['courses'])\n\n# @app.route('/article', methods=['GET', 'POST'])\n# def article(): \n# return render_template('courses/article-template.html')\n\n# Register a new user\n@app.route('/register', methods=['GET', 'POST'])\ndef register():\n \n # declare the Registration Form\n form = RegisterForm(request.form)\n msg = None\n\n if request.method == 'GET': \n return render_template( 'accounts/register.html', form=form, msg=msg )\n\n # check if both http method is POST and form is valid on submit\n if form.validate_on_submit():\n\n # assign form data to variables\n username = request.form.get('username', '', type=str)\n password = request.form.get('password', '', type=str) \n email = request.form.get('email', '', type=str) \n\n msg = user.create_user(email, password, username, db)\n \n if msg == None:\n doc_ref = db.collection(u'users').document(username)\n doc_ref.set({\n u'uname': username,\n })\n\n user.setupCourse(session['uname'], u'python', py_course['courses'], db)\n user.setupCourse(session['uname'], u'swift', sw_course['courses'], db)\n user.setupCourse(session['uname'], u'electrical-engineering', ee_course['courses'], db)\n\n return redirect(url_for('index'))\n\n \n return render_template( 'accounts/register.html', form=form, msg=msg )\n\n# Authenticate user\n@app.route('/login', methods=['GET', 'POST'])\ndef login():\n \n # Declare the login form\n form = LoginForm(request.form)\n\n # Flask message injected into the page, in case of any errors\n msg = None\n\n # check if both http method is POST and form is valid on submit\n if form.validate_on_submit():\n\n # assign form data to variables\n email = request.form.get('email', '', type=str) \n password = request.form.get('password', '', type=str) \n\n msg = user.login_user(email, password)\n\n if msg != None:\n return render_template( 'accounts/login.html', form=form, msg=msg )\n return redirect(url_for('index'))\n\n\n return render_template( 'accounts/login.html', form=form, msg=msg )\n\n# Logout user\n@app.route('/logout')\ndef logout():\n if 'uname' not in session:\n return redirect(url_for('index'))\n user.logout_user()\n return redirect(url_for('index'))\n\n@app.errorhandler(404)\ndef not_found(error):\n \"\"\"Page not found.\"\"\"\n return render_template(\"pages/error-404.html\")\n\n\n\n", "sub_path": "app/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 5352, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "app.models.User", "line_number": 24, "usage_type": "call"}, {"api_name": "firebase_admin.firestore.client", "line_number": 25, "usage_type": "call"}, {"api_name": "firebase_admin.firestore", "line_number": 25, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 29, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 27, "usage_type": "call"}, {"api_name": "app.app", "line_number": 27, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 33, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 31, "usage_type": "call"}, {"api_name": "app.app", "line_number": 31, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 37, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 35, "usage_type": "call"}, {"api_name": "app.app", "line_number": 35, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 42, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 43, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 43, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 44, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 40, "usage_type": "call"}, {"api_name": "app.app", "line_number": 40, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 48, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 46, "usage_type": "call"}, {"api_name": "app.app", "line_number": 46, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 52, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 53, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 53, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 55, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 57, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 57, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 58, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 61, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 62, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 50, "usage_type": "call"}, {"api_name": "app.app", "line_number": 50, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 66, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 67, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 67, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 69, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 71, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 71, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 72, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 73, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 73, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 75, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 76, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 64, "usage_type": "call"}, {"api_name": "app.app", "line_number": 64, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 80, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 81, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 81, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 83, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 85, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 85, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 86, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 87, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 87, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 89, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 90, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 78, "usage_type": "call"}, {"api_name": "app.app", "line_number": 78, "usage_type": "name"}, {"api_name": "app.forms.RegisterForm", "line_number": 101, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 101, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 101, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 104, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 104, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 105, "usage_type": "call"}, {"api_name": "flask.request.form.get", "line_number": 111, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 111, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 111, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 112, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 112, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 112, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 113, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 113, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 113, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 123, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 124, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 125, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 127, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 127, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 130, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 97, "usage_type": "call"}, {"api_name": "app.app", "line_number": 97, "usage_type": "name"}, {"api_name": "app.forms.LoginForm", "line_number": 137, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 137, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 137, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 146, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 146, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 146, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 147, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 147, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 147, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 152, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 153, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 153, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 156, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 133, "usage_type": "call"}, {"api_name": "app.app", "line_number": 133, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 161, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 162, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 162, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 164, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 164, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 159, "usage_type": "call"}, {"api_name": "app.app", "line_number": 159, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 169, "usage_type": "call"}, {"api_name": "app.app.errorhandler", "line_number": 166, "usage_type": "call"}, {"api_name": "app.app", "line_number": 166, "usage_type": "name"}]} +{"seq_id": "294800403", "text": "from tachycardia import is_tachycardic\nimport pytest\n@pytest.mark.parametrize(\"teststr,expected\",[\n (\"tachycardic\",True),\n (\"TACHYcarDic\", True),\n (\"\", False),\n (\"tttttttttttttttachycardic\", False),\n (\" tachycardic!!! \", True),\n (\" TA C H y!ardi??? c\", True),\n (\"tBchycDrdic\", True),\n (\"tbchycardLcs\", False),\n (10,False)\n])\ndef test(teststr,expected):\n res = is_tachycardic(teststr)\n assert res == expected\n", "sub_path": "test_tachy.py", "file_name": "test_tachy.py", "file_ext": "py", "file_size_in_byte": 442, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "tachycardia.is_tachycardic", "line_number": 15, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 3, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 3, "usage_type": "attribute"}]} +{"seq_id": "68090965", "text": "# -*- coding: utf-8 -*-\r\n\r\n\"\"\"\r\nTO-DO:\r\n - Remove free variable requirement\r\n - Allow for fractional atomic contribution\r\n - Fix checkbox layout\r\n\"\"\"\r\n\r\nimport re\r\nfrom sympy import symbols\r\n\r\n\r\nloc = '../CPL Energies/' # location of energies\r\n\r\nCu, Bi, W, O = symbols(['Cu','Bi','W','O'])\r\nfree, x, y, z = Cu, Bi, W, O\r\n\r\n\r\ndef getEnthalpy(_dict, f, m=None):\r\n \r\n # collect number of each species\r\n frac = dict()\r\n if not m:\r\n for b in bulks:\r\n if b in f:\r\n n = re.search(re.escape(b) + '[0-9]+', f)\r\n if n:\r\n frac[b] = int(re.search(r\"[0-9][0-9]*\", n.group(0)).group(0))\r\n else:\r\n frac[b] = 1\r\n elif re.match('Vac', f):\r\n frac.update((k, v - 1) if k == f.split('_')[1] else (k, v) for k, v in Super['fracs'].items())\r\n else:\r\n s = f.split('_')\r\n frac.update((k, v + 1) if k == s[0] else (k, v) for k, v in Super['fracs'].items())\r\n frac.update((k, v - 1) if k == s[1] else (k, v) for k, v in frac.items())\r\n _dict[f]['fracs'] = frac\r\n # calculate enthalpy\r\n return _dict[f]['E'] - sum(_dict[f]['fracs'][i] * bulks[i] for i in _dict[f]['fracs'])\r\n\r\n\r\n# Collect DFT energies and calculate enthalpies\r\nfrags = dict(); bulks = dict()\r\ntry:\r\n with open(loc + 'bulkE.txt', 'r') as b, open(loc + 'totalE.txt', 'r') as t:\r\n for line in b.readlines():\r\n i = line.split()\r\n if i[0] == 'O':\r\n bulks[i[0]] = float(i[1]) / 2\r\n else:\r\n bulks[i[0]] = float(i[1])\r\n for line in t.readlines():\r\n i = line.split()\r\n frags[i[0]] = {}\r\n frags[i[0]]['E'] = float(i[1])\r\n frags[i[0]]['H'] = round(getEnthalpy(frags, i[0]), 3)\r\nexcept IOError as e:\r\n print('\\nOperation failed: %s' % e.strerror)\r\n\r\nprim = next(iter(frags.keys()))\r\n\r\ndefects = dict()\r\ntry:\r\n with open(loc + 'defectE.txt', 'r') as d:\r\n for line in d.readlines():\r\n i = line.split()\r\n m = 1\r\n if i[0] == 'Super':\r\n Super = dict()\r\n Super['E'] = float(i[1])\r\n m = round(Super['E'] / frags[prim]['E'])\r\n frac = dict(); frac.update((k, m * v) for k, v in frags[prim]['fracs'].items())\r\n Super['fracs'] = frac\r\n Super['H'] = round(Super['E'] - sum(Super['fracs'][i] * bulks[i] for i in Super['fracs']), 3)\r\n else:\r\n defects[i[0]] = dict()\r\n defects[i[0]]['E'] = float(i[1])\r\n defects[i[0]]['H'] = round(getEnthalpy(defects, i[0], m) / m, 3)\r\n defects[i[0]]['diff'] = {k: v for k, v in zip(bulks, [s - d for s, d in zip(Super['fracs'].values(),\r\n defects[i[0]]['fracs'].values())]) if k in i[0].split('_')}\r\nexcept IOError as e:\r\n print('\\nOperation failed: %s' % e.strerror)\r\n\r\n# %% Define equations, inequalities, lines, and shading\r\nfrom sympy import Lt\r\nfrom sympy.solvers import solve\r\n\r\nfor j, f in enumerate(frags):\r\n ########################## EQUATIONS | INEQUALITIES #######################\r\n s = x; s -= s # define empty variable\r\n for l, v in frags[f]['fracs'].items():\r\n s += v * symbols(l)\r\n frags[f]['eq'] = s - frags[f]['H'] # generate equations\r\n if f == prim: # define primary line\r\n frags[f]['line'] = solve(frags[f]['eq'], y)[0].subs(z, 0)\r\n p = solve(frags[f]['eq'], z)[0] # solve in terms of z variable\r\n else:\r\n frags[f]['ineq'] = Lt(s, frags[f]['H']) # generate inequalities\r\n sol = frags[f]['eq'].subs(z, p) # solve by z back-substitution\r\n ############################# LINES | SHADES ##############################\r\n if str(y) in str(sol): # non-vertical lines\r\n frags[f]['line'] = solve(sol, y)[0]\r\n frags[f]['shade'] = solve(frags[f]['ineq'].subs(z, solve(frags[prim]['eq'], z)[0]), y)\r\n else: # vertical lines\r\n frags[f]['line'] = solve(sol, x)[0]\r\n frags[f]['shade'] = solve(frags[f]['ineq'].subs(z, solve(frags[prim]['eq'], z)[0]), x)\r\n\r\n# %% Define CPL\r\nimport numpy as np\r\nfrom sympy import lambdify\r\nfrom shapely.geometry import LineString, Polygon\r\nfrom descartes import PolygonPatch\r\nimport matplotlib.pyplot as plt\r\n\r\nfig, ax = plt.subplots(figsize=(10, 8))\r\nplt.subplots_adjust(left=0.1, bottom=0.1, right=0.85, top=0.85)\r\nax.xaxis.tick_top(); ax.xaxis.set_label_position('top')\r\nax.yaxis.tick_right(); ax.yaxis.set_label_position('right')\r\nax.spines['left'].set_visible(False)\r\nax.spines['bottom'].set_visible(False)\r\nplt.minorticks_on()\r\n\r\nfont = {'family': 'times new roman', 'size': 16}\r\naxes = {'fontsize': 20, 'labelpad': 20}\r\nticks = {'labelsize': 16}\r\nmath = {'rm': 'Roman'}\r\nopt = {'font': font, 'xtick': ticks, 'ytick': ticks, 'mathtext': math}\r\nfor i in opt: plt.rc(i, **opt[i])\r\n\r\nax.tick_params(axis='both', **ticks)\r\n\r\na = .75 # line opacity\r\na_s = 0.2 # shade opacity\r\n\r\ncolors = plt.cm.get_cmap('Dark2', len(frags) - 1)\r\n\r\ndef draw_CPL(c):\r\n ax.set_xlabel('$\\Delta\\mu_{%s}$' % str(x), **axes)\r\n ax.set_ylabel('$\\Delta\\mu_{%s}$' % str(y), **axes)\r\n lines = [];\r\n shades = []\r\n poly_x, poly_y, xy, v = np.zeros(4)\r\n for i, f in enumerate(frags):\r\n if i == 0: # primary line\r\n q = frags[f]['line'].subs(free, c)\r\n minX = round(float(solve(q, x)[0]), 3)\r\n minY = round(float(q.subs(x, 0)), 3)\r\n ax.axis([minX, 0.0, minY, 0.0])\r\n line = lambdify(x, q)\r\n t = np.array([minX, 0])\r\n lines.append(ax.plot(t, line(t), color='k', zorder=len(frags), label=f))\r\n primPoly = Polygon([[minX, 0], [0, 0], [0, minY]])\r\n else:\r\n l = frags[f]['line'].subs(free, c)\r\n s = str(frags[f]['shade'])\r\n if not str(y) in s: # vertical\r\n A = (l, 1); B = (l, minY - 1)\r\n seg = LineString([A, B])\r\n pts = seg.intersection(primPoly.boundary)\r\n t = np.array([p[0][0] for p in [p.coords.xy for p in pts]])\r\n v = np.array([p[1][0] for p in [p.coords.xy for p in pts]])\r\n if str(\">\") in s:\r\n xVal = 1\r\n else:\r\n xVal = minX - 1\r\n poly_x = np.array([A[0], xVal, xVal, B[0]])\r\n poly_y = np.array([A[1], A[1], B[1], B[1]])\r\n else: # non-vertical\r\n line = lambdify(x, l)\r\n A = (minX - 1000, line(minX - 1000)); B = (1000, line(1000))\r\n seg = LineString([A, B])\r\n pts = seg.intersection(primPoly.boundary)\r\n t = np.array([p[0][0] for p in [p.coords.xy for p in pts]])\r\n v = np.array([p[1][0] for p in [p.coords.xy for p in pts]])\r\n if not str(x) in s: # horizontal\r\n if str(\">\") in s:\r\n poly_x = np.array([A[0], A[0], B[0], B[0]])\r\n poly_y = np.array([A[1], 1, 1, B[1]])\r\n else:\r\n poly_x = np.array([A[0], A[0], B[0], B[0]])\r\n poly_y = np.array([A[1], minY - 1, minY - 1, B[1]])\r\n else: # slopped\r\n poly_x = [A[0], B[0]]; poly_y = [A[1], B[1]]\r\n if line(0) < line(1): # positive slope\r\n if str(\">\") in s:\r\n poly_x.append(A[0]); poly_y.append(B[1])\r\n else:\r\n poly_x.append(B[0]); poly_y.append(A[1])\r\n else: # negative slope\r\n if str(\">\") in s:\r\n poly_x.append(B[0]); poly_y.append(A[1])\r\n else:\r\n poly_x.append(A[0]); poly_y.append(B[1])\r\n\r\n lines.append(ax.plot(t, v, color=colors(i - 1), alpha=a, label=f, lw=2))\r\n poly = Polygon(np.array([[i, j] for i, j in zip(poly_x, poly_y)]))\r\n overlap = poly.intersection(primPoly)\r\n if overlap:\r\n shades.append(PolygonPatch(overlap, color=colors(i - 1), alpha=a_s, label=f))\r\n else:\r\n shades.append([])\r\n\r\n # mark single phase region\r\n # global SP\r\n # SP = Polygon(shades[0].get_path().vertices)\r\n # for act in shades[1:2]:\r\n # SP = SP.intersection(Polygon(act.get_path().vertices))\r\n # if SP and SP.geom_type != 'LineString':\r\n # if SP.geom_type != 'Polygon': SP = SP.geoms[1]\r\n # ax.add_patch(PolygonPatch(SP,color='k',alpha=0.75,label='SP')).set_hatch('x')\r\n\r\n return shades\r\n\r\nshades = draw_CPL(0.0)\r\n\r\n# %% Define widgets\r\nfrom widgets import Slider, PremiumCheckButtons\r\nfrom shapely.geometry import Point\r\n\r\ndef singlePhase():\r\n visibility = check.get_status()\r\n labels = [patch.get_label() for patch in ax.patches]\r\n global SP\r\n if 'SP' in labels: ax.patches.remove(ax.patches[labels.index('SP')])\r\n if visibility.count(True) > 1:\r\n if len(ax.patches) > 1 and len(ax.patches) == visibility.count(True):\r\n SP = Polygon(ax.patches[0].get_path().vertices)\r\n for act in ax.patches:\r\n SP = SP.intersection(Polygon(act.get_path().vertices))\r\n if SP and SP.geom_type != 'LineString':\r\n if SP.geom_type != 'Polygon': SP = SP.geoms[1]\r\n ax.add_patch(PolygonPatch(SP, color='k', alpha=0.75, label='SP')).set_hatch('x')\r\n\r\ndef update(val):\r\n plt.close(2)\r\n visibility = check.get_status()\r\n ax.lines.clear(); shades.clear(); ax.patches.clear()\r\n for vis, shade in zip(visibility, draw_CPL(sfree.val)):\r\n shades.append(shade)\r\n if shade:\r\n if vis:\r\n ax.add_patch(shade)\r\n shade.set_visible(True)\r\n else:\r\n shade.set_visible(False)\r\n singlePhase()\r\n\r\ndef shade(label):\r\n plt.close(2)\r\n i = [l.get_label() for l in ax.lines[1:]].index(re.sub('\\$_\\{([0-9]*)\\}\\$', r'\\1', label))\r\n if shades[i]:\r\n shades[i].set_visible(not shades[i].get_visible())\r\n if not shades[i] in ax.patches:\r\n ax.add_patch(shades[i])\r\n else:\r\n ax.patches.remove(shades[i])\r\n singlePhase()\r\n plt.draw()\r\n\r\ndef genDefect(event):\r\n plt.close(2)\r\n if event.xdata and 'SP' in [patch.get_label() for patch in ax.patches] \\\r\n and SP.contains(Point(event.xdata, event.ydata)):\r\n def delta(i, k):\r\n if symbols(k) == x:\r\n val = pts[:, 0][i]\r\n elif symbols(k) == y:\r\n val = pts[:, 1][i]\r\n elif symbols(k) == free:\r\n val = sfree.val\r\n else:\r\n val = solve(frags[prim]['eq'], z)[0].subs(\r\n {x: pts[:, 0][i], y: pts[:, 1][i], free: sfree.val})\r\n return val\r\n\r\n figDef = plt.figure(figsize=(10, 6)); gs = figDef.add_gridspec(2, 3)\r\n axDef = plt.subplot(gs[:, :2])\r\n axZoom = plt.subplot(gs[0, 2])\r\n\r\n bounds = SP.boundary.coords.xy\r\n pts = np.array([[x, y] for x, y in zip(bounds[0], bounds[1])])\r\n xDat = np.arange(1, len(pts))\r\n yDat = np.zeros([len(xDat), len(defects)])\r\n for i in range(len(xDat)):\r\n for j, d in zip(range(len(defects)), defects):\r\n yDat[i, j] = defects[d]['E'] - Super['E'] + \\\r\n sum([v * (bulks[k] + delta(i, k)) for k, v in defects[d]['diff'].items()])\r\n legDef = []\r\n for d in defects:\r\n if re.match('Vac', d):\r\n label = re.sub('Vac_(.*)', r'V$_{\\1}$', d)\r\n else:\r\n label = re.sub('_(.*)', r'$_{\\1}$', d)\r\n legDef.append(label)\r\n\r\n axZoom.plot(pts[:, 0], pts[:, 1], 'k--')\r\n axZoom.set_axis_off()\r\n for i, p in zip(xDat, pts):\r\n xs = 0.01\r\n ys = 0.01\r\n axZoom.text(p[0] + xs, p[1] + ys, i)\r\n\r\n axDef.plot(xDat, yDat)\r\n axDef.set_xlim([xDat[0], xDat[-1]])\r\n axDef.set_ylim(-0.5,3.0)\r\n axDef.set_xticks(xDat)\r\n axDef.set_xticklabels(xDat)\r\n for i in xDat:\r\n axDef.axvline(i, color='k', lw=0.5)\r\n axDef.set_ylabel('$\\Delta H_f$ (eV)')\r\n axDef.axhline(y=0, ls='--', lw=0.5, color='k')\r\n figDef.legend(legDef, ncol=1, prop={'size': 14}, loc='center', bbox_to_anchor=(0.78, 0.25))\r\n\r\n plt.show()\r\n\r\nif len(frags[prim]['fracs'].keys()) > 3:\r\n axSlider = plt.axes([0.1, 0.05, 0.75, 0.02]) # [left,bottom,width,height]\r\n sfree = Slider(axSlider, '$\\Delta\\mu_{%s}$' % str(free),\r\n round(float(solve(frags[prim]['line'].subs(x, 0))[0]), 0), 0.0,\r\n valinit=0.0, valstep=0.001, valfmt=\"%1.3f\")\r\n sfree.on_changed(update)\r\n sfree.label.set_size(20)\r\n\r\naxChecks = plt.axes([0.1, 0.1, 0.3, 0.4], frameon=False)\r\nvisibility = [shade.set_visible(False) if shade else [] for shade in shades]\r\ncheck = PremiumCheckButtons(axChecks, ax.lines[1:], visibility, loc=3, borderaxespad=0)\r\n\r\ncheck.on_clicked(shade)\r\n\r\ncid = fig.canvas.mpl_connect('button_press_event', genDefect)\r\n\r\nplt.show()", "sub_path": "CPL.py", "file_name": "CPL.py", "file_ext": "py", "file_size_in_byte": 13325, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "sympy.symbols", "line_number": 16, "usage_type": "call"}, {"api_name": "re.search", "line_number": 27, "usage_type": "call"}, {"api_name": "re.escape", "line_number": 27, "usage_type": "call"}, {"api_name": "re.search", "line_number": 29, "usage_type": "call"}, {"api_name": "re.match", "line_number": 32, "usage_type": "call"}, {"api_name": "sympy.symbols", "line_number": 93, "usage_type": "call"}, {"api_name": "sympy.solvers.solve", "line_number": 96, "usage_type": "call"}, {"api_name": "sympy.solvers.solve", "line_number": 97, "usage_type": "call"}, {"api_name": "sympy.Lt", "line_number": 99, "usage_type": "call"}, {"api_name": "sympy.solvers.solve", "line_number": 103, "usage_type": "call"}, {"api_name": "sympy.solvers.solve", "line_number": 104, "usage_type": "call"}, {"api_name": "sympy.solvers.solve", "line_number": 106, "usage_type": "call"}, {"api_name": "sympy.solvers.solve", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.minorticks_on", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 122, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rc", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm.get_cmap", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 136, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 143, "usage_type": "call"}, {"api_name": "sympy.solvers.solve", "line_number": 147, "usage_type": "call"}, {"api_name": "sympy.lambdify", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 151, "usage_type": "call"}, {"api_name": "shapely.geometry.Polygon", "line_number": 153, "usage_type": "call"}, {"api_name": "shapely.geometry.LineString", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 168, "usage_type": "call"}, {"api_name": "sympy.lambdify", "line_number": 170, "usage_type": "call"}, {"api_name": "shapely.geometry.LineString", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 182, "usage_type": "call"}, {"api_name": "shapely.geometry.Polygon", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 197, "usage_type": "call"}, {"api_name": "descartes.PolygonPatch", "line_number": 200, "usage_type": "call"}, {"api_name": "shapely.geometry.Polygon", "line_number": 228, "usage_type": "call"}, {"api_name": "shapely.geometry.Polygon", "line_number": 230, "usage_type": "call"}, {"api_name": "descartes.PolygonPatch", "line_number": 233, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.close", "line_number": 236, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 236, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 250, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 250, "usage_type": "name"}, {"api_name": "re.sub", "line_number": 251, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.draw", "line_number": 259, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 259, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 262, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 262, "usage_type": "name"}, {"api_name": "shapely.geometry.Point", "line_number": 264, "usage_type": "call"}, {"api_name": "sympy.symbols", "line_number": 266, "usage_type": "call"}, {"api_name": "sympy.symbols", "line_number": 268, "usage_type": "call"}, {"api_name": "sympy.symbols", "line_number": 270, "usage_type": "call"}, {"api_name": "sympy.solvers.solve", "line_number": 273, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 277, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 277, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 278, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 278, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 279, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 279, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 282, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 283, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 284, "usage_type": "call"}, {"api_name": "re.match", "line_number": 291, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 292, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 294, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 315, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 315, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 318, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 318, "usage_type": "name"}, {"api_name": "widgets.Slider", "line_number": 319, "usage_type": "call"}, {"api_name": "sympy.solvers.solve", "line_number": 320, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 325, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 325, "usage_type": "name"}, {"api_name": "widgets.PremiumCheckButtons", "line_number": 327, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 333, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 333, "usage_type": "name"}]} +{"seq_id": "431975563", "text": "#!/usr/bin/python\n# -*- coding: utf-8 -*-\n# Author: Zhiquan Chen\n# Date: \n\n\"\"\"\n[Function Description]\n------------------------------------------------------------------------------------------------------------------------\n将文件夹中的图片修改为相同的某一格式\n------------------------------------------------------------------------------------------------------------------------\n\"\"\"\n\nimport os\nimport time\nfrom PIL import Image\n\n\ndef image_format_modify(image_dir, image_suffix):\n \"\"\"\n 获取文件夹目录中的多个文件夹名称,得到各个文件夹的绝对路径\n \"\"\"\n image_list = os.listdir(image_dir)\n image_path_list = []\n for image in image_list:\n image_path = os.path.join(image_dir, image)\n image_path_list.append(image_path)\n for image_orginal in image_path_list:\n image_split = str.split(image_orginal, '.')\n if image_split[-1] != image_suffix:\n image_file = Image.open(image_orginal)\n image_modified = image_split[0] + '.' + image_suffix\n image_file.save(image_modified)\n image_file.close()\n os.remove(image_orginal)\n image_format_list = os.listdir(image_dir)\n print(image_format_list)\n\n\nif __name__ == '__main__':\n # 记录程序开始时间\n start = time.time()\n\n # 主程序 ###########################################################################################################\n image_directory = r'F:\\Guitar\\Guitar_Tabs\\云烟成雨'\n image_format = 'png'\n image_format_modify(image_directory, image_format)\n # #################################################################################################################\n\n # 统计计算时间\n end = time.time()\n cal_time = end - start\n m, s = divmod(cal_time, 60)\n h, m = divmod(m, 60)\n print(\"\\nTotal Computing Time: %02d:%02d:%02d\" % (h, m, s))", "sub_path": "Daily_Life/Image_Modification/Image_Format.py", "file_name": "Image_Format.py", "file_ext": "py", "file_size_in_byte": 1901, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "os.listdir", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 30, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 30, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 34, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 35, "usage_type": "call"}, {"api_name": "time.time", "line_number": 41, "usage_type": "call"}, {"api_name": "time.time", "line_number": 50, "usage_type": "call"}]} +{"seq_id": "201029369", "text": "import os\nimport torch\nimport torch.utils.data as Data\nimport utils.transforms as trans\nimport utils.utils as util\nimport layer.loss as ls\n\nimport shutil\nimport cfg.CDD as cfg\nimport dataset.rs as dates\nimport time\nimport datetime\nfrom funcs import validate\n\ndevice = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\nresume = 0\n\n\ndef main():\n ######### configs ###########\n best_metric = 0\n ###### load datasets ########\n train_transform_det = trans.Compose([trans.Scale(cfg.TRANSFROM_SCALES),])\n val_transform_det = trans.Compose([trans.Scale(cfg.TRANSFROM_SCALES),])\n train_data = dates.Dataset(cfg.TRAIN_DATA_PATH, cfg.TRAIN_LABEL_PATH, cfg.TRAIN_TXT_PATH, 'train', transform=True, transform_med=train_transform_det)\n train_loader = Data.DataLoader(train_data, batch_size=cfg.BATCH_SIZE, shuffle=True, num_workers=4, pin_memory=True)\n val_data = dates.Dataset(cfg.VAL_DATA_PATH, cfg.VAL_LABEL_PATH, cfg.VAL_TXT_PATH, 'val', transform=True, transform_med=val_transform_det)\n val_loader = Data.DataLoader(val_data, batch_size=1, shuffle=False, num_workers=1, pin_memory=True)\n ###### build models ########\n base_seg_model = 'resnet50'\n if base_seg_model == 'vgg':\n import model.siameseNet.d_aa as models\n pretrain_deeplab_path = os.path.join(cfg.PRETRAIN_MODEL_PATH, 'vgg16.pth')\n model = models.SiameseNet(norm_flag='l2')\n if resume:\n checkpoint = torch.load(cfg.TRAINED_BEST_PERFORMANCE_CKPT)\n model.load_state_dict(checkpoint['state_dict'])\n print('resume success')\n else:\n deeplab_pretrain_model = torch.load(pretrain_deeplab_path)\n model.init_parameters_from_deeplab(deeplab_pretrain_model)\n print('load vgg')\n else:\n import model.siameseNet.dares as models\n model = models.SiameseNet(norm_flag='l2')\n if resume:\n checkpoint = torch.load(cfg.TRAINED_BEST_PERFORMANCE_CKPT)\n model.load_state_dict(checkpoint['state_dict'])\n print('resume success')\n else:\n print('load resnet50')\n if device.type == 'cuda':\n print(\"use cuda!\")\n model = model.cuda()\n MaskLoss = ls.ContrastiveLoss1()\n ab_test_dir = os.path.join(cfg.SAVE_PRED_PATH, 'contrastive_loss')\n util.check_dir(ab_test_dir)\n save_change_map_dir = os.path.join(ab_test_dir, 'changemaps/')\n save_valid_dir = os.path.join(ab_test_dir,'valid_imgs')\n save_roc_dir = os.path.join(ab_test_dir,'roc')\n util.check_dir(save_change_map_dir),util.check_dir(save_valid_dir),util.check_dir(save_roc_dir)\n #########\n ######### optimizer ##########\n ######## how to set different learning rate for differernt layers #########\n optimizer = torch.optim.Adam(params=model.parameters(), lr=cfg.INIT_LEARNING_RATE, weight_decay=cfg.DECAY)\n ######## iter img_label pairs ###########\n loss_total = 0\n time_start = time.time()\n for epoch in range(60):\n for batch_idx, batch in enumerate(train_loader):\n step = epoch * len(train_loader) + batch_idx\n util.adjust_learning_rate(cfg.INIT_LEARNING_RATE, optimizer, step)\n model.train()\n img1, img2, label, filename, height, width = batch\n if device.type == 'cuda':\n img1, img2, label = img1.cuda(), img2.cuda(), label.cuda()\n label = label.float()\n out_conv5, out_fc, out_embedding = model(img1, img2)\n out_conv5_t0, out_conv5_t1 = out_conv5\n out_fc_t0, out_fc_t1 = out_fc\n out_embedding_t0, out_embedding_t1 = out_embedding\n label_rz_conv5 = util.rz_label(label, size=out_conv5_t0.data.cpu().numpy().shape[2:])\n label_rz_fc = util.rz_label(label, size=out_fc_t0.data.cpu().numpy().shape[2:])\n label_rz_embedding = util.rz_label(label, size=out_embedding_t0.data.cpu().numpy().shape[2:])\n if device.type == 'cuda':\n label_rz_conv5 = label_rz_conv5.cuda()\n label_rz_fc = label_rz_fc.cuda()\n label_rz_embedding = label_rz_embedding.cuda()\n contractive_loss_conv5 = MaskLoss(out_conv5_t0, out_conv5_t1, label_rz_conv5)\n contractive_loss_fc = MaskLoss(out_fc_t0, out_fc_t1, label_rz_fc)\n contractive_loss_embedding = MaskLoss(out_embedding_t0, out_embedding_t1, label_rz_embedding)\n loss = contractive_loss_conv5 + contractive_loss_fc + contractive_loss_embedding\n loss_total += loss.data.cpu()\n optimizer.zero_grad()\n loss.backward()\n optimizer.step()\n if (batch_idx) % 20 == 0:\n print(\"Epoch [%d/%d] Loss: %.4f Mask_Loss_conv5: %.4f Mask_Loss_fc: %.4f \"\n \"Mask_Loss_embedding: %.4f\" % (epoch, batch_idx,loss.item(),contractive_loss_conv5.item(),\n contractive_loss_fc.item(),contractive_loss_embedding.item()))\n if (batch_idx) % 1000 == 0:\n model.eval()\n current_metric = validate(model, val_loader, epoch, save_change_map_dir, save_roc_dir, cfg.TRANSFROM_SCALES)\n if current_metric > best_metric:\n torch.save({'state_dict': model.state_dict()}, os.path.join(ab_test_dir, 'model' + str(epoch) + '.pth'))\n shutil.copy(os.path.join(ab_test_dir, 'model' + str(epoch) + '.pth'), os.path.join(ab_test_dir, 'model_best.pth'))\n best_metric = current_metric\n current_metric = validate(model, val_loader, epoch, save_change_map_dir, save_roc_dir, cfg.TRANSFROM_SCALES)\n if current_metric > best_metric:\n torch.save({'state_dict': model.state_dict()}, os.path.join(ab_test_dir, 'model' + str(epoch) + '.pth'))\n shutil.copy(os.path.join(ab_test_dir, 'model' + str(epoch) + '.pth'), os.path.join(ab_test_dir, 'model_best.pth'))\n best_metric = current_metric\n if epoch % 5 == 0:\n torch.save({'state_dict': model.state_dict()}, os.path.join(ab_test_dir, 'model' + str(epoch) + '.pth'))\n elapsed = round(time.time() - time_start)\n elapsed = str(datetime.timedelta(seconds=elapsed))\n print('Elapsed {}'.format(elapsed))\n\n\nif __name__ == '__main__':\n torch.backends.cudnn.enabled = False \n main()\n", "sub_path": "train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 6383, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "torch.device", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 15, "usage_type": "attribute"}, {"api_name": "utils.transforms.Compose", "line_number": 23, "usage_type": "call"}, {"api_name": "utils.transforms", "line_number": 23, "usage_type": "name"}, {"api_name": "utils.transforms.Scale", "line_number": 23, "usage_type": "call"}, {"api_name": "cfg.CDD.TRANSFROM_SCALES", "line_number": 23, "usage_type": "attribute"}, {"api_name": "cfg.CDD", "line_number": 23, "usage_type": "name"}, {"api_name": "utils.transforms.Compose", "line_number": 24, "usage_type": "call"}, {"api_name": "utils.transforms", "line_number": 24, "usage_type": "name"}, {"api_name": "utils.transforms.Scale", "line_number": 24, "usage_type": "call"}, {"api_name": "cfg.CDD.TRANSFROM_SCALES", "line_number": 24, "usage_type": "attribute"}, {"api_name": "cfg.CDD", "line_number": 24, "usage_type": "name"}, {"api_name": "dataset.rs.Dataset", "line_number": 25, "usage_type": "call"}, {"api_name": "dataset.rs", "line_number": 25, "usage_type": "name"}, {"api_name": "cfg.CDD.TRAIN_DATA_PATH", "line_number": 25, "usage_type": "attribute"}, {"api_name": "cfg.CDD", "line_number": 25, "usage_type": "name"}, {"api_name": "cfg.CDD.TRAIN_LABEL_PATH", "line_number": 25, "usage_type": "attribute"}, {"api_name": "cfg.CDD.TRAIN_TXT_PATH", "line_number": 25, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 26, "usage_type": "name"}, {"api_name": "cfg.CDD.BATCH_SIZE", "line_number": 26, "usage_type": "attribute"}, {"api_name": "cfg.CDD", "line_number": 26, "usage_type": "name"}, {"api_name": "dataset.rs.Dataset", "line_number": 27, "usage_type": "call"}, {"api_name": "dataset.rs", "line_number": 27, "usage_type": "name"}, {"api_name": "cfg.CDD.VAL_DATA_PATH", "line_number": 27, "usage_type": "attribute"}, {"api_name": "cfg.CDD", "line_number": 27, "usage_type": "name"}, {"api_name": "cfg.CDD.VAL_LABEL_PATH", "line_number": 27, "usage_type": "attribute"}, {"api_name": "cfg.CDD.VAL_TXT_PATH", "line_number": 27, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 28, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "cfg.CDD.PRETRAIN_MODEL_PATH", "line_number": 33, "usage_type": "attribute"}, {"api_name": "cfg.CDD", "line_number": 33, "usage_type": "name"}, {"api_name": "model.siameseNet.d_aa", "line_number": 34, "usage_type": "name"}, {"api_name": "model.siameseNet.d_aa.SiameseNet", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 36, "usage_type": "call"}, {"api_name": "cfg.CDD.TRAINED_BEST_PERFORMANCE_CKPT", "line_number": 36, "usage_type": "attribute"}, {"api_name": "cfg.CDD", "line_number": 36, "usage_type": "name"}, {"api_name": "model.siameseNet.d_aa.load_state_dict", "line_number": 37, "usage_type": "call"}, {"api_name": "model.siameseNet.d_aa", "line_number": 37, "usage_type": "name"}, {"api_name": "torch.load", "line_number": 40, "usage_type": "call"}, {"api_name": "model.siameseNet.d_aa.init_parameters_from_deeplab", "line_number": 41, "usage_type": "call"}, {"api_name": "model.siameseNet.d_aa", "line_number": 41, "usage_type": "name"}, {"api_name": "model.siameseNet.d_aa", "line_number": 45, "usage_type": "name"}, {"api_name": "model.siameseNet.dares.SiameseNet", "line_number": 45, "usage_type": "call"}, {"api_name": "model.siameseNet.dares", "line_number": 45, "usage_type": "name"}, {"api_name": "torch.load", "line_number": 47, "usage_type": "call"}, {"api_name": "cfg.CDD.TRAINED_BEST_PERFORMANCE_CKPT", "line_number": 47, "usage_type": "attribute"}, {"api_name": "cfg.CDD", "line_number": 47, "usage_type": "name"}, {"api_name": "model.siameseNet.d_aa.load_state_dict", "line_number": 48, "usage_type": "call"}, {"api_name": "model.siameseNet.d_aa", "line_number": 48, "usage_type": "name"}, {"api_name": "model.siameseNet.d_aa", "line_number": 54, "usage_type": "name"}, {"api_name": "model.siameseNet.d_aa.cuda", "line_number": 54, "usage_type": "call"}, {"api_name": "layer.loss.ContrastiveLoss1", "line_number": 55, "usage_type": "call"}, {"api_name": "layer.loss", "line_number": 55, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "cfg.CDD.SAVE_PRED_PATH", "line_number": 56, "usage_type": "attribute"}, {"api_name": "cfg.CDD", "line_number": 56, "usage_type": "name"}, {"api_name": "utils.utils.check_dir", "line_number": 57, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 57, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "utils.utils.check_dir", "line_number": 61, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 61, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 65, "usage_type": "attribute"}, {"api_name": "model.siameseNet.d_aa.parameters", "line_number": 65, "usage_type": "call"}, {"api_name": "model.siameseNet.d_aa", "line_number": 65, "usage_type": "name"}, {"api_name": "cfg.CDD.INIT_LEARNING_RATE", "line_number": 65, "usage_type": "attribute"}, {"api_name": "cfg.CDD", "line_number": 65, "usage_type": "name"}, {"api_name": "cfg.CDD.DECAY", "line_number": 65, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 68, "usage_type": "call"}, {"api_name": "utils.utils.adjust_learning_rate", "line_number": 72, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 72, "usage_type": "name"}, {"api_name": "cfg.CDD.INIT_LEARNING_RATE", "line_number": 72, "usage_type": "attribute"}, {"api_name": "cfg.CDD", "line_number": 72, "usage_type": "name"}, {"api_name": "model.siameseNet.d_aa.train", "line_number": 73, "usage_type": "call"}, {"api_name": "model.siameseNet.d_aa", "line_number": 73, "usage_type": "name"}, {"api_name": "model.siameseNet.d_aa", "line_number": 78, "usage_type": "call"}, {"api_name": "utils.utils.rz_label", "line_number": 82, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 82, "usage_type": "name"}, {"api_name": "utils.utils.rz_label", "line_number": 83, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 83, "usage_type": "name"}, {"api_name": "utils.utils.rz_label", "line_number": 84, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 84, "usage_type": "name"}, {"api_name": "model.siameseNet.d_aa.eval", "line_number": 102, "usage_type": "call"}, {"api_name": "model.siameseNet.d_aa", "line_number": 102, "usage_type": "name"}, {"api_name": "funcs.validate", "line_number": 103, "usage_type": "call"}, {"api_name": "model.siameseNet.d_aa", "line_number": 103, "usage_type": "argument"}, {"api_name": "cfg.CDD.TRANSFROM_SCALES", "line_number": 103, "usage_type": "attribute"}, {"api_name": "cfg.CDD", "line_number": 103, "usage_type": "name"}, {"api_name": "torch.save", "line_number": 105, "usage_type": "call"}, {"api_name": "model.siameseNet.d_aa.state_dict", "line_number": 105, "usage_type": "call"}, {"api_name": "model.siameseNet.d_aa", "line_number": 105, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 105, "usage_type": "call"}, {"api_name": "os.path", "line_number": 105, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 106, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 106, "usage_type": "call"}, {"api_name": "os.path", "line_number": 106, "usage_type": "attribute"}, {"api_name": "funcs.validate", "line_number": 108, "usage_type": "call"}, {"api_name": "model.siameseNet.d_aa", "line_number": 108, "usage_type": "argument"}, {"api_name": "cfg.CDD.TRANSFROM_SCALES", "line_number": 108, "usage_type": "attribute"}, {"api_name": "cfg.CDD", "line_number": 108, "usage_type": "name"}, {"api_name": "torch.save", "line_number": 110, "usage_type": "call"}, {"api_name": "model.siameseNet.d_aa.state_dict", "line_number": 110, "usage_type": "call"}, {"api_name": "model.siameseNet.d_aa", "line_number": 110, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 110, "usage_type": "call"}, {"api_name": "os.path", "line_number": 110, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path", "line_number": 111, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 114, "usage_type": "call"}, {"api_name": "model.siameseNet.d_aa.state_dict", "line_number": 114, "usage_type": "call"}, {"api_name": "model.siameseNet.d_aa", "line_number": 114, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path", "line_number": 114, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 115, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 116, "usage_type": "call"}, {"api_name": "torch.backends", "line_number": 121, "usage_type": "attribute"}]} +{"seq_id": "585527886", "text": "import matplotlib.pyplot as plt\r\nimport numpy as np\r\nfrom DLCVDatasets import get_dataset\r\n\r\nif __name__ == \"__main__\":\r\n\r\n x_train, y_train, x_test, y_test, class_names = get_dataset( 'cifar10', range(10), training_size=1000, test_size=100 )\r\n\r\n # since the range of the images is 0 to 255, we reduce it to -0.5 to 0.5\r\n x_mean = 127.\r\n x_std = 255.\r\n x_train = (x_train - x_mean) / x_std\r\n x_test = (x_test - x_mean) / x_std\r\n\r\n data = x_train\r\n\r\n num_features = np.prod( x_train[0].shape ) # size of vector to classify\r\n num_classes = 10 # number of classes in CIFAR\r\n\r\n W = np.zeros([num_features, num_classes], np.float32) \r\n\r\n plt.figure()\r\n for class_chosen in range(10):\r\n\r\n # average image as a direction vector for linear classification\r\n theta = np.mean( x_train[y_train == class_chosen], axis = 0 ) \r\n\r\n W[:, class_chosen] = theta.flatten()\r\n\r\n # reshape and show average image vector\r\n plt.subplot(2, 5, class_chosen + 1)\r\n plt.imshow( ( (theta * x_std) + x_mean ).astype(np.uint8) ) # recast to uint8 to properly visualize the image\r\n plt.title(class_names[class_chosen])\r\n plt.show()\r\n\r\n # output training and test performance\r\n print('Training performance')\r\n print('Overall training classification rate: %5.2f' % np.mean((np.argmax(np.matmul(np.reshape(x_train, [x_train.shape[0], -1]), W), axis=1) == y_train).astype(np.float32)))\r\n for class_chosen in range(10):\r\n curr_data_vec = x_train[y_train == class_chosen].reshape( [-1, num_features] )\r\n print('Classification training rate (%s): %5.2f' % ( class_names[class_chosen], np.mean((np.argmax(np.matmul(curr_data_vec, W), axis=1) == class_chosen).astype(np.float32)) ) )\r\n\r\n print('Test performance')\r\n print('Overall test classification rate: %5.2f' % np.mean((np.argmax(np.matmul(np.reshape(x_test, [x_test.shape[0], -1]), W), axis=1) == y_test).astype(np.float32)))\r\n for class_chosen in range(10):\r\n curr_test_data_vec = x_test[y_test == class_chosen].reshape( [-1, num_features] )\r\n print('Classification test rate (%s): %5.2f' % ( class_names[class_chosen], np.mean((np.argmax(np.matmul(curr_test_data_vec, W), axis=1) == class_chosen).astype(np.float32)) ) )\r\n\r\n", "sub_path": "CV_DL/exercise02/exercise_02_linear_classification.py", "file_name": "exercise_02_linear_classification.py", "file_ext": "py", "file_size_in_byte": 2277, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "DLCVDatasets.get_dataset", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.prod", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 20, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 32, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.title", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 44, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 47, "usage_type": "attribute"}]} +{"seq_id": "479811929", "text": "from typing import Callable, Optional\n\nfrom pyconsoleapp import Component, builtin_components\n\n\nclass HeaderComponent(Component):\n \"\"\"Page Header. Includes title bar, navigation bar and message bar.\"\"\"\n\n _template = u'''{title_bar}\n{nav_bar}\n{single_hr}\n{message_bar}'''\n\n def __init__(self, **kwds):\n super().__init__(**kwds)\n self._title_bar = self.use_component(builtin_components.TitleBarComponent)\n self._nav_options = self.use_component(builtin_components.NavOptionsComponent)\n self._message_bar = self.use_component(builtin_components.MessageBarComponent)\n\n def printer(self, **kwds) -> str:\n return self._template.format(\n title_bar=self._title_bar.printer(),\n nav_bar=self._nav_options.printer(),\n single_hr=self.single_hr,\n message_bar=self._message_bar.printer()\n )\n\n def configure(self, go_back: Optional[Callable[[], None]] = None, **kwds):\n \"\"\"Configures the header component instnace.\"\"\"\n if go_back is not None:\n self._nav_options.configure(go_back=go_back)\n super().configure(**kwds)\n", "sub_path": "pyconsoleapp/builtin_components/header_component.py", "file_name": "header_component.py", "file_ext": "py", "file_size_in_byte": 1134, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pyconsoleapp.Component", "line_number": 6, "usage_type": "name"}, {"api_name": "pyconsoleapp.builtin_components.TitleBarComponent", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pyconsoleapp.builtin_components", "line_number": 16, "usage_type": "name"}, {"api_name": "pyconsoleapp.builtin_components.NavOptionsComponent", "line_number": 17, "usage_type": "attribute"}, {"api_name": "pyconsoleapp.builtin_components", "line_number": 17, "usage_type": "name"}, {"api_name": "pyconsoleapp.builtin_components.MessageBarComponent", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pyconsoleapp.builtin_components", "line_number": 18, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 28, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 28, "usage_type": "name"}]} +{"seq_id": "161762074", "text": "'''\nconverter.py [source_file] [output_file] [options]\n'''\n\n\nimport os\nimport sys\nimport json\n\n\nfrom utils import *\nfrom Parser import *\nfrom globalDefine import *\n\noption_triangulate = True\noption_forced_y_up = False\n\n\n\n\n\ndef write_file(filepath, content):\n index = filepath.rfind('/')\n dir = filepath[0:index]\n\n if not os.path.exists(dir):\n os.makedirs(dir)\n\n out = open(filepath, \"w\")\n out.write(content.encode('utf8', 'replace'))\n out.close()\n\n\ndef triangulate_node_hierarchy(node):\n node_attribute = node.GetNodeAttribute();\n\n if node_attribute:\n if node_attribute.GetAttributeType() == FbxNodeAttribute.eMesh or \\\n node_attribute.GetAttributeType() == FbxNodeAttribute.eNurbs or \\\n node_attribute.GetAttributeType() == FbxNodeAttribute.eNurbsSurface or \\\n node_attribute.GetAttributeType() == FbxNodeAttribute.ePatch:\n converter.Triangulate(node.GetNodeAttribute(), True);\n\n child_count = node.GetChildCount()\n for i in range(child_count):\n triangulate_node_hierarchy(node.GetChild(i))\n\ndef triangulate_scene(scene):\n node = scene.GetRootNode()\n if node:\n for i in range(node.GetChildCount()):\n triangulate_node_hierarchy(node.GetChild(i))\n\n\n\n\n\n\nif __name__ == \"__main__\":\n from optparse import OptionParser\n\n # try:\n from FbxCommon import *\n\n usage = \"Usage: %prog [source_file.fbx] [output_file.json] [options]\"\n parser = OptionParser(usage=usage)\n\n\n parser.add_option('-t', '--triangulate', action='store_true', dest='triangulate', help=\"force quad geometry into triangles\", default=True)\n parser.add_option('-y', '--force-y-up', action='store_true', dest='forceyup', help=\"ensure that the y axis shows up\", default=False)\n\n\n (options, args) = parser.parse_args()\n\n option_triangulate = options.triangulate\n option_forced_y_up = options.forceyup\n\n\n manager, scene = InitializeSdkObjects()\n\n sdkManager = manager\n\n converter = FbxGeometryConverter(manager)\n\n\n\n # The converter takes an FBX file as an argument.\n if len(args) > 1:\n print(\"\\nLoading file: %s\" % args[0])\n\n result = LoadScene(manager, scene, args[0])\n else:\n result = False\n print(\"\\nUsage: converter.py [source_file.fbx] [output_file.json]\\n\")\n\n\n if not result:\n print(\"\\nAn error occurred while loading the file...\")\n else:\n if option_triangulate:\n print(\"\\nForcing geometry to triangles\")\n triangulate_scene(scene)\n\n\n # axis_system = FbxAxisSystem.MayaYUp\n #\n # if not option_forced_y_up:\n # # According to asset's coordinate to convert scene\n # upVector = scene.GetGlobalSettings().GetAxisSystem().GetUpVector()\n # if upVector[0] == 3:\n # axis_system = FbxAxisSystem.MayaZUp\n #\n # axis_system.ConvertScene(scene)\n\n\n\n\n inputFolder = args[0].replace( \"\\\\\", \"/\" );\n index = args[0].rfind( \"/\" );\n inputFolder = inputFolder[:index]\n\n outputFolder = args[1].replace( \"\\\\\", \"/\" );\n index = args[1].rfind( \"/\" );\n outputFolder = outputFolder[:index]\n\n\n\n output_content = Parser(converter).parse(scene, os.path.join(os.getcwd(), args[0]))\n\n output_string = json.dumps(output_content, separators=(',', ': '), sort_keys=True)\n\n\n output_path = os.path.join(os.getcwd(), args[1])\n write_file(output_path, output_string)\n\n # Destroy all objects created by the FBX SDK.\n manager.Destroy()\n sys.exit(0)\n\n\n", "sub_path": "tool/converter/src/fbx/python/converter.py", "file_name": "converter.py", "file_ext": "py", "file_size_in_byte": 3613, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "os.path.exists", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 27, "usage_type": "call"}, {"api_name": "optparse.OptionParser", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 128, "usage_type": "call"}, {"api_name": "os.path", "line_number": 128, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 128, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 130, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 133, "usage_type": "call"}, {"api_name": "os.path", "line_number": 133, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 133, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 138, "usage_type": "call"}]} +{"seq_id": "9617828", "text": "from flask_restful import Resource, reqparse\nfrom flask import request\nfrom com_dayoung_api.cop.act.model.actor_ai import ActorAi\nimport json\nclass Chatbot(Resource):\n @staticmethod\n def post():\n print(\"들어옴\")\n ai = ActorAi()\n args = request.get_json()\n print(args)\n args = [args[i]['value'] for i in args.keys()]\n print(args)\n name = ai.train_actors(args)\n print(name)\n return name", "sub_path": "api/com_dayoung_api/cop/cht/resource/chatbot.py", "file_name": "chatbot.py", "file_ext": "py", "file_size_in_byte": 456, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "flask_restful.Resource", "line_number": 5, "usage_type": "name"}, {"api_name": "com_dayoung_api.cop.act.model.actor_ai.ActorAi", "line_number": 9, "usage_type": "call"}, {"api_name": "flask.request.get_json", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 10, "usage_type": "name"}]} +{"seq_id": "600374012", "text": "from data import get_conn, get_cursor\nfrom flask import Flask, render_template, request, redirect, url_for\nimport pymysql\nfrom dotenv import load_dotenv\nimport os\n\nload_dotenv()\n\nconn = get_conn(\"localhost\",\n os.environ.get(\"DB_USER\"),\n os.environ.get(\"DB_PASSWORD\"),\n \"Chinook\"\n )\n\ncursor = get_cursor(conn)\n\napp = Flask(__name__)\n\n@app.route('/genre/create')\ndef show_create_artist_form():\n return render_template('create_genre.template.html')\n\n\n# @app.route('/genre/create', methods=[\"POST\"])\n# def process_create_artist():\n# genre_name = request.form.get(\"genre_name\")\n\n# sql = f\"\"\"\n# insert into Genre (Name) values (\"{genre_name}\");\n# \"\"\"\n\n# cursor.execute(sql)\n\n\n# # Whenever we change the database, we have to remember to COMMIT\n# # the transactions\n# conn.commit()\n\n# return \"genre added\"\n\n\n\n# \"magic code\" -- boilerplate\nif __name__ == '__main__':\n app.run(host=os.environ.get('IP'),\n port=int(os.environ.get('PORT')),\n debug=True)\n", "sub_path": "examples/creategenre/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1066, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "dotenv.load_dotenv", "line_number": 7, "usage_type": "call"}, {"api_name": "data.get_conn", "line_number": 9, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 10, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 11, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 11, "usage_type": "attribute"}, {"api_name": "data.get_cursor", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 21, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 45, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 46, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 46, "usage_type": "attribute"}]} +{"seq_id": "276251831", "text": "import Files.functions as fc\nimport json\nimport pandas as pd\n#import datetime\nimport progressbar\nfrom tqdm import tqdm\nimport re\nfrom collections import Counter\nimport numpy as np\nfrom sklearn.cluster import KMeans as km\nfrom sklearn import tree\n\n\nresult1 = pd.read_csv(filepath_or_buffer=\"C://Users/zhang/Desktop/anaconda/pomdp_eole/extraction_results_23133.csv\")\nresult1.index = pd.Series(range(1, len(result1)+1))# 0123456 as index\ndata1 = pd.read_csv(filepath_or_buffer=\"C://Users/zhang/Desktop/anaconda/pomdp_eole/extraction_events_23133.csv\")\n# data1.drop([\"courseId\"])\n# courseId.drop([\"courseId\"])\nprint(\"Reading csv, JSON files and normalise them in DataFrame as input\")\nallEventList, finalExamList, resultExamList = fc.readJsonToList()\n\nprint(\"Add event time to results from original JSON file\")\nresult1 = fc.addTimeToResult(result1, resultExamList)\nUserId = list(set(data1[\"userId\"]))\n\n# 这个循环是带表头的,而表头中没有数字的存在,所以我们要从1开始循环。\nfor i in (range(1, len(result1))):\n result1.loc[i, 'examName'] = int(re.findall(r\"\\d+\\.?\\d*\", result1.loc[i]['examName'])[0])\n\nd1 = data1.copy(deep=True)\nr1 = result1.copy(deep=True)\n\n# I want to use this rename, but the interpreter don't know this function :/\n# r1.rename(columns={r1.columns{'examName': \"resourceId\"}), inplace=True)\nr1[\"resourceId\"] = r1['examName']\nr1 = r1.drop(columns=['examName'])\n\nd1 = d1.drop(columns=['resourceType', 'resourceName', 'resourceDesc'])\n\nr1 = r1.drop(index=0)\nr1['eventTime'].loc[-1] = '2018-10-02 21:35:05.755000'\nd1['score'] = 0\nd1['scoreMax'] = 0\nr1['action'] = \"Submitted\"\n\n# Drop the 'submitted' in study record because each visit of quiz has a unique resourceId, even the same resource\nd1 = (d1[d1['action'].isin(['Viewed'])])\nd1 = d1[['userId', 'courseId', 'score', 'scoreMax', 'eventTime', 'resourceId']]\n\n#Here, scoreMax is 18\nFinalExam = fc.treatFinalExam(finalExamList)\nFinalExam['courseId'] = d1.loc[0, 'courseId']\nFinalExam['scoreMax'] = 18\nFinalExam['eventTime'] = \"2019-10-02 21:35:05.755000\"\nFinalExam['resourceId'] = 666666 #标号我自己取的\nFinalExam = FinalExam[['userId', 'courseId', 'score', 'scoreMax', 'eventTime', 'resourceId']]\n\n\"\"\"\n 合成整体的计算数据,并按照学生的成绩类别进行分类\n\"\"\"\nframes = [d1, r1, FinalExam]\ndf_total = pd.concat(frames, sort=True)\nlabel = []\nq1TB = []\nq1PASS = []\nq1FAIL = []\nq2TB = []\nq2PASS = []\nq2FAIL = []\n\n\"\"\"Here is for the quiz part\"\"\"\nfor i in range(len(r1)):\n if r1.iloc[i]['score']>=8:\n label.append('TB')\n elif r1.iloc[i]['score']<8 and r1.iloc[i]['score']>=6:\n label.append('PASS')\n else:\n label.append('FAIL')\nr1['label'] = label\n\n\"\"\"\nHere is for the Final Exam Part\n\"\"\"\n\nFinalExam_label = []\nFinalTB = []\nFinalPASS = []\nFinalFAIL = []\nfor i in (FinalExam.score.astype(float)):\n if i >= 14:\n FinalExam_label.append('TB')\n elif i<14 and i>=8:\n FinalExam_label.append('PASS')\n else :\n FinalExam_label.append('FAIL')\nFinalExam['label'] = FinalExam_label\n\nQ1 = r1[r1['resourceId'] == 83669]\nQ2 = r1[r1['resourceId'] == 84682]\nq1 = []\nq2 = []\nfinal = []\nfor i in UserId:\n if i in list(Q1['userId']) and i in list(Q2['userId']) and i in list(FinalExam['userId']):\n\n a = float(Q1[Q1['userId'] == i].get('score'))\n q1.append(a)\n q2.append(float(Q2[Q2['userId'] == i].get('score')))\n final.append(float(FinalExam[FinalExam['userId'] == i].get('score')))\n\n\nd = {'userId':\"\" , 'score_Q1': \"\", 'scoreQ2': \"\", 'score_final': \"\"}\n\n\n\nx = np.array([q1, q2, final])\nx = x.reshape(len(x[0]), 3)\nkmeans = km(n_clusters=3, random_state=0).fit(x)\n\n# kmeans.labels_ 这个是分类好了的表格\n#quizAndExam KMEANS 效果贼差,决定不用了。\nquizAndExam = pd.DataFrame(data=x, columns={\"q1Score\", \"q2Score\", \"finalScore\"})\n\nquizAndExam['label'] = kmeans.labels_\n\nquizAndExam.to_csv(path_or_buf=\"成绩和分类.csv\")\nquizAndExam.to_excel('output1.xlsx', engine='xlsxwriter')\n\"\"\"\n三个学生的成绩等级。TB20个,B36个,F47个。咱大学的学生水平不行啊。\n\"\"\"\nTB = []\nB = []\nF = []\nfor i in range(len(FinalExam)):\n if FinalExam.iloc[i][\"label\"] == 'TB':\n TB.append(FinalExam.loc[i]['userId'])\n if FinalExam.iloc[i][\"label\"] == 'PASS':\n B.append(FinalExam.loc[i]['userId'])\n if FinalExam.iloc[i]['label'] == 'FAIL':\n F.append(FinalExam.loc[i]['userId'])\n\n#######################################################################################################################\n# 计算一下用所有数据的转移矩阵\n# 肯定是按照学生分的,先按照每个学生的情况将转移的次数加到矩阵里,然后在做那个矩阵除法就好了\n# 我们一共有117个学生,按照每个学生的名字做最外层循环。\n# 里面是transition matrix一样的计算过程\n\ntoCalculateMatrix = df_total[['eventTime', 'resourceId']].copy(deep=True)\ntoCalculateMatrix = toCalculateMatrix.drop(index=156)\ntoCalculateMatrix = toCalculateMatrix.sort_values(by='eventTime')\nnumIndex = 0\nresourceIndex = {}\n\n\n###########################################################################\n#存储resource的index编号,每个resource我都按照\n# 这一步能够简化计算action转移概率的 复杂度,非常重要。\nfor i in toCalculateMatrix.resourceId:\n if i not in resourceIndex.keys():\n resourceIndex[i] = numIndex\n numIndex += 1\n\"\"\"\n我们的数据将所有的学习记录都放在了一起,但是实际中的例子不是这样的。\n我以学生为单位,把每个学生的学习轨迹单独的计算出来,能够去掉影响resource转移的因素\n\"\"\"\ndf_tb = pd.DataFrame(columns=df_total.columns)\ndf_b = pd.DataFrame(columns=df_total.columns)\ndf_f = pd.DataFrame(columns=df_total.columns)\n\nfor i in TB:\n df_tb = pd.concat([df_tb, df_total[df_total['userId']==i]])\n\nfor i in B:\n df_b = pd.concat([df_b, df_total[df_total['userId']==i]])\n\nfor i in F:\n df_f = pd.concat([df_f, df_total[df_total['userId']==i]])\n# 这块矩阵可以改成按照学生的成绩等级分类,一共三个等级,TB,B,F\nM_tb = np.array(fc.transition_diff(resourceIndex, df_tb))\nM_b = fc.transition_diff(resourceIndex, df_b)\nM_f = fc.transition_diff(resourceIndex, df_f)\nS = []\n\nfor i in range(len(M_tb)):\n S.append(i)\n\n\nM1 = np.zeros([38,38])\nfor x in range(len(M_tb[1:])):\n for y in range(len(M_tb[x][1:])):\n M1[x][y]=(M_tb[x][y])", "sub_path": "Files/treat_data.py", "file_name": "treat_data.py", "file_ext": "py", "file_size_in_byte": 6442, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pandas.read_csv", "line_number": 14, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 15, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 16, "usage_type": "call"}, {"api_name": "Files.functions.readJsonToList", "line_number": 20, "usage_type": "call"}, {"api_name": "Files.functions", "line_number": 20, "usage_type": "name"}, {"api_name": "Files.functions.addTimeToResult", "line_number": 23, "usage_type": "call"}, {"api_name": "Files.functions", "line_number": 23, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 28, "usage_type": "call"}, {"api_name": "Files.functions.treatFinalExam", "line_number": 51, "usage_type": "call"}, {"api_name": "Files.functions", "line_number": 51, "usage_type": "name"}, {"api_name": "pandas.concat", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 116, "usage_type": "call"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 118, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 122, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 166, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 167, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 168, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 171, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 174, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 179, "usage_type": "call"}, {"api_name": "Files.functions.transition_diff", "line_number": 179, "usage_type": "call"}, {"api_name": "Files.functions", "line_number": 179, "usage_type": "name"}, {"api_name": "Files.functions.transition_diff", "line_number": 180, "usage_type": "call"}, {"api_name": "Files.functions", "line_number": 180, "usage_type": "name"}, {"api_name": "Files.functions.transition_diff", "line_number": 181, "usage_type": "call"}, {"api_name": "Files.functions", "line_number": 181, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 188, "usage_type": "call"}]} +{"seq_id": "110419218", "text": "#!/usr/bin/env python3\n# Explore the PC partition space\nimport click\nimport csv\nimport unidecode\nimport itertools\nimport sys\nimport time\nimport heapq\nfrom multiprocessing import Pool\nfrom tqdm import tqdm\nfrom warnings import warn\nfrom math import ceil, factorial as fac\n\nBIDS_COEF = 2\nTOPICS_COEF = 1\nCITES_COEF = 1\n\n\"\"\"\nSplit the PC into parts and examine the consequences. Run with --help for\nan explanation\n\"\"\"\n\ndef eprint(*args, **kwargs):\n print(*args, file=sys.stderr, **kwargs)\n\ndef to_int(x):\n if x:\n return int(x)\n return 0\n\ndef csv_to_dicts(fileobj, schema=None, sanitize=lambda x:x):\n result = []\n reader = csv.DictReader(fileobj, schema)\n return [sanitize(d) for d in csv.DictReader(fileobj, schema)]\n\ndef sanitize_members(d):\n d[\"email\"] = d.get(\"email\").lower()\n return d\n\ndef sanitize_score_input(d):\n d[\"email\"] = d.get(\"email\").lower()\n d[\"paper\"] = int(d.get(\"paper\"))\n d[\"score\"] = float(d.get(\"score\"))\n d[\"preference\"] = float(d.get(\"preference\"))\n d[\"topic_score\"] = float(d.get(\"topic_score\"))\n d[\"citations\"] = float(d.get(\"citations\"))\n return d\n\ndef sanitize_seed_input(d):\n d[\"email\"] = d.get(\"email\").lower()\n d[\"split\"] = d.get(\"split\").lower()\n return d\n\ndef build_affinities_dict(affinity_report, valid_members):\n scores = csv_to_dicts(affinity_report, sanitize=sanitize_score_input)\n\n pc_members = set(e[\"email\"] for e in valid_members)\n papers = {e[\"paper\"] for e in scores}\n affinities = {(e[\"paper\"], e[\"email\"]):e[\"score\"] for e in scores}\n bids = {(e[\"paper\"], e[\"email\"]):e[\"preference\"] for e in scores}\n topics = {(e[\"paper\"], e[\"email\"]):e[\"topic_score\"] for e in scores}\n cites = {(e[\"paper\"], e[\"email\"]):e[\"citations\"] for e in scores}\n\n affinities[\"id\"] = \"affs\"\n bids[\"id\"] = \"bids\"\n topics[\"id\"] = \"topics\"\n cites[\"id\"] = \"cites\"\n\n return pc_members,papers,affinities,bids,topics,cites\n\ndef iter_partitions(pc, seed, n):\n k = int(ceil(len(pc) / 2))\n for combo in itertools.combinations(pc, k):\n yield Partition(Group(combo), Group(pc - set(combo)), seed, top_n=n)\n\ndef filter_out_seed(pc, seed):\n eprint(\"there are\", len(pc), \"in the pc\")\n pc = {a for a in pc if a not in seed}\n eprint(\"there are\", len(pc), \"assignable members\")\n k = int(ceil(len(pc) / 2))\n remaining = (fac(k * 2)) // ((fac(k)**2))\n eprint(\"trying\", remaining, \"possible partitions\")\n return remaining, pc\n\nclass Group:\n def __init__(self, members):\n self.members = list(members)\n self.members.sort()\n\n def __contains__(self, key):\n if key in self.members:\n return True\n\n def __iter__(self):\n return iter(self.members)\n\n def iter_scores(self, paper, affs, pred=lambda x: True):\n return (affs.get((paper,mem), 0) for mem in self \\\n if pred(affs.get((paper,mem), 0)))\n\n def top_n_scores(self, paper, affs, n):\n return heapq.nlargest(n, self.iter_scores(paper, affs))\n\n def score(self, paper, affs, pred=lambda x:True):\n return sum(self.iter_scores(paper, affs, pred))\n\nclass MemoizedGroup(Group):\n def __init__(self, members, top_n=None):\n super(MemoizedGroup, self).__init__(members)\n self.scores = {}\n self.top_n_scores_store = {}\n\n def top_n_scores(self, paper, affs, n):\n affs_id = affs.get(\"id\", None)\n if not (paper, n, affs_id) in self.top_n_scores_store:\n self.top_n_scores_store[(paper,n,affs_id)] = \\\n super(MemoizedGroup,self).top_n_scores(paper, affs, n)\n return self.top_n_scores_store[(paper,n,affs_id)]\n\n def score(self, paper, affs, pred=lambda x:True):\n if not paper in self.scores:\n self.scores[paper] = super(MemoizedGroup,self).score(paper, affs,\n pred)\n return self.scores[paper]\n\nclass MultiGroup:\n def __init__(self, *args, top_n=None):\n self.groups = list(args)\n self.papers = []\n self.top_n = top_n\n\n def assign(self, p):\n self.papers.append(p)\n\n def __iter__(self):\n return itertools.chain(*self.groups)\n\n def __str__(self):\n return str(list(self))\n\n def top_n_scores(self, paper, affs, n):\n it = (g.top_n_scores(paper, affs, self.top_n) for g in self.groups)\n return heapq.nlargest(self.top_n, itertools.chain(*it))\n\n def score(self, paper, affs, pred=lambda x:True, tri_score=False):\n if self.top_n:\n if tri_score:\n bids = BIDS_COEF * \\\n sum(self.top_n_scores(paper, affs[0], self.top_n))\n topics = TOPICS_COEF * \\\n sum(self.top_n_scores(paper, affs[1], self.top_n))\n cites = CITES_COEF * \\\n sum(self.top_n_scores(paper, affs[2], self.top_n))\n if bids > self.top_n * BIDS_COEF:\n raise \"BIDs too large somewhere\"\n if topics > self.top_n * TOPICS_COEF:\n raise \"TOPICS too large somewhere\"\n if topics > self.top_n * CITES_COEF:\n raise \"CITES too large somewhere\"\n tmp = BIDS_COEF + TOPICS_COEF + CITES_COEF\n return (bids + topics + cites) / (tmp * self.top_n)\n else:\n return sum(self.top_n_scores(paper, affs, self.top_n))\n else:\n return sum(g.score(paper, affs, pred) for g in self.groups)\n\nclass Partition:\n def __init__(self, groupA, groupB, seed, top_n=None, save_papers=False):\n self.save_papers = save_papers\n if seed:\n self.groupA = MultiGroup(groupA, seed.groupA.groups[0], top_n=top_n)\n self.groupB = MultiGroup(groupB, seed.groupB.groups[0], top_n=top_n)\n else:\n self.groupA = MultiGroup(groupA, top_n=top_n)\n self.groupB = MultiGroup(groupB, top_n=top_n)\n\n def __contains__(self, key):\n return key in self.groupA or key in self.groupB\n\n def __iter__(self):\n return itertools.chain(self.groupA, self.groupB)\n\n def __str__(self):\n return \"=====Group A======\\n\" + str(self.groupA) + \\\n \"\\n=====Group B======\\n\" + str(self.groupB)\n def score(self, papers, affs, pred=lambda x:True, assignments=False,\n tri_score=False):\n # get the score for each paper in each partition, add the max\n result = 0\n l = []\n\n for p in papers:\n sA = self.groupA.score(p, affs, pred, tri_score=tri_score)\n sB = self.groupB.score(p, affs, pred, tri_score=tri_score)\n if assignments:\n if sA > sB:\n l.append('a')\n else:\n l.append('b')\n if self.save_papers:\n if sA > sB:\n self.groupB.assign(p)\n else:\n self.groupA.assign(p)\n result += max(sA, sB)\n if assignments:\n return result,l\n else:\n return result\n\ndef build_seed_part(fileobj, save_papers=False, top_n=None):\n if fileobj:\n d = csv_to_dicts(fileobj, sanitize=sanitize_seed_input)\n groupA = MemoizedGroup(e[\"email\"] for e in d if e[\"split\"] == 'a')\n groupB = MemoizedGroup(e[\"email\"] for e in d if e[\"split\"] == 'b')\n else:\n groupA = MemoizedGroup([])\n groupB = MemoizedGroup([])\n\n return Partition(groupA, groupB, None, top_n=top_n, save_papers=save_papers)\n\ndef true_pred(x):\n return True\n\ndef pos_pred(x):\n return x > 0\n\n# need this instead of closure for multiprocess\nclass PartitionProcessor:\n def __init__(self, papers, affinities, pred=true_pred, tri_score=False):\n self.papers = papers\n self.affinities = affinities\n self.pred = pred\n self.tri_score = tri_score\n def __call__(self, part):\n try:\n return part.score(self.papers, self.affinities, self.pred,\n tri_score=self.tri_score), part\n except KeyboardInterrupt:\n return None\n\n@click.group()\n@click.argument(\"pc-names\", type=click.File('r'))\n@click.argument(\"affinity-report\", type=click.File('r'))\n@click.pass_context\ndef cli(ctx, pc_names, affinity_report):\n \"\"\"\n Run with COMMAND --help to get the details for each mode every mode\n requires the following file names as arguments.\n\n PC_NAMES: a csv_file, each row is a pc member must have at least columns\n with the headers: \"first\",\"last\",\"email\",\"affiliation\" this can be\n downloaded from hotcrp\n\n AFFINITY_REPORT: the output of pc_paper_scores.py\n \"\"\"\n valid_emails = csv_to_dicts(pc_names, sanitize=sanitize_members)\n pc_members, papers, affinities,bids,topics,cites = build_affinities_dict(affinity_report, valid_emails)\n ctx.obj[\"pc_members\"] = pc_members\n ctx.obj[\"papers\"] = papers\n ctx.obj[\"affinities\"] = affinities\n ctx.obj[\"pc_info\"] = valid_emails\n ctx.obj[\"bids\"] = bids\n ctx.obj[\"topic_scores\"] = topics\n ctx.obj[\"cites\"] = cites\n\ntri_score_help = \"use the three scores in the affinity report rather than \" + \\\n \"just the combined value (uses the coefficients provided \" + \\\n \"in the top of this file)\"\n\n#pid,pc-email,affinity-score\n@click.command()\n@click.option(\"--seed-partition\", type=click.File('r'),\n help=\"csv file defining and initial fixed partition\")\n@click.option(\"-n\", type=int,\n help=\"if given only consider the best n reviewers\")\n@click.option(\"--full-report\", type=click.File('w'),\n help=\"dump every partition and score into this file\")\n@click.option(\"--positive-only/--no-positive-only\", default=False,\n help=\"only consider positive scores\")\n@click.option(\"--tri-score/--no-tri-score\", default=False,\n help=tri_score_help)\n@click.option(\"-j\", type=int, help=\"number of worker processes\")\n@click.pass_context\ndef search(ctx, seed_partition, n, full_report, j, tri_score, positive_only):\n \"\"\"\n exhaustively search all possible combinations of pc_members for an even\n partition and output the top scoring partition to stdout\n\n it's useful to provide a seed for the partition to limit the search space,\n a seed is a csv file with two columns, \"split\" and \"email\", where split\n contains either 'a' or 'b' and \"email\" is the email of a pc member\n \"\"\"\n seed_part = build_seed_part(seed_partition, top_n=n)\n pc_members = ctx.obj[\"pc_members\"]\n papers = ctx.obj[\"papers\"]\n score_pred = true_pred\n if positive_only:\n score_pred = pos_pred\n if tri_score:\n affinities = (ctx.obj[\"bids\"],\n ctx.obj[\"topic_scores\"],\n ctx.obj[\"cites\"])\n else:\n affinities = ctx.obj[\"affinities\"]\n\n remaining, filtered_emails = filter_out_seed(pc_members, seed_part)\n partitions = iter_partitions(filtered_emails, seed_part, n)\n score_part = PartitionProcessor(papers, affinities, score_pred, tri_score=tri_score)\n best_part = next(partitions)\n best_score,_ = score_part(best_part)\n\n if full_report:\n writer = csv.writer(full_report)\n writer.writerow([\"score\",\"partition\"])\n writer.writerow((best_score, best_part))\n\n pool = Pool(j)\n\n start = time.time()\n try:\n with tqdm(total=remaining) as pbar:\n for score,part in pool.imap_unordered(score_part, partitions, 1000):\n pbar.update(1)\n if score > best_score:\n best_score = score\n best_part = part\n if full_report:\n writer.writerow((score, part))\n except KeyboardInterrupt:\n eprint(\"interrupted!!!!\")\n end = time.time()\n\n eprint(\"took\", end - start, \"seconds\")\n eprint(\"best score is:\", best_score, \"produced by this split:\\n\", best_part)\n\n email_to_name = {p[\"email\"].lower():p for p in ctx.obj[\"pc_info\"]}\n for p in best_part.groupA:\n email_to_name[p][\"split\"] = 'a'\n for p in best_part.groupB:\n email_to_name[p][\"split\"] = 'b'\n writer = csv.DictWriter(sys.stdout, [\"first\",\"last\",\"email\",\"affiliation\",\"split\"])\n writer.writeheader()\n for p in best_part:\n writer.writerow(email_to_name[p.lower()])\n\n@click.command()\n@click.option(\"-n\", type=int,\n help=\"if given only consider the best n reviewers\")\n@click.option(\"--positive-only/--no-positive-only\", default=False,\n help=\"only consider positive scores\")\n@click.argument(\"part_csv\", type=click.File('r'))\n@click.pass_context\ndef examine(ctx, part_csv, n, positive_only):\n \"\"\"\n Examine a partition; count the papers assigned to each, regenerate the score\n\n PART_CSV a csv file generated by the \"search\" mode of this script\n \"\"\"\n papers = ctx.obj[\"papers\"]\n affinities = ctx.obj[\"affinities\"]\n part = build_seed_part(part_csv, True, top_n=n)\n\n score_pred = true_pred\n if positive_only:\n score_pred = pos_pred\n\n score = part.score(papers, affinities, score_pred)\n\n print(\"Partition:\\n\", part)\n print(\"SCORE is\", score)\n print(\"Count of papers:\")\n print(\"GroupA:\", len(part.groupA.papers), \"GroupB:\",\n len(part.groupB.papers))\n\n@click.command()\n@click.option(\"-n\", type=int,\n help=\"if given only consider the best n reviewers\")\n@click.option(\"--positive-only/--no-positive-only\", default=False,\n help=\"only consider positive scores\")\n@click.pass_context\ndef total(ctx, n, positive_only):\n \"\"\"\n Print the total possible affinity. Useful for judging how much partitioning\n hurts the assignments.\n \"\"\"\n papers = ctx.obj[\"papers\"]\n affinities = ctx.obj[\"affinities\"]\n pc_members = ctx.obj[\"pc_members\"]\n\n score_pred = true_pred\n if positive_only:\n score_pred = pos_pred\n\n group = Group(pc_members)\n\n if not n:\n score = sum(group.score(p, affinities, score_pred) for p in papers)\n else:\n score = sum(sum(group.top_n_scores(p, affinities, n)) for p in papers)\n\n print(len(pc_members), \"reviewers, and\", len(papers), \"papers\")\n print(\"Total affinity is\", score)\n\n@click.command()\n@click.option(\"-n\", type=int,\n help=\"if given only consider the best n reviewers\")\n@click.option(\"--positive-only/--no-positive-only\", default=False,\n help=\"only consider positive scores\")\n@click.argument(\"part_csv\", type=click.File('r'))\n@click.option(\"--tri-score/--no-tri-score\", default=False,\n help=tri_score_help)\n@click.pass_context\ndef papers(ctx, part_csv, n, positive_only, tri_score):\n \"\"\"\n For a particular partition (PART_CSV), generate a list of papers, what\n partition they're in and the penalty for being in that partition (where\n penalty is the difference between that partition's affinity and the total PC\n affinity for that paper)\n\n PART_CSV a csv file generated by the \"search\" mode of this script\n \"\"\"\n papers = ctx.obj[\"papers\"]\n pc_members = ctx.obj[\"pc_members\"]\n part = build_seed_part(part_csv, True, top_n=n)\n full_pc = MultiGroup(Group(pc_members), top_n=n)\n if tri_score:\n affinities = (ctx.obj[\"bids\"],\n ctx.obj[\"topic_scores\"],\n ctx.obj[\"cites\"])\n else:\n affinities = ctx.obj[\"affinities\"]\n\n score_pred = true_pred\n if positive_only:\n score_pred = pos_pred\n\n writer = csv.writer(sys.stdout)\n writer.writerow([\"paper\",\"part\",\"penalty\"])\n for p in papers:\n part_score,l = part.score([p], affinities, score_pred, assignments=True,\n tri_score=tri_score)\n total_score = full_pc.score(p, affinities, score_pred,\n tri_score=tri_score)\n if part_score > 1:\n eprint(\"partscore for paper\", p, \"is over 1:\", part_score)\n if total_score > 1:\n eprint(\"totalscore for paper\", p, \"is over 1:\", total_score)\n\n writer.writerow([p, l[0], int((total_score - part_score) * 1000)])\n\ncli.add_command(search)\ncli.add_command(examine)\ncli.add_command(total)\ncli.add_command(papers)\nif __name__ == '__main__':\n cli(obj={})\n", "sub_path": "explore_partition.py", "file_name": "explore_partition.py", "file_ext": "py", "file_size_in_byte": 16256, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "sys.stderr", "line_number": 25, "usage_type": "attribute"}, {"api_name": "csv.DictReader", "line_number": 34, "usage_type": "call"}, {"api_name": "csv.DictReader", "line_number": 35, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 73, "usage_type": "call"}, {"api_name": "itertools.combinations", "line_number": 74, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 81, "usage_type": "call"}, {"api_name": "math.factorial", "line_number": 82, "usage_type": "call"}, {"api_name": "heapq.nlargest", "line_number": 103, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 137, "usage_type": "call"}, {"api_name": "heapq.nlargest", "line_number": 144, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 144, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 182, "usage_type": "call"}, {"api_name": "click.group", "line_number": 243, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 244, "usage_type": "call"}, {"api_name": "click.File", "line_number": 244, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 245, "usage_type": "call"}, {"api_name": "click.File", "line_number": 245, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 246, "usage_type": "attribute"}, {"api_name": "csv.writer", "line_number": 315, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 319, "usage_type": "call"}, {"api_name": "time.time", "line_number": 321, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 323, "usage_type": "call"}, {"api_name": "time.time", "line_number": 333, "usage_type": "call"}, {"api_name": "csv.DictWriter", "line_number": 343, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 343, "usage_type": "attribute"}, {"api_name": "click.command", "line_number": 273, "usage_type": "call"}, {"api_name": "click.option", "line_number": 274, "usage_type": "call"}, {"api_name": "click.File", "line_number": 274, "usage_type": "call"}, {"api_name": "click.option", "line_number": 276, "usage_type": "call"}, {"api_name": "click.option", "line_number": 278, "usage_type": "call"}, {"api_name": "click.File", "line_number": 278, "usage_type": "call"}, {"api_name": "click.option", "line_number": 280, "usage_type": "call"}, {"api_name": "click.option", "line_number": 282, "usage_type": "call"}, {"api_name": "click.option", "line_number": 284, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 285, "usage_type": "attribute"}, {"api_name": "click.command", "line_number": 348, "usage_type": "call"}, {"api_name": "click.option", "line_number": 349, "usage_type": "call"}, {"api_name": "click.option", "line_number": 351, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 353, "usage_type": "call"}, {"api_name": "click.File", "line_number": 353, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 354, "usage_type": "attribute"}, {"api_name": "click.command", "line_number": 377, "usage_type": "call"}, {"api_name": "click.option", "line_number": 378, "usage_type": "call"}, {"api_name": "click.option", "line_number": 380, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 382, "usage_type": "attribute"}, {"api_name": "csv.writer", "line_number": 439, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 439, "usage_type": "attribute"}, {"api_name": "click.command", "line_number": 406, "usage_type": "call"}, {"api_name": "click.option", "line_number": 407, "usage_type": "call"}, {"api_name": "click.option", "line_number": 409, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 411, "usage_type": "call"}, {"api_name": "click.File", "line_number": 411, "usage_type": "call"}, {"api_name": "click.option", "line_number": 412, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 414, "usage_type": "attribute"}]} +{"seq_id": "504699095", "text": "from os.path import exists, join\nimport numpy as np\nfrom os import makedirs\nimport shutil\nfrom scipy.signal import convolve2d\nimport cv2\n\n\nfrom Asbestos_Utils import name_list, crop_save_remove_from_list, generate_complete_flipping, save_names, \\\n load_names, get_file_extension,generate_complete_flipping\n\n\ndef pad_image(image):\n sizex = image.shape[1]\n sizey = image.shape[0]\n extrapixX = 1600-sizex\n extrapixY = 256-sizey\n left = extrapixX//2\n right = extrapixX-left\n top = extrapixY//2\n bottom = extrapixY-top\n image = cv2.copyMakeBorder(image, top, bottom, left, right, 0,value = [255,255,255])\n return image\n\n\n\n\nclass DataPantograph(object):\n \"\"\"\n Class that produces a split of the data into the percentages that are introduced\n It also produces crops of the specified size.\n \"\"\"\n\n def __init__(self, path_name, data_name,size, data_augmentation=False, train_percentage=0.63, val_percentage=0.07,\n test_percentage=0.3,\n load_extension='.jpg',augmentation = False):\n\n self.path_name = path_name # Path were the data is stored\n self.data_name = data_name # Name given to the splitted data\n self.augmentation = int(data_augmentation)\n self.load_extension = load_extension\n self.size = size\n self.augmentation = augmentation\n # Percentages for train validation and test\n self.train_percentage = train_percentage # Train percentage\n self.val_percentage = val_percentage # Validation percentage\n self.test_percentage = test_percentage # Test percentage\n\n self.temp_train = './' + self.data_name + '/Temporal Train Crop Size'+str(self.size[0])+ 'x' + str(size[1])\n self.temp_val = './' + self.data_name + '/Temporal Val Crop Size'+str(self.size[0])+ 'x'+str(self.size[1])\n self.temp_test = './' + self.data_name + '/Temporal Test Crop Size'+str(self.size[0])+'x'+str(self.size[1])\n\n # Get the train, validation and test names with the data_name given\n self.train_names_original, self.val_names_original, self.test_names_original = self.train_val_test_names()\n self.get_chip_images(size)\n self.train_names = name_list(self.temp_train,load_extension)\n self.test_names = name_list(self.temp_test,load_extension)\n self.val_names = name_list(self.temp_val,load_extension)\n def train_val_test_names(self):\n \"\"\"\n Separate the data into train,validation and testing.\n Saves the train,validation and test names into a single file.\n :return:\n \"\"\"\n\n if exists('./' + self.data_name):\n print(\"The data set already exists. Loading the specified data ...\\n\")\n train_names, val_names, test_names = self.read_data_names()\n print('Size of training set: %d\\n' % (len(train_names)))\n print('Size of validation set: %d\\n' % (len(val_names)))\n print('Size of test set: %d\\n' % (len(test_names)))\n return train_names, val_names, test_names\n else:\n names = name_list(self.path_name,extension=self.load_extension )\n\n # Actions done to light hue images\n np.random.shuffle(names)\n test_names = names[0:int(len(names) * self.test_percentage)]\n val_names = names[\n int(len(names) * self.test_percentage):int(len(names) * self.test_percentage) + int(\n len(names) * self.val_percentage)]\n train_names = names[int(len(names) * self.test_percentage) + int(len(names) * self.val_percentage):]\n\n print('Size of training set: %d\\n' % (len(train_names)))\n print('Size of validation set: %d\\n' % (len(val_names)))\n print('Size of test set: %d\\n' % (len(test_names)))\n\n makedirs('./' + self.data_name)\n save_names(train_names, './' + self.data_name, 'Train Data')\n save_names(val_names, './' + self.data_name, 'Validation Data')\n save_names(test_names, './' + self.data_name, 'Test Data')\n return train_names, val_names, test_names\n\n def read_data_names(self):\n \"\"\"\n Separate the data into train,validation and testing.\n Saves the train,validation and test names into a single file.\n :return:\n \"\"\"\n train_names = load_names('./' + self.data_name, 'Train Data')\n val_names = load_names('./' + self.data_name, 'Validation Data')\n test_names = load_names('./' + self.data_name, 'Test Data')\n return train_names, val_names, test_names\n\n\n def get_chip_images(self,size):\n name_array = [self.train_names_original,self.val_names_original,self.test_names_original]\n folder_array = [self.temp_train,self.temp_val,self.temp_test]\n # Train data creation\n names = name_array[0]\n if not(exists(folder_array[0])):\n makedirs(folder_array[0])\n for name in names:\n image = cv2.imread(join(self.path_name,name))\n label = cv2.imread(join(self.path_name,name.replace(get_file_extension(name),'_FIB'+get_file_extension(name))),0)\n self.convolutional_selection(folder_array[0], size, image, label, name)\n if self.augmentation:\n generate_complete_flipping(folder_array[0],folder_array[0], data_type=self.load_extension,y_flip= False,xy_flip=False)\n # Validation Test Data\n for num_folder in range(2):\n folder = folder_array[num_folder+1]\n names = name_array[num_folder+1]\n if not(exists(folder)):\n makedirs(folder)\n for name in names:\n image = cv2.imread(join(self.path_name, name))\n label = cv2.imread(join(self.path_name, name.replace(get_file_extension(name), '_FIB' + get_file_extension(name))), 0)\n image = pad_image(image)\n label = pad_image(label)\n cv2.imwrite(join(folder, name), image)\n cv2.imwrite(join(folder,\n name.replace(get_file_extension(name),\n '_FIB' + get_file_extension(name))),label)\n\n def convolutional_selection(self,save_path, size, img, label, name):\n # Obtain the left top corner for the hard example window\n chip_names = []\n # Convolve the label to find the relevant spots\n print(self.gaussian_kernel(10).shape)\n print(label.shape)\n __, label = cv2.threshold(label, 127, 255, cv2.THRESH_BINARY_INV)\n convolved = convolve2d(label, self.gaussian_kernel((np.min(self.size))), mode='valid')\n __, label = cv2.threshold(label, 127, 255, cv2.THRESH_BINARY_INV)\n while np.sum(convolved) != 0:\n index1, index2 = np.unravel_index(convolved.argmax(), convolved.shape)\n chip_image = img[index1:(index1 + size[0]),\n index2:(index2 + size[1])] # Get the hard example window\n chip_label = label[index1:(index1 + size[0]),\n index2:(index2 + size[1])] # Get the hard example label window\n\n # Remove a certain search area for the convolution\n index1_rmv_conv_start = max([index1 - int(size[0] / 4) + 1, 0])\n index2_rmv_conv_start = max([index2 - int(size[1] / 4) + 1, 0])\n index1_rmv_conv_end = min([index1_rmv_conv_start + (size[0] - 2), img.shape[0]])\n index2_rmv_conv_end = min([index2_rmv_conv_start + (size[1] - 2), img.shape[1]])\n convolved[index1_rmv_conv_start:index1_rmv_conv_end, index2_rmv_conv_start:index2_rmv_conv_end] = 0\n\n save_name = name.replace(get_file_extension(name),\n '_H_' + str(index1) + '_' + str(index2) + get_file_extension(name))\n cv2.imwrite(join(save_path, save_name), chip_image)\n cv2.imwrite(join(save_path,\n save_name.replace(get_file_extension(save_name), '_FIB' + get_file_extension(save_name))),\n chip_label)\n return chip_names\n\n def gaussian_kernel(self, sigma):\n sizex = self.size[1]\n sizey = self.size[0]\n x, y = np.mgrid[-sizex/2:sizex/2, -sizey/2:sizey/2]\n g = np.exp(-(x ** 2 + y ** 2) / (2 * (sigma/2) ** 2))\n return g / g.sum()\n\n", "sub_path": "PantoMultiple/DataPantograph.py", "file_name": "DataPantograph.py", "file_ext": "py", "file_size_in_byte": 8358, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "cv2.copyMakeBorder", "line_number": 22, "usage_type": "call"}, {"api_name": "Asbestos_Utils.name_list", "line_number": 56, "usage_type": "call"}, {"api_name": "Asbestos_Utils.name_list", "line_number": 57, "usage_type": "call"}, {"api_name": "Asbestos_Utils.name_list", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 66, "usage_type": "call"}, {"api_name": "Asbestos_Utils.name_list", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 77, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 88, "usage_type": "call"}, {"api_name": "Asbestos_Utils.save_names", "line_number": 89, "usage_type": "call"}, {"api_name": "Asbestos_Utils.save_names", "line_number": 90, "usage_type": "call"}, {"api_name": "Asbestos_Utils.save_names", "line_number": 91, "usage_type": "call"}, {"api_name": "Asbestos_Utils.load_names", "line_number": 100, "usage_type": "call"}, {"api_name": "Asbestos_Utils.load_names", "line_number": 101, "usage_type": "call"}, {"api_name": "Asbestos_Utils.load_names", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 111, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 112, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 114, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 115, "usage_type": "call"}, {"api_name": "Asbestos_Utils.get_file_extension", "line_number": 115, "usage_type": "call"}, {"api_name": "Asbestos_Utils.generate_complete_flipping", "line_number": 118, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 123, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 124, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 126, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 126, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 127, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 127, "usage_type": "call"}, {"api_name": "Asbestos_Utils.get_file_extension", "line_number": 127, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 130, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 130, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 131, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 131, "usage_type": "call"}, {"api_name": "Asbestos_Utils.get_file_extension", "line_number": 132, "usage_type": "call"}, {"api_name": "Asbestos_Utils.get_file_extension", "line_number": 133, "usage_type": "call"}, {"api_name": "cv2.threshold", "line_number": 141, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY_INV", "line_number": 141, "usage_type": "attribute"}, {"api_name": "scipy.signal.convolve2d", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 142, "usage_type": "call"}, {"api_name": "cv2.threshold", "line_number": 143, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY_INV", "line_number": 143, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.unravel_index", "line_number": 145, "usage_type": "call"}, {"api_name": "Asbestos_Utils.get_file_extension", "line_number": 158, "usage_type": "call"}, {"api_name": "Asbestos_Utils.get_file_extension", "line_number": 159, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 160, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 160, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 161, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 161, "usage_type": "call"}, {"api_name": "Asbestos_Utils.get_file_extension", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.mgrid", "line_number": 169, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 170, "usage_type": "call"}]} +{"seq_id": "130821293", "text": "from django.conf import settings\nimport subprocess\nimport os\n\ndef runBash(cmd):\n process = subprocess.Popen(cmd, shell=True, stdout=subprocess.PIPE)\n out = process.stdout.read().strip()\n return out\n\nclass Build(object):\n def process_view(self, request, *args, **kwargs):\n if request.path.startswith('/build') or request.path.startswith('/built'):\n parts = request.path.split('/')\n path = '/%s' % '/'.join(parts[2:])\n name = path[1:][:path.find('.js')-1]\n out = '/%s/%s' % (settings.BUILTDIR, path)\n \n doContinue = True\n if request.path.startswith('/built') and os.path.exists(out):\n doContinue = False\n \n if doContinue:\n outDir = '/'.join(out.split('/')[:-1])\n if not os.path.exists(outDir):\n os.makedirs(outDir)\n \n path = path.replace('/examples', settings.EXAMPLEDIR)\n builder = '%s/build/build.sh' % os.environ['REQUIRE']\n \n command = '%s name=\"%s\" out=\"%s\" baseUrl=\"%s\" optimize=\"closure\" includeRequire=\"true\"' % (builder, name, out, settings.GMA)\n runBash(command)", "sub_path": "support/web/main/builder.py", "file_name": "builder.py", "file_ext": "py", "file_size_in_byte": 1238, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "subprocess.Popen", "line_number": 6, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.conf.settings.BUILTDIR", "line_number": 16, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 16, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 25, "usage_type": "call"}, {"api_name": "django.conf.settings.EXAMPLEDIR", "line_number": 27, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 27, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 28, "usage_type": "attribute"}, {"api_name": "django.conf.settings.GMA", "line_number": 30, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 30, "usage_type": "name"}]} +{"seq_id": "624962037", "text": "import numpy as np\n\nfrom Model import model\nfrom View.view import View\nimport pickle\nfrom PIL import Image\nimport os\n\nclass Controller:\n\n def __init__(self, pls_path: str):\n with open(pls_path, 'r') as pls:\n play_list = pls.read().splitlines()\n pickle_path = play_list[0]\n first_index = int(play_list[1])\n images_path = play_list[2:]\n self.tfl_man = TFL_Manager(pickle_path, first_index)\n\n self.run(images_path)\n\n def run(self, images_path: list):\n for image_path in images_path:\n self.tfl_man.tfl_detection(image_path)\n\n\nclass TFL_Manager:\n data, curr_frame_id = None, None\n prev_img, prev_tfl, prev_ax = None, None, None\n pp, focal = None, None\n EM = np.eye(4)\n\n def __init__(self, pkl_path: str, frame_id: int):\n self.curr_frame_id = frame_id\n with open(os.path.join(\"..\", pkl_path), 'rb') as pklfile:\n self.data = pickle.load(pklfile, encoding='latin1')\n self.pp, self.focal = self.data[\"principle_point\"], self.data[\"flx\"]\n\n def tfl_detection(self, image_path):\n img = np.array(Image.open(os.path.join(\"..\", image_path)))\n candidates, auxiliary = model.detect_candidates(img)\n View.draw_candidates(candidates, auxiliary)\n curr_tfl, curr_ax = model.filter_tfl(img, candidates, auxiliary)\n View.draw_traffic_lights(curr_tfl, curr_ax)\n if self.prev_img is not None:\n self.EM = np.dot(self.data['egomotion_' + str(self.curr_frame_id - 1) + '-' + str(self.curr_frame_id)], self.EM)\n distances = model.calc_distances(self.prev_tfl, curr_tfl, self.focal, self.pp)\n View.write_lengths(curr_tfl, curr_ax, distances)\n View.show(img)\n\n self.prev_img = img\n self.prev_tfl = curr_tfl\n self.prev_ax = curr_ax\n self.curr_frame_id += 1\n\n\n\n\n\n\ncontroller = Controller(r\"..\\play_list.pls\")", "sub_path": "Controller/controller.py", "file_name": "controller.py", "file_ext": "py", "file_size_in_byte": 1936, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "numpy.eye", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 39, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 39, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 39, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "Model.model.detect_candidates", "line_number": 40, "usage_type": "call"}, {"api_name": "Model.model", "line_number": 40, "usage_type": "name"}, {"api_name": "View.view.View.draw_candidates", "line_number": 41, "usage_type": "call"}, {"api_name": "View.view.View", "line_number": 41, "usage_type": "name"}, {"api_name": "Model.model.filter_tfl", "line_number": 42, "usage_type": "call"}, {"api_name": "Model.model", "line_number": 42, "usage_type": "name"}, {"api_name": "View.view.View.draw_traffic_lights", "line_number": 43, "usage_type": "call"}, {"api_name": "View.view.View", "line_number": 43, "usage_type": "name"}, {"api_name": "numpy.dot", "line_number": 45, "usage_type": "call"}, {"api_name": "Model.model.calc_distances", "line_number": 46, "usage_type": "call"}, {"api_name": "Model.model", "line_number": 46, "usage_type": "name"}, {"api_name": "View.view.View.write_lengths", "line_number": 47, "usage_type": "call"}, {"api_name": "View.view.View", "line_number": 47, "usage_type": "name"}, {"api_name": "View.view.View.show", "line_number": 48, "usage_type": "call"}, {"api_name": "View.view.View", "line_number": 48, "usage_type": "name"}]} +{"seq_id": "362024744", "text": "import praw\nimport time\n\n\ndef get_top_hot(reddit, event, return_data, amount):\n subreddit_name = event[\"subreddit\"]\n\n try:\n subreddit = reddit.subreddit(subreddit_name)\n except TypeError as e:\n print(\"Type Error:\", e)\n return {\n \"error\": True,\n \"message\": \"Invalid Subreddit Name\"\n }\n\n print(\"Target Subreddit: r/\"+subreddit_name)\n print(\"\")\n\n print(\"Top Ten Hot Posts:\")\n\n submissions = subreddit.hot(limit=10)\n main_data = []\n\n try:\n i = 1\n for submission in submissions:\n try:\n print(i, submission.title)\n submission_data = {\n \"error\": False,\n \"title\": submission.title,\n \"link\": submission.permalink,\n \"score\": submission.score,\n }\n main_data.append(submission_data)\n except UnicodeEncodeError:\n print(i, \"Unicode Encode Error\")\n submission_data = {\n \"error\": True,\n \"message\": \"Unicode Encode Error\",\n }\n main_data.append(submission_data)\n except:\n print(\"Some Error\")\n submission_data = {\n \"error\": True,\n \"message\": \"Some Error\",\n }\n main_data.append(submission_data)\n i += 1\n except:\n print(\"Cant find subreddit\")\n return_data[\"error\"] = True\n return_data[\"message\"] = \"Cant find subreddit\"\n\n return main_data\n\n\ndef all_comments(reddit, event, return_data):\n def getSubComments(comment, allComments, layer):\n nonlocal comments_count\n allComments.append([{\n \"layer\": layer,\n \"parent\": comment.body,\n \"child\": [],\n }])\n comments_count[\"num_total\"] += 1\n if hasattr(comment, \"replies\") and layer < comments_count[\"max_layer\"]:\n comments = comment.replies\n for comment in comments:\n getSubComments(\n comment, allComments[allComments.__len__()-1][0][\"child\"], layer + 1)\n\n submission_id = event[\"id\"]\n try:\n submission = reddit.submission(submission_id)\n submission_title = submission.title\n except:\n print(\"Invalid Id\")\n return {\n \"error\": True,\n \"message\": \"Invalid Submission Id\",\n }\n\n print(\"Target Submisssion Title: \"+submission_title)\n print(\"Target Submisssion Id: \"+submission_id)\n print(\"\\n========================================\\n\")\n\n comments_count = {\n \"id\": submission_id,\n \"num_top_level\": 0,\n \"num_total\": 0,\n \"max_top_level\": int(event[\"maxTop\"]),\n \"max_layer\": int(event[\"maxLayer\"]),\n \"url\": submission.url,\n \"text\": submission.selftext,\n }\n submission.comments.replace_more(limit=0)\n comments = submission.comments\n commentsList = []\n for comment in comments:\n if comments_count[\"num_top_level\"] > comments_count[\"max_top_level\"]:\n break\n getSubComments(comment, commentsList, 0)\n comments_count[\"num_top_level\"] += 1\n comments_count[\"comments\"] = commentsList\n return comments_count\n\n\ndef run(event):\n reddit = praw.Reddit(\n 'kinsaurralde', user_agent=\"get submissions and comments test script (by /u/kinsaurralde)\")\n\n print(\"\\n\\n\")\n print(\"Authenticated User: u/\"+str(reddit.user.me()))\n print(\"\")\n\n return_data = {\n \"version\": event[\"version\"],\n \"type\": event[\"type\"],\n }\n main_data = []\n\n if return_data[\"type\"] == \"top_ten_hot\":\n main_data = get_top_hot(reddit, event, return_data, 10)\n elif return_data[\"type\"] == \"comments\":\n main_data = all_comments(reddit, event, return_data)\n else:\n return_data[\"error\"] = True\n return_data[\"message\"] = \"Invalid Type\"\n\n return_data[\"body\"] = main_data\n\n return return_data\n\n\ndef event_error(event):\n try:\n required_version = str(event[\"version\"])\n required_type = str(event[\"type\"])\n if event[\"type\"] == \"top_ten_hot\":\n required_subreddit = str(event[\"subreddit\"])\n elif event[\"type\"] == \"comments\":\n required_id = str(event[\"id\"])\n required_max_top = int(event[\"maxTop\"])\n required_max_layer = int(event[\"maxLayer\"])\n return [False, \"\"]\n except KeyError as e:\n print(\"Missing Required Parameter:\", e)\n return [True, \"Missing Required Parameter\"]\n except ValueError as e:\n print(\"Wrong Parameter Type:\", e)\n return [True, \"Wrong Parameter Type\"]\n\n\ndef version_error(event):\n version = []\n for digit in str(event[\"version\"]).split(\".\"):\n if digit == \"TESTING\":\n version.append(999)\n else:\n try:\n version.append(int(digit))\n except ValueError as e:\n print(\"Invalid Version:\", e)\n version.append(0)\n if version.__len__() < 2:\n return [True, \"Invalid Version\"]\n if version[1] != 999 and version[0] < 1: # Current Version X in X.Y.Z\n return [True, \"Outdated Version\"]\n # Current Verion Y in X.Y.Z for top_ten_hot type\n elif event[\"type\"] == \"top_ten_hot\":\n if version[1] < 0:\n return [True, \"Outdated Version for Type\"]\n # Current Verion Y in X.Y.Z for comments type\n elif event[\"type\"] == \"comments\":\n if version[1] < 1:\n return [True, \"Outdated Version for Type\"]\n else:\n return [True, \"Invalid Type\"]\n return [False, \"\"]\n\n\ndef lambda_handler(event, context):\n start = time.time()\n error = False\n check_event = event_error(event) # Checks if required parameters present\n if check_event[0]:\n error = True\n message = check_event[1]\n else:\n check_version = version_error(event) # Checks if minimun version met\n if check_version[0]:\n error = True\n message = check_version[1]\n if error:\n main_data = message\n else:\n main_data = run(event)\n run_time = round((time.time() - start) * 1000)\n print(\"Total Run Time:\", run_time, \"ms\")\n return_data = {\n 'statusCode': 200,\n 'runTime': run_time,\n 'error': error,\n 'body': main_data\n }\n try:\n print(\"\\nReturned Data:\\n\")\n print(return_data)\n except:\n print(\"Printing Error\")\n return return_data\n\n\nlambda_handler({\n \"version\": \"1.1.0\",\n \"type\": \"comments\",\n \"subreddit\": \"all\",\n \"id\": \"\",\n \"maxTop\": 1,\n \"maxLayer\": 1,\n}, \"\")\n", "sub_path": "aws/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 6684, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "praw.Reddit", "line_number": 112, "usage_type": "call"}, {"api_name": "time.time", "line_number": 186, "usage_type": "call"}, {"api_name": "time.time", "line_number": 201, "usage_type": "call"}]} +{"seq_id": "457780403", "text": "# -*- coding: UTF-8 -*-\n__author__ = 'lowID'\nimport sql\nimport sqlite3\nimport json\n\nfrom utility import REST_API\nfrom errors import DatabaseException\n\n\nclass BaseCharacter(object):\n def __init__(self, property_dict):\n if isinstance(property_dict, dict) and len(property_dict) != 0:\n for k, v in property_dict.items():\n setattr(self, k, v)\n else:\n raise ValueError('Expected a dict input, get an %s.' % type(property_dict))\n\n\nclass Show(BaseCharacter):\n @classmethod\n def search(cls, db_conn, show_id):\n cursor = db_conn.cursor()\n #Get show info from database.\n try:\n cursor.execute(sql.SQLShow.get_show_by_id(show_id))\n show_info = cursor.fetchone()\n except sqlite3.OperationalError as e:\n raise DatabaseException('Database Error, message:<%s>.' % e.message)\n if show_info:\n #Cause the setting in database connection (row_factory), the cursor will returning a dict like object.\n #Use dict() function can cover it to a dict object.\n show_info = dict(show_info)\n cursor.execute(sql.SQLShow.get_play_urls_by_show_id(show_id))\n #Iterate the fetchall results, make every result covered to dict object, and then add the results to\n #show_info dict.\n play_urls = map(dict, cursor.fetchall())\n show_info['play_urls'] = play_urls\n return Show(show_info)\n else:\n return None\n\n def store(self, db_conn):\n cursor = db_conn.cursor()\n try:\n cursor.execute(sql.SQLShow.store_show(self))\n db_conn.commit()\n except Exception as e:\n raise DatabaseException('Database Error, message:<%s>.' % e.message)\n\n def withdraw(self, db_conn):\n cursor = db_conn.cursor()\n try:\n cursor.execute(sql.SQLShow.withdraw_show(self.show_id))\n db_conn.commit()\n except Exception as e:\n raise DatabaseException('Database Error, message:<%s>.' % e.message)\n\n def update(self, db_conn):\n cursor = db_conn.cursor()\n\n\nclass Shows(BaseCharacter):\n @staticmethod\n def count_shows(db_conn):\n cursor = db_conn.cursor()\n try:\n cursor.execute(sql.SQLShow.count_shows())\n return cursor.fetchone()[0]\n except Exception as e:\n raise DatabaseException('Database Error, message:<%s>.' % e.message)\n\n @staticmethod\n def get_shows(db_conn, show_filter='all'):\n cursor = db_conn.cursor()\n if show_filter == 'all':\n try:\n cursor.execute(sql.SQLShow.get_shows())\n return map(dict, cursor.fetchall())\n except Exception as e:\n raise DatabaseException('Database Error, message:<%s>.' % e.message)\n\n\n\nclass Channel(BaseCharacter):\n @staticmethod\n def get_all_channels(db_conn):\n cursor = db_conn.cursor()\n try:\n cursor.execute(sql.SQLChannel.get_all_channels())\n return cursor.fetchall()\n except Exception as e:\n raise DatabaseException('Database Error, message:<%s>.' % e.message)\n\n\nclass Author(BaseCharacter):\n pass\n\n\nclass User(BaseCharacter):\n pass", "sub_path": "character.py", "file_name": "character.py", "file_ext": "py", "file_size_in_byte": 3272, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "sql.SQLShow.get_show_by_id", "line_number": 26, "usage_type": "call"}, {"api_name": "sql.SQLShow", "line_number": 26, "usage_type": "attribute"}, {"api_name": "sqlite3.OperationalError", "line_number": 28, "usage_type": "attribute"}, {"api_name": "errors.DatabaseException", "line_number": 29, "usage_type": "call"}, {"api_name": "sql.SQLShow.get_play_urls_by_show_id", "line_number": 34, "usage_type": "call"}, {"api_name": "sql.SQLShow", "line_number": 34, "usage_type": "attribute"}, {"api_name": "sql.SQLShow.store_show", "line_number": 46, "usage_type": "call"}, {"api_name": "sql.SQLShow", "line_number": 46, "usage_type": "attribute"}, {"api_name": "errors.DatabaseException", "line_number": 49, "usage_type": "call"}, {"api_name": "sql.SQLShow.withdraw_show", "line_number": 54, "usage_type": "call"}, {"api_name": "sql.SQLShow", "line_number": 54, "usage_type": "attribute"}, {"api_name": "errors.DatabaseException", "line_number": 57, "usage_type": "call"}, {"api_name": "sql.SQLShow.count_shows", "line_number": 68, "usage_type": "call"}, {"api_name": "sql.SQLShow", "line_number": 68, "usage_type": "attribute"}, {"api_name": "errors.DatabaseException", "line_number": 71, "usage_type": "call"}, {"api_name": "sql.SQLShow.get_shows", "line_number": 78, "usage_type": "call"}, {"api_name": "sql.SQLShow", "line_number": 78, "usage_type": "attribute"}, {"api_name": "errors.DatabaseException", "line_number": 81, "usage_type": "call"}, {"api_name": "sql.SQLChannel.get_all_channels", "line_number": 90, "usage_type": "call"}, {"api_name": "sql.SQLChannel", "line_number": 90, "usage_type": "attribute"}, {"api_name": "errors.DatabaseException", "line_number": 93, "usage_type": "call"}]} +{"seq_id": "111340670", "text": "import numpy as np\nimport torch\nimport torch.optim as optim\nfrom matplotlib import pyplot as plt\nfrom torch import nn\nfrom torch.utils.data import DataLoader\nfrom torchvision import transforms\n\nfrom data_loader.data_loaders import SeismicDatasetLoader\nfrom model.model import Net\nfrom model.loss import softXEnt, softCrossEntropy\nfrom utils.util import ToTensor, Rescale, Normalize\nfrom scipy.signal import istft, resample\n\nTRAIN_DIR = 'train'\nPRED_DIR = 'pred'\nNOISE_DIR = 'Noise_waveforms'\npath = r'/home/antonio/SeismicSignalDenoising/data'\nsave_path = './denoising_net.pth'\n\nT = 30 # secunde\nFs = 100 # frecventa esantionare\nTs = 1 / Fs\nnt = int(T / Ts) #\nnperseg = 30\nnfft = 60\n\ntrain_size = 0.8\ntest_size = 0.2\n\ntensor = ToTensor()\nrescale = Rescale()\nnormalize = Normalize(0.5, 0.5)\n\ntransform = transforms.Compose([\n tensor,\n normalize\n # rescale\n # transforms.ToTensor(),\n # transforms.Normalize([0.5], [0.5])\n])\n\n\ndef main():\n train_dataset = SeismicDatasetLoader(root_dir=path, signal_dir=TRAIN_DIR, noise_dir=NOISE_DIR, snr=10, type='train',\n transform=transform)\n test_dataset = SeismicDatasetLoader(root_dir=path, signal_dir=PRED_DIR, noise_dir=NOISE_DIR, snr=10, type='test',\n transform=transform)\n\n # train_dataset, test_dataset = torch.utils.data.random_split(dataset, [int(round(train_size * dataset_size)),\n # int(round(test_size * dataset_size))])\n\n train_loader = DataLoader(train_dataset, batch_size=8, shuffle=True, num_workers=8)\n test_loader = DataLoader(test_dataset, batch_size=1, shuffle=False, num_workers=8)\n\n sample, stft_dict_tmp, mask, _, _, _ = train_dataset[8]\n plt.figure()\n plt.plot(sample['processed'])\n plt.figure()\n plt.plot(sample['signal'])\n # plt.show()\n device = torch.device(\"cuda:0\" if torch.cuda.is_available() else \"cpu\")\n\n net = Net().double()\n net.to(device)\n # criterion = nn.CrossEntropyLoss()\n criterion = softCrossEntropy()\n\n optimizer = optim.SGD(net.parameters(), lr=1e-3, momentum=0.9)\n\n MSE = []\n RMS_sig = 0\n RMS_noise = 0\n SNR_orig = []\n SNR = []\n\n for epoch in range(20):\n error_list = []\n running_loss = 0.0\n for i, data in enumerate(train_loader, 0):\n train_dataset.snr = np.random.randint(0, 13)\n with torch.enable_grad():\n\n sample, stft_dict, signal_mask, noise_mask, _, _ = data\n\n signal_mask = signal_mask.to(device)\n noise_mask = noise_mask.to(device)\n sample = sample['signal']\n inputs = stft_dict['Zxx_processed']\n\n composed_inputs = torch.stack((inputs.real, inputs.imag), 1)\n composed_inputs = composed_inputs.to(device)\n\n labels = torch.stack([signal_mask, noise_mask])\n # labels = signal_mask.view(signal_mask.size(0), -1)\n labels = labels.view(labels.size(1), labels.size(0), -1)\n labels = labels.long()\n\n optimizer.zero_grad()\n\n outputs = net(composed_inputs)\n\n # loss = criterion(outputs, labels)\n loss = criterion(outputs, labels)\n loss.backward()\n optimizer.step()\n\n running_loss += loss.item()\n if i % 20 == 19:\n print('[%d, %5d] loss: %.3f' % (epoch + 1, i + 1, running_loss / 20))\n running_loss = 0.0\n print('Finished Training')\n\n plt.figure()\n plt.plot(MSE, 'x')\n plt.figure()\n plt.plot(SNR_orig, SNR, 'x')\n plt.show()\n\n torch.save(net.state_dict(), save_path)\n\n model = Net()\n model.load_state_dict(torch.load(save_path))\n model.eval()\n\n SNR_mean = []\n with torch.no_grad():\n for i in range(13):\n test_dataset.snr = i\n for data in test_loader:\n sample, images, signal_labels, noise_labels, noise, noisy_snr = data\n noise_labels = noise_labels\n signal_labels = signal_labels\n sample = sample['signal']\n images = images['Zxx_processed']\n\n composed_images = torch.stack((images.real, images.imag), 1)\n composed_images = composed_images.to(device)\n\n sample = sample.squeeze(0)\n sample = sample.numpy()\n\n # plt.figure()\n # plt.plot(sample)\n # plt.title('Original signal')\n\n signal_labels = signal_labels.view(signal_labels.size(0), -1)\n signal_labels = signal_labels.squeeze(0)\n\n outputs = net(composed_images)\n outputs = outputs.view(outputs.size(0), outputs.size(1), 31, -1)\n\n signal_approx = composed_images * outputs[:, 0, :, :]\n\n new_signal_approx = signal_approx[:, 0, :, :] + 1j * signal_approx[:, 1, :, :]\n\n _, new_signal_approx = istft(new_signal_approx.cpu().detach().numpy(), fs=Fs, nperseg=nperseg,\n nfft=nfft,\n boundary='zeros')\n\n new_signal_approx = new_signal_approx.squeeze(0)\n rescaled_signal = 0.5 + (new_signal_approx - new_signal_approx.mean()) * (0.5 / new_signal_approx.std())\n\n # plt.figure()\n # plt.plot(rescaled_signal)\n # plt.title('Output Signal')\n # plt.show()\n\n snr_calculat = 20 * np.log10(np.std(rescaled_signal) / np.std(sample - rescaled_signal))\n SNR.append(snr_calculat)\n print(i)\n SNR_mean.append(sum(SNR) / len(SNR))\n SNR_orig.append(i)\n\n plt.figure()\n # plt.plot(MSE, 'x')\n plt.plot((SNR_mean,), '*')\n plt.legend(\"Blue - after denoising\")\n plt.plot((SNR_orig,), '*')\n plt.legend(\"Orange - original\")\n plt.show()\n\n\nif __name__ == '__main__':\n main()\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 6084, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "utils.util.ToTensor", "line_number": 31, "usage_type": "call"}, {"api_name": "utils.util.Rescale", "line_number": 32, "usage_type": "call"}, {"api_name": "utils.util.Normalize", "line_number": 33, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 35, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 35, "usage_type": "name"}, {"api_name": "data_loader.data_loaders.SeismicDatasetLoader", "line_number": 45, "usage_type": "call"}, {"api_name": "data_loader.data_loaders.SeismicDatasetLoader", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "torch.device", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 62, "usage_type": "attribute"}, {"api_name": "model.model.Net", "line_number": 64, "usage_type": "call"}, {"api_name": "model.loss.softCrossEntropy", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.optim.SGD", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 69, "usage_type": "name"}, {"api_name": "numpy.random.randint", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 81, "usage_type": "attribute"}, {"api_name": "torch.enable_grad", "line_number": 82, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 117, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 117, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "torch.save", "line_number": 120, "usage_type": "call"}, {"api_name": "model.model", "line_number": 122, "usage_type": "name"}, {"api_name": "model.model.Net", "line_number": 122, "usage_type": "call"}, {"api_name": "model.model.load_state_dict", "line_number": 123, "usage_type": "call"}, {"api_name": "model.model", "line_number": 123, "usage_type": "name"}, {"api_name": "torch.load", "line_number": 123, "usage_type": "call"}, {"api_name": "model.model.eval", "line_number": 124, "usage_type": "call"}, {"api_name": "model.model", "line_number": 124, "usage_type": "name"}, {"api_name": "torch.no_grad", "line_number": 127, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 137, "usage_type": "call"}, {"api_name": "scipy.signal.istft", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 169, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 175, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 175, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 177, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 177, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 178, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 179, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 179, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 180, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 181, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 181, "usage_type": "name"}]} +{"seq_id": "433885287", "text": "import requests\nimport smtplib \nfrom email.mime.multipart import MIMEMultipart\nfrom email.mime.text import MIMEText\nfrom email.header import Header\nfrom email.utils import formataddr\nimport time\n\n\ndef send_to_email(news):\n from_address = 'senderEmail'\n # list of users\n recpient_list = ['useremail1','useremail2','useremail2']\n msg = MIMEMultipart()\n msg['From']=formataddr((str(Header('PythoNews', 'utf-8')), '1mv19is404@sirmvit.edu'))\n \n msg['Subject']='Your Daily News'\n news_data = news\n html ='''\n \n PythoNews -Your Daily News\n \n \n \n '''\n for i in range(0, len(news_data)):\n html+=f'

{news_data[i][0]}

\\n

{news_data[i][1]}

\\nRead Full Post'\n html +=''\n msg.attach(MIMEText(html,'html'))\n # creates SMTP session \n s = smtplib.SMTP('smtp.gmail.com', 587) \n s.ehlo()\n s.starttls()\n s.ehlo()\n s.login(from_address,'Sender Password')\n text = msg.as_string()\n for i in recpient_list:\n msg['To']=i\n s.sendmail(from_address,i,text)\n s.quit()\n print(\"Your Daily News Sented\")\n \n\n \n \n\n\nurl = 'https://newsapi.org/v2/top-headlines?sources=google-news-in&apiKey='\n# you can got api from this website:https://newsapi.org\n\nresponse = requests.get(url)\ndata = response.json()\ndata=data['articles']\n\ntitles = list()\ndescription = list()\nurls = list()\n\nfor i in range(0,len(data)):\n titles.append(data[i]['title'])\n description.append(data[i]['description'])\n urls.append(data[i]['url'])\n\n\n\nparsedData = tuple(zip(titles,description,urls))\n\nwhile True:\n send_to_email(parsedData)\n time.sleep(3600) # 1hour = 3600 seconds\n # calling the same function after 1 hour\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "google_news.py", "file_name": "google_news.py", "file_ext": "py", "file_size_in_byte": 2231, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "email.mime.multipart.MIMEMultipart", "line_number": 14, "usage_type": "call"}, {"api_name": "email.utils.formataddr", "line_number": 15, "usage_type": "call"}, {"api_name": "email.header.Header", "line_number": 15, "usage_type": "call"}, {"api_name": "email.mime.text.MIMEText", "line_number": 47, "usage_type": "call"}, {"api_name": "smtplib.SMTP", "line_number": 49, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 69, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 88, "usage_type": "call"}]} +{"seq_id": "380407960", "text": "#!/usr/bin/python\n\nimport json\nimport datetime\nimport time\nimport os\nimport threading\nimport glob\nimport re\nimport yaml\nimport math\nimport blinker\nimport Queue\nimport weakref\nimport bitstring\nimport subprocess\n\nimport _strptime # Prevent import in thread\n\nfrom functools import partial\nfrom operator import itemgetter\nfrom collections import defaultdict\n\nimport illuminate\nfrom flask import Flask, url_for, redirect, jsonify, Response, request\n\n# Storage for run folders: (path, type of run)\nRUN_STORAGES = [(\"/data/runScratch.boston\", \"general\"), (\"/boston/diag/runs/veriseq\", \"nipt\")]\n\nSEQUENCER_LIST = [(seq['id'], (seq['type'], seq['name']))\n for seq in yaml.safe_load(open(os.path.join(os.path.dirname(__file__), \"sequencers.yaml\")))\n ]\n\nSEQUENCERS = dict(SEQUENCER_LIST)\n# Mark as cancelled if waiting for N times the measured cycle time\nCANCELLED_TIME_N_CYCLES = 3\n\napp = Flask(__name__)\ndb = None # Set on bottom of script\n\ndef updater():\n \"\"\"Updater background thread\"\"\"\n\n while True:\n db.update()\n time.sleep(61)\n\nMAX_ACTIVE_STREAMS = 19\nactive_streams = []\nKEEPALIVE_INTERVAL = 60 # Times sleep interval 61\n\ndef machine_id(run_id):\n return re.match(r\"\\d{6}_([A-Z0-9]+)_.*\", run_id).group(1)\n\nclass Database(object):\n \"\"\"Persistent storage for some run data.\"\"\"\n\n COUNT_FILE = \"/var/db/nsc-status/count.txt\"\n BOOKED_RUNS_FILE = \"/var/db/nsc-status/booked.txt\"\n CANCELLED_RUNS_FILE = \"/var/db/nsc-status/cancelled.txt\"\n\n def __init__(self):\n self.completed = set()\n self.status = {}\n self.basecount_signal = blinker.Signal()\n self.run_status_signal = blinker.Signal()\n self.machine_list_signal = blinker.Signal()\n self.keepalive_counter = 0\n try:\n with open(self.COUNT_FILE) as f:\n self.count = int(f.read())\n except IOError:\n self.count = 0\n try:\n with open(self.BOOKED_RUNS_FILE) as f:\n self.booked_runs = set(r.strip() for r in f.readlines())\n except IOError:\n self.booked_runs = set()\n try:\n with open(self.CANCELLED_RUNS_FILE) as f:\n self.cancelled_runs = set(r.strip() for r in f.readlines())\n except IOError:\n self.cancelled_runs = set()\n\n def update(self):\n with open(self.COUNT_FILE) as f:\n self.count = int(f.read())\n\n run_dirs_and_ids = [\n (rpath, os.path.basename(os.path.dirname(rpath)), run_type)\n for (run_storage, run_type) in RUN_STORAGES\n for rpath in glob.glob(os.path.join(run_storage, \"??????_*_*\", \"*\"))\n ]\n runs_on_storage = {\n run_id: (os.path.dirname(rpath), run_type)\n for (rpath, run_id, run_type) in run_dirs_and_ids\n if re.match(r\"[0-9]{6}_[A-Z0-9]+_[_A-Z0-9-]+$\", run_id)\n }\n\n new = set(runs_on_storage) - set(self.status.keys())\n\n for r_id in new:\n if machine_id(r_id) in SEQUENCERS:\n new_run = RunStatus(r_id, *runs_on_storage[r_id], start_cancelled=r_id in self.cancelled_runs)\n self.status[r_id] = new_run\n\n modified = False\n updated = []\n for r in self.status.values():\n if r.update():\n updated.append(r)\n if r.finished and not r.committed:\n try:\n if not r.run_id in self.booked_runs:\n self.increment(r.basecount)\n self.booked_runs.add(r.run_id)\n finally:\n r.committed = True\n modified = True\n\n missing = set(self.status.keys()) - set(runs_on_storage)\n for r_id in missing:\n if not self.status[r_id].is_fake:\n del self.status[r_id]\n self.booked_runs.discard(r_id)\n\n self.booked_runs &= set(self.status.keys())\n new_cancelled_runs = set(k for (k, v) in self.status.items() if v.cancelled)\n if new_cancelled_runs != self.cancelled_runs:\n self.cancelled_runs = new_cancelled_runs\n modified = True\n\n if modified:\n self.save()\n\n # Limit the number of NIPT runs shown per instrument\n per_instrument_counts = defaultdict(int)\n for run_id in sorted(self.status, reverse=True):\n run = self.status[run_id]\n if run.finished and run.run_type == \"nipt\":\n run.hidden = True\n\n if new or missing:\n self.machine_list_signal.send(self, data=self.machine_list)\n\n if updated or self.keepalive_counter > KEEPALIVE_INTERVAL:\n self.keepalive_counter = 0\n self.basecount_signal.send(self, data=self.global_base_count)\n for r in updated:\n self.run_status_signal.send(self, data=r.data_package)\n\n def increment(self, bases):\n self.count += bases\n\n def save(self):\n with open(self.COUNT_FILE, 'w') as f:\n f.write(str(int(self.count)))\n with open(self.BOOKED_RUNS_FILE, 'w') as f:\n f.writelines(\"\\n\".join(self.booked_runs))\n with open(self.CANCELLED_RUNS_FILE, 'w') as f:\n f.writelines(\"\\n\".join(self.cancelled_runs))\n\n @property\n def global_base_count(self):\n rate = sum(run.rate for run in self.status.values())\n count = self.count + sum(run.basecount for run in self.status.values() if not run.committed)\n return {'count': count, 'rate': rate}\n\n @property\n def machine_list(self):\n machines = {}\n for m_id, (m_type, m_name) in SEQUENCER_LIST:\n run_ids = [\n run_id for run_id in sorted(self.status.keys())[::-1]\n if re.match(\"\\\\d{6}_%s_.*\" % (m_id), run_id) and not self.status[run_id].hidden\n ]\n machines[m_id] = {\n 'id': m_id,\n 'name': m_name,\n 'type': m_type,\n 'run_ids': run_ids\n }\n\n # Sort machines with newest runs first\n #return list(reversed(sorted(machines.values(), key=lambda x: (not x['type'].startswith(\"-\"), x['run_ids']))))\n # Don't sort machines, use fixed order as in config\n return [machines[m_id] for m_id, _ in SEQUENCER_LIST]\n\n\n\ndef instrument_rate(m_id):\n instrument = SEQUENCERS[m_id][0]\n if instrument == \"hiseqx\":\n per_hour = 12500000000\n elif instrument == \"hiseq4k\" or instrument == \"hiseq3k\":\n per_hour = 8928571428\n elif instrument == \"hiseq\":\n per_hour = 3472222222\n elif instrument == \"nextseq\":\n per_hour = 4137931034\n elif instrument == \"miseq\":\n per_hour = 133928571\n elif instrument == \"novaseq\":\n per_hour = 3125000000\n return per_hour / 3600.0\n\nclass RunStatus(object):\n\n public = ['machine_id', 'run_id', 'run_dir', 'run_type', 'read_config', 'current_cycle',\n 'total_cycles', 'basecount', 'rate', 'finished', 'cancelled']\n\n def __init__(self, run_id, run_dir, run_type, start_cancelled=False):\n self.machine_id = machine_id(run_id)\n self.run_id = run_id\n self.run_dir = run_dir\n self.run_type = run_type\n self.hidden = False\n self.read_config = []\n self.current_cycle = 0\n self.total_cycles = 0\n self.clusters = 0\n\n self.last_update = time.time()\n self.booked = 0\n self.committed = False # Is base count added to grand total?\n self.start_time = 0\n self.cycle_arrival = {} # (cycle, time) pairs\n self.finished = False\n self.start_cancelled = start_cancelled\n self.cancelled = start_cancelled\n\n def set_metadata(self):\n try:\n ds = illuminate.InteropDataset(self.run_dir)\n except:\n return False\n self.read_config = list(ds.meta.read_config)\n self.total_cycles = sum(read['cycles'] for read in self.read_config)\n self.cycle_first_in_read_flag = [True] + sum(\n (\n [False]*(read['cycles']-1) + [True] for read in self.read_config),\n []\n )\n return self.total_cycles != 0\n\n def get_cycle(self):\n for cycle in range(self.current_cycle, self.total_cycles):\n test_paths = [\n os.path.join(\n self.run_dir, \"Data\", \"Intensities\", \"BaseCalls\", \"L001\",\n \"C{0}.1\".format(cycle+1)\n ),\n os.path.join(\n self.run_dir, \"Data\", \"Intensities\", \"BaseCalls\", \"L001\",\n \"{0:04d}.bcl.bgzf\".format(cycle+1)\n )\n ]\n if not any(os.path.exists(test_path) for test_path in test_paths):\n return cycle\n return self.total_cycles\n\n def get_clusters(self):\n instrument = SEQUENCERS[self.machine_id][0]\n if instrument == \"novaseq\":\n process = subprocess.Popen(['nsc-python27', \n '/opt/gls/clarity/customextensions/lims/base-counter/clusters-helper.py',\n self.run_dir],\n stdout=subprocess.PIPE, stderr=open(os.devnull, 'w'))\n try:\n out, _ = process.communicate()\n val = float(out)\n return None if math.isnan(val) else val\n except (subprocess.CalledProcessError, ValueError):\n return None\n try:\n ds = illuminate.InteropDataset(self.run_dir)\n all_df = ds.TileMetrics().df\n except (ValueError, TypeError, IOError, bitstring.ReadError, illuminate.InteropFileNotFoundError):\n return None # No information yet, or it's being written to\n return all_df[all_df.code == 103].sum().sum() # Number of clusters PF\n\n def check_finished(self):\n return os.path.exists(os.path.join(self.run_dir, \"RTAComplete.txt\"))\n\n def check_cancelled(self):\n \"\"\"Heuristic to determine if cancelled. If current cycle time is\n more than 2x last cycle time.\"\"\"\n\n if len(self.cycle_arrival) <= 1 and self.start_cancelled:\n return True\n if (not self.finished) and len(self.cycle_arrival) > 0:\n current_cycle_time = time.time() - self.cycle_arrival[self.current_cycle]\n if current_cycle_time > 7*3600:\n return True # Time based check: very long cycle time gets marked as a fail\n if len(self.cycle_arrival) > 2:\n cycle_rate, cycle_stride = self.get_cycle_rate()\n if self.cycle_first_in_read_flag[self.current_cycle]:\n first_cycle_in_read = 25 # Some slack on start of read 2, takes less than 25 cycles before writing on MiSeq\n else:\n first_cycle_in_read = 0\n cancelled = current_cycle_time > (CANCELLED_TIME_N_CYCLES+first_cycle_in_read) * (cycle_stride / cycle_rate)\n return cancelled\n return False\n\n def update(self):\n if self.finished:\n return\n now = time.time()\n updated = False\n initial_update = False\n # Get number of cycles, run metadata\n if not self.read_config:\n updated = self.set_metadata()\n initial_update = updated\n # If no metadata, the run hasn't really started yet.\n self.start_time = now\n if self.read_config:\n self.current_cycle = self.get_cycle()\n if self.cycle_arrival.setdefault(self.current_cycle, now) == now: # Add if not exists\n updated = True\n new_clusters = self.get_clusters()\n self.clusters = new_clusters or self.clusters\n if self.clusters:\n self.booked = self.current_cycle * self.clusters\n else:\n self.booked = 0\n if self.check_finished():\n self.finished = True\n updated = True\n new_cancelled = self.check_cancelled()\n if new_cancelled != self.cancelled:\n self.cancelled = new_cancelled\n updated = True\n if updated:\n if self.clusters != 0 or initial_update:\n self.last_update = now\n return updated\n\n def get_cycle_rate(self):\n # Difference in cycle, time\n # Here, we look at all cycles, including index cycles,\n # to estimate the speed of the sequencer\n cycle_arrival_list = sorted(self.cycle_arrival.items(), key=itemgetter(0))\n dcs, dts = zip(*[\n (\n (a2[0] - a1[0]),\n (a2[1] - a1[1])\n )\n for a1, a2 in\n zip(\n cycle_arrival_list,\n cycle_arrival_list[1:]\n )\n ])\n # Mean cycles per update for last 5 updates\n # Typically this will be unity, unless updates are run very infrequently\n mean_stride = sum(dcs[-5:]) / min(len(dcs), 5)\n # Total rate (index + data cycles per time)\n mean_cycle_rate = sum(dcs[-5:]) / sum(dts[-5:])\n return mean_cycle_rate, mean_stride\n\n @property\n def rate(self):\n \"\"\"Bases per second\"\"\"\n\n if self.finished or self.cancelled:\n return 0\n if len(self.cycle_arrival) > 2 and self.clusters != 0:\n mean_cycle_rate, mean_stride = self.get_cycle_rate()\n next_data_cycles = min(self.current_cycle+int(mean_stride), self.total_cycles)\n data_factor = (next_data_cycles - self.current_cycle) / mean_stride\n return self.clusters * mean_cycle_rate * data_factor\n elif len(self.cycle_arrival) > 0 and self.current_cycle != 0:\n return instrument_rate(self.machine_id)\n else:\n return 0\n\n @property\n def basecount(self):\n \"\"\"Estimated number of bases at this instant\"\"\"\n if len(self.cycle_arrival) > 3 or self.current_cycle > 29:\n return self.booked + (self.rate * (time.time() - self.last_update))\n elif len(self.cycle_arrival) >= 1:\n return self.rate * (time.time() - self.start_time)\n else:\n return 0\n\n @property\n def data_package(self):\n \"\"\"Dict to be sent to clients\"\"\"\n\n return dict((key, getattr(self, key)) for key in RunStatus.public)\n\n @property\n def is_fake(self):\n return False\n\nclass FakeRun(RunStatus):\n \"\"\"For testing, etc.\"\"\"\n\n def __init__(self, machine_id, num_cycles, start_cycle=0):\n run_id = datetime.datetime.now().strftime(\"%y%m%d_\" + machine_id + \"_FAKE_%f\")\n super(FakeRun, self).__init__(run_id, run_id, 'general')\n self.num_cycles = num_cycles\n self.start_time = time.time()\n self.start_cycle = start_cycle\n\n def set_metadata(self):\n self.read_config = True # See if we can get away with it\n self.total_cycles = self.num_cycles\n # Build look-up table for number of cycles -> number of data cycles\n self.data_cycles_lut = [(i, i) for i in xrange(self.num_cycles+1)]\n return True\n\n def get_clusters(self):\n m_t = SEQUENCERS[self.machine_id][0]\n if m_t == \"hiseq\":\n return 2e9\n elif m_t == \"hiseqx\":\n return 2.6e9\n elif m_t in [\"hiseq4k\", \"hiseq3k\"]:\n return 2.1e9\n elif \"nextseq\":\n return 400e6\n elif \"miseq\":\n return 25e6\n elif \"novaseq\":\n return 3.3e9 # 3.3 billion reads, spec S2 flow cell\n\n def get_cycle(self):\n speed = instrument_rate(self.machine_id) / self.get_clusters()\n return int(min(self.start_cycle + (time.time() - self.start_time) * speed, self.total_cycles))\n\n def check_finished(self):\n return False\n\n @property\n def is_fake(self):\n return True\n\n\nclass EventQueuer(object):\n \"\"\"Helper class encapsulates a single type of signal, conversion\n to queue data.\"\"\"\n\n def __init__(self, queue, ident):\n self.queue = queue\n self.ident = ident\n\n def __call__(self, sender, data):\n self.queue.put((self.ident, data))\n\nclass SseStream(object):\n \"\"\"Usees a Queue to translate between a signal (method call\n interface) and a generator protocol.\n\n This new version of SSE Stream can multiplex multiple signals\n onto the event stream, setting the id of each event as appropriate.\n The constructor argument is a list of event type sepecifications,\n encoded as tuples:\n (SIGNAL, ID)\n Every time a signal is received, the data of the event is JSON\n encoded and returned to the stream as a SSE, with the specified\n ID (\"event:\" line).\n\n The ID can be None, in which case no ID is sent.\n \"\"\"\n\n TERMINATE = object()\n\n def __init__(self, event_specs):\n self.queue = Queue.Queue()\n self.helpers = []\n for signal, ident in event_specs:\n qr = EventQueuer(self.queue, ident)\n signal.connect(qr)\n self.helpers.append(qr)\n\n def __iter__(self):\n return self\n\n def next(self):\n ident, data = self.queue.get(block=True)\n if ident is self.TERMINATE:\n raise StopIteration()\n event_str = \"\"\n if ident is not None:\n event_str = \"event: \" + ident + \"\\n\"\n event_str += 'data: ' + json.dumps(data) + '\\n\\n'\n return event_str\n\n def update(self, ident, sender=None, data=None):\n self.queue.put((ident, data))\n\n def terminate(self):\n self.queue.put((self.TERMINATE, None))\n\n\n@app.route(\"/\")\ndef get_main():\n return redirect(url_for('static', filename='index.html'))\n\n\n@app.route(\"/status\")\ndef status():\n global active_streams\n active_streams = [stream for stream in active_streams if stream() is not None]\n while len(active_streams) >= MAX_ACTIVE_STREAMS:\n (active_streams.pop(0))().terminate()\n events = [\n (db.basecount_signal, \"basecount\"),\n (db.run_status_signal, \"run_status\"),\n (db.machine_list_signal, \"machine_list\")\n ]\n stream = SseStream(events)\n # Send initial status\n stream.update(\"basecount\", data=db.global_base_count)\n stream.update(\"machine_list\", data=db.machine_list)\n for r in db.status.values():\n if not r.hidden:\n stream.update(\"run_status\", data=r.data_package)\n active_streams.append(weakref.ref(stream))\n return Response(stream, mimetype=\"text/event-stream\")\n\n@app.route(\"/machines\")\ndef machines():\n return jsonify(SEQUENCERS)\n\n@app.route(\"/fake\", methods=['POST'])\ndef fake():\n params = request.json\n run = FakeRun(params['machine'], int(params['cycles']), int(params['start_cycle']))\n db.status[run.run_id] = run\n db.update()\n return \"OK\"\n\n@app.route(\"/fake-runs\")\ndef fakes():\n return jsonify(runs=[run.run_id for run in db.status.values() if isinstance(run, FakeRun)])\n\n@app.route(\"/delete\", methods=['POST'])\ndef delete():\n params = request.json\n del db.status[params['id']]\n return \"OK\"\n\ndb = Database()\nupdater_thread = threading.Thread(target=updater, name=\"updater\")\nupdater_thread.daemon = True\nupdater_thread.start()\n\nif __name__ == \"__main__\":\n app.debug = True\n app.run(host=\"0.0.0.0\", port=5001, threaded=True)\n", "sub_path": "base-counter/server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 19293, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "yaml.safe_load", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 31, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 38, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 46, "usage_type": "call"}, {"api_name": "re.match", "line_number": 53, "usage_type": "call"}, {"api_name": "blinker.Signal", "line_number": 65, "usage_type": "call"}, {"api_name": "blinker.Signal", "line_number": 66, "usage_type": "call"}, {"api_name": "blinker.Signal", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path", "line_number": 90, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 90, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path", "line_number": 92, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}, {"api_name": "re.match", "line_number": 97, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 137, "usage_type": "call"}, {"api_name": "re.match", "line_number": 175, "usage_type": "call"}, {"api_name": "time.time", "line_number": 223, "usage_type": "call"}, {"api_name": "illuminate.InteropDataset", "line_number": 234, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 249, "usage_type": "call"}, {"api_name": "os.path", "line_number": 249, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 253, "usage_type": "call"}, {"api_name": "os.path", "line_number": 253, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 258, "usage_type": "call"}, {"api_name": "os.path", "line_number": 258, "usage_type": "attribute"}, {"api_name": "subprocess.Popen", "line_number": 265, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 268, "usage_type": "attribute"}, {"api_name": "os.devnull", "line_number": 268, "usage_type": "attribute"}, {"api_name": "math.isnan", "line_number": 272, "usage_type": "call"}, {"api_name": "subprocess.CalledProcessError", "line_number": 273, "usage_type": "attribute"}, {"api_name": "illuminate.InteropDataset", "line_number": 276, "usage_type": "call"}, {"api_name": "bitstring.ReadError", "line_number": 278, "usage_type": "attribute"}, {"api_name": "illuminate.InteropFileNotFoundError", "line_number": 278, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 283, "usage_type": "call"}, {"api_name": "os.path", "line_number": 283, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 283, "usage_type": "call"}, {"api_name": "time.time", "line_number": 292, "usage_type": "call"}, {"api_name": "time.time", "line_number": 308, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 343, "usage_type": "call"}, {"api_name": "time.time", "line_number": 382, "usage_type": "call"}, {"api_name": "time.time", "line_number": 384, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 402, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 402, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 405, "usage_type": "call"}, {"api_name": "time.time", "line_number": 432, "usage_type": "call"}, {"api_name": "Queue.Queue", "line_number": 472, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 489, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 501, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 501, "usage_type": "call"}, {"api_name": "weakref.ref", "line_number": 522, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 523, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 527, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 531, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 531, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 539, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 543, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 543, "usage_type": "name"}, {"api_name": "threading.Thread", "line_number": 548, "usage_type": "call"}]} +{"seq_id": "234548843", "text": "#import paramiko\nimport os\nimport time\nimport json\nimport sys\nsys.path.insert(0,'/pipeline/sharmi/allen_SB_code/celery/')\nfrom celery import Celery\n#from tasks import run_celerycommand\n\ndef parse_project_directory(line):\n\n proj = line.split(\"raw\")\n projectdirectory = proj[0]\n\n tok = line.split(\"/\")\n ribbondir = tok[len(tok)-2]\n sessiondir = tok[len(tok)-1]\n ribbon = int(ribbondir[6:])\n session = int(sessiondir[7:])\n\n return [projectdirectory, ribbon, session]\n\n\ndef get_channel_names(directory):\n #return os.listdir(directory)\n directory_list = list()\n for root, dirs, files in os.walk(directory, topdown=False):\n for name in dirs:\n directory_list.append(os.path.join(root, name))\n return dirs\n############### trying to read section number #####################\n#def get_section_num(directory):\n# with open('session_metadata.txt') as f:\n# content = f.readlines()\n# last_sect = int(content.split(\"Length\")[1])-1\n# return last_sect\n\n\ndef get_channel_nums(statetablefile):\n df=pd.read_csv(statetablefile)\n uniq_ch=df.groupby(['ch']).groups.keys()\n return uniq_ch\n\ndef parseprojectroot(projectdirectory):\n print (\"this is your directory: \" + projectdirectory)\n tok = projectdirectory.split(\"/\")\n dataind = tok.index('data')\n print (\"this is your root: \" + tok[dataind+1])\n return tok[dataind+1]\n\ndef parsefile(fname):\n\n with open(fname) as f:\n content = f.readlines()\n\n if len(content)>1:\n print (\"File is corrupted...\")\n else:\n #parse line\n\n fullline = content[0]\n fullinetok = fullline.split(\",\")\n section = int(fullinetok[1])\n owner = str(fullinetok[2])\n #owner = \"TESTEXPERIMENT\"\n\n line = fullinetok[0]\n\n proj = line.split(\"raw\")\n projectdirectory = proj[0]\n\n tok = line.split(\"/\")\n ribbondir = tok[len(tok)-2]\n sessiondir = tok[len(tok)-1]\n ribbon = int(ribbondir[6:])\n session = int(sessiondir[7:])\n\n return [projectdirectory, ribbon, session, section, owner,fullline]\n\ndef savemedianjson(med,medianfile,owner, project,acq_stack,median_stack,median_dir,minz,maxz,close_stack):\n\n med['render']['owner'] = owner\n med['render']['project'] = project\n med['input_stack'] = acq_stack\n med['output_stack'] = median_stack\n med['minZ'] = minz\n med['maxZ'] = maxz\n med['output_directory'] = median_dir\n med['close_stack'] = close_stack\n with open(medianfile, 'w') as outfile:\n json.dump(med, outfile,indent=4)\n\n\nif __name__ == \"__main__\":\n\n owner = \"Small_volumes_2018\"\n mediantemplate = \"template/median.json\"\n firstsection = 0\n lastsection = 35\n num_iter = 20\n #bgrd_size = 50\n\n curdir = os.getcwd()\n\n for i in range (0,1):\n\n with open(\"/pipeline/leila/stitching/confirm_data2process\") as f:\n alldirnames = f.readlines()\n\n for dirname in alldirnames:\n flatfield_dirname = \"%s/../../../../processed/Flatfield\" %dirname.strip('\\n')\n print (\"thiis is dirname: \" + dirname)\n print (flatfield_dirname)\n if not os.path.exists(flatfield_dirname):\n os.makedirs(flatfield_dirname)\n for sectnum in range(firstsection,lastsection+1):\n close_stack = False\n\n if sectnum==lastsection:\n close_stack = True\n\n projectdirectory = dirname.strip()\n project = parseprojectroot(projectdirectory)\n channels = get_channel_names(projectdirectory)\n [projectroot, ribbon,session] = parse_project_directory(projectdirectory)\n z = ribbon*100+sectnum\n print (\"this is your projectroot: \" + projectroot)\n\n\n\n #create file that consists of celery job commands\n filename = \"log/runme_sect_%s_%d_%d_%s.sh\"%(project, ribbon,session,sectnum)\n f = open(filename,'w')\n\n\n #create state table\n\n #projectroot = \"/nas/data/M246930_Scnn1a_4/\"\n statetablefile =projectroot+ \"scripts/statetable_ribbon_%d_session_%d_section_%d\"%(ribbon,session,sectnum)\n #statetablefile = \"/nas/data/M246930_Scnn1a_4/scripts/statetable_ribbon_%d_session_%d_section_%d\"%(ribbon,session,sectnum)\n print (projectroot)\n print (statetablefile)\n #exit(0)\n #make state table\n cmd = \"docker exec luigiscripts python make_state_table_ext_multi_pseudoz.py \"\n cmd = cmd + \"--projectDirectory %s \"%projectroot\n cmd = cmd + \"--outputFile %s \"%statetablefile\n cmd = cmd + \"--oneribbononly True \"\n cmd = cmd + \"--ribbon %d \"%ribbon\n cmd = cmd + \"--session %d \"%session\n cmd = cmd + \"--section %d \"%sectnum\n f.write(cmd+\"\\n\")\n #os.system(cmd)\n\n #upload acquisition stacks\n dcmd = \"docker exec renderapps_develop python -m renderapps.dataimport.create_fast_stacks \"\n dcmd = dcmd + \"--render.host ibs-forrestc-ux1 \"\n dcmd = dcmd + \"--render.client_scripts /var/www/render/render-ws-java-client/src/main/scripts \"\n dcmd = dcmd + \"--render.port 8988 \"\n dcmd = dcmd + \"--render.memGB 5G \"\n dcmd = dcmd + \"--log_level INFO \"\n dcmd = dcmd + \"--statetableFile %s \"%statetablefile\n dcmd = dcmd + \"--render.project %s \"%project\n dcmd = dcmd + \"--projectDirectory %s \"%projectroot\n dcmd = dcmd + \"--outputStackPrefix ACQ_S0%d\"%(session)\n dcmd = dcmd + \" --render.owner %s \"%owner\n f.write(dcmd+\"\\n\")\n #os.system(dcmd)\n #exit(0)\n #dcmd = dcmd + \"--projectDirectory %s \"%projectroot\n\n\n #print channels\n\n\n for ch in channels:\n print (\"Channel: \" + str(ch))\n #print \"PRINTING CHANNELS TEST\"\n #print ch\n medianfile = \"%s/log/median_%s_%s_%s_%s_%d.json\"%(curdir,project,ch,ribbon,session,sectnum)\n\n #stacks\n acq_stack = \"ACQ_S0%d_%s\"%(int(session),ch)\n median_stack = \"MED_S0%d_%s\"%(int(session),ch)\n\n #directories\n median_dir = \"%s/processed/Medians/\"%projectroot\n\n #psf file, scale factor, and background size for deconvolution\n print (close_stack)\n if close_stack:\n #create median file\n firstz = ribbon*100+firstsection\n lastz = ribbon*100+lastsection\n print (firstz)\n print (lastz)\n with open(mediantemplate) as json_data:\n med = json.load(json_data)\n savemedianjson(med,medianfile,owner, project,acq_stack,median_stack,median_dir,firstz,lastz,close_stack)\n\n\n\n #run\n mystr = \"DAPI\"\n cmd1 = \"docker exec rm_container python -m rendermodules.intensity_correction.calculate_multiplicative_correction --render.port 8988 --input_json %s\"%medianfile\n\n f.write(cmd1+\"\\n\")\n print (cmd1)\n\n f.close()\n rcmd = \"sh %s\"%filename\n print (rcmd)\n #result = run_celerycommand.apply_async(args=[rcmd,os.getcwd()])\n os.system(rcmd)\n time.sleep(2)\n", "sub_path": "Scripts/misc/runme_acq_med_original.py", "file_name": "runme_acq_med_original.py", "file_ext": "py", "file_size_in_byte": 9602, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "sys.path.insert", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.walk", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 91, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 103, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path", "line_number": 114, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 115, "usage_type": "call"}, {"api_name": "json.load", "line_number": 198, "usage_type": "call"}, {"api_name": "os.system", "line_number": 214, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 215, "usage_type": "call"}]} +{"seq_id": "577359747", "text": "import bamgen.bamgen\nimport coverview_.statistics\nimport coverview_.reads\nimport os\nimport pysam\nimport unittest\nimport uuid\n\n\ndef load_bam_into_read_array(file_name):\n read_array = coverview_.reads.pyReadArray()\n\n with pysam.AlignmentFile(file_name, 'rb') as bam_file:\n for read in bam_file:\n read_array.append(read)\n\n return read_array\n\n\nclass TestReadArray(unittest.TestCase):\n \"\"\"\n Here we are testing the ability to read a set of reads from a BAM file into\n memory using the ReadArray class, and to query sub-regions of the\n reads.\n \"\"\"\n def setUp(self):\n self.unique_bam_file_name = str(uuid.uuid4())\n self.unique_index_file_name = self.unique_bam_file_name + \".bai\"\n\n def tearDown(self):\n os.remove(self.unique_bam_file_name)\n os.remove(self.unique_index_file_name)\n\n def test_empty_read_array_counts_zero_reads_in_interval(self):\n read_sets = [\n (\"1\", 32, 100, 0)\n ]\n\n bamgen.bamgen.make_bam_file(self.unique_bam_file_name, read_sets)\n read_array = load_bam_into_read_array(self.unique_bam_file_name)\n\n assert read_array.count_reads_in_interval(32, 132) == 0\n\n def test_read_array_with_single_read_counts_one_for_interval_overlapping_read(self):\n read_sets = [\n (\"1\", 32, 100, 1)\n ]\n\n bamgen.bamgen.make_bam_file(self.unique_bam_file_name, read_sets)\n read_array = load_bam_into_read_array(self.unique_bam_file_name)\n\n assert read_array.count_reads_in_interval(32, 132) == 1\n\n def test_read_array_with_single_read_counts_zero_for_interval_not_overlapping_read(self):\n read_sets = [\n (\"1\", 32, 100, 1)\n ]\n\n bamgen.bamgen.make_bam_file(self.unique_bam_file_name, read_sets)\n read_array = load_bam_into_read_array(self.unique_bam_file_name)\n\n assert read_array.count_reads_in_interval(0, 31) == 0\n assert read_array.count_reads_in_interval(132, 133) == 0\n\n\nif __name__ == \"__main__\":\n unittest.main()\n", "sub_path": "test/unit/coverview/test_reads.py", "file_name": "test_reads.py", "file_ext": "py", "file_size_in_byte": 2039, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "coverview_.statistics.reads.pyReadArray", "line_number": 11, "usage_type": "call"}, {"api_name": "coverview_.statistics.reads", "line_number": 11, "usage_type": "attribute"}, {"api_name": "coverview_.statistics", "line_number": 11, "usage_type": "name"}, {"api_name": "pysam.AlignmentFile", "line_number": 13, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 20, "usage_type": "attribute"}, {"api_name": "uuid.uuid4", "line_number": 27, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 31, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 32, "usage_type": "call"}, {"api_name": "bamgen.bamgen.bamgen.make_bam_file", "line_number": 39, "usage_type": "call"}, {"api_name": "bamgen.bamgen.bamgen", "line_number": 39, "usage_type": "attribute"}, {"api_name": "bamgen.bamgen", "line_number": 39, "usage_type": "name"}, {"api_name": "bamgen.bamgen.bamgen.make_bam_file", "line_number": 49, "usage_type": "call"}, {"api_name": "bamgen.bamgen.bamgen", "line_number": 49, "usage_type": "attribute"}, {"api_name": "bamgen.bamgen", "line_number": 49, "usage_type": "name"}, {"api_name": "bamgen.bamgen.bamgen.make_bam_file", "line_number": 59, "usage_type": "call"}, {"api_name": "bamgen.bamgen.bamgen", "line_number": 59, "usage_type": "attribute"}, {"api_name": "bamgen.bamgen", "line_number": 59, "usage_type": "name"}, {"api_name": "unittest.main", "line_number": 67, "usage_type": "call"}]} +{"seq_id": "454210115", "text": "from flask import Flask,request\nfrom flask_cors import CORS\nfrom pymongo import MongoClient\nfrom flask.json import jsonify\nimport json\n\nfrom constants import *\nfrom models import Events\n\napp = Flask(__name__)\n\nclient = MongoClient(HOST + '://' + USERNAME_PASSWORD + '@' + URL + '/' + DB_NAME)\ndb = client[DB_NAME]\nCORS(app)\n\n@app.route(\"/\")\ndef home():\n return \"AiBizConnect API\"\n\n@app.route(\"/addData\", methods = [\"POST\"])\ndef addData():\n # Raise 500 error if event name not given\n if not json.loads(request.data).get('eventName'):\n raise AssertionError\n\n # Add data and convert to Event model class\n event = Events(json.loads(request.data))\n event = json.dumps(event.__dict__)\n event = json.loads(event)\n data = db[EVENT_TABLE].insert({'event':event})\n return \"Added Successfully\"\n\n\n@app.route('/getData', methods = ['GET'])\ndef getData():\n data = db[EVENT_TABLE].find({})\n events_obj = []\n for val in data:\n event = Events(val['event'])\n event = json.dumps(event.__dict__)\n event = json.loads(event)\n events_obj.append(event)\n return jsonify(events_obj)\n\nif __name__ == \"__main__\":\n app.run()\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1174, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "flask.Flask", "line_number": 10, "usage_type": "call"}, {"api_name": "pymongo.MongoClient", "line_number": 12, "usage_type": "call"}, {"api_name": "flask_cors.CORS", "line_number": 14, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 23, "usage_type": "call"}, {"api_name": "flask.request.data", "line_number": 23, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 23, "usage_type": "name"}, {"api_name": "models.Events", "line_number": 27, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 27, "usage_type": "call"}, {"api_name": "flask.request.data", "line_number": 27, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 27, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 28, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 29, "usage_type": "call"}, {"api_name": "models.Events", "line_number": 39, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 40, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 41, "usage_type": "call"}, {"api_name": "flask.json.jsonify", "line_number": 43, "usage_type": "call"}]} +{"seq_id": "554647373", "text": "# -*- coding:utf-8 -*-\n# todo :\n# DataPreProcess.__instance = object.__new__(cls, *args, **kwd)\n# use feature union\n# use abstract class\nimport os\nimport traceback\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom sklearn import preprocessing\nfrom DataPreProcess.CorrelationComputer import CorrelationComputer\nfrom DataSet.DataLoader import DataLoader\nfrom DataPreProcess.DataCleaner import DataCleaner\nfrom DataPreProcess.DataConverter import DataConverter\nfrom DataPreProcess.DataAnalyzer import DataAnalyzer\nfrom Util.Logger.Logger import Logger\n\n\nclass DataPreProcess(object):\n \"\"\"\n 输入:特征名称、特征矩阵、保存路径?,返回新的特征名称、特征矩阵,附加信息?还是直接log或者print、或者出图?\n \"\"\"\n __instance = None\n\n def __new__(cls, *args, **kwd):\n if DataPreProcess.__instance is None:\n DataPreProcess.__instance = object.__new__(cls)\n # DataPreProcess.__instance = object.__new__(cls, *args, **kwd) # 修改这行之后,还是单例么?测试\n return DataPreProcess.__instance\n\n def __init__(self, file_path, sheet_name, index_name, output_dir):\n self._loader = DataLoader()\n self._cleaner = DataCleaner\n self._converter = DataConverter()\n self._analyzer = DataAnalyzer()\n self._logger = Logger(os.path.realpath(__file__)).get_logger()\n self._htic = self._loader.read_excel(file_path, sheet_name, index_name)\n self._correlation_computer = CorrelationComputer()\n self._output_dir = output_dir\n\n def get_scalered_max_min_features(self, data):\n \"\"\"\n 返回最大最小规范化后的特征矩阵\n :param data: array-like形式的特征矩阵\n \"\"\"\n return preprocessing.MinMaxScaler().fit(data).transform(data)\n\n def get_normalized_features(self, norm, data):\n \"\"\"\n 返回归一化后的特征矩阵\n :param norm: 值为“L1”, 或“L2”\n :param data: array-like形式的特征矩阵\n \"\"\"\n return preprocessing.Normalizer(norm).fit(data).transform(data)\n\n def get_standardlized_features(self, data):\n \"\"\"\n 返回z-score标准化后的特征矩阵\n :param data: array-like形式的特征矩阵\n \"\"\"\n s = preprocessing.StandardScaler().fit(data)\n standarded_data = s.transform(data)\n return standarded_data, s.std_, s.mean_\n\n def get_binarized_targets_with_threold(self, target, threold):\n \"\"\"\n 返回根据��值二值化后的二分类目标向量,返回结果为单列的二维数组\n 注意:这里设置copy为False,会直接将target也二值化,但仍然保留一维数组形式\n :param target: 目标向量\n :param threold: 阈值\n \"\"\"\n return preprocessing.Binarizer(threold, copy=False).fit(target).transform(target)\n\n def get_binarized_target_with_threshold_and_NULL(self, target, threold):\n \"\"\"\n 为在精准率、召回率等评价指标中使用默认的pos_label=1,小于阈值依然为0,大于阈值依然为1,空值特设为2, 返回阈值二值化后的三分类目标向量\n :param target: 目标向量\n :param threold: 阈值\n \"\"\"\n try:\n for i in range(len(target)):\n if np.isnan(target[i]):\n target[i] = 2\n elif float(target[i]) < threold:\n target[i] = 0\n elif float(target[i]) >= threold:\n target[i] = 1\n except:\n self._logger.exception('get_binarized_target_with_threshold_and_NULL occur except: threold: ' + str(threold) + ' ' + traceback.format_exc())\n return target.astype(np.int16)\n\n def process_target(self, type=0, threshold=0.1):\n # 0代表二分类\n\n # 类别特征二值化\n self.get_binarized_targets_with_threold(self._htic.target, threshold)\n\n # 基于z-score的标准化\n self._htic.data, std, mean = self.get_standardlized_features(self._htic.data)\n\n # 删除最低特征值方差的特征\n self._htic.feature_names = np.delete(self._htic.feature_names, np.s_[std.argmin(): std.argmin() + 1], axis=0)\n self._htic.data = np.delete(self._htic.data, np.s_[std.argmin():std.argmin() + 1], axis=1)\n\n # 基于特征与目标变量相关性评价,做单变量特征选择\n methods, method_scores = self.get_feature_coefficent_based_different_methods()\n self.plot_feature_coefficent(methods, self._htic.feature_names, method_scores)\n output_dir = os.path.join(self._output_dir, str(threshold))\n if not os.path.isdir(output_dir):\n os.makedirs(output_dir)\n plt.savefig(os.path.join(output_dir, 'Feature Scores.png'))\n\n # 删除比较低的几个特征,目前需要结合特征评分图来手工输入,以后设计为自动选择30%\n deleted__feature_names = ['C', 'ALPHA', 'BETA', 'GAMMA', 'BADER'] # band-gap\n deleted__feature_names = ['C', 'ALPHA', 'GAMMA', 'ACNA', 'BADER', 'AEN'] # vk\n self._htic.feature_names, self._htic.data = \\\n self._delete_low_relative_feature_names_and_features(deleted__feature_names)\n\n # 基于归特征淘汰(RFE)、L1-based、随机sparse、树、PipeLine的特征选择\n # http://d0evi1.com/sklearn/feature_selection/\n\n return self._htic\n\n def process_regression_data(self):\n # 特征标准化\n self._htic.data, std, mean = self.get_standardlized_features(self._htic.data)\n\n # 删除最低方差特征\n self._htic.feature_names = np.delete(self._htic.feature_names, np.s_[std.argmin(): std.argmin() + 1], axis=0)\n self._htic.data = np.delete(self._htic.data, np.s_[std.argmin():std.argmin() + 1], axis=1)\n\n # 基于特征与目标变量相关性评价,做单变量特征选择\n methods, method_scores = self.get_feature_coefficent_based_different_methods()\n self.plot_feature_coefficent(methods, self._htic.feature_names, method_scores)\n # output_dir = os.path.join(self._output_dir, str(threshold))\n # if not os.path.isdir(output_dir):\n # os.makedirs(output_dir)\n # plt.savefig(os.path.join(output_dir, 'Feature Scores.png'))\n\n # 删除相关性比较低的几个特征\n # todo: 目前需要结合特征评分图来手工输入,以后设计为自动选择30%\n deleted__feature_names = ['C', 'ALPHA', 'BETA', 'GAMMA', 'BADER'] # band-gap\n self._htic.feature_names, self._htic.data = \\\n self._delete_low_relative_feature_names_and_features(deleted__feature_names)\n return self._htic\n\n def process_binary_class_data(self, threshold=0.1):\n # 类别特征二值化\n self.get_binarized_targets_with_threold(self._htic.target, threshold)\n\n # 基于z-score的标准化\n self._htic.data, std, mean = self.get_standardlized_features(self._htic.data)\n\n # 删除最低特征值方差的特征\n self._htic.feature_names = np.delete(self._htic.feature_names, np.s_[std.argmin(): std.argmin() + 1], axis=0)\n self._htic.data = np.delete(self._htic.data, np.s_[std.argmin():std.argmin() + 1], axis=1)\n\n # 基于特征与目标变量相关性评价,做单变量特征选择\n methods, method_scores = self.get_feature_coefficent_based_different_methods()\n self.plot_feature_coefficent(methods, self._htic.feature_names, method_scores)\n output_dir = os.path.join(self._output_dir, str(threshold))\n if not os.path.isdir(output_dir):\n os.makedirs(output_dir)\n plt.savefig(os.path.join(output_dir, 'Feature Scores.png'))\n\n # 删除比较低的几个特征,目前需要结合特征评分图来手工输入,以后设计为自动选择30%\n deleted__feature_names = ['C', 'ALPHA', 'BETA', 'GAMMA', 'BADER'] # band-gap\n self._htic.feature_names, self._htic.data = \\\n self._delete_low_relative_feature_names_and_features(deleted__feature_names)\n\n # 基于归特征淘汰(RFE)、L1-based、随机sparse、树、PipeLine的特征选择\n # http://d0evi1.com/sklearn/feature_selection/\n\n return self._htic\n # return self._htic.indices, self._htic.feature_names, self._htic.data, self._htic.target_names, self._htic.target\n\n def process_3_class_data(self, threshold=600000):\n # 目标变量离散化,缺失值为0,小于阈值为1,大于阈值为2\n self._loader.binarizer_target_with_NULL(threshold)\n\n # 基于z-score的标准化\n self._htic.data, std, mean = self.get_standardlized_features(self._htic.data)\n\n # 删除最低特征值方差的特征\n self._htic.feature_names = np.delete(self._htic.feature_names, np.s_[std.argmin(): std.argmin() + 1], axis=0)\n self._htic.data = np.delete(self._htic.data, np.s_[std.argmin():std.argmin() + 1], axis=1)\n\n # 基于特征与目标变量相关性评价,做单变量特征选择\n methods, method_scores = self.get_feature_coefficent_based_different_methods()\n self.plot_feature_coefficent(methods, self._htic.feature_names, method_scores)\n output_dir = os.path.join(self._output_dir, str(threshold))\n if not os.path.isdir(output_dir):\n os.makedirs(output_dir)\n plt.savefig(os.path.join(output_dir, 'Feature Scores.png'))\n\n # 删除比较低的几个特征,目前需要结合特征评分图来手工输入,以后设计为自动选择30%\n deleted__feature_names = ['C', 'ALPHA', 'GAMMA', 'ACNA', 'BADER', 'AEN'] # vk\n self._htic.feature_names, self._htic.data = \\\n self._delete_low_relative_feature_names_and_features(deleted__feature_names)\n\n # 基于归特征淘汰(RFE)、L1-based、随机sparse、树、PipeLine的特征选择\n # http://d0evi1.com/sklearn/feature_selection/\n\n return self._htic\n # return self._htic.indices, self._htic.feature_names, self._htic.data, self._htic.target_names, self._htic.target\n\n def get_feature_coefficent_based_different_methods(self):\n \"\"\" 返回基于相关性的单变量特征选择所用的方法及其绝对值评分 \"\"\"\n self.generate_feature_score(self._htic.feature_names, self._htic.data, self._htic.target_names, self._htic.target)\n methods = list()\n method_scores = list()\n for i, compute_type in enumerate(self._correlation_computer.features_scores):\n methods.append(compute_type)\n abs_ndarray = np.abs(list(self._correlation_computer.features_scores[compute_type].values()))\n method_scores.append(abs_ndarray)\n return methods, method_scores\n\n def generate_feature_score(self, feature_names, features, target_names, target):\n # 特征评价, pandas pearson\n feature_names, scores = self._correlation_computer.get_correlation_by_pandas_pearson(feature_names, features, target_names, target, 2, 'pearson')\n self._correlation_computer.add_feature_scores('pd_pearson', feature_names, scores)\n\n # 特征评价, pandas kendall\n feature_names, scores = self._correlation_computer.get_correlation_by_pandas_pearson(feature_names, features, target_names, target, 2, 'kendall')\n self._correlation_computer.add_feature_scores('pd_kendall', feature_names, scores)\n\n # 特征评价, pandas spearman\n feature_names, scores = self._correlation_computer.get_correlation_by_pandas_pearson(feature_names, features, target_names, target, 2, 'spearman')\n self._correlation_computer.add_feature_scores('pd_spearman', feature_names, scores)\n\n # # 特征评价, scikit rfe\n # sk_rfe, data = corr_computer.select_k_best_by_rfe(normed_data, target, 2)\n # corr_computer.add_feature_scores('sk_rfe', feature_names, sk_rfe.ranking_)\n\n # # 特征评价, scikit mic\n # sk_mic, data = corr_computer.select_k_best_by_mic(normed_data, target, 2)\n # corr_computer.add_feature_scores('sk_mic', feature_names, sk_mic.scores_)\n\n # # 特征评价, scikit pearson\n # sk_pearson, data = corr_computer.select_k_best_by_pearson(normed_data, target, 'all')\n # corr_computer.add_feature_scores('sk_pearson', feature_names, sk_pearson.scores_)\n\n # #特征评价\n # sk_tree, data = corr_computer.select_k_best_by_tree(normed_data, target, 2)\n # print(feature_names)\n # print(sk_tree)\n\n # # 特征评价\n # sk_lr, data = corr_computer.select_k_best_by_logistic_regression(normed_data, target, 2)\n # print(feature_names)\n # print(sk_lr)\n\n def plot_feature_coefficent(self, methods, feature_names, method_scores):\n \"\"\"\n 画出特征评分图\n :param methods: 特征评分所用的方法列表\n :param method_scores: 每个方法所对应的不同的特征的评分列表\n \"\"\"\n # plt.subplot(1, 1, 1)\n plt.clf()\n plt.figure(num=\"Feature Scores\", figsize=(12, 8))\n index = np.arange(len(feature_names)) * 2\n bar_width = 0.2\n colors = ['b', 'g', 'r', 'c', 'm', 'y', 'k', 'w']\n for i in range(len(methods)):\n plt.bar(index + i * bar_width, method_scores[i], bar_width, alpha=0.9, color=colors[i], label=methods[i])\n plt.xlabel('features')\n plt.ylabel('scores')\n plt.title('Scores of feature')\n plt.xticks(index + bar_width, feature_names)\n plt.ylim()\n plt.legend()\n # plt.show()\n\n def _delete_low_relative_feature_names_and_features(self, deleted__feature_names):\n for name in deleted__feature_names:\n index = np.where(self._htic.feature_names == name)\n self._htic.feature_names = np.delete(self._htic.feature_names, index, axis=0)\n self._htic.data = np.delete(self._htic.data, index, axis=1)\n return self._htic.feature_names, self._htic.data\n\n def rank_feature_scores_by_sbs(self, feature_names, features, target):\n # 通过SBS算法,选取寻找最优子集,寻找最高平跟标准\n from sklearn.neighbors import KNeighborsClassifier\n from Example.DataPreProcess.sbs import SBS\n\n clf = KNeighborsClassifier(n_neighbors=2)\n # clf = LogisticRegression(C=100) # 不管值是0.1 1 10 100,在特殊为10、11、12、13时最大\n\n sbs = SBS(clf, k_features=1)\n sbs.fit(features, target)\n k_feat = [len(k) for k in sbs.subsets_] # sbs.subsets_ 中数据个数为14,13,12,11...,k_feat中值为14,13,12,11...\n # for k in sbs.subsets_:\n # print(k)\n # print(sbs.scores_)\n # print(k_feat)\n plt.plot(k_feat, sbs.scores_, marker = 'o')\n plt.xlim(0, 16, 1) # 这里假如不限定xlim的话,则自动默认按照k_feat 进行排列,对应的调整后面的scores\n plt.ylim(0.1, 1.1) # 这里如果不限定ylim的话,会自动的按照sbs.scores_的范围来显示\n plt.ylabel('Recall score')\n plt.xlabel('number of features')\n plt.grid()\n plt.show()\n\n # 显示最多个数序号为11时,实际为sbs.subsets_中[0]14[1]13[2]12[3]11 特征数即14-11,显示个数为2时,实际为14-3=11号位置处\n k4 = list(sbs.subsets_[0])\n print(k4)\n print(feature_names[k4])\n pass\n\n def rank_feature_scores_by_random_forest(self, feature_names, features, target):\n # 通过随机森林判定特征重要性\n from sklearn.ensemble import RandomForestClassifier\n forest = RandomForestClassifier(n_estimators=1000,\n random_state=14,\n n_jobs=1)\n forest.fit(features, target)\n importances = forest.feature_importances_\n indices = np.argsort(importances)[::-1]\n for f in range(features.shape[1]):\n print('%2d %-*s %f' % (f + 1, 30, feature_names[f], importances[indices[f]]))\n\n plt.title('Feature Importance')\n plt.bar(range(features.shape[1]), importances[indices], color=['b'], align='center')\n plt.xticks(range(features.shape[1]), feature_names, rotation=90)\n plt.xlim([-1, features.shape[1]])\n plt.tight_layout()\n plt.show()\n\n def select_feature_based_l1(self, feature_names, features, target):\n # todo 待测试\n \"\"\" 返回重要性大于阈值的特征,此处例子为1.25*mean\"\"\"\n from sklearn.svm import LinearSVC\n from sklearn.feature_selection import SelectFromModel\n # 基于L1的LinearSVC选择特征\n lsvc = LinearSVC(C=0.01, penalty='l1', dual=False).fit(features, target)\n model = SelectFromModel(lsvc, prefit=True)\n X_lsvc = model.transform(features)\n # X_lsvc.shape\n return X_lsvc\n\n def select_feature_based_tree(self, feature_names, features, target):\n # todo 待测试\n \"\"\" 返回重要性大于阈值的特征,此处例子为1.25*mean\"\"\"\n from sklearn.ensemble import ExtraTreesClassifier\n from sklearn.feature_selection import SelectFromModel\n # 基于树模型进行模型选择\n clf = ExtraTreesClassifier()\n clf = clf.fit(features,target)\n # 打印特征重要性(数值越高特征越重要)\n print(clf.feature_importances_)\n\n # 选择特征重要性为1.25倍均值的特征\n model = SelectFromModel(clf, threshold='1.25*mean', prefit=True)\n # 返回所选的特征\n X_trees = model.transform(features)\n # X_trees.shape\n return X_trees\n\n\nif __name__ == '__main__':\n dp = DataPreProcess(r'D:\\File\\Thesis\\A_Framework_of_HTIC\\01_Data\\training256.xls',\n 'band_gap_0713',\n 'formula')\n simple_names, feature_names, features, target_name, target = dp.process_binary_class_data(threshold=0.1)\n # dp.rank_feature_scores_by_sbs(feature_names, features, target)\n # dp.rank_feature_scores_by_random_forest(feature_names, features, target)\n\n # dp = DataPreProcess(r'D:\\File\\Thesis\\A_Framework_of_HTIC\\01_Data\\training256.xls',\n # 'vk_0713',\n # 'formula')\n # simple_names, feature_names, features, target_name, target = dp.process_3_class_data(threshold=0.1)\n\n\n\n# 后续可补充:\n# process_mutilabel_class_data() 对目标变量,先进行OneHotEncoder,然后处理\n\n# 删除特征值的方差达不到最低标准的特征,VarianceThreshold默认删除0方差特征,返回大于阈值的特征,可以基于需要判断的概率来推测方差,比如伯努利型\n# features = feature_selection.VarianceThreshold(threshold=(.8 * (1 - .8))).transform(self._htic.data)\n\n# 在单变量特征选择过程中,可以将评分结果同意缩放到[0,1]区间,以方便对比结果\n# maxmined_values = self.preprocessing_maxmin_features(list(self.corr_computer.features_scores[compute_type].values()))\n# abs_ndarray = np.abs(maxmined_values)\n# score_max.append(maxmined_values)\n# print(list(self.corr_computer.features_scores[compute_type].values()), compute_type)\n# print(list(maxmined_values), compute_type, 'after maxmin methods')\n\n# testdata = pd.DataFrame({'pet': ['cat', 'dog', 'dog', 'fish'],'age': [4 , 6, 3, 3], 'salary':[4, 5, 1, 1]})\n# 对某一数字型类别特征独热编码,如化合物类型,OneHotEncoder输入必须是二维数组,前提是训练集中已经包含了所有可类别结果\n# example_result = OneHotEncoder(sparse=False).fit_transform(testdata[['age']]\n# a2 = OneHotEncoder(sparse=False).fit_transform(testdata[['salary']])\n# final_output = numpy.hstack((a1, a2))\n\n# 对目标变量或者特征标签离散化,LabelEncoder只支持一维数组\n# a = LabelEncoder().fit_transform(testdata['pet'])\n# 顺便进行独热编码\n# OneHotEncoder(sparse=False).fit_transform(a.reshape(-1, 1)) # -1 代表自动根据一维数组总数和行列总数来推测\n\n# 对某个特征进行二值化,必要时进行独热编码,a one-vs-all fashion\n# LabelBinarizer().fit_transform(testdata['pet'])\n\n# 多标签二值化并独热编码,如将两个特征的值组,作为一个对象7联合在一起编码\n# MultiLabelBinarizer().fit_transform(testdata[['age','salary']].values", "sub_path": "DataPreProcess/DataPreprocess.py", "file_name": "DataPreprocess.py", "file_ext": "py", "file_size_in_byte": 21497, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "DataSet.DataLoader.DataLoader", "line_number": 33, "usage_type": "call"}, {"api_name": "DataPreProcess.DataCleaner.DataCleaner", "line_number": 34, "usage_type": "name"}, {"api_name": "DataPreProcess.DataConverter.DataConverter", "line_number": 35, "usage_type": "call"}, {"api_name": "DataPreProcess.DataAnalyzer.DataAnalyzer", "line_number": 36, "usage_type": "call"}, {"api_name": "Util.Logger.Logger.Logger", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.realpath", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "DataPreProcess.CorrelationComputer.CorrelationComputer", "line_number": 39, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 47, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 47, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.Normalizer", "line_number": 55, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 55, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 62, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 62, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.Binarizer", "line_number": 73, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 73, "usage_type": "name"}, {"api_name": "numpy.isnan", "line_number": 83, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 91, "usage_type": "attribute"}, {"api_name": "numpy.delete", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.s_", "line_number": 103, "usage_type": "attribute"}, {"api_name": "numpy.delete", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.s_", "line_number": 104, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path", "line_number": 109, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 110, "usage_type": "call"}, {"api_name": "os.path", "line_number": 110, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path", "line_number": 112, "usage_type": "attribute"}, {"api_name": "numpy.delete", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.s_", "line_number": 130, "usage_type": "attribute"}, {"api_name": "numpy.delete", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.s_", "line_number": 131, "usage_type": "attribute"}, {"api_name": "numpy.delete", "line_number": 156, "usage_type": "call"}, {"api_name": "numpy.s_", "line_number": 156, "usage_type": "attribute"}, {"api_name": "numpy.delete", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.s_", "line_number": 157, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 162, "usage_type": "call"}, {"api_name": "os.path", "line_number": 162, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path", "line_number": 163, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 165, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 165, "usage_type": "call"}, {"api_name": "os.path", "line_number": 165, "usage_type": "attribute"}, {"api_name": "numpy.delete", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.s_", "line_number": 186, "usage_type": "attribute"}, {"api_name": "numpy.delete", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.s_", "line_number": 187, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 192, "usage_type": "call"}, {"api_name": "os.path", "line_number": 192, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 193, "usage_type": "call"}, {"api_name": "os.path", "line_number": 193, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 194, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 195, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 195, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 195, "usage_type": "call"}, {"api_name": "os.path", "line_number": 195, "usage_type": "attribute"}, {"api_name": "numpy.abs", "line_number": 215, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 261, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 261, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 262, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 262, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 263, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 267, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 267, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 268, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 268, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 269, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 269, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 270, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 270, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 271, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 271, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 272, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 272, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 273, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 273, "usage_type": "name"}, {"api_name": "numpy.where", "line_number": 278, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 279, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 280, "usage_type": "call"}, {"api_name": "sklearn.neighbors.KNeighborsClassifier", "line_number": 288, "usage_type": "call"}, {"api_name": "Example.DataPreProcess.sbs.SBS", "line_number": 291, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 298, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 298, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 299, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 299, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 300, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 300, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 301, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 301, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 302, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 302, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 303, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 303, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 304, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 304, "usage_type": "name"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 315, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 320, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 324, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 324, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 325, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 325, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 326, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 326, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 327, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 327, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 328, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 328, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 329, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 329, "usage_type": "name"}, {"api_name": "sklearn.svm.LinearSVC", "line_number": 337, "usage_type": "call"}, {"api_name": "sklearn.feature_selection.SelectFromModel", "line_number": 338, "usage_type": "call"}, {"api_name": "sklearn.ensemble.ExtraTreesClassifier", "line_number": 349, "usage_type": "call"}, {"api_name": "sklearn.feature_selection.SelectFromModel", "line_number": 355, "usage_type": "call"}, {"api_name": "{'KNeighborsClassifier': 'sklearn.neighbors.KNeighborsClassifier', 'SBS': 'Example.DataPreProcess.sbs.SBS', 'RandomForestClassifier': 'sklearn.ensemble.RandomForestClassifier', 'LinearSVC': 'sklearn.svm.LinearSVC', 'SelectFromModel': 'sklearn.feature_selection.SelectFromModel', 'ExtraTreesClassifier': 'sklearn.ensemble.ExtraTreesClassifier'}", "line_number": 363, "usage_type": "call"}]} +{"seq_id": "487188562", "text": "import configparser\nimport psycopg2\nfrom sql_queries import create_table_queries, drop_table_queries\n\n\ndef drop_tables(cur, conn):\n \n '''This function connects to a database, and drops any existing tables found in the drop_table_queries \n list in the sql_queries.py script.'''\n \n for query in drop_table_queries:\n try:\n cur.execute(query)\n conn.commit()\n except psycopg2.Error as e:\n print(\"Error: Issue dropping table: \" + query)\n print(e)\n print(\"Existing tables dropped\")\n\n\ndef create_tables(cur, conn):\n \n '''This function connects to a database and creates the tables defined in the create_table_queries list found in \n the sql_queries.py script.'''\n \n for query in create_table_queries:\n try:\n cur.execute(query)\n conn.commit()\n except psycopg2.Error as e:\n print(\"Error: Issue creating table: \" + query)\n print(e)\n print(\"New tables created\")\n\n\ndef main():\n \n '''Main function to create tables.\n This function reads a config file with the Redshift paramaters defined and connects to the database, drops \n any existing tables and creates tables in sql_queries.py file and closes the database connection.\n \n '''\n config = configparser.ConfigParser()\n config.read('dwh.cfg')\n\n conn = psycopg2.connect(\"host={} dbname={} user={} password={} port={}\".format(*config['CLUSTER'].values()))\n cur = conn.cursor()\n\n drop_tables(cur, conn)\n create_tables(cur, conn)\n\n conn.close()\n\n\nif __name__ == \"__main__\":\n main()", "sub_path": "create_tables.py", "file_name": "create_tables.py", "file_ext": "py", "file_size_in_byte": 1625, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "sql_queries.drop_table_queries", "line_number": 11, "usage_type": "name"}, {"api_name": "psycopg2.Error", "line_number": 15, "usage_type": "attribute"}, {"api_name": "sql_queries.create_table_queries", "line_number": 26, "usage_type": "name"}, {"api_name": "psycopg2.Error", "line_number": 30, "usage_type": "attribute"}, {"api_name": "configparser.ConfigParser", "line_number": 43, "usage_type": "call"}, {"api_name": "psycopg2.connect", "line_number": 46, "usage_type": "call"}]} +{"seq_id": "252590548", "text": "import numpy as np\nimport cv2\nfrom common_func import *\n\n\nclass predExtractor:\n \"\"\"\n This class further extract information based on\n the result of recognizer apply with sepecific\n rules as constrains\n \"\"\"\n def __init__(self, predictions, probabilities, boxes):\n \"\"\" Initializer \"\"\"\n self.preds = predictions\n self.probs = probabilities\n self.perplexity = self.get_correlation_m\n self.boxes = boxes\n\n def dump_preds(self):\n \"\"\" dump predictions if perplexity is high \"\"\"\n self.preds = np.zeros((self.preds.shape), dtype=self.preds.dtype)\n\n def zero_removal(self):\n for i in range(len(self.preds.flatten())):\n self.probs[i, 0] = -np.float('inf')\n self.preds[i] = np.argmax(self.probs[i,:])\n\n def get_pred(self):\n self.preds = np.argmax(self.probs, axis=1)\n return self.preds\n\n def number_redundancy_removal3(self):\n duplicates_num, duplicates = self.find_duplicates()\n missed = self.find_missed_number()\n\n def fresh(row):\n for col in range(10):\n if col not in missed:\n self.probs[row, col] = -np.float('inf')\n\n while max(duplicates_num)>1:\n sorted_p = np.sort(self.probs)\n differences = sorted_p[:,9] - sorted_p[:,8]\n freshed_rows = []\n\n for item in duplicates:\n if item!='x' and duplicates.count(item)>1:\n\n # get all the item's index\n indexs = []\n while item in duplicates:\n i = duplicates.index(item)\n indexs.append(i)\n duplicates[i] = \"x\"\n\n # find max prob for this redundancy item\n max_p = None\n col = item\n for row in indexs:\n if max_p is None:\n max_p = differences[row]\n if differences[row] < max_p:\n fresh(row)\n freshed_rows.append(row)\n else:\n max_p = differences[row]\n\n for row in indexs:\n if differences[row] < max_p:\n fresh(row)\n freshed_rows.append(row)\n\n self.get_pred()\n # reset freshed rows' preds\n for row in freshed_rows:\n self.preds[row] = 0\n duplicates_num, duplicates = self.find_duplicates()\n missed = self.find_missed_number()\n\n # print \"pred:\"\n # print self.preds\n # print \"duplicates:\"\n # print duplicates\n # print \"duplicates_num:\"\n # print duplicates_num\n # print self.probs\n # raw_input(\"----\")\n\n missed = self.find_missed_number()\n # print missed\n preds_list = list(self.preds.flatten())\n if 0 in preds_list:\n self.preds[preds_list.index(0)] = missed[0]\n # self.zero_removal()\n\n\n def number_redundancy_removal2(self):\n duplicates_num, duplicates = self.find_duplicates()\n while max(duplicates_num)>1:\n sorted_p = np.sort(self.probs)\n differences = sorted_p[:,9] - sorted_p[:,8]\n\n for item in duplicates:\n if item!='x' and duplicates.count(item)>1:\n indexs = []\n # print \"item: \", item\n\n # get all the item's index\n while item in duplicates:\n i = duplicates.index(item)\n indexs.append(i)\n duplicates[i] = \"x\"\n\n # find max prob for this redundancy item\n max_p = None\n col = item\n for row in indexs:\n if max_p is None:\n max_p = differences[row]\n if differences[row] < max_p:\n self.probs[row, col] = -np.float('inf')\n else:\n max_p = differences[row]\n\n\n for row in indexs:\n if differences[row] < max_p:\n self.probs[row, col] = -np.float('inf')\n\n self.get_pred()\n duplicates_num, duplicates = self.find_duplicates()\n\n # print \"pred:\"\n # print self.preds\n # print \"duplicates:\"\n # print duplicates\n # print \"duplicates_num:\"\n # print duplicates_num\n # print self.probs\n # raw_input(\"----\")\n\n\n\n def number_redundancy_removal(self):\n duplicates_num, duplicates = self.find_duplicates()\n while max(duplicates_num)>1:\n for item in duplicates:\n if item!='x' and duplicates.count(item)>1:\n indexs = []\n # print \"item: \", item\n\n # get all the item's index\n while item in duplicates:\n i = duplicates.index(item)\n indexs.append(i)\n duplicates[i] = \"x\"\n\n # find max prob for this redundancy item\n max_p = None\n col = item\n for row in indexs:\n if max_p is None:\n max_p = self.probs[row, col]\n if self.probs[row, col] < max_p:\n self.probs[row, col] = -np.float('inf')\n else:\n max_p = self.probs[row, col]\n\n\n for row in indexs:\n if self.probs[row, col] < max_p:\n self.probs[row, col] = -np.float('inf')\n\n self.get_pred()\n duplicates_num, duplicates = self.find_duplicates()\n\n # print \"duplicates:\"\n # print duplicates\n # print \"duplicates_num:\"\n # print duplicates_num\n # print self.probs\n # raw_input(\"----\")\n\n # duplicates_num, duplicates = self.find_duplicates()\n\n def find_missed_number(self):\n contains = list(self.preds.flatten())\n missed = []\n for i in range(1,10):\n if i not in contains:\n missed.append(i)\n return missed\n\n def find_duplicates(self):\n duplicates = []\n duplicates_num = [0, 0, 0, 0, 0, 0, 0, 0, 0, 0]\n for i in range(len(self.preds.flatten())):\n duplicates.append(self.preds[i])\n duplicates_num[self.preds[i]] += 1\n\n # for item in duplicates:\n # if duplicates.count(item)<2:\n # duplicates.pop(duplicates.index(item))\n return duplicates_num, duplicates\n\n def get_perplexity(self):\n \"\"\"\n This method get the average normalized value\n from correlation matrix\n \"\"\"\n m = self.get_correlation_m(self.probs)\n return np.average(m)\n\n def get_correlation_m(self,matrix=None):\n \"\"\"\n This function calculate the correlation matrix\n \"\"\"\n if matrix is None:\n matrix=self.probs\n # correlation matrix calculation\n matrix2 = np.dot(matrix, np.transpose(matrix))\n normed = np.linalg.norm(matrix, axis=1)\n normed2 = np.dot(normed.reshape((-1,1)), normed.reshape((1,-1)))\n correlation = matrix2 / normed2\n return correlation\n def test_function(self):\n new=np.argmax(self.probs,axis=0)[1:]\n if len(np.unique(new))==len(self.preds):\n for i in range(8):\n self.preds[new[i]]=i+1\n else:\n to_do=[0,1,2,3,4,5,6,7,8]\n ans=[0,0,0,0,0,0,0,0,0]\n while(len(to_do)!=0):\n temp=np.argmax(self.probs[to_do[0]])\n if(temp not in ans):\n ans[to_do[0]]=temp\n del to_do[0]\n else:\n index=to_do[0]\n t_index=ans.index(temp)\n t_temp=np.argmax(self.probs[t_index])\n self.probs[index].sort()\n origin_sec=self.probs[index][-3]\n self.probs[t_index].sort()\n to_sec=self.probs[t_index][-3]\n if self.probs[index][temp]>self.probs[t_index][t_temp]:\n ans[t_index]=0\n ans[index]=temp\n del to_do[0]\n to_do.append(t_index)\n else:\n ans[index]=np.where(self.probs[index]==origin_sec)[0][0]\n del to_do[0]\n self.preds=np.array(ans)\n\n def draw_pred(self, target_img):\n \"\"\"\n This function draw predicted number in target_img\n \"\"\"\n i = 0\n for box in self.boxes:\n draw_pred_point = (int(sort_box_points(box)[1][0]),int(sort_box_points(box)[1][1]))\n cv2.putText(target_img,\n str(self.preds[i]), draw_pred_point,\n cv2.FONT_HERSHEY_PLAIN, 3,\n draw_number_color,\n 5)\n i += 1\n", "sub_path": "rune/prediction_selection.py", "file_name": "prediction_selection.py", "file_ext": "py", "file_size_in_byte": 9386, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "numpy.zeros", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 212, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 222, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 223, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 227, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 228, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 235, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 242, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 253, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 255, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 264, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_PLAIN", "line_number": 266, "usage_type": "attribute"}]} +{"seq_id": "52702490", "text": "from datetime import datetime\n\nfrom flask import current_app, flash, g, redirect, render_template, request, url_for\nfrom flask_login import current_user, login_required\n\nfrom app import db\nfrom app.main import bp\nfrom app.main.forms import EditProfileForm, MessageForm, PostForm, SearchForm\nfrom app.models import Message, Post, User\n\n\n@bp.before_app_request\ndef before_request():\n if current_user.is_authenticated:\n current_user.last_seen = datetime.utcnow()\n db.session.commit()\n g.search_form = SearchForm()\n\n\n@bp.route('/', methods=['GET', 'POST'])\n@bp.route('/index', methods=['GET', 'POST'])\n@login_required\ndef index():\n form = PostForm()\n if form.validate_on_submit():\n post = Post(body=form.post.data, author=current_user)\n db.session.add(post)\n db.session.commit()\n flash('successfully created a blog post')\n return redirect(url_for('main.index'))\n\n page = request.args.get('page', 1, type=int)\n posts = current_user.followed_posts().paginate(page, current_app.config['POSTS_PER_PAGE'], False)\n next_url = url_for('main.index', page=posts.next_num) if posts.has_next else None\n prev_url = url_for('main.index', page=posts.prev_num) if posts.has_prev else None\n return render_template('index.html', title='Home', posts=posts.items, form=form,\n next_url=next_url, prev_url=prev_url)\n\n\n@bp.route('/user/')\n@login_required\ndef user(username):\n user = User.query.filter_by(username=username).first_or_404()\n page = request.args.get('page', 1, type=int)\n posts = user.posts.order_by(Post.timestamp.desc()).paginate(page, current_app.config['POSTS_PER_PAGE'], False)\n next_url = url_for('main.user', username=username, page=posts.next_num) if posts.has_next else None\n prev_url = url_for('main.user', username=username, page=posts.prev_num) if posts.has_prev else None\n return render_template('user.html', user=user, posts=posts.items, next_url=next_url, prev_url=prev_url)\n\n\n@bp.route('/edit_profile', methods=['GET', 'POST'])\n@login_required\ndef edit_profile():\n form = EditProfileForm(original_username=current_user.username)\n if form.validate_on_submit():\n current_user.username = form.username.data\n current_user.about_me = form.about_me.data\n db.session.commit()\n flash('Your changes have been saved.')\n return redirect(url_for('main.edit_profile'))\n elif request.method == 'GET':\n form.username.data = current_user.username\n form.about_me.data = current_user.about_me\n return render_template('edit_profile.html', title='Edit Profile', form=form)\n\n\n@bp.route('/follow/')\n@login_required\ndef follow(username):\n user = User.query.filter_by(username=username).first()\n if not user:\n flash(f'User {username} not found')\n return redirect('index')\n elif user == current_user:\n flash(f'you cannot follow yourself')\n return redirect(url_for('main.index', username=user.username))\n\n current_user.follow(user)\n db.session.commit()\n flash(f'Successfully following {username}')\n return redirect(url_for('main.user', username=username))\n\n\n@bp.route('/unfollow/')\n@login_required\ndef unfollow(username):\n user = User.query.filter_by(username=username).first()\n if user is None:\n flash(f'User {username} not found.')\n return redirect(url_for('main.index'))\n if user == current_user:\n flash('You cannot unfollow yourself!')\n return redirect(url_for('main.user', username=username))\n current_user.unfollow(user)\n db.session.commit()\n flash(f'You are not following {username}.')\n return redirect(url_for('user', username=username))\n\n\n@bp.route('/explore')\n@login_required\ndef explore():\n page = request.args.get('page', 1, type=int)\n posts = Post.query.order_by(Post.timestamp.desc()).paginate(page, current_app.config['POSTS_PER_PAGE'], False)\n next_url = url_for('explore', page=posts.next_num) if posts.has_next else None\n prev_url = url_for('explore', page=posts.prev_num) if posts.has_prev else None\n return render_template('index.html', title='Explore', posts=posts.items, next_url=next_url, prev_url=prev_url)\n\n\n@bp.route('/search')\n@login_required\ndef search():\n if not g.search_form.validate():\n return redirect(url_for('main.explore'))\n\n page = request.args.get('page', 1, type=int)\n posts, total = Post.search(g.search_form.q.data, page, current_app.config['POSTS_PER_PAGE'])\n next_url = url_for('main.search', q=g.search_form.q.data, page=page + 1) \\\n if total > page * current_app.config['POSTS_PER_PAGE'] else None\n prev_url = url_for('main.search', q=g.search_form.q.data, page=page - 1) \\\n if page > 1 else None\n return render_template('search.html', title='Search', posts=posts,\n next_url=next_url, prev_url=prev_url)\n\n\n@bp.route('/user//popup')\n@login_required\ndef user_popup(username):\n user = User.query.filter_by(username=username).first_or_404()\n return render_template('user_popup.html', user=user)\n\n\n@bp.route('/send_message/', methods=['GET', 'POST'])\n@login_required\ndef send_message(recipient):\n user = User.query.filter_by(username=recipient).first_or_404()\n form = MessageForm()\n\n if form.validate_on_submit():\n msg = Message(author=current_user, recipient=user, body=form.message.data)\n db.session.add(msg)\n db.session.commit()\n flash('Your message has been sent.')\n return redirect(url_for('main.user', username=recipient))\n\n return render_template('send_message.html', title='Send Message', form=form, recipient=recipient)\n\n\n@bp.route('/messages')\n@login_required\ndef messages():\n current_user.last_message_read_time = datetime.utcnow()\n db.session.commit()\n page = request.args.get('page', 1, type=int)\n messages = current_user.messages_received.order_by(\n Message.timestamp.desc()).paginate(\n page, current_app.config['POSTS_PER_PAGE'], False)\n next_url = url_for('main.messages', page=messages.next_num) \\\n if messages.has_next else None\n prev_url = url_for('main.messages', page=messages.prev_num) \\\n if messages.has_prev else None\n return render_template('messages.html', messages=messages.items,\n next_url=next_url, prev_url=prev_url)\n", "sub_path": "services/web/app/main/routes.py", "file_name": "routes.py", "file_ext": "py", "file_size_in_byte": 6410, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "flask_login.current_user.is_authenticated", "line_number": 14, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 14, "usage_type": "name"}, {"api_name": "flask_login.current_user.last_seen", "line_number": 15, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 15, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 15, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 15, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 16, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 16, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 16, "usage_type": "name"}, {"api_name": "flask.g.search_form", "line_number": 17, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 17, "usage_type": "name"}, {"api_name": "app.main.forms.SearchForm", "line_number": 17, "usage_type": "call"}, {"api_name": "app.main.bp.before_app_request", "line_number": 12, "usage_type": "attribute"}, {"api_name": "app.main.bp", "line_number": 12, "usage_type": "name"}, {"api_name": "app.main.forms.PostForm", "line_number": 24, "usage_type": "call"}, {"api_name": "app.models.Post", "line_number": 26, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 26, "usage_type": "name"}, {"api_name": "app.db.session.add", "line_number": 27, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 27, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 27, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 28, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 28, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 28, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 30, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 30, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 32, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 32, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 32, "usage_type": "name"}, {"api_name": "flask_login.current_user.followed_posts", "line_number": 33, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 33, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 33, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 33, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 34, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 35, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 36, "usage_type": "call"}, {"api_name": "app.main.bp.route", "line_number": 20, "usage_type": "call"}, {"api_name": "app.main.bp", "line_number": 20, "usage_type": "name"}, {"api_name": "app.main.bp.route", "line_number": 21, "usage_type": "call"}, {"api_name": "app.main.bp", "line_number": 21, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 22, "usage_type": "name"}, {"api_name": "app.models.User.query.filter_by", "line_number": 43, "usage_type": "call"}, {"api_name": "app.models.User.query", "line_number": 43, "usage_type": "attribute"}, {"api_name": "app.models.User", "line_number": 43, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 44, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 44, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 44, "usage_type": "name"}, {"api_name": "app.models.Post.timestamp.desc", "line_number": 45, "usage_type": "call"}, {"api_name": "app.models.Post.timestamp", "line_number": 45, "usage_type": "attribute"}, {"api_name": "app.models.Post", "line_number": 45, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 45, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 45, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 46, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 47, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 48, "usage_type": "call"}, {"api_name": "app.main.bp.route", "line_number": 40, "usage_type": "call"}, {"api_name": "app.main.bp", "line_number": 40, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 41, "usage_type": "name"}, {"api_name": "app.main.forms.EditProfileForm", "line_number": 54, "usage_type": "call"}, {"api_name": "flask_login.current_user.username", "line_number": 54, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 54, "usage_type": "name"}, {"api_name": "flask_login.current_user.username", "line_number": 56, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 56, "usage_type": "name"}, {"api_name": "flask_login.current_user.about_me", "line_number": 57, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 57, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 58, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 58, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 58, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 60, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 60, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 61, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 61, "usage_type": "name"}, {"api_name": "flask_login.current_user.username", "line_number": 62, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 62, "usage_type": "name"}, {"api_name": "flask_login.current_user.about_me", "line_number": 63, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 63, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 64, "usage_type": "call"}, {"api_name": "app.main.bp.route", "line_number": 51, "usage_type": "call"}, {"api_name": "app.main.bp", "line_number": 51, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 52, "usage_type": "name"}, {"api_name": "app.models.User.query.filter_by", "line_number": 70, "usage_type": "call"}, {"api_name": "app.models.User.query", "line_number": 70, "usage_type": "attribute"}, {"api_name": "app.models.User", "line_number": 70, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 72, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 73, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 74, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 75, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 76, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 76, "usage_type": "call"}, {"api_name": "flask_login.current_user.follow", "line_number": 78, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 78, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 79, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 79, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 80, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 81, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 81, "usage_type": "call"}, {"api_name": "app.main.bp.route", "line_number": 67, "usage_type": "call"}, {"api_name": "app.main.bp", "line_number": 67, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 68, "usage_type": "name"}, {"api_name": "app.models.User.query.filter_by", "line_number": 87, "usage_type": "call"}, {"api_name": "app.models.User.query", "line_number": 87, "usage_type": "attribute"}, {"api_name": "app.models.User", "line_number": 87, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 89, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 90, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 90, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 91, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 92, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 93, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 93, "usage_type": "call"}, {"api_name": "flask_login.current_user.unfollow", "line_number": 94, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 94, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 95, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 95, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 95, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 96, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 97, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 97, "usage_type": "call"}, {"api_name": "app.main.bp.route", "line_number": 84, "usage_type": "call"}, {"api_name": "app.main.bp", "line_number": 84, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 85, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 103, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 103, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 103, "usage_type": "name"}, {"api_name": "app.models.Post.query.order_by", "line_number": 104, "usage_type": "call"}, {"api_name": "app.models.Post.query", "line_number": 104, "usage_type": "attribute"}, {"api_name": "app.models.Post", "line_number": 104, "usage_type": "name"}, {"api_name": "app.models.Post.timestamp.desc", "line_number": 104, "usage_type": "call"}, {"api_name": "app.models.Post.timestamp", "line_number": 104, "usage_type": "attribute"}, {"api_name": "flask.current_app.config", "line_number": 104, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 104, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 105, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 106, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 107, "usage_type": "call"}, {"api_name": "app.main.bp.route", "line_number": 100, "usage_type": "call"}, {"api_name": "app.main.bp", "line_number": 100, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 101, "usage_type": "name"}, {"api_name": "flask.g.search_form.validate", "line_number": 113, "usage_type": "call"}, {"api_name": "flask.g.search_form", "line_number": 113, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 113, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 114, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 114, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 116, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 116, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 116, "usage_type": "name"}, {"api_name": "app.models.Post.search", "line_number": 117, "usage_type": "call"}, {"api_name": "app.models.Post", "line_number": 117, "usage_type": "name"}, {"api_name": "flask.g.search_form", "line_number": 117, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 117, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 117, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 117, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 119, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 119, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 118, "usage_type": "call"}, {"api_name": "flask.g.search_form", "line_number": 118, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 118, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 120, "usage_type": "call"}, {"api_name": "flask.g.search_form", "line_number": 120, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 120, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 122, "usage_type": "call"}, {"api_name": "app.main.bp.route", "line_number": 110, "usage_type": "call"}, {"api_name": "app.main.bp", "line_number": 110, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 111, "usage_type": "name"}, {"api_name": "app.models.User.query.filter_by", "line_number": 129, "usage_type": "call"}, {"api_name": "app.models.User.query", "line_number": 129, "usage_type": "attribute"}, {"api_name": "app.models.User", "line_number": 129, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 130, "usage_type": "call"}, {"api_name": "app.main.bp.route", "line_number": 126, "usage_type": "call"}, {"api_name": "app.main.bp", "line_number": 126, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 127, "usage_type": "name"}, {"api_name": "app.models.User.query.filter_by", "line_number": 136, "usage_type": "call"}, {"api_name": "app.models.User.query", "line_number": 136, "usage_type": "attribute"}, {"api_name": "app.models.User", "line_number": 136, "usage_type": "name"}, {"api_name": "app.main.forms.MessageForm", "line_number": 137, "usage_type": "call"}, {"api_name": "app.models.Message", "line_number": 140, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 140, "usage_type": "name"}, {"api_name": "app.db.session.add", "line_number": 141, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 141, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 141, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 142, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 142, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 142, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 143, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 144, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 144, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 146, "usage_type": "call"}, {"api_name": "app.main.bp.route", "line_number": 133, "usage_type": "call"}, {"api_name": "app.main.bp", "line_number": 133, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 134, "usage_type": "name"}, {"api_name": "flask_login.current_user.last_message_read_time", "line_number": 152, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 152, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 152, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 152, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 153, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 153, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 153, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 154, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 154, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 154, "usage_type": "name"}, {"api_name": "flask_login.current_user.messages_received.order_by", "line_number": 155, "usage_type": "call"}, {"api_name": "flask_login.current_user.messages_received", "line_number": 155, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 155, "usage_type": "name"}, {"api_name": "app.models.Message.timestamp.desc", "line_number": 156, "usage_type": "call"}, {"api_name": "app.models.Message.timestamp", "line_number": 156, "usage_type": "attribute"}, {"api_name": "app.models.Message", "line_number": 156, "usage_type": "name"}, {"api_name": "flask.current_app.config", "line_number": 157, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 157, "usage_type": "name"}, {"api_name": "flask.url_for", "line_number": 158, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 160, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 162, "usage_type": "call"}, {"api_name": "app.main.bp.route", "line_number": 149, "usage_type": "call"}, {"api_name": "app.main.bp", "line_number": 149, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 150, "usage_type": "name"}]} +{"seq_id": "69533763", "text": "#sort el ninio events\nimport numpy as np\nimport xarray as xr\nimport pandas as pd\nimport plumb_flux\nimport os\n\nos.environ['HDF5_USE_FILE_LOCKING'] = 'FALSE'\nRUTA='/storage/shared/glusterfs/acrcc/users/vg140344/data/'\n#abro el archivo de geopotencial y junto la coordenada year y numbre\nwinds = xr.open_dataset(RUTA + 'winds200.nc')\nwinds.coords['year'] = np.arange(1981, 2017)\nwinds = winds.stack(realiz = ['year', 'number'])\nwinds = winds.transpose('month', 'realiz', 'latitude', 'longitude')\nhgt = xr.open_dataset(RUTA + 'hgt200.nc')\nhgt.coords['year'] = np.arange(1981, 2017)\nhgt = hgt.stack(realiz = ['year', 'number'])\nhgt = hgt.transpose('month', 'realiz', 'latitude', 'longitude')\nhgt.z.values = hgt.z.values / 9.8\nmonth = ['Aug', 'Sep', 'Oct', 'Nov']\nseas = ['ASO', 'SON']\npx = []\npy = []\nfor i in np.arange(0,4):\n\tpfx, pfy, lat = plumb_flux.ComputePlumbFluxes(winds.u.values[i, :, :, :],\n\t\t\t\t\t\t winds.v.values[i,:,:,:], hgt.z.values[i, :, :, :],\n\t\t\t\t\t\t hgt.latitude.values, hgt.longitude.values)\n\tpfx = xr.DataArray(pfx, coords=[np.arange(51 * 36), lat,\n\t\t\t\t\thgt.longitude.values], dims=['realiz', 'latitude', 'logitude'])\n\tpfy = xr.DataArray(pfy, coords=[np.arange(51 * 36), lat,\n\t\t\t\t\thgt.longitude.values], dims=['realiz', 'latitude', 'logitude'])\n\tpx.append(pfx)\n\tpy.append(pfy)\n\npx = xr.concat(px, dim='month')\npx ['month'] = np.array([8, 9, 10, 11])\npy = xr.concat(py, dim='month')\npy ['month'] = np.array([8, 9, 10, 11])\n\npx.to_netcdf(RUTA + 'monthly_plumb_xflux.nc')\npy.to_netcdf(RUTA + 'monthly_plumb_yflux.nc')\n\n\npx = []\npy = []\t\t\nfor i in np.arange(0,2):\n\tif i == 0:\n\t\tvar1 = hgt.sel(**{'month':slice(8, 10)}).mean(dim='month')\n\t\tvar2 = winds.sel(**{'month':slice(8, 10)}).mean(dim='month')\n\n\telse:\n\t\tvar1 = hgt.sel(**{'month':slice(9, 11)}).mean(dim='month')\n\t\tvar2 = winds.sel(**{'month':slice(9, 11)}).mean(dim='month')\n\tpfx, pfy, lat = plumb_flux.ComputePlumbFluxes(var2.u.values, var2.v.values,\n\t\t\t\t\t\t var1.z.values,\n\t\t\t\t\t\t var1.latitude.values, var1.longitude.values)\n\tpfx = xr.DataArray(pfx, coords=[np.arange(51 * 36), lat,\n\t\t\t\t\thgt.longitude.values], dims=['realiz', 'latitude', 'logitude'])\n\tpfy = xr.DataArray(pfy, coords=[np.arange(51 * 36), lat,\n\t\t\t\t\thgt.longitude.values], dims=['realiz', 'latitude', 'logitude'])\n\tpx.append(pfx)\n\tpy.append(pfy)\n\npx = xr.concat(px, dim='season')\npx ['season'] = np.array(['ASO', 'SON'])\npy = xr.concat(py, dim='season')\npy ['season'] = np.array(['ASO', 'SON'])\n\npx.to_netcdf(RUTA + 'seasonal_plumb_xflux.nc')\npy.to_netcdf(RUTA + 'seasonal_plumb_yflux.nc')\n\n\n", "sub_path": "extras/arrange_data/compute_plumb_fluxes.py", "file_name": "compute_plumb_fluxes.py", "file_ext": "py", "file_size_in_byte": 2534, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "os.environ", "line_number": 8, "usage_type": "attribute"}, {"api_name": "xarray.open_dataset", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 12, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 24, "usage_type": "call"}, {"api_name": "plumb_flux.ComputePlumbFluxes", "line_number": 25, "usage_type": "call"}, {"api_name": "xarray.DataArray", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 28, "usage_type": "call"}, {"api_name": "xarray.DataArray", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 30, "usage_type": "call"}, {"api_name": "xarray.concat", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 36, "usage_type": "call"}, {"api_name": "xarray.concat", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 46, "usage_type": "call"}, {"api_name": "plumb_flux.ComputePlumbFluxes", "line_number": 54, "usage_type": "call"}, {"api_name": "xarray.DataArray", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 57, "usage_type": "call"}, {"api_name": "xarray.DataArray", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 59, "usage_type": "call"}, {"api_name": "xarray.concat", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 65, "usage_type": "call"}, {"api_name": "xarray.concat", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 67, "usage_type": "call"}]} +{"seq_id": "251106236", "text": "import json\nimport os\nimport numpy as np\nfrom scipy.stats import ttest_rel\nfrom itertools import combinations\n\n\ndef process_json():\n curr_path = os.path.join(os.getcwd(), \"listwise_ltr/json_files/\")\n\n json_files = [file for file in os.listdir(\n curr_path) if file.endswith(\".json\")]\n # print(json_files)\n ndcg_on_test = dict()\n\n for file in json_files:\n if not \"TEST\" in file:\n continue\n if \"ERR\" in file:\n continue\n print(\"\\n\"+file)\n with open(os.path.join(curr_path, file)) as f:\n scores_dict = json.load(f)\n ndcg_score = scores_dict[\"ndcg\"][0]\n print(f\"nDCG score: {ndcg_score}\")\n ndcg_on_test[file] = ndcg_score\n\n max_ndcg_40_key = max(ndcg_on_test.keys(),\n key=lambda key: ndcg_on_test[key])\n max_value = ndcg_on_test[max_ndcg_40_key]\n average_ndcg_40 = np.mean(list(ndcg_on_test.values()))\n var_ndcg_40 = np.var(list(ndcg_on_test.values()))\n\n return max_ndcg_40_key, max_value, average_ndcg_40, var_ndcg_40\n\n\nif __name__ == \"__main__\":\n best_setting, best_ndcg, average, variance = process_json()\n print(\n f\"\\nbest config: {best_setting}\\nNDCG score on test: {best_ndcg}\\naverage NDCG on test: {average}\\nNDCG variance: {variance}\")\n", "sub_path": "practical3/listwise_ltr/process_json.py", "file_name": "process_json.py", "file_ext": "py", "file_size_in_byte": 1296, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "os.path.join", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 9, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.var", "line_number": 32, "usage_type": "call"}]} +{"seq_id": "12118026", "text": "import numpy as np\nimport networkx as nx\nfrom scipy.special import binom\nfrom skimage import measure\nimport sys\nimport itertools\nimport time\n\n\ndef memoize(f):\n memo = {}\n def helper(x, y):\n try:\n return memo[(x,y)]\n except:\n memo[(x,y)] = f(x, y)\n return memo[(x,y)]\n return helper\n\ndef memoize2(f):\n memo = {}\n def helper(x, y):\n try:\n return memo[(x,y)]\n except:\n memo[(x,y)] = f(x, y)\n return memo[(x,y)]\n return helper\n\n@memoize\ndef binomial(n,k):\n return int(binom(n,k))\n\n\ndef num_regions(p):\n '''\n counts the number of regions in p\n \n Parameters\n ----------\n p : list\n cube coordinates\n \n Returns\n -------\n int\n number of regions found from the cubes\n '''\n A = polycube_to_graph(p, len(p))\n \n G = nx.from_numpy_matrix(A)\n \n return nx.number_connected_components(G)\n\n\ndef is_connected(p):\n '''\n checks if the cubes are connected in one component\n \n Parameters\n ----------\n p : list\n cube coordinates\n \n Returns\n -------\n bool\n True if there is only one component\n '''\n A = polycube_to_graph(p, len(p))\n \n G = nx.from_numpy_matrix(A)\n \n return nx.is_connected(G)\n\n@memoize2\ndef convert_to_base(b, num):\n '''\n converts num in base 10 to b\n this works for base 2 through 10\n \n Parameters\n ----------\n b : int\n base to convert to\n num : int\n number to convert\n \n Returns\n -------\n str\n num in base b\n '''\n\n if num < b:\n return str(num)\n else:\n return convert_to_base(b, num//b) + str(num%b)\n \n\ndef unrank_kSubset(r, k, n):\n '''\n bijection that turns an integer to a k-subset\n of n elements\n \n Parameters\n ----------\n r : int\n rank of the subset\n k : int\n size of the subset \n n : int\n size of the original set\n \n Returns\n -------\n list\n k-subset defined by (r,k,n) parameters\n '''\n T = [0 for _ in range(k)]\n \n for i in range(1,k+1):\n \n c = binomial(n, k+1-i)\n \n while c > r:\n \n n -= 1\n c = binomial(n, k+1-i)\n \n T[i-1] = n\n r -= c\n return T\n\n\ndef checkTree(edgeList, n):\n '''\n checks if a given edge list is a tree\n \n Parameters\n ----------\n edgeList : list\n edgelist of the graph\n n : int\n number of nodes the graph is on\n \n Returns\n -------\n bool\n True if edgeList defines a tree\n '''\n A = np.zeros((n,n))\n \n for edge in edgeList:\n A[edge] = 1\n \n G = nx.from_numpy_matrix(A)\n \n return nx.is_tree(G)\n \n \ndef brute_force_trees(n, E):\n '''\n Brute force search through all (n-1)-subsets in (E)\n \n Parameters\n ----------\n n : int\n Number of nodes (cubes)\n E : list\n edge list of the graph\n \n Returns\n -------\n list\n list containing the edge lists of the spanning trees\n '''\n \n validTrees = []\n \n m = len(E)\n \n numSubsets = binomial(m, n-1)\n \n for i in range(numSubsets):\n \n subset = unrank_kSubset(i, n-1, m)\n \n edgeSubset = []\n \n for j in subset[::-1]:\n \n edgeSubset.append(E[j])\n \n if checkTree(edgeSubset, n):\n \n validTrees.append(edgeSubset)\n \n return validTrees\n\n\ndef check_all_isomorphism(p1, p2):\n \"\"\"\n Performs all 48 symmetries on polycube2 (p2)\n and checks if it's equal to polycube1 (p1)\n \n 48 symmetries come from the octahedral group with reflections\n (O48)\n \n Parameters\n ----------\n p1 : list\n cube coordinates\n p2 : list\n cube coordinates\n \n Returns\n -------\n bool\n True if p1 and p2 are isomorphic under O48 symmetries\n \"\"\"\n \n #use p1 as the comparison\n fixedCoordinates = set(p1)\n \n #get all coordinates\n coordinateSet = p2\n \n #get all x coordinates in a list, all y in a list, all z in a list\n coordinatesSplit = list(zip(*coordinateSet))\n \n #offset to reflect each coordinate by \n reflectPoints = [max(coor) for coor in coordinatesSplit]\n \n coorReflect = list(itertools.product('01', repeat=3))\n \n permutations = list(itertools.permutations('012'))\n \n #check each permutation of the coordinates\n for permute in permutations:\n \n #permutation indices\n i,j,k = permute[0], permute[1], permute[2]\n i = int(i)\n j = int(j)\n k = int(k)\n \n #get reflection args\n a,b,c = reflectPoints[i], reflectPoints[j], reflectPoints[k]\n \n #check each possible reflection of each coordinate\n for reflect in coorReflect:\n \n rX = int(reflect[0])\n rY = int(reflect[1])\n rZ = int(reflect[2])\n \n newCoordSet = set()\n \n for coord in coordinateSet:\n \n #permute the coordinates\n x,y,z = coord[i], coord[j], coord[k]\n \n #reflect the coordinates\n if rX:\n x = a-x\n if rY:\n y = b-y\n if rZ:\n z = c-z\n \n #add new coordinate under the map\n newCoord = (x, y, z)\n newCoordSet.add(newCoord)\n \n #check if mapping equals the comparison set\n if newCoordSet == fixedCoordinates:\n return True\n \n #at this point, none match\n return False\n\n\ndef check_poly_isomorphism(p1, p2):\n \"\"\"\n Performs all 8 symmetries on polycube2 (p2)\n and checks if it's equal to polycube1 (p1)\n \n 8 symmetries come from dihedral group of order 8 (D8)\n (treat the xy-plane as the square under symmetries)\n \n Parameters\n ----------\n p1 : list\n cube coordinates\n p2 : list\n cube coordinates\n \n Returns\n -------\n bool\n True if p1 and p2 are isomorphic under D8 symmetries in xy-plane\n \"\"\"\n \n #use p1 as the comparison\n fixedCoordinates = set(p1)\n \n #get all coordinates\n coordinateSet = p2\n \n #get all x coordinates in a list, all y in a list, all z in a list\n coordinatesSplit = list(zip(*coordinateSet))\n \n #offset to reflect each coordinate by \n reflectPoints = [max(coor) for coor in coordinatesSplit]\n \n #symmetries only on x and y\n #coorReflect = list(itertools.product('01', repeat=2))\n #permutations = list(itertools.permutations('01'))\n coorReflect = [(0,0), (0,1), (1,0), (1,1)]\n permutations = [(0,1), (1,0)]\n \n #check each permutation of the coordinates\n for permute in permutations:\n \n #permutation indices\n i,j = permute[0], permute[1]\n \n #get reflection args\n a,b = reflectPoints[i], reflectPoints[j]\n \n #check each possible reflection of each coordinate\n for reflect in coorReflect:\n \n rX, rY = reflect[0], reflect[1]\n \n newCoordSet = set()\n \n for coord in coordinateSet:\n \n #permute x and y coordinates\n x,y,z = coord[i], coord[j], coord[2]\n \n #reflect the coordinates\n if rX:\n x = a-x\n if rY:\n y = b-y\n \n #add new coordinate under the map\n newCoord = (x, y, z)\n newCoordSet.add(newCoord)\n \n #check if mapping equals the comparison set\n if newCoordSet == fixedCoordinates:\n return True\n \n #at this point, none match\n return False\n\n\ndef get_polycubes_of_size(k):\n '''\n finds all polycube structures of size k that\n exists in a 5x5x5 bounding box\n \n Parameters\n ----------\n k : int\n Number of polycubes\n \n Returns\n -------\n list\n list that contains lists of polycube cube coordinates\n '''\n \n boxDim = min(5, k)\n boxSize = boxDim**3\n \n totalPossibilities = binomial(boxSize, k)\n \n returnUnique = []\n \n for r in range(totalPossibilities):\n \n #get a list of coordinate ranks in base 10\n coordinateRanks = unrank_kSubset(r=r, k=k, n=boxSize)\n \n #convert coordinates ranks to base k (coordinates in bounding box)\n coordinateStrings = [convert_to_base(boxDim, x).zfill(3) for x in coordinateRanks]\n \n #split string into integer coordinates \n coordinates = [tuple([int(x) for x in str]) for str in coordinateStrings]\n \n #we want one connected component\n if not is_connected(coordinates):\n continue\n \n #this handles translational symmetries\n minX = min(coordinates, key=lambda x: x[0])[0]\n minY = min(coordinates, key=lambda x: x[1])[1]\n minZ = min(coordinates, key=lambda x: x[2])[2]\n coordinates = [(x[0]-minX, x[1]-minY, x[2]-minZ) for x in coordinates]\n \n #check against previously added polycubes\n unique = True\n for alreadyAdded in returnUnique:\n if check_poly_isomorphism(alreadyAdded, coordinates):\n #if isomorphic, skip\n unique = False\n break \n \n if unique:\n returnUnique.append(coordinates)\n \n return returnUnique\n\n\ndef matrix_minor(A, r, c):\n '''\n Removes a row and a column from an adjacency matrix\n \n Parameters\n ----------\n A : numpy 2-d array\n adjacency matrix\n r : int\n row index to delete\n c : int\n column index to delete\n \n Returns\n -------\n numpy 2-d array\n adjacency matrix after deletions\n '''\n #delete row\n Ar = np.delete(A, r, 0)\n \n #delete column\n Arc = np.delete(Ar, c, 1)\n \n return Arc\n\n\ndef taxi_distance(v1, v2):\n '''\n taxi cab metric distance between v1 and v2\n \n Parameters\n ----------\n v1 : list\n first point\n v2 : list\n second point\n \n Returns\n -------\n float\n taxi cab distance between v1 and v2 \n '''\n \n #assert len(v1)==len(v2), \"taxi_distance: v1 and v2 are not of the same dimension\"\n \n d = 0\n for x,y in zip(v1, v2):\n if x <= y:\n d += (y-x)\n else:\n d += (x-y)\n \n return d\n \n \ndef polycube_to_graph(p, n):\n '''\n converts a polycube into a graph by connecting\n polycubes (nodes) if they share a face\n \n Parameters\n ----------\n p : list\n list of cube coordinates\n \n Returns\n -------\n numpy 2-d array\n adjacency matrix of the cubes\n '''\n A = np.zeros((n, n))\n \n for i in range(n):\n for j in range(n):\n if taxi_distance(p[i], p[j]) == 1:\n A[i, j] = 1\n \n return A\n \n\ndef get_edge_list(G):\n '''\n Returns an edge list of the given graph\n \n Parameters\n ----------\n G : networkx graph object\n graph to find the edge list of\n \n Returns\n -------\n list\n edge list where edges are 2-tuples of ints\n '''\n edges = []\n \n for line in nx.generate_edgelist(G):\n \n line = line.split()\n \n x = int(line[0])\n y = int(line[1])\n \n edges.append((x, y))\n \n return edges \n \n \ndef main(test):\n \n d1 = [(0,0,0), \n (0,0,1),\n (0,1,1),\n (0,1,2),\n (1,1,1),\n (1,1,2)]\n\n d2 = [(0,0,0), \n (1,0,0),\n (2,0,0),\n (3,0,0),\n (4,0,0),\n (5,0,0)]\n \n d3 = [(0,0,0), \n (0,1,0),\n (0,2,0),\n (0,3,0),\n (0,4,0),\n (0,5,0)]\n\n d4 = [(0,0,0), \n (1,0,0),\n (1,1,0),\n (2,1,0),\n (1,1,1),\n (2,1,1)]\n \n if test == 'all':\n #all checks\n print(check_all_isomorphism(d1, d2))\n print(check_all_isomorphism(d1, d3))\n print(check_all_isomorphism(d1, d4))\n print(check_all_isomorphism(d2, d3))\n print(check_all_isomorphism(d2, d4))\n print(check_all_isomorphism(d3, d4))\n \n elif test == 'poly':\n #polycube checks\n print(check_poly_isomorphism(d1, d2))\n print(check_poly_isomorphism(d1, d3))\n print(check_poly_isomorphism(d1, d4))\n print(check_poly_isomorphism(d2, d3))\n print(check_poly_isomorphism(d2, d4))\n print(check_poly_isomorphism(d3, d4))\n \n elif int(test):\n \n numCubes = int(test)\n \n begin = time.time()\n listOfPolycubes = get_polycubes_of_size(numCubes)\n \n print('Number of Structures Found: %d'%len(listOfPolycubes))\n \n print('Structure : Number of Articulations : Trees')\n print('-------------------------------------')\n for polycube in listOfPolycubes:\n \n #kirchoff's matrix-tree theorem to count \n # #number of spanning trees\n A = polycube_to_graph(polycube, numCubes)\n G = nx.from_numpy_matrix(A)\n #Q = nx.laplacian_matrix(G).todense()\n #Q_star = matrix_minor(Q, 1, 1)\n #numTrees = int(np.linalg.det(Q_star))\n \n edges = get_edge_list(G)\n trees = brute_force_trees(numCubes, edges)\n numTrees = len(trees)\n \n print(polycube, ': ', numTrees, ': ', trees)\n\n print('Time to run: %0.2f seconds'%(time.time() - begin))\n \n\nif __name__ == '__main__':\n \n test = sys.argv[1]\n \n main(test)", "sub_path": "enumeratePolycubes.py", "file_name": "enumeratePolycubes.py", "file_ext": "py", "file_size_in_byte": 14146, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "scipy.special.binom", "line_number": 32, "usage_type": "call"}, {"api_name": "networkx.from_numpy_matrix", "line_number": 51, "usage_type": "call"}, {"api_name": "networkx.number_connected_components", "line_number": 53, "usage_type": "call"}, {"api_name": "networkx.from_numpy_matrix", "line_number": 72, "usage_type": "call"}, {"api_name": "networkx.is_connected", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 152, "usage_type": "call"}, {"api_name": "networkx.from_numpy_matrix", "line_number": 157, "usage_type": "call"}, {"api_name": "networkx.is_tree", "line_number": 159, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 235, "usage_type": "call"}, {"api_name": "itertools.permutations", "line_number": 237, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 440, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 443, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 492, "usage_type": "call"}, {"api_name": "networkx.generate_edgelist", "line_number": 518, "usage_type": "call"}, {"api_name": "time.time", "line_number": 582, "usage_type": "call"}, {"api_name": "networkx.from_numpy_matrix", "line_number": 594, "usage_type": "call"}, {"api_name": "time.time", "line_number": 605, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 610, "usage_type": "attribute"}]} +{"seq_id": "561735955", "text": "import operator\nfrom django.db import models\n\n\nclass Season(models.Model):\n name = models.CharField(max_length=150)\n start_date = models.DateField('Start date')\n end_date = models.DateField('End date')\n season_round = models.IntegerField(max_length=1)\n\n def __unicode__(self):\n return str(self.start_date.year) + ' Round ' + str(self.season_round)\n\n def get_stats(self):\n \"\"\"\n Generates the season stats\n \"\"\"\n player_count = 0\n results_count = 0\n total_games_count = 0.0\n for ladder in self.ladder_set.all():\n player_count += ladder.league_set.count()\n results_count += ladder.result_set.count() / 2\n total_games_count += (ladder.league_set.count() * (ladder.league_set.count() - 1)) / 2\n\n percentage_played = (results_count / total_games_count) * 100\n\n return {\n 'divisions': self.ladder_set.count(),\n 'percentage_played': \"{0:.2f}\".format(percentage_played),\n 'total_games_count': total_games_count,\n 'results_count': results_count,\n 'player_count': player_count\n }\n\n def get_leader_stats(self):\n \"\"\"\n Generates the list of leaders for current season\n \"\"\"\n current_leaders = {}\n\n for ladder in self.ladder_set.all():\n current_leaders[ladder.id] = ladder.get_leader()\n\n return {\n 'current_leaders': current_leaders,\n }\n\n\nclass Player(models.Model):\n first_name = models.CharField(max_length=100)\n last_name = models.CharField(max_length=100)\n home_phone = models.CharField(max_length=100, blank=True)\n mobile_phone = models.CharField(max_length=100, blank=True)\n email = models.CharField(max_length=100, blank=True)\n junior = models.BooleanField(default=False)\n\n def __unicode__(self):\n return self.first_name + ' ' + self.last_name\n\n\nclass Ladder(models.Model):\n season = models.ForeignKey(Season)\n division = models.CharField(max_length=11)\n ladder_type = models.CharField(max_length=100)\n\n\n def __unicode__(self):\n return str(self.season.start_date.year) + ' Round ' + str(self.season.season_round) + ' - Division: ' + str(self.division)\n\n def get_leader(self):\n \"\"\"\n Finds the leader of the ladder\n \"\"\"\n totals = {}\n for result in self.result_set.filter(ladder=self):\n try:\n if result.result == 9:\n totals[result.player] += int(result.result) + 3\n else:\n totals[result.player] += int(result.result) + 1\n except KeyError:\n if result.result == 9:\n totals[result.player] = int(result.result) + 3\n else:\n totals[result.player] = int(result.result) + 1\n\n\n if totals:\n player = max(totals.iteritems(), key=operator.itemgetter(1))[0]\n else:\n return {'player': 'No Results', 'player_id': '../#', 'total': '-', 'division': self.division}\n\n\n return {'player': player.__str__(), 'player_id': player.id, 'total': totals[player], 'division': self.division}\n\n def get_latest_results(self):\n \"\"\"\n Gets latest results for the ladder\n \"\"\"\n results = {}\n for result in self.result_set.filter(ladder=self).order_by('-date_added')[:10]: # [:10] to limit to 5\n\n try:\n opponent = self.result_set.filter(ladder=self, player=result.opponent, opponent=result.player)[0]\n except IndexError:\n continue # this exception happens if result does not have opponent\n player_opponent_index = ''.join(str(e) for e in sorted([result.player.id, opponent.player.id]))\n try:\n if results[player_opponent_index]:\n continue\n except KeyError:\n results[player_opponent_index] = {'player': result.player, 'player_result': result.result,\n 'opponent_result': opponent.result, 'opponent': opponent.player,\n 'date_added': result.date_added}\n\n ordered_results = {}\n i = 0\n for key in sorted(results, key=lambda x: (results[x]['date_added']), reverse=True):\n ordered_results[i] = results[key]\n i += 1\n\n return ordered_results.items()\n\n def get_stats(self):\n \"\"\"\n Generates the stats for current division\n \"\"\"\n total_matches_played = 0.00\n total_matches = self.league_set.count() * (self.league_set.count() - 1) / 2\n total_matches_played += self.result_set.count() / 2\n perc_matches_played = (total_matches_played / total_matches) * 100\n\n return {\n 'total_matches_played': total_matches_played,\n 'total_matches': total_matches,\n 'perc_matches_played': perc_matches_played\n }\n\n\nclass League(models.Model):\n ladder = models.ForeignKey(Ladder)\n player = models.ForeignKey(Player)\n sort_order = models.IntegerField(default=0)\n\n def __unicode__(self):\n return self.player.first_name + ' ' + self.player.last_name\n\n def player_stats(self):\n \"\"\"\n Generates the player stats for player listings\n \"\"\"\n total_points = 0.00\n games = 0.00\n won_count = 0\n for result in self.player.result_player.filter(player=self.player, ladder=self.ladder):\n\n if result.result == 9:\n total_points += (result.result + 2 + 1) # two for winning one for playing\n won_count += 1\n else:\n total_points += (result.result + 1) # one for playing\n\n games += 1\n\n # work out points per match\n if games > 0:\n pointsdivgames = total_points / games\n percplayed = games / (self.ladder.league_set.count() - 1) * 100\n else:\n percplayed = pointsdivgames = 0\n\n return {\n 'total_points': total_points,\n 'games': games,\n 'pointsdivgames': pointsdivgames,\n 'won_count': won_count,\n 'percplayed': percplayed\n }\n\n\nclass Result(models.Model):\n ladder = models.ForeignKey(Ladder)\n player = models.ForeignKey(Player, related_name='result_player')\n opponent = models.ForeignKey(Player, related_name='result_opponent')\n result = models.IntegerField()\n date_added = models.DateField('Date added')\n inaccurate_flag = models.BooleanField()\n\n def __unicode__(self):\n return self.player.first_name + ' ' + self.player.last_name + ' vs ' + self.opponent.first_name + ' ' + self.opponent.last_name + ' score: ' + str(\n self.result)\n", "sub_path": "ladder/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 6797, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "django.db.models.Model", "line_number": 5, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 5, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 6, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 6, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 7, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 8, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 9, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 9, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 50, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 50, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 51, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 51, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 52, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 52, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 53, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 53, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 54, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 54, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 55, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 55, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 56, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 56, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 62, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 62, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 63, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 63, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 64, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 64, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 65, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 65, "usage_type": "name"}, {"api_name": "operator.itemgetter", "line_number": 90, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 141, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 141, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 142, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 142, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 143, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 143, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 144, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 144, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 182, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 182, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 183, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 183, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 184, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 184, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 185, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 185, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 186, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 186, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 187, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 187, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 188, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 188, "usage_type": "name"}]} +{"seq_id": "400468302", "text": "from __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\nfrom __future__ import unicode_literals\n\nimport functools\nimport logging\nimport trollius as asyncio\n\nfrom asyncio import From\nfrom collections import defaultdict\nfrom thrift.Thrift import TMessageType, TApplicationException\nfrom thrift.transport.TTransport import TTransportException\nfrom thrift.transport.THeaderTransport import THeaderTransport\nfrom thrift.protocol.THeaderProtocol import THeaderProtocolFactory\nfrom thrift.util.async_common import (\n FramedProtocol,\n THeaderProtocol,\n TReadOnlyBuffer,\n TWriteOnlyBuffer,\n WrappedTransportFactory,\n)\n\n\n__all__ = [\n 'ThriftClientProtocolFactory',\n]\n\nlogger = logging.getLogger(__name__)\n\n\ndef ThriftClientProtocolFactory(client_class, loop=None, timeouts=None,\n client_type=None):\n return functools.partial(\n ThriftHeaderClientProtocol, client_class, loop, timeouts, client_type\n )\n\n\nclass SenderTransport(object):\n MAX_QUEUE_SIZE = 1024\n\n def __init__(self, trans, loop=None):\n self._queue = asyncio.Queue(maxsize=self.MAX_QUEUE_SIZE)\n self._trans = trans\n self._loop = loop or asyncio.get_event_loop()\n self._consumer = asyncio.Task(self._send(), loop=self._loop)\n self._producers = []\n\n def __del__(self):\n if not self._consumer.done() or not self._consumer.cancelled():\n logger.debug(\n 'SenderTransport did not finish properly'\n ' as the consumer asyncio.Task is still pending.'\n ' Please make sure to call .close() on this object.'\n )\n\n def send_message(self, msg):\n self._producers.append(\n asyncio.Task(self._queue.put(msg), loop=self._loop),\n )\n\n def _clean_producers(self):\n self._producers = [\n p for p in self._producers if not p.done() and not p.cancelled()\n ]\n\n @asyncio.coroutine\n def _send(self):\n while True:\n msg = yield From(self._queue.get())\n self._clean_producers()\n self._trans.write(msg)\n\n def close(self):\n self._consumer.cancel()\n for producer in self._producers:\n if not producer.done() and not producer.cancelled():\n producer.cancel()\n\n\nclass ThriftHeaderClientProtocol(FramedProtocol):\n DEFAULT_TIMEOUT = 60.0\n _exception_serializer = None\n\n def __init__(self, client_class,\n loop=None,\n timeouts=None,\n client_type=None):\n FramedProtocol.__init__(self, loop=loop)\n self._client_class = client_class\n self.client = None\n self.transport = None\n if timeouts is None:\n timeouts = {}\n default_timeout = timeouts.get('') or self.DEFAULT_TIMEOUT\n self.timeouts = defaultdict(lambda: default_timeout)\n self.timeouts.update(timeouts)\n self.pending_tasks = {}\n self.client_type = client_type\n\n def connection_made(self, transport):\n assert self.transport is None, \"Transport already instantiated here.\"\n assert self.client is None, \"Client already instantiated here.\"\n # asyncio.Transport\n self.transport = transport\n # Thrift transport\n self.thrift_transport = SenderTransport(self.transport, self.loop)\n self.client = self._client_class(\n self.thrift_transport,\n WrappedTransportFactory(self),\n THeaderProtocolFactory(client_type=self.client_type))\n\n def connection_lost(self, exc):\n for fut in self.client._futures.values():\n te = TTransportException(\n type=TTransportException.END_OF_FILE,\n message=\"Connection closed\")\n if not fut.done():\n fut.set_exception(te)\n\n def update_pending_tasks(self, seqid, task):\n no_longer_pending = [\n _seqid for _seqid, _task in self.pending_tasks.items()\n if _task.done() or _task.cancelled()\n ]\n for _seqid in no_longer_pending:\n del self.pending_tasks[_seqid]\n assert seqid not in self.pending_tasks, (\n \"seqid already pending for timeout\"\n )\n self.pending_tasks[seqid] = task\n\n def schedule_timeout(self, fname, seqid):\n timeout = self.timeouts[fname]\n if not timeout:\n return\n\n exc = TApplicationException(\n TApplicationException.TIMEOUT, \"Call to {} timed out\".format(fname)\n )\n serialized_exc = self.serialize_texception(fname, seqid, exc)\n timeout_task = asyncio.Task(\n self.message_received(serialized_exc, delay=timeout),\n loop=self.loop,\n )\n self.update_pending_tasks(seqid, timeout_task)\n\n @asyncio.coroutine\n def message_received(self, frame, delay=0):\n tmi = TReadOnlyBuffer(frame)\n iprot = THeaderProtocol(tmi)\n (fname, mtype, rseqid) = iprot.readMessageBegin()\n\n if delay:\n yield From(asyncio.sleep(delay))\n else:\n try:\n timeout_task = self.pending_tasks.pop(rseqid)\n except KeyError:\n # Task doesn't have a timeout or has already been cancelled\n # and pruned from `pending_tasks`.\n pass\n else:\n timeout_task.cancel()\n\n method = getattr(self.client, \"recv_\" + fname.decode(), None)\n if method is None:\n logger.error(\"Method %r is not supported\", fname)\n self.transport.abort()\n else:\n try:\n method(iprot, mtype, rseqid)\n except (\n asyncio.futures.InvalidStateError,\n asyncio.CancelledError,\n ) as e:\n logger.warning(\"Method %r cancelled: %s\", fname, str(e))\n\n def close(self):\n for task in self.pending_tasks.values():\n if not task.done() and not task.cancelled():\n task.cancel()\n self.transport.abort()\n self.thrift_transport.close()\n\n @classmethod\n def serialize_texception(cls, fname, seqid, exception):\n \"\"\"This saves us a bit of processing time for timeout handling by\n reusing the Thrift structs involved in exception serialization.\n\n NOTE: this is not thread-safe nor is it meant to be.\n \"\"\"\n # the serializer is a singleton\n if cls._exception_serializer is None:\n buffer = TWriteOnlyBuffer()\n transport = THeaderTransport(buffer)\n cls._exception_serializer = THeaderProtocol(transport)\n else:\n transport = cls._exception_serializer.trans\n buffer = transport.getTransport()\n buffer.reset()\n\n serializer = cls._exception_serializer\n serializer.writeMessageBegin(fname, TMessageType.EXCEPTION, seqid)\n exception.write(serializer)\n serializer.writeMessageEnd()\n serializer.trans.flush()\n return buffer.getvalue()\n", "sub_path": "thrift/lib/py/server/TTrolliusServer.py", "file_name": "TTrolliusServer.py", "file_ext": "py", "file_size_in_byte": 7052, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "logging.getLogger", "line_number": 29, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 34, "usage_type": "call"}, {"api_name": "trollius.Queue", "line_number": 43, "usage_type": "call"}, {"api_name": "trollius.get_event_loop", "line_number": 45, "usage_type": "call"}, {"api_name": "trollius.Task", "line_number": 46, "usage_type": "call"}, {"api_name": "trollius.Task", "line_number": 59, "usage_type": "call"}, {"api_name": "asyncio.From", "line_number": 70, "usage_type": "call"}, {"api_name": "trollius.coroutine", "line_number": 67, "usage_type": "attribute"}, {"api_name": "thrift.util.async_common.FramedProtocol", "line_number": 81, "usage_type": "name"}, {"api_name": "thrift.util.async_common.FramedProtocol.__init__", "line_number": 89, "usage_type": "call"}, {"api_name": "thrift.util.async_common.FramedProtocol", "line_number": 89, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 96, "usage_type": "call"}, {"api_name": "thrift.util.async_common.WrappedTransportFactory", "line_number": 110, "usage_type": "call"}, {"api_name": "thrift.protocol.THeaderProtocol.THeaderProtocolFactory", "line_number": 111, "usage_type": "call"}, {"api_name": "thrift.transport.TTransport.TTransportException", "line_number": 115, "usage_type": "call"}, {"api_name": "thrift.transport.TTransport.TTransportException.END_OF_FILE", "line_number": 116, "usage_type": "attribute"}, {"api_name": "thrift.transport.TTransport.TTransportException", "line_number": 116, "usage_type": "name"}, {"api_name": "thrift.Thrift.TApplicationException", "line_number": 138, "usage_type": "call"}, {"api_name": "thrift.Thrift.TApplicationException.TIMEOUT", "line_number": 139, "usage_type": "attribute"}, {"api_name": "thrift.Thrift.TApplicationException", "line_number": 139, "usage_type": "name"}, {"api_name": "trollius.Task", "line_number": 142, "usage_type": "call"}, {"api_name": "thrift.util.async_common.TReadOnlyBuffer", "line_number": 150, "usage_type": "call"}, {"api_name": "thrift.util.async_common.THeaderProtocol", "line_number": 151, "usage_type": "call"}, {"api_name": "asyncio.From", "line_number": 155, "usage_type": "call"}, {"api_name": "trollius.sleep", "line_number": 155, "usage_type": "call"}, {"api_name": "trollius.futures", "line_number": 174, "usage_type": "attribute"}, {"api_name": "trollius.CancelledError", "line_number": 175, "usage_type": "attribute"}, {"api_name": "trollius.coroutine", "line_number": 148, "usage_type": "attribute"}, {"api_name": "thrift.util.async_common.TWriteOnlyBuffer", "line_number": 195, "usage_type": "call"}, {"api_name": "thrift.transport.THeaderTransport.THeaderTransport", "line_number": 196, "usage_type": "call"}, {"api_name": "thrift.util.async_common.THeaderProtocol", "line_number": 197, "usage_type": "call"}, {"api_name": "thrift.Thrift.TMessageType.EXCEPTION", "line_number": 204, "usage_type": "attribute"}, {"api_name": "thrift.Thrift.TMessageType", "line_number": 204, "usage_type": "name"}]} +{"seq_id": "152914629", "text": "# All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\nimport json\nimport pytest\nimport io\n\nfrom cloudevents.sdk import exceptions\nfrom cloudevents.sdk import marshaller\n\nfrom cloudevents.sdk.event import v01\nfrom cloudevents.sdk.event import v02\n\nfrom cloudevents.sdk.converters import binary\nfrom cloudevents.sdk.converters import structured\n\nfrom cloudevents.tests import data\n\n\ndef test_binary_converter_upstream():\n m = marshaller.NewHTTPMarshaller(\n [binary.NewBinaryHTTPCloudEventConverter()])\n event = m.FromRequest(v02.Event(), data.headers, None, lambda x: x)\n assert event is not None\n assert event.Get(\"type\") == (data.ce_type, True)\n assert event.Get(\"id\") == (data.ce_id, True)\n\n\ndef test_structured_converter_upstream():\n m = marshaller.NewHTTPMarshaller(\n [structured.NewJSONHTTPCloudEventConverter()])\n event = m.FromRequest(\n v02.Event(),\n {\"Content-Type\": \"application/cloudevents+json\"},\n io.StringIO(json.dumps(data.ce)),\n lambda x: x.read(),\n )\n\n assert event is not None\n assert event.Get(\"type\") == (data.ce_type, True)\n assert event.Get(\"id\") == (data.ce_id, True)\n\n\ndef test_binary_converter_v01():\n m = marshaller.NewHTTPMarshaller(\n [binary.NewBinaryHTTPCloudEventConverter()])\n\n pytest.raises(\n exceptions.UnsupportedEventConverter,\n m.FromRequest,\n v01.Event,\n {},\n None,\n lambda x: x,\n )\n\n\ndef test_unsupported_converter_v01():\n m = marshaller.NewHTTPMarshaller(\n [structured.NewJSONHTTPCloudEventConverter()])\n\n pytest.raises(\n exceptions.UnsupportedEventConverter,\n m.FromRequest,\n v01.Event,\n {},\n None,\n lambda x: x,\n )\n\n\ndef test_structured_converter_v01():\n m = marshaller.NewHTTPMarshaller(\n [structured.NewJSONHTTPCloudEventConverter()])\n event = m.FromRequest(\n v01.Event(),\n {\"Content-Type\": \"application/cloudevents+json\"},\n io.StringIO(json.dumps(data.ce)),\n lambda x: x.read(),\n )\n\n assert event is not None\n assert event.Get(\"type\") == (data.ce_type, True)\n assert event.Get(\"id\") == (data.ce_id, True)\n\n\ndef test_default_http_marshaller_with_structured():\n m = marshaller.NewDefaultHTTPMarshaller()\n\n event = m.FromRequest(\n v02.Event(),\n {\"Content-Type\": \"application/cloudevents+json\"},\n io.StringIO(json.dumps(data.ce)),\n lambda x: x.read(),\n )\n assert event is not None\n assert event.Get(\"type\") == (data.ce_type, True)\n assert event.Get(\"id\") == (data.ce_id, True)\n\n\ndef test_default_http_marshaller_with_binary():\n m = marshaller.NewDefaultHTTPMarshaller()\n\n event = m.FromRequest(\n v02.Event(), data.headers,\n io.StringIO(json.dumps(data.body)),\n json.load\n )\n assert event is not None\n assert event.Get(\"type\") == (data.ce_type, True)\n assert event.Get(\"data\") == (data.body, True)\n assert event.Get(\"id\") == (data.ce_id, True)\n\n\ndef test_unsupported_event_configuration():\n m = marshaller.NewHTTPMarshaller(\n [binary.NewBinaryHTTPCloudEventConverter()])\n pytest.raises(\n exceptions.UnsupportedEventConverter,\n m.FromRequest,\n v01.Event(),\n {\"Content-Type\": \"application/cloudevents+json\"},\n io.StringIO(json.dumps(data.ce)),\n lambda x: x.read(),\n )\n\n\ndef test_invalid_data_unmarshaller():\n m = marshaller.NewDefaultHTTPMarshaller()\n pytest.raises(\n exceptions.InvalidDataUnmarshaller,\n m.FromRequest,\n v01.Event(), {}, None, None\n )\n\n\ndef test_invalid_data_marshaller():\n m = marshaller.NewDefaultHTTPMarshaller()\n pytest.raises(\n exceptions.InvalidDataMarshaller, m.ToRequest, v01.Event(), \"blah\", None\n )\n", "sub_path": "cloudevents/tests/test_event_from_request_converter.py", "file_name": "test_event_from_request_converter.py", "file_ext": "py", "file_size_in_byte": 4341, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "cloudevents.sdk.marshaller.NewHTTPMarshaller", "line_number": 32, "usage_type": "call"}, {"api_name": "cloudevents.sdk.marshaller", "line_number": 32, "usage_type": "name"}, {"api_name": "cloudevents.sdk.converters.binary.NewBinaryHTTPCloudEventConverter", "line_number": 33, "usage_type": "call"}, {"api_name": "cloudevents.sdk.converters.binary", "line_number": 33, "usage_type": "name"}, {"api_name": "cloudevents.sdk.event.v02.Event", "line_number": 34, "usage_type": "call"}, {"api_name": "cloudevents.sdk.event.v02", "line_number": 34, "usage_type": "name"}, {"api_name": "cloudevents.tests.data.headers", "line_number": 34, "usage_type": "attribute"}, {"api_name": "cloudevents.tests.data", "line_number": 34, "usage_type": "name"}, {"api_name": "cloudevents.tests.data.ce_type", "line_number": 36, "usage_type": "attribute"}, {"api_name": "cloudevents.tests.data", "line_number": 36, "usage_type": "name"}, {"api_name": "cloudevents.tests.data.ce_id", "line_number": 37, "usage_type": "attribute"}, {"api_name": "cloudevents.tests.data", "line_number": 37, "usage_type": "name"}, {"api_name": "cloudevents.sdk.marshaller.NewHTTPMarshaller", "line_number": 41, "usage_type": "call"}, {"api_name": "cloudevents.sdk.marshaller", "line_number": 41, "usage_type": "name"}, {"api_name": "cloudevents.sdk.converters.structured.NewJSONHTTPCloudEventConverter", "line_number": 42, "usage_type": "call"}, {"api_name": "cloudevents.sdk.converters.structured", "line_number": 42, "usage_type": "name"}, {"api_name": "cloudevents.sdk.event.v02.Event", "line_number": 44, "usage_type": "call"}, {"api_name": "cloudevents.sdk.event.v02", "line_number": 44, "usage_type": "name"}, {"api_name": "io.StringIO", "line_number": 46, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 46, "usage_type": "call"}, {"api_name": "cloudevents.tests.data.ce", "line_number": 46, "usage_type": "attribute"}, {"api_name": "cloudevents.tests.data", "line_number": 46, "usage_type": "name"}, {"api_name": "cloudevents.tests.data.ce_type", "line_number": 51, "usage_type": "attribute"}, {"api_name": "cloudevents.tests.data", "line_number": 51, "usage_type": "name"}, {"api_name": "cloudevents.tests.data.ce_id", "line_number": 52, "usage_type": "attribute"}, {"api_name": "cloudevents.tests.data", "line_number": 52, "usage_type": "name"}, {"api_name": "cloudevents.sdk.marshaller.NewHTTPMarshaller", "line_number": 56, "usage_type": "call"}, {"api_name": "cloudevents.sdk.marshaller", "line_number": 56, "usage_type": "name"}, {"api_name": "cloudevents.sdk.converters.binary.NewBinaryHTTPCloudEventConverter", "line_number": 57, "usage_type": "call"}, {"api_name": "cloudevents.sdk.converters.binary", "line_number": 57, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 59, "usage_type": "call"}, {"api_name": "cloudevents.sdk.exceptions.UnsupportedEventConverter", "line_number": 60, "usage_type": "attribute"}, {"api_name": "cloudevents.sdk.exceptions", "line_number": 60, "usage_type": "name"}, {"api_name": "cloudevents.sdk.event.v01.Event", "line_number": 62, "usage_type": "attribute"}, {"api_name": "cloudevents.sdk.event.v01", "line_number": 62, "usage_type": "name"}, {"api_name": "cloudevents.sdk.marshaller.NewHTTPMarshaller", "line_number": 70, "usage_type": "call"}, {"api_name": "cloudevents.sdk.marshaller", "line_number": 70, "usage_type": "name"}, {"api_name": "cloudevents.sdk.converters.structured.NewJSONHTTPCloudEventConverter", "line_number": 71, "usage_type": "call"}, {"api_name": "cloudevents.sdk.converters.structured", "line_number": 71, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 73, "usage_type": "call"}, {"api_name": "cloudevents.sdk.exceptions.UnsupportedEventConverter", "line_number": 74, "usage_type": "attribute"}, {"api_name": "cloudevents.sdk.exceptions", "line_number": 74, "usage_type": "name"}, {"api_name": "cloudevents.sdk.event.v01.Event", "line_number": 76, "usage_type": "attribute"}, {"api_name": "cloudevents.sdk.event.v01", "line_number": 76, "usage_type": "name"}, {"api_name": "cloudevents.sdk.marshaller.NewHTTPMarshaller", "line_number": 84, "usage_type": "call"}, {"api_name": "cloudevents.sdk.marshaller", "line_number": 84, "usage_type": "name"}, {"api_name": "cloudevents.sdk.converters.structured.NewJSONHTTPCloudEventConverter", "line_number": 85, "usage_type": "call"}, {"api_name": "cloudevents.sdk.converters.structured", "line_number": 85, "usage_type": "name"}, {"api_name": "cloudevents.sdk.event.v01.Event", "line_number": 87, "usage_type": "call"}, {"api_name": "cloudevents.sdk.event.v01", "line_number": 87, "usage_type": "name"}, {"api_name": "io.StringIO", "line_number": 89, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 89, "usage_type": "call"}, {"api_name": "cloudevents.tests.data.ce", "line_number": 89, "usage_type": "attribute"}, {"api_name": "cloudevents.tests.data", "line_number": 89, "usage_type": "name"}, {"api_name": "cloudevents.tests.data.ce_type", "line_number": 94, "usage_type": "attribute"}, {"api_name": "cloudevents.tests.data", "line_number": 94, "usage_type": "name"}, {"api_name": "cloudevents.tests.data.ce_id", "line_number": 95, "usage_type": "attribute"}, {"api_name": "cloudevents.tests.data", "line_number": 95, "usage_type": "name"}, {"api_name": "cloudevents.sdk.marshaller.NewDefaultHTTPMarshaller", "line_number": 99, "usage_type": "call"}, {"api_name": "cloudevents.sdk.marshaller", "line_number": 99, "usage_type": "name"}, {"api_name": "cloudevents.sdk.event.v02.Event", "line_number": 102, "usage_type": "call"}, {"api_name": "cloudevents.sdk.event.v02", "line_number": 102, "usage_type": "name"}, {"api_name": "io.StringIO", "line_number": 104, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 104, "usage_type": "call"}, {"api_name": "cloudevents.tests.data.ce", "line_number": 104, "usage_type": "attribute"}, {"api_name": "cloudevents.tests.data", "line_number": 104, "usage_type": "name"}, {"api_name": "cloudevents.tests.data.ce_type", "line_number": 108, "usage_type": "attribute"}, {"api_name": "cloudevents.tests.data", "line_number": 108, "usage_type": "name"}, {"api_name": "cloudevents.tests.data.ce_id", "line_number": 109, "usage_type": "attribute"}, {"api_name": "cloudevents.tests.data", "line_number": 109, "usage_type": "name"}, {"api_name": "cloudevents.sdk.marshaller.NewDefaultHTTPMarshaller", "line_number": 113, "usage_type": "call"}, {"api_name": "cloudevents.sdk.marshaller", "line_number": 113, "usage_type": "name"}, {"api_name": "cloudevents.sdk.event.v02.Event", "line_number": 116, "usage_type": "call"}, {"api_name": "cloudevents.sdk.event.v02", "line_number": 116, "usage_type": "name"}, {"api_name": "cloudevents.tests.data.headers", "line_number": 116, "usage_type": "attribute"}, {"api_name": "cloudevents.tests.data", "line_number": 116, "usage_type": "name"}, {"api_name": "io.StringIO", "line_number": 117, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 117, "usage_type": "call"}, {"api_name": "cloudevents.tests.data.body", "line_number": 117, "usage_type": "attribute"}, {"api_name": "cloudevents.tests.data", "line_number": 117, "usage_type": "name"}, {"api_name": "json.load", "line_number": 118, "usage_type": "attribute"}, {"api_name": "cloudevents.tests.data.ce_type", "line_number": 121, "usage_type": "attribute"}, {"api_name": "cloudevents.tests.data", "line_number": 121, "usage_type": "name"}, {"api_name": "cloudevents.tests.data.body", "line_number": 122, "usage_type": "attribute"}, {"api_name": "cloudevents.tests.data", "line_number": 122, "usage_type": "name"}, {"api_name": "cloudevents.tests.data.ce_id", "line_number": 123, "usage_type": "attribute"}, {"api_name": "cloudevents.tests.data", "line_number": 123, "usage_type": "name"}, {"api_name": "cloudevents.sdk.marshaller.NewHTTPMarshaller", "line_number": 127, "usage_type": "call"}, {"api_name": "cloudevents.sdk.marshaller", "line_number": 127, "usage_type": "name"}, {"api_name": "cloudevents.sdk.converters.binary.NewBinaryHTTPCloudEventConverter", "line_number": 128, "usage_type": "call"}, {"api_name": "cloudevents.sdk.converters.binary", "line_number": 128, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 129, "usage_type": "call"}, {"api_name": "cloudevents.sdk.exceptions.UnsupportedEventConverter", "line_number": 130, "usage_type": "attribute"}, {"api_name": "cloudevents.sdk.exceptions", "line_number": 130, "usage_type": "name"}, {"api_name": "cloudevents.sdk.event.v01.Event", "line_number": 132, "usage_type": "call"}, {"api_name": "cloudevents.sdk.event.v01", "line_number": 132, "usage_type": "name"}, {"api_name": "io.StringIO", "line_number": 134, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 134, "usage_type": "call"}, {"api_name": "cloudevents.tests.data.ce", "line_number": 134, "usage_type": "attribute"}, {"api_name": "cloudevents.tests.data", "line_number": 134, "usage_type": "name"}, {"api_name": "cloudevents.sdk.marshaller.NewDefaultHTTPMarshaller", "line_number": 140, "usage_type": "call"}, {"api_name": "cloudevents.sdk.marshaller", "line_number": 140, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 141, "usage_type": "call"}, {"api_name": "cloudevents.sdk.exceptions.InvalidDataUnmarshaller", "line_number": 142, "usage_type": "attribute"}, {"api_name": "cloudevents.sdk.exceptions", "line_number": 142, "usage_type": "name"}, {"api_name": "cloudevents.sdk.event.v01.Event", "line_number": 144, "usage_type": "call"}, {"api_name": "cloudevents.sdk.event.v01", "line_number": 144, "usage_type": "name"}, {"api_name": "cloudevents.sdk.marshaller.NewDefaultHTTPMarshaller", "line_number": 149, "usage_type": "call"}, {"api_name": "cloudevents.sdk.marshaller", "line_number": 149, "usage_type": "name"}, {"api_name": "pytest.raises", "line_number": 150, "usage_type": "call"}, {"api_name": "cloudevents.sdk.exceptions.InvalidDataMarshaller", "line_number": 151, "usage_type": "attribute"}, {"api_name": "cloudevents.sdk.exceptions", "line_number": 151, "usage_type": "name"}, {"api_name": "cloudevents.sdk.event.v01.Event", "line_number": 151, "usage_type": "call"}, {"api_name": "cloudevents.sdk.event.v01", "line_number": 151, "usage_type": "name"}]} +{"seq_id": "155930599", "text": "from django.views.generic import TemplateView, FormView, UpdateView\nfrom .forms import NarrateForm, TranslateForm\nfrom django.contrib.messages.views import SuccessMessageMixin\nfrom django.urls import reverse_lazy\nfrom django.contrib.auth.mixins import LoginRequiredMixin, UserPassesTestMixin\nfrom django.shortcuts import render\nfrom .models import (\n AMLVideo,\n HeaderText,\n Volunteers,\n Footer,\n Process,\n Team,\n AboutText,\n About,\n Narrate,\n Translate,\n VolunteerText,\n)\nfrom .forms import AMLVideoFilterForm\n\n\ndef home(request):\n filter_form = AMLVideoFilterForm(request.GET)\n videos = AMLVideo.objects.all()\n header = HeaderText.objects.all()\n volunteers = Volunteers.objects.all()\n footer = Footer.objects.all()\n process = Process.objects.all()\n\n # Get category from filter\n category = filter_form.data.get('category')\n if category:\n videos = videos.filter(\n category__exact=category\n )\n\n # Get language from filter\n language = filter_form.data.get('language')\n if language:\n videos = videos.filter(\n language__exact=language\n )\n\n # Get level from filter\n level = filter_form.data.get('level')\n if level:\n videos = videos.filter(\n level__exact=level\n )\n\n videos = videos.order_by(\"-category\", \"-language\", \"-level\")\n\n context = {\n 'videos': videos,\n 'filter_form': filter_form,\n 'level': level,\n 'language': language,\n 'category': category,\n 'volunteers': volunteers,\n 'footer': footer,\n 'process': process,\n 'header': header,\n }\n return render(request, 'home.html', context)\n\n\ndef about(request):\n team = Team.objects.all()\n footer = Footer.objects.all()\n text = AboutText.objects.all()\n about = About.objects.all()\n header = HeaderText.objects.all()\n context = {\n 'team': team,\n 'footer': footer,\n 'text': text,\n 'about': about,\n 'header': header,\n }\n return render(request, 'about.html', context)\n\n\ndef volunteer(request):\n volunteer = Volunteers.objects.all()\n footer = Footer.objects.all()\n text = VolunteerText.objects.all()\n context = {\n 'volunteer': volunteer,\n 'footer': footer,\n 'text': text,\n }\n return render(request, 'volunteers.html', context)\n\n\nclass ProfilePageView(TemplateView):\n template_name = 'profile.html'\n\n def get(self, request, *args, **kwargs):\n add_narration = NarrateForm(self.request.GET or None)\n add_translation = TranslateForm(self.request.GET or None)\n footer = Footer.objects.all()\n context = self.get_context_data(**kwargs)\n context['add_narration'] = add_narration\n context['add_translation'] = add_translation\n context['footer'] = footer\n return self.render_to_response(context)\n\n\nclass NarrateFormView(FormView):\n form_class = NarrateForm\n template_name = 'profile.html'\n footer = Footer.objects.all()\n\n def post(self, request, *args, **kwargs):\n add_narration = self.form_class(request.POST)\n add_translation = TranslateForm()\n if add_narration.is_valid():\n add_narration.save()\n return self.render_to_response(\n self.get_context_data(\n success=True\n )\n )\n else:\n return self.render_to_response(\n self.get_context_data(\n add_narration=add_narration,\n )\n )\n\n\nclass NarrateUpdate(SuccessMessageMixin, UpdateView):\n model = Narrate\n fields = ['title', 'body']\n template_name = 'narrate_update_form.html'\n success_url = reverse_lazy('narrate-status')\n success_message = \"Narration successfully updated\"\n\n def get(self, request, *args, **kwargs):\n footer = Footer.objects.all()\n context = self.get_context_data(**kwargs)\n context['footer'] = footer\n return self.render_to_response(context)\n\n\ndef submittedView(request):\n footer = Footer.objects.all()\n return render(request, 'submission.html', {'footer': footer, })\n\n\nclass TranslateFormView(FormView):\n form_class = TranslateForm\n template_name = 'profile.html'\n\n def post(self, request, *args, **kwargs):\n add_translation = self.form_class(request.POST)\n add_narration = NarrateForm()\n if add_translation.is_valid():\n add_translation.save()\n return self.render_to_response(\n self.get_context_data(\n success=True\n )\n )\n else:\n return self.render_to_response(\n self.get_context_data(\n add_translation=add_translation,\n )\n )\n\n\nclass TranslateUpdate(SuccessMessageMixin, UpdateView):\n model = Translate\n fields = ['title', 'body']\n template_name = 'translate_update_form.html'\n success_url = reverse_lazy('translate-status')\n success_message = \"Translation successfully updated\"\n\n def get(self, request, *args, **kwargs):\n footer = Footer.objects.all()\n context = self.get_context_data(**kwargs)\n context['footer'] = footer\n return self.render_to_response(context)\n\n\ndef submittedView(request):\n return render(request, 'submission.html')\n\n\ndef TranslationStatus(request):\n translation = Translate.objects.all()\n footer = Footer.objects.all()\n context = {'translation': translation, 'footer': footer, }\n return render(request, 'translate_status.html', context)\n\n\ndef NarrationStatus(request):\n narration = Narrate.objects.all()\n footer = Footer.objects.all()\n context = {'narration': narration, 'footer': footer, }\n return render(request, 'narrate_status.html', context)\n", "sub_path": "pages/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 5854, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "forms.AMLVideoFilterForm", "line_number": 24, "usage_type": "call"}, {"api_name": "models.AMLVideo.objects.all", "line_number": 25, "usage_type": "call"}, {"api_name": "models.AMLVideo.objects", "line_number": 25, "usage_type": "attribute"}, {"api_name": "models.AMLVideo", "line_number": 25, "usage_type": "name"}, {"api_name": "models.HeaderText.objects.all", "line_number": 26, "usage_type": "call"}, {"api_name": "models.HeaderText.objects", "line_number": 26, "usage_type": "attribute"}, {"api_name": "models.HeaderText", "line_number": 26, "usage_type": "name"}, {"api_name": "models.Volunteers.objects.all", "line_number": 27, "usage_type": "call"}, {"api_name": "models.Volunteers.objects", "line_number": 27, "usage_type": "attribute"}, {"api_name": "models.Volunteers", "line_number": 27, "usage_type": "name"}, {"api_name": "models.Footer.objects.all", "line_number": 28, "usage_type": "call"}, {"api_name": "models.Footer.objects", "line_number": 28, "usage_type": "attribute"}, {"api_name": "models.Footer", "line_number": 28, "usage_type": "name"}, {"api_name": "models.Process.objects.all", "line_number": 29, "usage_type": "call"}, {"api_name": "models.Process.objects", "line_number": 29, "usage_type": "attribute"}, {"api_name": "models.Process", "line_number": 29, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 65, "usage_type": "call"}, {"api_name": "models.Team.objects.all", "line_number": 69, "usage_type": "call"}, {"api_name": "models.Team.objects", "line_number": 69, "usage_type": "attribute"}, {"api_name": "models.Team", "line_number": 69, "usage_type": "name"}, {"api_name": "models.Footer.objects.all", "line_number": 70, "usage_type": "call"}, {"api_name": "models.Footer.objects", "line_number": 70, "usage_type": "attribute"}, {"api_name": "models.Footer", "line_number": 70, "usage_type": "name"}, {"api_name": "models.AboutText.objects.all", "line_number": 71, "usage_type": "call"}, {"api_name": "models.AboutText.objects", "line_number": 71, "usage_type": "attribute"}, {"api_name": "models.AboutText", "line_number": 71, "usage_type": "name"}, {"api_name": "models.About.objects.all", "line_number": 72, "usage_type": "call"}, {"api_name": "models.About.objects", "line_number": 72, "usage_type": "attribute"}, {"api_name": "models.About", "line_number": 72, "usage_type": "name"}, {"api_name": "models.HeaderText.objects.all", "line_number": 73, "usage_type": "call"}, {"api_name": "models.HeaderText.objects", "line_number": 73, "usage_type": "attribute"}, {"api_name": "models.HeaderText", "line_number": 73, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 81, "usage_type": "call"}, {"api_name": "models.Volunteers.objects.all", "line_number": 85, "usage_type": "call"}, {"api_name": "models.Volunteers.objects", "line_number": 85, "usage_type": "attribute"}, {"api_name": "models.Volunteers", "line_number": 85, "usage_type": "name"}, {"api_name": "models.Footer.objects.all", "line_number": 86, "usage_type": "call"}, {"api_name": "models.Footer.objects", "line_number": 86, "usage_type": "attribute"}, {"api_name": "models.Footer", "line_number": 86, "usage_type": "name"}, {"api_name": "models.VolunteerText.objects.all", "line_number": 87, "usage_type": "call"}, {"api_name": "models.VolunteerText.objects", "line_number": 87, "usage_type": "attribute"}, {"api_name": "models.VolunteerText", "line_number": 87, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 93, "usage_type": "call"}, {"api_name": "django.views.generic.TemplateView", "line_number": 96, "usage_type": "name"}, {"api_name": "forms.NarrateForm", "line_number": 100, "usage_type": "call"}, {"api_name": "forms.TranslateForm", "line_number": 101, "usage_type": "call"}, {"api_name": "models.Footer.objects.all", "line_number": 102, "usage_type": "call"}, {"api_name": "models.Footer.objects", "line_number": 102, "usage_type": "attribute"}, {"api_name": "models.Footer", "line_number": 102, "usage_type": "name"}, {"api_name": "django.views.generic.FormView", "line_number": 110, "usage_type": "name"}, {"api_name": "forms.NarrateForm", "line_number": 111, "usage_type": "name"}, {"api_name": "models.Footer.objects.all", "line_number": 113, "usage_type": "call"}, {"api_name": "models.Footer.objects", "line_number": 113, "usage_type": "attribute"}, {"api_name": "models.Footer", "line_number": 113, "usage_type": "name"}, {"api_name": "forms.TranslateForm", "line_number": 117, "usage_type": "call"}, {"api_name": "django.contrib.messages.views.SuccessMessageMixin", "line_number": 133, "usage_type": "name"}, {"api_name": "django.views.generic.UpdateView", "line_number": 133, "usage_type": "name"}, {"api_name": "models.Narrate", "line_number": 134, "usage_type": "name"}, {"api_name": "django.urls.reverse_lazy", "line_number": 137, "usage_type": "call"}, {"api_name": "models.Footer.objects.all", "line_number": 141, "usage_type": "call"}, {"api_name": "models.Footer.objects", "line_number": 141, "usage_type": "attribute"}, {"api_name": "models.Footer", "line_number": 141, "usage_type": "name"}, {"api_name": "models.Footer.objects.all", "line_number": 148, "usage_type": "call"}, {"api_name": "models.Footer.objects", "line_number": 148, "usage_type": "attribute"}, {"api_name": "models.Footer", "line_number": 148, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 149, "usage_type": "call"}, {"api_name": "django.views.generic.FormView", "line_number": 152, "usage_type": "name"}, {"api_name": "forms.TranslateForm", "line_number": 153, "usage_type": "name"}, {"api_name": "forms.NarrateForm", "line_number": 158, "usage_type": "call"}, {"api_name": "django.contrib.messages.views.SuccessMessageMixin", "line_number": 174, "usage_type": "name"}, {"api_name": "django.views.generic.UpdateView", "line_number": 174, "usage_type": "name"}, {"api_name": "models.Translate", "line_number": 175, "usage_type": "name"}, {"api_name": "django.urls.reverse_lazy", "line_number": 178, "usage_type": "call"}, {"api_name": "models.Footer.objects.all", "line_number": 182, "usage_type": "call"}, {"api_name": "models.Footer.objects", "line_number": 182, "usage_type": "attribute"}, {"api_name": "models.Footer", "line_number": 182, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 189, "usage_type": "call"}, {"api_name": "models.Translate.objects.all", "line_number": 193, "usage_type": "call"}, {"api_name": "models.Translate.objects", "line_number": 193, "usage_type": "attribute"}, {"api_name": "models.Translate", "line_number": 193, "usage_type": "name"}, {"api_name": "models.Footer.objects.all", "line_number": 194, "usage_type": "call"}, {"api_name": "models.Footer.objects", "line_number": 194, "usage_type": "attribute"}, {"api_name": "models.Footer", "line_number": 194, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 196, "usage_type": "call"}, {"api_name": "models.Narrate.objects.all", "line_number": 200, "usage_type": "call"}, {"api_name": "models.Narrate.objects", "line_number": 200, "usage_type": "attribute"}, {"api_name": "models.Narrate", "line_number": 200, "usage_type": "name"}, {"api_name": "models.Footer.objects.all", "line_number": 201, "usage_type": "call"}, {"api_name": "models.Footer.objects", "line_number": 201, "usage_type": "attribute"}, {"api_name": "models.Footer", "line_number": 201, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 203, "usage_type": "call"}]} +{"seq_id": "563911785", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nimport uuid\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ]\n\n operations = [\n migrations.CreateModel(\n name='Author',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('full_name', models.CharField(max_length=256, unique=True, null=True)),\n ('brief', models.TextField(max_length=2000, null=True, blank=True)),\n ('discoverable', models.BooleanField(default=True, help_text=b'If checked off, all books associated with this author will be discoverable to clients')),\n ],\n options={\n 'verbose_name': 'Author',\n 'verbose_name_plural': 'Authors',\n },\n ),\n migrations.CreateModel(\n name='Book',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('identification', models.UUIDField(default=uuid.uuid4, null=True, verbose_name=b'Book ID')),\n ('title', models.CharField(max_length=256, null=True)),\n ('subtitle', models.CharField(max_length=256, null=True, blank=True)),\n ('pdf_file', models.FileField(upload_to=b'/Users/fadihannaal-kass/Desktop/ebookify/database/books', null=True, verbose_name=b'PDF File', blank=True)),\n ('epub_file', models.FileField(upload_to=b'/Users/fadihannaal-kass/Desktop/ebookify/database/books', null=True, verbose_name=b'EPUB File', blank=True)),\n ('mobi_file', models.FileField(upload_to=b'/Users/fadihannaal-kass/Desktop/ebookify/database/books', null=True, verbose_name=b'MOBI File', blank=True)),\n ('cover', models.ImageField(null=True, upload_to=b'/Users/fadihannaal-kass/Desktop/ebookify/database/covers', blank=True)),\n ('description', models.TextField(max_length=5000, null=True, blank=True)),\n ('num_pages', models.IntegerField(default=0, help_text=b\"Leave at 0 if you don't want the number of pages to be shown\", verbose_name=b'number of Pages')),\n ('num_views', models.IntegerField(default=0, verbose_name=b'Views')),\n ('num_downloads', models.IntegerField(default=0, verbose_name=b'Downloads')),\n ('discoverable', models.BooleanField(default=True, help_text=b'Make this book discoverable to view and download')),\n ('recommended', models.BooleanField(default=False, help_text=b'Recommend this book to your library visitors')),\n ('authors', models.ManyToManyField(to='Attributes.Author')),\n ],\n options={\n 'verbose_name': 'Book',\n 'verbose_name_plural': 'Books',\n },\n ),\n migrations.CreateModel(\n name='Category',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('name', models.CharField(max_length=256, unique=True, null=True)),\n ('brief', models.TextField(max_length=2000, null=True, blank=True)),\n ('discoverable', models.BooleanField(default=True, help_text=b'If checked off, all books associated with this category will be discoverable to clients')),\n ],\n options={\n 'verbose_name': 'Category',\n 'verbose_name_plural': 'Categories',\n },\n ),\n migrations.CreateModel(\n name='Contributor',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('full_name', models.CharField(max_length=256, unique=True, null=True)),\n ('discoverable', models.BooleanField(default=True, help_text=b'If checked off, all books associated with this contributor will be discoverable to clients')),\n ],\n options={\n 'verbose_name': 'Contributor',\n 'verbose_name_plural': 'Contributors',\n },\n ),\n migrations.CreateModel(\n name='Language',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('name', models.CharField(max_length=256, unique=True, null=True)),\n ('discoverable', models.BooleanField(default=True, help_text=b'If checked off, all books associated with this language will be discoverable to clients')),\n ],\n options={\n 'verbose_name': 'Language',\n 'verbose_name_plural': 'Languages',\n },\n ),\n migrations.AddField(\n model_name='book',\n name='categories',\n field=models.ManyToManyField(to='Attributes.Category'),\n ),\n migrations.AddField(\n model_name='book',\n name='contributors',\n field=models.ManyToManyField(to='Attributes.Contributor', blank=True),\n ),\n migrations.AddField(\n model_name='book',\n name='language',\n field=models.ForeignKey(to='Attributes.Language', null=True),\n ),\n ]\n", "sub_path": "Attributes/migrations/0001_initial.py", "file_name": "0001_initial.py", "file_ext": "py", "file_size_in_byte": 5410, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 27, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 27, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.models.UUIDField", "line_number": 31, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 31, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 31, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 32, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 32, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 33, "usage_type": "name"}, {"api_name": "django.db.models.FileField", "line_number": 34, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 34, "usage_type": "name"}, {"api_name": "django.db.models.FileField", "line_number": 35, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 35, "usage_type": "name"}, {"api_name": "django.db.models.FileField", "line_number": 36, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 36, "usage_type": "name"}, {"api_name": "django.db.models.ImageField", "line_number": 37, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 37, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 38, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 38, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 39, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 39, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 40, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 40, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 41, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 41, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 42, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 42, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 43, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 43, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 44, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 44, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 51, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 51, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 54, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 54, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 55, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 55, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 56, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 56, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 57, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 57, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 64, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 64, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 67, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 67, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 68, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 68, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 69, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 69, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 76, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 76, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 79, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 79, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 80, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 80, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 81, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 81, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 88, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 88, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 91, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 91, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 93, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 93, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 96, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 96, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 98, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 98, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 101, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 101, "usage_type": "name"}]} +{"seq_id": "230349159", "text": "import os\nimport json\nimport torch\nimport numpy as np\nfrom torch.utils.data import Dataset\nimport pandas as pd\nimport argparse\nimport copy\nfrom collections import Counter \n\nimport os\nimport json\nimport torch\nimport numpy as np\nfrom torch.utils.data import Dataset\nimport pandas as pd\nimport argparse\nimport copy\nfrom collections import Counter\nimport random\n\nclass MOVIE(Dataset):\n def __init__(self, args, df, unique_item_num):\n super().__init__()\n\n self.m_data_dir = args.data_dir\n self.m_batch_size = args.batch_size\n\n self.m_sample_num = len(df)\n print(\"sample num\", self.m_sample_num)\n\n self.m_batch_num = int(self.m_sample_num/self.m_batch_size)\n print(\"batch num\", self.m_batch_num)\n\n if (self.m_sample_num/self.m_batch_size-self.m_batch_num) > 0:\n self.m_batch_num += 1\n \n self.m_user_batch_list = []\n # self.m_pos_item_batch_list = []\n # self.m_neg_item_batch_list = []\n self.m_item_batch_list = []\n\n self.m_user2uid = {}\n self.m_item2iid = {}\n\n # userid_list = df.userid.tolist()\n # pos_itemid_list = df.pos_itemid.tolist()\n # neg_itemidlist_list = df.neg_itemid.tolist()\n\n user_itemlist_dict = dict(df.groupby(\"userid\").itemid.apply(list))\n unique_userid_list = list(user_itemlist_dict.keys())\n unique_user_num = len(unique_userid_list)\n\n # assert unique_user_num != df.userid.nunique()\n self.m_unique_item_num = unique_item_num\n\n for user_index in range(unique_user_num):\n user_id = unique_userid_list[user_index]\n itemid_list = user_itemlist_dict[user_id]\n \n self.m_user_batch_list.append(user_id)\n self.m_item_batch_list.append(itemid_list)\n # self.m_pos_item_batch_list.append(pos_itemid)\n # self.m_neg_item_batch_list.append(neg_itemid_list)\n\n print(\"... load train data ...\", len(self.m_user_batch_list), len(self.m_item_batch_list))\n\n def __len__(self):\n return len(self.m_user_batch_list)\n\n def __getitem__(self, idx):\n if torch.is_tensor(idx):\n idx = idx.tolist()\n \n i = idx\n\n user_i = self.m_user_batch_list[i]\n itemlist_i = self.m_item_batch_list[i]\n \n return user_i, itemlist_i, self.m_unique_item_num\n \n @staticmethod\n def collate(batch):\n batch_size = len(batch)\n\n user_iter = []\n positem_iter = []\n negitem_iter = []\n\n positemnum_iter = []\n\n unique_item_num = 0\n for i in range(batch_size):\n sample_i = batch[i]\n \n itemlist_i = sample_i[1]\n positemnum_iter.append(len(itemlist_i))\n\n unique_item_num = sample_i[2]\n\n maxitemnum_iter = max(positemnum_iter)\n\n unique_itemlist = range(0, unique_item_num)\n unique_itemset = set(unique_itemlist)\n neg_samplenum = 500\n\n for i in range(batch_size):\n sample_i = batch[i]\n \n user_i = sample_i[0]\n user_iter.append(user_i)\n\n itemlist_i = copy.deepcopy(sample_i[1])\n negitemset_i = set(itemlist_i)^unique_itemset\n negitemlist_i = list(negitemset_i)\n\n sampled_negitemlist_i = random.sample(negitemlist_i, k=neg_samplenum)\n\n itemnum_i = len(itemlist_i) \n itemlist_i = itemlist_i+[0]*(maxitemnum_iter-itemnum_i)\n positem_iter.append(itemlist_i)\n\n negitem_iter.append(sampled_negitemlist_i) \n\n user_iter_tensor = torch.from_numpy(np.array(user_iter)).long()\n positem_iter_tensor = torch.from_numpy(np.array(positem_iter)).long()\n negitem_iter_tensor = torch.from_numpy(np.array(negitem_iter)).long()\n\n positemnum_iter_tensor = torch.from_numpy(np.array(positemnum_iter)).long()\n\n return user_iter_tensor, positem_iter_tensor, negitem_iter_tensor, positemnum_iter_tensor\n\nclass MOVIE_TEST(Dataset):\n def __init__(self, args, train_df, df):\n super().__init__()\n\n self.m_data_dir = args.data_dir\n self.m_batch_size = args.batch_size\n\n self.m_sample_num = len(df)\n print(\"sample num\", self.m_sample_num)\n\n self.m_batch_num = int(self.m_sample_num/self.m_batch_size)\n print(\"batch num\", self.m_batch_num)\n\n if (self.m_sample_num/self.m_batch_size-self.m_batch_num) > 0:\n self.m_batch_num += 1\n \n self.m_user_batch_list = []\n self.m_item_batch_list = []\n self.m_maskitem_batch_list = []\n\n self.m_user2uid = {}\n self.m_item2iid = {}\n\n userid_list = df.userid.tolist()\n itemid_list = df.itemid.tolist()\n\n user_itemlist_dict = dict(df.groupby(\"userid\").itemid.apply(list))\n unique_userid_list = list(user_itemlist_dict.keys())\n unique_user_num = len(unique_userid_list)\n print(\"unique_user_num\", unique_user_num)\n print(\"user num unique in df\", df.userid.nunique())\n\n assert unique_user_num == df.userid.nunique()\n\n user_maskitemlist_dict = dict(train_df.groupby(\"userid\").itemid.apply(list))\n\n for user_index in range(unique_user_num):\n user_id = unique_userid_list[user_index]\n itemid_list = user_itemlist_dict[user_id]\n maskitemid_list = user_maskitemlist_dict[user_id]\n\n self.m_user_batch_list.append(user_id)\n self.m_item_batch_list.append(itemid_list)\n self.m_maskitem_batch_list.append(maskitemid_list)\n\n print(\"... load train data ...\", len(self.m_user_batch_list), len(self.m_item_batch_list))\n\n def __len__(self):\n return len(self.m_user_batch_list)\n\n def __getitem__(self, idx):\n if torch.is_tensor(idx):\n idx = idx.tolist()\n \n i = idx\n\n user_i = self.m_user_batch_list[i]\n item_i = self.m_item_batch_list[i]\n maskitem_i = self.m_maskitem_batch_list[i]\n\n return user_i, item_i, maskitem_i\n \n @staticmethod\n def collate(batch):\n batch_size = len(batch)\n\n user_iter = []\n item_iter = []\n itemnum_iter = []\n max_item_num = 0\n\n maskitem_iter = []\n max_maskitem_num = 0\n\n for i in range(batch_size):\n sample_i = batch[i]\n item_i = sample_i[1]\n max_item_num = max(max_item_num, len(item_i))\n\n maskitem_i = sample_i[2]\n max_maskitem_num = max(max_maskitem_num, len(maskitem_i))\n\n pad_item_id = 0\n for i in range(batch_size):\n sample_i = batch[i]\n \n user_i = sample_i[0]\n user_iter.append(user_i)\n\n item_i = copy.deepcopy(sample_i[1])\n itemnum_i = len(item_i)\n itemnum_iter.append(itemnum_i)\n\n item_i = item_i+[pad_item_id]*(max_item_num-itemnum_i)\n item_iter.append(item_i)\n\n maskitem_i = copy.deepcopy(sample_i[2])\n maskitemnum_i = len(maskitem_i)\n maskitem_i = maskitem_i+[pad_item_id]*(max_maskitem_num-maskitemnum_i)\n maskitem_iter.append(maskitem_i)\n\n user_iter_tensor = torch.from_numpy(np.array(user_iter)).long()\n item_iter_tensor = torch.from_numpy(np.array(item_iter)).long()\n \n itemnum_iter_tensor = torch.from_numpy(np.array(itemnum_iter)).long()\n maskitem_iter_tensor = torch.from_numpy(np.array(maskitem_iter)).long()\n\n return user_iter_tensor, item_iter_tensor, maskitem_iter_tensor, itemnum_iter_tensor\n", "sub_path": "sampleBPR/BPR/movie.py", "file_name": "movie.py", "file_ext": "py", "file_size_in_byte": 7577, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "torch.utils.data.Dataset", "line_number": 22, "usage_type": "name"}, {"api_name": "torch.is_tensor", "line_number": 72, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 113, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 117, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 125, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 126, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 127, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 129, "usage_type": "call"}, {"api_name": "torch.utils.data.Dataset", "line_number": 133, "usage_type": "name"}, {"api_name": "torch.is_tensor", "line_number": 184, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 222, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 229, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 234, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 234, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 235, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 235, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 237, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 237, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 238, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 238, "usage_type": "call"}]} +{"seq_id": "439426351", "text": "import astropy.coordinates as coord\nimport astropy.units as u\nimport numpy as np\nfrom .config import rsun, vcirc\n\nro = rsun\nvo = vcirc\n\n\ndef gala_to_galpy_orbit(w):\n from galpy.orbit import Orbit\n\n # PhaseSpacePosition or Orbit:\n cyl = w.cylindrical\n\n R = cyl.rho.to_value(ro).T\n phi = cyl.phi.to_value(u.rad).T\n z = cyl.z.to_value(ro).T\n\n vR = cyl.v_rho.to_value(vo).T\n vT = (cyl.rho * cyl.pm_phi).to_value(vo, u.dimensionless_angles()).T\n vz = cyl.v_z.to_value(vo).T\n\n o = Orbit(np.array([R, vR, vT, z, vz, phi]).T, ro=ro, vo=vo)\n\n if hasattr(w, 't'):\n o.t = w.t.to_value(u.Myr)\n\n return o\n\n\ndef get_staeckel_actions(w, potential):\n from galpy.actionAngle import estimateDeltaStaeckel, actionAngleStaeckel\n\n R = w.cylindrical.rho.to_value(ro)\n z = w.z.to_value(ro)\n delta = estimateDeltaStaeckel(potential, R, z)\n\n o = gala_to_galpy_orbit(w)\n aAS = actionAngleStaeckel(pot=potential, delta=delta)\n return np.squeeze(aAS(o)) * ro * vo\n\n\ndef get_staeckel_aaf(w, potential):\n from galpy.actionAngle import estimateDeltaStaeckel, actionAngleStaeckel\n\n R = w.cylindrical.rho.to_value(ro)\n z = w.z.to_value(ro)\n delta = estimateDeltaStaeckel(potential, R, z)\n\n o = gala_to_galpy_orbit(w)\n aAS = actionAngleStaeckel(pot=potential, delta=delta)\n\n aaf = aAS.actionsFreqsAngles(o)\n aaf = {'actions': np.squeeze(aaf[:3]) * ro * vo,\n 'freqs': np.squeeze(aaf[3:6]) * vo / ro,\n 'angles': coord.Angle(np.squeeze(aaf[6:]) * u.rad)}\n\n return aaf\n", "sub_path": "thriftshop/galpy_helpers.py", "file_name": "galpy_helpers.py", "file_ext": "py", "file_size_in_byte": 1542, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "config.rsun", "line_number": 6, "usage_type": "name"}, {"api_name": "config.vcirc", "line_number": 7, "usage_type": "name"}, {"api_name": "astropy.units.rad", "line_number": 17, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 17, "usage_type": "name"}, {"api_name": "astropy.units.dimensionless_angles", "line_number": 21, "usage_type": "call"}, {"api_name": "astropy.units", "line_number": 21, "usage_type": "name"}, {"api_name": "galpy.orbit.Orbit", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 24, "usage_type": "call"}, {"api_name": "astropy.units.Myr", "line_number": 27, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 27, "usage_type": "name"}, {"api_name": "galpy.actionAngle.estimateDeltaStaeckel", "line_number": 37, "usage_type": "call"}, {"api_name": "galpy.actionAngle.actionAngleStaeckel", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 41, "usage_type": "call"}, {"api_name": "galpy.actionAngle.estimateDeltaStaeckel", "line_number": 49, "usage_type": "call"}, {"api_name": "galpy.actionAngle.actionAngleStaeckel", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 56, "usage_type": "call"}, {"api_name": "astropy.coordinates.Angle", "line_number": 57, "usage_type": "call"}, {"api_name": "astropy.coordinates", "line_number": 57, "usage_type": "name"}, {"api_name": "numpy.squeeze", "line_number": 57, "usage_type": "call"}, {"api_name": "astropy.units.rad", "line_number": 57, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 57, "usage_type": "name"}]} +{"seq_id": "622994005", "text": "from banal import ensure_list\n\nfrom followthemoney import model\nfrom followthemoney.compare import compare\n\n\nclass Result(object):\n\n def __init__(self, enricher, subject):\n self.enricher = enricher\n self._entities = {}\n self._candidate = None\n self._subject = None\n self.set_subject(subject)\n\n def make_entity(self, schema):\n return model.make_entity(schema, key_prefix=self.enricher.key_prefix)\n\n def add_entity(self, entity):\n if entity is None or entity.id is None:\n return\n if entity.id in self._entities:\n self._entities[entity.id].merge(entity)\n else:\n self._entities[entity.id] = entity\n\n def set_candidate(self, entity):\n if entity is None or entity.id is None:\n return\n self.add_entity(entity)\n self._candidate = entity.id\n\n def set_subject(self, entity):\n if entity is None or entity.id is None:\n return\n self.add_entity(entity)\n self._subject = entity.id\n\n @property\n def entities(self):\n return self._entities.values()\n\n @property\n def subject(self):\n return self._entities.get(self._subject)\n\n @property\n def candidate(self):\n return self._entities.get(self._candidate)\n\n @property\n def score(self):\n if self.subject is None or self.candidate is None:\n return 0.0\n if self.subject.id == self.candidate.id:\n return 1.0\n return compare(model, self.subject, self.candidate)\n\n def to_dict(self):\n return {\n 'entities': [e.to_dict() for e in self.entities],\n 'subject': self._subject,\n 'candidate': self._candidate,\n 'enricher': self.enricher.name\n }\n\n @classmethod\n def from_dict(cls, enricher, data):\n result = cls(enricher, None)\n entities = ensure_list(data.get('entities'))\n entities = [model.get_proxy(e) for e in entities]\n entities = {e.id: e for e in entities}\n result._entities = entities\n result._subject = data.get('subject')\n result._candidate = data.get('candidate')\n return result\n", "sub_path": "enrich/followthemoney_enrich/result.py", "file_name": "result.py", "file_ext": "py", "file_size_in_byte": 2190, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "followthemoney.model.make_entity", "line_number": 17, "usage_type": "call"}, {"api_name": "followthemoney.model", "line_number": 17, "usage_type": "name"}, {"api_name": "followthemoney.compare.compare", "line_number": 57, "usage_type": "call"}, {"api_name": "followthemoney.model", "line_number": 57, "usage_type": "argument"}, {"api_name": "banal.ensure_list", "line_number": 70, "usage_type": "call"}, {"api_name": "followthemoney.model.get_proxy", "line_number": 71, "usage_type": "call"}, {"api_name": "followthemoney.model", "line_number": 71, "usage_type": "name"}]} +{"seq_id": "167739550", "text": "import discord\nfrom discord.ext import commands\nimport asyncio\nimport os\nimport json\nimport random\n\nclass Duel(commands.Cog):\n currentPlaying = []\n lightAtk = {\"{} threw a jab at {}!\": \"https://thumbs.gfycat.com/PartialGenuineAbyssiniancat-max-1mb.gif\",\n \" {} smacked {} on the head!\": \"https://media1.giphy.com/media/Xj7aX90bEKoec/giphy.gif\", \n \"{} threw a right hook at {}!\":\"https://media1.giphy.com/media/3ornk9vVI7DsFYNAR2/giphy.gif\", \n \"{} threw a karate reverse punch at {}!\": \"https://66.media.tumblr.com/9e3ff679d0ad3ca25b553f6d55881055/tumblr_ocmsthRUSu1su3rezo1_400.gif\",\n \"{} kicked {}'s leg!\": \"http://theselfdefenceexpert.com/wp-content/uploads/2014/01/low-kick-gif-alves.gif\",\n \"{} used a palm strike on {}!\": \"https://media1.giphy.com/media/OLHy9ERaFUvzW/giphy.gif\",\n \"{} threw a capoeira handstand kick at {}!\": \"https://66.media.tumblr.com/282c8847ebbe7422135a0a57821dc44c/tumblr_p7hg2kQLuj1v6w3juo4_500.gif\",\n \"{} thew elbow strikes at {}!\": \"https://66.media.tumblr.com/c4f24c77e3d5dfc0e80ab8583fc65080/tumblr_pcqp5pM1R01rmrpdmo2_400.gif\",\n \"{} grabbed two swords and started stabbing {}!\": \"https://thumbs.gfycat.com/AcidicImmenseHapuku-size_restricted.gif\",\n \"{} landed two devastating punches at {}'s stomach!\": \"https://media.giphy.com/media/nD2mLa2S7Cz6/giphy.gif\",\n \"{} karate chopped {}!\": \"https://media1.tenor.com/images/a300542f540340d9b9b59eaf6a69d352/tenor.gif?itemid=5137929\",\n \"{} threw {} to the ground!\": \"https://media1.giphy.com/media/9lLbEiVBO3OY8/giphy.gif\",\n \"{} did a takedown on {}!\": \"http://i40.tinypic.com/291doup.gif\",\n \"{} slapped {}!\": \"https://thumbs.gfycat.com/DecentBelatedAntbear-size_restricted.gif\",\n \"{} threw a flurry of Wing Chun punches at {}!\": \"https://media2.giphy.com/media/Cf3ZcH9D3uHtK/giphy.gif\",\n \"{} kicked {}'s ass...literally.\": \"https://media2.giphy.com/media/xT5LMY2kXko8zjyiwU/source.gif\",\n \"{} headbutted {}!\": \"https://media1.tenor.com/images/b943da307642d8dedf3a495b6bdb48c4/tenor.gif?itemid=5114718\",\n \"{} punched {}'s throat!\": \"https://media.giphy.com/media/13HXKG2HGN8aPK/giphy.gif\"\n }\n\n heavyAtk = {\"{} threw a Muay Thai jumping elbow at {}!\": \"https://cdn.ebaumsworld.com/mediaFiles/picture/31430/81968730.gif\",\n \"{} threw a front kick straight into {}'s face!\": \"https://media.giphy.com/media/xT1R9JqCleRlz4jnvq/giphy.gif\",\n \"{} faked a spin kick and then threw another spin kick at {}!\": \"https://i.pinimg.com/originals/26/b3/72/26b372434422e37d39da16d99c792a87.gif\",\n \"{} threw an urumawashi geri (Karate hook kick) at {}!\": \"https://66.media.tumblr.com/92eae13144a2c558cddc611b7860ebc7/tumblr_ot6sigmrzp1smsd90o1_400.gif\",\n \"{} sweeped and slammed {} into the ground!\": \"https://66.media.tumblr.com/07ac5647d522cb71a06371def2216d32/tumblr_op12h97In01v6w3juo1_500.gif\",\n \"{} threw a flying knee at {}!\": \"https://i0.wp.com/78.media.tumblr.com/0f4fc9616b3113ad92eeb771950910e2/tumblr_oo0bwozKxH1rmrpdmo1_400.gif?resize=350%2C200&ssl=1\",\n \"{} lands a big uppercut at {}!\": \"https://media1.giphy.com/media/AM0NtWwYBdoxW/giphy.gif\",\n \"{} lands a flurry of punches at {}!\": \"https://i.redd.it/ouqn4rpssg901.gif\",\n \"{} lands a meia lua on {}'s face!\": \"http://i.imgur.com/6tpk9QG.gif\",\n \"{} hit {} with a flurry of punches and a spinning jump kick!\": \"https://static.fjcdn.com/gifs/Good+reaction+gif+brilredbig3also+what+would+you+call+the+finishing_1970c0_5251657.gif\",\n \"{} landed a fancy kick at {}'s face!\": \"https://66.media.tumblr.com/77e4d275d8c5541030364e95a250e4b2/tumblr_oqvll1zWfV1rmrpdmo1_400.gif\",\n \"{} side kicked {}!\": \"http://24.media.tumblr.com/f5da63478bde392418aebe858e19279e/tumblr_mgdokseHXu1r2xo28o1_500.gif\",\n \"{} lands an axe kick on {}'s face!\": \"https://media3.giphy.com/media/rQX1Ogknv7boI/giphy.gif\",\n \"{} lands a big roundhouse kick on {}!\": \"http://giant.gfycat.com/HarmlessBlueArawana.gif\",\n \"{} did some fancy flying kick at {}!\": \"https://moviesfilmsandflix.files.wordpress.com/2018/02/scott-adkins-kick-gif.gif\",\n \"{} kicked {} in the crotch!\": \"https://media1.tenor.com/images/60b54c9f2604313502c1971f4d3e7195/tenor.gif?itemid=5535907\",\n \"{} just punched {}'s....oof\": \"https://i.makeagif.com/media/10-25-2015/bsXHxC.gif\",\n \"{} lands a Do Mawashi Kaiten Geri on {}!\": \"https://i.pinimg.com/originals/70/3d/cf/703dcf86945f375ecfef5141e8e2aff5.gif\",\n \"{} did a fancy takedown on {}!\": \"https://i.makeagif.com/media/3-05-2016/5D6gCK.gif\",\n \"{} kicked {} multiple times in the chest!\": \"https://moviesfilmsandflix.files.wordpress.com/2017/09/tumblr_ou2idnirmf1v6w3juo4_500.gif\",\n \"{} distracted {} with a suitcase and threw a side kick!\": \"https://cdn.playbuzz.com/cdn/41ffd427-5d74-4fb7-8ea4-1ed2cb98eb87/0c74ec17-04ad-4c62-9e12-9e2fd5d693b4.gif\",\n \"{} did a flying capoeira kick at {}!\": \"https://i.imgflip.com/1tfgax.gif\",\n \"{} threw a 360 spin kick at {}!\": \"https://media.giphy.com/media/1QNIxRWZGKpgI/giphy.gif\",\n \"{} knocked {} out with a spinning kick thing!\": \"https://66.media.tumblr.com/d559f50518c1e57d8d1fb732fc6e55ce/tumblr_mqwnhx9ZTC1r61a7no2_250.gif\",\n \"{} misses a leg kick and still hits {} with a roundhouse kick!\": \"https://66.media.tumblr.com/tumblr_m3afej40m91ro43xzo1_400.gif\",\n \"{} hammer kicked a board and punched {} through the board!\": \"https://66.media.tumblr.com/2f3d497c2ae87e71037aaa06bf4724de/tumblr_ori1nrncjh1rmrpdmo1_400.gif\",\n \"{} punched {} through a statue!\": \"https://66.media.tumblr.com/0b595d67b9ecd1f56832c0d9cc84c3f0/tumblr_p2968hqraP1rmrpdmo4_400.gif\",\n \"{} used judo throw on {}!\": \"https://thumbs.gfycat.com/ImpishCompassionateAustraliancurlew-size_restricted.gif\",\n \"{} kicked {}'s legs while he/she was on a table! Don't stand on furnitures kids.\": \"https://66.media.tumblr.com/c76d877a93f725e446583f1af1cf2283/tumblr_or4b6u4Xno1syv3zao1_400.gif\",\n \"{} threw a spinning head kick at {}!\":\"https://media.giphy.com/media/gasmaAkDqzcFW/giphy.gif\",\n \"{} smashed {}'s head with two sticks!\":\"https://i.gifer.com/Im8u.gif\",\n \"{} threw big hooks at {}!\":\"https://i.makeagif.com/media/10-04-2015/2rtp-R.gif\",\n \"{} got some mad anime skills and {} took some heavy hits! Thank god the sword was blunt...\":\"https://i.gifer.com/Gpw9.gif\",\n \"{} pulled a Jackie Chan on {}!\":\"https://66.media.tumblr.com/4c9878322d4bc5315cc5fa6207b66e3f/tumblr_onooa9fsWG1v6w3juo1_500.gif\"\n }\n \n blocks = {\"{} blocked all of {}'s attacks!\": \"https://66.media.tumblr.com/6e29fc2555517475c055300480e56922/tumblr_p1643yvSA81rmrpdmo1_400.gif\",\n \"{} parried {}'s attacks!\": \"https://i2.wp.com/warriorpunch.com/wp-content/uploads/2017/08/mirror.gif?resize=595%2C335&ssl=1\",\n \"{} dodged all of {}'s attacks!\": \"https://media0.giphy.com/media/Tsl0wXsKDiu7S/giphy.gif\",\n \"{} grabbed a shield and blocked {}'s attacks!\": \"https://media.giphy.com/media/l0HlymZ7Jv6JoiYjC/giphy.gif\",\n \"{} blocked {}'s knife attack!\": \"https://media.giphy.com/media/E8FjnBF1TOpO0/giphy.gif\",\n \"{} pulled a Neo and blocked {}'s attacks with ease!\": \"https://media0.giphy.com/media/HTjQEEGqyQTXq/giphy.gif\",\n \"{} evaded {}'s attacks!\":\"https://media1.tenor.com/images/82e9b9f2152e7e375ac845aa52de8870/tenor.gif?itemid=7588651\",\n \"{} caught {}'s kick!\": \"https://media.giphy.com/media/7XujHz25hunSg/giphy.gif\",\n \"{} dodged {}'s leg sweep!\": \"http://mmafury.com/wp-content/uploads/2015/02/Muay-Thai-Fighter-goes-for-a-leg-sweep-but-nope...quick-little-jump-nice-balance-GIF.gif\",\n \"{} blocked {}'s attacks!\": \"https://66.media.tumblr.com/8addede13d1c6dd47bb33c2e7440136b/tumblr_pk3de2L4Sb1v6w3juo3_500.gif\",\n \"{} disarmed {}!\": \"https://media2.giphy.com/media/MnefNOdkdnDr2/giphy.gif\",\n \"{} blocked all of {}'s punches with elbows!\": \"https://i.imgur.com/nxxApiV.gif?noredirect\",\n \"{} ducked {}'s roundhouse kick!\": \"https://i.pinimg.com/originals/1f/70/d8/1f70d861132c952b7f2aeb794778025a.gif\",\n \"{} got drunk and dodged {}'s attacks!\": \"https://66.media.tumblr.com/dcc6b61ed548038911e915ea94c88078/tumblr_pbvql6zaHu1xthcgdo3_500.gif\",\n \"{} got mad at {} for interfering with his/her shopping!\": \"https://thumbs.gfycat.com/BraveFailingCrayfish-size_restricted.gif\",\n \"{} dodged all of {}'s punches!\": \"https://media1.tenor.com/images/70d2210f5419789e4342c603c210a09f/tenor.gif?itemid=11739995\",\n \"{} parried {}'s attacks!\": \"https://66.media.tumblr.com/3ad228014391bc9a8d87bd521b6c3d66/tumblr_n3m70z1qYv1tw9yl5o4_400.gif\",\n \"{} blocked and dodged {}'s three-section staff attacks! \": \"https://i2.wp.com/78.media.tumblr.com/316497e5ade9bcb0c5ba94d9f12cbc68/tumblr_p96jtrEehi1wstc5to1_400.gif?w=605&ssl=1\",\n \"{} broke {}'s sword!\":\"https://thumbs.gfycat.com/MatureDecisiveHedgehog-size_restricted.gif\",\n \"{} ducked {}'s spinning kick!\":\"https://thumbs.gfycat.com/PoorReliableAddax-size_restricted.gif\",\n \"{} caught {}'s punch!\":\"https://media0.giphy.com/media/5jYeXVJRkf0mfOont4/giphy.gif\",\n \"{} parried all of {}'s punches and kicks!\":\"https://i.imgur.com/axgo0ye.gif?noredirect\"\n }\n\n nothingUrl = [\"https://thumbs.gfycat.com/PassionateRectangularBrahmancow-small.gif\", \"https://media.tenor.com/images/eb84105a1eb998307819b358ac485528/tenor.gif\", \"https://media0.giphy.com/media/tXL4FHPSnVJ0A/giphy.gif\", \n \"https://media0.giphy.com/media/l2JhpjWPccQhsAMfu/giphy.gif\",\n \"https://i.kym-cdn.com/photos/images/newsfeed/001/057/927/eac.gif\",\n \"https://thumbs.gfycat.com/BogusFarawayBream-max-1mb.gif\"\n ]\n def __init__(self, bot):\n self.bot = bot\n \n #linkboi command\n @commands.command(pass_context = True)\n async def give(self, ctx, user: discord.Member, amount):\n if ctx.author.id != 217380909815562241:\n await ctx.send(\"You're not my creator -.-\")\n else:\n cog = self.bot.get_cog(\"mafia\")\n await ctx.send(\"{} Mafia Points has been added to {}'s account sir.\".format(str(amount), user.name))\n self.addPoint(user.id, int(amount), cog)\n \n #linkboi command\n @commands.command(pass_context = True)\n async def deduct(self, ctx, user: discord.Member, amount):\n if ctx.author.id != 217380909815562241:\n await ctx.send(\"You're not my creator -.-\")\n else:\n cog = self.bot.get_cog(\"mafia\")\n await ctx.send(\"{} Mafia Points has been removed from {}'s account sir.\".format(str(amount), user.name))\n self.addPoint(user.id, -1*int(amount), cog)\n \n\n @commands.command(pass_context= True)\n async def transfer(self, ctx, user:discord.Member, amount:int):\n if user == ctx.author:\n await ctx.channel.send(\"Don't give money to yourself. That's just sad.\")\n return\n \n if user.bot:\n await ctx.channel.send(\"Don't give a bot money. They prob don't want it lmao.\")\n return\n if amount<0:\n await ctx.channel.send(\"Lol you can't give negative money. Go back to school kid.\")\n return\n cog = self.bot.get_cog(\"mafia\")\n cog.checkFile(user.id)\n cog.checkFile(ctx.author.id)\n giver = cog.findUser(ctx.author.id)\n receiver = cog.findUser(user.id)\n\n\n if giver.points < amount:\n await ctx.channel.send(\"Boi, you don't have that much money. Nice try.\")\n return\n \n receiver.points += amount\n giver.points -= amount\n\n cog.editFile(receiver)\n cog.editFile(giver)\n\n embed = discord.Embed(title = str(amount) + \":moneybag: have been transferred from \" + ctx.author.name + \" to \" + user.name + \"'s account.\", colour = discord.Colour.green())\n embed.add_field(name = ctx.author.name + \"'s balance:\", value = str(giver.points) + \":moneybag:\")\n embed.add_field(name = user.name + \"'s balance:\", value = str(receiver.points) + \":moneybag: \")\n embed.set_thumbnail(url = user.avatar_url)\n embed.set_image(url = \"https://www.retailgazette.co.uk/wp-content/uploads/shutterstock_465600824.jpg\")\n await ctx.channel.send(embed = embed)\n\n @commands.command(pass_context = True)\n async def duel(self, ctx, victim: discord.Member):\n channel = ctx.channel\n player1 = ctx.author\n if victim.bot:\n await channel.send(\"You can't challenge a bot. You'll definitely lose!\")\n elif victim in self.currentPlaying:\n await channel.send(\"{} is currently in a duel!\".format(victim.name))\n elif player1 in self.currentPlaying:\n await channel.send(\"{} is currently in a duel!\".format(player1.name))\n elif victim != player1:\n try:\n challenge = discord.Embed(title = \"{}, {} has challenged you to a duel. Do you accept?\".format(victim.name, player1.name), description = \"y/n\", colour = discord.Colour.red())\n challenge.set_author(name = player1.name, icon_url=player1.avatar_url)\n challenge.set_thumbnail(url = victim.avatar_url)\n challenge.set_image(url = \"https://i.ytimg.com/vi/cb5DITStXlI/maxresdefault.jpg\")\n await channel.send(embed = challenge)\n try:\n self.currentPlaying.append(victim)\n self.currentPlaying.append(player1)\n answer = await self.bot.wait_for('message', check=lambda message: message.author == victim and (message.content == \"y\" or message.content == \"n\" or message.content == \"yes\" or message.content == \"no\"), timeout = 30)\n if answer.content == \"y\" or answer.content == \"yes\":\n \n supportChannel = self.bot.get_channel(550923896858214446)\n await supportChannel.send(\"{} started duel with {}!\".format(player1.name, victim.name))\n await channel.send(\"Let the duel begin!\")\n await self.playDuel(player1, victim, channel)\n await supportChannel.send(\"{} ended duel with {}!\".format(player1.name, victim.name))\n \n else:\n\n await channel.send(\"{} declined the duel. Guess someone's too scared.\".format(victim.name))\n \n except asyncio.TimeoutError:\n await channel.send(\"The session expired.\")\n self.currentPlaying.remove(victim)\n self.currentPlaying.remove(player1)\n except:\n if victim in self.currentPlaying:\n self.currentPlaying.remove(victim)\n if player1 in self.currentPlaying:\n self.currentPlaying.remove(player1)\n else:\n await channel.send(\"You can't challenge yourself dummy.\")\n \n async def playDuel(self, player1, player2, channel):\n\n moveList = [None, self.lightAtk, self.heavyAtk, self.blocks]\n p1Health = 100\n p2Health = 100\n p1Disabled = False\n p2Disabled = False\n await channel.send(\"{} please check your DM!\".format(player1.name))\n while p1Health !=0 and p2Health!=0:\n if not p1Disabled:\n p1Moves = await self.getMoves(player1)\n else:\n p1Moves = None\n\n if not p2Disabled:\n p2Moves = await self.getMoves(player2)\n else:\n p2Moves = None\n p1Disabled = False\n p2Disabled = False\n count = 0\n\n await asyncio.sleep(2)\n #determines if p1 wins this round\n p1GetsTurn = await self.isWinner(player1, player2, p1Moves, p2Moves, count, channel)\n \n if p1GetsTurn == None:\n embed= discord.Embed(title = \"Lol nothing happened.\", colour= discord.Colour.red())\n embed.set_image(url = random.choice(self.nothingUrl))\n await channel.send(embed = embed)\n await asyncio.sleep(3)\n elif p1GetsTurn:\n p1MoveInt = int(p1Moves)\n damage = None\n if p1MoveInt == 1:\n damage = random.randint(10, 20)\n p2Health -= damage\n elif p1MoveInt == 2:\n damage = random.randint(21, 30)\n p2Health -= damage\n elif p1MoveInt == 3:\n p2Disabled = True\n if p2Health < 0:\n p2Health = 0\n moveType = moveList[p1MoveInt]\n content = random.choice(list(moveType.keys()))\n\n embed = discord.Embed(title = content.format(player1.name, player2.name), colour = discord.Colour.red())\n if damage != None:\n embed.add_field(name = \"{} took {} damage!\".format(player2.name, str(damage)), value = \"{} now has {} health!\".format(player2.name, str(p2Health)))\n if p1MoveInt == 3:\n embed.add_field(name = \"{} is stunned next turn!\".format(player2.name), value = \"Oh boi...\")\n embed.set_image(url = moveType[content])\n embed.set_author(name = player1.name, icon_url=player1.avatar_url)\n await channel.send(embed = embed)\n await asyncio.sleep(3)\n else:\n if p2Moves != None:\n p2MoveInt = int(p2Moves)\n damage = None\n if p2MoveInt == 1:\n damage = random.randint(10, 20)\n p1Health -= damage\n elif p2MoveInt == 2:\n damage = random.randint(21, 30)\n p1Health -= damage\n elif p2MoveInt == 3:\n p1Disabled = True\n if p1Health < 0:\n p1Health = 0\n moveType = moveList[p2MoveInt]\n content = random.choice(list(moveType.keys()))\n\n embed = discord.Embed(title = content.format(player2.name, player1.name), colour = discord.Colour.red())\n if damage != None:\n embed.add_field(name = \"{} took {} damage!\".format(player1.name, str(damage)), value = \"{} now has {} health!\".format(player1.name, str(p1Health)))\n embed.set_image(url = moveType[content])\n embed.set_author(name = player2.name, icon_url=player2.avatar_url)\n if p2MoveInt == 3:\n embed.add_field(name = \"{} is stunned next turn!\".format(player1.name), value = \"Oh boi...\")\n await channel.send(embed = embed)\n await asyncio.sleep(3)\n else:\n await channel.send(\"Guess no one's playing. Game over then.\")\n \n break \n cog = self.bot.get_cog(\"mafia\")\n cog.checkFile(player1.id)\n cog.checkFile(player2.id)\n \n if p1Health == 0:\n randPoints = random.randint(10, 30)\n self.addPoint(player2.id, randPoints, cog)\n embed = discord.Embed(title = \"{} wins the duel!\".format(player2.name), colour = discord.Colour.blue())\n embed.set_thumbnail(url = player2.avatar_url)\n embed.add_field(name = \"{} received: \".format(player2.name), value = \"{} Mafia points.\".format(randPoints))\n elif p2Health == 0:\n randPoints = random.randint(10, 30)\n self.addPoint(player1.id, randPoints, cog)\n embed = discord.Embed(title =\"{} wins the duel!\".format(player1.name), colour = discord.Colour.blue())\n embed.set_thumbnail(url = player1.avatar_url)\n embed.add_field(name = \"{} received: \".format(player1.name), value = \"{} Mafia points.\".format(randPoints))\n else:\n embed = discord.Embed(title = \"No contest.\")\n embed.set_image(url = \"https://thumbs.gfycat.com/FailingFavorableBaleenwhale-size_restricted.gif\")\n await channel.send(embed = embed)\n \n \n\n\n \n async def getMoves(self, member):\n moves = None\n count = 1\n\n embed = discord.Embed(title = \"Alright fighter, choose your move.\", description = \"Enter the number associated with your choice.\", colour = discord.Colour.red())\n embed.add_field(name = \"1. Light attack\", value = \"Beats a heavy attack, but can be blocked.\")\n embed.add_field(name = \"2. Heavy attack\", value = \"Beats a block, but can be interupted by a light attack.\")\n embed.add_field(name = \"3. Block\", value = \"Blocks light attack and stuns the opponent, but cannot block heavy attack.\")\n embed.set_image(url = \"https://art.ngfiles.com/images/332000/332918_phatalphd_cowboy-standoff.jpg?f1419452016\")\n await member.send(embed = embed)\n try:\n answer = await self.bot.wait_for('message', check=lambda message: message.author == member and (message.content == \"1\" or message.content == \"2\" or message.content == \"3\"), timeout = 30)\n moves = answer.content\n await member.send(\"Got it. Return to the channel.\")\n\n except asyncio.TimeoutError:\n await member.send(\"Boi you afk. You gonna die.\")\n moves = None\n return moves\n \n async def isWinner(self, player1, player2, playerMoves1, playerMoves2, count, channel):\n #returns whether parameter player1 gets move\n moveNames = [\"light attack\", \"heavy attack\", \"block\"]\n if playerMoves1 == None:\n return False\n elif playerMoves2 == None:\n return True\n person = playerMoves1\n personInt = int(person)\n\n opponent = playerMoves2\n oppInt = int(opponent)\n embed = discord.Embed(title = \"{} chose a {} and {} chose a {}!\".format(player1.name, moveNames[personInt-1], player2.name, moveNames[oppInt-1]), colour = discord.Colour.red())\n await channel.send(embed = embed)\n\n if person == opponent: #If attacks are the same\n return None\n elif person == \"3\" and opponent == \"1\":\n return True\n elif (person == \"1\" and opponent == \"3\") or (int(person)> int(opponent)):\n return False\n else:\n return True\n \n def addPoint(self, userID, points, cog):\n user = cog.findUser(userID)\n user.points += points\n cog.editFile(user)\n\n\n\ndef setup(bot):\n bot.add_cog(Duel(bot))\n", "sub_path": "MINIGAMES/duel.py", "file_name": "duel.py", "file_ext": "py", "file_size_in_byte": 22379, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "discord.ext.commands.Cog", "line_number": 8, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 8, "usage_type": "name"}, {"api_name": "discord.Member", "line_number": 100, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.command", "line_number": 99, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 99, "usage_type": "name"}, {"api_name": "discord.Member", "line_number": 110, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.command", "line_number": 109, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 109, "usage_type": "name"}, {"api_name": "discord.Member", "line_number": 120, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 148, "usage_type": "call"}, {"api_name": "discord.Colour.green", "line_number": 148, "usage_type": "call"}, {"api_name": "discord.Colour", "line_number": 148, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.command", "line_number": 119, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 119, "usage_type": "name"}, {"api_name": "discord.Member", "line_number": 156, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 167, "usage_type": "call"}, {"api_name": "discord.Colour.red", "line_number": 167, "usage_type": "call"}, {"api_name": "discord.Colour", "line_number": 167, "usage_type": "attribute"}, {"api_name": "asyncio.TimeoutError", "line_number": 188, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.command", "line_number": 155, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 155, "usage_type": "name"}, {"api_name": "asyncio.sleep", "line_number": 222, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 227, "usage_type": "call"}, {"api_name": "discord.Colour.red", "line_number": 227, "usage_type": "call"}, {"api_name": "discord.Colour", "line_number": 227, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 228, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 230, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 235, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 238, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 245, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 247, "usage_type": "call"}, {"api_name": "discord.Colour.red", "line_number": 247, "usage_type": "call"}, {"api_name": "discord.Colour", "line_number": 247, "usage_type": "attribute"}, {"api_name": "asyncio.sleep", "line_number": 255, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 261, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 264, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 271, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 273, "usage_type": "call"}, {"api_name": "discord.Colour.red", "line_number": 273, "usage_type": "call"}, {"api_name": "discord.Colour", "line_number": 273, "usage_type": "attribute"}, {"api_name": "asyncio.sleep", "line_number": 281, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 291, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 293, "usage_type": "call"}, {"api_name": "discord.Colour.blue", "line_number": 293, "usage_type": "call"}, {"api_name": "discord.Colour", "line_number": 293, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 297, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 299, "usage_type": "call"}, {"api_name": "discord.Colour.blue", "line_number": 299, "usage_type": "call"}, {"api_name": "discord.Colour", "line_number": 299, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 303, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 315, "usage_type": "call"}, {"api_name": "discord.Colour.red", "line_number": 315, "usage_type": "call"}, {"api_name": "discord.Colour", "line_number": 315, "usage_type": "attribute"}, {"api_name": "asyncio.TimeoutError", "line_number": 326, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 343, "usage_type": "call"}, {"api_name": "discord.Colour.red", "line_number": 343, "usage_type": "call"}, {"api_name": "discord.Colour", "line_number": 343, "usage_type": "attribute"}]} +{"seq_id": "642263217", "text": "#!/usr/bin/env python\n#-*- coding:utf-8 -*-#\n# Copyright (c) 2010-2018 Shanghai Flaginfo Information Incorporated Technology Co.,LTD.\n# All rights reserved.\n#\n\"\"\"\nCreated on 2019-2-19\n\n@author: zhangzhuo\n\nusage : \n1) development\n python offline_recommend.py dev\n2) test\n python offline_recommend.py test\n3) production\n python offline_recommend.py prod\n or\n python offline_recommend.py\n\"\"\"\nimport sys\nimport math\nimport pymongo\nimport time\nimport redis\nimport json\nimport ast\nfrom configparser import ConfigParser\n# import logging\nimport logging.config\nfrom kafka import KafkaConsumer\nfrom kafka import KafkaProducer\nfrom kafka.errors import KafkaError\nimport threading\nfrom time import sleep\n\n# initialize logger\nlogging.config.fileConfig(\"logging.cfg\")\nlogger = logging.getLogger(\"root\")\n\n# constant definitions\nACTIVE_UID_SET_KEY = \"flaginfo:platform:behavior:active:user:ids\"\n\n# global variables\nconsumer_thread_lock = threading.Lock()\nconsumer_thread_item_list = []\nconsumer_thread_welfare_list = []\n\nclass RuntimeContext(object):\n \"\"\" 运行环境\n \"\"\"\n \n def __init__(self):\n \"\"\" 初始化运行环境\n \"\"\"\n # configuration initialization\n config_parser = ConfigParser()\n config_file = self.get_config_file_name()\n config_parser.read(config_file, encoding=\"UTF-8\")\n sections = config_parser.sections()\n\n kafka_section = sections[0]\n news_analysis_topic_name = config_parser.get(kafka_section, \"news_analysis_topic\")\n kafka_bootstrap_servers = config_parser.get(kafka_section, \"kafka_bootstrap_servers\")\n self.news_recommend_topic_name = config_parser.get(kafka_section, \"news_recommend_topic\")\n kafka_group_id = config_parser.get(kafka_section, \"group_id\")\n \n redis_section = sections[1]\n redis_host = config_parser.get(redis_section, \"redis_host\")\n redis_port = config_parser.get(redis_section, \"redis_port\")\n redis_database = config_parser.get(redis_section, \"redis_database\")\n \n mongodb_section = sections[2]\n mongodb_uri = config_parser.get(mongodb_section, \"mongodb_uri\")\n mongodb_dbname = config_parser.get(mongodb_section, \"mongodb_dbname\")\n MONGODB_COLLECTION_TAG_INTEREST = \"tagInterest\"\n \n recommend_section = sections[3]\n self.batch_analysis_item_number = int(config_parser.get(recommend_section, \"batch_analysis_item_number\"))\n self.default_recommend_ratio = float(config_parser.get(recommend_section, \"default_recommend_ratio\"))\n # 用户画像更新周期,单位:秒\n self.user_personas_refresh_time = int(config_parser.get(recommend_section, \"user_personas_refresh_time\"))\n self.max_recommend_time_span = int(config_parser.get(recommend_section, \"max_recommend_time_span\"))\n logger.debug(\"configuration initialization ok\")\n\n # environment initialization\n self.kafka_consumer = KafkaConsumer(news_analysis_topic_name, group_id = kafka_group_id,\n bootstrap_servers = kafka_bootstrap_servers,\n auto_offset_reset = \"earliest\")\n self.kafka_producer = KafkaProducer(bootstrap_servers = kafka_bootstrap_servers)\n\n redis_pool = redis.ConnectionPool(host=redis_host,port=redis_port,db=redis_database)\n self.redis_template = redis.StrictRedis(connection_pool=redis_pool)\n\n mongodb_client = pymongo.MongoClient(mongodb_uri)\n mongodb_database = mongodb_client[mongodb_dbname]\n self.tag_interest_collection = mongodb_database[MONGODB_COLLECTION_TAG_INTEREST]\n\n logger.debug(\"environment initialization ok\")\n \n def get_config_file_name(self):\n \"\"\" 根据命令行参数,获取配置文件名\n \"\"\"\n argv = sys.argv\n config_type = \"prod\" # default configuration type\n if None != argv and len(argv) > 1 :\n config_type = argv[1]\n config_file = config_type + \".cfg\"\n logger.info(\"get_config_file_name() return : \" + config_file)\n return config_file\n\nclass ItemBasedCF(object):\n \"\"\" ItemCF算法\n \"\"\"\n \n def __init__(self, runtime_context):\n # key: 用户ID -> 标签名, value: 用户对标签的兴趣指数\n self.tag_interest = dict()\n # key: 标签名 -> 标签名, value: 标签之间的相关指数\n self.item_cf_dict = dict()\n self.runtime_context = runtime_context\n\n def load_user_tag_interest(self):\n \"\"\" 加载用户画像中的兴趣指数标签\n \"\"\"\n try:\n # 获取Redis活跃用户列表\n active_uid_set = self.runtime_context.redis_template.smembers(ACTIVE_UID_SET_KEY)\n if None == active_uid_set :\n logger.warn(\"no active user\")\n return\n for uid in active_uid_set:\n user_id = str(uid, encoding = \"utf8\")\n self.tag_interest.setdefault(user_id,{})\n result = self.runtime_context.tag_interest_collection.find_one({\"userId\":user_id})\n if result == None or \"interest\" not in list(result.keys()):\n continue\n #self.tag_interest[user_id][''] = 0\n elif not result[\"interest\"]:\n continue\n #self.tag_interest[user_id][''] = 0\n else:\n for interest in result[\"interest\"]:\n self.tag_interest[user_id][interest[\"name\"]] = float(interest[\"value\"])\n except Exception as e:\n logger.error(\"用户画像获取失败\", e)\n \n def item_similarity_for_tags(self):\n \"\"\" 建立标签-标签的共现矩阵\n \"\"\"\n item_dict = dict() #物品-物品的共现矩阵\n item_count = dict() #物品被多少个不同用户购买\n # 统计物品之间关联相同用户的次数\n for user, tags in self.tag_interest.items():\n for i in tags.keys():\n item_count.setdefault(i,0)\n item_count[i] += 1\n item_dict.setdefault(i,{})\n for j in tags.keys():\n if i == j : continue\n item_dict[i].setdefault(j,0)\n item_dict[i][j] += 1\n\n #计算相似度矩阵\n for i,related_items in item_dict.items():\n self.item_cf_dict.setdefault(i,{})\n for j,cij in related_items.items():\n if item_count[i] == 0 or item_count[j] == 0 :\n self.item_cf_dict[i][j] = 0\n else :\n self.item_cf_dict[i][j] = cij / (math.sqrt(item_count[i] * item_count[j]))\n\n return self.item_cf_dict\n\n def get_item_cf_tags(self,user_id,item_cf_top_n=3,recommend_count=10):\n \"\"\" 根据用户的兴趣指数列表,使用ItemCF算法推荐TopN的标签\n \n Args:\n item_cf_top_n : 标签相似度TopN\n recommend_count : 推荐标签数量\n \n Return:\n key : 标签名\n value : 标签推荐指数\n \"\"\"\n rank = dict()\n action_item = self.tag_interest[user_id] # 用户兴趣指数列表\n for tag_name, tag_rating in action_item.items():\n for j_tag_name, j_tag_rating in sorted(self.item_cf_dict[tag_name].items(),key=lambda x:x[1],reverse=True)[0:item_cf_top_n]:\n if j_tag_name in action_item.keys():\n continue\n rank.setdefault(j_tag_name,0)\n rank[j_tag_name] += tag_rating * j_tag_rating\n return dict(sorted(rank.items(),key=lambda x:x[1],reverse=True)[0:recommend_count])\n\n def do_recommend_item_cf(self, news_label):\n \"\"\" 根据用户画像中的兴趣指数标签和备选内容的标签的相似度进行推荐\n \n Args:\n news_label : 备选内容的标签\n \n Return:\n key: 用户ID\n value: 用户的推荐指数\n \"\"\"\n # ItemCF算法根据用户画像算出每个用户的推荐标签列表\n # TODO(shanshu) 此处计算量大,必须优化,考虑增量更新或随用户画像定时更新\n recommend_tags = dict()\n for user_id, tags in self.tag_interest.items():\n recommend_tags[user_id] = self.get_item_cf_tags(user_id)\n\n # 列表转字典。key: 标签名->用户ID, value: 推荐指数\n tag_user_dict = dict()\n for user_id, tags in recommend_tags.items():\n for tag_name in tags.keys():\n tag_user_dict.setdefault(tag_name,{})\n tag_user_dict[tag_name][user_id] = tags[tag_name]\n\n # 对每组标签的用户集合进行排序\n for tag_name, users in tag_user_dict.items():\n tag_user_dict[tag_name] = sorted(tag_user_dict[tag_name].items(),key=lambda x:x[1],reverse=True)\n\n #根据新闻数据标签的值计算用户推荐值\n rec = dict()\n recommend_list = dict()\n for name, value in news_label.items():\n rec.setdefault(name,{})\n if name in tag_user_dict.keys():\n rec[name] = tag_user_dict[name]\n for user_id,tag_rating in rec[name]:\n if user_id in recommend_list.keys():\n recommend_list[user_id] = recommend_list[user_id] + tag_rating*value\n else:\n recommend_list[user_id] = tag_rating*value\n\n recommend_list = sorted(recommend_list.items(),key=lambda x:x[1],reverse=True)\n return recommend_list\n\n def analyze_user_label_rating(self,news_label):\n \"\"\" 计算指定内容标签的用户推荐指数\n \n Args:\n news_label : 指定内容的标签\n\n Return:\n key: 用户ID\n value: 用户的推荐指数\n \"\"\"\n # key: 标签名->用户名, value: 用户对标签的兴趣指数\n item_user_dict = dict()\n for user_id, tags in self.tag_interest.items(): # TODO(shanshu) 此处对每一条新闻都要计算,计算量过大,必须优化\n for name in tags.keys():\n item_user_dict.setdefault(name,{})\n item_user_dict[name][user_id] = tags[name]\n #对每���标签的用户集合进行排序\n for items,user_id in item_user_dict.items():\n item_user_dict[items] = sorted(item_user_dict[items].items(),key=lambda x:x[1],reverse=True)\n #根据新闻数据标签的值计算用户推荐值\n rec = dict()\n user_rating_list = dict()\n for name, value in news_label.items():\n rec.setdefault(name,{})\n if name in item_user_dict.keys():\n rec[name] = item_user_dict[name]\n for user_id,tag_rating in rec[name]:\n if user_id in user_rating_list.keys():\n user_rating_list[user_id] = user_rating_list[user_id] + tag_rating*value\n else:\n user_rating_list[user_id] = tag_rating*value\n\n user_rating_list = sorted(user_rating_list.items(),key=lambda x:x[1],reverse=True)\n logger.debug(user_rating_list)\n return user_rating_list\n\n def content_base_recommend(self,news_total):\n \"\"\" 计算指定内容标签的用户推荐指数\n \n Args:\n news_total : 内容列表\n \n Return:\n key: 用户ID->内容ID\n value: 推荐指数\n \"\"\"\n news_count = len(news_total)\n recommend_count = round(news_count * self.runtime_context.default_recommend_ratio)\n # TODO(shanshu) 根据用户画像,每个用户采用不同的比率推荐\n if recommend_count < 1 :\n recommend_count = 1\n try:\n recommemd_list = list()\n for user_id, user_interest_tags in self.tag_interest.items():\n # key: 用户ID->内容ID, value: 推荐指数\n user_item_rating = dict()\n user_item_rating.setdefault(user_id,{})\n for news in news_total:\n user_item_rating[user_id][news[\"id\"]] = 0\n for tag_name, tag_value in news[\"tags\"].items():\n for user_interest_tag_name, user_interest_rating in user_interest_tags.items():\n if tag_name == user_interest_tag_name:\n user_item_rating[user_id][news[\"id\"]] = user_item_rating[user_id][news[\"id\"]] + user_interest_rating * tag_value\n user_item_rating[user_id] = dict(sorted(user_item_rating[user_id].items(),key=lambda x:x[1],reverse=True)[0:recommend_count])\n recommemd_list.append(user_item_rating)\n logger.debug(\"recommend for user \" + user_id + \" ,count:\" + str(len(user_item_rating[user_id])))\n return recommemd_list\n except Exception as e:\n logger.error(\"content_base_recommend()发生异常\", e)\n def search_news_category(self,itemId,news_total):\n \"\"\" 计算指定内容ID的分类名称\n \n Args:\n news_total : 内容列表\n itemId:内容ID\n Return:\n category_name:分类名称\n \"\"\"\n try:\n category_name = \" \"\n for news in news_total:\n if itemId == news['id']:\n category_name = news['category']+ news['subCategory']\n app_id = news['appId']\n sp_id = news['spId']\n itemType = news['itemType']\n return category_name,app_id,sp_id,itemType\n except Exception as e:\n logger.error (\"search_news_category erro:\", e)\n return \" \"\n\ndef main_old():\n runtime_context = RuntimeContext()\n item_cf = ItemBasedCF(runtime_context)\n last_user_personas_refresh_time = 0\n for message_obj in runtime_context.kafka_consumer:\n try:\n # 解析消息内容\n message = message_obj.value\n message = message.decode(\"utf-8\")\n #print (s1)#########\n news_obj = ast.literal_eval(message.strip())\n news_id = news_obj[\"id\"]\n news_label = news_obj[\"tags\"]\n # 定期刷新用户画像\n time_passed = time.time() - last_user_personas_refresh_time\n logger.debug(\"[user_personas_refresh]time_passed : \" + str(time_passed))\n if time_passed > runtime_context.user_personas_refresh_time:\n logger.debug(\"refresh user personas begin\")\n item_cf.load_user_tag_interest()\n last_user_personas_refresh_time = time.time()\n # 基于标签的推荐\n RecommandUserLabelList = item_cf.analyze_user_label_rating(news_label)\n userLabelListToDict = dict()\n for i,j in RecommandUserLabelList:\n userLabelListToDict[i] = j\n userLabelList = dict()\n userLabelList[\"id\"] = news_id\n userLabelList[\"type\"] = 1\n userLabelList[\"recommendationRate\"] = userLabelListToDict\n json_message = json.dumps(userLabelList) # 将dict类型的数据转成str\n runtime_context.kafka_producer.send(runtime_context.news_recommend_topic_name, json_message.encode(\"utf-8\"))\n runtime_context.kafka_producer.flush()\n # ItemCF推荐\n # TODO(shanshu) 共现矩阵计算量大,需要优化为增量式更新,避免每次大量计算\n item_cf.item_similarity_for_tags()\n # TODO(shanshu) 每次只计算一条内容,重复计算太多,需要优化为批量计算\n item_cf_recommend_list = item_cf.do_recommend_item_cf(news_label)\n item_cf_recommend_dict = dict()\n for user_id, rating in item_cf_recommend_list:\n item_cf_recommend_dict[user_id] = rating\n recommend_message = dict()\n recommend_message[\"id\"] = news_id\n recommend_message[\"type\"] = 2\n recommend_message[\"recommendationRate\"] = item_cf_recommend_dict\n json_message = json.dumps(recommend_message) # 将dict类型的数据转成str\n runtime_context.kafka_producer.send(runtime_context.news_recommend_topic_name, json_message.encode(\"utf-8\"))\n runtime_context.kafka_producer.flush()\n \n except Exception as e:\n logger.error(\"main()发生异常\", e)\n\nclass KafkaConsumerThread(threading.Thread):\n \"\"\" Kafka消费线程\n \"\"\"\n\n def __init__(self, threadID, name, runtime_context):\n \"\"\" 构造函数\n \n Args :\n threadID : 线程编号(数字)\n name : 线程名\n runtime_context : 运行环境(RuntimeContext对象)\n \"\"\"\n threading.Thread.__init__(self)\n self.threadID = threadID\n self.name = name\n self.runtime_context = runtime_context\n\n def run(self):\n logger.info(self.name + \" started\")\n self.exit_flag = False\n self.sleep_duration = 10; # TODO 延迟时间改为可配置\n while not self.exit_flag:\n try:\n for message_obj in self.runtime_context.kafka_consumer:\n # 解析消息内容\n consumer_thread_lock.acquire()\n try:\n message = message_obj.value\n message_utf8 = message.decode(\"UTF-8\")\n news_obj = ast.literal_eval(message_utf8.strip())\n if news_obj[\"itemType\"] == 3:\n consumer_thread_welfare_list.append(news_obj)\n else:\n consumer_thread_item_list.append(news_obj)\n news_num = len(consumer_thread_item_list)\n finally:\n consumer_thread_lock.release()\n # 防止消息过度堆积,延迟处理\n if news_num >= self.runtime_context.batch_analysis_item_number:\n sleep(self.sleep_duration)\n except Exception as e:\n logger.error(\"KafkaConsumerThread发生异常\", e)\n sleep(self.sleep_duration)\n\ndef main():\n runtime_context = RuntimeContext()\n item_cf = ItemBasedCF(runtime_context)\n news_num = 0\n last_user_personas_refresh_time = 0\n last_recommend_time = 0\n # 启动Kafka消费线程\n kafka_consumer_thread = KafkaConsumerThread(1, \"KafkaConsumerThread-1\", runtime_context)\n kafka_consumer_thread.start()\n logger.info(\"offline recommend service started\")\n exit_flag = False\n while not exit_flag:\n try:\n consumer_thread_lock.acquire()\n \n news_num = len(consumer_thread_item_list)\n if news_num > 0:\n # 批量推荐\n # 达到一定时间后,或者数量到达一定上限后,触发推荐\n time_passed = time.time() - last_recommend_time\n if (time_passed >= runtime_context.max_recommend_time_span and not last_recommend_time == 0) or news_num >= runtime_context.batch_analysis_item_number:\n last_recommend_time = time.time()\n # 定期刷新用户画像\n time_passed = time.time() - last_user_personas_refresh_time\n logger.debug(\"[user_personas_refresh]time_passed : \" + str(time_passed))\n if time_passed > runtime_context.user_personas_refresh_time:\n logger.debug(\"refresh user personas begin\")\n item_cf.load_user_tag_interest()\n logger.debug(\"refresh user personas end\")\n last_user_personas_refresh_time = time.time()\n # 执行推荐算法\n #print (consumer_thread_item_list)\n recommend_list = item_cf.content_base_recommend(consumer_thread_item_list)\n # 发布推荐消息\n for sub_list in recommend_list:\n recommend_message = {}\n recommemd_message_items = []\n for user_id, item_list in sub_list.items():\n recommend_message[\"version\"] = 1\n recommend_message[\"userId\"] = user_id\n #处理非福利类数据\n for item_id, rating in item_list.items():\n item_dict = {}\n item_dict[\"itemId\"] = item_id\n item_dict[\"rating\"] = rating\n item_dict[\"categoryName\"],item_dict[\"appId\"],item_dict[\"spId\"],item_dict[\"itemType\"] = item_cf.search_news_category(item_id,consumer_thread_item_list)\n recommemd_message_items.append(item_dict)\n #处理福利类数据\n if consumer_thread_welfare_list:\n for item in consumer_thread_welfare_list:\n item_dict = {}\n item_dict[\"itemId\"] = item[\"id\"]\n item_dict[\"rating\"] = 1\n item_dict[\"categoryName\"] = item[\"category\"]\n item_dict[\"appId\"] = item[\"appId\"]\n item_dict[\"spId\"] = item[\"spId\"]\n item_dict[\"itemType\"] = item[\"itemType\"]\n recommemd_message_items.append(item_dict) \n #recommend_message[\"itemType\"] = 0\n recommend_message[\"list\"] = recommemd_message_items\n json_message = json.dumps(recommend_message,ensure_ascii=False).encode(\"utf-8\") # 将dict类型的数据转成str\n logger.info (recommend_message)\n try:\n runtime_context.kafka_producer.send(runtime_context.news_recommend_topic_name, json_message)\n runtime_context.kafka_producer.flush()\n logger.debug(\"publish recommend message ok. user: \" + user_id + \", item count: \" + str(len(recommemd_message_items)))\n except KafkaError as e:\n logger.error(\"kafka send error:\",e) \n #清空福利类数据集合列表\n consumer_thread_welfare_list.clear()\n news_num = 0\n consumer_thread_item_list.clear()\n except Exception as e:\n logger.error(\"main()发生异常\", e)\n finally:\n consumer_thread_lock.release()\n # 每次完成一次推荐后,降低CPU负载\n sleep_duration = 5; # TODO 延迟时间改为可配置\n sleep(sleep_duration)\n\nif __name__ == \"__main__\":\n main()\n", "sub_path": "offline-recommend-analysis-service/tags/v1.2.1-20190219-1/offline_recommend.py", "file_name": "offline_recommend.py", "file_ext": "py", "file_size_in_byte": 23180, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "logging.config.config.fileConfig", "line_number": 38, "usage_type": "call"}, {"api_name": "logging.config.config", "line_number": 38, "usage_type": "attribute"}, {"api_name": "logging.config", "line_number": 38, "usage_type": "name"}, {"api_name": "logging.config.getLogger", "line_number": 39, "usage_type": "call"}, {"api_name": "logging.config", "line_number": 39, "usage_type": "name"}, {"api_name": "threading.Lock", "line_number": 45, "usage_type": "call"}, {"api_name": "configparser.ConfigParser", "line_number": 57, "usage_type": "call"}, {"api_name": "kafka.KafkaConsumer", "line_number": 87, "usage_type": "call"}, {"api_name": "kafka.KafkaProducer", "line_number": 90, "usage_type": "call"}, {"api_name": "redis.ConnectionPool", "line_number": 92, "usage_type": "call"}, {"api_name": "redis.StrictRedis", "line_number": 93, "usage_type": "call"}, {"api_name": "pymongo.MongoClient", "line_number": 95, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 104, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 171, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 340, "usage_type": "call"}, {"api_name": "time.time", "line_number": 344, "usage_type": "call"}, {"api_name": "time.time", "line_number": 349, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 359, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 374, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 381, "usage_type": "attribute"}, {"api_name": "threading.Thread.__init__", "line_number": 393, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 393, "usage_type": "attribute"}, {"api_name": "ast.literal_eval", "line_number": 410, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 420, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 423, "usage_type": "call"}, {"api_name": "time.time", "line_number": 444, "usage_type": "call"}, {"api_name": "time.time", "line_number": 446, "usage_type": "call"}, {"api_name": "time.time", "line_number": 448, "usage_type": "call"}, {"api_name": "time.time", "line_number": 454, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 485, "usage_type": "call"}, {"api_name": "kafka.errors.KafkaError", "line_number": 491, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 503, "usage_type": "call"}]} +{"seq_id": "48325500", "text": "from django.conf import settings\nfrom django.conf.urls.static import static\nfrom django.contrib import admin\nfrom django.urls import include, path\n\nurlpatterns = [\n path('about/', include('about.urls')),\n path(\"admin/\", admin.site.urls),\n path(\"api/\", include(\"api.urls\")),\n path(\"auth/\", include(\"users.urls\")),\n path(\"auth/\", include(\"django.contrib.auth.urls\")),\n path(\"\", include(\"recepies.urls\"), name='index'),\n ]\n\nurlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)\nurlpatterns += static(settings.STATIC_URL, document_root=settings.STATIC_ROOT)\n", "sub_path": "foodgram/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 598, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 7, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 8, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 8, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 9, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 10, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 11, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.static.static", "line_number": 15, "usage_type": "call"}, {"api_name": "django.conf.settings.MEDIA_URL", "line_number": 15, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 15, "usage_type": "name"}, {"api_name": "django.conf.settings.MEDIA_ROOT", "line_number": 15, "usage_type": "attribute"}, {"api_name": "django.conf.urls.static.static", "line_number": 16, "usage_type": "call"}, {"api_name": "django.conf.settings.STATIC_URL", "line_number": 16, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 16, "usage_type": "name"}, {"api_name": "django.conf.settings.STATIC_ROOT", "line_number": 16, "usage_type": "attribute"}]} +{"seq_id": "191596939", "text": "import json\n\nfrom django.core import serializers\nfrom django.http import HttpResponse\nfrom django.http import JsonResponse\nfrom django.shortcuts import get_list_or_404, get_object_or_404\nfrom django.views.decorators.csrf import csrf_exempt\n\nfrom api.models import Collection, PieceLike, Rank\nfrom api.models import Piece, Category, Artist, Comments\n\n\n###########################################\n# Resource for operations with Piece class\n###########################################\n\n@csrf_exempt\ndef pieces_list(request):\n pieces_list = Piece.objects.all()\n pieces_result = []\n for piece in pieces_list:\n piece.artist_name = piece.artist.name_artistic\n pieces_result.append(piece)\n return HttpResponse(serializers.serialize(\"json\", pieces_result))\n\n\n@csrf_exempt\ndef collection_by_artist(request, artist_name):\n collection = get_list_or_404(Collection.objects.filter(name=artist_name))\n return HttpResponse(serializers.serialize(\"json\", collection))\n\n\n@csrf_exempt\ndef piece_by_id(request, piece_id):\n piece_result = Piece.objects.get(pk=piece_id)\n artist = Artist.objects.get(pk=piece_result.artist.id)\n piece_result.artist_name = artist.name\n piece = []\n piece.append(piece_result)\n return HttpResponse(serializers.serialize(\"json\", piece))\n\n\n@csrf_exempt\ndef piece_by_category(request, category_id):\n category = Category.objects.get(pk=category_id)\n piece = get_list_or_404(Piece.objects.filter(category=category))\n return HttpResponse(serializers.serialize(\"json\", piece))\n\n\ndef update_from_request(selected_piece, request):\n if request.name is not None:\n selected_piece.name = request.name\n if request.url is not None:\n selected_piece.url = request.url\n if request.image_cover is not None:\n selected_piece.image_cover = request.image_cover\n if request.duration is not None:\n selected_piece.duration = request.duration\n if request.category is not None:\n cat = get_object_or_404(Category, pk=request.category)\n selected_piece.category = cat\n if request.lyrics is not None:\n selected_piece.lyrics = request.lyrics\n return selected_piece\n\n\n@csrf_exempt\ndef update_piece(request):\n if request.method == \"POST\":\n jsonPiece = json.loads(request.body)\n piece_id = jsonPiece['body']['pk']\n pieces = get_list_or_404(Piece.objects.filter(pk=piece_id))\n if len(pieces) == 0:\n return JsonResponse({\"mensaje\": \"There are no pieces with id\" + piece_id})\n else:\n request = PieceRequest(jsonPiece)\n selected_piece = pieces[0]\n selected_piece = update_from_request(selected_piece, request)\n selected_piece.save()\n return JsonResponse({\"mensaje\": \"successfully updated\"})\n\n\n@csrf_exempt\ndef add_piece(request):\n if request.method == 'POST':\n jsonPiece = json.loads(request.body)\n new_piece = Piece(\n name=jsonPiece['body']['name'],\n url=jsonPiece['body']['sound'],\n image_cover=jsonPiece['body']['cover'],\n duration=jsonPiece['body']['duration'],\n category=get_object_or_404(Category.objects.filter(id=jsonPiece['body']['category'])),\n artist=Artist.objects.get(userId__username=jsonPiece['body']['artist']),\n );\n new_piece.save();\n return HttpResponse(serializers.serialize(\"json\", [new_piece]))\n\n\n@csrf_exempt\ndef like_piece(request, piece_id):\n if request.method == 'POST':\n json_body = json.loads(request.body)\n username = json_body['username']\n piece = get_object_or_404(Piece, pk=piece_id)\n new_like = PieceLike(piece=piece, username=username)\n new_like.save()\n return JsonResponse({\"mensaje\": \"successfully liked\"})\n\n\n@csrf_exempt\ndef unlike_piece(request, piece_id):\n if request.method == 'POST':\n json_body = json.loads(request.body)\n username = json_body['username']\n piece = get_object_or_404(Piece, pk=piece_id)\n like = PieceLike.objects.filter(piece=piece, username=username)\n like.delete()\n return JsonResponse({\"mensaje\": \"successfully unliked\"})\n\n\n@csrf_exempt\ndef is_liked_piece_by_username(request, piece_id):\n if request.method == 'POST':\n json_body = json.loads(request.body)\n username = json_body['username']\n piece = get_object_or_404(Piece, pk=piece_id)\n like = PieceLike.objects.filter(piece=piece, username=username)\n if len(like) > 0:\n return JsonResponse({\"liked\": True})\n else:\n return JsonResponse({\"liked\": False})\n\n\n@csrf_exempt\ndef likes_by_piece(request, piece_id):\n if request.method == 'GET':\n piece = get_object_or_404(Piece, pk=piece_id)\n likes = PieceLike.objects.filter(piece=piece)\n if likes is not None:\n return JsonResponse({\"likes\": len(likes)})\n else:\n return JsonResponse({\"likes\": 0})\n\n\n@csrf_exempt\ndef get_most_voted(request):\n if request.method == 'GET':\n pieceLikes = PieceLike.objects.all();\n if pieceLikes is not None:\n validated_pieces = []\n answer = []\n for pl in pieceLikes:\n piece = pl.piece\n if piece.name not in validated_pieces:\n validated_pieces.append(piece.name)\n rank = Rank(piece_name=piece.name, likes_number=len(PieceLike.objects.filter(piece=piece)))\n answer.append(rank)\n return HttpResponse(serializers.serialize(\"json\", answer))\n\n\nclass PieceRequest:\n def __init__(self, json_piece):\n self.name = json_piece['body']['fields']['name']\n self.url = json_piece['body']['fields']['url']\n self.image_cover = json_piece['body']['fields']['image_cover']\n self.duration = json_piece['body']['fields']['duration']\n self.category = json_piece['body']['fields']['category']\n self.lyrics = json_piece['body']['fields']['lyrics']\n\n\n@csrf_exempt\ndef add_comment(request, piece_id):\n if request.method == 'POST':\n json_body = json.loads(request.body)\n email = json_body['body']['email']\n text = json_body['body']['text']\n piece = get_object_or_404(Piece, pk=piece_id)\n new_comment = Comments(piece=piece, email=email, text=text)\n new_comment.save()\n return JsonResponse({\"mensaje\": \"ok\"})\n\n\n@csrf_exempt\ndef comments_piece(request, piece_id):\n if request.method == 'GET':\n piece_obj = Piece.objects.filter(pk=piece_id)\n comments = Comments.objects.filter(piece=piece_obj)\n if comments is not None:\n return HttpResponse(serializers.serialize(\"json\", comments))\n else:\n return JsonResponse({\"comments\": 'No comments'})\n", "sub_path": "api/resources/pieces_resource.py", "file_name": "pieces_resource.py", "file_ext": "py", "file_size_in_byte": 6802, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "api.models.Piece.objects.all", "line_number": 19, "usage_type": "call"}, {"api_name": "api.models.Piece.objects", "line_number": 19, "usage_type": "attribute"}, {"api_name": "api.models.Piece", "line_number": 19, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 24, "usage_type": "call"}, {"api_name": "django.core.serializers.serialize", "line_number": 24, "usage_type": "call"}, {"api_name": "django.core.serializers", "line_number": 24, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 17, "usage_type": "name"}, {"api_name": "django.shortcuts.get_list_or_404", "line_number": 29, "usage_type": "call"}, {"api_name": "api.models.Collection.objects.filter", "line_number": 29, "usage_type": "call"}, {"api_name": "api.models.Collection.objects", "line_number": 29, "usage_type": "attribute"}, {"api_name": "api.models.Collection", "line_number": 29, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 30, "usage_type": "call"}, {"api_name": "django.core.serializers.serialize", "line_number": 30, "usage_type": "call"}, {"api_name": "django.core.serializers", "line_number": 30, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 27, "usage_type": "name"}, {"api_name": "api.models.Piece.objects.get", "line_number": 35, "usage_type": "call"}, {"api_name": "api.models.Piece.objects", "line_number": 35, "usage_type": "attribute"}, {"api_name": "api.models.Piece", "line_number": 35, "usage_type": "name"}, {"api_name": "api.models.Artist.objects.get", "line_number": 36, "usage_type": "call"}, {"api_name": "api.models.Artist.objects", "line_number": 36, "usage_type": "attribute"}, {"api_name": "api.models.Artist", "line_number": 36, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 40, "usage_type": "call"}, {"api_name": "django.core.serializers.serialize", "line_number": 40, "usage_type": "call"}, {"api_name": "django.core.serializers", "line_number": 40, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 33, "usage_type": "name"}, {"api_name": "api.models.Category.objects.get", "line_number": 45, "usage_type": "call"}, {"api_name": "api.models.Category.objects", "line_number": 45, "usage_type": "attribute"}, {"api_name": "api.models.Category", "line_number": 45, "usage_type": "name"}, {"api_name": "django.shortcuts.get_list_or_404", "line_number": 46, "usage_type": "call"}, {"api_name": "api.models.Piece.objects.filter", "line_number": 46, "usage_type": "call"}, {"api_name": "api.models.Piece.objects", "line_number": 46, "usage_type": "attribute"}, {"api_name": "api.models.Piece", "line_number": 46, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 47, "usage_type": "call"}, {"api_name": "django.core.serializers.serialize", "line_number": 47, "usage_type": "call"}, {"api_name": "django.core.serializers", "line_number": 47, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 43, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 60, "usage_type": "call"}, {"api_name": "api.models.Category", "line_number": 60, "usage_type": "argument"}, {"api_name": "json.loads", "line_number": 70, "usage_type": "call"}, {"api_name": "django.shortcuts.get_list_or_404", "line_number": 72, "usage_type": "call"}, {"api_name": "api.models.Piece.objects.filter", "line_number": 72, "usage_type": "call"}, {"api_name": "api.models.Piece.objects", "line_number": 72, "usage_type": "attribute"}, {"api_name": "api.models.Piece", "line_number": 72, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 74, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 80, "usage_type": "call"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 67, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 86, "usage_type": "call"}, {"api_name": "api.models.Piece", "line_number": 87, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 92, "usage_type": "call"}, {"api_name": "api.models.Category.objects.filter", "line_number": 92, "usage_type": "call"}, {"api_name": "api.models.Category.objects", "line_number": 92, "usage_type": "attribute"}, {"api_name": "api.models.Category", "line_number": 92, "usage_type": "name"}, {"api_name": "api.models.Artist.objects.get", "line_number": 93, "usage_type": "call"}, {"api_name": "api.models.Artist.objects", "line_number": 93, "usage_type": "attribute"}, {"api_name": "api.models.Artist", "line_number": 93, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 96, "usage_type": "call"}, {"api_name": "django.core.serializers.serialize", "line_number": 96, "usage_type": "call"}, {"api_name": "django.core.serializers", "line_number": 96, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 83, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 102, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 104, "usage_type": "call"}, {"api_name": "api.models.Piece", "line_number": 104, "usage_type": "argument"}, {"api_name": "api.models.PieceLike", "line_number": 105, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 107, "usage_type": "call"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 99, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 113, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 115, "usage_type": "call"}, {"api_name": "api.models.Piece", "line_number": 115, "usage_type": "argument"}, {"api_name": "api.models.PieceLike.objects.filter", "line_number": 116, "usage_type": "call"}, {"api_name": "api.models.PieceLike.objects", "line_number": 116, "usage_type": "attribute"}, {"api_name": "api.models.PieceLike", "line_number": 116, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 118, "usage_type": "call"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 110, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 124, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 126, "usage_type": "call"}, {"api_name": "api.models.Piece", "line_number": 126, "usage_type": "argument"}, {"api_name": "api.models.PieceLike.objects.filter", "line_number": 127, "usage_type": "call"}, {"api_name": "api.models.PieceLike.objects", "line_number": 127, "usage_type": "attribute"}, {"api_name": "api.models.PieceLike", "line_number": 127, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 129, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 131, "usage_type": "call"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 121, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 137, "usage_type": "call"}, {"api_name": "api.models.Piece", "line_number": 137, "usage_type": "argument"}, {"api_name": "api.models.PieceLike.objects.filter", "line_number": 138, "usage_type": "call"}, {"api_name": "api.models.PieceLike.objects", "line_number": 138, "usage_type": "attribute"}, {"api_name": "api.models.PieceLike", "line_number": 138, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 140, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 142, "usage_type": "call"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 134, "usage_type": "name"}, {"api_name": "api.models.PieceLike.objects.all", "line_number": 148, "usage_type": "call"}, {"api_name": "api.models.PieceLike.objects", "line_number": 148, "usage_type": "attribute"}, {"api_name": "api.models.PieceLike", "line_number": 148, "usage_type": "name"}, {"api_name": "api.models.Rank", "line_number": 156, "usage_type": "call"}, {"api_name": "api.models.PieceLike.objects.filter", "line_number": 156, "usage_type": "call"}, {"api_name": "api.models.PieceLike.objects", "line_number": 156, "usage_type": "attribute"}, {"api_name": "api.models.PieceLike", "line_number": 156, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 158, "usage_type": "call"}, {"api_name": "django.core.serializers.serialize", "line_number": 158, "usage_type": "call"}, {"api_name": "django.core.serializers", "line_number": 158, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 145, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 174, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 177, "usage_type": "call"}, {"api_name": "api.models.Piece", "line_number": 177, "usage_type": "argument"}, {"api_name": "api.models.Comments", "line_number": 178, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 180, "usage_type": "call"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 171, "usage_type": "name"}, {"api_name": "api.models.Piece.objects.filter", "line_number": 186, "usage_type": "call"}, {"api_name": "api.models.Piece.objects", "line_number": 186, "usage_type": "attribute"}, {"api_name": "api.models.Piece", "line_number": 186, "usage_type": "name"}, {"api_name": "api.models.Comments.objects.filter", "line_number": 187, "usage_type": "call"}, {"api_name": "api.models.Comments.objects", "line_number": 187, "usage_type": "attribute"}, {"api_name": "api.models.Comments", "line_number": 187, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 189, "usage_type": "call"}, {"api_name": "django.core.serializers.serialize", "line_number": 189, "usage_type": "call"}, {"api_name": "django.core.serializers", "line_number": 189, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 191, "usage_type": "call"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 183, "usage_type": "name"}]} +{"seq_id": "321455522", "text": "\"\"\"Detects text in the file.\"\"\"\nfrom google.cloud import vision\nimport io\nclient = vision.ImageAnnotatorClient()\npath = './bubcar/resources/test5.jpg'\n\nwith io.open(path, 'rb') as image_file:\n content = image_file.read()\n\nimage = vision.types.Image(content=content)\n\nprice_candidate = []\ncard_number_candidate = []\ndate_candidate = []\n\nresponse = client.text_detection(image=image)\ntexts = response.text_annotations\nprint('Texts:')\n\nfor text in texts:\n content = text.description\n content = content.replace(',','')\n print('\\n\"{}\"'.format(content))\n\n\n\n\n\nif response.error.message:\n raise Exception(\n '{}\\nFor more info on error messages, check: '\n 'https://cloud.google.com/apis/design/errors'.format(\n response.error.message))", "sub_path": "image detection.py", "file_name": "image detection.py", "file_ext": "py", "file_size_in_byte": 766, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "google.cloud.vision.ImageAnnotatorClient", "line_number": 4, "usage_type": "call"}, {"api_name": "google.cloud.vision", "line_number": 4, "usage_type": "name"}, {"api_name": "io.open", "line_number": 7, "usage_type": "call"}, {"api_name": "google.cloud.vision.types.Image", "line_number": 10, "usage_type": "call"}, {"api_name": "google.cloud.vision.types", "line_number": 10, "usage_type": "attribute"}, {"api_name": "google.cloud.vision", "line_number": 10, "usage_type": "name"}]} +{"seq_id": "171615873", "text": "from flask import Flask, url_for, request, Markup, render_template, flash\nimport os\nimport json\n\napp = Flask(__name__)\n\nwith open('county_demographics.json') as demographics_data:\n counties = json.load(demographics_data)\n\ndef get_state_options():\n listOfStates = []\n for county in counties:\n if county['State'] not in listOfStates:\n listOfStates.append(county['State'])\n options = \"\"\n for state in listOfStates:\n options = options + Markup(\"\")\n return options\n \ndef get_state_facts(state):\n income = 0\n my_county = counties[0]['County']\n \n fact = \"\"\n for county in counties:\n if state == county['State']:\n income = county['Income']['Median Houseold Income']\n my_county = county['County']\n fact = fact + Markup(\"

\" + \"Median household income for \" + my_county + \" in \" + state + \" is \" + \"$\" + str(income) + \"

\")\n return fact\n \n@app.route(\"/\")\ndef render_main():\n return render_template('index.html', options = get_state_options())\n\n@app.route(\"/response\")\ndef render_response():\n returnState = request.args['returnState']\n return render_template('index.html', stateFact = get_state_facts(returnState), options = get_state_options())\n \nif __name__==\"__main__\":\n app.run(debug=False)\n", "sub_path": "webapp.py", "file_name": "webapp.py", "file_ext": "py", "file_size_in_byte": 1355, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "flask.Flask", "line_number": 5, "usage_type": "call"}, {"api_name": "json.load", "line_number": 8, "usage_type": "call"}, {"api_name": "flask.Markup", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.Markup", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 34, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 38, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 38, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 39, "usage_type": "call"}]} +{"seq_id": "364516068", "text": "#!/usr/bin/python\n# -*- coding: utf8 -*-\n\n\"\"\"\n author:nango\n\nupdate_time: 2015/7/21--22:05\n\napis:\n /campus/login\n /campus/logout\n\"\"\"\n\nimport route\nimport web\nimport random\nimport time\nfrom database import *\nfrom output import *\nfrom encrypt import *\n\n\n@route.route('/campus/food/list')\nclass FoodList:\n def POST(self):\n input = web.input(campus_id = None, food_id = None)\n\n if input.campus_id == None:\n return output(110)\n\n try:\n input.campus_id = int(input.campus_id)\n if input.food_id != None:\n input.food_id = int(input.food_id)\n except:\n output(111)\n\n db = getDb()\n # 查看campus_id 是否存在\n results = db.select('campus', vars={'campus_id' : input.campus_id},\n where=\"campus_id=$campus_id\", what=\"campus_id\")\n\n if len(results) != 1:\n return output(460)\n\n if input.food_id == None:\n results = db.select('food', vars={'campus_id': input.campus_id},\n where=\"campus_id=$campus_id\")\n else:\n results = db.select('food', vars={'food_id': input.food_id, 'campus_id': input.campus_id},\n where=\"food_id=$food_id and campus_id=$campus_id\")\n if len(results) == 0:\n return output(463)\n\n resul = []\n for i in results:\n is_served = True\n if i.is_served == 'no':\n is_served = False\n\n is_sold_out = True\n if i.is_sold_out == 'no':\n is_sold_out = False\n resul.append(\n {\"id\": i.food_id, \"name\": i.food_name, \"price\": i.food_price,\n \"desc\": i.food_desc, \"img_url\": i.food_img_url, \"is_sold_out\": is_sold_out,\n \"is_served\": is_served})\n return output(200, resul)", "sub_path": "sites/campus/food/list.py", "file_name": "list.py", "file_ext": "py", "file_size_in_byte": 1890, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "web.input", "line_number": 26, "usage_type": "call"}, {"api_name": "route.route", "line_number": 23, "usage_type": "call"}]} +{"seq_id": "527081176", "text": "import discord\nfrom discord.ext import commands\nfrom .utils import checks\nfrom .lib import chatterbotapi as cb\nimport time\n\nfactory = cb.ChatterBotFactory()\n\n\nclass ConversationInfo:\n def __init__(self, session, channel):\n self.session = session\n self.channel = channel\n self.last_message_time = time.time()\n\n\nclass Conversation:\n \"\"\"Because sometimes DankBot gets lonely.\"\"\"\n\n def __init__(self, bot):\n self.bot = bot\n self.conversations = {}\n\n @commands.command(pass_context=True, name='talk')\n async def start_convo(self, ctx: commands.Context):\n \"\"\"Starts talking with DankBot.\"\"\"\n if ctx.message.author.id in self.conversations:\n return\n await self.bot.say('Starting a conversation with {}.'.format(ctx.message.author.mention))\n chatter_bot = factory.create(cb.ChatterBotType.CLEVERBOT)\n session = chatter_bot.create_session()\n self.conversations[ctx.message.author.id] = ConversationInfo(session, ctx.message.channel)\n\n @commands.command(pass_context=True, name='shutup')\n async def end_convo(self, ctx: commands.Context):\n \"\"\"Tells DankBot to shut up and stop talking.\"\"\"\n if ctx.message.author.id not in self.conversations:\n return\n await self.bot.say('Ending a conversation with {}'.format(ctx.message.author.mention))\n self.conversations.pop(ctx.message.author.id)\n\n @commands.command(name='stopall')\n @checks.is_owner()\n async def stop_all(self):\n \"\"\"Brutally cuts out DankBot's voice box.\"\"\"\n await self.bot.say('Ending all conversations.')\n self.conversations.clear()\n\n async def on_message(self, message: discord.Message):\n if message.content.startswith(self.bot.command_prefix):\n return\n author_id = message.author.id\n if author_id in self.conversations and self.conversations[author_id].channel == message.channel:\n convo = self.conversations[author_id]\n if convo.last_message_time + 120 < time.time():\n self.conversations.pop(author_id)\n else:\n convo.last_message_time = time.time()\n await self.bot.send_message(message.channel, convo.session.think(message.content))\n\n\ndef setup(bot):\n bot.add_cog(Conversation(bot))\n", "sub_path": "cogs/conversation.py", "file_name": "conversation.py", "file_ext": "py", "file_size_in_byte": 2328, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "lib.chatterbotapi.ChatterBotFactory", "line_number": 7, "usage_type": "call"}, {"api_name": "lib.chatterbotapi", "line_number": 7, "usage_type": "name"}, {"api_name": "time.time", "line_number": 14, "usage_type": "call"}, {"api_name": "discord.ext.commands.Context", "line_number": 25, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 25, "usage_type": "name"}, {"api_name": "lib.chatterbotapi.ChatterBotType", "line_number": 30, "usage_type": "attribute"}, {"api_name": "lib.chatterbotapi", "line_number": 30, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 24, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 24, "usage_type": "name"}, {"api_name": "discord.ext.commands.Context", "line_number": 35, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 35, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 34, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 34, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 42, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 42, "usage_type": "name"}, {"api_name": "utils.checks.is_owner", "line_number": 43, "usage_type": "call"}, {"api_name": "utils.checks", "line_number": 43, "usage_type": "name"}, {"api_name": "discord.Message", "line_number": 49, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 55, "usage_type": "call"}, {"api_name": "time.time", "line_number": 58, "usage_type": "call"}]} +{"seq_id": "158074712", "text": "# -*- coding: utf-8 -*-\n\n# Define your item pipelines here\n#\n# Don't forget to add your pipeline to the ITEM_PIPELINES setting\n# See: http://doc.scrapy.org/en/latest/topics/item-pipeline.html\n\nimport os\nimport sys\nimport pymongo\n\nreload(sys)\nsys.setdefaultencoding(\"utf-8\")\n\n\n# file version\nclass FilePipeline(object):\n def process_item(self, item, spider):\n title = item['title']\n context = item['context']\n file_dir = 'NetEaseNews'\n if not os.path.exists(file_dir):\n os.mkdir(file_dir)\n if context is not '':\n with open(file_dir + '/' + title, 'w') as f:\n f.write(context)\n f.close()\n return item\n\n\n# db version\nclass MongoPipeline(object):\n\n def __init__(self):\n self.client = pymongo.MongoClient()\n self.db = self.client['netease']\n self.collection = self.db['news']\n\n def process_item(self, item, spider):\n self.collection.insert_one(item)\n return item\n", "sub_path": "news_netease/pipelines.py", "file_name": "pipelines.py", "file_ext": "py", "file_size_in_byte": 996, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "sys.setdefaultencoding", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 23, "usage_type": "call"}, {"api_name": "pymongo.MongoClient", "line_number": 35, "usage_type": "call"}]} +{"seq_id": "106794827", "text": "from gui.Gui import MyWindow\r\nfrom PyQt5 import QtWidgets\r\nimport sys\r\nimport consultas\r\nfrom errores import NotFound, NotAcceptable, GenomeError, BadRequest\r\n\r\nlista_escritura = []\r\nclass T03Window(MyWindow):\r\n def __init__(self):\r\n super().__init__()\r\n\r\n def process_query(self, query_array):\r\n # Agrega en pantalla la solucion. Muestra los graficos!!\r\n for i in range(len(query_array)):\r\n text = \"-------------------- Consulta \" + str(i + 1) + \\\r\n \" --------------------\" + \"\\n\"\r\n lista_escritura.append(text)\r\n self.add_answer(text)\r\n if query_array[i][0] == 'pariente_de':\r\n try:\r\n nombre = query_array[i][2]\r\n id = consultas.id_getter(nombre)\r\n grado = query_array[i][1]\r\n lista = consultas.pariente_de(grado, id)\r\n for j in range(len(lista)):\r\n self.add_answer(lista[j] + '\\n')\r\n lista_escritura.append((lista[j] + '\\n'))\r\n if len(lista) == 0:\r\n self.add_answer(\"Error: Not Acceptable\" + '\\n')\r\n self.add_answer(NotFound(\"Falta Respuesta #NotFound\") + '\\n')\r\n lista_escritura.append((\"Falta Respuesta #NotFound\" + '\\n'))\r\n except:\r\n print(NotFound(\"No tiene parientes de grado{}\".format(query_array[0])))\r\n\r\n\r\n elif query_array[i][0] == 'índice_de_tamaño':\r\n try:\r\n id = consultas.id_getter(query_array[i][1])\r\n ans = consultas.índice_de_tamaño(id)\r\n self.add_answer(ans + '\\n')\r\n lista_escritura.append((ans + '\\n'))\r\n except:\r\n print(NotFound(\"Error: Not Acceptable\"))\r\n\r\n elif query_array[i][0] == 'ascendencia':\r\n try:\r\n nombre = query_array[i][1]\r\n id = consultas.id_getter(nombre)\r\n lista = consultas.ascendencia(id)\r\n lista[0]\r\n for j in range(len(lista)):\r\n self.add_answer(lista[j] + '\\n')\r\n lista_escritura.append((lista[j] + '\\n'))\r\n except:\r\n print(NotFound(\"No tiene ascendencia\"))\r\n self.add_answer('Error: Not Acceptable' + '\\n')\r\n lista_escritura.append(('Error: Not Acceptable' + '\\n'))\r\n\r\n elif query_array[i][0] == 'gemelo_genético':\r\n self.add_answer('No alcancé a hacerla' + '\\n')\r\n lista_escritura.append(('No alcancé a hacerla' + '\\n'))\r\n\r\n\r\n elif query_array[i][0] == 'valor_característica':\r\n try:\r\n ans = consultas.valor_característica(query_array[i][1], query_array[i][2])\r\n if isinstance(ans, tuple):\r\n ans = ans[0]\r\n #Convierte la tupla de la guata en un string\r\n if type(ans) != list:\r\n ans = [ans]\r\n for j in range(len(ans)):\r\n self.add_answer(str(ans[j]))\r\n lista_escritura.append((str(ans[j])))\r\n if j < len(ans) -1 :\r\n self.add_answer(', ')\r\n lista_escritura.append(', ')\r\n\r\n self.add_answer('\\n')\r\n lista_escritura.append('\\n')\r\n except:\r\n print(NotFound(query_array[i][0], query_array[i][1]))\r\n\r\n elif query_array[i][0] == 'min':\r\n try:\r\n ans = consultas.min(query_array[1])\r\n self.add_answer(ans + '\\n')\r\n lista_escritura.append(ans + '\\n')\r\n except:\r\n print(NotFound(query_array[i][0], query_array[i][1]))\r\n\r\n elif query_array[i][0] == 'max':\r\n try:\r\n ans = consultas.max(query_array[1])\r\n self.add_answer(ans + '\\n')\r\n lista_escritura.append(ans + '\\n')\r\n except:\r\n NotFound(\"máximo no se puede entregar\")\r\n\r\n elif query_array[i][0] == 'prom':\r\n ans = consultas.prom(query_array[1])\r\n self.add_answer(ans + '\\n')\r\n lista_escritura.append('\\n')\r\n else:\r\n print(BadRequest(query_array[i]))\r\n\r\n\r\n def save_file(self, query_array):\r\n # Crea un archivo con la solucion. NO muestra los graficos!!\r\n file = open('resultados.txt', 'w')\r\n print(\"\".join(lista_escritura), file = file)\r\n file.close()\r\n\r\nif __name__ == '__main__':\r\n def hook(type, value, traceback):\r\n print(type)\r\n print(value)\r\n print(traceback)\r\n\r\n\r\n sys.__excepthook__ = hook\r\n\r\n app = QtWidgets.QApplication(sys.argv)\r\n window = T03Window()\r\n sys.exit(app.exec_())\r\n", "sub_path": "test_1/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 5124, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "gui.Gui.MyWindow", "line_number": 8, "usage_type": "name"}, {"api_name": "consultas.id_getter", "line_number": 22, "usage_type": "call"}, {"api_name": "consultas.pariente_de", "line_number": 24, "usage_type": "call"}, {"api_name": "errores.NotFound", "line_number": 30, "usage_type": "call"}, {"api_name": "errores.NotFound", "line_number": 33, "usage_type": "call"}, {"api_name": "consultas.id_getter", "line_number": 38, "usage_type": "call"}, {"api_name": "consultas.índice_de_tamaño", "line_number": 39, "usage_type": "call"}, {"api_name": "errores.NotFound", "line_number": 43, "usage_type": "call"}, {"api_name": "consultas.id_getter", "line_number": 48, "usage_type": "call"}, {"api_name": "consultas.ascendencia", "line_number": 49, "usage_type": "call"}, {"api_name": "errores.NotFound", "line_number": 55, "usage_type": "call"}, {"api_name": "consultas.valor_característica", "line_number": 66, "usage_type": "call"}, {"api_name": "errores.NotFound", "line_number": 82, "usage_type": "call"}, {"api_name": "consultas.min", "line_number": 86, "usage_type": "call"}, {"api_name": "errores.NotFound", "line_number": 90, "usage_type": "call"}, {"api_name": "consultas.max", "line_number": 94, "usage_type": "call"}, {"api_name": "errores.NotFound", "line_number": 98, "usage_type": "call"}, {"api_name": "consultas.prom", "line_number": 101, "usage_type": "call"}, {"api_name": "errores.BadRequest", "line_number": 105, "usage_type": "call"}, {"api_name": "sys.__excepthook__", "line_number": 121, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 123, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 123, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 123, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 125, "usage_type": "call"}]} +{"seq_id": "98391513", "text": "\nimport numpy as np\nfrom . import BiQuad\n\nimport numba\nfrom tqdm import tqdm\n\nBiQuad_type = numba.deferred_type()\nBiQuad_type.define(BiQuad.BiQuad.class_type.instance_type)\n\nspec = [\n ('filters', BiQuad_type[:]),\n]\n\n# simple LF-Rd wavetable oscillator based on Fant 1985, Fant 1995\nclass SerialFilterBank:\n\n\n def tick(self, x, verbose=False):\n \n if np.isscalar(x):\n y = x\n for filter in self.filters:\n y = filter.tick(y)\n return y\n else:\n \n y = x\n for i in tqdm(range(len(y)), disable=(not verbose)): \n for filter in self.filters:\n y[i] = filter.tick(y[i])\n return y\n\n\n #processes signal with per-sample pole zero updates\n def tick_pz(self, x, poles, zeros, verbose = False):\n\n\n if np.isscalar(x):\n self.set_pz_conjugates(poles, zeros)\n y = x\n for filter in self.filters:\n y = filter.tick(y)\n return y\n else:\n \n y = x\n for i in tqdm(range(len(y)), disable=(not verbose)): \n self.set_pz_conjugates(poles[i,:], zeros[i,:])\n for filter in self.filters: \n y[i] = filter.tick(y[i])\n return y\n\n\n\n # updates coefficients with array of poles zeros which are extended\n # to include conjuate pairs\n def set_pz_conjugates(self, poles, zeros):\n\n for i in range(len(poles)):\n p0 = poles[i]\n z0 = zeros[i]\n\n re_p0 = np.real(p0)\n im_p0 = np.imag(p0)\n re_z0 = np.real(z0)\n im_z0 = np.imag(z0)\n\n #self.filters[i].a1 = -np.real(np.conj(p0) + p0)\n #self.filters[i].a2 = np.real(np.conj(p0) * p0)\n \n self.filters[i].b0 = 1\n #self.filters[i].b1 = -np.real(np.conj(z0) + z0)\n #self.filters[i].b2 = np.real(np.conj(z0) * z0)\n \n self.filters[i].a1 = -(re_p0 + re_p0)\n self.filters[i].a2 = re_p0 * re_p0 + im_p0 * im_p0\n \n self.filters[i].b1 = -(re_z0 + re_z0)\n self.filters[i].b2 = re_z0 * re_z0 + im_z0 * im_z0\n\n\n\n def __init__(self, num_sos):\n \n self.filters = []\n for i in range(num_sos):\n self.filters.append(BiQuad.BiQuad())\n\n\n ", "sub_path": "hsvs/tools/SerialFilterBank.py", "file_name": "SerialFilterBank.py", "file_ext": "py", "file_size_in_byte": 2428, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "numba.deferred_type", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.isscalar", "line_number": 21, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.isscalar", "line_number": 39, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.real", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.imag", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.real", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.imag", "line_number": 67, "usage_type": "call"}]} +{"seq_id": "508696822", "text": "# encoding=utf8\nimport urllib2\nfrom bs4 import BeautifulSoup\n# import urllib.requests\nimport socket\nimport requests\nsession = requests.session()\nUser_Agent = 'Mozilla/5.0 (Windows NT 6.3; WOW64; rv:43.0) Gecko/20100101 Firefox/43.0'\nheader = {}\nheader['User-Agent'] = User_Agent\n\n'''\n获取所有代理IP地址 http://ip.zdaye.com/\n'''\n\n\ndef getProxyIp():\n proxy = []\n for i in range(1, 10):\n try:\n url = 'http://www.kuaidaili.com/free/inha/' + str(i)\n # req = urllib2.Request(url, headers=header)\n res = session.get(url).content\n soup = BeautifulSoup(res, \"html.parser\")\n ips = soup.findAll('tr')\n for x in range(1, len(ips)):\n ip = ips[x]\n tds = ip.findAll(\"td\")\n ip_temp = tds[0].text + \"\\t\" + tds[1].text\n proxy.append(ip_temp)\n except:\n continue\n return proxy\n\n\n'''\n验证获得的代理IP地址是否可用\n'''\n\ndef validateIp(proxy):\n url = \"http://www.bjtu.edu.cn\"\n f = open(\"C:/Users/DELL/Desktop/ip.txt\", \"w\")\n socket.setdefaulttimeout(3)\n for i in range(0, len(proxy)):\n try:\n ip = proxy[i].strip().split(\"\\t\")\n proxy_host = \"http://\" + ip[0] + \":\" + ip[1]\n proxy_temp = {\"http\": proxy_host}\n session.get(url, proxies=proxy_temp)\n f.write(proxy[i] + '\\n')\n print(proxy[i])\n except Exception:\n continue\n f.close()\n'''\n更换ip \n# proxie = { \n# 'http' : 'http://122.193.14.102:80'\n# } \n# url = 'xxx'\n# \n# response = s.get(url, verify=False, proxies = proxie, timeout = 20)\n'''\n\nif __name__ == '__main__':\n proxy = getProxyIp()\n validateIp(proxy)", "sub_path": "IP_download.py", "file_name": "IP_download.py", "file_ext": "py", "file_size_in_byte": 1742, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "requests.session", "line_number": 7, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 24, "usage_type": "call"}, {"api_name": "socket.setdefaulttimeout", "line_number": 43, "usage_type": "call"}]} +{"seq_id": "87593090", "text": "import numpy as np\nimport tensorflow as tf\nimport cv2\nimport os\n\nos.chdir(\"/home/cloud-user/CNN/Images/n02088364-beagle\")\ni = 0\nfor image in os.listdir():\n\tim = cv2.imread(image)\n\tprint(np.shape(im))\n\tgray = cv2.cvtColor(im, cv2.COLOR_BGR2GRAY)\n\tprint(np.shape(gray))\n\tgray = cv2.resize(gray, (28,28))\n\tprint(np.shape(gray))\n\tif i ==10:\n\t\texit()\n\ti+=1\n", "sub_path": "dogs.py", "file_name": "dogs.py", "file_ext": "py", "file_size_in_byte": 352, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "os.chdir", "line_number": 6, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 8, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 11, "usage_type": "attribute"}, {"api_name": "numpy.shape", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 14, "usage_type": "call"}]} +{"seq_id": "265794238", "text": "from bs4 import BeautifulSoup\n\nimport urllib2\n\nopener = urllib2.build_opener()\nopener.addheaders = [('User-agent', 'Mozilla/5.0')]\n\nurl = \"http://scholar.google.com/scholar\"\n\nappendix = \"?q=computer&btnG=&as_sdt=1%2C5&as_sdtp=\"\n\ndata = opener.open(url + appendix).read()\n\nsoup = BeautifulSoup(data)\nfor div in soup.find_all('div', {'class' : 'gs_r'}):\n print(div.prettify())\n", "sub_path": "print_all.py", "file_name": "print_all.py", "file_ext": "py", "file_size_in_byte": 378, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "urllib2.build_opener", "line_number": 5, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 14, "usage_type": "call"}]} +{"seq_id": "525651675", "text": "import numpy as np\nimport tensorflow as tf\nfrom PIL import Image\n\nnumber = ['0', '1', '2', '3', '4', '5', '6', '7', '8', '9']\nalphabet = ['a', 'b', 'c', 'd', 'e', 'f', 'g', 'h', 'i', 'j', 'k', 'l', 'm', 'n', 'o', 'p', 'q', 'r', 's', 't', 'u',\n 'v', 'w', 'x', 'y', 'z']\nAlphabet = ['A', 'B', 'C', 'D', 'E', 'F', 'G', 'H', 'I', 'J', 'K', 'L', 'M', 'N', 'O', 'P', 'Q', 'R', 'S', 'T', 'U',\n 'V', 'W', 'X', 'Y', 'Z']\n\nchar_set = number\n\n##图片高\nIMAGE_HEIGHT = 50\n##图片宽\nIMAGE_WIDTH = 200\n##验证码长度\nMAX_CAPTCHA = 5\n##验证码选择空间\nCHAR_SET_LEN = len(char_set)\n##提前定义变量空间\nX = tf.placeholder(tf.float32, [None, IMAGE_HEIGHT * IMAGE_WIDTH])\nY = tf.placeholder(tf.float32, [None, MAX_CAPTCHA * CHAR_SET_LEN])\nkeep_prob = tf.placeholder(tf.float32) ##节点保留率\n\n\ndef initTable():\n table = []\n for i in range(256):\n if i < 90:\n table.append(0)\n else:\n table.append(1)\n return table\n\n\ndef preHandleImage(imagePath):\n im = Image.open(imagePath)\n im = im.convert('L')\n img = im.point(initTable(), '1')\n img = np.array(img)\n img = 1 * (img.flatten())\n return img\n\n\n##彩色图转化为灰度图\ndef convert2gray(img):\n if len(img.shape) > 2:\n gray = np.mean(img, -1)\n # 上面的转法较快,正规转法如下\n # r, g, b = img[:,:,0], img[:,:,1], img[:,:,2]\n # gray = 0.2989 * r + 0.5870 * g + 0.1140 * b\n # print(gray)\n return gray\n else:\n return img\n\n\n##获取字符在 字符域中下标\ndef getPos(char_set=char_set, char=None):\n return char_set.index(char)\n\n\n##验证码字符转换为长向量\ndef text2vec(text):\n text_len = len(text)\n if text_len > MAX_CAPTCHA:\n raise ValueError('验证码最长4个字符')\n\n vector = np.zeros(MAX_CAPTCHA * CHAR_SET_LEN)\n \"\"\"\n def char2pos(c): \n if c =='_': \n k = 62 \n return k \n k = ord(c)-48 \n if k > 9: \n k = ord(c) - 55 \n if k > 35: \n k = ord(c) - 61 \n if k > 61: \n raise ValueError('No Map') \n return k \n \"\"\"\n for i, c in enumerate(text):\n idx = i * CHAR_SET_LEN + getPos(char=c)\n vector[idx] = 1\n return vector\n\n\n##卷积层 附relu max_pool drop操作\ndef conn_layer(w_alpha=0.01, b_alpha=0.1, _keep_prob=0.75, input=None, last_size=None, cur_size=None):\n w_c1 = tf.Variable(w_alpha * tf.random_normal([3, 3, last_size, cur_size]))\n b_c1 = tf.Variable(b_alpha * tf.random_normal([cur_size]))\n conv1 = tf.nn.relu(tf.nn.bias_add(tf.nn.conv2d(input, w_c1, strides=[1, 1, 1, 1], padding='SAME'), b_c1))\n conv1 = tf.nn.max_pool(conv1, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME')\n conv1 = tf.nn.dropout(conv1, keep_prob=_keep_prob)\n return conv1\n\n\n##对卷积层到全链接层的数据进行变换\ndef _get_conn_last_size(input):\n shape = input.get_shape().as_list()\n dim = 1\n for d in shape[1:]:\n dim *= d\n input = tf.reshape(input, [-1, dim])\n return input, dim\n\n\n##全链接层\ndef _fc_layer(w_alpha=0.01, b_alpha=0.1, input=None, last_size=None, cur_size=None):\n w_d = tf.Variable(w_alpha * tf.random_normal([last_size, cur_size]))\n b_d = tf.Variable(b_alpha * tf.random_normal([cur_size]))\n fc = tf.nn.bias_add(tf.matmul(input, w_d), b_d)\n return fc\n\n\n##构建前向传播网络\ndef crack_captcha_cnn(w_alpha=0.01, b_alpha=0.1):\n x = tf.reshape(X, shape=[-1, IMAGE_HEIGHT, IMAGE_WIDTH, 1])\n\n conv1 = conn_layer(input=x, last_size=1, cur_size=32)\n conv2 = conn_layer(input=conv1, last_size=32, cur_size=64)\n conn3 = conn_layer(input=conv2, last_size=64, cur_size=128)\n\n input, dim = _get_conn_last_size(conn3)\n\n fc_layer1 = _fc_layer(input=input, last_size=dim, cur_size=1024)\n fc_layer1 = tf.nn.relu(fc_layer1)\n fc_layer1 = tf.nn.dropout(fc_layer1, keep_prob)\n\n fc_out = _fc_layer(input=fc_layer1, last_size=1024, cur_size=MAX_CAPTCHA * CHAR_SET_LEN)\n return fc_out\n\n\n##反向传播\ndef back_propagation():\n output = crack_captcha_cnn()\n ##学习率\n loss = tf.reduce_mean(tf.nn.sigmoid_cross_entropy_with_logits(labels=Y, logits=output))\n optm = tf.train.AdamOptimizer(learning_rate=0.001).minimize(loss)\n predict = tf.reshape(output, [-1, MAX_CAPTCHA, CHAR_SET_LEN])\n max_idx_p = tf.arg_max(predict, 2)\n max_idx_l = tf.arg_max(tf.reshape(Y, [-1, MAX_CAPTCHA, CHAR_SET_LEN]), 2)\n accuracy = tf.reduce_mean(tf.cast(tf.equal(max_idx_p, max_idx_l), tf.float32))\n return loss, optm, accuracy\n\n\noutput = crack_captcha_cnn()\nsaver = tf.train.Saver()\nsess = tf.InteractiveSession()\nsess.run(tf.initialize_all_variables())\n# with tf.Session() as sess:\npath = './susongwuyou.model-1200'\nsaver.restore(sess, path)\n\n\n##测试训练模型\ndef crack_captcha(captcha_image):\n imag = preHandleImage(captcha_image)\n predict = tf.argmax(tf.reshape(output, [-1, MAX_CAPTCHA, CHAR_SET_LEN]), 2)\n text_list = sess.run(predict, feed_dict={X: [imag], keep_prob: 1})\n text = text_list[0].tolist()\n result = ''.join(map(str, text))\n return result\n", "sub_path": "susongwuyoutemp/susongwuyou.py", "file_name": "susongwuyou.py", "file_ext": "py", "file_size_in_byte": 5175, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "tensorflow.placeholder", "line_number": 22, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 22, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 23, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 23, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 24, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 38, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 38, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 70, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 93, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 93, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 94, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 94, "usage_type": "call"}, {"api_name": "tensorflow.nn.relu", "line_number": 95, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 95, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.bias_add", "line_number": 95, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d", "line_number": 95, "usage_type": "call"}, {"api_name": "tensorflow.nn.max_pool", "line_number": 96, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 96, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.dropout", "line_number": 97, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 97, "usage_type": "attribute"}, {"api_name": "tensorflow.reshape", "line_number": 107, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 113, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 113, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 114, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 114, "usage_type": "call"}, {"api_name": "tensorflow.nn.bias_add", "line_number": 115, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 115, "usage_type": "attribute"}, {"api_name": "tensorflow.matmul", "line_number": 115, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 121, "usage_type": "call"}, {"api_name": "tensorflow.nn.relu", "line_number": 130, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 130, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.dropout", "line_number": 131, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 131, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_mean", "line_number": 141, "usage_type": "call"}, {"api_name": "tensorflow.nn.sigmoid_cross_entropy_with_logits", "line_number": 141, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 141, "usage_type": "attribute"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 142, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 142, "usage_type": "attribute"}, {"api_name": "tensorflow.reshape", "line_number": 143, "usage_type": "call"}, {"api_name": "tensorflow.arg_max", "line_number": 144, "usage_type": "call"}, {"api_name": "tensorflow.arg_max", "line_number": 145, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 145, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 146, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 146, "usage_type": "call"}, {"api_name": "tensorflow.equal", "line_number": 146, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 146, "usage_type": "attribute"}, {"api_name": "tensorflow.train.Saver", "line_number": 151, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 151, "usage_type": "attribute"}, {"api_name": "tensorflow.InteractiveSession", "line_number": 152, "usage_type": "call"}, {"api_name": "tensorflow.initialize_all_variables", "line_number": 153, "usage_type": "call"}, {"api_name": "tensorflow.argmax", "line_number": 162, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 162, "usage_type": "call"}]} +{"seq_id": "604500126", "text": "from django.utils.datetime_safe import datetime\nfrom rest_framework import status\nfrom rest_framework.authentication import get_authorization_header\nfrom rest_framework.authtoken.views import ObtainAuthToken\nfrom rest_framework.response import Response\nfrom rest_framework.authtoken.models import Token\n\n\nclass ObtainExpiringAuthToken(ObtainAuthToken):\n def post(self, request):\n serializer = self.serializer_class(data=request.DATA)\n if serializer.is_valid():\n token, created = Token.objects.get_or_create(user=serializer.object['user'])\n if not created:\n token.created = datetime.utcnow()\n token.save()\n return Response({'token': token.key},\n status=status.HTTP_201_CREATED)\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n def delete(self, request):\n auth = get_authorization_header(request).split()\n if not auth or auth[0].lower() != b'token':\n return Response({\"status\": \"No token provided.\"},\n status=status.HTTP_401_UNAUTHORIZED)\n if len(auth) == 1:\n msg = 'Invalid token header. No credentials provided.'\n return Response({\"status\": msg}, status=status.HTTP_400_BAD_REQUEST)\n elif len(auth) > 2:\n msg = 'Invalid token header. Token string should not contain spaces.'\n return Response({\"status\": msg}, status=status.HTTP_400_BAD_REQUEST)\n # The token was provided, so we are going to delete it\n try:\n token = Token.objects.get(key=auth[1])\n token.delete()\n return Response({'status': \"Token deleted.\".format(auth[1])})\n except Token.DoesNotExist:\n return Response({\"status\": \"Token does not exist.\"},\n status=status.HTTP_401_UNAUTHORIZED)\n", "sub_path": "public/neumeeditor/views/authentication.py", "file_name": "authentication.py", "file_ext": "py", "file_size_in_byte": 1890, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "rest_framework.authtoken.views.ObtainAuthToken", "line_number": 9, "usage_type": "name"}, {"api_name": "rest_framework.authtoken.models.Token.objects.get_or_create", "line_number": 13, "usage_type": "call"}, {"api_name": "rest_framework.authtoken.models.Token.objects", "line_number": 13, "usage_type": "attribute"}, {"api_name": "rest_framework.authtoken.models.Token", "line_number": 13, "usage_type": "name"}, {"api_name": "django.utils.datetime_safe.datetime.utcnow", "line_number": 15, "usage_type": "call"}, {"api_name": "django.utils.datetime_safe.datetime", "line_number": 15, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 17, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 18, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 18, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 19, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 19, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 19, "usage_type": "name"}, {"api_name": "rest_framework.authentication.get_authorization_header", "line_number": 22, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 24, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_401_UNAUTHORIZED", "line_number": 25, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 25, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 28, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 28, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 28, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 31, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 31, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 31, "usage_type": "name"}, {"api_name": "rest_framework.authtoken.models.Token.objects.get", "line_number": 34, "usage_type": "call"}, {"api_name": "rest_framework.authtoken.models.Token.objects", "line_number": 34, "usage_type": "attribute"}, {"api_name": "rest_framework.authtoken.models.Token", "line_number": 34, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 36, "usage_type": "call"}, {"api_name": "rest_framework.authtoken.models.Token.DoesNotExist", "line_number": 37, "usage_type": "attribute"}, {"api_name": "rest_framework.authtoken.models.Token", "line_number": 37, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 38, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_401_UNAUTHORIZED", "line_number": 39, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 39, "usage_type": "name"}]} +{"seq_id": "403200545", "text": "import telegram\nfrom telegram import ChatAction, InlineKeyboardButton,InlineKeyboardMarkup\nfrom telegram.ext import Updater, CommandHandler,MessageHandler,Filters,CallbackQueryHandler,PicklePersistence\nfrom telegram.ext.dispatcher import run_async\nimport logging\nimport os\nimport requests\nimport base64\n\nlogging.basicConfig(format='%(asctime)s - %(name)s - %(levelname)s - %(message)s',\n level=logging.INFO)\n\nlogger = logging.getLogger(__name__)\n\n@run_async\ndef start(update,context):\n\tname=update.message.chat.first_name\n\tupdate.message.reply_text(\"Hi! \"+name+\"\\nWelcome to Plant Identifier Bot, Send a clear image of a plant and i will try to recognise the plant in it\\nWant to support this bot? click /donate \\nRate this bot by clicking --> /rate\")\n\n@run_async\t\ndef rate(update,context):\n update.message.reply_text(\"To give ratings to this bot click here\\nhttps://t.me/BotsArchive/1504\")\n\n@run_async \ndef donate(update,context):\n update.message.reply_text(\"If you really liked ❤️ this bot and want to support it and help me to pay for the server expenses , you can donate me any amount you wish because _every penny counts\\.\\.\\._\\nYour support can really help this bot to run *24x7* \\n*Payment Options* \\n\\n1\\. [Paypal](https://paypal.me/yamit11) \\n\\n2\\. UPI : `amity11@kotak` \\n\\n3\\. [Debit/Credit cards/UPI](https://rzp.io/l/amity11)\\nIf you want different payment option please contact @amit\\_y11\",parse_mode=telegram.ParseMode.MARKDOWN_V2)\n \ndef encode_files(file_names):\n files_encoded = []\n for file_name in file_names:\n with open(file_name, \"rb\") as file:\n files_encoded.append(base64.b64encode(file.read()).decode(\"ascii\"))\n return files_encoded\n \nkeyboard=[[InlineKeyboardButton(\"➕Add Image ➕\", callback_data=\"add\"), InlineKeyboardButton (\"⚙️ Process ⚙️\",callback_data=\"process\")]]\nreply_markup = InlineKeyboardMarkup(keyboard)\n\n@run_async \ndef identify (update,context):\n\t####################\n\t#Getting user info\n global chat_id\n chat_id=update.effective_chat.id\n #####################\n\n check=context.bot.get_chat_member(\"@botsbyamit\",update.effective_chat)\n if check['status']==\"member\" or check['status']==\"creator\":\n try:\n images = context.user_data['images']\n except KeyError:\n images = context.user_data['images'] = []\n i=len(images)+1\n file=\"plant\"+str(i)+\".jpg\"\n \n file_id = update.message.photo[-1].file_id\n try:\n fileid=context.user_data['fileid']\n except KeyError:\n fileid=context.user_data['fileid']=[] \n fileid.insert(i-1,file_id)\n \n newFile=context.bot.get_file(file_id)\n newFile.download(file)\n images.insert(i-1,file)\n context.bot.send_message(chat_id=chat_id,text=\"Click Add image option to add another image of the plant or click process button to start processing with the sended image\",reply_markup=reply_markup)\n else:\n update.message.reply_text(\"You need to be a member of @botsbyamit in order to use this bot.\\n\\nPlease join @botsbyamit and send your image again to continue.\")\n\n \n#############################\n@run_async\ndef button(update,context):\n images=context.user_data['images']\n query=update.callback_query\n query.answer()\n if query.data==\"process\":\n try:\n query.edit_message_text(\"Processing .....\")\n context.bot.send_chat_action(chat_id=update.effective_message.chat_id ,action=telegram.ChatAction.TYPING)\n api_key = \"Your Api key from Plant.id\"\n image=encode_files(images)\n json_data = {\"images\": image,\"modifiers\": [\"similar_images\"],\"plant_details\": [\"common_names\", \"url\", \"wiki_description\", \"taxonomy\"]}\n response = requests.post(\"https://api.plant.id/v2/identify\", json=json_data, headers={\"Content-Type\": \"application/json\",\"Api-Key\": api_key}).json()\n data=response ['suggestions'][0]\n name=data['plant_name']\n probability=data['probability']\n data1=data['plant_details']\n scientific_name=data1['scientific_name']\n common=data1['common_names']\n try:\n common=\", \".join(common)\n except:\n common=\"None\"\n url=data1['url']\n \n \n message=\"Plant Name : \"+name+\"\\n\\nProbability : \"+str(probability*100)+\"\\n\\nScientific name : \"+scientific_name+\"\\n\\nCommon Names : \"+common+\"\\n\\nWikipedia Page : \"+url+\"\\n\\npowered by Plant.id Api\\n\\n\"\n query.edit_message_text(message,parse_mode=telegram.ParseMode.HTML)\n images.clear()\n except Exception as e:\n query.edit_message_text(\"Limit for this week has been reached please check back again next week\")\n images.clear()\n \n elif query.data==\"add\":\n query.edit_message_text(text=\"Okay send another image of the same plant\")\n return images\n\n@run_async\ndef clear(update,context):\n\ttry:\n\t\timages=context.user_data['images']\n\texcept KeyError:\n\t\tupdate.message.reply_text(\"You haven't sended any image of the plant yet\")\n\timages.clear()\n\tupdate.message.reply_text(\"All images removed\")\n\n@run_async\ndef remove(update,context):\n try:\n images=context.user_data['images']\n except KeyError:\n update.message.reply_text(\"You haven't sended any image of the plant yet\")\n if len(images)==0:\n update.message.reply_text(\"You haven't sended any image of the plant yet\")\n else:\n images.pop()\n update.message.reply_text(\"Last image sended by you is been removed\",reply_markup=reply_markup)\n\t\ndef getinfo(update,context):\n\ttry:\n\t\timages=context.user_data['images']\n\texcept KeyError:\n\t\tupdate.message.reply_text(\"You have not sended any images yet\")\n\tinfo=len(images)\n\tupdate.message.reply_text(\"You have sended \"+str(info)+\" images\")\n\t\n\t\n\npersistence=PicklePersistence('plantdata')\ndef main(): \n token=\"Your Bot Token\"\n updater = Updater(token,use_context=True, persistence=persistence)\n dp=updater.dispatcher\n dp.add_handler(CommandHandler('start',start))\n dp.add_handler(CommandHandler('donate',donate))\n dp.add_handler(CommandHandler('rate',rate))\n dp.add_handler(MessageHandler(Filters.photo, identify))\n dp.add_handler(CallbackQueryHandler(button))\n dp.add_handler(CommandHandler ('removeall',clear))\n dp.add_handler(CommandHandler ('removelast', remove))\n dp.add_handler(CommandHandler (\"getinfo\",getinfo))\n updater.start_polling()\n updater.idle()\n \n\t\nif __name__==\"__main__\":\n\tmain()\n", "sub_path": "plant.py", "file_name": "plant.py", "file_ext": "py", "file_size_in_byte": 6720, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "logging.basicConfig", "line_number": 10, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 11, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 13, "usage_type": "call"}, {"api_name": "telegram.ext.dispatcher.run_async", "line_number": 15, "usage_type": "name"}, {"api_name": "telegram.ext.dispatcher.run_async", "line_number": 20, "usage_type": "name"}, {"api_name": "telegram.ParseMode", "line_number": 26, "usage_type": "attribute"}, {"api_name": "telegram.ext.dispatcher.run_async", "line_number": 24, "usage_type": "name"}, {"api_name": "base64.b64encode", "line_number": 32, "usage_type": "call"}, {"api_name": "telegram.InlineKeyboardButton", "line_number": 35, "usage_type": "call"}, {"api_name": "telegram.InlineKeyboardMarkup", "line_number": 36, "usage_type": "call"}, {"api_name": "telegram.ext.dispatcher.run_async", "line_number": 38, "usage_type": "name"}, {"api_name": "telegram.ChatAction", "line_number": 79, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 83, "usage_type": "call"}, {"api_name": "telegram.ParseMode", "line_number": 98, "usage_type": "attribute"}, {"api_name": "telegram.ext.dispatcher.run_async", "line_number": 71, "usage_type": "name"}, {"api_name": "telegram.ext.dispatcher.run_async", "line_number": 108, "usage_type": "name"}, {"api_name": "telegram.ext.dispatcher.run_async", "line_number": 117, "usage_type": "name"}, {"api_name": "telegram.ext.PicklePersistence", "line_number": 139, "usage_type": "call"}, {"api_name": "telegram.ext.Updater", "line_number": 142, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 144, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 145, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 146, "usage_type": "call"}, {"api_name": "telegram.ext.MessageHandler", "line_number": 147, "usage_type": "call"}, {"api_name": "telegram.ext.Filters.photo", "line_number": 147, "usage_type": "attribute"}, {"api_name": "telegram.ext.Filters", "line_number": 147, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackQueryHandler", "line_number": 148, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 149, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 150, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 151, "usage_type": "call"}]} +{"seq_id": "581791413", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Mar 31 16:29:37 2018\n\n@author: lnogga\n\"\"\"\n\nfrom torch.autograd import Variable\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torch.optim as optim\n\n\n\nclass Net(nn.Module):\n def __init__(self):\n super(Net, self).__init__()\n self.fc1 = nn.Linear(2, 8)\n self.fc2 = nn.Linear(8, 1)\n\n def forward(self, x):\n x = x.view(-1, 2)\n x = F.relu(self.fc1(x))\n x = F.relu(self.fc2(x))\n return x\n\n#trainset = \n#validset\n#testset =\n\n\nnet = Net()\ncriterion = nn.MSELoss()\noptimizer = optim.Adam()\n\nfor epoch in range(2): # loop over the dataset multiple times\n\n running_loss = 0.0\n for i, data in enumerate(trainset):\n # get the inputs\n inputs, labels = data\n\n # wrap them in Variable\n inputs, labels = Variable(inputs), Variable(labels)\n\n # zero the parameter gradients\n optimizer.zero_grad()\n\n # forward + backward + optimize\n outputs = net(inputs)\n loss = criterion(outputs, labels)\n loss.backward()\n optimizer.step()\n\n # print statistics\n running_loss += loss.data[0]\n if i % 2000 == 1999: # print every 2000 mini-batches\n print('[%d, %5d] loss: %.3f' %\n (epoch + 1, i + 1, running_loss / 2000))\n running_loss = 0.0\n\nprint('Finished Training')", "sub_path": "function_approximation.py", "file_name": "function_approximation.py", "file_ext": "py", "file_size_in_byte": 1412, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "torch.nn.Module", "line_number": 16, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 16, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 24, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 25, "usage_type": "name"}, {"api_name": "torch.nn.MSELoss", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 34, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.autograd.Variable", "line_number": 45, "usage_type": "call"}]} +{"seq_id": "602287232", "text": "# -*- coding: utf-8 -*-\n# This program is free software: you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation, either version 3 of the License, or\n# (at your option) any later version.\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU General Public License for more details.\n# You should have received a copy of the GNU General Public License\n# along with this program. If not, see .\n\n\"\"\"Contains classes and functions for a the project settings\n\n.. module:: project_settings\n :synopsis: Classes and functions for a the project settings\n\n.. moduleauthor:: Karsten Bock \n\"\"\"\nfrom __future__ import print_function\n\nfrom future import standard_library\nstandard_library.install_aliases()\nimport os\n\nimport PyCEGUI\n\nfrom .dialog import Dialog\nfrom .common import (cb_cut_copy_paste, is_dir_path_valid, select_path,\n ask_create_path)\n\n\nclass ProjectSettings(Dialog):\n\n \"\"\"Class that displays a project settings dialog\"\"\"\n\n def __init__(self, app, project_settings, project_dir):\n \"\"\"Constructor\"\"\"\n Dialog.__init__(self, app)\n self.d_camera_editor = None\n self.agt_path_editor = None\n self.agt_path_cb = None\n self.p_name_editor = None\n self.d_camera_cb = None\n self.obj_nspace_editor = None\n self.obj_nspace_cb = None\n self.comp_file_editor = None\n self.comp_file_cb = None\n self.act_file_editor = None\n self.act_file_cb = None\n self.syst_file_editor = None\n self.syst_file_cb = None\n self.project_settings = project_settings\n self.project_dir = project_dir\n self.beh_file_editor = None\n self.beh_file_cb = None\n self.comb_file_editor = None\n self.comb_file_cb = None\n\n def setup_dialog(self, root):\n \"\"\"Sets up the dialog windows\n\n Args:\n\n root: The root window to which the windows should be added\n \"\"\"\n self.window.setArea(PyCEGUI.UDim(0, 3), PyCEGUI.UDim(0, 4),\n PyCEGUI.UDim(0.4, 3), PyCEGUI.UDim(0.55, 4))\n self.window.setMinSize(PyCEGUI.USize(PyCEGUI.UDim(0.4, 3),\n PyCEGUI.UDim(0.55, 4)))\n self.window.setText(_(\"Project Settings\"))\n\n font = root.getFont()\n\n browse_button_width = 50\n margin = 5\n cb_width = 15\n evt_select_state_changed = PyCEGUI.ToggleButton.EventSelectStateChanged\n evt_btn_clicked = PyCEGUI.ButtonBase.EventMouseClick\n evt_text_accepted = PyCEGUI.Editbox.EventTextAccepted\n evt_key_down = PyCEGUI.Window.EventKeyDown\n\n vert_margin = PyCEGUI.UBox(PyCEGUI.UDim(0, margin), PyCEGUI.UDim(0, 0),\n PyCEGUI.UDim(0, margin), PyCEGUI.UDim(0, 0))\n horz_margin = PyCEGUI.UBox(PyCEGUI.UDim(0, 0), PyCEGUI.UDim(0, margin),\n PyCEGUI.UDim(0, 0), PyCEGUI.UDim(0, margin))\n\n p_name_layout = root.createChild(\"HorizontalLayoutContainer\")\n p_name_layout.setMargin(vert_margin)\n p_name_layout.setHeight(PyCEGUI.UDim(0.05, 0))\n p_name_label = p_name_layout.createChild(\"TaharezLook/Label\")\n p_name_label.setMargin(horz_margin)\n p_name_label.setText(_(\"Projectname\"))\n p_name_label.setProperty(\"HorzFormatting\", \"LeftAligned\")\n text_width = font.getTextExtent(p_name_label.getText())\n p_name_label.setWidth(PyCEGUI.UDim(0, text_width))\n p_name_editor = p_name_layout.createChild(\"TaharezLook/Editbox\")\n p_name_editor.setMargin(horz_margin)\n p_name_editor.setWidth(PyCEGUI.UDim(1.0, -(text_width + 4 * margin)))\n p_name_editor.subscribeEvent(evt_key_down,\n cb_cut_copy_paste)\n self.p_name_editor = p_name_editor\n\n agt_path_layout = root.createChild(\"HorizontalLayoutContainer\")\n agt_path_layout.setMargin(vert_margin)\n agt_path_layout.setHeight(PyCEGUI.UDim(0.05, 0))\n agt_path_cb = agt_path_layout.createChild(\"TaharezLook/Checkbox\")\n agt_path_cb.setMargin(horz_margin)\n agt_path_cb.setText(_(\"Agent Object Path\"))\n text_width = font.getTextExtent(agt_path_cb.getText()) + cb_width\n agt_path_cb.setWidth(PyCEGUI.UDim(0, text_width))\n agt_path_edit_layout = agt_path_layout.createChild(\"HorizontalLayout\"\n \"Container\")\n agt_path_edit_layout.setWidth((PyCEGUI.UDim(1.0, -\n (text_width +\n 4 * margin))))\n agt_path_edit_layout.setDisabled(True)\n agt_path_cb.subscribeEvent(evt_select_state_changed,\n (lambda args:\n cb_op_cb_selection_changed(\n args,\n agt_path_edit_layout)))\n self.agt_path_cb = agt_path_cb\n agt_path_editor = agt_path_edit_layout.createChild(\"TaharezLook/\"\n \"Editbox\")\n agt_path_editor.setMargin(horz_margin)\n agt_path_editor.setWidth(PyCEGUI.UDim(1.0, - (text_width + 4 * margin +\n browse_button_width)))\n agt_path_editor.subscribeEvent(evt_text_accepted,\n self.cb_agent_path_accepted)\n agt_path_editor.subscribeEvent(evt_key_down,\n cb_cut_copy_paste)\n self.agt_path_editor = agt_path_editor\n agt_path_browse = agt_path_edit_layout.createChild(\"TaharezLook/\"\n \"Button\")\n agt_path_browse.setWidth(PyCEGUI.UDim(0.0, browse_button_width))\n agt_path_browse.setText(\"...\")\n agt_path_browse.subscribeEvent(evt_btn_clicked,\n self.cb_agent_path_browse_clicked)\n\n d_camera_layout = root.createChild(\"HorizontalLayoutContainer\")\n d_camera_layout.setMargin(vert_margin)\n d_camera_layout.setHeight(PyCEGUI.UDim(0.05, 0))\n d_camera_cb = d_camera_layout.createChild(\"TaharezLook/Checkbox\")\n d_camera_cb.setMargin(horz_margin)\n d_camera_cb.setText(_(\"Default Camera\"))\n text_width = font.getTextExtent(d_camera_cb.getText()) + cb_width\n d_camera_cb.setWidth(PyCEGUI.UDim(0, text_width))\n self.d_camera_cb = d_camera_cb\n d_camera_editor = d_camera_layout.createChild(\"TaharezLook/Editbox\")\n d_camera_editor.setMargin(horz_margin)\n d_camera_editor.setWidth(PyCEGUI.UDim(1.0, -(text_width + 4 * margin)))\n d_camera_editor.setDisabled(True)\n d_camera_editor.subscribeEvent(evt_key_down,\n cb_cut_copy_paste)\n d_camera_cb.subscribeEvent(evt_select_state_changed,\n (lambda args:\n cb_op_cb_selection_changed(\n args,\n d_camera_editor)))\n self.d_camera_editor = d_camera_editor\n\n obj_nspace_layout = root.createChild(\"HorizontalLayoutContainer\")\n obj_nspace_layout.setMargin(vert_margin)\n obj_nspace_layout.setHeight(PyCEGUI.UDim(0.05, 0))\n obj_nspace_cb = obj_nspace_layout.createChild(\"TaharezLook/Checkbox\")\n obj_nspace_cb.setMargin(horz_margin)\n obj_nspace_cb.setText(_(\"Object Namespace\"))\n text_width = font.getTextExtent(obj_nspace_cb.getText()) + cb_width\n obj_nspace_cb.setWidth(PyCEGUI.UDim(0, text_width))\n self.obj_nspace_cb = obj_nspace_cb\n obj_nspace_editor = obj_nspace_layout.createChild(\n \"TaharezLook/Editbox\")\n obj_nspace_editor.setMargin(horz_margin)\n obj_nspace_editor.setWidth(PyCEGUI.UDim(1.0,\n -(text_width + 4 * margin)))\n obj_nspace_editor.setDisabled(True)\n obj_nspace_editor.subscribeEvent(evt_key_down,\n cb_cut_copy_paste)\n obj_nspace_cb.subscribeEvent(evt_select_state_changed,\n (lambda args:\n cb_op_cb_selection_changed(\n args,\n obj_nspace_editor)))\n self.obj_nspace_editor = obj_nspace_editor\n\n filetypes = [('yaml files', '.yaml'), ('all files', '.*')]\n\n comp_file_layout = root.createChild(\"HorizontalLayoutContainer\")\n comp_file_layout.setMargin(vert_margin)\n comp_file_layout.setHeight(PyCEGUI.UDim(0.05, 0))\n comp_file_cb = comp_file_layout.createChild(\"TaharezLook/Checkbox\")\n comp_file_cb.setMargin(horz_margin)\n comp_file_cb.setText(_(\"Components File\"))\n text_width = font.getTextExtent(comp_file_cb.getText()) + cb_width\n comp_file_cb.setWidth(PyCEGUI.UDim(0, text_width))\n comp_file_edit_layout = comp_file_layout.createChild(\"HorizontalLayout\"\n \"Container\")\n comp_file_edit_layout.setWidth((PyCEGUI.UDim(1.0, -\n (text_width +\n 4 * margin))))\n comp_file_edit_layout.setDisabled(True)\n comp_file_cb.subscribeEvent(evt_select_state_changed,\n (lambda args:\n cb_op_cb_selection_changed(\n args,\n comp_file_edit_layout)))\n self.comp_file_cb = comp_file_cb\n comp_file_editor = comp_file_edit_layout.createChild(\"TaharezLook/\"\n \"Editbox\")\n comp_file_editor.setMargin(horz_margin)\n comp_file_editor.setWidth(PyCEGUI.UDim(1.0,\n - (text_width + 4 * margin +\n browse_button_width)))\n comp_file_editor.subscribeEvent(evt_text_accepted,\n self.cb_open_file_path_accepted)\n comp_file_editor.subscribeEvent(evt_key_down,\n cb_cut_copy_paste)\n self.comp_file_editor = comp_file_editor\n comp_file_browse = comp_file_edit_layout.createChild(\"TaharezLook/\"\n \"Button\")\n comp_file_browse.setWidth(PyCEGUI.UDim(0.0, browse_button_width))\n comp_file_browse.setText(\"...\")\n comp_file_browse.subscribeEvent(\n evt_btn_clicked,\n lambda args: self.cb_open_file_browse_clicked(args,\n comp_file_editor,\n filetypes\n ))\n\n act_file_layout = root.createChild(\"HorizontalLayoutContainer\")\n act_file_layout.setMargin(vert_margin)\n act_file_layout.setHeight(PyCEGUI.UDim(0.05, 0))\n act_file_cb = act_file_layout.createChild(\"TaharezLook/Checkbox\")\n act_file_cb.setMargin(horz_margin)\n act_file_cb.setText(_(\"Actions File\"))\n text_width = font.getTextExtent(act_file_cb.getText()) + cb_width\n act_file_cb.setWidth(PyCEGUI.UDim(0, text_width))\n act_file_edit_layout = act_file_layout.createChild(\"HorizontalLayout\"\n \"Container\")\n act_file_edit_layout.setWidth((PyCEGUI.UDim(1.0, -\n (text_width +\n 4 * margin))))\n act_file_edit_layout.setDisabled(True)\n act_file_cb.subscribeEvent(evt_select_state_changed,\n (lambda args:\n cb_op_cb_selection_changed(\n args,\n act_file_edit_layout)))\n self.act_file_cb = act_file_cb\n act_file_editor = act_file_edit_layout.createChild(\"TaharezLook/\"\n \"Editbox\")\n act_file_editor.setMargin(horz_margin)\n act_file_editor.setWidth(PyCEGUI.UDim(1.0,\n - (text_width + 4 * margin +\n browse_button_width)))\n act_file_editor.subscribeEvent(evt_text_accepted,\n self.cb_open_file_path_accepted)\n act_file_editor.subscribeEvent(evt_key_down,\n cb_cut_copy_paste)\n self.act_file_editor = act_file_editor\n act_file_browse = act_file_edit_layout.createChild(\"TaharezLook/\"\n \"Button\")\n act_file_browse.setWidth(PyCEGUI.UDim(0.0, browse_button_width))\n act_file_browse.setText(\"...\")\n act_file_browse.subscribeEvent(\n evt_btn_clicked,\n lambda args: self.cb_open_file_browse_clicked(args,\n act_file_editor,\n filetypes\n ))\n\n syst_file_layout = root.createChild(\"HorizontalLayoutContainer\")\n syst_file_layout.setMargin(vert_margin)\n syst_file_layout.setHeight(PyCEGUI.UDim(0.05, 0))\n syst_file_cb = syst_file_layout.createChild(\"TaharezLook/Checkbox\")\n syst_file_cb.setMargin(horz_margin)\n syst_file_cb.setText(_(\"Systems File\"))\n text_width = font.getTextExtent(syst_file_cb.getText()) + cb_width\n syst_file_cb.setWidth(PyCEGUI.UDim(0, text_width))\n syst_file_edit_layout = syst_file_layout.createChild(\"HorizontalLayout\"\n \"Container\")\n syst_file_edit_layout.setWidth((PyCEGUI.UDim(1.0, -\n (text_width +\n 4 * margin))))\n syst_file_edit_layout.setDisabled(True)\n syst_file_cb.subscribeEvent(evt_select_state_changed,\n (lambda args:\n cb_op_cb_selection_changed(\n args,\n syst_file_edit_layout)))\n self.syst_file_cb = syst_file_cb\n syst_file_editor = syst_file_edit_layout.createChild(\"TaharezLook/\"\n \"Editbox\")\n syst_file_editor.setMargin(horz_margin)\n syst_file_editor.setWidth(PyCEGUI.UDim(1.0,\n - (text_width + 4 * margin +\n browse_button_width)))\n syst_file_editor.subscribeEvent(evt_text_accepted,\n self.cb_open_file_path_accepted)\n syst_file_editor.subscribeEvent(evt_key_down,\n cb_cut_copy_paste)\n self.syst_file_editor = syst_file_editor\n syst_file_browse = syst_file_edit_layout.createChild(\"TaharezLook/\"\n \"Button\")\n syst_file_browse.setWidth(PyCEGUI.UDim(0.0, browse_button_width))\n syst_file_browse.setText(\"...\")\n syst_file_browse.subscribeEvent(\n evt_btn_clicked,\n lambda args: self.cb_open_file_browse_clicked(args,\n syst_file_editor,\n filetypes\n ))\n\n beh_file_layout = root.createChild(\"HorizontalLayoutContainer\")\n beh_file_layout.setMargin(vert_margin)\n beh_file_layout.setHeight(PyCEGUI.UDim(0.05, 0))\n beh_file_cb = beh_file_layout.createChild(\"TaharezLook/Checkbox\")\n beh_file_cb.setMargin(horz_margin)\n beh_file_cb.setText(_(\"Behaviours File\"))\n text_width = font.getTextExtent(beh_file_cb.getText()) + cb_width\n beh_file_cb.setWidth(PyCEGUI.UDim(0, text_width))\n beh_file_edit_layout = beh_file_layout.createChild(\"HorizontalLayout\"\n \"Container\")\n beh_file_edit_layout.setWidth((PyCEGUI.UDim(1.0, -\n (text_width +\n 4 * margin))))\n beh_file_edit_layout.setDisabled(True)\n beh_file_cb.subscribeEvent(evt_select_state_changed,\n (lambda args:\n cb_op_cb_selection_changed(\n args,\n beh_file_edit_layout)))\n self.beh_file_cb = beh_file_cb\n beh_file_editor = beh_file_edit_layout.createChild(\"TaharezLook/\"\n \"Editbox\")\n beh_file_editor.setMargin(horz_margin)\n beh_file_editor.setWidth(PyCEGUI.UDim(1.0,\n - (text_width + 4 * margin +\n browse_button_width)))\n beh_file_editor.subscribeEvent(evt_text_accepted,\n self.cb_open_file_path_accepted)\n beh_file_editor.subscribeEvent(evt_key_down,\n cb_cut_copy_paste)\n self.beh_file_editor = beh_file_editor\n beh_file_browse = beh_file_edit_layout.createChild(\"TaharezLook/\"\n \"Button\")\n beh_file_browse.setWidth(PyCEGUI.UDim(0.0, browse_button_width))\n beh_file_browse.setText(\"...\")\n beh_file_browse.subscribeEvent(\n evt_btn_clicked,\n lambda args: self.cb_open_file_browse_clicked(args,\n beh_file_editor,\n filetypes\n ))\n\n comb_file_layout = root.createChild(\"HorizontalLayoutContainer\")\n comb_file_layout.setMargin(vert_margin)\n comb_file_layout.setHeight(PyCEGUI.UDim(0.05, 0))\n comb_file_cb = comb_file_layout.createChild(\"TaharezLook/Checkbox\")\n comb_file_cb.setMargin(horz_margin)\n comb_file_cb.setText(_(\"Combined File\"))\n text_width = font.getTextExtent(comb_file_cb.getText()) + cb_width\n comb_file_cb.setWidth(PyCEGUI.UDim(0, text_width))\n comb_file_edit_layout = comb_file_layout.createChild(\"HorizontalLayout\"\n \"Container\")\n comb_file_edit_layout.setWidth((PyCEGUI.UDim(1.0, -\n (text_width +\n 4 * margin))))\n comb_file_edit_layout.setDisabled(True)\n comb_file_cb.subscribeEvent(evt_select_state_changed,\n (lambda args:\n cb_op_cb_selection_changed(\n args,\n comb_file_edit_layout)))\n self.comb_file_cb = comb_file_cb\n comb_file_editor = comb_file_edit_layout.createChild(\"TaharezLook/\"\n \"Editbox\")\n comb_file_editor.setMargin(horz_margin)\n comb_file_editor.setWidth(PyCEGUI.UDim(1.0,\n - (text_width + 4 * margin +\n browse_button_width)))\n comb_file_editor.subscribeEvent(evt_text_accepted,\n self.cb_open_file_path_accepted)\n comb_file_editor.subscribeEvent(evt_key_down,\n cb_cut_copy_paste)\n self.comb_file_editor = comb_file_editor\n comb_file_browse = comb_file_edit_layout.createChild(\"TaharezLook/\"\n \"Button\")\n comb_file_browse.setWidth(PyCEGUI.UDim(0.0, browse_button_width))\n comb_file_browse.setText(\"...\")\n comb_file_browse.subscribeEvent(\n evt_btn_clicked,\n lambda args: self.cb_open_file_browse_clicked(args,\n comb_file_editor,\n filetypes\n ))\n\n if self.project_settings:\n self.fill_fields()\n\n def get_values(self):\n \"\"\"Returns the values of the dialog fields\"\"\"\n values = {}\n try:\n values[\"ProjectName\"] = self.p_name_editor.getText()\n if self.agt_path_cb.isSelected():\n values[\"AgentObjectsPath\"] = self.agt_path_editor.getText()\n if self.d_camera_cb.isSelected():\n values[\"Camera\"] = self.d_camera_editor.getText()\n if self.obj_nspace_cb.isSelected():\n values[\"ObjectNamespace\"] = self.obj_nspace_editor.getText()\n if self.comp_file_cb.isSelected():\n values[\"ComponentsFile\"] = self.comp_file_editor.getText()\n if self.act_file_cb.isSelected():\n values[\"ActionsFile\"] = self.act_file_editor.getText()\n if self.syst_file_cb.isSelected():\n values[\"SystemsFile\"] = self.syst_file_editor.getText()\n if self.beh_file_cb.isSelected():\n values[\"BehavioursFile\"] = self.beh_file_editor.getText()\n if self.comb_file_cb.isSelected():\n values[\"CombinedFile\"] = self.comb_file_editor.getText()\n except AttributeError:\n print(\"Please call setup_dialog before trying to get the values\")\n return values\n\n def validate(self):\n \"\"\"Check if the current state of the dialog fields is valid\"\"\"\n try:\n if not self.p_name_editor.getText().strip():\n return False\n if self.agt_path_cb.isSelected():\n if not self.agt_path_editor.getText().strip():\n return False\n path = self.agt_path_editor.getText()\n if not os.path.isabs(path):\n path = os.path.join(self.project_dir,\n self.agt_path_editor.getText())\n if not os.path.exists(path):\n return False\n if not os.path.isdir(path):\n return False\n if self.d_camera_cb.isSelected():\n if not self.d_camera_editor.getText().strip():\n return False\n if self.comp_file_cb.isSelected():\n if not self.comp_file_editor.getText().strip():\n return False\n path = self.comp_file_editor.getText()\n if not os.path.isabs(path):\n path = os.path.join(self.project_dir,\n path)\n if not os.path.exists(path):\n return False\n if not os.path.isfile(path):\n return False\n if self.act_file_cb.isSelected():\n if not self.act_file_editor.getText().strip():\n return False\n path = self.act_file_editor.getText()\n if not os.path.isabs(path):\n path = os.path.join(self.project_dir,\n path)\n if not os.path.exists(path):\n return False\n if not os.path.isfile(path):\n return False\n if self.syst_file_cb.isSelected():\n if not self.syst_file_editor.getText().strip():\n return False\n path = self.syst_file_editor.getText()\n if not os.path.isabs(path):\n path = os.path.join(self.project_dir,\n path)\n if not os.path.exists(path):\n return False\n if not os.path.isfile(path):\n return False\n if self.beh_file_cb.isSelected():\n if not self.beh_file_editor.getText().strip():\n return False\n path = self.beh_file_editor.getText()\n if not os.path.isabs(path):\n path = os.path.join(self.project_dir,\n path)\n if not os.path.exists(path):\n return False\n if not os.path.isfile(path):\n return False\n if self.comb_file_cb.isSelected():\n if not self.comb_file_editor.getText().strip():\n return False\n path = self.comb_file_editor.getText()\n if not os.path.isabs(path):\n path = os.path.join(self.project_dir,\n path)\n if not os.path.exists(path):\n return False\n if not os.path.isfile(path):\n return False\n return True\n except AttributeError:\n print (\"Please call setup_dialog before trying to validate \"\n \"the values\")\n return False\n\n def fill_fields(self):\n \"\"\"Fills the fields from the data in the project file\"\"\"\n\n if \"ProjectName\" in self.project_settings:\n self.p_name_editor.setText(self.project_settings[\"ProjectName\"])\n if \"Camera\" in self.project_settings:\n self.d_camera_editor.setText(self.project_settings[\"Camera\"])\n self.d_camera_cb.setSelected(True)\n if \"ObjectNamespace\" in self.project_settings:\n self.obj_nspace_editor.setText(self.project_settings[\"Object\"\n \"Namespace\"])\n self.obj_nspace_cb.setSelected(True)\n if \"AgentObjectsPath\" in self.project_settings:\n self.agt_path_editor.setText(self.project_settings[\"AgentObjects\"\n \"Path\"])\n self.agt_path_cb.setSelected(True)\n if \"ComponentsFile\" in self.project_settings:\n self.comp_file_editor.setText(self.project_settings[\"Components\"\n \"File\"])\n self.comp_file_cb.setSelected(True)\n if \"ActionsFile\" in self.project_settings:\n self.act_file_editor.setText(self.project_settings[\"Actions\"\n \"File\"])\n self.act_file_cb.setSelected(True)\n if \"SystemsFile\" in self.project_settings:\n self.syst_file_editor.setText(self.project_settings[\"Systems\"\n \"File\"])\n self.syst_file_cb.setSelected(True)\n if \"BehavioursFile\" in self.project_settings:\n self.beh_file_editor.setText(self.project_settings[\"Behaviours\"\n \"File\"])\n self.beh_file_cb.setSelected(True)\n if \"CombinedFile\" in self.project_settings:\n self.comb_file_editor.setText(self.project_settings[\"Combined\"\n \"File\"])\n self.comb_file_cb.setSelected(True)\n\n def make_relative_to_project(self, abspath):\n \"\"\"Makes a path relative tot he current loaded project, if its inside\n the project directory.\n\n Returns: If the path is inside the project directory if will return a\n relative version of it, otherwise it will return the full path.\n \"\"\"\n if abspath.startswith(self.project_dir):\n return os.path.relpath(abspath, self.project_dir)\n return abspath\n\n def check_agent_path(self, new_path):\n \"\"\"Checks if path is valid and returns either the path, in relative\n form if possible, or None if invalid.\"\"\"\n new_path = os.path.normpath(new_path)\n abspath = None\n if os.path.isabs(new_path):\n abspath = new_path\n new_path = self.make_relative_to_project(new_path)\n else:\n abspath = os.path.join(self.project_dir, new_path)\n if not is_dir_path_valid(abspath):\n return None\n return new_path\n\n def cb_agent_path_browse_clicked(self, args):\n \"\"\"Callback for click on the browse button of the agent path\"\"\"\n initialdir = (self.agt_path_editor.getText().strip() or\n self.project_dir)\n if not os.path.isabs(initialdir):\n initialdir = os.path.join(self.project_dir, initialdir)\n initialdir = os.path.normpath(initialdir)\n selected_path = select_path(\"Select Agent Object Path\", initialdir)\n if not selected_path:\n return True\n checked_path = self.check_agent_path(selected_path)\n if not checked_path:\n import tkinter.messagebox\n tkinter.messagebox.showerror(_(\"Invalid path\"),\n _(\"%s is not a valid path\") % selected_path)\n return True\n abspath = checked_path\n if not os.path.isabs(abspath):\n abspath = os.path.join(self.project_dir, abspath)\n if not os.path.exists(abspath):\n if not ask_create_path(abspath):\n return\n self.agt_path_editor.setText(checked_path)\n\n def cb_agent_path_accepted(self, args):\n \"\"\"Callback when a text was 'accepted' in the agent path editbox\"\"\"\n new_path = self.check_agent_path(self.agt_path_editor.getText())\n if new_path is not None:\n self.agt_path_editor.setText(new_path)\n self.agt_path_editor.setCaretIndex(0)\n\n def check_open_file_path(self, path):\n \"\"\"Check a file path that should be opened\n\n Args:\n\n path: The path to check\n \"\"\"\n path = os.path.normpath(path)\n if os.path.isabs(path):\n check_path = path\n path = self.make_relative_to_project(path)\n else:\n check_path = os.path.join(self.project_dir, path)\n if not os.path.exists(check_path):\n path = None\n elif not os.path.isfile(check_path):\n path = None\n return path\n\n def check_and_set_open_file_path(self, target, path):\n \"\"\"Checks a path that should be opened and sets the contents\n of an editbox to it.\n\n Args:\n target: The editox that should be set\n\n path: The path that should be checked\"\"\"\n path = self.check_open_file_path(path)\n if path is not None:\n target.setText(path)\n target.setCaretIndex(0)\n\n def cb_open_file_path_accepted(self, args):\n \"\"\"Callback when a text was 'accepted' a file path editbox\"\"\"\n self.check_and_set_open_file_path(args.window, args.window.getText())\n\n def cb_open_file_browse_clicked(self, args, target, filetypes):\n \"\"\"Called when a button to browse a file to open was clicked\n\n Args:\n\n args: EventArgs passy ba cegui\n\n target: The editbox whose text should be set to the selected file\n\n filetypes: The available file types\n \"\"\"\n import tkinter.filedialog\n selected_file = tkinter.filedialog.askopenfilename(\n filetypes=filetypes,\n title=\"Open file\",\n initialdir=self.project_dir)\n if not selected_file:\n return\n self.check_and_set_open_file_path(target, selected_file)\n\n\ndef cb_op_cb_selection_changed(args, target):\n \"\"\"Called when the selection state of a checkbox in front of an option\n was changed\n\n Args:\n\n args: WindowEventArgs passed by cegui\n\n target: The window to enabled/disable by the checkbox\n \"\"\"\n window = args.window\n target.setDisabled(not window.isSelected())\n", "sub_path": "editor/project_settings.py", "file_name": "project_settings.py", "file_ext": "py", "file_size_in_byte": 32903, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "future.standard_library.install_aliases", "line_number": 23, "usage_type": "call"}, {"api_name": "future.standard_library", "line_number": 23, "usage_type": "name"}, {"api_name": "dialog.Dialog", "line_number": 33, "usage_type": "name"}, {"api_name": "dialog.Dialog.__init__", "line_number": 39, "usage_type": "call"}, {"api_name": "dialog.Dialog", "line_number": 39, "usage_type": "name"}, {"api_name": "PyCEGUI.UDim", "line_number": 67, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 68, "usage_type": "call"}, {"api_name": "PyCEGUI.USize", "line_number": 69, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 69, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 70, "usage_type": "call"}, {"api_name": "PyCEGUI.ToggleButton", "line_number": 78, "usage_type": "attribute"}, {"api_name": "PyCEGUI.ButtonBase", "line_number": 79, "usage_type": "attribute"}, {"api_name": "PyCEGUI.Editbox", "line_number": 80, "usage_type": "attribute"}, {"api_name": "PyCEGUI.Window", "line_number": 81, "usage_type": "attribute"}, {"api_name": "PyCEGUI.UBox", "line_number": 83, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 83, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 84, "usage_type": "call"}, {"api_name": "PyCEGUI.UBox", "line_number": 85, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 85, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 86, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 90, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 96, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 99, "usage_type": "call"}, {"api_name": "common.cb_cut_copy_paste", "line_number": 101, "usage_type": "argument"}, {"api_name": "PyCEGUI.UDim", "line_number": 106, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 111, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 114, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 127, "usage_type": "call"}, {"api_name": "common.cb_cut_copy_paste", "line_number": 132, "usage_type": "argument"}, {"api_name": "PyCEGUI.UDim", "line_number": 136, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 143, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 148, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 152, "usage_type": "call"}, {"api_name": "common.cb_cut_copy_paste", "line_number": 155, "usage_type": "argument"}, {"api_name": "PyCEGUI.UDim", "line_number": 165, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 170, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 175, "usage_type": "call"}, {"api_name": "common.cb_cut_copy_paste", "line_number": 179, "usage_type": "argument"}, {"api_name": "PyCEGUI.UDim", "line_number": 191, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 196, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 199, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 212, "usage_type": "call"}, {"api_name": "common.cb_cut_copy_paste", "line_number": 218, "usage_type": "argument"}, {"api_name": "PyCEGUI.UDim", "line_number": 222, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 233, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 238, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 241, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 254, "usage_type": "call"}, {"api_name": "common.cb_cut_copy_paste", "line_number": 260, "usage_type": "argument"}, {"api_name": "PyCEGUI.UDim", "line_number": 264, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 275, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 280, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 283, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 296, "usage_type": "call"}, {"api_name": "common.cb_cut_copy_paste", "line_number": 302, "usage_type": "argument"}, {"api_name": "PyCEGUI.UDim", "line_number": 306, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 317, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 322, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 325, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 338, "usage_type": "call"}, {"api_name": "common.cb_cut_copy_paste", "line_number": 344, "usage_type": "argument"}, {"api_name": "PyCEGUI.UDim", "line_number": 348, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 359, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 364, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 367, "usage_type": "call"}, {"api_name": "PyCEGUI.UDim", "line_number": 380, "usage_type": "call"}, {"api_name": "common.cb_cut_copy_paste", "line_number": 386, "usage_type": "argument"}, {"api_name": "PyCEGUI.UDim", "line_number": 390, "usage_type": "call"}, {"api_name": "os.path.isabs", "line_number": 436, "usage_type": "call"}, {"api_name": "os.path", "line_number": 436, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 437, "usage_type": "call"}, {"api_name": "os.path", "line_number": 437, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 439, "usage_type": "call"}, {"api_name": "os.path", "line_number": 439, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 441, "usage_type": "call"}, {"api_name": "os.path", "line_number": 441, "usage_type": "attribute"}, {"api_name": "os.path.isabs", "line_number": 450, "usage_type": "call"}, {"api_name": "os.path", "line_number": 450, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 451, "usage_type": "call"}, {"api_name": "os.path", "line_number": 451, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 453, "usage_type": "call"}, {"api_name": "os.path", "line_number": 453, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 455, "usage_type": "call"}, {"api_name": "os.path", "line_number": 455, "usage_type": "attribute"}, {"api_name": "os.path.isabs", "line_number": 461, "usage_type": "call"}, {"api_name": "os.path", "line_number": 461, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 462, "usage_type": "call"}, {"api_name": "os.path", "line_number": 462, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 464, "usage_type": "call"}, {"api_name": "os.path", "line_number": 464, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 466, "usage_type": "call"}, {"api_name": "os.path", "line_number": 466, "usage_type": "attribute"}, {"api_name": "os.path.isabs", "line_number": 472, "usage_type": "call"}, {"api_name": "os.path", "line_number": 472, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 473, "usage_type": "call"}, {"api_name": "os.path", "line_number": 473, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 475, "usage_type": "call"}, {"api_name": "os.path", "line_number": 475, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 477, "usage_type": "call"}, {"api_name": "os.path", "line_number": 477, "usage_type": "attribute"}, {"api_name": "os.path.isabs", "line_number": 483, "usage_type": "call"}, {"api_name": "os.path", "line_number": 483, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 484, "usage_type": "call"}, {"api_name": "os.path", "line_number": 484, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 486, "usage_type": "call"}, {"api_name": "os.path", "line_number": 486, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 488, "usage_type": "call"}, {"api_name": "os.path", "line_number": 488, "usage_type": "attribute"}, {"api_name": "os.path.isabs", "line_number": 494, "usage_type": "call"}, {"api_name": "os.path", "line_number": 494, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 495, "usage_type": "call"}, {"api_name": "os.path", "line_number": 495, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 497, "usage_type": "call"}, {"api_name": "os.path", "line_number": 497, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 499, "usage_type": "call"}, {"api_name": "os.path", "line_number": 499, "usage_type": "attribute"}, {"api_name": "os.path.relpath", "line_number": 552, "usage_type": "call"}, {"api_name": "os.path", "line_number": 552, "usage_type": "attribute"}, {"api_name": "os.path.normpath", "line_number": 558, "usage_type": "call"}, {"api_name": "os.path", "line_number": 558, "usage_type": "attribute"}, {"api_name": "os.path.isabs", "line_number": 560, "usage_type": "call"}, {"api_name": "os.path", "line_number": 560, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 564, "usage_type": "call"}, {"api_name": "os.path", "line_number": 564, "usage_type": "attribute"}, {"api_name": "common.is_dir_path_valid", "line_number": 565, "usage_type": "call"}, {"api_name": "os.path.isabs", "line_number": 573, "usage_type": "call"}, {"api_name": "os.path", "line_number": 573, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 574, "usage_type": "call"}, {"api_name": "os.path", "line_number": 574, "usage_type": "attribute"}, {"api_name": "os.path.normpath", "line_number": 575, "usage_type": "call"}, {"api_name": "os.path", "line_number": 575, "usage_type": "attribute"}, {"api_name": "common.select_path", "line_number": 576, "usage_type": "call"}, {"api_name": "tkinter.messagebox.messagebox.showerror", "line_number": 582, "usage_type": "call"}, {"api_name": "tkinter.messagebox.messagebox", "line_number": 582, "usage_type": "attribute"}, {"api_name": "tkinter.messagebox", "line_number": 582, "usage_type": "name"}, {"api_name": "os.path.isabs", "line_number": 586, "usage_type": "call"}, {"api_name": "os.path", "line_number": 586, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 587, "usage_type": "call"}, {"api_name": "os.path", "line_number": 587, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 588, "usage_type": "call"}, {"api_name": "os.path", "line_number": 588, "usage_type": "attribute"}, {"api_name": "common.ask_create_path", "line_number": 589, "usage_type": "call"}, {"api_name": "os.path.normpath", "line_number": 607, "usage_type": "call"}, {"api_name": "os.path", "line_number": 607, "usage_type": "attribute"}, {"api_name": "os.path.isabs", "line_number": 608, "usage_type": "call"}, {"api_name": "os.path", "line_number": 608, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 612, "usage_type": "call"}, {"api_name": "os.path", "line_number": 612, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 613, "usage_type": "call"}, {"api_name": "os.path", "line_number": 613, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 615, "usage_type": "call"}, {"api_name": "os.path", "line_number": 615, "usage_type": "attribute"}, {"api_name": "tkinter.messagebox.filedialog.askopenfilename", "line_number": 648, "usage_type": "call"}, {"api_name": "tkinter.messagebox.filedialog", "line_number": 648, "usage_type": "attribute"}, {"api_name": "tkinter.messagebox", "line_number": 648, "usage_type": "name"}]} +{"seq_id": "21438442", "text": "#!/usr/bin/env python3\nimport argparse\nimport sys\n\nif __name__ == \"__main__\":\n parser = argparse.ArgumentParser()\n parser.add_argument(\"output\", type=str, help=\"Output file path\")\n args = parser.parse_args()\n\n entries = {}\n\n for line in sys.stdin:\n node, entry = line.rstrip(\"\\n\").split(maxsplit=1)\n if entry not in entries:\n entries[entry] = []\n entries[entry].append(node)\n\n with open(args.output, \"w\", encoding=\"utf-8\") as output_file, open(args.output + \".nodes\", \"w\", encoding=\"utf-8\") as output_nodes_file:\n for entry, nodes in sorted(sorted(entries.items(), key=lambda x: x[0]), key=lambda x: len(x[1]), reverse=True):\n print(\"{:6d} {}\".format(len(nodes), entry), file=output_file)\n print(\"{:6d} {} -> {}\".format(len(nodes), entry.split(\" ->\")[0], \" \".join(sorted(nodes))), file=output_nodes_file)\n", "sub_path": "WorkData/tools/tagging/data_consistency/consistency_vertical_grouper.py", "file_name": "consistency_vertical_grouper.py", "file_ext": "py", "file_size_in_byte": 884, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 12, "usage_type": "attribute"}]} +{"seq_id": "591438689", "text": "import pip\nfrom pip._internal import main as pip\n\ndef install(domname):\n subprocess.call([sys.executable, \"-m\", \"pip\", \"install\",\"-U\", \"-r\", \"requirements.txt\"])\n\ndef domain_name(domname):\n if hasattr(pip, 'main'):\n pip.main(['main', domname])\n else:\n pip._internal.main(['install', domname])\n\nif __name__ == '__main__':\n domain_name('argh') ", "sub_path": "pipfunc.py", "file_name": "pipfunc.py", "file_ext": "py", "file_size_in_byte": 371, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "pip._internal.main", "line_number": 8, "usage_type": "argument"}, {"api_name": "pip._internal.main.main", "line_number": 9, "usage_type": "call"}, {"api_name": "pip._internal.main", "line_number": 9, "usage_type": "name"}, {"api_name": "pip._internal.main._internal.main", "line_number": 11, "usage_type": "call"}, {"api_name": "pip._internal.main._internal", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pip._internal.main", "line_number": 11, "usage_type": "name"}]} +{"seq_id": "431155041", "text": "import qrcode \nimport uuid\n\n# img = qrcode.make('hello, qrcode')\n# img.save('test.png')\n\n# qr = qrcode.QRCode( \n # version=1, \n # error_correction=qrcode.constants.ERROR_CORRECT_L, \n # box_size=10, \n # border=4, \n# ) \n# qr.add_data('hello, qrcode') \n# qr.make(fit=True) \n# img = qr.make_image()\n# img.save('123.png')\n\n# 参数含义:\n# version:值为1~40的整数,控制二维码的大小(最小值是1,是个12×12的矩阵)。 如果想让程序自动确定,将值设置为 None 并使用 fit 参数即可。\n\n# error_correction:控制二维码的错误纠正功能。可取值下列4个常量。\n#   ERROR_CORRECT_L:大约7%或更少的错误能被纠正。\n#   ERROR_CORRECT_M(默认):大约15%或更少的错误能被纠正。\n#   ROR_CORRECT_H:大约30%或更少的错误能被纠正。\n\n# box_size:控制二维码中每个小格子包含的像素数。\n\n# border:控制边框(二维码与图片边界的距离)包含的格子数(默认为4,是相关标准规定的最小值)\n\n#for i in range(0,100):\n#\tguid = uuid.uuid1()\n# img = qrcode.make(\"https://60.28.163.133:10010/Pages/Login.aspx\")\n# img.save('login.png')\n\nf = open(\"JoinStr.txt\",encoding='utf-8',errors='ignore');\nLINES = f.readlines();\nf.close();\n\ni = 1\nfor line in LINES:\n\t#guid = uuid.uuid1()\n\tline = line.replace(\"\\n\",\"\")\n\tif len(line) > 0:\n\t\timg = qrcode.make(line)\n\t\timg.save(str(i)+'.png')\n\ti = i+1\n\n", "sub_path": "快速生成二维码/快速生成二维码.py", "file_name": "快速生成二维码.py", "file_ext": "py", "file_size_in_byte": 1456, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "qrcode.make", "line_number": 44, "usage_type": "call"}]} +{"seq_id": "74163698", "text": "import torch\nimport random\n\n\nclass Genetic:\n mutation_prob = 3\n\n def __init__(self, bots: list, escalate_mutations, reset_mutations):\n self.bots = bots\n if escalate_mutations:\n self.mutation_prob += 2\n if reset_mutations:\n self.mutation_prob = 3\n self.fitnesses = [bot.player.time_lived for bot in self.bots]\n self.n_input, self.n_hidden_1, self.n_hidden_2, self.n_output, self.layers_num = self.bots[0].get_net_structure()\n self.neurons_num_in_layers = [self.n_input, self.n_hidden_1, self.n_hidden_2, self.n_output]\n self.boundaries = []\n self.genomes = []\n self.selection()\n self.store_genomes()\n self.crossover()\n self.mutation()\n self.refresh_bots()\n\n def selection(self):\n self.bots.sort(key=lambda fit: fit.player.time_lived, reverse=True)\n\n def store_genomes(self):\n for bot in self.bots:\n genome = []\n for layer in bot.net.layers:\n weight = layer.weight\n bias = layer.bias\n genome.append(torch.cat((weight, bias.unsqueeze(1)), 1))\n self.genomes.append(genome)\n\n def create_children_genomes(self, parent1, parent2):\n child1 = []\n child2 = []\n for i in range(self.layers_num):\n child1.append(torch.cat((parent1[i][:self.boundaries[i]], parent2[i][self.boundaries[i]:]), 0))\n child2.append(torch.cat((parent2[i][:self.boundaries[i]], parent1[i][self.boundaries[i]:]), 0))\n return child1, child2\n\n def crossover(self):\n if random.random() * 100 > 90:\n return\n\n self.boundaries = []\n for n in self.neurons_num_in_layers[1:]: # 1: потому что инпут слой не учитывается\n bound = (random.random() - 0.5) / 2 + 0.5\n bound = round(n * bound)\n self.boundaries.append(bound)\n\n self.genomes[-1], self.genomes[-2] = self.create_children_genomes(self.genomes[0], self.genomes[1])\n self.genomes[-3], self.genomes[-4] = self.create_children_genomes(self.genomes[0], self.genomes[2])\n self.genomes[-5], self.genomes[-6] = self.create_children_genomes(self.genomes[1], self.genomes[2])\n\n # boundary1 = (random.random() - 0.5) / 2 + 0.5\n # boundary1 = round(self.n_hidden * boundary1)\n # boundary2 = (random.random() - 0.5) / 2 + 0.5\n # boundary2 = round(self.n_hidden * boundary2)\n # boundary3 = (random.random() - 0.5) / 2 + 0.5\n # boundary3 = round(self.n_output * boundary3)\n #\n # # new_genomexy: x - net layer number, y - child number\n # new_genome11 = torch.cat((self.genomes[0][0][:boundary1], self.genomes[1][0][boundary1:]), 0)\n # new_genome21 = torch.cat((self.genomes[0][1][:boundary2], self.genomes[1][1][boundary2:]), 0)\n # new_genome31 = torch.cat((self.genomes[0][2][:boundary3], self.genomes[1][2][boundary3:]), 0)\n #\n # new_genome12 = torch.cat((self.genomes[1][0][:boundary1], self.genomes[0][0][boundary1:]), 0)\n # new_genome22 = torch.cat((self.genomes[1][1][:boundary2], self.genomes[0][1][boundary2:]), 0)\n #\n # new_genome13 = torch.cat((self.genomes[0][0][:boundary1], self.genomes[2][0][boundary1:]), 0)\n # new_genome23 = torch.cat((self.genomes[0][1][:boundary2], self.genomes[2][1][boundary2:]), 0)\n #\n # new_genome14 = torch.cat((self.genomes[2][0][:boundary1], self.genomes[0][0][boundary1:]), 0)\n # new_genome24 = torch.cat((self.genomes[2][1][:boundary2], self.genomes[0][1][boundary2:]), 0)\n #\n # new_genome15 = torch.cat((self.genomes[1][0][:boundary1], self.genomes[2][0][boundary1:]), 0)\n # new_genome25 = torch.cat((self.genomes[1][1][:boundary2], self.genomes[2][1][boundary2:]), 0)\n #\n # new_genome16 = torch.cat((self.genomes[2][0][:boundary1], self.genomes[1][0][boundary1:]), 0)\n # new_genome26 = torch.cat((self.genomes[2][1][:boundary2], self.genomes[1][1][boundary2:]), 0)\n #\n # self.genomes[-1] = [new_genome11, new_genome21]\n # self.genomes[-2] = [new_genome12, new_genome22]\n # self.genomes[-3] = [new_genome13, new_genome23]\n # self.genomes[-4] = [new_genome14, new_genome24]\n # self.genomes[-5] = [new_genome15, new_genome25]\n # self.genomes[-6] = [new_genome16, new_genome26]\n\n def mutation(self):\n for genome in self.genomes:\n for i in range(self.layers_num):\n self.mutate_genome(genome[i])\n\n def mutate_genome(self, genome):\n for i in range(genome.shape[0]):\n max_gen = genome[i].max()\n min_gen = genome[i].min()\n for j in range(genome.shape[1]):\n if random.random() * 100 < self.mutation_prob:\n genome[i][j] = self.mutation_fun2(max_gen, min_gen, genome[i][j])\n\n def mutation_fun1(self, max_gen, min_gen, gen):\n return min_gen + max_gen - gen\n\n def mutation_fun2(self, max_gen, min_gen, gen):\n span = float(max_gen - min_gen)\n delta_gen = span / 5\n r = random.random()\n if random.random() > 0.5:\n if gen + delta_gen < max_gen:\n return gen + r * delta_gen\n else:\n return gen - r * delta_gen\n else:\n if gen - delta_gen > min_gen:\n return gen - r * delta_gen\n else:\n return gen + r * delta_gen\n\n def refresh_bots(self):\n for i in range(len(self.bots)):\n weights = []\n biases = []\n for j in range(self.layers_num):\n weights.append(self.genomes[i][j][:, :-1])\n biases.append(self.genomes[i][j][:, -1])\n self.bots[i].set_net_parameters(weights, biases)\n\n def get_bots(self):\n return self.bots\n", "sub_path": "genetic.py", "file_name": "genetic.py", "file_ext": "py", "file_size_in_byte": 5908, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "torch.cat", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 42, "usage_type": "call"}, {"api_name": "random.random", "line_number": 46, "usage_type": "call"}, {"api_name": "random.random", "line_number": 51, "usage_type": "call"}, {"api_name": "random.random", "line_number": 103, "usage_type": "call"}, {"api_name": "random.random", "line_number": 112, "usage_type": "call"}, {"api_name": "random.random", "line_number": 113, "usage_type": "call"}]} +{"seq_id": "372621989", "text": "# Advent of Code - Day 14 - Part One\n\nfrom typing import Dict, List\n\n\nclass Polymerizer:\n template: str\n rules: Dict[str, str]\n polymer: str\n\n def __init__(self, input: List[str]) -> None:\n self.template = input[0]\n self.rules = {}\n for ln in input[2:]:\n src, el = ln.split(\" -> \")\n self.rules[src] = el\n self.polymer = self.template\n\n def step(self) -> None:\n buf = list(self.polymer)\n cur = 1\n while cur < len(buf):\n pair = buf[cur-1] + buf[cur]\n buf.insert(cur, self.rules[pair])\n cur += 2\n self.polymer = \"\".join(buf)\n\n\ndef result(input):\n polymerizer = Polymerizer(input)\n for _ in range(10):\n polymerizer.step()\n\n elements = {}\n for el in polymerizer.polymer:\n if el not in elements.keys():\n elements[el] = 0\n elements[el] += 1\n\n most = max(elements.values())\n least = min(elements.values())\n return most - least\n", "sub_path": "2021/aoc/day14/part1.py", "file_name": "part1.py", "file_ext": "py", "file_size_in_byte": 998, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "typing.Dict", "line_number": 8, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 11, "usage_type": "name"}]} +{"seq_id": "102201462", "text": "from utility import *\nimport json\n\ndef createGetScript(endpoint, params):\n script = 'curl '+start_url+endpoint+'?'\n keys = params.keys()\n values = params.values()\n pair = [keys[i]+'='+values[i] for i in range(len(keys))]\n evil_param = '&'.join(pair)\n script+=evil_param\n return script\n\ndef createPostScript(endpoint, params):\n keys = params.keys()\n values = params.values()\n pair = [keys[i]+'='+values[i] for i in range(len(keys))]\n evil_param = '&'.join(pair)\n script = 'curl -d ' + '\"'+ evil_param +'\" '+'-X POST '+start_url+endpoint\n return script\n\ndef genDT(endpoint,params,method):\n scope = {\n 'class':DT,\n 'results':{\n start_url: [\n {\n 'endpoint': endpoint,\n 'params': params,\n 'method': method\n }\n ]\n }\n } \n\n script = ''\n if method == 'GET':\n script = createGetScript(endpoint, params)\n else:\n script = createPostScript(endpoint, params)\n \n return scope, script\n\ndef genSI(endpoint, params, method):\n scope = {\n 'class':SI,\n 'results':{\n start_url: [\n {\n 'endpoint': endpoint,\n 'params': params,\n 'method': method\n }\n ]\n }\n }\n \n script = ''\n if method == 'POST':\n script = createPostScript(endpoint,params)\n else:\n script = createGetScript(endpoint,params)\n \n return scope, script\n\ndef genSCI(endpoint, params, method):\n scope = {\n 'class':SCI,\n 'results':{\n start_url: [\n {\n 'endpoint': endpoint,\n 'params': params,\n 'method': method\n }\n ]\n }\n }\n\n script = ''\n if method == 'POST':\n script = createPostScript(endpoint,params)\n else:\n script = createGetScript(endpoint,params)\n \n return scope, script\n\ndef genSSCI():\n pass\n\ndef genCSRF():\n pass\n\ndef genOR(endpoint, params, method):\n scope = {\n 'class':OR,\n 'results':{\n start_url: [\n {\n 'endpoint': endpoint,\n 'params': params,\n 'method': method\n }\n ]\n }\n } \n\n script = ''\n if method == 'GET':\n script = createGetScript(endpoint, params)\n else:\n script = createPostScript(endpoint, params)\n return scope, script \n\nrender = {\n DT: genDT,\n SI: genSI,\n CSRF: genCSRF,\n OR: genOR,\n SSCI: genSSCI,\n SCI: genSCI\n}\n\nclass generator(object):\n def __init__(self,category):\n self.scope = {}\n self.category = category\n self.cate_str = '_'.join(category.split(' '))\n self.count = 0\n \n def updateScope(self,scope):\n if(self.count):\n self.scope['results'][start_url]+=scope['results'][start_url]\n else:\n self.scope=scope\n self.count += 1\n \n def saveScript(self,script):\n script_name = 'result/'+self.cate_str+'_attack'+str(self.count)+'.sh'\n with open(script_name, 'w') as f:\n f.write(script)\n\n def saveScope(self):\n file_name = 'result/'+self.cate_str+'_scope.json'\n with open(file_name,'w+') as f:\n json.dump(self.scope,f)\n\n", "sub_path": "benchmarkSpider/benchmarkSpider/generator.py", "file_name": "generator.py", "file_ext": "py", "file_size_in_byte": 3452, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "json.dump", "line_number": 145, "usage_type": "call"}]} +{"seq_id": "256704570", "text": "import logging\n\nfrom pylons import config, request, response, session, tmpl_context as c\n\nfrom cyberweb import model\nfrom cyberweb.model import meta\n\nfrom datetime import datetime\nfrom time import sleep\n\nfrom threading import Thread\nimport paramiko\n\nimport jodis.sshresource as myssh\nfrom time import sleep\n\n\nlog = logging.getLogger(__name__)\n\nclass Gccom:\n # Eventually, this is pulled from the database\n gccom_comm_acct = ''\n gcem_dir = 'GCEMproj'\n gccom_dir = 'gccom'\n\n #set up names of and parameter data. configuration status == 0/1 == y/n\n #note: eventually will use this to list all grids, select one, and print an image of the grid.\n # note: for temps (ldc1, ldc2), boxgrid.dat and grid.dat are same.\n bath_grid = {\n 'hdrs' : { 'name':'Grid Name','IMax':0,'JMax':0,'KMax':0,'img':'Image File','desc':'Description',\n 'fname':'filename'},\n 'cube-box' : { 'name':'Cube Box','IMax':32,'JMax':32,'KMax':32,'img':'gridbox',\n 'desc':'Cube Grid Box for test cases tests',\n 'fname':'grid.32x32x32.dat'},\n 'long-box' : { 'name': 'Long Box','IMax':33,'JMax':33,'KMax':97,'img':'long-box',\n 'desc':'Long Box for test cases tests',\n 'fname':'longboxgrid.33x33x97.dat'},\n 'seamount' : { 'name':'Simple Seamount','IMax':97,'JMax':33,'KMax':33,'img':'seamount',\n 'desc':'Simple Seamount description',\n 'fname':'Seamount.97x33x33.dat'},\n 'channel' : { 'name':'Channel','IMax':0,'JMax':0,'KMax':0,'img':'channel',\n 'desc':'Channel Box for test cases tests',\n 'fname':'filename'},\n 'montbay' : { 'name':'MontereyBay','IMax':0,'JMax':0,'KMax':0,'img':'image',\n 'desc':'description',\n 'fname':'filename'},\n }\n ## note: there is aproblem here in that i am hardcoding elements in exec code. \n ## need to move to looking at header, mathching index of key and then using that.\n model_info = {\n 'hdrs' : { 'desc' :'Description', \n 'st' :'status',\n 'modeldir' :'execdir',\n 'modelname' :'modelname',\n 'grid_key' :'grid_key' \n },\n 'ldc1' : { 'desc' :'Lid Driven Cavity Test Case 1', \n 'st' : 0, \n 'modeldir' :'cwproj/gccom/ldc1',\n 'modelname' :'gccom.ldc1',\n 'grid_key' :'cube-box' \n },\n 'ldc2' : { 'desc' :'Lid Driven Cavity Test Case 2', \n 'st' : 1, \n 'modeldir' :'cwproj/gccom/ldc2',\n 'modelname' :'gccom.ldc2',\n 'grid_key' :'cube-box' \n },\n 'temperature1' : { 'desc' :'Temperature Test Case 1', \n 'st' : 1 , \n 'modeldir' :'cwproj/gccom/temperature1',\n 'modelname' :'gccom.temperature1',\n 'grid_key' :'cube-box' \n },\n 'temperature2' : { 'desc' :'Temperature Test Case 2', \n 'st' : 1, \n 'modeldir' :'cwproj/gccom/temperature2',\n 'modelname' :'gccom.temperature2' ,\n 'grid_key' :'long-box' \n },\n 'seamount1' : { 'desc' :'Simple Seamount Test Case 1', \n 'st' : 1, \n 'modeldir' :'cwproj/gccom/seamount1',\n 'modelname' :'gccom.seamount1' ,\n 'grid_key' :'seamount' \n },\n } \n model_params = {\n 'hdrs' : [ 'Parameter','Description','Value' ],\n 'ldc2' : [\n ['IterM','Max Iteration','100.000'],\n ['dt','Time Step (dt)','0.0001'],\n ['MaxFileNo','Max File No','10.000'],\n ['wrthz','Writeout freq.','10.00'],\n ['omp','Omega Pres.','1.0'],\n ['epsp','eps. Pres.','0.0000001'],\n ['itemp','SOR Max Iter P.','5000.0'],\n ['Re','Reynolds Number','1000.0'],\n ['PrT','PrT Number','5.0'],\n ['PrS','PrS Number','5.0'],\n ['Ros','Rossby Number','0.0'],\n ['Fr','Froud Number','0.003'],\n ['UStar','Velocity Scale','0.1'],\n ['LStar','Length Scale','1.0'],\n ['TStar','Temp. Scale','40.0'],\n ['SStar','Salinity Scale','42.0'],\n\n ],\n 'seamount1' : [\n ['IterM','Max Iteration','100.000'],\n ['dt','Time Step (dt)','0.0001'],\n ['MaxFileNo','Max File No','10.000'],\n ['wrthz','Writeout freq.','10.00'],\n ['omp','Omega Pres.','1.0'],\n ['epsp','eps. Pres.','0.0000001'],\n ['itemp','SOR Max Iter P.','5000.0'],\n ['Re','Reynolds Number','1000.0'],\n ['PrT','PrT Number','5.0'],\n ['PrS','PrS Number','0.0'],\n ['Ros','Rossby Number','0.0'],\n ['Fr','Froud Number','0.00'],\n ['UStar','Velocity Scale','0.1'],\n ['LStar','Length Scale','1.0'],\n ['TStar','Temp. Scale','10.0'],\n ['SStar','Salinity Scale','42.0'],\n ],\n 'temperature1' : [\n ['IterM','Max Iteration','100.000'],\n ['dt','Time Step (dt]','0.0001'],\n ['MaxFileNo','Max File No','10.000'],\n ['wrthz','Writeout freq.','10.00'],\n ['omp','Omega Pres.','1.0'],\n ['epsp','eps. Pres.','0.0000001'],\n ['itemp','SOR Max Iter P.','1000.0'],\n ['Re','Reynolds Number','1000.0'],\n ['PrT','PrT Number','5.0'],\n ['PrS','PrS Number','5.0'],\n ['Ros','Rossby Number','0.0'],\n ['Fr','Froud Number','0.003'],\n ['UStar','Velocity Scale','0.1'],\n ['LStar','Length Scale','1.0'],\n ['TStar','Temp. Scale','10.0'],\n ['SStar','Salinity Scale','42.0'],\n ],\n 'temperature2' : [\n ['IterM','Max Iteration','100.000'],\n ['dt','Time Step (dt]','0.0001'],\n ['MaxFileNo','Max File No','10.000'],\n ['wrthz','Writeout freq.','10.00'],\n ['omp','Omega Pres.','1.0'],\n ['epsp','eps. Pres.','0.0000001'],\n ['itemp','SOR Max Iter P.','5000.0'],\n ['Re','Reynolds Number','1000.0'],\n ['PrT','PrT Number','5.0'],\n ['PrS','PrS Number','0.0'],\n ['Ros','Rossby Number','0.0'],\n ['Fr','Froud Number','1.00'],\n ['UStar','Velocity Scale','0.1'],\n ['LStar','Length Scale','1.0'],\n ['TStar','Temp. Scale','10.0'],\n ['SStar','Salinity Scale','42.0'],\n ],\n }\n\n def __init__(self):\n # Resource list will eventually be pulled from database\n # @todo: Need to define the parameter list and grid files for each application.\n iccom_comm_acct = 'mary'\n", "sub_path": "build/lib.linux-x86_64-2.7/cyberweb/lib/gccom.py", "file_name": "gccom.py", "file_ext": "py", "file_size_in_byte": 8187, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "logging.getLogger", "line_number": 18, "usage_type": "call"}]} +{"seq_id": "114102717", "text": "# -*- coding=utf8 -*-\n\n#\n# 微服务\n#\n\nimport threading\nimport socket\nimport json\nimport selectors\nimport queue\n\nclass ServiceServer:\n def __init__(self,appName,port,servercenter):\n self.redata = {}\n self.redata['port'] = port\n self.redata['app'] = appName\n self.port = port\n self.servercenteraddr = servercenter\n self.workerQueue = queue.Queue(1024)\n self.sel = selectors.DefaultSelector()\n self.registerServer()\n self.initThreadPool()\n self.tcp()\n\n def registerServer(self):\n sd = socket.socket(socket.AF_INET,socket.SOCK_DGRAM)\n print(self.port)\n sd.sendto(json.dumps(self.redata).encode('utf8'),self.servercenteraddr)\n print(sd.recv(1024))\n sd.close()\n threading.Timer(3,self.registerServer).start()\n\n def tcp(self):\n sd = socket.socket(socket.AF_INET,socket.SOCK_STREAM)\n sd.setsockopt(socket.SOL_SOCKET,socket.SO_REUSEPORT,1)\n sd.setsockopt(socket.SOL_SOCKET,socket.SO_REUSEADDR,1)\n sd.bind(('0.0.0.0',self.port))\n sd.listen(-1)\n sd.setblocking(False)\n self.sel.register(sd,selectors.EVENT_READ,self.accept)\n while True:\n events = self.sel.select()\n for key,mask in events:\n print(key)\n func = key.data\n func(key.fileobj,mask)\n\n def read(self,sock:socket.socket,mask):\n data = sock.recv(4096)\n print(data)\n if data is None:\n self.sel.unregister(sock)\n else:\n self.workerQueue.put((data.decode('utf8') ,sock))\n\n def accept(self,sock:socket.socket,mask):\n con,addr = sock.accept()\n con.setblocking(False)\n self.sel.register(con,selectors.EVENT_READ,self.read)\n\n\n def hadel(self,data,con:socket.socket):\n con.send(json.dumps({\"a\":\"b\",\"list\" : [1,2,3,4]}).encode('utf8'))\n\n def initThreadPool(self):\n for i in range(20):\n t = threading.Thread(target=self.run,args=(self.workerQueue,))\n t.start()\n\n\n def run(self,q:queue.Queue):\n while True:\n msg,con = q.get()\n try:\n self.hadel(json.loads(msg),con)\n except Exception as e:\n print(e)\n q.task_done()\n\nif __name__ == '__main__':\n ServiceServer('AppName',9004,('127.0.0.1',9003))", "sub_path": "ServiceServer.py", "file_name": "ServiceServer.py", "file_ext": "py", "file_size_in_byte": 2378, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "queue.Queue", "line_number": 20, "usage_type": "call"}, {"api_name": "selectors.DefaultSelector", "line_number": 21, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 27, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 27, "usage_type": "attribute"}, {"api_name": "socket.SOCK_DGRAM", "line_number": 27, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 29, "usage_type": "call"}, {"api_name": "threading.Timer", "line_number": 32, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 35, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 35, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 35, "usage_type": "attribute"}, {"api_name": "socket.SOL_SOCKET", "line_number": 36, "usage_type": "attribute"}, {"api_name": "socket.SO_REUSEPORT", "line_number": 36, "usage_type": "attribute"}, {"api_name": "socket.SOL_SOCKET", "line_number": 37, "usage_type": "attribute"}, {"api_name": "socket.SO_REUSEADDR", "line_number": 37, "usage_type": "attribute"}, {"api_name": "selectors.EVENT_READ", "line_number": 41, "usage_type": "attribute"}, {"api_name": "socket.socket", "line_number": 49, "usage_type": "attribute"}, {"api_name": "socket.socket", "line_number": 57, "usage_type": "attribute"}, {"api_name": "selectors.EVENT_READ", "line_number": 60, "usage_type": "attribute"}, {"api_name": "socket.socket", "line_number": 63, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 64, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 68, "usage_type": "call"}, {"api_name": "queue.Queue", "line_number": 72, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 76, "usage_type": "call"}]} +{"seq_id": "35060634", "text": "# Copyright 2019 StreamSets Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport logging\nimport os\n\nfrom streamsets.testframework.markers import azure, sdc_min_version\nfrom streamsets.testframework.utils import get_random_string\n\n\nlogger = logging.getLogger(__name__)\nlogger.setLevel(logging.DEBUG)\n\nADLS_GEN1_ORIGIN = 'com_streamsets_pipeline_stage_origin_datalake_gen1_DataLakeDSource'\n\n\n@azure('datalake')\n@sdc_min_version('3.9.0')\ndef test_initial_scan(sdc_builder, sdc_executor, azure, benchmark):\n \"\"\"Performance test for ADLS Gen1 origin.\n\n The test populates a random ADLS folder with a total of 1000 random files evenly distributed in 100\n subdirectories. Then it is measured the time a pipeline takes to recursively scan this directory tree and\n populate the internal queue of files found. This corresponds to the time the pipeline takes to reach the\n RUNNING state.\n\n Pipeline: adls_orig >> trash\n\n \"\"\"\n directory_name = f'/stf_perf_{get_random_string()}'\n fs = azure.datalake.file_system\n\n try:\n # Build the origin pipeline\n builder = sdc_builder.get_pipeline_builder()\n trash = builder.add_stage('Trash')\n adls_orig = builder.add_stage(name=ADLS_GEN1_ORIGIN)\n adls_orig.set_attributes(data_format='TEXT',\n files_directory=directory_name,\n file_name_pattern='*',\n read_order='TIMESTAMP',\n process_subdirectories=True)\n adls_orig >> trash\n pipeline = builder.build().configure_for_environment(azure)\n\n # Populate the Azure directory employed in the test.\n _adls_populate_dir(fs, directory_name, num_dirs=100, num_files=10)\n\n def benchmark_pipeline(executor, pipeline):\n executor.add_pipeline(pipeline)\n executor.start_pipeline(pipeline).wait_for_status('RUNNING', timeout_sec=1200)\n executor.stop_pipeline(pipeline).wait_for_stopped()\n executor.remove_pipeline(pipeline)\n\n benchmark.pedantic(benchmark_pipeline, args=(sdc_executor, pipeline), rounds=5)\n\n finally:\n logger.info('Azure Data Lake directory %s and underlying files will be deleted.', directory_name)\n if fs.exists(directory_name):\n fs.rm(directory_name, recursive=True)\n\n\ndef _adls_populate_dir(adls_client, path, num_dirs=100, num_files=10):\n \"\"\"Populate a directory with random subdirectories and files. If `path` does not exists, it is created by the\n function. Then it is populated with `num_dir` subdirectories, each one containing `num_files` random\n files. The content of each file is just its path.\n\n \"\"\"\n if not adls_client.exists(path):\n adls_client.mkdir(path)\n\n for _ in range(num_dirs):\n folder_name = get_random_string(length=10)\n for _ in range(num_files):\n file_name = \"{}.txt\".format(get_random_string(length=10))\n file_path = os.path.join(path, folder_name, file_name)\n try:\n logger.info(\"Creating new file: %s...\", file_path)\n _adls_create_file(adls_client, file_path, file_path)\n except Exception as e:\n logger.error(\"Could not create blob: %s: %s\", file_path, str(e))\n\n\ndef _adls_create_file(adls_client, file_content, file_path):\n \"\"\"Create a file in ADLS with the specified content. If the file already exist, it is truncated before\n writing `file_content`.\n\n \"\"\"\n tmp_file = 'tmp.txt'\n try:\n with open(tmp_file, 'w') as f:\n f.write(f'{file_content}')\n adls_client.put(tmp_file, file_path)\n except Exception as e:\n raise RuntimeError(f'Could not create file: {file_path}, {str(e)}')\n\n os.remove(tmp_file)\n", "sub_path": "performance/test_adls_gen1_origin.py", "file_name": "test_adls_gen1_origin.py", "file_ext": "py", "file_size_in_byte": 4297, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "logging.getLogger", "line_number": 22, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 23, "usage_type": "attribute"}, {"api_name": "streamsets.testframework.utils.get_random_string", "line_number": 41, "usage_type": "call"}, {"api_name": "streamsets.testframework.markers.azure.datalake", "line_number": 42, "usage_type": "attribute"}, {"api_name": "streamsets.testframework.markers.azure", "line_number": 42, "usage_type": "name"}, {"api_name": "streamsets.testframework.markers.azure", "line_number": 55, "usage_type": "argument"}, {"api_name": "streamsets.testframework.markers.azure", "line_number": 28, "usage_type": "call"}, {"api_name": "streamsets.testframework.markers.sdc_min_version", "line_number": 29, "usage_type": "call"}, {"api_name": "streamsets.testframework.utils.get_random_string", "line_number": 84, "usage_type": "call"}, {"api_name": "streamsets.testframework.utils.get_random_string", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path", "line_number": 87, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 108, "usage_type": "call"}]} +{"seq_id": "620743621", "text": "from collections import defaultdict\nfrom itertools import combinations\nfrom prime import Prime\n\n\ndef ans():\n groups = defaultdict(set)\n for prime in Prime.gen_nums(10000):\n groups[''.join(sorted(str(prime)))].add(prime)\n for set_ in groups.values():\n for comb in combinations(set_, 3):\n seq = sorted(list(comb))\n if seq[0] < 1000:\n continue\n if (\n seq[2] + seq[0] == 2 * seq[1] and\n seq[0] != 1487\n ):\n return ''.join(str(n) for n in seq)\n \n\nif __name__ == '__main__':\n print(ans())\n", "sub_path": "src/049.py", "file_name": "049.py", "file_ext": "py", "file_size_in_byte": 616, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "collections.defaultdict", "line_number": 7, "usage_type": "call"}, {"api_name": "prime.Prime.gen_nums", "line_number": 8, "usage_type": "call"}, {"api_name": "prime.Prime", "line_number": 8, "usage_type": "name"}, {"api_name": "itertools.combinations", "line_number": 11, "usage_type": "call"}]} +{"seq_id": "281497528", "text": "import numpy as np\nfrom skimage.color import rgb2hsv\nfrom skimage.exposure import rescale_intensity\nfrom scipy.ndimage import convolve\n\n# kernel for final convolution\ndisc = np.array([\n [0, 0, 1, 0, 0],\n [1, 1, 1, 1, 1],\n [1, 1, 1, 1, 1],\n [1, 1, 1, 1, 1],\n [0, 0, 1, 0, 0]], dtype=np.uint8)\n\n\ndef normalize_hsv(hsv):\n \"\"\"It extends the Hue value to 0-179 and S&V values to 0-255\n\n Parameters\n ----------\n hsv : HSV image array\n HSV image array obtained as result of color.rgb2hsv()\n\n Returns\n -------\n out : HSV image array\n It is an uint8 image array in HSV format\n\n \"\"\"\n hsv[:, :, 0] = hsv[:, :, 0]*179\n hsv[:, :, 1:] = hsv[:, :, 1:]*255\n return hsv.astype('uint8')\n\n\ndef histogram_backproject(img1, img2):\n \"\"\"Return the image after backprojection of img2 on img1.\n\n Parameters\n ----------\n img1 : array\n Image array on which img2 is backprojected\n img2 : array\n Image array whose histogram is backprojected\n\n Returns\n -------\n out : Image array\n Single channel image array\n\n References\n ----------\n .. [1] \"Indexing via color histograms\", M.J.Swain & D.H.Ballard, IEEE, 1990\n\n \"\"\"\n\n # Both image should be of uint8 dtype\n assert (img1.dtype == np.uint8 and img2.dtype == np.uint8),\\\n \" both images should be of np.uint8 dtype \"\n\n shape1, shape2 = img1.shape, img2.shape\n\n # Both images should be single or 3-channel\n assert len(shape1) == len(shape2),\\\n \"both images should be 1-channel or 3-channel\"\n\n # for grayscale image take 1D histogram of intensity values\n if len(shape1) < 3:\n\n # find histograms\n hist1 = np.bincount(img1.ravel(), minlength=256)\n hist2 = np.bincount(img2.ravel(), minlength=256)\n\n # find their ratio hist2/hist1\n R = np.float64(hist2) / (hist1 + 1)\n\n # Now apply this ratio as the palette to original image, img1\n B = R[img1]\n B = np.minimum(B, 1)\n B = rescale_intensity(B, out_range=(0, 255))\n B = np.uint8(B)\n B = convolve(B, disc)\n return B\n\n # if color image, take 2D histogram\n else:\n # convert images to hsv plane\n hsv_img1 = rgb2hsv(img1)\n hsv_img2 = rgb2hsv(img2)\n hsv_img1 = normalize_hsv(hsv_img1)\n hsv_img2 = normalize_hsv(hsv_img2)\n\n # find their color 2D histograms\n h1, s1, v1 = np.dsplit(hsv_img1, (1, 2))\n hist1, _, _ = np.histogram2d(h1.ravel(), s1.ravel(),\n [180, 256], [[0, 180], [0, 256]])\n h2, s2, v2 = np.dsplit(hsv_img2, (1, 2))\n hist2, _, _ = np.histogram2d(h2.ravel(), s2.ravel(),\n [180, 256], [[0, 180], [0, 256]])\n\n # find their ratio hist2/hist1\n R = np.float64(hist2) / (hist1 + 1)\n\n # backproject\n B = R[h1.ravel(), s1.ravel()]\n B = np.minimum(B, 1)\n B = B.reshape(img1.shape[:2])\n B = rescale_intensity(B, out_range=(0, 255))\n B = np.uint8(B)\n B = convolve(B, disc)\n return B\n", "sub_path": "scikit_roughworks/backproject2.py", "file_name": "backproject2.py", "file_ext": "py", "file_size_in_byte": 3153, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "numpy.array", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 12, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 56, "usage_type": "attribute"}, {"api_name": "numpy.bincount", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.bincount", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 77, "usage_type": "call"}, {"api_name": "skimage.exposure.rescale_intensity", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 79, "usage_type": "call"}, {"api_name": "scipy.ndimage.convolve", "line_number": 80, "usage_type": "call"}, {"api_name": "skimage.color.rgb2hsv", "line_number": 86, "usage_type": "call"}, {"api_name": "skimage.color.rgb2hsv", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.dsplit", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.histogram2d", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.dsplit", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.histogram2d", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 104, "usage_type": "call"}, {"api_name": "skimage.exposure.rescale_intensity", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 107, "usage_type": "call"}, {"api_name": "scipy.ndimage.convolve", "line_number": 108, "usage_type": "call"}]} +{"seq_id": "651930063", "text": "from __future__ import absolute_import, print_function\n\n# Import modules\nfrom tweepy.streaming import StreamListener\nfrom tweepy import OAuthHandler\nfrom tweepy import Stream\n\n# Ben Oostendorp's Twitter credentials\nconsumer_key = \"PAUelIu5JVwCzcbuAIn7lD9dM\"\nconsumer_secret = \"gWeFWtvsvt6U3OEkypUqYTzzyNwy4t3BYrarLKXOBRC6HY5s7l\"\naccess_token = \"1056813780238254080-b2XfWWnhRlmbAWNanuOWxfMxCE3paM\"\naccess_token_secret = \"Dx2wl2sKOosNj25lh1I93Z2OVOklpOHIc0mydvM7VdFad\"\n\n\nmin_follower_count = 10\nmin_keyword_count = 2\n\n\nclass wrapper(StreamListener):\n\n def on_status(self, status):\n\n # vars not being used right now:\n # id_str = status.id_str\n # created = status.created_at\n # text = status.text\n # fav = status.favorite_count\n # name = status.user.screen_name\n # description = status.user.description\n # loc = status.user.location\n # user_created = status.user.created_at\n\n followers = status.user.followers_count\n\n search_term = keywords[0]\n\n print(\"checking tweets for: \")\n print(search_term + \" . . . \")\n\n if followers < min_follower_count:\n return\n if search_term not in status.text:\n return\n\n keyword_count = 0\n\n for keyword in keywords:\n if keyword in status.text:\n keyword_count = keyword_count + 1\n print(\"keyword #\")\n print(keyword_count)\n\n\n if keyword_count >= min_keyword_count:\n print(\"criteria met!\")\n print(status.text)\n exit(1)\n return status.text\n\n\n\nclass search:\n def __init__(self, keywords):\n\n #setup keys\n wrapper_instance = wrapper()\n twitter_setup()\n\n # tweet stream will not be open that long, because it will return the text from\n # the most first notable tweet based on search term. Notable tweets decided from follower count of poster,\n # and number of key word hits.\n tweet_stream = Stream(twitter_setup(), wrapper_instance)\n return tweet_stream.filter(track=keywords)\n\n\n\n\n\n\ndef twitter_setup():\n auth = OAuthHandler(consumer_key, consumer_secret)\n auth.set_access_token(access_token, access_token_secret)\n return auth\n\nif __name__ == '__main__':\n keywords = [\"Boeing\", \"747\", \"saftey\", \"issue\"]\n search(keywords)\n", "sub_path": "Legacy/twitter_wrapper.py", "file_name": "twitter_wrapper.py", "file_ext": "py", "file_size_in_byte": 2361, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "tweepy.streaming.StreamListener", "line_number": 19, "usage_type": "name"}, {"api_name": "tweepy.Stream", "line_number": 72, "usage_type": "call"}, {"api_name": "tweepy.OAuthHandler", "line_number": 81, "usage_type": "call"}]} +{"seq_id": "372782730", "text": "from iexfinance import *\nfrom datetime import *\nimport matplotlib.pyplot as plt\nimport requests\nimport pandas\nimport json\nimport algorithm\nimport random\nimport numpy\nimport copy\n\n\nwhile True:\n\n print(\"****************************************\")\n inputstock = input(\"Enter your ticker symbol: \" )\n stock = inputstock.upper()\n\n\n today = date.today()\n lastweek = date.today() - timedelta(7)\n\n #print(lastweek)\n print(today)\n\n lastmonth = date.today() - timedelta(30)\n lastyear = date.today() - timedelta(365)\n\n period = input(\"Enter time period (Week, Month, or Year): \")\n if (period.lower() == \"week\"):\n start = lastweek\n p = 7\n if (period.lower() == \"month\"):\n start = lastmonth\n p = 30\n if (period.lower() == \"year\"):\n start = lastyear\n p = 365\n end = today\n\n #dtstart = datetime(2015, 1, 1)\n\n data = get_historical_data(stock, date.today() - timedelta(1825), end, output_format='pandas')\n #print(data)\n #a = type(data)\n #print(a)\n\n data = data.values.tolist()\n datalist = []\n for i in range(len(data[0])):\n datalist.append([])\n for i in range(len(datalist)):\n for j in range(len(data)):\n datalist[i].append(0)\n for i in range(len(datalist)):\n for j in range(len(datalist[i])):\n datalist[i][j]=data[j][i]\n #print(datalist[3])\n \n predictiondata = algorithm.projections(datalist[3], p)\n #print(datalist)\n #print(predictiondata)\n\n i = 0\n k = 0\n \n #This code is only needed if data implemented has days the market is closed\n '''\n for i in range(len(datalist[3])):\n j = (date.today() - timedelta(i)).weekday()\n if j == 5 or j == 6:\n #print(i)\n datalist[3][i]=-1\n k+=1\n \n for j in range(k):\n datalist[3].remove(-1)\n '''\n histset = datalist[3]\n normset = []\n bullset = []\n bearset = []\n \n for i in range(len(histset) - 1):\n normset.append(numpy.nan)\n bullset.append(numpy.nan)\n bearset.append(numpy.nan)\n\n normset.append(histset[len(histset)-1])\n bullset.append(histset[len(histset)-1])\n bearset.append(histset[len(histset)-1])\n \n #print(datalist[3])\n normset = normset + predictiondata[0]\n bullset = bullset + predictiondata[1]\n bearset = bearset + predictiondata[2]\n \n plt.plot(histset)\n plt.plot(normset)\n plt.plot(bullset)\n plt.plot(bearset)\n\n plt.title('Time Series Chart For ' + (stock))\n plt.xlim(len(histset) - p, len(histset) + p)\n plt.show()\n \n userquit = input(\"Type quit to exit the program or press enter to continue\")\n if userquit.lower() == (\"quit\"):\n break\n\nexit()\n\n#The API iexfinance was used in this piece of code\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2781, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "algorithm.projections", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 85, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 86, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 87, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}]} +{"seq_id": "146977504", "text": "import os\nimport time\n\nimport pandas as pd\nimport pytest\nimport tarfile\nimport shutil\nimport unittest\nimport requests\nimport threading\nfrom pyramid import testing\nfrom pyramid.paster import get_app\n\ndataDir = os.path.join('jbrowse', 'jbrowse', 'data')\ndbDir = os.path.join(dataDir, 'db')\ndbTar = os.path.join('data', 'db.tar.gz')\ntestDataPath = os.path.join('tests', 'data')\n\nif os.path.exists(dbDir):\n shutil.rmtree(dbDir)\n\nif not os.path.exists(dbDir):\n with tarfile.open(dbTar) as tar:\n tar.extractall(dataDir)\n\nfrom core.util import PLConfig as cfg\n\ncfg.testing()\nsleepTime = 600\n\nbaseUrl = 'https://peaklearner.rc.nau.edu'\n\n\n\nclass PeakLearnerTests(unittest.TestCase):\n user = 'tristanmillerschool@gmail.com'\n hub = 'H3K4me3_TDH_ENCODE'\n hubURL = '%s/%s/%s/' % (baseUrl, user, hub)\n axlTrackURL = '%s%s/' % (hubURL, 'aorta_ENCFF502AXL')\n modelSumsUrl = '%smodelSums/' % axlTrackURL\n jobsURL = '/Jobs/'\n queueUrl = '%squeue/' % jobsURL\n\n def setUp(self):\n self.config = testing.setUp()\n app = get_app('production.ini')\n from webtest import TestApp\n\n self.testapp = TestApp(app)\n\n\n def test_makeTestDataFiles(self):\n out = self.testapp.get(self.queueUrl)\n\n while out.status_code != 404:\n job = out.json\n user = job['user']\n hub = job['hub']\n track = job['track']\n taskId = job['taskId']\n jobId = job['id']\n\n params = {'ref': 'chr3', 'start': 93504854}\n\n if job['type'] == 'model':\n params['penalty'] = job['penalty']\n\n with requests.get(self.modelSumsUrl, params=params, headers={'Accept': 'application/json'}) as r:\n print(r.status_code)\n\n\n\n\n\n break\n\n assert 1 == 0\n\n", "sub_path": "tests/makeTestDataFiles.py", "file_name": "makeTestDataFiles.py", "file_ext": "py", "file_size_in_byte": 1800, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "tarfile.open", "line_number": 23, "usage_type": "call"}, {"api_name": "core.util.PLConfig.testing", "line_number": 28, "usage_type": "call"}, {"api_name": "core.util.PLConfig", "line_number": 28, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 35, "usage_type": "attribute"}, {"api_name": "pyramid.testing.setUp", "line_number": 45, "usage_type": "call"}, {"api_name": "pyramid.testing", "line_number": 45, "usage_type": "name"}, {"api_name": "pyramid.paster.get_app", "line_number": 46, "usage_type": "call"}, {"api_name": "webtest.TestApp", "line_number": 49, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 68, "usage_type": "call"}]} +{"seq_id": "360591602", "text": "from latin import LatinNoun\nfrom kivy.app import App\nfrom kivy.uix.boxlayout import BoxLayout\nfrom kivy.uix.floatlayout import FloatLayout\nfrom kivy.core.window import Window\nfrom kivy.properties import StringProperty, NumericProperty\nfrom random import shuffle\n\n\ndef shuffle_list(noun_obj):\n noun_list = noun_obj.declined_list[:]\n shuffle(noun_list)\n return noun_list\n\ndef make_parsing_round_set(noun_obj):\n list_to_shuffle = []\n poppable_set = set()\n for declined_list in noun_obj.declined_list[:]:\n list_to_shuffle.append(declined_list[2])\n shuffle(list_to_shuffle)\n for e in list_to_shuffle:\n poppable_set.add(e)\n return poppable_set\n\ndef find_all_matches(noun_selected_parsing, noun_obj):\n matches = []\n for t in noun_obj.declined_list:\n if t[2] == noun_selected_parsing:\n matches.append(t)\n return matches\n\nclass DeclensionGame(BoxLayout, FloatLayout):\n round = NumericProperty(1)\n round_name = StringProperty(\"Declining\")\n game_state = NumericProperty(0)\n word1_showing = StringProperty(\"\")\n word2_showing = StringProperty(\"\")\n word3_showing = StringProperty(\"\")\n stem1 = StringProperty(\"\")\n stem2 = StringProperty(\"\")\n stem3 = StringProperty(\"\")\n answer1 = StringProperty(\"\")\n answer2 = StringProperty(\"\")\n answer3 = StringProperty(\"\")\n\n # lexical_entry, transl_sg, transl_pl\n noun1 = [\"rosa, rosae, f., rose\", \"rose\", \"roses\"]\n noun2 = [\"amicus, amici, m., friend\", \"friend\", \"friends\"]\n noun3 = [\"labor, laboris, m., labor, work\", \"labor\", \"labors\"]\n\n # create class objects and shuffled lists\n noun1_obj = LatinNoun(*noun1)\n noun2_obj = LatinNoun(*noun2)\n noun3_obj = LatinNoun(*noun3)\n\n noun1_shuffled = shuffle_list(noun1_obj)\n noun2_shuffled = shuffle_list(noun2_obj)\n noun3_shuffled = shuffle_list(noun3_obj)\n\n # pop initial nouns\n noun1_selected = noun1_shuffled.pop()\n noun2_selected = noun2_shuffled.pop()\n noun3_selected = noun3_shuffled.pop()\n\n # things for parsing round\n noun1_set = make_parsing_round_set(noun1_obj)\n noun2_set = make_parsing_round_set(noun2_obj)\n noun3_set = make_parsing_round_set(noun3_obj)\n\n noun1_selected_parsing = noun1_set.pop()\n noun2_selected_parsing = noun2_set.pop()\n noun3_selected_parsing = noun3_set.pop()\n\n noun1_possibs = find_all_matches(noun1_selected_parsing, noun1_obj)\n noun2_possibs = find_all_matches(noun2_selected_parsing, noun2_obj)\n noun3_possibs = find_all_matches(noun3_selected_parsing, noun3_obj)\n\n def display_possibles(self, noun_possibs):\n possibilities = []\n for p in noun_possibs:\n possibilities.append(\"{} {}\\n\".format(p[0], p[1]))\n return \"\".join(possibilities)\n\n def change_round(self):\n # change round\n if self.round > 2:\n self.round = 1\n else:\n self.round += 1\n # change round name\n if self.round == 1:\n self.round_name = \"Declining\"\n elif self.round == 2:\n self.round_name = \"Parsing\"\n else:\n self.round_name = \"Translation\"\n\n def revert_game_state(self):\n if self.game_state >= 1:\n self.game_state -= 1\n self.show_words()\n\n def change_game_state(self):\n if self.game_state > 5:\n self.game_state = 1\n # get new word (Rounds 1 & 3)\n if self.round != 2:\n if len(self.noun1_shuffled) == 0:\n self.noun1_shuffled = shuffle_list(self.noun1_obj)\n if len(self.noun2_shuffled) == 0:\n self.noun2_shuffled = shuffle_list(self.noun2_obj)\n if len(self.noun3_shuffled) == 0:\n self.noun3_shuffled = shuffle_list(self.noun3_obj)\n self.noun1_selected = self.noun1_shuffled.pop()\n self.noun2_selected = self.noun2_shuffled.pop()\n self.noun3_selected = self.noun3_shuffled.pop()\n\n # get new word (Round 2)\n else:\n if len(self.noun1_set) == 0:\n self.noun1_set = make_parsing_round_set(self.noun1_obj)\n if len(self.noun2_set) == 0:\n self.noun2_set = make_parsing_round_set(self.noun2_obj)\n if len(self.noun3_set) == 0:\n self.noun3_set = make_parsing_round_set(self.noun3_obj)\n self.noun1_selected_parsing = self.noun1_set.pop()\n self.noun2_selected_parsing = self.noun2_set.pop()\n self.noun3_selected_parsing = self.noun3_set.pop()\n self.noun1_possibs = find_all_matches(self.noun1_selected_parsing, self.noun1_obj)\n self.noun2_possibs = find_all_matches(self.noun2_selected_parsing, self.noun2_obj)\n self.noun3_possibs = find_all_matches(self.noun3_selected_parsing, self.noun3_obj)\n else: # advance game state\n self.game_state += 1\n self.show_words()\n\n def show_words(self):\n # SELECT THE INDEX BASED ON ROUND\n # Round 1: Declining\n if self.round == 1:\n self.stem1 = f\"{self.noun1_selected[0]} {self.noun1_selected[1]}\"\n self.stem2 = f\"{self.noun2_selected[0]} {self.noun2_selected[1]}\"\n self.stem3 = f\"{self.noun3_selected[0]} {self.noun3_selected[1]}\"\n self.answer1 = self.noun1_selected[2]\n self.answer2 = self.noun2_selected[2]\n self.answer3 = self.noun3_selected[2]\n\n # Round 2: Parsing\n elif self.round == 2:\n self.stem1 = self.noun1_selected_parsing\n self.stem2 = self.noun2_selected_parsing\n self.stem3 = self.noun3_selected_parsing\n self.answer1 = self.display_possibles(self.noun1_possibs)\n self.answer2 = self.display_possibles(self.noun2_possibs)\n self.answer3 = self.display_possibles(self.noun3_possibs)\n\n # Round 3: Translation\n else: # self.round == 3:\n self.stem1 = f\"[size=70]{self.noun1_selected[2]}[/size], \" \\\n f\"\\n[size=50]{self.noun1_selected[0]} \" \\\n f\"{self.noun1_selected[1]}[/size]\\n\"\n self.stem2 = f\"[size=70]{self.noun2_selected[2]}[/size], \" \\\n f\"\\n[size=50]{self.noun2_selected[0]} \" \\\n f\"{self.noun2_selected[1]}[/size]\\n\"\n self.stem3 = f\"[size=70]{self.noun3_selected[2]}[/size], \" \\\n f\"\\n[size=50]{self.noun3_selected[0]} \" \\\n f\"{self.noun3_selected[1]}[/size]\\n\"\n self.answer1 = self.noun1_selected[-1]\n self.answer2 = self.noun2_selected[-1]\n self.answer3 = self.noun3_selected[-1]\n\n # SHOW WORDS\n if self.round != 2:\n if self.game_state == 0:\n self.word1_showing = \"\"\n self.word2_showing = \"\"\n self.word3_showing = \"\"\n # Show First Word\n elif self.game_state == 1:\n self.word1_showing = self.stem1\n self.word2_showing = \"\"\n self.word3_showing = \"\"\n elif self.game_state == 2:\n self.word1_showing = \"{}\\n[size=70]{}[/size]\".format(self.stem1, self.answer1)\n self.word2_showing = \"\"\n self.word3_showing = \"\"\n # Show Second Word\n elif self.game_state == 3:\n self.word2_showing = self.stem2\n self.word3_showing = \"\"\n elif self.game_state == 4:\n self.word2_showing = \"{}\\n[size=70]{}[/size]\".format(self.stem2, self.answer2)\n self.word3_showing = \"\"\n # Show Third Word\n elif self.game_state == 5:\n self.word3_showing = self.stem3\n elif self.game_state == 6:\n self.word3_showing = \"{}\\n[size=70]{}[/size]\".format(self.stem3, self.answer3)\n else:\n if self.game_state == 0:\n self.word1_showing = \"\"\n self.word2_showing = \"\"\n self.word3_showing = \"\"\n # Show First Word\n elif self.game_state == 1:\n self.word1_showing = \"[size=70]{}[/size]\".format(self.stem1)\n self.word2_showing = \"\"\n self.word3_showing = \"\"\n elif self.game_state == 2:\n self.word1_showing = \"[size=70]{}[/size]\\n{}\".format(self.stem1, self.answer1)\n self.word2_showing = \"\"\n self.word3_showing = \"\"\n # Show Second Word\n elif self.game_state == 3:\n self.word2_showing = \"[size=70]{}[/size]\".format(self.stem2)\n self.word3_showing = \"\"\n elif self.game_state == 4:\n self.word2_showing = \"[size=70]{}[/size]\\n{}\".format(self.stem2, self.answer2)\n self.word3_showing = \"\"\n # Show Third Word\n elif self.game_state == 5:\n self.word3_showing = \"[size=70]{}[/size]\".format(self.stem3)\n elif self.game_state == 6:\n self.word3_showing = \"[size=70]{}[/size]\\n{}\".format(self.stem3, self.answer3)\n\n\nclass DeclensionApp(App):\n def build(self):\n Window.bind(on_key_down=self.key_down)\n Window.fullscreen = 'auto'\n return DeclensionGame()\n\n def key_down(self, key, scancode=None, *_):\n if scancode == 281: # PAGE_DOWN\n DeclensionGame.change_game_state(self.root)\n elif scancode == 98: # B\n DeclensionGame.change_round(self.root)\n\n\nif __name__ == \"__main__\":\n DeclensionApp().run()", "sub_path": "declension.py", "file_name": "declension.py", "file_ext": "py", "file_size_in_byte": 9668, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "random.shuffle", "line_number": 12, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 20, "usage_type": "call"}, {"api_name": "kivy.uix.boxlayout.BoxLayout", "line_number": 32, "usage_type": "name"}, {"api_name": "kivy.uix.floatlayout.FloatLayout", "line_number": 32, "usage_type": "name"}, {"api_name": "kivy.properties.NumericProperty", "line_number": 33, "usage_type": "call"}, {"api_name": "kivy.properties.StringProperty", "line_number": 34, "usage_type": "call"}, {"api_name": "kivy.properties.NumericProperty", "line_number": 35, "usage_type": "call"}, {"api_name": "kivy.properties.StringProperty", "line_number": 36, "usage_type": "call"}, {"api_name": "kivy.properties.StringProperty", "line_number": 37, "usage_type": "call"}, {"api_name": "kivy.properties.StringProperty", "line_number": 38, "usage_type": "call"}, {"api_name": "kivy.properties.StringProperty", "line_number": 39, "usage_type": "call"}, {"api_name": "kivy.properties.StringProperty", "line_number": 40, "usage_type": "call"}, {"api_name": "kivy.properties.StringProperty", "line_number": 41, "usage_type": "call"}, {"api_name": "kivy.properties.StringProperty", "line_number": 42, "usage_type": "call"}, {"api_name": "kivy.properties.StringProperty", "line_number": 43, "usage_type": "call"}, {"api_name": "kivy.properties.StringProperty", "line_number": 44, "usage_type": "call"}, {"api_name": "latin.LatinNoun", "line_number": 52, "usage_type": "call"}, {"api_name": "latin.LatinNoun", "line_number": 53, "usage_type": "call"}, {"api_name": "latin.LatinNoun", "line_number": 54, "usage_type": "call"}, {"api_name": "kivy.app.App", "line_number": 226, "usage_type": "name"}, {"api_name": "kivy.core.window.Window.bind", "line_number": 228, "usage_type": "call"}, {"api_name": "kivy.core.window.Window", "line_number": 228, "usage_type": "name"}, {"api_name": "kivy.core.window.Window.fullscreen", "line_number": 229, "usage_type": "attribute"}, {"api_name": "kivy.core.window.Window", "line_number": 229, "usage_type": "name"}]} +{"seq_id": "552528084", "text": "import numpy as np\nimport pickle\nimport matplotlib.pyplot as plt\nimport sklearn.utils as sklearn\nimport sklearn.linear_model as lin\nimport bonnerlib2\t\nfrom sklearn import datasets\nfrom sklearn.neural_network import MLPClassifier\nfrom sklearn.preprocessing import StandardScaler\nfrom sklearn.utils.extmath import softmax\nfrom sklearn.utils import gen_batches\n\n########### QUESTION 1 #############\n\ncolors = np.array(['r','b'])\n\nX_train, t_train = datasets.make_moons(n_samples=200, noise=0.2)\nX_test, t_test = datasets.make_moons(n_samples=10000, noise=0.2)\n\nfig1 = plt.figure()\nplt.scatter(X_train[:, 0], X_train[:, 1], color=colors[t_train],s=2)\nfig1.suptitle(\"Figure 1, Question 1(a): Moons Training Data\")\n\nfig2 = plt.figure()\nplt.scatter(X_test[:, 0], X_test[:, 1], color=colors[t_test],s=2)\nfig2.suptitle(\"Figure 2, Question 1(a): Moons Test Data\")\n\ndef fitMoons():\n\tfig_contour = plt.figure()\n\tfig_contour.suptitle(\"Figure 3, Question 1(b): Contour plots for various training sessions\")\n\n\tclf = MLPClassifier(hidden_layer_sizes=[3],\n\t\t\t\t\tactivation='tanh',\n\t\t\t\t\tsolver='sgd',\n\t\t\t\t\tlearning_rate_init=0.01,\n\t\t\t\t\ttol=10.0**(-20),\n\t\t\t\t\tmax_iter=10000)\n\n\terrTestList = []\n\terrTestMin = np.Inf\n\terrTestCLF = np.Inf\n\tfor x in range(9):\n\t\tcurrent_clf = clf.fit(X_train, t_train)\n\t\ttesting_err = 1 - clf.score(X_test, t_test)\n\t\terrTestList.append(testing_err)\n\t\tprint(\"Testing error for training session %s = %s\" % (x+1, testing_err))\n\t\tax = fig_contour.add_subplot(3, 3, x+1)\n\t\tax.scatter(X_train[:, 0], X_train[:, 1], color=colors[t_train],s=2)\n\t\tbonnerlib2.dfContour(current_clf, ax)\n\n\t\tif testing_err < errTestMin:\n\t\t\terrTestMin = testing_err\n\t\t\terrTestCLF = current_clf\n\t\n\tfig_best = plt.figure()\n\tfig_best.suptitle(\"Figure 4, Question 1(b): Contour plot for best training session\")\t\t\n\tprint(\"Smallest test error = %s\" % (errTestMin))\n\tax = fig_best.add_subplot(1, 1, 1)\n\tax.scatter(X_train[:, 0], X_train[:, 1], color=colors[t_train],s=2)\n\tbonnerlib2.dfContour(errTestCLF, ax)\n\nfitMoons()\n\n###########\tEND OF QUESTION 1 ############\n\n########### QUESTION 3 #############\n\nwith open('mnist.pickle','rb') as f:\n\tdata = pickle.load(f)\n\ndef flatten(data):\n\tX = np.vstack(data)\n\tt = np.zeros(np.shape(X)[0], dtype='int')\n\tm1 = 0\n\tm2 = 0\n\tfor i in range(0, len(data)):\n\t\tm = np.shape(data[i])[0]\n\t\tm2 = m2 + m\n\t\tt[m1:m2] = i\n\t\tm1 = m1 + m\n\treturn X, t\n\n# Question 3(a) and (b)\nX, t = flatten(data['training'])\nX_shuffled, t_shuffled = sklearn.shuffle(X, t)\nX_test, t_test = flatten(data['testing'])\n\nscalar = StandardScaler()\nscalar.fit(X,t)\nX_normal = scalar.transform(X_shuffled)\nX_test_normal = scalar.transform(X_test)\n\ndef displaySample(N, D):\n\tarray = sklearn.resample(D, replace='False', n_samples=N)\n\tfig_digit = plt.figure()\n\tfor row in range(N):\n\t\tax = fig_digit.add_subplot(np.ceil(np.sqrt(N)), np.ceil(np.sqrt(N)), row+1)\n\t\timage = np.reshape(array[row], (28, 28))\n\t\tim = plt.imshow(image, cmap='Greys', interpolation='nearest')\n\t\tplt.axis('off')\n\treturn ax\n\n# creates matrix of target values where\n# each row is (t1, t2, ..., tk) \n# and t_k = 1 if x is in that class\nt_test_prime = np.zeros(shape=(len(t_test), 10))\nt_test_prime[np.arange(len(t_test)), t_test] = 1\nt_test_matrix = np.array(t_test_prime.astype(int))\n\nt_shuffled_prime = np.zeros(shape=(len(t_shuffled), 10))\nt_shuffled_prime[np.arange(len(t_shuffled)), t_shuffled] = 1 \nt_shuffled_matrix = np.array(t_shuffled_prime.astype(int))\n\n# Question 3(c)\nax = displaySample(16, X_normal)\nplt.suptitle(\"Question 3(c): some normalized MNIST digits\")\n\n# Question 3(d)\nclf = MLPClassifier(hidden_layer_sizes=[100],\n\t\t\t\tactivation='tanh',\n\t\t\t\tsolver='sgd',\n\t\t\t\t#batch_size=60000,\n\t\t\t\tbatch_size=200,\n\t\t\t\ttol=0.0,\n\t\t\t\tmax_iter=5,\n\t\t\t\twarm_start=True,\n\t\t\t\tlearning_rate_init=0.18,\n\t\t\t\tmomentum=0.65,\n\t\t\t\talpha=0.165)\n\ntrainErrList = []\ntestErrList = []\nfor x in range(50):\n\tclf.fit(X_normal, t_shuffled)\n\ttesting_err = round((1 - clf.score(X_test_normal, t_test))*100, 2)\n\ttraining_err = round((1 - clf.score(X_normal, t_shuffled))*100, 2)\n\ttrainErrList.append(training_err)\n\ttestErrList.append(testing_err)\n\tprint(\"Testing error for training session %s = %s\" % (x+1, testing_err))\n\tprint(\"Training error for training session %s = %s\" % (x+1, training_err))\n\ncoefs = clf.coefs_\nbias = clf.intercepts_\npredict_proba = clf.predict_proba(X_test_normal)\n\n# Question 3(f)\nfig5 = plt.figure()\t\nfig5.suptitle(\"Figure 5, Question 3: training and test error in batch mode\")\nplt.plot(trainErrList, color='orange')\nplt.plot(testErrList, color='blue')\nplt.xlabel('training iterations')\nplt.ylabel('error')\n\nfig6 = plt.figure()\nfig6.suptitle(\"Figure 6: Question 3: test error during last 500 iterations of batch training\")\nplt.plot(testErrList[-500:], color='blue')\nplt.xlabel('training iterations')\nplt.ylabel('error')\n\n###########\tEND OF QUESTION 3 ############\n\n########### QUESTION 4 #############\n\ndef tanh(x):\n\treturn np.tanh(x)\n\ndef softmax(x):\n\treturn np.exp(x) / np.exp(x).sum(keepdims=True, axis=1)\n\n# Question 4(a)\ndef predict(X, W1, W2, b1, b2):\n\thidden_matrix = tanh(np.dot(X, W1) + b1.reshape(1, -1))\n\toutput_matrix = softmax(np.dot(hidden_matrix, W2) + b2.reshape(1, -1))\n\treturn hidden_matrix, output_matrix\n\nh1, output = predict(X_test_normal, coefs[0], coefs[1], bias[0], bias[1])\n\n# Question 4(b)\nprint(np.sum((output - predict_proba)**2))\n\n# Question 4(c)\ndef gradient(H, Y, T):\t\n\toutput_error = Y - T\n\tDW = (np.matmul(output_error.T, H)) / len(X_normal)\n\tDb = np.mean(output_error, 0)\n\treturn DW.T, Db.reshape(-1,1)\n\n# Question 4(d)\ntrainingErrList = []\ntestingErrList = []\nmeanLoss = []\ndef bgd(W1, b1, lrate, sigma, K):\n\tW2 = sigma * np.random.randn(100, 10) + 0\n\tb2 = np.zeros(shape = [10, 1])\n\tfor i in range(K+1):\n\t\thidden_grad, output_grad = predict(X_normal, W1, W2, b1, b2)\n\t\ttest_hidden, test_output = predict(X_test_normal, W1, W2, b1, b2)\n\t\tdW2, db2 = gradient(hidden_grad, output_grad, t_shuffled_matrix)\n\t\tW2 = W2 - lrate*dW2\n\t\tb2 = b2 - lrate*db2\n\t\tif np.mod(i, 5) == 0:\n\t\t\tmeanloss = round((np.mean(-t_shuffled_matrix * np.log(output_grad))*100), 2)\n\t\t\ttesting_err = round((np.sum((t_test_matrix - test_output)**2)/len(t_test_matrix)*100), 2)\n\t\t\ttraining_err = round((np.sum((t_shuffled_matrix - output_grad)**2)/len(t_shuffled_matrix)*100), 2)\n\t\t\tmeanLoss.append(meanloss)\n\t\t\ttrainingErrList.append(training_err)\n\t\t\ttestingErrList.append(testing_err)\n\t\t\tprint(\"Iteration number = %s\" %(i))\n\t\t\tprint(\"Testing error for training session %s = %s\" % (i, testing_err))\n\t\t\tprint(\"Training error for training session %s = %s\" % (i, training_err))\n\t\t\tprint(\"Mean training loss for training session %s = %s\" % (i, meanloss))\n\n# Question 4(e)\nbgd(coefs[0], bias[0], 0.18, 0.01, 1000)\n\n# Question 4(d) graph outputs\nfig7 = plt.figure()\t\nfig7.suptitle(\"Figure 7, Question 4(d): training and test error for batch gradient descent\")\nplt.plot(trainingErrList, color='orange')\nplt.plot(testingErrList, color='blue')\nplt.xlabel('iterations')\nplt.ylabel('error')\n\nfig8 = plt.figure()\nfig8.suptitle(\"Figure 8: Question 4(d): mean training loss for batch gradient descent\")\nplt.plot(meanLoss, color='orange')\nplt.xlabel('iterations')\nplt.ylabel('loss')\n\nfig9 = plt.figure()\nfig9.suptitle(\"Figure 9: Question 4(d): training and test error for last 500 epochs of bgd\")\nplt.plot(trainingErrList[-100:], color='orange')\nplt.plot(testingErrList[-100:], color='blue')\nplt.xlabel('iterations')\nplt.ylabel('error')\n\nfig10 = plt.figure()\nfig10.suptitle(\"Figure 10: Question 4(d): mean training loss for last 500 epochs of bgd\")\nplt.plot(meanLoss[-100:], color='orange')\nplt.xlabel('iterations')\nplt.ylabel('loss')\n\n# Question 4(g)\nstochtrainingErrList = []\nstochtestingErrList = []\nstochmeanLoss = []\ndef sgd(W1, b1, lrate, alpha, sigma, K, batchSize, mom):\n\tW2 = sigma * np.random.randn(100, 10) + 0\n\tb2 = np.zeros(shape = [10, 1])\n\tvdW2 = np.zeros(shape=[100, 10])\n\tvdb2 = np.zeros(shape=[10, 1])\n\tfor i in range(K+1):\n\t\tfor batch_slice in gen_batches(len(X_normal), batchSize):\n\t\t\tX_train_batch = X_normal[batch_slice]\n\t\t\tT_train_batch = t_shuffled_matrix[batch_slice]\n\t\t\thidden_grad, output_grad = predict(X_train_batch, W1, W2, b1, b2)\n\t\t\tdW2, db2 = gradient(hidden_grad, output_grad, T_train_batch)\n\t\t\tdW2 = dW2 + alpha*W2\n\t\t\tvdW2 = mom*vdW2 + dW2\n\t\t\tvdb2 = mom*vdb2 + db2\n\t\t\tW2 = W2 - lrate*vdW2\n\t\t\tb2 = b2 - lrate*vdb2\n\t\tif np.mod(i, 5) == 0:\n\t\t\ttest_hidden, test_output = predict(X_test_normal, W1, W2, b1, b2)\n\t\t\thidden_grad, output_grad = predict(X_normal, W1, W2, b1, b2)\n\t\t\tmeanloss = round((np.mean(-t_shuffled_matrix * np.log(output_grad))*100), 2)\n\t\t\tmeanloss = round((meanloss + (0.5*alpha) * (np.mean(np.square(W2)))), 2)\n\t\t\ttraining_err = round((np.sum((t_shuffled_matrix - output_grad)**2)/len(t_shuffled_matrix)*100), 2)\n\t\t\ttesting_err = round((np.sum((t_test_matrix - test_output)**2)/len(t_test_matrix)*100), 2)\n\t\t\tstochmeanLoss.append(meanloss)\n\t\t\tstochtrainingErrList.append(training_err)\n\t\t\tstochtestingErrList.append(testing_err)\n\t\t\tprint(\"Iteration number = %s\" %(i))\n\t\t\tprint(\"Testing error for training session %s = %s\" % (i, stochtestingErrList[-1]))\n\t\t\tprint(\"Training error for training session %s = %s\" % (i, stochtrainingErrList[-1]))\n\t\t\tprint(\"Mean training loss for training session %s = %s\" % (i, stochmeanLoss[-1]))\n\tprint(\"Minimum test error during %s epochs of training = %s\" %(K, min(stochtestingErrList)))\n\n# Question 4(h)\nsgd(coefs[0], bias[0], 0.18, 0.0001, 0.01, 50, 3000, 0.99)\n\n# Question 4(g) graph outputs\nfig11 = plt.figure()\t\nfig11.suptitle(\"Figure 11, Question 4(g): training and test error for stochastic gradient descent\")\nplt.plot(stochtrainingErrList, color='orange')\nplt.plot(stochtestingErrList, color='blue')\nplt.xlabel('iterations')\nplt.ylabel('error')\n\nfig12 = plt.figure()\nfig12.suptitle(\"Figure 12: Question 4(g): mean training loss for stochastic gradient descent\")\nplt.plot(stochmeanLoss, color='orange')\nplt.xlabel('iterations')\nplt.ylabel('loss')\n\nplt.show()", "sub_path": "A3/source.py", "file_name": "source.py", "file_ext": "py", "file_size_in_byte": 9882, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "numpy.array", "line_number": 15, "usage_type": "call"}, {"api_name": "sklearn.datasets.make_moons", "line_number": 17, "usage_type": "call"}, {"api_name": "sklearn.datasets", "line_number": 17, "usage_type": "name"}, {"api_name": "sklearn.datasets.make_moons", "line_number": 18, "usage_type": "call"}, {"api_name": "sklearn.datasets", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "sklearn.neural_network.MLPClassifier", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.Inf", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.Inf", "line_number": 41, "usage_type": "attribute"}, {"api_name": "bonnerlib2.dfContour", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "bonnerlib2.dfContour", "line_number": 60, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 77, "usage_type": "call"}, {"api_name": "sklearn.utils.shuffle", "line_number": 85, "usage_type": "call"}, {"api_name": "sklearn.utils", "line_number": 85, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 88, "usage_type": "call"}, {"api_name": "sklearn.utils.resample", "line_number": 94, "usage_type": "call"}, {"api_name": "sklearn.utils", "line_number": 94, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}, {"api_name": "numpy.ceil", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.suptitle", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "sklearn.neural_network.MLPClassifier", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 147, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 149, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 151, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 152, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 156, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 157, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}, {"api_name": "numpy.tanh", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 172, "usage_type": "call"}, {"api_name": "sklearn.utils.extmath.softmax", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 185, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 193, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 193, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.mod", "line_number": 201, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 202, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 202, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 204, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 217, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 217, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 219, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 219, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 220, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 220, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 221, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 221, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 222, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 222, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 224, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 224, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 226, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 226, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 227, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 227, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 228, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 228, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 230, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 230, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 232, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 232, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 233, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 233, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 234, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 234, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 235, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 235, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 237, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 237, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 239, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 239, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 240, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 240, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 241, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 241, "usage_type": "name"}, {"api_name": "numpy.random.randn", "line_number": 248, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 248, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 249, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 250, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 251, "usage_type": "call"}, {"api_name": "sklearn.utils.gen_batches", "line_number": 253, "usage_type": "call"}, {"api_name": "numpy.mod", "line_number": 263, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 266, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 266, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 267, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 267, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 268, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 269, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 283, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 283, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 285, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 285, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 286, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 286, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 287, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 287, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 288, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 288, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 290, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 290, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 292, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 292, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 293, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 293, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 294, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 294, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 296, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 296, "usage_type": "name"}]} +{"seq_id": "40515260", "text": "import importlib\nimport time, sys\nimport logging\nimport os.path\nfrom multiprocessing import Process\n\nfrom bitshares.notify import Notify\nfrom bitshares.instance import shared_bitshares_instance\n\nlog = logging.getLogger(__name__)\n\n\n# FIXME: currently static list of bot strategies: ? how to enumerate bots available and deploy new bot strategies.\n\nSTRATEGIES=[('dexbot.strategies.echo',\"Echo Test\"),\n ('dexbot.strategies.follow_orders',\"Haywood's Follow Orders\")]\n\nlog_bots = logging.getLogger('dexbot.per_bot')\n# NOTE this is the special logger for per-bot events\n# it returns LogRecords with extra fields: botname, account, market and is_disabled\n# is_disabled is a callable returning True if the bot is currently disabled.\n# GUIs can add a handler to this logger to get a stream of events re the running bots.\n\n\nclass BotInfrastructure(Process):\n\n bots = dict()\n\n def __init__(\n self,\n config,\n bitshares_instance=None,\n gui_data=None\n ):\n super().__init__()\n # BitShares instance\n self.bitshares = bitshares_instance or shared_bitshares_instance()\n\n self.config = config\n\n # set the module search path\n user_bot_path = os.path.expanduser(\"~/bots\")\n if os.path.exists(user_bot_path):\n sys.path.append(user_bot_path)\n\n # Load all accounts and markets in use to subscribe to them\n accounts = set()\n markets = set()\n \n # Initialize bots:\n for botname, bot in config[\"bots\"].items():\n if \"account\" not in bot:\n log_bots.critical(\"Bot has no account\",extra={'botname':botname,'account':'unknown','market':'unknown','is_dsabled':(lambda: True)})\n continue\n if \"market\" not in bot:\n log_bots.critical(\"Bot has no market\",extra={'botname':botname,'account':bot['account'],'market':'unknown','is_disabled':(lambda: True)})\n continue\n try:\n klass = getattr(\n importlib.import_module(bot[\"module\"]),\n 'Strategy'\n )\n self.bots[botname] = klass(\n config=config,\n name=botname,\n bitshares_instance=self.bitshares,\n gui_data=gui_data\n )\n markets.add(bot['market'])\n accounts.add(bot['account'])\n except:\n log_bots.exception(\"Bot initialisation\",extra={'botname':botname,'account':bot['account'],'market':'unknown','is_disabled':(lambda: True)})\n\n if len(markets) == 0:\n log.critical(\"No bots to launch, exiting\")\n sys.exit(70) # 70= \"Software error\" in /usr/include/sysexts.h\n # Create notification instance\n # Technically, this will multiplex markets and accounts and\n # we need to demultiplex the events after we have received them\n self.notify = Notify(\n markets=list(markets),\n accounts=list(accounts),\n on_market=self.on_market,\n on_account=self.on_account,\n on_block=self.on_block,\n bitshares_instance=self.bitshares\n )\n\n # Events\n def on_block(self, data):\n for botname, bot in self.config[\"bots\"].items():\n if (not botname in self.bots) or self.bots[botname].disabled:\n continue\n try:\n self.bots[botname].ontick(data)\n except Exception as e:\n self.bots[botname].error_ontick(e)\n self.bots[botname].log.exception(\"in .tick()\")\n\n def on_market(self, data):\n if data.get(\"deleted\", False): # no info available on deleted orders\n return\n for botname, bot in self.config[\"bots\"].items():\n if self.bots[botname].disabled:\n continue\n if bot[\"market\"] == data.market:\n try:\n self.bots[botname].onMarketUpdate(data)\n except Exception as e:\n self.bots[botname].error_onMarketUpdate(e)\n self.bots[botname].log.exception(\".onMarketUpdate()\")\n\n def on_account(self, accountupdate):\n account = accountupdate.account\n for botname, bot in self.config[\"bots\"].items():\n if self.bots[botname].disabled:\n self.bots[botname].log.info(\"bot disabled\" % botname)\n continue\n if bot[\"account\"] == account[\"name\"]:\n try:\n self.bots[botname].onAccount(accountupdate)\n except Exception as e:\n self.bots[botname].error_onAccount(e)\n self.bots[botname].log.exception(\".onAccountUpdate()\")\n\n def run(self):\n self.notify.listen()\n", "sub_path": "dexbot/bot.py", "file_name": "bot.py", "file_ext": "py", "file_size_in_byte": 4823, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "logging.getLogger", "line_number": 10, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 18, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 25, "usage_type": "name"}, {"api_name": "bitshares.instance.shared_bitshares_instance", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.path.expanduser", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 42, "usage_type": "name"}, {"api_name": "os.path.path.exists", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 43, "usage_type": "name"}, {"api_name": "sys.path.append", "line_number": 44, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "importlib.import_module", "line_number": 60, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 76, "usage_type": "call"}, {"api_name": "bitshares.notify.Notify", "line_number": 80, "usage_type": "call"}]} +{"seq_id": "434564993", "text": "\"\"\"\nThis file is for models creation, which consults options\nand creates each encoder and decoder accordingly.\n\"\"\"\nimport torch\nimport torch.nn as nn\n\nimport onmt\nimport onmt.io\nimport onmt.Models\nimport onmt.modules\nfrom onmt.RLModels import StateEncoder, UtteranceEncoder, StateUtteranceEncoder, \\\n MeanEncoder, RNNEncoder, \\\n PolicyDecoder, PolicyModel, ValueModel, ValueDecoder\nfrom onmt.Utils import use_gpu\n\nfrom cocoa.io.utils import read_pickle\nfrom neural import make_model_mappings\n\n\ndef make_embeddings(opt, word_dict, emb_length, for_encoder=True):\n return nn.Embedding(len(word_dict), emb_length)\n\n\ndef make_encoder(opt, embeddings, output_size, fix_emb=False,):\n \"\"\"\n Various encoder dispatcher function.\n Args:\n opt: the option in current environment.\n embeddings (Embeddings): vocab embeddings for this encoder.\n \"\"\"\n\n encoder = StateEncoder(embeddings, output_size=output_size,\n state_length=opt.state_length, extra_size=3 if opt.dia_num>0 else 0,\n fix_emb=fix_emb)\n if opt.use_utterance:\n\n # TODO: use function to get the size?\n bert_output_size = 768\n\n if opt.bert_encoder == 'mean':\n bert_encoder = MeanEncoder(bert_output_size, output_size)\n else:\n bert_encoder = RNNEncoder(bert_output_size, output_size)\n uencoder = UtteranceEncoder(bert_encoder, output_size, output_size=output_size, model_path=opt.bert_model_path, use_gpu=len(opt.gpuid)>0)\n encoder = StateUtteranceEncoder(encoder, uencoder, input_size=output_size*2, output_size=output_size)\n\n return encoder\n\n\ndef make_decoder(opt, encoder_size, intent_size, output_value=False):\n \"\"\"\n Various decoder dispatcher function.\n Args:\n opt: the option in current environment.\n embeddings (Embeddings): vocab embeddings for this decoder.\n \"\"\"\n if output_value:\n return ValueDecoder(encoder_size=encoder_size)\n return PolicyDecoder(encoder_size=encoder_size, intent_size=intent_size)\n\n\ndef load_test_model(model_path, opt, dummy_opt):\n if model_path is not None:\n print('Load model from {}.'.format(model_path))\n checkpoint = torch.load(model_path, map_location=lambda storage, loc: storage)\n\n model_opt = checkpoint['opt']\n for arg in dummy_opt:\n if arg not in model_opt:\n model_opt.__dict__[arg] = dummy_opt[arg]\n else:\n print('Build model from scratch.')\n checkpoint = None\n model_opt = opt\n\n mappings = read_pickle('{}/vocab.pkl'.format(model_opt.mappings))\n\n # mappings = read_pickle('{0}/{1}/vocab.pkl'.format(model_opt.mappings, model_opt.model))\n mappings = make_model_mappings(model_opt.model, mappings)\n\n model, critic = make_base_model(model_opt, mappings, use_gpu(opt), checkpoint)\n model.eval()\n critic.eval()\n return mappings, model, model_opt, critic\n\n\ndef make_base_model(model_opt, mappings, gpu, checkpoint=None):\n \"\"\"\n Args:\n model_opt: the option loaded from checkpoint.\n fields: `Field` objects for the model.\n gpu(bool): whether to use gpu.\n checkpoint: the model gnerated by train phase, or a resumed snapshot\n model from a stopped training.\n Returns:\n the NMTModel.\n \"\"\"\n # Make encoder.\n src_dict = mappings['src_vocab']\n src_embeddings = make_embeddings(model_opt, src_dict, model_opt.word_vec_size)\n encoder = make_encoder(model_opt, src_embeddings, model_opt.hidden_size)\n # print('encoder', encoder)\n\n # Make decoder.\n tgt_dict = mappings['tgt_vocab']\n\n\n decoder = make_decoder(model_opt, model_opt.hidden_size, len(tgt_dict))\n # print('decoder', decoder)\n\n\n model = PolicyModel(encoder, decoder)\n model.model_type = 'text'\n\n # Make Critic.\n # critic_embeddings = src_embeddings\n critic_embeddings = make_embeddings(model_opt, src_dict, model_opt.word_vec_size)\n value_encoder = make_encoder(model_opt, critic_embeddings, model_opt.hidden_size, fix_emb=True)\n value_decoder = make_decoder(model_opt, model_opt.hidden_size, len(tgt_dict), output_value=True)\n critic = ValueModel(value_encoder, value_decoder)\n # model.critic = critic\n\n # Load the model states from checkpoint or initialize them.\n if checkpoint is not None:\n print('Loading model parameters.')\n model.load_state_dict(checkpoint['model'])\n\n # Get parameters of critic model\n # if hasattr(model, 'critic'):\n if critic is not None:\n if checkpoint.get('critic') is not None:\n print('Loading critic model parameters.')\n critic.load_state_dict(checkpoint['critic'])\n else:\n print('Intializing critic parameters.')\n for p in critic.parameters():\n p.data.uniform_(-model_opt.param_init, model_opt.param_init)\n else:\n if model_opt.param_init != 0.0:\n print('Intializing model parameters.')\n for p in model.parameters():\n p.data.uniform_(-model_opt.param_init, model_opt.param_init)\n\n # Get parameters of critic model\n if critic is not None:\n print('Intializing critic parameters.')\n for p in critic.parameters():\n p.data.uniform_(-model_opt.param_init, model_opt.param_init)\n\n wordvec = {'utterance': model_opt.pretrained_wordvec[0]}\n if len(model_opt.pretrained_wordvec) > 1:\n wordvec['kb'] = model_opt.pretrained_wordvec[1]\n\n def load_wordvec(embeddings, name):\n embeddings.load_pretrained_vectors(\n wordvec[name], model_opt.fix_pretrained_wordvec)\n\n # Don't need pretrained word vec for LFs\n if not model_opt.model in ('lf2lf',):\n load_wordvec(model.encoder.embeddings, 'utterance')\n if hasattr(model, 'context_embedder'):\n load_wordvec(model.context_embedder.embeddings, 'utterance')\n if hasattr(model, 'kb_embedder') and model.kb_embedder is not None:\n load_wordvec(model.kb_embedder.embeddings, 'kb')\n\n if model_opt.model == 'seq2seq':\n load_wordvec(model.decoder.embeddings, 'utterance')\n\n # Make the whole model leverage GPU if indicated to do so.\n if gpu:\n model.cuda()\n critic.cuda()\n else:\n model.cpu()\n critic.cpu()\n\n return model, critic\n\n\ndef make_critic_model(model_opt, mappings, gpu, encoder=None):\n # Make encoder.\n if encoder is None:\n src_dict = mappings['src_vocab']\n src_embeddings = make_embeddings(model_opt, src_dict, model_opt.word_vec_size)\n encoder = make_encoder(model_opt, src_embeddings, model_opt.hidden_size)\n # print('encoder', encoder)\n\n # Make decoder.\n tgt_dict = mappings['tgt_vocab']\n\n decoder = make_decoder(model_opt, model_opt.hidden_size, len(tgt_dict), output_value=True)\n # print('decoder', decoder)\n\n model = PolicyModel(encoder, decoder)\n\n # Make the whole model leverage GPU if indicated to do so.\n if gpu:\n model.cuda()\n else:\n model.cpu()\n\n return model\n", "sub_path": "craigslistbargain/neural/rl_model_builder.py", "file_name": "rl_model_builder.py", "file_ext": "py", "file_size_in_byte": 7195, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "torch.nn.Embedding", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "name"}, {"api_name": "onmt.RLModels.StateEncoder", "line_number": 33, "usage_type": "call"}, {"api_name": "onmt.RLModels.MeanEncoder", "line_number": 42, "usage_type": "call"}, {"api_name": "onmt.RLModels.RNNEncoder", "line_number": 44, "usage_type": "call"}, {"api_name": "onmt.RLModels.UtteranceEncoder", "line_number": 45, "usage_type": "call"}, {"api_name": "onmt.RLModels.StateUtteranceEncoder", "line_number": 46, "usage_type": "call"}, {"api_name": "onmt.RLModels.ValueDecoder", "line_number": 59, "usage_type": "call"}, {"api_name": "onmt.RLModels.PolicyDecoder", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 66, "usage_type": "call"}, {"api_name": "cocoa.io.utils.read_pickle", "line_number": 77, "usage_type": "call"}, {"api_name": "neural.make_model_mappings", "line_number": 80, "usage_type": "call"}, {"api_name": "onmt.Utils.use_gpu", "line_number": 82, "usage_type": "call"}, {"api_name": "onmt.RLModels.PolicyModel", "line_number": 113, "usage_type": "call"}, {"api_name": "onmt.RLModels.ValueModel", "line_number": 121, "usage_type": "call"}, {"api_name": "onmt.RLModels.PolicyModel", "line_number": 195, "usage_type": "call"}]} +{"seq_id": "270828184", "text": "import datetime\nfrom openpyxl import Workbook\n\nwb = Workbook()\n# default sheet\nws = wb.active\nws['A1'] = 42\nws.append([1, 2, 3])\nws['A2'] = datetime.datetime.now()\nwb.save(\"sample.xlsx\")\n", "sub_path": "python/openpyxl/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 187, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "openpyxl.Workbook", "line_number": 4, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 9, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 9, "usage_type": "attribute"}]} +{"seq_id": "35016938", "text": "\n# coding: utf-8\n\n# # Inhalt\n# \n# Analyse der Sensordaten\n# Analyse der Performance des Förderbandes\n# Leistungsverbrauch / Leistungsersparnis\n# Diverse Kennzahlen\n# \n# ### Geprüft wird das Förderband zwischen den beiden Sensoren\n# ![title](res/b06_and_b09.png)\n\n# ### Initialisierung\n\n# In[1]:\n\n\nget_ipython().run_line_magic('run', '../Setup.ipynb')\n\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport matplotlib\nimport datetime \nfrom matplotlib import dates\nfrom datetime import timedelta\nfrom sklearn import preprocessing\n\nfrom isac.plotting import plot_helper as isac_plotter\nfrom isac.database.connection import database_connector as connector\nfrom isac import configuration\nfrom data_compare import sensor_data_provider as sdp \n\n\n# In[2]:\n\n\ndataProvider = sdp.SensorDataProvider()\n\nsensor_in_name = \"Bandbero_99_B06_Meldungen\"\nsensor_out_name = \"Bandbero_99_B09_Meldungen\"\npowermeter_name = \"PÜberwachung_ZS_B103_Leistungen\"\n\n# Sensordaten\nb06 = dataProvider.getSensorData(sensor_in_name)\nb09 = dataProvider.getSensorData(sensor_out_name)\nb06['Zeit'] = pd.to_datetime(b06['Zeit'], unit='ms')\nb09['Zeit'] = pd.to_datetime(b09['Zeit'], unit='ms')\n\n# Leistungsdaten\npowermeter = dataProvider.getSensorData(powermeter_name)\npowermeter['Zeit'] = pd.to_datetime(powermeter['Zeit'], unit='ms')\n\n\n# ### Analyse der in der Datenbank gespeicherten Sensordaten\n# Immer, wenn ein Sensor ein Signal geliefert hat, bedeutet das, dass ein Werkstückträger von einem der Sensoren erfasst wurde.\n# Stellt man die beiden Sensoren vor und nach dem Förderband gegenüber, kann man berechnen, wieviele Werkstückträger sich zu jeder Zeit auf dem Förderband befinden.\n# \n# #### Interpretation der Signale:\n# Liefert der Sensor am Förderbandbeginn ein Signal, hat ein Werkstückträger das Förderband erreicht.\n# Liefert der Sensor am Förderbandende ein Signal hat ein Werkstückträger das Förderband verlassen.\n# \n\n# In[3]:\n\n\nanzahl=50\nfig, axes = plt.subplots(figsize=(50,5))\nb06.head(anzahl).plot(drawstyle=\"steps-post\", ls=\"-\", label=\"Sensor Förderbandbeginn\", color='cyan', ax=axes, x='Zeit', y='Wert', linewidth=1.5);\nb09.head(anzahl).plot(drawstyle=\"steps-post\", ls=\"--\", label=\"Sensor Förderbandende\", color='red', ax=axes, x='Zeit', y='Wert', linewidth=1.5);\n\naxes.legend(loc='upper left')\n\nisac_plotter.savePlotLocal(\"res-gen/plot-1.png\")\n\n\n# #### Remodellierung der Signale am Förderbandende\n# Wenn der Sensor am Förderbandende ein Signal liefert, hat ein Werkstückträger das Förderband verlassen.\n# Jeder Spike des sich am Förderbandende befindenden Sensors wird deshalb als Verlassen eines Werkstückträgers vom Förderband interpretiert.\n# \n# Der Sensor am Förderbandende liefert negative Werte bei einem erkannten Werkstückträger zurück, um die Anzahl der sich auf dem Förderband befindenden Werkstückträger berechnen zu können.\n\n# In[4]:\n\n\nb09[\"Wert\"] = -b09[\"Wert\"]\nb09.head()\n\n\n# #### Standardfall Sensorsignale\n# Im Normalfall sollte jeder Sensor beim Durchgang eines Werkstückträgers zwei Signale liefern:\n# \n# Erkennen eines neuen Werkstückträgers (Wert: 1)\n# Verlassen dieses Werkstückträgers (Wert: 0)\n# \n# Zusätzlich liefert ein Sensor bei Aktivierung des Gesamtsystems ein \"Initialisierungs-Signal\" (Wert: 0).\n# \n# #### Erkennung fehlerhafter Sensordaten\n# Sollten Werkstückträger nicht korrekt erkannt worden sein oder ein technischer Defekt vorliegen, kann es vorkommen, dass Signale nicht oder zu oft getriggert werden.\n# \n# Um das zu erkennen, werden die vorliegenden Sensordaten auf Gültigkeit überprüft. Nach jedem Eingang eines Werkstückträgers (Spike) muss auch zwingendermaßen wieder ein Verlassen-Signal folgen. Mehrere Signale des gleichen Typs hintereinander sind nicht valide.\n# \n# Diese False-Positives und False-Negatives können erkannt werden, indem man prüft, ob sich Signale wiederholen. Wiederholen sich Signale eines Sensors, befinden sie sich außerhalb des Gültigkeitsbereiches [0;1] siehe lila-farbene Grafik.\n\n# In[5]:\n\n\ndef plotErrors(dataframe):\n for i in range(0, len(dataframe)):\n current_value = abs(dataframe.loc[i, 'Wert'])\n if (current_value == 0):\n current_value = -1\n\n if (i == 0):\n value_before = 1\n dataframe.loc[i, 'recognize_error'] = value_before + current_value\n else:\n value_before = dataframe.loc[i - 1, 'recognize_error']\n dataframe.loc[i, 'recognize_error'] = value_before + current_value\n \n fig, axes = plt.subplots(figsize=(20,5))\n hfmt = dates.DateFormatter('%H:%M')\n axes.xaxis.set_major_formatter(hfmt)\n dataframe.plot(drawstyle=\"steps-post\", label='Gelieferte Signale des Sensors', color='purple', ax=axes, x='Zeit', y='recognize_error', linewidth=1.0);\n axes.legend(loc='upper left')\n\n\n# #### Beispiel fehlerhafte Signale Bandbero_99_B06\n# ![title](res/ISAC_FOERDERBAND.Bandbero_99_B06_Meldungen.png)\n\n# #### Signale am Förderbandbeginn: Bandbero_99_B06\n\n# In[6]:\n\n\nplotErrors(b06)\nisac_plotter.savePlotLocal(\"res-gen/plot-2.png\")\n\n\n# #### Signale am Förderbandende: Bandbero_99_B09\n\n# In[7]:\n\n\nplotErrors(b09)\nisac_plotter.savePlotLocal(\"res-gen/plot-3.png\")\n\n\n# #### Fehlerhafte Sensordaten korrigieren\n# Nach Sichtung der gelieferten Daten konnten wir feststellen, dass sich immer nur einzelne False-Negatives in den Testdaten befanden. Um diese zu korrigieren, änderten wir die Datensätze der erkannten False-Negatives.\n# \n# Ein Datensatz zählt als fehlerhaft, wenn er dem zeitlich vorherigen entspricht.\n# \n# Hinweis:\n# In den gelieferten Daten handelte es sich immer nur um einzelne Datensätze, die korrigiert werden mussten (Dementsprechend waren drei Datensätze hintereinander entweder False oder True). Sollten sich andere Fehler eingeschlichen haben, lässt sich dieser nicht mit diesem Ansatz korrigieren.\n\n# In[8]:\n\n\ndef fixErrors(dataframe, value_positive):\n for i in range(0, len(dataframe)):\n # Vorheriger Wert\n if (i == 0):\n pass\n else:\n value_before = dataframe.loc[i - 1, 'Wert']\n value_current = dataframe.loc[i, 'Wert']\n \n if (value_before == value_current):\n if (value_current == 0):\n dataframe.loc[i, 'Wert'] = value_positive\n else:\n dataframe.loc[i, 'Wert'] = 0\n \n return dataframe\n\n\n# #### Korrigierte Signale am Förderbandende: Bandbero_99_B06\n\n# In[9]:\n\n\nfixedDataframe_1 = fixErrors(b06, 1)\nplotErrors(fixedDataframe_1)\nisac_plotter.savePlotLocal(\"res-gen/plot-4.png\")\n\n\n# #### Korrigierte Signale am Förderbandende: Bandbero_99_B09\n\n# In[10]:\n\n\nfixedDataframe_2 = fixErrors(b09, -1)\nplotErrors(fixedDataframe_2)\nisac_plotter.savePlotLocal(\"res-gen/plot-5.png\")\n\n\n# ### Performance des Förderbandes\n# \n# Zur Berechnung von weiteren Kennzahlen werden die Datensätze der beiden Sensoren zusammengefasst und in einem Datenframe gespeichert. Damit können im Weiteren folgende Aussagen getroffen werden:\n# \n# Anzahl Werkstückträger auf dem Förderband\n# Aktivität des Förderbandes\n# Leistungsersparnis durch Abschaltung bei Inaktivität\n\n# In[11]:\n\n\n# Übernehmen der korrigierten Datensätze\nb06 = fixedDataframe_1[[\"Wert\", \"Zeit\", '_id']]\nb09 = fixedDataframe_2[[\"Wert\", \"Zeit\", '_id']]\n\n# Zusammenfassen der Datensätze aus beiden Sensoren\nresult = pd.concat([b06, b09])\nresult = result.sort_values(by=['Zeit'])\nresult = result.reset_index(drop=True)\nresult.head()\n\n\n# #### Anzahl der Werkstückträger auf dem Förderband\n# Die Anzahl der Werkstückträger ergibt sich aus den Signalen der beiden Sensoren am Förderbandbeginn und Förderbandende.\n# Erkennt der Sensor am Förderbandbeginn einen Werkstückträger, erhöht sich die Anzahl der Werkstückträger auf dem Förderband. Erreicht dieser Werkstückträger den Sensor am Förderbandende, verlässt er das Förderband.\n\n# In[12]:\n\n\nfor i in range(0, len(result)):\n # Vorheriger Wert\n if (i == 0):\n blocks_before = 0\n result.loc[i, 'blocksBetween'] = blocks_before + result.loc[i, 'Wert']\n else:\n blocks_before = result.loc[i - 1, 'blocksBetween']\n # Anzahl der Blöcke dazwischen\n result.loc[i, 'blocksBetween'] = blocks_before + result.loc[i - 1, 'Wert']\n\n\n# #### Inaktivität Förderband\n# Das Förderband wird unnötig angetrieben, wenn sich kein Werkstückträger auf dem Förderband, d.h. zwischen den beiden Sensoren am Anfang und Ende des Förderbandes, befindet.\n# \n# Befindet sich kein Werkstückträger auf dem Förderband könnte das Förderband inaktiv geschalten werden.\n\n# In[13]:\n\n\n# Konstante zur Erkennung des IDLE-Status\nCONST_IS_IDLE = 0.5\n\nfor i in range(0, len(result)):\n if (result.loc[i, 'blocksBetween'] == 0):\n result.loc[i, 'is_idle'] = CONST_IS_IDLE\n else:\n result.loc[i, 'is_idle'] = 0\n\nresult.head(8)\n\n\n# ### Visualisierung der Werkstückträger auf dem Förderband\n\n# In[14]:\n\n\nfig, axes = plt.subplots(figsize=(60,10))\nresult['Zeit'] = pd.to_datetime(result['Zeit'], unit='ms')\nhfmt = dates.DateFormatter('%H:%M')\naxes.xaxis.set_major_formatter(hfmt)\n\nanzahl = 2000\naxes.fill_between(result['Zeit'].values, result['is_idle'], 0, facecolor='grey', alpha=0.5, step=\"post\")\n\nresult.head(anzahl * 2).plot(label=\"Inaktivität Förderband (potentiell)\", drawstyle=\"steps-post\", color='grey', ax=axes, x='Zeit', y='is_idle');\nb06.head(anzahl).plot(label=\"Sensor Förderbandbeginn\", drawstyle=\"steps-post\", color='cyan', ax=axes, x='Zeit', y='Wert', linewidth=1.0);\nb09.head(anzahl).plot(label=\"Sensor Förderbandende\", drawstyle=\"steps-post\", color='red', ax=axes, x='Zeit', y='Wert', linewidth=1.0);\nresult.head(anzahl * 2).plot(label=\"Anzahl Werkstückträger auf Förderband\", drawstyle=\"steps-post\", color='blue', ax=axes, x='Zeit', y='blocksBetween');\n\naxes.legend(loc='upper left')\n\nisac_plotter.savePlotLocal(\"res-gen/plot-6.png\")\n\n\n# ### Kumulierte potentielle Aktivität bzw. Inaktivität des Förderbandes\n# Das Förderband ist aktiv, wenn sich mindestens ein Werkstückträger auf dem Förderband befindet.\n\n# In[15]:\n\n\nfor i in range(0, len(result)):\n if (i == 0):\n # Init\n result.loc[i, 'time_idle'] = timedelta()\n result.loc[i, 'time_active'] = timedelta()\n result.loc[i, 'time_sum'] = timedelta()\n continue\n \n if (result.loc[i, 'is_idle'] == 0):\n result.loc[i, 'time_active'] = result.loc[i, 'Zeit'] - result.loc[i - 1, 'Zeit'] + result.loc[i - 1, 'time_active']\n result.loc[i, 'time_idle'] = result.loc[i - 1, 'time_idle']\n else:\n result.loc[i, 'time_idle'] = result.loc[i, 'Zeit'] - result.loc[i - 1, 'Zeit'] + result.loc[i - 1, 'time_idle']\n result.loc[i, 'time_active'] = result.loc[i - 1, 'time_active']\n \n result.loc[i, 'time_sum'] = result.loc[i, 'time_active'] + result.loc[i, 'time_idle']\n\n\n# In[16]:\n\n\nlabels = 'Inaktiv', 'Aktiv'\ncolors = ['grey', 'green']\n\nsizes = [result['time_idle'].max().microseconds, # Förderband Inaktiv\n result['time_active'].max().microseconds] # Förderband Aktiv\nexplode = (0, 0.1)\n\nfig, ax = plt.subplots()\nax.pie(sizes, explode=explode, labels=labels, autopct='%1.1f %%', startangle=90, colors=colors)\nax.axis('equal')\n\nplt.title(\"Förderbandaktivität während Versuchsdauer\")\n\nisac_plotter.savePlotLocal(\"res-gen/plot-7.png\")\n\nplt.show()\n\n\n# In[17]:\n\n\nfig, axes = plt.subplots(figsize=(20,5))\nhfmt = dates.DateFormatter('%H:%M')\naxes.xaxis.set_major_formatter(hfmt)\n\ndef versuchsdauer_pct(in_value_in_seconds, pos=None):\n return str((pos - 1) * 12.5) + ' %'\n\naxes.yaxis.set_major_formatter(plt.FuncFormatter(versuchsdauer_pct))\naxes.set_ylabel(\"Versuchsdauer\")\n\nresult.plot(label=\"Gesamte Laufzeit\", color='orange', ax=axes, x='Zeit', y='time_sum');\nresult.plot(label=\"Aktiv\", color='green', ax=axes, x='Zeit', y='time_active');\nresult.plot(label=\"Inaktivität (potentiell)\", color='grey', ax=axes, x='Zeit', y='time_idle');\n\naxes.legend(loc='upper left')\n\nisac_plotter.savePlotLocal(\"res-gen/plot-8.png\")\n\n\n# ### Übersicht Verbrauch\n\n# In[18]:\n\n\nfig, axes = plt.subplots(figsize=(20, 5))\naxes.set_ylabel(\"Watt\")\nhfmt = dates.DateFormatter('%H:%M')\naxes.xaxis.set_major_formatter(hfmt)\n\npowermeter.plot(label=\"Leistungsverbrauch\", color='orange', ax=axes, x='Zeit', y='Wert');\naxes.fill_between(powermeter['Zeit'].values, powermeter['Wert'], facecolor='orange', alpha=0.6,)\n\naxes.legend(loc='upper left')\n\nisac_plotter.savePlotLocal(\"res-gen/plot-9.png\")\n\n\n# ### Leistungsverbrauch im Vergleich zu auf dem Förderband befindlichen Werkstückträgern\n# Frage: Ist die Anzahl der Werkstückträger auf dem Förderband im Leistungsverbrauch erkennbar?\n# \n# Antwort: Nach Sichtung der Daten sind keine relevanten Spikes im Verbrauch in Abhängigkeit zu den auf dem Förderband befindlichen Werkstückträgern sichtbar.\n\n# In[19]:\n\n\n# Normalisierung des Leistungsverbrauches\npowermeter_norm = powermeter[100:].copy()\n\nx = powermeter_norm[[\"Wert\"]].values\nmin_max_scaler = preprocessing.MinMaxScaler()\nx_scaled = min_max_scaler.fit_transform(x)\npowermeter_norm[\"Wert\"] = pd.DataFrame(x_scaled)\n\n\n# In[20]:\n\n\nfig, axes = plt.subplots(figsize=(40,5))\nhfmt = dates.DateFormatter('%H:%M')\naxes.xaxis.set_major_formatter(hfmt)\n\naxes.fill_between(result['Zeit'].values, result['is_idle'], 0, facecolor='grey', alpha=0.5, step=\"post\")\nresult.plot(label=\"Inaktivität (potentiell)\", drawstyle=\"steps-post\", color='grey', ax=axes, x='Zeit', y='is_idle');\nresult.plot(label=\"Anzahl Werkstückträger auf Förderband\", drawstyle=\"steps-post\", color='blue', ax=axes, x='Zeit', y='blocksBetween');\npowermeter_norm.plot(label=\"Leistungsverbrauch (normalisiert)\", drawstyle=\"steps-post\", color='orange', ax=axes, x='Zeit', y='Wert');\n\naxes.legend(loc='upper left')\n\nisac_plotter.savePlotLocal(\"res-gen/plot-10.png\")\n\n\n# ### Leistungsersparnis durch Abschaltung bei Inaktivität des Förderbands\n# Um die potentielle Ersparnis bei der Abschaltung des Förderbandes bei Inaktivität zu berechnen, werden die Sensordaten mit den Leistungsverbrauchsdaten miteinander verknüpft.\n# \n# Anhand der Sensordaten kann gefiltert werden, wann das Förderband abgeschaltet werden könnte und somit in Abhängigkeit der derzeit verbrauchten Leistung die potentielle Leistungsersparnis berechnet werden.\n\n# In[21]:\n\n\n# Zusammenfügen der Daten\npowermeter[\"is_powermeter_sensor\"] = 0\nmerged_frames = pd.concat([result, powermeter], sort=True)\nmerged_frames = merged_frames.sort_values(by=['Zeit'])\nmerged_frames = merged_frames.reset_index(drop=True)\n\n\n# #### Berechnung der Zeiten, in denen das Förderband inaktiv sein könnte\n\n# In[22]:\n\n\ndef getInactivePowermeterDataframe(dataframe):\n is_idle = True\n for i in range(0, len(dataframe)):\n value_is_powermeter = dataframe.loc[i, 'is_powermeter_sensor']\n if (value_is_powermeter == 0):\n # Powermeter Sensoreintrag\n if is_idle:\n dataframe.loc[i, 'is_idle'] = 0.5\n else:\n dataframe.loc[i, 'is_idle'] = 0\n else:\n # Anderer Sensoreintrag\n is_idle = dataframe.loc[i, 'is_idle'] == 0.5\n \n return dataframe\n\npowermeter_inactive = getInactivePowermeterDataframe(merged_frames)\n\n\n# #### Berechnung der Leistung bei Inaktivität\n\n# In[23]:\n\n\ndef calculatePowermeterPower(df_powermeter):\n previous_index = 0\n for i in range(0, len(df_powermeter)):\n current_index = powermeter_inactive.index.values.astype(int)[i]\n \n if (i == 0):\n # Init\n powermeter_inactive.loc[current_index, 'amount_possible'] = 0.0\n powermeter_inactive.loc[current_index, 'amount_idle'] = 0.0\n powermeter_inactive.loc[current_index, 'amount_active'] = 0.0\n previous_index = current_index\n continue\n \n # Insgesamt möglich\n powermeter_inactive.loc[current_index, 'amount_possible'] = powermeter_inactive.loc[previous_index, 'amount_possible'] + (powermeter_inactive.loc[current_index, 'Zeit'] - powermeter_inactive.loc[previous_index, 'Zeit']).microseconds * powermeter_inactive.loc[previous_index, 'Wert']\n \n # Unterscheidung Idle / Möglich\n if (powermeter_inactive.loc[current_index, 'is_idle'] == 0.5):\n # Idle\n powermeter_inactive.loc[current_index, 'amount_idle'] = powermeter_inactive.loc[previous_index, 'amount_idle'] + (powermeter_inactive.loc[current_index, 'Zeit'] - powermeter_inactive.loc[previous_index, 'Zeit']).microseconds * powermeter_inactive.loc[previous_index, 'Wert']\n powermeter_inactive.loc[current_index, 'amount_active'] = powermeter_inactive.loc[previous_index, 'amount_active']\n powermeter_inactive.loc[current_index, 'dWert'] = 0\n else:\n # Working\n powermeter_inactive.loc[current_index, 'amount_idle'] = powermeter_inactive.loc[previous_index, 'amount_idle']\n powermeter_inactive.loc[current_index, 'amount_active'] = powermeter_inactive.loc[previous_index, 'amount_active'] + (powermeter_inactive.loc[current_index, 'Zeit'] - powermeter_inactive.loc[previous_index, 'Zeit']).microseconds * powermeter_inactive.loc[previous_index, 'Wert']\n powermeter_inactive.loc[current_index, 'dWert'] = powermeter_inactive.loc[current_index, 'Wert']\n previous_index = current_index \n return df_powermeter\n\npowermeter_inactive = calculatePowermeterPower(powermeter_inactive)\n\n\n# ### Vergleich Leistungsverbrauch zu potienteller Leistungsersparnis\n\n# In[24]:\n\n\nfig, axes = plt.subplots(figsize=(40,10))\naxes.set_ylabel(\"Watt\")\n\nactive_pm = powermeter_inactive[powermeter_inactive[\"is_powermeter_sensor\"] == 0.0].copy()\ninactive_pm = powermeter_inactive[powermeter_inactive[\"is_powermeter_sensor\"] == 0.0][powermeter_inactive[\"is_idle\"]== 0.5].copy()\n\ninactive_pm.plot(label=\"Leistungsverbrauch (potentielle Ersparnis)\", color='grey', ax=axes, x='Zeit', y='dWert', drawstyle=\"steps-post\");\naxes.fill_between(inactive_pm['Zeit'].values, inactive_pm['Wert'], facecolor='grey', alpha=1, step=\"post\")\n\nactive_pm.plot(label=\"Leistungsverbrauch\", color='orange', ax=axes, x='Zeit', y='dWert', drawstyle=\"steps-post\");\naxes.fill_between(active_pm['Zeit'].values, active_pm['dWert'], facecolor='orange', alpha=1, step=\"post\")\n\naxes.legend(loc='upper left')\n\nisac_plotter.savePlotLocal(\"res-gen/plot-11.png\")\n\n\n# In[25]:\n\n\nlabels = 'Inaktiv', 'Aktiv'\ncolors = ['grey', 'orange']\n\nsizes = [inactive_pm['amount_idle'].max(), # Förderband Inaktiv\n active_pm['amount_active'].max()] # Förderband Aktiv\nexplode = (0, 0.1)\n\nfig, ax = plt.subplots()\nax.pie(sizes, explode=explode, labels=labels, autopct='%1.1f %%', startangle=90, colors=colors)\nax.axis('equal')\n\nplt.title(\"Leistungsverbrauch Förderband während Versuchsdauer\")\n\nisac_plotter.savePlotLocal(\"res-gen/plot-12.png\")\n\nplt.show()\n\n\n# ### Berechnung Leistungsverbrauch\n\n# In[26]:\n\n\ndef wms_to_kwh(value):\n value_watt_seconds = value * 0.000001\n value_watt_hours = value_watt_seconds / 3600\n value_kilo_watt_hours = value_watt_hours / 1000\n return value_kilo_watt_hours\n\n\n# ##### Berechnung Leistungsverbrauch Förderband Gesamt\n\n# In[27]:\n\n\n# watt_micro_seconds\namount_idle_max = powermeter_inactive['amount_idle'].max()\namount_active_max = powermeter_inactive['amount_active'].max()\n\nvalue_kwh_active_and_idle = wms_to_kwh(amount_idle_max + amount_active_max)\nvalue_kwh_idle = wms_to_kwh(amount_idle_max)\nvalue_kwh_active = wms_to_kwh(amount_active_max)\nstr(value_kwh_active_and_idle) + ' kWh'\n\n\n# ##### Berechnung Leistungsverbrauch Förderband inaktiv\n\n# In[28]:\n\n\nvalue_kwh_idle = wms_to_kwh(amount_idle_max)\nstr(value_kwh_idle) + ' kWh'\n\n\n# ##### Berechnung Leistungsverbrauch Förderband aktiv\n\n# In[29]:\n\n\nvalue_kwh_active = wms_to_kwh(amount_active_max)\nstr(value_kwh_active) + ' kWh'\n\n\n# ### Mögliche Leistungsersparnis, wenn das Förderband bei Inaktivität abgeschaltet wird\n# Das Förderband gilt als inaktiv, wenn sich kein Werkstückträger auf dem Förderband befindet\n\n# In[30]:\n\n\nprint('Förderband Gesamtverbrauch')\ntime_start = powermeter['Zeit'][0]\ntime_end = powermeter['Zeit'][powermeter['Zeit'].index[-1]]\ntime_delta = time_end - time_start\nprint(str('Versuchsdauer: ' + str(time_delta)))\nprint(str(value_kwh_active_and_idle) + ' kWh')\nprint('')\n\nprint('Förderband Verbrauch mit transportiertem Werkstückträger')\nprint(str(value_kwh_active) + ' kWh')\nprint('')\n\nprint('Förderband Verbrauch ohne transportierten Werkstückträger')\nprint(str(value_kwh_idle) + ' kWh')\nprint('')\n\nprint('Mögliche Einsparung:')\navoidable_power_usage = value_kwh_idle / (value_kwh_active_and_idle / 100)\nprint(str(value_kwh_idle) + ' kWh')\nprint(\"%.2f\" % round(avoidable_power_usage, 2) + ' %')\nprint('')\n\n", "sub_path": "p4/Demonstrator/images/jupyter-notebooks/projects/transport-module-power-analysis/notebooks/data_compare/Analyze_Unnecessary_Worktime.py", "file_name": "Analyze_Unnecessary_Worktime.py", "file_ext": "py", "file_size_in_byte": 20865, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "data_compare.sensor_data_provider.SensorDataProvider", "line_number": 38, "usage_type": "call"}, {"api_name": "data_compare.sensor_data_provider", "line_number": 38, "usage_type": "name"}, {"api_name": "pandas.to_datetime", "line_number": 47, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 48, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "isac.plotting.plot_helper.savePlotLocal", "line_number": 74, "usage_type": "call"}, {"api_name": "isac.plotting.plot_helper", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 121, "usage_type": "name"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 122, "usage_type": "name"}, {"api_name": "isac.plotting.plot_helper.savePlotLocal", "line_number": 137, "usage_type": "call"}, {"api_name": "isac.plotting.plot_helper", "line_number": 137, "usage_type": "name"}, {"api_name": "isac.plotting.plot_helper.savePlotLocal", "line_number": 146, "usage_type": "call"}, {"api_name": "isac.plotting.plot_helper", "line_number": 146, "usage_type": "name"}, {"api_name": "isac.plotting.plot_helper.savePlotLocal", "line_number": 185, "usage_type": "call"}, {"api_name": "isac.plotting.plot_helper", "line_number": 185, "usage_type": "name"}, {"api_name": "isac.plotting.plot_helper.savePlotLocal", "line_number": 195, "usage_type": "call"}, {"api_name": "isac.plotting.plot_helper", "line_number": 195, "usage_type": "name"}, {"api_name": "pandas.concat", "line_number": 214, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 263, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 263, "usage_type": "name"}, {"api_name": "pandas.to_datetime", "line_number": 264, "usage_type": "call"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 265, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 265, "usage_type": "name"}, {"api_name": "isac.plotting.plot_helper.savePlotLocal", "line_number": 278, "usage_type": "call"}, {"api_name": "isac.plotting.plot_helper", "line_number": 278, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 290, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 291, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 292, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 315, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 315, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 319, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 319, "usage_type": "name"}, {"api_name": "isac.plotting.plot_helper.savePlotLocal", "line_number": 321, "usage_type": "call"}, {"api_name": "isac.plotting.plot_helper", "line_number": 321, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 323, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 323, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 329, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 329, "usage_type": "name"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 330, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 330, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.FuncFormatter", "line_number": 336, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 336, "usage_type": "name"}, {"api_name": "isac.plotting.plot_helper.savePlotLocal", "line_number": 345, "usage_type": "call"}, {"api_name": "isac.plotting.plot_helper", "line_number": 345, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 353, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 353, "usage_type": "name"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 355, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 355, "usage_type": "name"}, {"api_name": "isac.plotting.plot_helper.savePlotLocal", "line_number": 363, "usage_type": "call"}, {"api_name": "isac.plotting.plot_helper", "line_number": 363, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 378, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 378, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 380, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 386, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 386, "usage_type": "name"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 387, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 387, "usage_type": "name"}, {"api_name": "isac.plotting.plot_helper.savePlotLocal", "line_number": 397, "usage_type": "call"}, {"api_name": "isac.plotting.plot_helper", "line_number": 397, "usage_type": "name"}, {"api_name": "pandas.concat", "line_number": 410, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 482, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 482, "usage_type": "name"}, {"api_name": "isac.plotting.plot_helper.savePlotLocal", "line_number": 496, "usage_type": "call"}, {"api_name": "isac.plotting.plot_helper", "line_number": 496, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 509, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 509, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 513, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 513, "usage_type": "name"}, {"api_name": "isac.plotting.plot_helper.savePlotLocal", "line_number": 515, "usage_type": "call"}, {"api_name": "isac.plotting.plot_helper", "line_number": 515, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 517, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 517, "usage_type": "name"}]} +{"seq_id": "636114892", "text": "from django.db import models\nfrom django.utils.translation import ugettext_lazy as _\nfrom djinn_contenttypes.registry import CTRegistry\nfrom djinn_contenttypes.models.publishable import PublishableContent\nfrom djinn_contenttypes.models.attachment import ImgAttachment\nfrom djinn_contenttypes.models.commentable import Commentable\n\n\nclass News(PublishableContent, Commentable):\n\n \"\"\" News content type \"\"\"\n\n text = models.TextField(null=True, blank=True)\n\n images = models.ManyToManyField(ImgAttachment)\n\n show_images = models.BooleanField(default=True)\n\n is_global = models.BooleanField(default=False)\n\n create_tmp_object = True\n\n def documents(self):\n\n return self.get_related(relation_type=\"related_document\")\n\n @property\n def date(self):\n\n return self.publish_from or self.created\n\n class Meta:\n app_label = \"djinn_news\"\n\n\nCTRegistry.register(\n \"news\",\n {\"class\": News,\n \"app\": \"djinn_news\",\n \"group_add\": True,\n \"label\": _(\"News\")}\n )\n", "sub_path": "djinn_news/models/news.py", "file_name": "news.py", "file_ext": "py", "file_size_in_byte": 1011, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "djinn_contenttypes.models.publishable.PublishableContent", "line_number": 9, "usage_type": "name"}, {"api_name": "djinn_contenttypes.models.commentable.Commentable", "line_number": 9, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 15, "usage_type": "call"}, {"api_name": "djinn_contenttypes.models.attachment.ImgAttachment", "line_number": 15, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "djinn_contenttypes.registry.CTRegistry.register", "line_number": 36, "usage_type": "call"}, {"api_name": "djinn_contenttypes.registry.CTRegistry", "line_number": 36, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 41, "usage_type": "call"}]} +{"seq_id": "214107344", "text": "import gdata.docs\r\nimport gdata.docs.service\r\nimport gdata.spreadsheet.service\r\nimport re, os\r\n\r\n#Gets data from a specific spreadsheet with the required title\r\ndef getdata(title,chartname):\r\n gd_client = gdata.spreadsheet.service.SpreadsheetsService()\r\n #gd_client.source = 'appspotsite.com'\r\n gd_client.email = 'webwizarddummy@gmail.com'\r\n gd_client.password = 'webwizarddummy'\r\n gd_client.ProgrammaticLogin()\r\n q = gdata.spreadsheet.service.DocumentQuery()\r\n q['title'] = chartname\r\n q['title-exact'] = 'true'\r\n feed = gd_client.GetSpreadsheetsFeed(query=q)\r\n spreadsheet_id = feed.entry[0].id.text.rsplit('/',1)[1]\r\n #spreadsheet_id = '0AsgIq4778ozjdE9oZFZ0Sk4xbU5GOUpzcHhpenZhWEE'\r\n feed = gd_client.GetWorksheetsFeed(spreadsheet_id)\r\n worksheet_id = feed.entry[0].id.text.rsplit('/',1)[1]\r\n rows = gd_client.GetListFeed(spreadsheet_id, worksheet_id).entry\r\n for row in rows:\r\n if row.custom['title'].text == title:\r\n #print row.custom['content'].text\r\n return row\r\n \r\n#getdata('about')", "sub_path": "getdata.py", "file_name": "getdata.py", "file_ext": "py", "file_size_in_byte": 1076, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "gdata.docs.spreadsheet.service.SpreadsheetsService", "line_number": 8, "usage_type": "call"}, {"api_name": "gdata.docs.spreadsheet", "line_number": 8, "usage_type": "attribute"}, {"api_name": "gdata.docs", "line_number": 8, "usage_type": "name"}, {"api_name": "gdata.docs.spreadsheet.service.DocumentQuery", "line_number": 13, "usage_type": "call"}, {"api_name": "gdata.docs.spreadsheet", "line_number": 13, "usage_type": "attribute"}, {"api_name": "gdata.docs", "line_number": 13, "usage_type": "name"}]} +{"seq_id": "610022165", "text": "import boto3\nimport os\nimport requests\nfrom requests_oauthlib import OAuth1\nimport pickle\n\ns3 = boto3.client('s3')\nbucket_name = os.environ['bucket_name']\n\nAPI_KEY = os.environ['API_KEY']\nAPI_SECRET = os.environ['API_SECRET']\nACCESS_TOKEN = os.environ['ACCESS_TOKEN']\nACCESS_TOKEN_SECRET = os.environ['ACCESS_TOKEN_SECRET']\n\ndef main(event, context):\n\n auth = OAuth1(API_KEY, API_SECRET, ACCESS_TOKEN, ACCESS_TOKEN_SECRET)\n\n url = 'https://api.twitter.com/1.1/statuses/home_timeline.json?count=200'\n\n new_tweets = requests.get(url, auth=auth).json()\n\n try:\n pickled = s3.get_object(\n Bucket=bucket_name,\n Key=\"tweet_collection.pickle\"\n )\n tweet_collection = pickle.load(pickled)\n except:\n tweet_collection = []\n\n for new_tweet in new_tweets:\n new = True\n for old_tweet in tweet_collection:\n if new_tweet['id'] == old_tweet['id']:\n new = False\n if new: \n tweet_collection.append(new_tweet)\n \n pickled_new = pickle.dumps(tweet_collection)\n\n response = s3.put_object(\n Bucket=bucket_name,\n Body=pickled_new,\n Key=\"tweet_collection.pickle\"\n )\n return response", "sub_path": "collect_tweets/lambda-handler.py", "file_name": "lambda-handler.py", "file_ext": "py", "file_size_in_byte": 1228, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "boto3.client", "line_number": 7, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 13, "usage_type": "attribute"}, {"api_name": "requests_oauthlib.OAuth1", "line_number": 17, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 21, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 28, "usage_type": "call"}, {"api_name": "pickle.dumps", "line_number": 40, "usage_type": "call"}]} +{"seq_id": "201287163", "text": "from collections import defaultdict\nimport operator\n\nfrom six import string_types as basestring\n\nfrom ...utils import groupby\nfrom ...structure import named_structures, Monosaccharide, Substituent, Anomer, Stem, RingType, SuperClass\nfrom ...algorithms.similarity import (monosaccharide_similarity, has_substituent,\n has_modification, has_monosaccharide, is_generic_monosaccharide)\nfrom ...composition.composition_transform import strip_derivatization\nfrom .synonyms import monosaccharides as monosaccharide_synonyms\n\n\ndef has_ambiguity(node):\n ambiguous = node.stem[0] is Stem.x or node.anomer is Anomer.x or\\\n node.superclass is SuperClass.x or node.ring_type is RingType.x\n return ambiguous\n\n\n# A static copy of monosaccharide names to structures for copy-free comparison\nmonosaccharides = dict(named_structures.monosaccharides)\nmonosaccharides_ordered = sorted(list(monosaccharides.items()), key=lambda x: has_ambiguity(x[1]))\n\n\ndef get_preferred_name(name, selector=min, key=len):\n '''\n Given a name, of its synonyms, find the name that satisfies the `selector`\n criterion function (:func:`min`) based on some `key` function of the name (:func:`len`)\n\n Parameters\n ----------\n name: str\n Given name to compare to synonyms\n selector: function\n Function to use to select the preferred name by some statistic\n key: function\n Function to use to convert names into statistics\n\n Returns\n -------\n str\n '''\n preferred_name = selector(monosaccharide_synonyms.get(name, [name]) + [name], key=key)\n return preferred_name\n\n\ndef is_a(node, target, tolerance=0, include_modifications=True, include_substituents=True, exact=True, short_circuit=False):\n '''\n Perform a semi-fuzzy match between `node` and `target` where node is the unqualified\n residue queried and target is the known residue to be matched against\n\n Parameters\n ----------\n node: Monosaccharide or Substituent\n Object to be identified\n target: Monosaccharide, Substituent or str\n The reference type. May be a |str| object which is used to look up a |Monosaccharide| by name in\n :obj:`glypy.monosaccharides`\n tolerance: int\n The error tolerance for the search\n include_modifications: bool\n Whether or not to include modifications in comparison. Defaults to |True|\n include_substituents: bool\n Whether or not to include substituents in comparison. Defaults to |True|\n exact: bool\n Whether or not to penalize for unmatched attachments. Defaults to |True|\n\n Returns\n -------\n bool\n\n '''\n res = 0\n qs = 0\n if isinstance(target, basestring):\n target = monosaccharides[target]\n\n if isinstance(node, Substituent):\n if not isinstance(target, Substituent):\n return False\n else:\n res += node.name == target.name\n qs += 1\n else:\n if not isinstance(target, Monosaccharide):\n return False\n res, qs = monosaccharide_similarity(node, target, include_modifications=include_modifications,\n include_substituents=include_substituents,\n include_children=False, exact=exact,\n short_circuit_after=tolerance if short_circuit else None)\n threshold = (qs - res) <= tolerance\n return threshold\n\n\ndef identify(node, blacklist=None, tolerance=0, include_modifications=True, include_substituents=True):\n '''\n Attempt to find a common usage name for the given |Monosaccharide|, `node`. The name is determined by\n performing an incremental comparison of the traits of `node` with each named residue in the database\n accessed at :obj:`glypy.monosaccharides`.\n\n Parameters\n ----------\n node: Monosaccharide\n Object to be identified\n blacklist: list\n The set of all monosaccharides to not attempt matching against, because they are too general.\n tolerance: int\n The error tolerance for the search\n include_modifications: bool\n Whether or not to include modifications in comparison. Defaults to |True|\n include_substituents: bool\n Whether or not to include substituents in comparison. Defaults to |True|\n\n Returns\n -------\n str\n\n Raises\n ------\n IdentifyException:\n When a suitable name cannot be found.\n\n See Also\n --------\n is_a\n preferred_name\n monosaccharide_similarity\n '''\n if blacklist is None:\n blacklist = {\"Pen\", \"Hex\", \"Hep\", \"Oct\", \"Non\"}\n for name, structure in monosaccharides_ordered:\n if name in blacklist:\n continue\n if is_a(node, structure, tolerance, include_modifications, include_substituents):\n return get_preferred_name(name)\n raise IdentifyException(\"Could not identify {}\".format(node))\n\n\nclass IdentifyException(Exception):\n pass\n\n\ndef naive_name_monosaccharide(monosaccharide):\n '''\n Generate a generic name for `monosaccharide`, based loosely on IUPAC\n naming schema without including information about linkage.\n\n The tendency for monosaccharides of superclass > 7 to have special names,\n which will be used preferentially if possible.\n\n Parameters\n ----------\n monosaccharide: Monosaccharide\n\n Returns\n -------\n str:\n A simple name based on `SuperClass`, modifications, and substituents.\n\n See Also\n --------\n :func:`glypy.io.nomenclature.identity.identify`\n\n '''\n try:\n c = monosaccharide.clone()\n if not isinstance(c, Monosaccharide):\n return None\n strip_derivatization(c)\n try:\n if monosaccharide.superclass.value > 6:\n return identify(c, tolerance=0)\n except:\n pass\n c.anomer = None\n return identify(c)\n except IdentifyException:\n try:\n c.stem = None\n c.configuration = None\n return identify(c)\n except IdentifyException:\n return \"\".join(mod.name for mod in list(c.modifications.values()) if mod.name != 'aldi') +\\\n c.superclass.name.title() + ''.join([''.join(map(str.title, subst.name.split(\"_\")))[:3]\n for p, subst in c.substituents()])\n\n\ndef split_along_axis(monosaccharides, axis):\n getter = operator.attrgetter(axis)\n groups = groupby(monosaccharides, getter)\n return groups\n\n\ndef residue_list_to_tree(monosaccharides, axes=('anomer', 'superclass', 'stem', 'configuration')):\n root = split_along_axis(monosaccharides, axes[0])\n if len(axes) > 1:\n for level, group in list(root.items()):\n root[level] = residue_list_to_tree(group, axes[1:])\n return root\n", "sub_path": "glypy/io/nomenclature/identity.py", "file_name": "identity.py", "file_ext": "py", "file_size_in_byte": 6836, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "structure.Stem.x", "line_number": 15, "usage_type": "attribute"}, {"api_name": "structure.Stem", "line_number": 15, "usage_type": "name"}, {"api_name": "structure.Anomer.x", "line_number": 15, "usage_type": "attribute"}, {"api_name": "structure.Anomer", "line_number": 15, "usage_type": "name"}, {"api_name": "structure.SuperClass.x", "line_number": 16, "usage_type": "attribute"}, {"api_name": "structure.SuperClass", "line_number": 16, "usage_type": "name"}, {"api_name": "structure.RingType.x", "line_number": 16, "usage_type": "attribute"}, {"api_name": "structure.RingType", "line_number": 16, "usage_type": "name"}, {"api_name": "structure.named_structures.monosaccharides", "line_number": 21, "usage_type": "attribute"}, {"api_name": "structure.named_structures", "line_number": 21, "usage_type": "name"}, {"api_name": "synonyms.monosaccharides.get", "line_number": 43, "usage_type": "call"}, {"api_name": "synonyms.monosaccharides", "line_number": 43, "usage_type": "name"}, {"api_name": "six.string_types", "line_number": 75, "usage_type": "argument"}, {"api_name": "structure.Substituent", "line_number": 78, "usage_type": "argument"}, {"api_name": "structure.Substituent", "line_number": 79, "usage_type": "argument"}, {"api_name": "structure.Monosaccharide", "line_number": 85, "usage_type": "argument"}, {"api_name": "algorithms.similarity.monosaccharide_similarity", "line_number": 87, "usage_type": "call"}, {"api_name": "structure.Monosaccharide", "line_number": 167, "usage_type": "argument"}, {"api_name": "composition.composition_transform.strip_derivatization", "line_number": 169, "usage_type": "call"}, {"api_name": "operator.attrgetter", "line_number": 189, "usage_type": "call"}, {"api_name": "utils.groupby", "line_number": 190, "usage_type": "call"}]} +{"seq_id": "484955809", "text": "# coding: utf-8\nimport logging\n\nimport requests\n\nfrom constants import DEFAULT_IMAGE\nfrom raters import get_released_date\n\n\ndef get_data(user_params):\n params = {\n \"plot\": \"short\",\n \"r\": \"json\"\n }\n\n params.update(user_params)\n req = requests.get(\"http://www.omdbapi.com/\", params=params)\n\n if req.status_code != 200:\n logging.warn(\"Status %s:: Cannot get data for params %s\", req.status_code, params)\n return {}\n\n data = req.json()\n if data[\"Response\"] == 'False' or data.get(\"Type\") == \"episode\" or data[\"imdbRating\"] == \"N/A\":\n return {}\n\n return {\n \"title\": data[\"Title\"],\n \"imdb_id\": data[\"imdbID\"],\n \"description\": data[\"Plot\"],\n \"release_date\": get_released_date(data),\n \"genre\": data[\"Genre\"],\n \"director\": data[\"Director\"],\n \"cast\": data[\"Actors\"],\n \"imdb_rating\": data[\"imdbRating\"],\n \"poster\": DEFAULT_IMAGE\n }\n\n\ndef get_data_by_title_and_year(title, year):\n return get_data({\"t\": title, \"y\": year})\n\n\ndef get_data_by_imdb_id(imdb_id):\n return get_data({\"i\": imdb_id})\n", "sub_path": "omdbapi.py", "file_name": "omdbapi.py", "file_ext": "py", "file_size_in_byte": 1110, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "requests.get", "line_number": 17, "usage_type": "call"}, {"api_name": "logging.warn", "line_number": 20, "usage_type": "call"}, {"api_name": "raters.get_released_date", "line_number": 31, "usage_type": "call"}, {"api_name": "constants.DEFAULT_IMAGE", "line_number": 36, "usage_type": "name"}]} +{"seq_id": "163009126", "text": "import os\nfrom os.path import join\nimport json\nimport shutil\nimport argparse\nimport multiprocessing\nfrom itertools import repeat\nfrom tqdm import tqdm\nfrom auxiliaries import validate_exists_and_dir, create_or_recreate_dir, imap_wrapper\n\n\ndef filter_invalid_images(images_dir, descs_dir, inv_images_dir, inv_descs_dir, num_processes):\n validate_exists_and_dir(images_dir, 'images_dir')\n validate_exists_and_dir(descs_dir, 'descs_dir')\n # Create the result dirs\n create_or_recreate_dir(inv_images_dir)\n create_or_recreate_dir(inv_descs_dir)\n\n # Find all the descriptions\n images_fnames = os.listdir(images_dir)\n descs_fnames = os.listdir(descs_dir)\n\n # Find which descriptions are invalid\n # And mark them and their corresponding image for moving\n src_paths = []\n dst_paths = []\n for image_fname, desc_fname in zip(images_fnames, descs_fnames):\n valid_desc = True\n desc_path = join(descs_dir, desc_fname)\n # Load the json\n desc = json.load(open(desc_path))\n # Validate the description\n try:\n label = desc['meta']['clinical']['benign_malignant']\n if label not in {'benign', 'malignant'}:\n valid_desc = False\n except KeyError:\n valid_desc = False\n if not valid_desc:\n # The description is invalid.\n # Mark it and its corresponding image for moving\n image_path = join(images_dir, image_fname)\n src_paths += [image_path, desc_path]\n dst_paths += [join(inv_images_dir, image_fname), join(inv_descs_dir, desc_fname)]\n\n # Move the invalid descriptions and images to the filtered directories\n p = multiprocessing.Pool(processes=num_processes)\n list(tqdm(p.imap(imap_wrapper, zip(repeat(shutil.move), src_paths, dst_paths)), total=len(src_paths), desc='Filtering invalid images'))\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument('--images-dir', type=str, help='Directory which holds the images, and only them', required=True)\n parser.add_argument('--descs-dir', type=str, help='Directory which holds the descriptions of the images and only them', required=True)\n parser.add_argument('--inv-dir', type=str, help='Directory to store the filtered out invalid data', required=True)\n parser.add_argument('--p', type=int, help='Number of processes to use in parallel', default=16)\n args = parser.parse_args()\n\n inv_images_dir = join(args.inv_dir, 'images')\n inv_descs_dir = join(args.inv_dir, 'descs')\n\n filter_invalid_images(images_dir=args.images_dir, descs_dir=args.descs_dir, inv_images_dir=inv_images_dir, inv_descs_dir=inv_descs_dir, num_processes=args.p)\n", "sub_path": "filter_invalid.py", "file_name": "filter_invalid.py", "file_ext": "py", "file_size_in_byte": 2718, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "auxiliaries.validate_exists_and_dir", "line_number": 13, "usage_type": "call"}, {"api_name": "auxiliaries.validate_exists_and_dir", "line_number": 14, "usage_type": "call"}, {"api_name": "auxiliaries.create_or_recreate_dir", "line_number": 16, "usage_type": "call"}, {"api_name": "auxiliaries.create_or_recreate_dir", "line_number": 17, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 20, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "json.load", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 47, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 48, "usage_type": "call"}, {"api_name": "auxiliaries.imap_wrapper", "line_number": 48, "usage_type": "argument"}, {"api_name": "itertools.repeat", "line_number": 48, "usage_type": "call"}, {"api_name": "shutil.move", "line_number": 48, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 60, "usage_type": "call"}]} +{"seq_id": "447025435", "text": "# Copyright © 2012-2023 jrnl contributors\n# License: https://www.gnu.org/licenses/gpl-3.0.html\n\nimport logging\nimport os\nimport subprocess\nimport sys\nimport tempfile\nfrom pathlib import Path\n\nfrom jrnl.exception import JrnlException\nfrom jrnl.messages import Message\nfrom jrnl.messages import MsgStyle\nfrom jrnl.messages import MsgText\nfrom jrnl.os_compat import on_windows\nfrom jrnl.os_compat import split_args\nfrom jrnl.output import print_msg\n\n\ndef get_text_from_editor(config: dict, template: str = \"\") -> str:\n suffix = \".jrnl\"\n if config[\"template\"]:\n template_filename = Path(config[\"template\"]).name\n suffix = \"-\" + template_filename\n filehandle, tmpfile = tempfile.mkstemp(prefix=\"jrnl\", text=True, suffix=suffix)\n os.close(filehandle)\n\n with open(tmpfile, \"w\", encoding=\"utf-8\") as f:\n if template:\n f.write(template)\n\n try:\n subprocess.call(split_args(config[\"editor\"]) + [tmpfile])\n except FileNotFoundError:\n raise JrnlException(\n Message(\n MsgText.EditorMisconfigured,\n MsgStyle.ERROR,\n {\"editor_key\": config[\"editor\"]},\n )\n )\n\n with open(tmpfile, \"r\", encoding=\"utf-8\") as f:\n raw = f.read()\n os.remove(tmpfile)\n\n if not raw:\n raise JrnlException(Message(MsgText.NoTextReceived, MsgStyle.NORMAL))\n\n return raw\n\n\ndef get_text_from_stdin() -> str:\n print_msg(\n Message(\n MsgText.WritingEntryStart,\n MsgStyle.TITLE,\n {\n \"how_to_quit\": MsgText.HowToQuitWindows\n if on_windows()\n else MsgText.HowToQuitLinux\n },\n )\n )\n\n try:\n raw = sys.stdin.read()\n except KeyboardInterrupt:\n logging.error(\"Write mode: keyboard interrupt\")\n raise JrnlException(\n Message(MsgText.KeyboardInterruptMsg, MsgStyle.ERROR_ON_NEW_LINE),\n Message(MsgText.JournalNotSaved, MsgStyle.WARNING),\n )\n\n return raw\n", "sub_path": "jrnl/editor.py", "file_name": "editor.py", "file_ext": "py", "file_size_in_byte": 2030, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pathlib.Path", "line_number": 23, "usage_type": "call"}, {"api_name": "tempfile.mkstemp", "line_number": 25, "usage_type": "call"}, {"api_name": "os.close", "line_number": 26, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 33, "usage_type": "call"}, {"api_name": "jrnl.os_compat.split_args", "line_number": 33, "usage_type": "call"}, {"api_name": "jrnl.exception.JrnlException", "line_number": 35, "usage_type": "call"}, {"api_name": "jrnl.messages.Message", "line_number": 36, "usage_type": "call"}, {"api_name": "jrnl.messages.MsgText.EditorMisconfigured", "line_number": 37, "usage_type": "attribute"}, {"api_name": "jrnl.messages.MsgText", "line_number": 37, "usage_type": "name"}, {"api_name": "jrnl.messages.MsgStyle.ERROR", "line_number": 38, "usage_type": "attribute"}, {"api_name": "jrnl.messages.MsgStyle", "line_number": 38, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 45, "usage_type": "call"}, {"api_name": "jrnl.exception.JrnlException", "line_number": 48, "usage_type": "call"}, {"api_name": "jrnl.messages.Message", "line_number": 48, "usage_type": "call"}, {"api_name": "jrnl.messages.MsgText.NoTextReceived", "line_number": 48, "usage_type": "attribute"}, {"api_name": "jrnl.messages.MsgText", "line_number": 48, "usage_type": "name"}, {"api_name": "jrnl.messages.MsgStyle.NORMAL", "line_number": 48, "usage_type": "attribute"}, {"api_name": "jrnl.messages.MsgStyle", "line_number": 48, "usage_type": "name"}, {"api_name": "jrnl.output.print_msg", "line_number": 54, "usage_type": "call"}, {"api_name": "jrnl.messages.Message", "line_number": 55, "usage_type": "call"}, {"api_name": "jrnl.messages.MsgText.WritingEntryStart", "line_number": 56, "usage_type": "attribute"}, {"api_name": "jrnl.messages.MsgText", "line_number": 56, "usage_type": "name"}, {"api_name": "jrnl.messages.MsgStyle.TITLE", "line_number": 57, "usage_type": "attribute"}, {"api_name": "jrnl.messages.MsgStyle", "line_number": 57, "usage_type": "name"}, {"api_name": "jrnl.os_compat.on_windows", "line_number": 60, "usage_type": "call"}, {"api_name": "jrnl.messages.MsgText.HowToQuitWindows", "line_number": 59, "usage_type": "attribute"}, {"api_name": "jrnl.messages.MsgText", "line_number": 59, "usage_type": "name"}, {"api_name": "jrnl.messages.MsgText.HowToQuitLinux", "line_number": 61, "usage_type": "attribute"}, {"api_name": "jrnl.messages.MsgText", "line_number": 61, "usage_type": "name"}, {"api_name": "sys.stdin.read", "line_number": 67, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 67, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 69, "usage_type": "call"}, {"api_name": "jrnl.exception.JrnlException", "line_number": 70, "usage_type": "call"}, {"api_name": "jrnl.messages.Message", "line_number": 71, "usage_type": "call"}, {"api_name": "jrnl.messages.MsgText.KeyboardInterruptMsg", "line_number": 71, "usage_type": "attribute"}, {"api_name": "jrnl.messages.MsgText", "line_number": 71, "usage_type": "name"}, {"api_name": "jrnl.messages.MsgStyle.ERROR_ON_NEW_LINE", "line_number": 71, "usage_type": "attribute"}, {"api_name": "jrnl.messages.MsgStyle", "line_number": 71, "usage_type": "name"}, {"api_name": "jrnl.messages.Message", "line_number": 72, "usage_type": "call"}, {"api_name": "jrnl.messages.MsgText.JournalNotSaved", "line_number": 72, "usage_type": "attribute"}, {"api_name": "jrnl.messages.MsgText", "line_number": 72, "usage_type": "name"}, {"api_name": "jrnl.messages.MsgStyle.WARNING", "line_number": 72, "usage_type": "attribute"}, {"api_name": "jrnl.messages.MsgStyle", "line_number": 72, "usage_type": "name"}]} +{"seq_id": "566064649", "text": "import json\nimport requests\nfrom PIL import Image\nfrom bs4 import BeautifulSoup\nfrom requests import RequestException\n\nfrom comiccrawl.compress import compress_folder\nfrom comiccrawl.utils import make_new_folder, save_image_from_url, delete_folder\n\n\ndef crawl_chapter(start_url, end_url, next_text, base_folder, chapter_num):\n \"\"\"\n Iterates over all the images of a chapter in the webcomic and download\n each of it pages.\n\n This function will create a new folder using the chapter_num parameter and\n all the images will be downloaded there.\n\n If there is some text in a page it will call the text2pic service\n TODO: take the text2pic host from config file. now is on localhost:5000\n\n After downloading the whole chapter into a folder it will make a CBZ\n file from it,\n\n :param start_url: first page of this chapter\n :param end_url: last page\n :param next_text: the text of the \"next page\" link to iterate the chapter\n :param base_folder: the folder where the webcomic is being downloaded\n :param chapter_num: the number of this chapter\n :return:\n\n \"\"\"\n current_page = start_url\n page_index = 1\n new_folder = make_new_folder(base_folder, chapter_num)\n\n image_index = 0\n while current_page != end_url:\n # get current page\n try:\n page = requests.get(current_page)\n if page.status_code != requests.codes.ok:\n break\n\n soup = BeautifulSoup(page.text, \"html.parser\")\n print('curr: ' + current_page)\n # get the url for the \"Next page\" link\n next_link = soup.find('a', href=True, text=next_text)\n if next_link:\n next_link = next_link['href']\n else:\n next_link = ''\n print('next link: ' + next_link)\n # get the current image\n img_link = soup.find('div', {'class': 'comic-table'}).findNext('div', {'id': 'comic'}).find('img')['src']\n print('img link: ' + img_link)\n # some pages have text\n #
\n image_text = soup.find('div', {'class': 'entry'}).get_text()\n print('image_text: ' + image_text)\n img_path = save_image_from_url(img_link, image_index, new_folder)\n\n image_text = image_text.strip()\n if image_text != '':\n # I want the text printed in the same image size as the last downloaded image\n # so I extract the size from it.\n im = Image.open(img_path)\n image_size = im.size # (width,height) tuple\n # get the image from the flask utility\n try:\n get_image_from_text2pic(image_text, image_size, image_index, new_folder)\n except RequestException as ex:\n print(\"Exception connecting to text2pic\")\n print(ex)\n # next loop step\n current_page = next_link\n page_index += 1\n image_index += 1\n except RequestException as ex: # this covers everything\n print(\"Couldn´t get page\" + current_page)\n print(ex)\n break\n except Exception as inst:\n print(type(inst)) # the exception instance\n print(inst.args) # arguments stored in .args\n print(inst) # __str__ allows args to be printed directly,\n # but may be overridden in exception subclasses#\n # end of chapter loop\n # create zip\n compress_folder(new_folder)\n delete_folder(new_folder)\n\n\ndef get_image_from_text2pic(text, size, image_index, new_folder):\n \"\"\"\n\n Sends some text to the text2pic service that embeds the text in a series\n of images of the given size\n\n The images are named using the image_index so they will appear after\n the original image\n\n :param text: the text to put into images\n :param size: the image size\n :param image_index: current image index\n :param new_folder: destination folder to store images\n :return:\n \"\"\"\n text2pic_host = 'http://localhost:5000'\n # compose input json for text2pic\n w_margin = size[0] * 0.15\n h_margin = size[1] * 0.15\n data = {'text': text, 'width': size[0], 'height': size[1],\n 'margin-width': w_margin, 'margin-height': h_margin,\n 'font': 'arial.ttf', 'font-size': 32}\n json_data = json.dumps(data)\n headers = {'Content-type': 'application/json'}\n img_request = requests.post(text2pic_host + '/text2pic', headers=headers, data=json_data, stream=True)\n if img_request.status_code == 200:\n # get json body and iterate\n # the json response is an array of image relative urls\n json_resp = img_request.json()\n for item in json_resp['images']:\n current_img_url = text2pic_host + item['filename']\n save_image_from_url(current_img_url, image_index, new_folder, True)\n else:\n # bad request\n print('Error in text from ' + image_index + ' image in chapter ' + new_folder + ')')\n print(img_request.text)\n\n\ndef get_end_of_chapter(start_url, next_chapter):\n page = requests.get(start_url)\n if page.status_code != requests.codes.ok:\n return ''\n\n soup = BeautifulSoup(page.text, \"html.parser\")\n # get the url for the \"Next page\" link\n next_link = soup.find('a', href=True, string=next_chapter)\n if next_link:\n next_link = next_link['href']\n else:\n next_link = ''\n return next_link\n", "sub_path": "comiccrawl/chapterthread.py", "file_name": "chapterthread.py", "file_ext": "py", "file_size_in_byte": 5525, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "comiccrawl.utils.make_new_folder", "line_number": 35, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 41, "usage_type": "call"}, {"api_name": "requests.codes", "line_number": 42, "usage_type": "attribute"}, {"api_name": "bs4.BeautifulSoup", "line_number": 45, "usage_type": "call"}, {"api_name": "comiccrawl.utils.save_image_from_url", "line_number": 61, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 67, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 67, "usage_type": "name"}, {"api_name": "requests.RequestException", "line_number": 72, "usage_type": "name"}, {"api_name": "requests.RequestException", "line_number": 79, "usage_type": "name"}, {"api_name": "comiccrawl.compress.compress_folder", "line_number": 90, "usage_type": "call"}, {"api_name": "comiccrawl.utils.delete_folder", "line_number": 91, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 116, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 118, "usage_type": "call"}, {"api_name": "comiccrawl.utils.save_image_from_url", "line_number": 125, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 133, "usage_type": "call"}, {"api_name": "requests.codes", "line_number": 134, "usage_type": "attribute"}, {"api_name": "bs4.BeautifulSoup", "line_number": 137, "usage_type": "call"}]} +{"seq_id": "199421809", "text": "# -*- coding: utf-8 -*-\n#\n# This file is part of hepcrawl.\n# Copyright (C) 2015, 2016, 2017 CERN.\n#\n# hepcrawl is a free software; you can redistribute it and/or modify it\n# under the terms of the Revised BSD License; see LICENSE file for\n# more details.\n\n\"\"\"Spider for PHENIX.\"\"\"\n\nfrom __future__ import absolute_import, division, print_function\n\nfrom urlparse import urljoin\n\nfrom scrapy import Request\nfrom scrapy.spiders import XMLFeedSpider\n\nfrom ..items import HEPRecord\nfrom ..loaders import HEPLoader\nfrom ..utils import ParsedItem\n\n\nclass PhenixSpider(XMLFeedSpider):\n\n \"\"\"PHENIX crawler\n\n Scrapes theses metadata from `PHENIX experiment web page`_.\n\n 1. ``PhenixSpider.parse()`` iterates through every record on the html page and yields\n a ``HEPRecord``.\n\n Examples:\n ::\n\n $ scrapy crawl phenix\n\n Using source file and output directory::\n\n $ scrapy crawl phenix -a source_file=file://`pwd`/tests/responses/phenix/test_list.html -s \"JSON_OUTPUT_DIR=tmp/\"\n\n .. _PHENIX experiment web page:\n http://www.phenix.bnl.gov/WWW/talk/theses.php\n \"\"\"\n\n name = 'phenix'\n start_urls = [\"http://www.phenix.bnl.gov/WWW/talk/theses.php\"]\n domain = \"http://www.phenix.bnl.gov\"\n iterator = \"html\"\n itertag = \"//table//td/ul/li\"\n\n def __init__(self, source_file=None, *args, **kwargs):\n \"\"\"Construct PHENIX spider\"\"\"\n super(PhenixSpider, self).__init__(*args, **kwargs)\n self.source_file = source_file\n\n def start_requests(self):\n \"\"\"You can also run the spider on local test files\"\"\"\n if self.source_file:\n yield Request(self.source_file)\n elif self.start_urls:\n for url in self.start_urls:\n yield Request(url)\n\n @staticmethod\n def parse_datablock(node):\n \"\"\"Get data out of the text block where there's\n title, affiliation and year\n \"\"\"\n datablock = node.xpath(\"./text()\").extract()[0]\n datalist = datablock.strip().split(\",\")\n\n thesis_type = None\n if \"Ph.D.\" in datablock:\n thesis_type = \"PhD\"\n\n title = datablock.split('\"')[1]\n datalist = [el for el in datalist if \"archive\" not in el]\n year = datalist.pop().strip()\n affline = datalist.pop().strip()\n stop_words = {\"Ph.D.\", \"Master\", \"thesis\", \"at\"}\n affiliation = \" \".join(\n [w for w in affline.split() if w not in stop_words])\n\n return title, year, affiliation, thesis_type\n\n def get_authors(self, node):\n \"\"\"Return authors dictionary \"\"\"\n author = node.xpath(\"./b/text()\").extract()\n authors = []\n _, _, affiliation, _ = self.parse_datablock(node)\n\n for aut in author:\n authors.append({\n 'raw_name': aut,\n 'affiliations': [{\"value\": affiliation}]\n })\n\n return authors\n\n def add_fft_file(self, pdf_files, file_access, file_type):\n \"\"\"Create a structured dictionary and add to ``files`` item.\"\"\"\n file_dicts = []\n for link in pdf_files:\n file_dict = {\n \"access\": file_access,\n \"description\": self.name.title(),\n \"url\": urljoin(self.domain, link),\n \"type\": file_type,\n }\n file_dicts.append(file_dict)\n return file_dicts\n\n def parse_node(self, response, node):\n \"\"\"Parse PHENIX web page into a ``HEPrecord``.\"\"\"\n record = HEPLoader(item=HEPRecord(), selector=node, response=response)\n title, year, _, thesis_type = self.parse_datablock(node)\n\n if not thesis_type:\n return None\n\n pdf_files = node.xpath(\".//a/@href\").extract()\n record.add_value('additional_files', self.add_fft_file(pdf_files, \"HIDDEN\", \"Fulltext\"))\n record.add_value('authors', self.get_authors(node))\n record.add_value('date_published', year)\n record.add_value('thesis', {'degree_type': thesis_type})\n record.add_value('title', title)\n record.add_value('urls', self.start_urls)\n record.add_value('source', 'PHENIX')\n record.add_value('collections', ['HEP', 'THESIS'])\n\n parsed_item = ParsedItem(\n record=record.load_item(),\n record_format='hepcrawl',\n )\n\n return parsed_item\n", "sub_path": "hepcrawl/spiders/phenix_spider.py", "file_name": "phenix_spider.py", "file_ext": "py", "file_size_in_byte": 4350, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "scrapy.spiders.XMLFeedSpider", "line_number": 24, "usage_type": "name"}, {"api_name": "scrapy.Request", "line_number": 60, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 63, "usage_type": "call"}, {"api_name": "urlparse.urljoin", "line_number": 108, "usage_type": "call"}, {"api_name": "loaders.HEPLoader", "line_number": 116, "usage_type": "call"}, {"api_name": "items.HEPRecord", "line_number": 116, "usage_type": "call"}, {"api_name": "utils.ParsedItem", "line_number": 132, "usage_type": "call"}]} +{"seq_id": "31223550", "text": "from __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\nimport os\nimport sys\nimport random\n\nimport librosa\nimport numpy as np\nimport tensorflow as tf\n\nfrom utils import audio_process\n\n\nQUANTIZATION_CHANNELS = 256\nMU = QUANTIZATION_CHANNELS - 1\n'''\n VCTK specifications:\n - 109 native speakers of English with various accents.\n - Each speaker reads out about 400 stences: most of which were selected from a newspaper\n plus Rainbow Passage and an elicitation paragraph intended to identify the speaker's\n accents.\n - 96 kHz versions of the recordings.\n'''\n\n\n\n# The maximum value in VCTK waveforms is 1.1599158.\n# The minimum value in VCTK waveforms is -1.1816651.\nMAX = 1.2\n\n\ndef get_dataset_list(dir_name):\n audio_file_names = []\n txt_file_names = []\n audio_full_paths = {}\n txt_full_paths = {}\n dataset_list = []\n for [path, dir, files] in os.walk(dir_name):\n for file_name in files:\n ext = os.path.splitext(file_name)[-1]\n if ext == \".wav\":\n full_file_name = path + \"/\" + file_name\n full_file_name_ = full_file_name.split(\"/\")\n file_name_ = (\"/\" +\n full_file_name_[-2] +\n \"/\" +\n full_file_name_[-1]).replace(\".wav\", \"\")\n audio_full_paths[file_name_] = full_file_name\n audio_file_names.append(file_name_)\n elif ext == \".txt\":\n full_file_name = path + \"/\" + file_name\n full_file_name_ = full_file_name.split(\"/\")\n file_name_ = (\"/\" +\n full_file_name_[-2] +\n \"/\" +\n full_file_name_[-1]).replace(\".txt\", \"\")\n txt_full_paths[file_name_] = full_file_name\n txt_file_names.append(file_name_)\n audio_file_names = set(audio_file_names)\n txt_file_names = set(txt_file_names)\n file_names = list(audio_file_names & txt_file_names)\n file_names.sort()\n for file_name in file_names:\n file_name_ = file_name.split(\"/\")\n audio_full_path = audio_full_paths[file_name]\n txt_full_path = txt_full_paths[file_name]\n dataset_list.append({\"audio_file_path\": audio_full_path,\n \"txt_file_path\": txt_full_path,\n \"personal_id\": file_name_[1]})\n return dataset_list\n\n\ndef get_char_dict(dataset_list):\n txt_file_names = [data[\"txt_file_path\"] for data in dataset_list]\n char_dict = set()\n for file_name in txt_file_names:\n with open(file=file_name, mode=\"rt\", encoding=\"utf-8\") as f:\n for line in f:\n char_dict.update(set(line))\n char_dict = list(set(char_dict))\n reverse_dict = {char_dict[i]: i for i in range(len(char_dict))}\n return char_dict, reverse_dict\n\n\ndef get_id_dict(dataset_list):\n id_dict = [data[\"personal_id\"] for data in dataset_list]\n id_dict = list(set(id_dict))\n reverse_id_dict = {id_dict[i]: i for i in range(len(id_dict))}\n return id_dict, reverse_id_dict\n\n\ndef get_txt(file_path):\n txt_list = []\n with open(file=file_path, mode=\"rt\", encoding=\"utf-8\") as f:\n for line in f:\n txt_list.append(line)\n txt_ = \" \".join(txt_list)\n return txt_\n\n\nclass DatasetLoader(object):\n def __init__(self,\n dir_name,\n batch_size,\n samplinig_rate,\n quantization_channels,\n sample_size,\n silence_threshold):\n self.dir_name = dir_name\n self.dataset_list = get_dataset_list(self.dir_name)\n self.batch_size = batch_size\n self.quantization_channels = quantization_channels\n self.sample_size = sample_size\n self.silence_threshold = silence_threshold\n # For batch generating or training. Not implemented yet\n self.audio_cutting = False\n self.max_ = MAX\n self.mu = quantization_channels - 1\n self.sampling_rate = samplinig_rate\n self.num_data = len(self.dataset_list)\n self.char_dict, self.reverse_dict = get_char_dict(self.dataset_list)\n self.txt_one_hot_table = np.eye(len(self.char_dict))\n self.id_dict, self.reverse_id_dict= get_id_dict(self.dataset_list)\n self.id_one_hot_table = np.eye(len(self.id_dict))\n\n def generate_data(self):\n while True:\n random_dataset_list = self.dataset_list[:]\n random.shuffle(random_dataset_list)\n for dataset in random_dataset_list:\n file_path = dataset[\"audio_file_path\"]\n audio, _ = librosa.load(file_path,\n sr=self.sampling_rate)\n # For batch generating or training. Not implemented yet\n if self.audio_cutting:\n length = np.shape(audio)[0]\n if length >= self.sample_size:\n audio = audio[:self.sample_size]\n else:\n audio = np.pad(audio, pad_width=[0, self.sample_size - length],\n mode=\"constant\")\n if self.silence_threshold is not None:\n audio = audio_process.remove_silence(audio=audio,\n threshold=self.silence_threshold)\n id_ = self.id_encode(dataset[\"personal_id\"])\n yield audio, id_, file_path\n\n def get_batch_(self):\n random_batch_idx = np.random.choice(a=self.num_data,\n size=self.batch_size,\n replace=False)\n dataset_list_ = []\n for i in random_batch_idx:\n dataset_list_.append(self.dataset_list[i])\n audio_batch = []\n for i in random_batch_idx:\n audio, _ = librosa.load(self.dataset_list[i][\"audio_file_path\"],\n sr=self.sampling_rate)\n quantized_audio = audio_process.quantization(audio,\n self.quantization_channels,\n MAX,\n self.mu)\n one_hot_audio = audio_process.audio_one_hot_encoding(quantized_audio,\n self.quantization_channels)\n audio_batch.append(one_hot_audio)\n txt_batch = []\n for i in random_batch_idx:\n txt_ = get_txt(file_path=self.dataset_list[i][\"txt_file_path\"])\n txt_one_hot = []\n for ch in txt_:\n one_hot = self.txt_one_hot_encoding(ch)\n txt_one_hot.append(one_hot)\n txt_batch.append(txt_one_hot)\n id_batch = []\n for i in random_batch_idx:\n id_ = self.id_one_hot_encoding(self.dataset_list[i][\"personal_id\"])\n id_batch.append(id_)\n return dataset_list_, \\\n np.asarray(audio_batch), \\\n np.asarray(txt_batch), \\\n np.asarray(id_batch)\n\n def create_dataset(self):\n self.dataset = tf.data.Dataset.\\\n from_generator(generator=self.generate_data,\n output_types=(tf.float32,\n tf.float32,\n tf.string),\n output_shapes=(tf.TensorShape([None]),\n tf.TensorShape([len(self.id_dict)]),\n None))\n self.next_audio_batch, self.next_id_batch, self.next_file_paths = \\\n self.dataset.batch(self.batch_size).make_one_shot_iterator().get_next()\n\n def get_next_batch(self):\n return self.next_audio_batch, self.next_id_batch, self.next_file_paths\n\n def txt_one_hot_encoding(self, character):\n return self.txt_one_hot_table[self.reverse_dict[character]]\n\n def id_encode(self, id):\n return self.id_one_hot_table[self.reverse_id_dict[id]]\n\n def id_decode(self, one_hot):\n return self.id_dict[int(np.argmax(one_hot))]\n\n def get_char_dict(self):\n return self.char_dict\n\n def get_id_dict(self):\n return self.id_dict\n\n def get_id_cardinality(self):\n return len(self.id_dict)\n\n def generate_audio_from_one_hot(self, one_hot):\n print(\"not implemented\")\n\n\n", "sub_path": "dataset_loader.py", "file_name": "dataset_loader.py", "file_ext": "py", "file_size_in_byte": 8578, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "os.walk", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.eye", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.eye", "line_number": 125, "usage_type": "call"}, {"api_name": "random.shuffle", "line_number": 130, "usage_type": "call"}, {"api_name": "librosa.load", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 141, "usage_type": "call"}, {"api_name": "utils.audio_process.remove_silence", "line_number": 144, "usage_type": "call"}, {"api_name": "utils.audio_process", "line_number": 144, "usage_type": "name"}, {"api_name": "numpy.random.choice", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 150, "usage_type": "attribute"}, {"api_name": "librosa.load", "line_number": 158, "usage_type": "call"}, {"api_name": "utils.audio_process.quantization", "line_number": 160, "usage_type": "call"}, {"api_name": "utils.audio_process", "line_number": 160, "usage_type": "name"}, {"api_name": "utils.audio_process.audio_one_hot_encoding", "line_number": 164, "usage_type": "call"}, {"api_name": "utils.audio_process", "line_number": 164, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 182, "usage_type": "call"}, {"api_name": "tensorflow.data.Dataset.from_generator", "line_number": 185, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 185, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 187, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 188, "usage_type": "attribute"}, {"api_name": "tensorflow.string", "line_number": 189, "usage_type": "attribute"}, {"api_name": "tensorflow.TensorShape", "line_number": 190, "usage_type": "call"}, {"api_name": "tensorflow.TensorShape", "line_number": 191, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 206, "usage_type": "call"}]} +{"seq_id": "98294964", "text": "import re\nimport requests\nfrom bs4 import BeautifulSoup\nfrom fake_useragent import UserAgent\nimport sqlite3\nimport random\n\nimport socket \nsocket.setdefaulttimeout(15)\n\ndef crawl():\n\tconn = sqlite3.connect('proxy.db')\n\tcs = conn.cursor()\n\tcs.execute('''\n\t\tcreate table if not exists proxy (\n\t\t\thost varchar(20),\n\t\t\tport varchar(10),\n\t\t\tproto varchar (10),\n\t\t\tunique(host, port, proto)\n\t\t)\n\t''')\n\txici(conn, cs)\n\tcn(conn, cs)\n\ndef xici(conn, cs):\n\t\n\tua = UserAgent().random\n\theaders = {'user-agent': ua}\n\trsp = requests.get('http://www.xicidaili.com/', headers=headers)\n\tif rsp.status_code == 200:\n\t\tprint(\"成功爬取西刺 !\")\n\t\n\tsoup = BeautifulSoup(rsp.text, 'html.parser')\n\tfor i in soup.find_all('tr', class_=\"subtitle\"):\n\t\ti.decompose()\n\t\t\n\ttrs = soup.find_all('tr', class_=re.compile(r'\\s*'))\n\tfor tr in trs:\n\t\ttds = tr.find_all('td')\n\t\thost = tds[1].text\n\t\tport = tds[2].text\n\t\tproto = tds[5].text\n\t\tif proto == 'socks4/5':\n\t\t\tcontinue\n\t\tcs.execute('insert or ignore into proxy(host, port, proto) values (?, ?, ?)',\n\t\t\t(host, port, proto))\n\t\tconn.commit()\n\t\n\t\ndef cn(conn, cs):\n\tua = UserAgent().random\n\theaders = {\"user-agent\": ua}\n\ttry:\n\t\tr = requests.get(\"http://cn-proxy.com/\", headers=headers, timeout=15)\n\texcept Exception:\n\t\tprint('无法翻墙,爬取 cn-proxy 失败!')\n\t\tcs.close()\n\t\tconn.close()\n\t\treturn\n\n\tif r.status_code == 200:\n\t\tprint(\"成功爬取 cn-proxy !\")\n\n\tsoup = BeautifulSoup(r.text, \"html.parser\")\n\n\tfor tbody in soup.find_all(\"tbody\"):\n\t\tfor tr in tbody.find_all(\"tr\"):\n\t\t\ttds = tr.find_all(\"td\")\n\t\t\thost = tds[0].text\n\t\t\tport = tds[1].text\n\t\t\tcs.execute(\n\t\t\t\t\"insert or ignore into proxy (host, port, proto) values (?, ?, ?)\", (host, port, 'http')\n\t\t\t)\n\t\t\tconn.commit()\n\n\tcs.close()\n\tconn.close()\n\n\t\ndef get():\n\tconn = sqlite3.connect(\"proxy.db\")\n\tcs = conn.cursor()\n\tcs.execute(\"select * from proxy\")\n\trecord = random.choice(cs.fetchall())\n\n\tproxies = {\n\t\t\"http\": \"{}://{}:{}\".format(record[2], record[0], record[1]),\n\t\t\"https\": \"{}://{}:{}\".format(record[2], record[0], record[1]),\n\t}\n\n\ttry:\n\t\tr = requests.get(\"https://www.baidu.com/\", proxies=proxies, timeout=15)\n\texcept Exception:\n\t\tcs.execute(\"delete from proxy where host=?\", (record[0],))\n\t\tconn.commit()\n\t\treturn get()\n\t\t\n\tcs.close()\n\tconn.close()\n\treturn proxies\n\t\nif __name__ == '__main__':\n\tcrawl()", "sub_path": "proxy.py", "file_name": "proxy.py", "file_ext": "py", "file_size_in_byte": 2300, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "socket.setdefaulttimeout", "line_number": 9, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 12, "usage_type": "call"}, {"api_name": "fake_useragent.UserAgent", "line_number": 27, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 29, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 33, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 37, "usage_type": "call"}, {"api_name": "fake_useragent.UserAgent", "line_number": 51, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 54, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 64, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 81, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 84, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 92, "usage_type": "call"}]} +{"seq_id": "545117311", "text": "from keras.models import Model\nfrom keras.layers import Dense, Dropout, Flatten\nfrom keras.layers.convolutional import Conv2D, ZeroPadding2D\nfrom keras.layers.pooling import MaxPooling2D, AveragePooling2D\nfrom keras.optimizers import Adam\nfrom keras.models import Sequential\nimport numpy\nimport pandas\nimport sys\n\n# build model\nmodel = Sequential()\nmodel.add(Conv2D(filters = 64, kernel_size=(5,5), padding = 'valid', input_shape = (48, 48, 1), activation='relu')) \nmodel.add(ZeroPadding2D(padding = (2, 2), data_format = 'channels_last'))\nmodel.add(MaxPooling2D(pool_size = (5, 5), strides = (2, 2)))\nmodel.add(ZeroPadding2D(padding = (1, 1), data_format = 'channels_last'))\nmodel.add(Conv2D(64, (3, 3), activation = 'relu'))\nmodel.add(ZeroPadding2D(padding = (1, 1), data_format = 'channels_last'))\nmodel.add(Conv2D(64, (3, 3), activation = 'relu'))\nmodel.add(AveragePooling2D(pool_size = (3, 3), strides=(2, 2)))\nmodel.add(ZeroPadding2D(padding = (1, 1), data_format = 'channels_last'))\nmodel.add(Conv2D(128, (3, 3), activation = 'relu'))\nmodel.add(ZeroPadding2D(padding = (1, 1), data_format = 'channels_last'))\nmodel.add(Conv2D(128, (3, 3), activation = 'relu'))\nmodel.add(ZeroPadding2D(padding = (1, 1), data_format = 'channels_last'))\nmodel.add(AveragePooling2D(pool_size = (3, 3), strides = (2, 2)))\nmodel.add(Flatten())\nmodel.add(Dense(1024, activation = 'relu'))\nmodel.add(Dropout(0.5))\nmodel.add(Dense(1024, activation = 'relu'))\nmodel.add(Dropout(0.5))\nmodel.add(Dense(7, activation = 'softmax'))\nmodel.summary()\n\n# read in data and split data\nx_item = []\ntmp = pandas.read_csv(sys.argv[1], usecols=[1]).values.tolist()\nfor i in tmp:\n\tx_item.append(i[0].split(' '))\nx_item = numpy.array(x_item).reshape(-1,48,48,1)\ntmp = pandas.read_csv(sys.argv[1], usecols=[0])\ntmp = numpy.array(tmp)\ny_item = numpy.zeros((len(tmp), 7))\nfor i in range(0, len(tmp)):\n\ty_item[i][tmp[i][0]] = 1\n\n# set learning rate and crossentropy\nopt = Adam(lr = 0.00001)\nmodel.compile(loss='categorical_crossentropy', optimizer=opt, metrics=['accuracy'])\nmodel.fit(x = x_item, y = y_item, epochs = 150, batch_size = 128)\n\n# save the model\nmodel.save('./model.h5') ", "sub_path": "CNN/train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 2145, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "keras.models.Sequential", "line_number": 12, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 13, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.ZeroPadding2D", "line_number": 14, "usage_type": "call"}, {"api_name": "keras.layers.pooling.MaxPooling2D", "line_number": 15, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.ZeroPadding2D", "line_number": 16, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 17, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.ZeroPadding2D", "line_number": 18, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 19, "usage_type": "call"}, {"api_name": "keras.layers.pooling.AveragePooling2D", "line_number": 20, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.ZeroPadding2D", "line_number": 21, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 22, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.ZeroPadding2D", "line_number": 23, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 24, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.ZeroPadding2D", "line_number": 25, "usage_type": "call"}, {"api_name": "keras.layers.pooling.AveragePooling2D", "line_number": 26, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 27, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 28, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 29, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 30, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 31, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 32, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 37, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 37, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 40, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 41, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 43, "usage_type": "call"}, {"api_name": "keras.optimizers.Adam", "line_number": 48, "usage_type": "call"}]} +{"seq_id": "382694606", "text": "import xarray as xr\nimport pandas as pd\nimport numpy as np\nimport xskillscore as xs\n\nimport json\nimport cartopy.crs as ccrs\nimport matplotlib.pyplot as plt\n\nfrom S2S.local_configuration import config\nfrom S2S.graphics import latex, graphics\n\nfrom matplotlib.colors import BoundaryNorm\n\nfrom S2S.data_handler import ERA5, BarentsWatch\nfrom S2S.process import Hindcast, Observations, Grid2Point\nfrom S2S import models\nimport S2S.scoring as sc\n\ndef plus_minus_15_days(t_start,t_end):\n if t_start[1]==1:\n t_start = (t_start[0]-1,12,15)\n else:\n t_start = (t_start[0],t_start[1]-1,15)\n\n if t_end[1]==12:\n t_end = (t_end[0]+1,1,15)\n else:\n t_end = (t_end[0],t_end[1]+1,15)\n\n return t_start,t_end\n\n# bounds = (0,28,55,75)\nbounds = (0,28,55,75)\nvar = 'sst'\n\nclim_t_start = (2000,1,1)\nclim_t_end = (2021,1,4)\n\nsteps = pd.to_timedelta([9,16,23,30,37],'D')\n\nhigh_res = True\n\nmparser = {\n '1':'JAN','2':'FEB','3':'MAR',\n '4':'APR','5':'MAY','6':'JUN',\n '7':'JUL','8':'AUG','9':'SEP',\n '10':'OCT','11':'NOV','12':'DEC'\n }\nmonths = ['01','02','03','04','05','06','07','08','09','10','11','12']\n\nmae_fc, mae_clim, mae_pers = [], [], []\n################################################################################\nbw = BarentsWatch().load('all',no=0).sortby('time')[var]\n\nt_start = (2020,12,1) #can start with 8\nt_end = (2021,1,1)\nmodel = 'CY47R1'\n\nhh = Hindcast(\n var,\n (2020,1,23),\n (2020,1,24),\n bounds,\n high_res=high_res,\n steps=steps,\n process=False,\n download=False,\n split_work=False,\n )\nadd_month = hh.add_month\nsmaller_than = hh.smaller_than\ndel hh\n\n# assign new end time, one month after start time\nt_end = add_month(t_start)\n\n# until end time is reached; load one month at the time\nwhile smaller_than(t_end,(2021,4,1)):\n\n month = str(t_start[1])\n print(month)\n\n nk = []\n for loc in bw.location.values:\n\n fname = config['NORKYST'] +\\\n 'NorKyst800_' +\\\n str(loc) +\\\n '.nc'\n\n nk.append(xr.open_dataset(fname)[var].drop('radius'))\n nk = xr.concat(nk,'location')\n\n ts,te = plus_minus_15_days(t_start,t_end)\n hindcast = Hindcast(\n var,\n ts,\n te,\n bounds,\n high_res = high_res,\n steps = steps,\n process = True,\n download = False,\n split_work = False,\n period = [nk.time.min(),nk.time.max()]\n )\n print([nk.time.min(),nk.time.max()])\n # update times to the next month\n t_start = t_end\n t_end = add_month(t_end)\n\n observations = Observations(\n name='NorKyst-800',\n observations=nk,\n forecast=hindcast,\n process=True\n )\n del nk\n\n hindcast = Grid2Point(observations,hindcast)\\\n .correlation(step_dependent=True)\n\n mae_fc = xs.mae(\n hindcast.data_a.mean('member'),\n observations.data_a,\n dim=[]\n )\n\n mae_fc = mae_fc.where(mae_fc.time.dt.month==int(month),drop=True)\\\n .mean('time',skipna=True)\n\n crps_fc = sc.crps_ensemble(observations.data_a,hindcast.data_a)\n\n crps_fc = crps_fc.where(crps_fc.time.dt.month==int(month),drop=True)\\\n .mean('time',skipna=True)\n\n crps_clim = xs.crps_gaussian(\n observations.data_a,\n xr.zeros_like(observations.data_a),\n xr.ones_like(observations.data_a),\n dim=[]\n )\n\n crps_clim = crps_clim.where(crps_clim.time.dt.month==int(month),drop=True)\\\n .mean('time',skipna=True)\n\n del hindcast\n\n pers = models.persistence(\n init_value = observations.init_a,\n observations = observations.data_a\n )\n\n mae_pers = xs.mae(\n pers,\n observations.data_a,\n dim=[]\n )\n\n mae_pers = mae_pers.where(mae_pers.time.dt.month==int(month),drop=True)\\\n .mean('time',skipna=True)\n\n del pers\n\n mae_clim = xs.mae(\n xr.zeros_like(observations.data_a),\n observations.data_a,\n dim=[]\n )\n\n mae_clim = mae_clim.where(mae_clim.time.dt.month==int(month),drop=True)\\\n .mean('time',skipna=True)\n\n del observations\n\n for step in steps:\n\n for ref_fc,lab in zip([mae_clim,mae_pers],['CLIM','PERS']):\n\n latex.set_style(style='white')\n fig,ax = plt.subplots(1,1,\\\n figsize=latex.set_size(width=345,subplots=(1,1),fraction=0.95),\\\n subplot_kw=dict(projection=ccrs.NorthPolarStereo()))\n\n ss = ( 1 - mae_fc/ref_fc ).sel(step=step)\n\n cmap = latex.skill_cmap().reversed()\n levels = np.arange(-1.,1.05,0.05)\n norm = BoundaryNorm(levels,cmap.N)\n\n cs = ax.scatter(\n ss.lon.values,\n ss.lat.values,\n c=ss.values,\n s=1.1,\n cmap=cmap,\n norm=norm,\n alpha=0.95,\n transform=ccrs.PlateCarree()\n )\n ax.coastlines(resolution='10m', color='grey',\\\n linewidth=0.2)\n ax.set_title(mparser[month] + ' MAEss EC vs. '+lab+', NorKyst, lt:'\\\n +str(step.days))\n fig.colorbar(cs,ax=ax)\n graphics.save_fig(fig,\n model+'mae_skill_map_NorKyst_month'+month+lab+str(step.days)\n )\n\n latex.set_style(style='white')\n fig,ax = plt.subplots(1,1,\\\n figsize=latex.set_size(width=345,subplots=(1,1),fraction=0.95),\\\n subplot_kw=dict(projection=ccrs.NorthPolarStereo()))\n\n ss = ( 1 - crps_fc/crps_clim ).sel(step=step)\n\n cmap = latex.skill_cmap().reversed()\n levels = np.arange(-1.,1.05,0.05)\n norm = BoundaryNorm(levels,cmap.N)\n\n cs = ax.scatter(\n ss.lon.values,\n ss.lat.values,\n c=ss.values,\n s=1.1,\n cmap=cmap,\n norm=norm,\n alpha=0.95,\n transform=ccrs.PlateCarree()\n )\n ax.coastlines(resolution='10m', color='grey',\\\n linewidth=0.2)\n ax.set_title(mparser[month] + ' CRPss EC vs. CLIM, NorKyst, lt:'\\\n +str(step.days))\n fig.colorbar(cs,ax=ax)\n graphics.save_fig(fig,\n model+'crps_skill_map_NorKyst_month'+month+'CLIM'+str(step.days)\n )\n", "sub_path": "scripts/Henrik/skill_on_norkyst_CY47R1.py", "file_name": "skill_on_norkyst_CY47R1.py", "file_ext": "py", "file_size_in_byte": 7655, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pandas.to_timedelta", "line_number": 40, "usage_type": "call"}, {"api_name": "S2S.data_handler.BarentsWatch", "line_number": 54, "usage_type": "call"}, {"api_name": "S2S.process.Hindcast", "line_number": 60, "usage_type": "call"}, {"api_name": "S2S.local_configuration.config", "line_number": 87, "usage_type": "name"}, {"api_name": "xarray.open_dataset", "line_number": 92, "usage_type": "call"}, {"api_name": "xarray.concat", "line_number": 93, "usage_type": "call"}, {"api_name": "S2S.process.Hindcast", "line_number": 96, "usage_type": "call"}, {"api_name": "S2S.process.Observations", "line_number": 113, "usage_type": "call"}, {"api_name": "S2S.process.Grid2Point", "line_number": 121, "usage_type": "call"}, {"api_name": "xskillscore.mae", "line_number": 124, "usage_type": "call"}, {"api_name": "S2S.scoring.crps_ensemble", "line_number": 133, "usage_type": "call"}, {"api_name": "S2S.scoring", "line_number": 133, "usage_type": "name"}, {"api_name": "xskillscore.crps_gaussian", "line_number": 138, "usage_type": "call"}, {"api_name": "xarray.zeros_like", "line_number": 140, "usage_type": "call"}, {"api_name": "xarray.ones_like", "line_number": 141, "usage_type": "call"}, {"api_name": "S2S.models.persistence", "line_number": 150, "usage_type": "call"}, {"api_name": "S2S.models", "line_number": 150, "usage_type": "name"}, {"api_name": "xskillscore.mae", "line_number": 155, "usage_type": "call"}, {"api_name": "xskillscore.mae", "line_number": 166, "usage_type": "call"}, {"api_name": "xarray.zeros_like", "line_number": 167, "usage_type": "call"}, {"api_name": "S2S.graphics.latex.set_style", "line_number": 181, "usage_type": "call"}, {"api_name": "S2S.graphics.latex", "line_number": 181, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 182, "usage_type": "name"}, {"api_name": "S2S.graphics.latex.set_size", "line_number": 183, "usage_type": "call"}, {"api_name": "S2S.graphics.latex", "line_number": 183, "usage_type": "name"}, {"api_name": "cartopy.crs.NorthPolarStereo", "line_number": 184, "usage_type": "call"}, {"api_name": "cartopy.crs", "line_number": 184, "usage_type": "name"}, {"api_name": "S2S.graphics.latex.skill_cmap", "line_number": 188, "usage_type": "call"}, {"api_name": "S2S.graphics.latex", "line_number": 188, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.colors.BoundaryNorm", "line_number": 190, "usage_type": "call"}, {"api_name": "cartopy.crs.PlateCarree", "line_number": 200, "usage_type": "call"}, {"api_name": "cartopy.crs", "line_number": 200, "usage_type": "name"}, {"api_name": "S2S.graphics.graphics.save_fig", "line_number": 207, "usage_type": "call"}, {"api_name": "S2S.graphics.graphics", "line_number": 207, "usage_type": "name"}, {"api_name": "S2S.graphics.latex.set_style", "line_number": 211, "usage_type": "call"}, {"api_name": "S2S.graphics.latex", "line_number": 211, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 212, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 212, "usage_type": "name"}, {"api_name": "S2S.graphics.latex.set_size", "line_number": 213, "usage_type": "call"}, {"api_name": "S2S.graphics.latex", "line_number": 213, "usage_type": "name"}, {"api_name": "cartopy.crs.NorthPolarStereo", "line_number": 214, "usage_type": "call"}, {"api_name": "cartopy.crs", "line_number": 214, "usage_type": "name"}, {"api_name": "S2S.graphics.latex.skill_cmap", "line_number": 218, "usage_type": "call"}, {"api_name": "S2S.graphics.latex", "line_number": 218, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 219, "usage_type": "call"}, {"api_name": "matplotlib.colors.BoundaryNorm", "line_number": 220, "usage_type": "call"}, {"api_name": "cartopy.crs.PlateCarree", "line_number": 230, "usage_type": "call"}, {"api_name": "cartopy.crs", "line_number": 230, "usage_type": "name"}, {"api_name": "S2S.graphics.graphics.save_fig", "line_number": 237, "usage_type": "call"}, {"api_name": "S2S.graphics.graphics", "line_number": 237, "usage_type": "name"}]} +{"seq_id": "309319049", "text": "import sys\nfrom setuptools import setup, find_packages\n\ninstall_requires = ['requests']\nif sys.version_info < (3, 4):\n install_requires.append('enum34')\n\nsetup(\n name = 'netbluemind4',\n packages = find_packages(),\n version = '4.1.48893',\n description = 'Automatically generated client for BlueMind >= 4 REST API. Check netbluemind for older releases',\n author = 'BlueMind team',\n author_email = 'contact@bluemind.net',\n url = 'http://git.blue-mind.net/bluemind/',\n keywords = ['bluemind', 'rest', 'api', 'mail', 'groupware'],\n classifiers = [\n 'Development Status :: 5 - Production/Stable',\n 'Intended Audience :: Developers',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n ],\n install_requires=install_requires\n)\n", "sub_path": "pypi_install_script/netbluemind4-4.1.48893.tar/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 814, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "sys.version_info", "line_number": 5, "usage_type": "attribute"}, {"api_name": "setuptools.setup", "line_number": 8, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 10, "usage_type": "call"}]} +{"seq_id": "456025787", "text": "from typing import List\n\nclass Solution:\n def minPathSum(self, grid: List[List[int]]) -> int:\n m = len(grid)\n n = len(grid[0])\n dp = [[0 for _ in range(n)] for _ in range(m)]\n # Handle corner cases\n dp[0][0] = grid[0][0]\n if m == 1 and n > 1:\n for i in range(1, n):\n dp[0][i] = dp[0][i-1] + grid[0][i]\n \n elif n == 1 and m > 1:\n for i in range(1, m):\n dp[i][0] = dp[i-1][0] + grid[i][0]\n\n elif n > 1 and m > 1:\n # Handle boundary cases\n for i in range(1, n):\n dp[0][i] = dp[0][i-1] + grid[0][i]\n for i in range(1, m):\n dp[i][0] = dp[i-1][0] + grid[i][0]\n # DP\n for i in range(1, m):\n for j in range(1,n):\n dp[i][j] = min(dp[i-1][j], dp[i][j-1]) + grid[i][j]\n return dp[-1][-1]\n\nif __name__==\"__main__\":\n solution = Solution()\n grid = [[1,2,3],[4,5,6]]\n print(solution.minPathSum(grid=grid))", "sub_path": "leetcode/64_minimum_path_sum.py", "file_name": "64_minimum_path_sum.py", "file_ext": "py", "file_size_in_byte": 1047, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "typing.List", "line_number": 4, "usage_type": "name"}]} +{"seq_id": "461938549", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Nov 13 11:12:31 2019\n\n@author: Ankit\n\"\"\"\n\nimport json\nimport requests\nimport urllib.parse\nfrom pandas.io.json import json_normalize \nimport pandas as pd\nimport numpy as np\nimport copy\nfrom datetime import datetime\n\n#from flask import Flask , jsonify\n\n\n\nfrom aes256 import aes256\n\n#muffview\n\n \nclass Targetvalue:\n \n def ret(self , value,level,ulimit,llimit):\n \n value = int (value)\n level = int (level)\n ulimit = float(ulimit)\n llimit = float(llimit)\n #band = int(band)\n \n\n #url = \"https://er15.xyz:4436/api/Customers/GetCustDetailLabel?month=10&year=2019\"\n url = \"https://er15.xyz:4436/api/Customers/CRMLevelCustomerDetail?month=12&year=2019\"\n \n resp = requests.get(url)\n json_data = resp.json()\n \n if(json_data['Status'] =='OK'):\n redisAesKey = datetime.today().strftime('%Y%m%d') + \"1201\"\n jso = aes256().decrypt(json_data['Data'],redisAesKey)\n js = json.loads(jso)\n df = json_normalize(js)\n # df = df.loc[df['IsActive'] == True]\n \n # #json_data = requests.get(url).json()\n # json_data = resp.json()\n # df=json_normalize(json_data)\n \n \n #ACTIVE = df[(df.IsActive == True)]\n #INACTIVE = df[(df.IsActive == False)]\n #df=ACTIVE\n #print(df.head(2))\n df['SelfOrderPercentage']=((df['Selfordercount'])/(df['Selfordercount']+df['Salespersonordercount'])*100)\n df['SalesOrderPercentage']=(df['Salespersonordercount'])/(df['Selfordercount']+df['Salespersonordercount'])*100\n df.fillna(0)\n df = df.replace(np.NaN,0)\n df.rename(columns = {\"kkVolumn\":\"KKvolume\"},inplace = True)\n \n ###### LEVEL DEFINITIONS BEING DEFINED\n df.loc[df.Volume == 0, 'levels'] = 'Level 0'\n df.loc[df.Volume >= 1, 'levels'] = 'Level 1'\n df.loc[(df.Volume >= 10000) & (df.OrderCount >= 3) & (df.BrandCount >= 5), 'levels'] = 'Level 2'\n df.loc[(df.Volume >= 20000) & (df.OrderCount >= 5) & (df.BrandCount >= 10) & (df.KKvolume >= 2000), 'levels'] = 'Level 3'\n df.loc[(df.Volume >= 30000) & (df.OrderCount >= 8) & (df.BrandCount >= 20) & (df.KKvolume >= 8000) & ((df.Selfordercount/(df.Salespersonordercount+df.Selfordercount))*100 > 30), 'levels'] = 'Level 4'\n df.loc[(df.Volume >= 75000) & (df.OrderCount >= 12) & (df.BrandCount >= 40) & (df.KKvolume >= 15000) & ((df.Selfordercount/(df.Salespersonordercount+df.Selfordercount))*100 > 60), 'levels'] = 'Level 5'\n df.round(2)\n #\n l0=df[df['levels'] == 'Level 0']\n l1=df[df['levels'] == 'Level 1']\n l2=df[df['levels'] == 'Level 2']\n l3=df[df['levels'] == 'Level 3']\n l4=df[df['levels'] == 'Level 4']\n l5=df[df['levels'] == 'Level 5']\n \n \n \n l=[l0,l1,l2,l3,l4,l5]\n \n \n \n \n i=level\n \n if ( value == 0):\n l[i]['Target'] = 0\n \n # elif (value != 0 & percentage == 0):\n # l[i]['Target'] = [(int(j) + value) for j in l[i]['Volume']]\n \n else:\n \n l[i]['Target'] = [(int(j) + value) for j in l[i]['Volume']]\n \n \n l[i].sort_values(by='Volume', ascending=False)\n df1 = l[i][['SkCode','Cityid','WarehouseName','WarehouseId','levels','Volume','Target']]\n df1 = df1[(df1['Volume'] < ulimit) & (df1['Volume'] >= llimit) ]\n df1 = df1.sort_values(by='Volume', ascending=False)\n #df1['Band'] = band \n #df1= df1.to_json(orient='records')\n #df_list = df1.values.tolist()\n df1 = df1.to_dict('records')\n return df1\n \n \n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "Targetvalue.py", "file_name": "Targetvalue.py", "file_ext": "py", "file_size_in_byte": 3861, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "requests.get", "line_number": 40, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 44, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 44, "usage_type": "name"}, {"api_name": "aes256.aes256", "line_number": 45, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 46, "usage_type": "call"}, {"api_name": "pandas.io.json.json_normalize", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.NaN", "line_number": 62, "usage_type": "attribute"}]} +{"seq_id": "362913729", "text": "from tabulate import tabulate\n\njugadores = []\npuntajeParcial = [['1'],['2'],['3'],['4'],['5'],['6'],['E'],['F'],['P'],['G'],['GD'],['']]\n\ndef ingresoCantidadJug ():#Esta función es para obtener y retornar la cantidad de jugadores ingresada por el usuario.\n cantidad = int(input(\"Por favor, ingrese la cantidad de jugadores: \"))\n return cantidad\n\ndef ingresoNombres (cantidad):#Esta función sirva para agregar los nombres de los jugadores a la lista Vacía de jugadores\n for cantidad in range(0, cantidad):\n jugadores.append(input(\"Por favor ingrese nombre del jugador: \"))\n\ndef insertarColumnas (cantidad):#Sirve para agregar X cantidad de \"espacios\" agregando un 0 por cada jugador en cada \"Vagón\" de la lista puntajeParcial\n for t in range (0,cantidad):#Por ejemplo si ponemos 3 en cantidad se va a agregar 1 cero por iteración en los 12 vagones (de 1-6, E,F,P,G,GD y el espacio del resultado final)\n for posiciones in range (0,12):#Estos son los doce \"vagones\"\n puntajeParcial[posiciones].append(0)#acá agrega el valor 0 en el vagón X(cambia según la iteración)\n\ndef sumaPuntajeFinal(cantidad,puntajeParcial):# Se debe agregar esta función solo al final, porque si los valores están vacíos no se pueden sumar\n for nume in range(1, cantidad+1):# Debe estar completo el ingreso de los 11 valores por jugador, lo que no tenga completo se debe agregar 0\n resultado = ((puntajeParcial[0][nume])+(puntajeParcial[1][nume])+(puntajeParcial[2][nume])+\n (puntajeParcial[3][nume])+(puntajeParcial[4][nume])+(puntajeParcial[5][nume])+\n (puntajeParcial[6][nume])+(puntajeParcial[7][nume])+(puntajeParcial[8][nume])+\n (puntajeParcial[9][nume])+(puntajeParcial[10][nume]))\n puntajeParcial[11][nume] = (resultado)\n cadenaPuntaje =(\" PUNTAJE FINAL \")\n print(\"\\n\"+(cadenaPuntaje.center(50,\"=\")+\"\\n\"))\n mostrarPuntajeParcial()\n\ndef anotacion (nroJugador,categoria,puntos):#Esto es para ingresar los puntos, hay que cambiar las opciones a un diccionario para validar si hay un error de tipeo.\n #global categoria\n #global puntos\n #categoria = input(\"Ingrese categoria a anotar (Ingresando 1, 2, 3, 4, 5, 6, E, F, P, G, GD): \")\n #puntos = int(input(\"Ingrese cantidad de puntos obtenidos: \"))\n if categoria == 1:\n puntajeParcial[0][nroJugador] = puntos\n elif categoria == 2:\n puntajeParcial[1][nroJugador] = puntos\n elif categoria == 3:\n puntajeParcial[2][nroJugador] = puntos\n elif categoria == 4:\n puntajeParcial[3][nroJugador] = puntos\n elif categoria == 5:\n puntajeParcial[4][nroJugador] = puntos\n elif categoria == 6:\n puntajeParcial[5][nroJugador] = puntos\n elif categoria == 'E':\n puntajeParcial[6][nroJugador] = puntos\n elif categoria == 'F':\n puntajeParcial[7][nroJugador] = puntos\n elif categoria == 'P':\n puntajeParcial[8][nroJugador] = puntos\n elif categoria == 'G':\n puntajeParcial[9][nroJugador] = puntos\n elif categoria == 'GD':\n puntajeParcial[10][nroJugador] = puntos\n\ndef mostrarPuntajeParcial():#Esto muestra el tablero con los resultados parciales\n print(tabulate(puntajeParcial, jugadores))\n\ndef mostrarGanador (puntajeParcial,cantidad,jugadores):#Muestra que jugador ganó y con cuantos púntos\n listaPuntos = (puntajeParcial[11])\n soloPuntos = (puntajeParcial[11][1:])\n puntajeOrdenado = (sorted(soloPuntos,reverse=True))\n cadenaResultados = (\" RESULTADOS FINALES \")\n print(\"\\n\" + (cadenaResultados.center(50, \"*\") + \"\\n\"))\n for i in range(0,len(puntajeOrdenado)):\n ganador = (puntajeOrdenado[i])\n if ganador in listaPuntos:\n ordenGanadores = listaPuntos.index(ganador)\n print(\"En \", i + 1, \"puesto: \", (str(jugadores[ordenGanadores - 1])), \"con \", ganador, \"puntos\")\n", "sub_path": "tablero.py", "file_name": "tablero.py", "file_ext": "py", "file_size_in_byte": 3884, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "tabulate.tabulate", "line_number": 59, "usage_type": "call"}]} +{"seq_id": "36259258", "text": "#!/usr/bin/python3\n# -*- coding: utf-8 -*-\n\"\"\"\n@File Name : PicPy\n@Author : LeeCQ\n@Date-Time : 2021/7/11 23:30\n\"\"\"\nfrom pathlib import Path\nfrom argparse import ArgumentParser\n\nfrom requests import post\n\n\ndef img_upload(file_path):\n _f = Path(file_path)\n files = {\"file\": (_f.name, _f.open('rb'), \"image/jpeg\")}\n return post('http://img.p.leecq.cn:8080/upload', files=files)\n\n\ndef args_from_cli():\n \"\"\"解析命令行参数\"\"\"\n p_args = ArgumentParser(description='Python图片上传接口')\n p_args.add_argument('file_path', help='文件的绝对路径')\n return p_args.parse_args()\n\n\ndef main():\n \"\"\"MAN\"\"\"\n _args = args_from_cli()\n _http = img_upload(_args.file_path)\n if _http.status_code == 200:\n return _http.text\n else:\n raise Exception(f'{_http.status_code}: {_http.text}')\n\n\nif __name__ == '__main__':\n print(main())\n", "sub_path": "client/PicPy.py", "file_name": "PicPy.py", "file_ext": "py", "file_size_in_byte": 886, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pathlib.Path", "line_number": 15, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 17, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 22, "usage_type": "call"}]} +{"seq_id": "434956878", "text": "__author__ = 'thor'\n\nfrom numpy import *\nimport numpy as np\nimport pandas as pd\nfrom collections import Counter, OrderedDict, defaultdict\nimport string\nimport random as rnd\nimport itertools\nfrom ut.util.uiter import all_subsets_of\n\nfrom ut.stats.bin_est.set_est import Shapley as Shapley_1\n\n# from ut.daf.manip import rollin_col\n\n\ndef compute_shapley_values_from_coalition_values(coalition_values, normalize=False):\n coalition_values = pd.DataFrame(index=coalition_values.keys(),\n data=coalition_values.values(),\n columns=['value'])\n se = Shapley_1(coalition_values, success='value')\n se.change_type_of_d_index(tuple)\n shapley_values = se.compute_shapley_values()\n if normalize:\n return _normalize_dict_values(shapley_values)\n else:\n return shapley_values\n\n\ndef _normalize_dict_values(d):\n value_sum = float(np.sum(d.values()))\n return {k: v / value_sum for k, v in d.items()}\n\n\ndef all_subsets_iterator(superset):\n return itertools.chain(\n *itertools.imap(lambda subset_size: itertools.combinations(superset, subset_size),\n range(1, len(superset))))\n\n\nclass ShapleyDataModel(object):\n def __init__(self, data=None, data_type=None):\n \"\"\"\n Inputs:\n * item_seperator will be used to construct string hashes from lists.\n You should choose a character that never shows up in the items, or you'll get problems.\n Other attributes:\n * coalition_obs is a Counter of coalitions\n * coalition_values is also a Counter of coalitions, but it counts not only\n the coalition_obs, but all non-empty subsets of the latter.\n \"\"\"\n self.coalition_obs = Counter()\n self.item_list = []\n self._coalition_size_map = None\n if data is not None:\n # if data_type not given, determine\n if data_type is None:\n if isinstance(data, Counter):\n data_type = 'coalition_obs'\n else:\n data_type = 'item_collections'\n\n # according to type, process and set data\n if data_type == 'coalition_obs':\n self.coalition_obs = data\n elif data_type == 'coalition_obs_collection':\n for d in data:\n self.absorb_coalition_obs(d)\n elif data_type == 'item_collections':\n for d in data:\n self.absorb_coalition(d)\n\n @staticmethod\n def coalition_of(iter_of_items):\n return tuple(unique(iter_of_items))\n\n def absorb_coalition(self, collection_of_items_of_single_coalition):\n \"\"\"\n Updates the self.coalition_obs with the input coalition (a list of items)\n \"\"\"\n self.coalition_obs.update([self.coalition_of(collection_of_items_of_single_coalition)])\n\n def absorb_coalition_obs(self, coalition_obs_dict):\n \"\"\"\n Updates the self.coalition_obs with the input dict of coalition: obs_value\n \"\"\"\n self.absorb_coalition_and_value(coalition_obs_dict.keys()[0], coalition_obs_dict.values()[0])\n\n def absorb_coalition_and_value(self, coalition, value):\n \"\"\"\n Updates the self.coalition_obs with the input dict of coalition: obs_value\n \"\"\"\n self.coalition_obs.update({self.coalition_of(coalition): value})\n\n def coalition_values(self, verbose=False):\n \"\"\"\n Computes the self.coalition_values attribute.\n To do this, we accumulate the counts of all subsets of each unique coalition.\n \"\"\"\n coalition_contributions = Counter(self.coalition_obs)\n\n if verbose:\n print(self.coalition_values)\n\n for coalition, count in self.coalition_obs.iteritems(): # for every coalition\n # ... get list corresponding to the key\n coalition = self._key_to_list(coalition)\n # ... get all non-empty strict subsets of this list, and assign the mother coalition count\n subset_counts = \\\n {self._list_to_key(sub_coalition): count\n for sub_coalition in all_subsets_iterator(coalition)}\n # ... update the coalition_values counter with these counts\n coalition_contributions.update(subset_counts)\n\n if verbose:\n print(\" after {} contributions:\\n {}\" \\\n .format(coalition, self.coalition_values))\n\n return coalition_contributions\n\n def coalition_size_map(self):\n if not self._coalition_size_map:\n self._coalition_size_map = defaultdict(dict)\n for coalition, count in self.coalition_obs.iteritems():\n self._coalition_size_map[len(coalition)].update({coalition: count})\n self._coalition_size_map = OrderedDict(sorted(self._coalition_size_map.items(), key=lambda t: t[0]))\n return self._coalition_size_map\n\n def mk_poset(self):\n d = defaultdict(list)\n _coalition_size_map = self.coalition_size_map()\n coalition_sizes = sorted(_coalition_size_map.keys())\n # TODO: Finish, if necessary\n\n def mk_item_list(self):\n self.item_list = unique(concatenate(self.coalition_obs.keys()))\n\n\ndef _test_shapley_data_model():\n list_of_coalitions = [['A', 'B', 'C'], ['A', 'C', 'B'], ['B', 'A', 'C'], ['A', 'A', 'B', 'C'],\n ['C', 'A'], ['B', 'C'], ['C', 'B'], ['C', 'B'], ['A']]\n dm = ShapleyDataModel() # initialize the data model\n\n for coalition in list_of_coalitions: # count the coalitions\n dm.absorb_coalition(coalition)\n assert dm.coalition_obs == Counter({('A', 'B', 'C'): 4, ('B', 'C'): 3, ('A',): 1, ('A', 'C'): 1}), \\\n \"Unexpected result for dm.coalition_obs\"\n\n\n print(\"All good in _test_shapley_data_model\")\n\n\ndef rand_shapley_values(items=3):\n if isinstance(items, int):\n items = ','.join(string.ascii_uppercase[:items]).split(',')\n if isinstance(items, list):\n items = {items[i]: 2**i for i in range(len(items))}\n return items\n\n\nclass LinearValuedCoalitionGenerator(object):\n def __init__(self, shapley_values=3, normalize=False):\n shapley_values = shapley_values or 3\n if not isinstance(shapley_values, dict):\n shapley_values = rand_shapley_values(items=shapley_values)\n self.shapley_values = shapley_values\n if normalize:\n self.shapley_values = _normalize_dict_values(self.shapley_values)\n\n @staticmethod\n def coalition_of(coalition):\n return tuple(sort(coalition))\n\n def coalition_value(self, coalition):\n return sum([self.shapley_values[item] for item in coalition])\n\n def rand_coalition(self):\n return self.coalition_of(rnd.sample(self.shapley_values.keys(), rnd.randint(1, len(self.shapley_values))))\n\n def rand_coalition_obs(self):\n coalition = self.rand_coalition()\n return {coalition: self.coalition_value(coalition)}\n\n def rand_coalition_obs_cum(self, n_draws=None):\n n_draws = n_draws or len(self.shapley_values) / 2\n coalition_obs = Counter()\n for x in itertools.starmap(self.rand_coalition_obs, itertools.repeat([], n_draws)):\n coalition_obs.update(x)\n return coalition_obs\n\n def coalition_values(self):\n return {self.coalition_of(coalition): self.coalition_value(coalition)\n for coalition in all_subsets_of(self.shapley_values.keys(), include_empty_set=False)}\n\n\n\n\n# class ShapleyDataModel_old(object):\n# def __init__(self, item_seperator=','):\n# \"\"\"\n# Inputs:\n# * item_seperator will be used to construct string hashes from lists.\n# You should choose a character that never shows up in the items, or you'll get problems.\n# Other attributes:\n# * coalition_obs is a Counter of coalitions\n# * coalition_values is also a Counter of coalitions, but it counts not only\n# the coalition_obs, but all non-empty subsets of the latter.\n# \"\"\"\n# self.coalition_obs = Counter()\n# self.coalition_values = None\n# self.item_seperator = item_seperator\n# self.contribution_df = None\n# self.item_list = []\n#\n# def absorb_coalition(self, coalition):\n# \"\"\"\n# Updates the self.coalition_obs with the input coalition (a list of items)\n# \"\"\"\n# self.coalition_obs.update([self._list_to_key(coalition)])\n#\n# def mk_coalition_size_map(self):\n#\n# d = defaultdict(list)\n# for coalition, count in self.coalition_obs.iteritems():\n# d[len(self._key_to_list(coalition))].append({coalition: count})\n# return d\n#\n# def mk_coalition_contributions(self, verbose=False):\n# \"\"\"\n# Computes the self.coalition_values attribute.\n# To do this, we accumulate the counts of all subsets of each unique coalition.\n# \"\"\"\n# # init with coalition_obs\n# self.coalition_values = Counter(self.coalition_obs)\n# if verbose:\n# print(self.coalition_values)\n# for coalition, count in self.coalition_obs.iteritems(): # for every coalition\n# # get list corresponding to the key\n# coalition = self._key_to_list(coalition)\n# # get all non-empty strict subsets of this list,\n# # and assign the mother coalition count\n# subset_counts = \\\n# {self._list_to_key(sub_coalition): count\n# for sub_coalition in all_subsets_iterator(coalition)}\n# # update the coalition_values counter with these counts\n# self.coalition_values.update(subset_counts)\n# if verbose:\n# print(\" after {} contributions:\\n {}\" \\\n# .format(coalition, self.coalition_values))\n#\n# def mk_item_list(self):\n# self.item_list = list(unique(self.item_seperator.join(dm.coalition_obs.keys()) \\\n# .split(self.item_seperator)))\n#\n# # def all_supersets_iterator(self, subset):\n#\n# # subset = dm\n#\n# def mk_contribution_df(self):\n# self._fill_counters()\n# self.contribution_df = \\\n# pd.DataFrame(index=self.coalition_values.keys(), columns=dm.item_list)\n# for coalition in self.contribution_df.index.values:\n# print self._remove_and_remain_dicts(coalition)\n# for rr in self._remove_and_remain_dicts(coalition):\n# # the contribution of each item is the total contribution\n# # minus what the contribution would be without this item\n# contribution = \\\n# self.coalition_values[coalition] \\\n# - self.coalition_values[rr['remaining']]\n# # enter this in the contribution_df\n# self.contribution_df.loc[coalition, rr['removed']] = contribution\n#\n# def _fill_counters(self):\n# \"\"\"\n# adds missing item combinations to counters, giving them 0 count\n# \"\"\"\n# self.mk_item_list()\n# zero_counts = {k: 0 for k in itertools.imap(self._list_to_key,\n# all_subsets_iterator(self.item_list))\n# }\n# self.coalition_obs.update(zero_counts)\n# self.coalition_values.update(zero_counts)\n#\n# def _list_to_key(self, coalition):\n# \"\"\"\n# Transforms a list of strings to a comma (or item_seperator) separated string\n# of unique items of the input list.\n# \"\"\"\n# return self.item_seperator.join(unique(coalition))\n#\n# def _key_to_list(self, coalition_key):\n# \"\"\"\n# Inverse of _list_to_key:\n# Returns a list from a character (item_seperator) seperated string of items.\n# \"\"\"\n# return coalition_key.split(self.item_seperator)\n#\n# def _remove_and_remain_dicts(self, superset):\n# \"\"\"\n# Returns a list of {removed, remaining} dicts listing all (keys of) superset - item\n# sets for every item in superset.\n# Returns an empty list if the input superset has only one element.\n# Example:\n# self._remove_and_remain_dicts('A,B,C')\n# returns\n# [{'remaining': 'B,C', 'removed': 'A'},\n# {'remaining': 'A,B', 'removed': 'C'},\n# {'remaining': 'A,C', 'removed': 'B'}]\n# \"\"\"\n# superset = set(self._key_to_list(superset))\n# if len(superset) > 1:\n# return [{'removed': x,\n# 'remaining': self._list_to_key(\n# list(superset.difference(x)))}\n# for x in superset]\n# else:\n# return list() # return empty list if superset has only one element\n#\n#\n# def _test_shapley_data_model():\n# list_of_coalitions = [['A', 'B', 'C'], ['A', 'C', 'B'], ['B', 'A', 'C'], ['A', 'A', 'B', 'C'],\n# ['C', 'A'], ['B', 'C'], ['C', 'B'], ['C', 'B'], ['A']]\n# dm = ShapleyDataModel_old() # initialize the data model\n#\n# for coalition in list_of_coalitions: # count the coalitions\n# dm.absorb_coalition(coalition)\n# assert dm.coalition_obs == Counter({'A,B,C': 4, 'B,C': 3, 'A': 1, 'A,C': 1}), \\\n# \"Unexpected result for dm.coalition_obs\"\n#\n# dm.mk_coalition_contributions()\n# assert dm.coalition_values \\\n# == Counter({'C': 8, 'B': 7, 'B,C': 7, 'A': 6, 'A,C': 5, 'A,B,C': 4, 'A,B': 4}), \\\n# \"Unexpected result for dm.coalition_values\"\n#\n# print(\"All good in _test_shapley_data_model\")\n", "sub_path": "stats/bin_est/shapley.py", "file_name": "shapley.py", "file_ext": "py", "file_size_in_byte": 13699, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pandas.DataFrame", "line_number": 18, "usage_type": "call"}, {"api_name": "ut.stats.bin_est.set_est.Shapley", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 31, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 36, "usage_type": "call"}, {"api_name": "itertools.imap", "line_number": 37, "usage_type": "call"}, {"api_name": "itertools.combinations", "line_number": 37, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 52, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 58, "usage_type": "argument"}, {"api_name": "collections.Counter", "line_number": 100, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 123, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 126, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 130, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 146, "usage_type": "call"}, {"api_name": "string.ascii_uppercase", "line_number": 155, "usage_type": "attribute"}, {"api_name": "random.sample", "line_number": 178, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 178, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 186, "usage_type": "call"}, {"api_name": "itertools.starmap", "line_number": 187, "usage_type": "call"}, {"api_name": "itertools.repeat", "line_number": 187, "usage_type": "call"}, {"api_name": "ut.util.uiter.all_subsets_of", "line_number": 193, "usage_type": "call"}]} +{"seq_id": "643197544", "text": "\"\"\" A class for testing a SSD model on a video file or webcam \"\"\"\nimport cv2\nfrom keras.applications.imagenet_utils import preprocess_input\nfrom keras.preprocessing import image\nimport numpy as np\nfrom model.ssd300MobileNet import SSD\nfrom model.utils.ssd_utils import BBoxUtility\nimport matplotlib.pyplot as plt\nfrom settings import *\nimport time\n\nclass MobileNetTest(object):\n \"\"\" Class for testing a trained SSD model on a video file and show the\n result in a window. Class is designed so that one VideoTest object\n can be created for a model, and the same object can then be used on\n multiple videos and webcams.\n\n Arguments:\n class_names: A list of strings, each containing the name of a class.\n The first name should be that of the background class\n which is not used.\n\n model: An SSD model. It should already be trained for\n images similar to the video to test on.\n\n input_shape: The shape that the model expects for its input,\n as a tuple, for example (300, 300, 3)\n\n bbox_util: An instance of the BBoxUtility class in ssd_utils.py\n The BBoxUtility needs to be instantiated with\n the same number of classes as the length of\n class_names.\n \"\"\"\n def __init__(self, class_names, weight_path, input_shape):\n self.class_names = class_names\n self.num_classes = len(class_names)\n self.input_shape = input_shape\n self.model = SSD(self.input_shape, num_classes=self.num_classes)\n self.model.load_weights(weight_path)\n self.model._make_predict_function()\n self.bbox_util = BBoxUtility(self.num_classes)\n\n # Create unique and somewhat visually distinguishable bright\n # colors for the different classes.\n self.class_colors = []\n for i in range(0, self.num_classes):\n # This can probably be written in a more elegant manner\n hue = 255 * i / self.num_classes\n col = np.zeros((1, 1, 3)).astype(\"uint8\")\n col[0][0][0] = hue\n col[0][0][1] = 128 # Saturation\n col[0][0][2] = 255 # Value\n cvcol = cv2.cvtColor(col, cv2.COLOR_HSV2BGR)\n col = (int(cvcol[0][0][0]), int(cvcol[0][0][1]), int(cvcol[0][0][2]))\n self.class_colors.append(col)\n\n def run(self, frame, frame_num, conf_thresh=0.6):\n \"\"\" Runs the test on a video (or webcam)\n\n # Arguments\n conf_thresh: Threshold of confidence. Any boxes with lower confidence\n are not visualized.\n \"\"\"\n output_list = list()\n im_size = (self.input_shape[0], self.input_shape[1])\n resized = cv2.resize(frame, im_size)\n orig_image = cv2.cvtColor(resized, cv2.COLOR_RGB2BGR)\n to_draw = resized\n\n # Use model to predict\n inputs = [image.img_to_array(orig_image)]\n tmp_inp = np.array(inputs)\n x = preprocess_input(tmp_inp)\n y = self.model.predict(x)\n results = self.bbox_util.detection_out(y)\n\n if len(results) > 0 and len(results[0]) > 0:\n # Interpret output, only one frame is used\n det_label = results[0][:, 0]\n det_conf = results[0][:, 1]\n det_xmin = results[0][:, 2]\n det_ymin = results[0][:, 3]\n det_xmax = results[0][:, 4]\n det_ymax = results[0][:, 5]\n\n top_indices = [i for i, conf in enumerate(det_conf) if conf >= conf_thresh]\n\n top_conf = det_conf[top_indices]\n top_label_indices = det_label[top_indices].tolist()\n top_xmin = det_xmin[top_indices]\n top_ymin = det_ymin[top_indices]\n top_xmax = det_xmax[top_indices]\n top_ymax = det_ymax[top_indices]\n\n output_list.append(frame_num)\n for i in range(top_conf.shape[0]):\n if (top_conf[i] < 0.9):\n continue\n xmin = int(round(top_xmin[i] * to_draw.shape[1]))\n ymin = int(round(top_ymin[i] * to_draw.shape[0]))\n xmax = int(round(top_xmax[i] * to_draw.shape[1]))\n ymax = int(round(top_ymax[i] * to_draw.shape[0]))\n\n # Draw the box on top of the to_draw image\n class_num = int(top_label_indices[i])\n cv2.rectangle(to_draw, (xmin, ymin), (xmax, ymax),\n self.class_colors[class_num], 2)\n text = self.class_names[class_num] + \" \" + ('%.2f' % top_conf[i])\n output_list.append(self.class_names[class_num])\n\n text_top = (xmin, ymin - 10)\n text_bot = (xmin + 80, ymin + 5)\n text_pos = (xmin + 5, ymin)\n cv2.rectangle(to_draw, text_top, text_bot, self.class_colors[class_num], -1)\n cv2.putText(to_draw, text, text_pos, cv2.FONT_HERSHEY_SIMPLEX, 0.35, (0, 0, 0), 1)\n\n cv2.imshow(\"SSD result\", to_draw)\n cv2.waitKey(10)\n\n def draw_fps(self, fps_time_slot):\n x_range = [index for index, value in enumerate(fps_time_slot)]\n y_fps = [value[1] for index, value in enumerate(fps_time_slot)]\n plt.ylim((0, 5))\n plt.plot(x_range, y_fps, c='r')\n plt.xlabel('time')\n plt.ylabel('fps')\n plt.show()\n\n\n", "sub_path": "mobilenettest.py", "file_name": "mobilenettest.py", "file_ext": "py", "file_size_in_byte": 5441, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "model.ssd300MobileNet.SSD", "line_number": 38, "usage_type": "call"}, {"api_name": "model.utils.ssd_utils.BBoxUtility", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 49, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 53, "usage_type": "call"}, {"api_name": "cv2.COLOR_HSV2BGR", "line_number": 53, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 66, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 67, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2BGR", "line_number": 67, "usage_type": "attribute"}, {"api_name": "keras.preprocessing.image.img_to_array", "line_number": 71, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 71, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 72, "usage_type": "call"}, {"api_name": "keras.applications.imagenet_utils.preprocess_input", "line_number": 73, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 106, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 114, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 115, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 115, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 117, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 123, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 123, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 124, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 125, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}]} +{"seq_id": "56883559", "text": "#!/usr/bin/env python3\nfrom multiprocessing import Process, Value\nimport time\nimport sys\nimport xmlrpc.client\n\ndef call_rpc(errors, i, num):\n try:\n for j in range(0, num):\n s = xmlrpc.client.ServerProxy('https://localhost:8000')\n s.test(i)\n except Exception as Ex:\n errors.value += 1\n\ndef jobs_process(errors, process_n, num):\n for i in range(process_n):\n p = Process(target=call_rpc, args=(errors, i, num))\n p.start()\n\nif __name__ == '__main__':\n process_n, num_n = sys.argv[1:]\n errors = Value('i', 0)\n intprocess = int(process_n)\n num = int(num_n)\n start = time.time()\n jobs = Process(target=jobs_process, args=(errors, intprocess, num))\n jobs.start()\n jobs.join()\n took = time.time() - start\n print(\"Total jobs: %s\" % (process_n))\n print(\"RPC Errors: %s\" % (errors.value))\n print(\"Elapsed time: %s\" % (took))\n sys.exit(0)\n\n \n \n", "sub_path": "pileus-siml/OR/clientPerf2.py", "file_name": "clientPerf2.py", "file_ext": "py", "file_size_in_byte": 943, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "xmlrpc.client.client.ServerProxy", "line_number": 10, "usage_type": "call"}, {"api_name": "xmlrpc.client.client", "line_number": 10, "usage_type": "attribute"}, {"api_name": "xmlrpc.client", "line_number": 10, "usage_type": "name"}, {"api_name": "multiprocessing.Process", "line_number": 17, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 21, "usage_type": "attribute"}, {"api_name": "multiprocessing.Value", "line_number": 22, "usage_type": "call"}, {"api_name": "time.time", "line_number": 25, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 26, "usage_type": "call"}, {"api_name": "time.time", "line_number": 29, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 33, "usage_type": "call"}]} +{"seq_id": "527266734", "text": "import matplotlib.pyplot as plt\nimport numpy as np\n\ndef read_file(filename):\n infile = open(filename,\"r\")\n first_line = infile.readline()\n if first_line[0] == \"|\":\n N = eval(first_line.split()[1][2:])\n epsilon_max = eval(first_line.split()[3][12:])\n epsilon_tot = eval(first_line.split()[5][12:])\n cpu_time = eval(first_line.split()[7][9:])\n log_10_h = eval(first_line.split()[9][10:])\n\n x = []; v = []; u = []\n infile.readline()\n for line in infile:\n inline = line.split()\n x.append(eval(inline[0]))\n v.append(eval(inline[1]))\n u.append(eval(inline[2]))\n if filename[7:10] == \"spe\":\n FLOPS = 7*N\n if filename[7:10] == \"gen\":\n FLOPS = N*11\n infile.close()\n return x,v,u,FLOPS,N,epsilon_max,epsilon_tot,cpu_time,log_10_h\n\n if first_line[0] == \"N\":\n N = []; epsilon_max = []; log_10_h = []; cpu_time = [];\n N.append(eval(first_line.split()[0][2:]))\n epsilon_max.append(eval(first_line.split()[1][12:]))\n log_10_h.append(eval(first_line.split()[2][10:]))\n cpu_time.append(eval(first_line.split()[3][9:]))\n for line in infile:\n N.append(eval(line.split()[0][2:]))\n epsilon_max.append(eval(line.split()[1][12:]))\n log_10_h.append(eval(line.split()[2][10:]))\n cpu_time.append(eval(line.split()[3][9:]))\n return N, epsilon_max, log_10_h, cpu_time\n\n\nfilenames_1 = [\"./data/genN10.txt\",\"./data/genN100.txt\",\"./data/genN1000.txt\",\\\n\"./data/speN10.txt\",\"./data/speN100.txt\",\"./data/speN1000.txt\"]\n\nfilenames_2 = [\"./data/gen_stats.txt\",\"./data/spe_stats.txt\"]\n\n\nfor filename in filenames_1:\n x,v,u,FLOPS,N,epsilon_max,epsilon_tot,cpu_time,log_10_h = read_file(filename)\n plt.xlabel(\"x\")\n plt.ylabel(\"u(x), v(x)\")\n\n plt.plot(x,v,label=r\"Numerical solution, $n={:}$ steps\".format(N))\n plt.plot(x,u,label=\"Analytic solution\")\n if filename[7:10] == \"spe\":\n type = \"Special algo\"\n if filename[7:10] == \"gen\":\n type = \"General algo\"\n #plt.title(r\"{:} $FLOPS={:}$\".format(type,FLOPS))\n plt.legend() ; plt.grid()\n argv = \"save\"\n if argv == \"plot\":\n plt.show()\n if argv == \"save\":\n plt.savefig(\"./figures/1b_{:}_{:}.png\".format(type[0:3],N))\n plt.clf()\n\nn_algos = []\ncpu_time_algos = []\nfor filename in filenames_2:\n N, epsilon_max, log_10_h, cpu_time = read_file(filename)\n cpu_time_algos.append(cpu_time)\n n_algos.append(N)\n plt.xlabel(r\"$log_{10}(h)$\")\n plt.ylabel(r\"$\\varepsilon$\")\n plt.plot(log_10_h,epsilon_max)\n if filename[7:10] == \"spe\":\n type = \"Special algo\"\n if filename[7:10] == \"gen\":\n type = \"General algo\"\n\n plt.title(\"{:}\".format(type))\n plt.grid(); #plt.legend()\n\n argv = \"save\"\n if argv == \"plot\":\n plt.show()\n if argv == \"save\":\n plt.savefig(\"./figures/1d_{:}_eps.png\".format(type[0:3]))\n plt.clf()\n\nfilenames = [\"./data/LU10.txt\",\"./data/LU100.txt\",\"./data/LU1000.txt\"]\ncpu_time_LU = []\nfor filename in filenames:\n infile = open(filename,\"r\")\n first_line = infile.readline().split()\n x = []; u = []; v = []\n for i in infile:\n line = i.split()\n x.append(eval(line[0]))\n v.append(eval(line[1]))\n u.append(eval(line[2]))\n cpu_time_LU.append(eval(first_line[-1][9:]))\n\n plt.plot(x,v,label=\"Numerical, n= {:}\".format(filename[9:11]))\n plt.plot(x,u,label=\"Analytic\")\n plt.xlabel(\"x\")\n plt.ylabel(\"u(x), v(x)\")\n plt.grid()\n plt.legend()\n plt.savefig(\"./figures/LU{:}.png\".format(len(x)-2))\n plt.clf()\n\nn = []\nfor i in range(1,8):\n n.append(int(i))\ncpu_time_gen = np.log10(cpu_time_algos[0])\ncpu_time_spe = np.log10(cpu_time_algos[1])\ncpu_time_LU = np.log10(cpu_time_LU)\n\nplt.plot(n,cpu_time_gen,label=\"General solution\")\nplt.plot(n,cpu_time_spe,label=\"Specialized solution\")\nplt.plot(n[0:3],cpu_time_LU,label=\"LU-decomposition\")\n\nplt.xlabel(\"n\")\nplt.ylabel(r\"$log_10$(CPU-time) [s]\")\nplt.xticks(ticks = n,labels=[r\"$10^1$\",r\"$10^2$\",r\"$10^3$\",r\"$10^4$\",r\"$10^5$\",r\"$10^6$\",r\"$10^7$\"])\nplt.legend()\nplt.savefig(\"./figures/CPU_times\")\nplt.clf()\n", "sub_path": "Project_1/reader.py", "file_name": "reader.py", "file_ext": "py", "file_size_in_byte": 4199, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "matplotlib.pyplot.xlabel", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 106, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 110, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "numpy.log10", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 121, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 122, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 123, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 123, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 125, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}]} +{"seq_id": "239212899", "text": "# Copyright 2014-2020 Scalyr Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"\nBenchmarks which compare various compression algorithms.\n\nNOTE: We also want to measure CPU utilization for those benchmarks which means we should also run\nthem using \"time.process_time\" timer which contains sum of system and user CPU time and not wall\nclock time.\n\nThis way we get accurate CPU utilization information.\n\"\"\"\n\nfrom __future__ import absolute_import\nfrom __future__ import print_function\n\nif False:\n from typing import Tuple\n from typing import Callable\n\nimport sys\nimport zlib\nimport bz2\nimport functools\n\nimport pytest\n\ntry:\n import snappy\nexcept ImportError:\n snappy = None\n\ntry:\n import zstandard\nexcept ImportError:\n zstandard = None\n\ntry:\n import brotli\nexcept ImportError:\n brotli = None\n\ntry:\n import lz4framed as lz4\nexcept ImportError:\n lz4 = None\n\n\nfrom .utils import read_bytes_from_log_fixture_file\nfrom .time_utils import process_time\n\n\n# fmt: off\n@pytest.mark.parametrize(\"log_tuple\",\n [\n (\"agent_debug_5_mb.log.gz\", 3 * 1024),\n (\"agent_debug_5_mb.log.gz\", 10 * 1024),\n (\"agent_debug_5_mb.log.gz\", 500 * 1024),\n (\"json_log_5_mb.log.gz\", 3 * 1024),\n (\"json_log_5_mb.log.gz\", 10 * 1024),\n (\"json_log_5_mb.log.gz\", 500 * 1024),\n (\"add_events_request_10_events.log.gz\", -1),\n (\"add_events_request_100_events.log.gz\", -1),\n (\"add_events_request_200_events.log.gz\", -1),\n (\"add_events_request_10_events_with_attributes.log.gz\", -1),\n (\"add_events_request_100_events_with_attributes.log.gz\", -1),\n (\"add_events_request_200_events_with_attributes.log.gz\", -1),\n ],\n ids=[\n \"agent_debug_log_3k\",\n \"agent_debug_log_10k\",\n \"agent_debug_log_500k\",\n \"json_log_3k\",\n \"json_log_10k\",\n \"json_log_500k\",\n \"add_events_10_events\",\n \"add_events_100_events\",\n \"add_events_200_events\",\n \"add_events_10_events_with_attributes\",\n \"add_events_100_events_with_attributes\",\n \"add_events_200_events_with_attributes\",\n ],\n)\n# fmt: on\n@pytest.mark.parametrize(\n \"compression_algorithm_tuple\",\n [\n (\"deflate\", {\"level\": 3}),\n (\"deflate\", {\"level\": 6}),\n (\"deflate\", {\"level\": 9}),\n (\"bz2\", {}),\n (\"snappy\", {}),\n (\"zstandard\", {\"level\": 3}),\n (\"zstandard\", {\"level\": 5}),\n (\"zstandard\", {\"level\": 10}),\n (\"zstandard\", {\"level\": 12}),\n (\"brotli\", {\"quality\": 3}),\n (\"brotli\", {\"quality\": 5}),\n (\"brotli\", {\"quality\": 8}),\n (\"lz4\", {}),\n ],\n ids=[\n \"deflate_level_3\",\n \"deflate_level_6_default\",\n \"deflate_level_9\",\n \"bz2\",\n \"snappy\",\n \"zstandard_level_3_default\",\n \"zstandard_level_5\",\n \"zstandard_level_10\",\n \"zstandard_level_12\",\n \"brotli_quality_3\",\n \"brotli_quality_5\",\n \"brotli_quality_8\",\n \"lz4\",\n ],\n)\n@pytest.mark.benchmark(group=\"compress\", timer=process_time)\ndef test_compress_bytes(benchmark, compression_algorithm_tuple, log_tuple):\n _test_compress_bytes(benchmark, compression_algorithm_tuple, log_tuple)\n\n\n# fmt: off\n@pytest.mark.parametrize(\"log_tuple\",\n [\n (\"agent_debug_5_mb.log.gz\", 3 * 1024),\n (\"agent_debug_5_mb.log.gz\", 10 * 1024),\n (\"agent_debug_5_mb.log.gz\", 500 * 1024),\n (\"json_log_5_mb.log.gz\", 3 * 1024),\n (\"json_log_5_mb.log.gz\", 10 * 1024),\n (\"json_log_5_mb.log.gz\", 500 * 1024),\n (\"add_events_request_10_events.log.gz\", -1),\n (\"add_events_request_100_events.log.gz\", -1),\n (\"add_events_request_200_events.log.gz\", -1),\n (\"add_events_request_10_events_with_attributes.log.gz\", -1),\n (\"add_events_request_100_events_with_attributes.log.gz\", -1),\n (\"add_events_request_200_events_with_attributes.log.gz\", -1),\n ],\n ids=[\n \"agent_debug_log_3k\",\n \"agent_debug_log_10k\",\n \"agent_debug_log_500k\",\n \"json_log_3k\",\n \"json_log_10k\",\n \"json_log_500k\",\n \"add_events_10_events\",\n \"add_events_100_events\",\n \"add_events_200_events\",\n \"add_events_10_events_with_attributes\",\n \"add_events_100_events_with_attributes\",\n \"add_events_200_events_with_attributes\",\n ],\n)\n@pytest.mark.parametrize(\"compression_algorithm_tuple\",\n [\n (\"deflate\", {\"level\": 3}),\n (\"deflate\", {\"level\": 6}),\n (\"deflate\", {\"level\": 9}),\n (\"bz2\", {}),\n (\"snappy\", {}),\n (\"zstandard\", {\"level\": 3}),\n (\"zstandard\", {\"level\": 5}),\n (\"zstandard\", {\"level\": 10}),\n (\"zstandard\", {\"level\": 12}),\n (\"brotli\", {\"quality\": 3}),\n (\"brotli\", {\"quality\": 5}),\n (\"brotli\", {\"quality\": 8}),\n (\"lz4\", {}),\n ],\n ids=[\n \"deflate_level_3\",\n \"deflate_level_6_default\",\n \"deflate_level_9\",\n \"bz2\",\n \"snappy\",\n \"zstandard_level_3_default\",\n \"zstandard_level_5\",\n \"zstandard_level_10\",\n \"zstandard_level_12\",\n \"brotli_quality_3\",\n \"brotli_quality_5\",\n \"brotli_quality_8\",\n \"lz4\",\n ],\n)\n# fmt: on\n@pytest.mark.benchmark(group=\"decompress\", timer=process_time)\ndef test_decompress_bytes(benchmark, compression_algorithm_tuple, log_tuple):\n _test_decompress_bytes(benchmark, compression_algorithm_tuple, log_tuple)\n\n\ndef _test_compress_bytes(benchmark, compression_algorithm_tuple, log_tuple):\n compression_algorithm, kwargs = compression_algorithm_tuple\n\n file_name, bytes_to_read = log_tuple\n data = read_bytes_from_log_fixture_file(file_name, bytes_to_read)\n\n compress_func, decompress_func = _get_compress_and_decompress_func(\n compression_algorithm, kwargs\n )\n\n def run_benchmark():\n # Work around for Python <= 3.6 where compress is not a keyword argument, but a regular argument\n if sys.version_info < (3, 6, 0) and compression_algorithm == \"deflate\":\n result = compress_func(data, kwargs[\"level\"])\n else:\n result = compress_func(data)\n return result\n\n result = benchmark.pedantic(run_benchmark, iterations=10, rounds=20)\n\n size_before_compression = len(data)\n size_after_compression = len(result)\n compression_ratio = float(size_before_compression) / size_after_compression\n\n benchmark.stats.size_before_compression = size_before_compression\n benchmark.stats.size_after_compression = size_after_compression\n benchmark.stats.stats.compression_ratio = compression_ratio\n\n assert result is not None\n # assert correctness\n assert size_after_compression < size_before_compression\n assert data == decompress_func(result)\n\n\ndef _test_decompress_bytes(benchmark, compression_algorithm_tuple, log_tuple):\n compression_algorithm, kwargs = compression_algorithm_tuple\n\n file_name, bytes_to_read = log_tuple\n data = read_bytes_from_log_fixture_file(file_name, bytes_to_read)\n\n compress_func, _ = _get_compress_and_decompress_func(compression_algorithm, kwargs)\n\n compressed_data = compress_func(data)\n assert compressed_data != data\n\n # NOTE: We intentionally request new decompression function so we get new zstandard context for\n # decompression (this way we avoid dictionary being already populated).\n _, decompress_func = _get_compress_and_decompress_func(\n compression_algorithm, kwargs\n )\n\n def run_benchmark():\n result = decompress_func(compressed_data)\n return result\n\n result = benchmark.pedantic(run_benchmark, iterations=10, rounds=20)\n\n size_before_decompression = len(compressed_data)\n size_after_decompression = len(result)\n\n assert result is not None\n # assert correctness\n assert result != compressed_data\n assert size_after_decompression > size_before_decompression\n assert data == result\n\n\ndef _get_compress_and_decompress_func(compression_algorithm, kwargs):\n # type: (str, dict) -> Tuple[Callable, Callable]\n if compression_algorithm == \"deflate\":\n if sys.version_info < (3, 6, 0):\n # Work around for Python <= 3.6 where compress is not a keyword argument, but a regular\n # argument\n compress_func = zlib.compress # type: ignore\n else:\n compress_func = functools.partial(zlib.compress, **kwargs) # type: ignore\n decompress_func = zlib.decompress # type: ignore\n elif compression_algorithm == \"bz2\":\n compress_func = functools.partial(bz2.compress, **kwargs) # type: ignore\n decompress_func = bz2.decompress # type: ignore\n elif compression_algorithm == \"snappy\":\n compress_func = functools.partial(snappy.compress, **kwargs) # type: ignore\n decompress_func = snappy.decompress # type: ignore\n elif compression_algorithm == \"zstandard\":\n compressor = zstandard.ZstdCompressor(**kwargs)\n decompressor = zstandard.ZstdDecompressor()\n compress_func = compressor.compress # type: ignore\n decompress_func = decompressor.decompress # type: ignore\n elif compression_algorithm == \"brotli\":\n compress_func = functools.partial(brotli.compress, **kwargs) # type: ignore\n decompress_func = brotli.decompress # type: ignore\n elif compression_algorithm == \"lz4\":\n compress_func = functools.partial(lz4.compress, **kwargs) # type: ignore\n decompress_func = lz4.decompress # type: ignore\n else:\n raise ValueError(\"Unsupported algorithm: %s\" % (compression_algorithm))\n\n return compress_func, decompress_func\n", "sub_path": "benchmarks/micro/test_compression_algorithms.py", "file_name": "test_compression_algorithms.py", "file_ext": "py", "file_size_in_byte": 10117, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "pytest.mark.parametrize", "line_number": 65, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 65, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 96, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 96, "usage_type": "attribute"}, {"api_name": "pytest.mark.benchmark", "line_number": 129, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 129, "usage_type": "attribute"}, {"api_name": "time_utils.process_time", "line_number": 129, "usage_type": "name"}, {"api_name": "pytest.mark.parametrize", "line_number": 135, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 135, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 165, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 165, "usage_type": "attribute"}, {"api_name": "pytest.mark.benchmark", "line_number": 198, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 198, "usage_type": "attribute"}, {"api_name": "time_utils.process_time", "line_number": 198, "usage_type": "name"}, {"api_name": "utils.read_bytes_from_log_fixture_file", "line_number": 207, "usage_type": "call"}, {"api_name": "sys.version_info", "line_number": 215, "usage_type": "attribute"}, {"api_name": "utils.read_bytes_from_log_fixture_file", "line_number": 241, "usage_type": "call"}, {"api_name": "sys.version_info", "line_number": 273, "usage_type": "attribute"}, {"api_name": "zlib.compress", "line_number": 276, "usage_type": "attribute"}, {"api_name": "functools.partial", "line_number": 278, "usage_type": "call"}, {"api_name": "zlib.compress", "line_number": 278, "usage_type": "attribute"}, {"api_name": "zlib.decompress", "line_number": 279, "usage_type": "attribute"}, {"api_name": "functools.partial", "line_number": 281, "usage_type": "call"}, {"api_name": "bz2.compress", "line_number": 281, "usage_type": "attribute"}, {"api_name": "bz2.decompress", "line_number": 282, "usage_type": "attribute"}, {"api_name": "functools.partial", "line_number": 284, "usage_type": "call"}, {"api_name": "snappy.compress", "line_number": 284, "usage_type": "attribute"}, {"api_name": "snappy.decompress", "line_number": 285, "usage_type": "attribute"}, {"api_name": "zstandard.ZstdCompressor", "line_number": 287, "usage_type": "call"}, {"api_name": "zstandard.ZstdDecompressor", "line_number": 288, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 292, "usage_type": "call"}, {"api_name": "brotli.compress", "line_number": 292, "usage_type": "attribute"}, {"api_name": "brotli.decompress", "line_number": 293, "usage_type": "attribute"}, {"api_name": "functools.partial", "line_number": 295, "usage_type": "call"}, {"api_name": "lz4framed.compress", "line_number": 295, "usage_type": "attribute"}, {"api_name": "lz4framed.decompress", "line_number": 296, "usage_type": "attribute"}]} +{"seq_id": "202449053", "text": "class AppModel(models.Model):\n\n \"\"\"\n save时处理\n \"\"\"\n def create_thumbnail(self, image: ImageFieldFile):\n from PIL import Image\n from io import BytesIO\n from django.core.files.base import ContentFile\n thumbnail_size = 100, 120\n\n tiny_img = Image.open(image)\n tiny_img.thumbnail(thumbnail_size)\n tiny_img.save(image.file, format=tiny_img.format)\n f = BytesIO()\n tiny_img.save(f, format=tiny_img.format)\n image.save(image.name, ContentFile(f.getvalue()), save=False)\n\n\n\nclass DefectImageSerializer(serializers.ModelSerializer):\n\n def create(self, validated_data):\n create = super().create(validated_data)\n if not create.image:\n return create\n\n from PIL import Image\n from pathlib import Path\n from io import BytesIO\n thumbnail_size = 100, 120\n\n filepath = str(create.image.file)\n tiny_img = Image.open(filepath)\n tiny_img.thumbnail(thumbnail_size)\n\n file = Path(filepath)\n thumb_name = f'{file.stem}_thumbnail{file.suffix}'\n imgio = BytesIO()\n\n tiny_img.save(imgio, format=tiny_img.format)\n tiny_img.close()\n create.image_thumbnail.save(thumb_name, imgio)\n create.save()\n return create\n\n class Meta:\n model = models.DefectImage\n fields = '__all__'", "sub_path": "python_01/image/drf_image_create_thumbnail.py", "file_name": "drf_image_create_thumbnail.py", "file_ext": "py", "file_size_in_byte": 1380, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "PIL.Image.open", "line_number": 12, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 12, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 15, "usage_type": "call"}, {"api_name": "django.core.files.base.ContentFile", "line_number": 17, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 34, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 34, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 37, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 39, "usage_type": "call"}]} +{"seq_id": "246925229", "text": "#importation des donnees order book en format json_data\n#construction d'une dataframe\n#rajout des colonnes volume total style somme execl\n#rajout d'une colonne detection des walls\n#enregistrement des donnees traitees dans data_treated.JSON\n\nimport json\nimport pandas as pd\nimport numpy as np\nimport sys\nimport os.path\nhomedir = os.path.expanduser(\"~\")\n################################################################################\n #fonctions\n################################################################################\n#lecture du fichier JSON\ndef dataquisition(data):\n with open(data) as json_data:\n data_dict = json.load(json_data)\n json_data.close()\n #transformation JSON en 2 dataframe\n bids = data_dict['bids']\n asks = data_dict['asks']\n #print(bids[1]['price'])\n listbids = []\n listasks = []\n\n #transformation en dictionnaire\n for bid in bids :\n listbids.append({'value' : float(bid['price']), 'bidsvolume' :float(bid['size'])})\n for ask in asks :\n listasks.append({'value' : float(ask['price']), 'asksvolume' :float(ask['size'])})\n\n #ajout de la liste de dicitonnaire dans les dataframes\n dfbids = pd.DataFrame(listbids)\n dfasks = pd.DataFrame(listasks)\n return [dfbids,dfasks]\n\n#calcul des volumes vente et achat\ndef calculVolumeGlobal (dfbids,dfasks):\n #calcul du volume global et detection des walls\n dfbids['bidstotalvolume'] = dfbids['bidsvolume'].cumsum(axis = 0)\n dfasks['askstotalvolume'] = dfasks['asksvolume'].cumsum(axis = 0)\n #le 0.1 correspond a 10% du volume global\n comparator = lambda x: 1000 if x>0.1 else 0\n\n\n askstotal = dfasks['askstotalvolume'].iloc[-1]\n dfasks['askswall'] = dfasks['asksvolume'].multiply(1/askstotal)\n dfasks['askswall'] = dfasks['askswall'].apply(comparator)\n\n bidstotal = dfbids['bidstotalvolume'].iloc[-1]\n dfbids['bidswall'] = dfbids['bidsvolume'].multiply(1/bidstotal)\n dfbids['bidswall'] = dfbids['bidswall'].apply(comparator)\n\n return [bidstotal, askstotal]\n#tendance\n #achat/vente/nondefini\ndef tendance (bidstotal,askstotal,seuilvente = 51,seuilachat = 51):\n TACHAT = 1\n TVENTE = 2\n TUNDEFINED = 3\n total = askstotal + bidstotal\n pourcentagevente = int(askstotal/total*100)\n pourcentageachat = int(bidstotal/total*100)\n\n _tendance = TUNDEFINED\n\n if pourcentagevente > seuilvente :\n _tendance = TVENTE\n if pourcentageachat > seuilachat and _tendance != TVENTE:\n _tendance = TACHAT\n\n return _tendance\n\n#filtreWall\ndef filtreWall(df,type):\n if (type == \"bids\"):\n dfbids_filtre = df.drop(df[df.bidswall == 1000].index)\n return dfbids_filtre\n if (type == \"asks\"):\n dfasks_filtre = df.drop(df[df.askswall == 1000].index)\n return dfasks_filtre\n else : print(\"erreur\")\n\n#placement_ordre\ndef placement_ordre(tendance,dfbids,dfasks):\n TACHAT = 1\n TVENTE = 2\n TUNDEFINED = 3\n\n def fAchat():\n dwall = dfbids.drop(dfbids[dfbids.bidswall != 1000].index)\n dwall = dwall.sort_values(by = 'value')\n if(dwall.empty):\n dwall = dfasks[\"value\"].iloc[-1]-0.00000001\n return[\"bid\",float(dwall)]\n else:\n #print dwall\n dwall = dwall[\"value\"].iloc[-1]\n return [\"bid\",float(dwall)-0.00000001]\n\n def fVente():\n #print(dfasks)\n dwall = dfasks.drop(dfasks[dfasks.askswall != 1000].index)\n\n if (dwall.empty):\n dwall = dfasks[\"value\"].iloc[0]+0.00000001\n return [\"ask\",float(dwall)]\n else:\n dwall = dwall.sort_values(by = 'value')\n dwall = dwall[\"value\"].iloc[0]\n return [\"ask\",float(dwall)+0.00000001]\n\n\n\n def fRien():\n return [\"undefined\",-1.0]\n\n switcher={\n TVENTE: fVente,\n TACHAT: fAchat,\n TUNDEFINED: fRien\n }\n func = switcher.get(tendance, lambda: \"argumentinvalide\")\n return func()\n\n################################################################################\n #programme\n################################################################################\nif __name__ == '__main__':\n import glob, os\n os.chdir(homedir+\"/server_crypto/orderbook_data/\")\n for file in glob.glob(\"*.json\"):\n timestamp = os.path.splitext(os.path.basename(file))[0]\n #aquisition des donnees\n [dfbids,dfasks] = dataquisition(file)\n #ajout des wall\n [askstotal,bidstotal] = calculVolumeGlobal(dfbids,dfasks)\n #print(dfbids)\n tendavantfiltrage = tendance(bidstotal,askstotal,50,50)\n #print(tendavantfiltrage)\n\n #filtrage des wall\n dfasks_filtre = filtreWall(dfasks,\"asks\")\n dfbids_filtre = filtreWall(dfbids,\"bids\")\n #tendance\n [askstotal_f,bidstotal_f] = calculVolumeGlobal(dfbids_filtre,dfasks_filtre)\n tendance_f = tendance(bidstotal_f,askstotal_f,50,50)\n\n #renvoi un tuple contenant l'action a executer (vendrvalorisation.pye/acheter/rien)\n #ainsi que prix a rentrer\n [_tendance, _prix] = placement_ordre(tendavantfiltrage,dfbids,dfasks)\n\n ##################################################################################\n #sauvegarde de l'algo\n\n import time\n now = int(time.time())\n date = now-now%60\n\n exists = os.path.isfile(homedir +'/server_crypto/data/algosignal_ws.json')\n def dataAlgo(_tendance,_prix):\n if exists:\n with open(homedir +'/server_crypto/data/algosignal_ws.json', 'r') as json_algo:\n algosignal = json.load(json_algo)\n algosignal[\"algosignal\"].append({\"tendance\":_tendance,\"prix\":_prix,\"time\":int(timestamp)})\n json_algo.close()\n algosignal[\"algosignal\"] = {frozenset(item.items()) : item for item in algosignal[\"algosignal\"]}.values()\n\n with open(homedir +'/server_crypto/data/algosignal_ws.json', 'w') as json_algo:\n json.dump(algosignal, json_algo)\n json_algo.close()\n #ajouter un element au fichier avec la date la tendnace et le prix\n else:\n with open(homedir +'/server_crypto/data/algosignal_ws.json', 'w') as json_algo:\n algosignal = {\"algosignal\":[{\"tendance\":_tendance,\"prix\":_prix,\"time\":int(timestamp)}]}\n json.dump(algosignal, json_algo)\n json_algo.close()\n\n def mergetlohcv():\n dates =[]\n match_tlohcv=[]\n with open(homedir +'/server_crypto/data/algosignal_ws.json', 'r') as data_lohcv:\n _signal = json.load(data_lohcv)\n for date in _signal[\"algosignal\"]:\n\n dates.append(int(date['time']))\n with open(homedir +'/server_crypto/data/data_received_tohlcv_ws.json', 'r') as data_lo:\n tlohc = json.load(data_lo)\n\n for ele in tlohc:\n\n if ele[\"time\"] in dates:\n match_tlohcv.append(ele)\n data_lo.close()\n data_lohcv.close()\n with open(homedir +'/server_crypto/data/algosignal_ws.json', 'w') as data_l:\n _signal[\"tohlcv\"] = match_tlohcv\n # match_tlohcv\n json.dump(_signal,data_l)\n\n ################################################################################\n dataAlgo(_tendance,_prix)\n mergetlohcv()\n #Affichage des donnees\n ################################################################################\n\n #concatenation de bids et asks + reindexage\n dfall = pd.concat([dfbids,dfasks],sort=False,ignore_index=True)\n dfall = dfall.sort_values(by=['value']).reset_index(drop = True).replace(0,value = np.nan)\n\n dfall_filtered = pd.concat([dfbids_filtre,dfasks_filtre],sort=False,ignore_index=True,)\n dfall_filtered = dfall_filtered.sort_values(by=['value']).reset_index(drop = True).replace(0,value = np.nan)\n #suppression des walls\n\n\n\n #dfall_filtered.to_json(homedir+\"/my-app/src/assets/data_treated_filtered.json\",orient = 'table',index = False)\n\n #dfall.to_json(homedir+\"/my-app/src/assets/data_treated.json\",orient = 'table',index = False)\n\n\n sys.exit()\n", "sub_path": "walldetection_websocket.py", "file_name": "walldetection_websocket.py", "file_ext": "py", "file_size_in_byte": 8485, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "os.path.path.expanduser", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 12, "usage_type": "name"}, {"api_name": "json.load", "line_number": 19, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 35, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 36, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 133, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 134, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 135, "usage_type": "call"}, {"api_name": "os.path", "line_number": 135, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 135, "usage_type": "call"}, {"api_name": "time.time", "line_number": 159, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 162, "usage_type": "call"}, {"api_name": "os.path", "line_number": 162, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 166, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 172, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 178, "usage_type": "call"}, {"api_name": "json.load", "line_number": 185, "usage_type": "call"}, {"api_name": "json.load", "line_number": 190, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 201, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 210, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 211, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 214, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 224, "usage_type": "call"}]} +{"seq_id": "112427530", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nimport scipy.optimize as op\nimport emcee\nimport corner\nimport yaml\nimport likelihood.uniform as l\nimport pickle\nimport utils as u\n\n# load MC samples, names of satellites\nsample = 'fritzplusMCs'\ntag = 'uniform_noMCsystem'\nMC_dwarfs = np.load('data/sampling/'+sample+'.npy')\nwith open('data/sampling/names_key.pkl', 'rb') as f:\n names = pickle.load(f)[sample]\nassert MC_dwarfs.shape[0] == len(names)\n\n\"\"\"\n# cut based on distances\ndists = MC_dwarfs[:,6,:]\ndists = np.median(dists, axis=1)\ninc = dists < 100\nMC_dwarfs = MC_dwarfs[inc]\n# \"\"\"\n\n\"\"\"\n# use satellites from Cautun & Frenk (2017)\ncautun = ['Sagittarius I', 'LMC', 'SMC', 'Draco I', 'Ursa Minor', 'Sculptor',\n 'Carina I', 'Fornax', 'Leo II', 'Leo I']\ncautun = np.array([names.index(sat) for sat in cautun])\nMC_dwarfs = MC_dwarfs[cautun]\n# \"\"\"\n\n# \"\"\"\n# ignore satellites by name\n# ignoresats = ['Horologium I', 'Carina II', 'Carina III', 'Hydrus I']\n# ignoresats = ['LMC', 'SMC']\nignoresats = ['Horologium I', 'Carina II', 'Carina III', 'Hydrus I', 'LMC',\n 'SMC']\nignore = [names.index(sat) for sat in ignoresats]\nMC_dwarfs = np.delete(MC_dwarfs, ignore, axis=0)\n# \"\"\"\n\n# data and covariances for each satellite\nMC_vels = MC_dwarfs[:,9:12,:]\nvels = np.mean(MC_vels, axis=2)\nvel_covs = np.array([np.cov(dwarf) for dwarf in MC_vels])\n\n# Initialize walkers by randomly sampling prior\nnwalkers = 100\np0 = l.sample_prior(nwalkers=nwalkers)\nndim = len(p0[0])\n\n# Set up and run MCMC\nsampler = emcee.EnsembleSampler(nwalkers, ndim, l.lnprob, args=(vels,vel_covs))\npos, prob, state = sampler.run_mcmc(p0, 500)\n\n# Look by eye at the burn-in\nstepnum = np.arange(0,500,1)+1\nstepnum = np.array([stepnum for i in range(nwalkers)])\nplt.plot(stepnum, sampler.chain[:,:,0]);\n\nprint(\"Mean acceptance fraction: {0:.3f}\"\n .format(np.mean(sampler.acceptance_fraction)))\n\n# if needed, reset and run chain for new sample\nsampler.reset()\npos, prob, state = sampler.run_mcmc(pos, 500)\n\n# Flatten the chain and remove burn-in\nburnin = 0\nsamples = sampler.chain[:, burnin:, :].reshape((-1, ndim))\n\n# Make corner plot\nfig = corner.corner(samples, labels=[r\"$v_r$\", r\"$v_\\theta$\", r\"$v_\\phi$\",\n r\"$\\sigma_r$\", r\"$\\sigma_\\theta$\", r\"$\\sigma_\\phi$\"],\n quantiles=[0.16, 0.5, 0.84],\n show_titles=True, title_kwargs={\"fontsize\": 12})\n\nfig.savefig('figures/cornerplots/'+tag+'.png', bbox_inches='tight')\nnp.save(u.SIM_DIR+'beta/mcmc/data/'+tag, samples)\n", "sub_path": "likelihood_uniform.py", "file_name": "likelihood_uniform.py", "file_ext": "py", "file_size_in_byte": 2533, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "numpy.load", "line_number": 14, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.cov", "line_number": 48, "usage_type": "call"}, {"api_name": "likelihood.uniform.sample_prior", "line_number": 52, "usage_type": "call"}, {"api_name": "likelihood.uniform", "line_number": 52, "usage_type": "name"}, {"api_name": "emcee.EnsembleSampler", "line_number": 56, "usage_type": "call"}, {"api_name": "likelihood.uniform.lnprob", "line_number": 56, "usage_type": "attribute"}, {"api_name": "likelihood.uniform", "line_number": 56, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 65, "usage_type": "call"}, {"api_name": "corner.corner", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 82, "usage_type": "call"}, {"api_name": "utils.SIM_DIR", "line_number": 82, "usage_type": "attribute"}]} +{"seq_id": "501390061", "text": "import myread as load\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\nimport cv2\r\nimport math as mt\r\nfrom mpl_toolkits.mplot3d import Axes3D\r\nfrom scipy import ndimage\r\nimport filtertd as filt\r\nimport time\r\nimport peakutils\r\ndef getSymAngle(images):\r\n# output = load.xypt('data_dvslaser/bottle_wallet_phone.aedat')\r\n\r\n\t\r\n\tdisplayimages=[[0 for i in range(128)] for j in range(128)]\r\n\r\n\r\n\timages = ndimage.filters.median_filter(images,size = 3)\r\n\r\n\t# for i in range(128):\r\n\t# \tfor j in range(128):\r\n\t# \t\tdisplayimages[i][j]=images[127-i][j]\r\n\r\n\tfilteredimages=[[0 for i in range(128)] for j in range(128)]\r\n\tt1 = time.time()\r\n\tdistancetransformimage=ndimage.distance_transform_edt(np.logical_not(images))\r\n\tprint(time.time() - t1)\r\n\tsigma=5.0\r\n\tcomx=0.0\r\n\tcomy=0.0\r\n\ttotalevents=0\r\n\r\n\tfor i in range(128):\r\n\t\tfor j in range(128):\r\n\t\t\t# intensity=images[i][j]\r\n\t\t\tintensity= distancetransformimage[i][j]\r\n\t\t\tfilteredimages[i][j] = mt.exp(-(intensity/sigma)**2)\r\n\t\t\tcomx+=j*filteredimages[i][j]\r\n\t\t\tcomy+=i*filteredimages[i][j]\r\n\t\t\ttotalevents+=filteredimages[i][j]\r\n\r\n\tfilteredimages = np.array(filteredimages)\r\n\r\n\t# # for i in range(\r\n\r\n\tfor i in range(128):\r\n\t\tfor j in range(128):\r\n\t\t\tdisplayimages[i][j]=filteredimages[127-i][j]\r\n\r\n\r\n\t# for i in range(128):\r\n\t# \tfor j in range(128):\r\n\t# \t\tcomx+=j*filteredimages[i][j]\r\n\t# \t\tcomy+=i*filteredimages[i][j]\r\n\t# \t\ttotalevents+=filteredimages[i][j]\r\n\r\n\tcomx=mt.floor(comx/totalevents)\r\n\tcomy=mt.floor(comy/totalevents)\r\n\r\n\tS=[0 for i in range(36)]\r\n\tfor var in range(0,36):\r\n\t\ttheta=5*var\r\n\t\ttheta = 180-theta\r\n\t\tdenom=0\r\n\t\t# print(theta)\r\n\t\tfor i in range(128):\r\n\t\t\tfor j in range(128):\r\n\t\t\t\tp=i-comy\r\n\t\t\t\tq=j-comx\r\n\t\t\t\tradiantheta = theta*mt.pi/180\r\n\t\t\t\tnewx= 2*(comx+p*mt.sin(radiantheta)*mt.cos(radiantheta) + q*((mt.cos(radiantheta))**2)) - j\r\n\t\t\t\tnewy = 2*(comy+q*mt.sin(radiantheta)*mt.cos(radiantheta) + p*((mt.sin(radiantheta))**2)) - i\r\n\t\t\t\t\r\n\t\t\t\t# newx = (mt.cos(2*radiantheta)*i + mt.sin(2*radiantheta)*j) \r\n\t\t\t\t# newy = (mt.sin(2*radiantheta)*i - mt.cos(2*radiantheta)*j)\r\n\r\n\t\t\t\tnewx=int(newx)\r\n\t\t\t\tnewy=int(newy)\r\n\t\t\t\t# print(theta)\r\n\t\t\t\t# print(j,i)\r\n\t\t\t\t# print(newx,newy)\r\n\t\t\t\t# print(\"fwfe\")\r\n\r\n\t\t\t\tdenom += (filteredimages[i][j])**2\r\n\t\t\t\tif(newy>=0 and newy<128 and newx>=0 and newx<128):\r\n\t\t\t\t\t#print(filteredimages[i][j], \" \", filteredimages[newy][newx])\r\n\t\t\t\t\tS[var] += filteredimages[i][j]*filteredimages[newy][newx]\r\n\t\tS[var]=S[var]/denom\r\n\r\n\tS_Axis = []\r\n\tAngle = []\r\n\r\n\tprint(S)\r\n\thalf_length = int(((len(S)/2)-1))\r\n\r\n\tsnew = S+S[0:half_length]\r\n\tindexes = peakutils.indexes(snew, thres=0.05/max(snew))\r\n\r\n\tprint(indexes)\r\n\r\n\tfor idx in indexes:\r\n\t\tidx = idx % 36\r\n\t\tAngle.append(idx)\r\n\r\n\t# for i in range(0,36):\r\n\t# \tleft = S[i-1]\r\n\t# \tif (i == 35):\r\n\t# \t\tright = S[0]\r\n\t# \telse:\r\n\t# \t\tright = S[i+1]\r\n\t# \tif(S[i]>left and S[i]>right):\r\n\t# \t\tS_Axis.append(S[i])\r\n\t# \t\tAngle.append(i)\r\n\r\n\t# print(S)\r\n\r\n\t# sortedS = sorted(S_Axis)\r\n\t# largestS = sortedS[-1]\r\n\t# secondS = sortedS[-2]\r\n\t# # thirdS = sortedS[-3]\r\n\r\n\t# i1 = S_Axis.index(largestS)\r\n\t# i2 = S_Axis.index(secondS)\r\n\t# # i3 = S_Axis.index(thirdS)\r\n\r\n\t# print(\"First symmetry axis = \" , Angle[i1]*5)\r\n\t# print(\"Second symmetry axis = \", Angle[i2]*5)\r\n\t# print(\"Third symmetry axis = \", Angle[i3]*2)\r\n\r\n\r\n\t# displayimages= np.array(displayimages)\r\n\t# maxi =np.amax(displayimages)\r\n\t# # print(maxi)\r\n\t# displayimages= np.divide(displayimages,maxi)\r\n\t# cv2.namedWindow('image', cv2.WINDOW_NORMAL)\r\n\t# cv2.imshow('image', displayimages)\r\n\t# cv2.waitKey(0)\r\n\r\n\t# return(Angle[i1]*5, Angle[i2]*5)\r\n\r\n\twdk=[0 for i in range(len(Angle))]\r\n\tcounter=0\r\n\tfor var in Angle:\r\n\t\twk=[[0 for i in range(128)] for j in range(128)]\r\n\t\ttheta = 5*var\r\n\t\ttheta = 180 - theta\r\n\t\tfor i in range(128):\r\n\t\t\tfor j in range(128):\r\n\t\t\t\tp=i-comy\r\n\t\t\t\tq=j-comx\r\n\t\t\t\tradiantheta = theta*mt.pi/180\r\n\t\t\t\tnewx= 2*(comx+p*mt.sin(radiantheta)*mt.cos(radiantheta) + q*((mt.cos(radiantheta))**2)) - j\r\n\t\t\t\tnewy = 2*(comy+q*mt.sin(radiantheta)*mt.cos(radiantheta) + p*((mt.sin(radiantheta))**2)) - i\r\n\r\n\t\t\t\t# newx = (mt.cos(2*radiantheta)*i + mt.sin(2*radiantheta)*j) \r\n\t\t\t\t# newy = (mt.sin(2*radiantheta)*i - mt.cos(2*radiantheta)*j) \r\n\t\t\t\tnewx=int(newx)\r\n\t\t\t\tnewy=int(newy)\r\n\t\t\t\tif(newy>=0 and newy<128 and newx>=0 and newx<128):\r\n\t\t\t\t\twk[i][j]= filteredimages[i][j]*filteredimages[newy][newx]\r\n\r\n\t\twk = np.array(wk)\r\n\t\twk_disp = np.array(wk)\r\n\t\tmaxi =np.amax(wk_disp)\r\n\t\t# print(maxi)\r\n\t\twk_disp= np.divide(wk_disp,maxi)\r\n\t\tcv2.namedWindow('mirror', cv2.WINDOW_NORMAL)\r\n\t\tprint(theta)\r\n\t\tcv2.imshow('mirror', wk_disp)\r\n\t\tcv2.waitKey(0)\r\n\r\n\r\n\t\tdenom = 0\r\n\r\n\t\twk = cv2.Sobel(wk, cv2.CV_64F, 0, 1, ksize=3)\r\n\t\twk = np.abs(wk)\r\n\t\tfor y in range(128):\r\n\t\t\tfor x in range(128):\r\n\t\t\t\twkxy = wk[y][x]\r\n\t\t\t\tp = x - comx\r\n\t\t\t\tq = y - comy\r\n\t\t\t\tDist = np.abs((q-(p*mt.tan(radiantheta)))*mt.cos(radiantheta))\r\n\t\t\t\twdk[counter]+= Dist*wkxy\r\n\t\t\t\t# wdk[counter]+= wkxy*(y-x*mt.tan(radiantheta)+comx*mt.tan(radiantheta)-comy)/(1+(mt.tan(radiantheta))**2)\r\n\t\t\t\tdenom+=wkxy\r\n\t\twdk[counter]=wdk[counter]/(denom)\r\n\t\t# print(Angle[counter]*5, wdk[counter])\r\n\t\tcounter+=1\r\n\r\n\t# print(np.array(Angle)*5)\r\n\t# print(wdk)\r\n\r\n\r\n\r\n\r\n\t# wdk=[0 for i in range(len(maxima))]\r\n\t# counter=0\r\n\t# for var in maxima:\r\n\t# \ttheta=5*var\r\n\t# \tdenom=0\r\n\t# \tfor i in range(128):\r\n\t# \t\tfor j in range(128):\r\n\t# \t\t\twkxy = wk[i][j]\r\n\t# \t\t\twdk[counter]+= wkxy*(i-j*mt.tan(theta)+comx*mt.tan(theta)-comy)/(1+(mt.tan(theta))**2)\r\n\t# \t\t\tdenom+=wkxy\r\n\t# \twdk[counter]=wdk[counter]/denom\r\n\t# \tcounter+=1\r\n\r\n\t# print(\"hi\")\r\n\t# print(wdk)\r\n\r\n\r\n\tdisplayimages= np.array(filteredimages)\r\n\tmaxi =np.amax(filteredimages)\r\n\t# print(maxi)\r\n\tdisplayimages= np.divide(filteredimages,maxi)\r\n\tcv2.namedWindow('filtimage', cv2.WINDOW_NORMAL)\r\n\tcv2.imshow('filtimage', displayimages)\r\n\tcv2.waitKey(0)\r\n\r\n\tmin_dist = 100000\r\n\tmin_idx = -1\r\n\r\n\tprint(np.array(Angle) * 5)\r\n\tprint(wdk)\r\n\r\n\tfor i in range(len(wdk)):\r\n\t\tif (wdk[i] < min_dist):\r\n\t\t\tmin_dist = wdk[i]\r\n\t\t\tmin_idx = i\r\n\r\n\tprint(\"dist: \", min_dist, Angle[min_idx]*5)\r\n\treturn(Angle[min_idx]*5)\r\n\r\n\t\r\n", "sub_path": "VisionPickPlaceDemo/symmetry.py", "file_name": "symmetry.py", "file_ext": "py", "file_size_in_byte": 5997, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "scipy.ndimage.filters.median_filter", "line_number": 18, "usage_type": "call"}, {"api_name": "scipy.ndimage.filters", "line_number": 18, "usage_type": "attribute"}, {"api_name": "scipy.ndimage", "line_number": 18, "usage_type": "name"}, {"api_name": "time.time", "line_number": 25, "usage_type": "call"}, {"api_name": "scipy.ndimage.distance_transform_edt", "line_number": 26, "usage_type": "call"}, {"api_name": "scipy.ndimage", "line_number": 26, "usage_type": "name"}, {"api_name": "numpy.logical_not", "line_number": 26, "usage_type": "call"}, {"api_name": "time.time", "line_number": 27, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 42, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 57, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 58, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 70, "usage_type": "attribute"}, {"api_name": "math.sin", "line_number": 71, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 71, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 72, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 72, "usage_type": "call"}, {"api_name": "peakutils.indexes", "line_number": 97, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 151, "usage_type": "attribute"}, {"api_name": "math.sin", "line_number": 152, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 152, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 153, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 166, "usage_type": "call"}, {"api_name": "cv2.namedWindow", "line_number": 167, "usage_type": "call"}, {"api_name": "cv2.WINDOW_NORMAL", "line_number": 167, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 169, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 170, "usage_type": "call"}, {"api_name": "cv2.Sobel", "line_number": 175, "usage_type": "call"}, {"api_name": "cv2.CV_64F", "line_number": 175, "usage_type": "attribute"}, {"api_name": "numpy.abs", "line_number": 176, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 182, "usage_type": "call"}, {"api_name": "math.tan", "line_number": 182, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 182, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 214, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 216, "usage_type": "call"}, {"api_name": "cv2.namedWindow", "line_number": 217, "usage_type": "call"}, {"api_name": "cv2.WINDOW_NORMAL", "line_number": 217, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 218, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 219, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 224, "usage_type": "call"}]} +{"seq_id": "378034539", "text": "from numpy.random import seed\nseed(5)\nfrom tensorflow import set_random_seed\nset_random_seed(8)\n\n\nimport tensorflow as tf\nfrom keras.backend.tensorflow_backend import set_session\n#config = tf.ConfigProto()\n#config.gpu_options.per_process_gpu_memory_fraction = 1\n#set_session(tf.Session(config=config))\n\nconfig = tf.ConfigProto()\nconfig.gpu_options.allow_growth = True\nset_session(tf.Session(config=config))\n#import os\n#os.environ[\"CUDA_DEVICE_ORDER\"] = \"PCI_BUS_ID\" # see issue #152\n#os.environ[\"CUDA_VISIBLE_DEVICES\"] = \"\"\nimport os\nos.environ['TF_CPP_MIN_LOG_LEVEL'] = '2'\n\nfrom model import *\nfrom data import saveResult\nimport os\nfrom os import listdir\nfrom os.path import isfile, join\nfrom PIL import Image\nimport keras as k\nfrom skimage import transform\nfrom loadImages import *\nimport operator\nfrom ctImageClass import *\n#os.environ[\"CUDA_VISIBLE_DEVICES\"] = \"0\"\n\n\n#data_gen_args = dict(rotation_range=0.2,\n# width_shift_range=0.05,\n# height_shift_range=0.05,\n# shear_range=0.05,\n# zoom_range=0.05,\n# horizontal_flip=True,\n# fill_mode='nearest')\n\n\n\n#načtení všech obrázků ve složce\npathToImages = ['Q:\\Matula\\dataproKNS']\n \ndoYouWantToResizeImages = True\nnormalize = False\nres = (400, 400)\n\n\nimagesInArray = loadImages(pathToImages[0], normalize=normalize, imload=True, resize = doYouWantToResizeImages, size =res, train = True)\nmasksInArray = loadImages(pathToImages[0], normalize=normalize, maskload=True, resize = doYouWantToResizeImages, size =res, train = True)\nif len(pathToImages)>1:\n for path in pathToImages[1:]:\n imagesInArray = np.concatenate((imagesInArray ,loadImages(path, normalize=normalize, imload=True, resize = doYouWantToResizeImages, size=res, train = True)), axis = 0)\n masksInArray = np.concatenate((masksInArray, loadImages(path, normalize=normalize, maskload=True, resize = doYouWantToResizeImages, size=res, train = True)), axis = 0)\n \n\n \nwhitePixelCount = []\nfor i in range(0, len(masksInArray)):\n whitePixelCount.append(masksInArray[i].sum())\n#\nwhitePixelCount = sorted(enumerate(whitePixelCount), key=operator.itemgetter(1))\n#\nn = 2500\nmaxIndeces = [i[0] for i in whitePixelCount[len(whitePixelCount)-n-1: len(whitePixelCount)-1]]\npretrainMasks = []\npretrainIms = []\npretrainMasks = masksInArray[maxIndeces]\npretrainIms = imagesInArray[maxIndeces]\n\npretrainIms = np.expand_dims(pretrainIms, axis =3)\npretrainMasks = np.expand_dims(pretrainMasks, axis = 3)\n\n \n \nmasksInArray = np.expand_dims(masksInArray, axis =3)\nimagesInArray = np.expand_dims(imagesInArray, axis = 3)\n#myGene = trainGenerator(2,'data/membrane/train','image','label',data_gen_args,save_to_dir = None)\n#\nmodel = unet()\n#model_checkpoint = ModelCheckpoint('model_cartilage.hdf5', monitor='loss',verbose=1, save_best_only=True)\n#model.load_weights('weights_wo_green.h5')\n#model.fit_generator(myGene,steps_per_epoch=300,epochs=1,callbacks=[model_checkpoint])\n#tbCallBack = k.callbacks.TensorBoard(log_dir='./Graph', histogram_freq=0, write_graph=True, write_images=True)\n\n#from ctImageClass import *\n#\n#newmodel = imageDataGenerator(model=model, batchSize = 1)\n#newmodel.getPaths(pathToImages)\n#\n#newmodel.trainModel()\n\nmodel.fit(imagesInArray, masksInArray, epochs = 100, batch_size = 4)\n#model.save('pretrain.h5')\nmodel.save('jiriho_metaloartefakty.h5')\n\n#model = load_model('model_smarter_250e.h5')\npathToTestImages = ['Q:\\Matula\\pro_testy']\n\ntestImagesInArray = loadImages(pathToTestImages[0], normalize=normalize, imload=True, resize = doYouWantToResizeImages, size =res, test = True)\ntestMasksInArray = loadImages(pathToTestImages[0], normalize=normalize, maskload=True, resize = doYouWantToResizeImages, size =res, test = True)\nif len(pathToTestImages)>1:\n for path in pathToTestImages[1:]:\n testImagesInArray = np.concatenate((testImagesInArray ,loadImages(path, normalize=normalize, imload=True, resize = doYouWantToResizeImages, size=res, test = True)), axis = 0)\n testMasksInArray = np.concatenate((testMasksInArray, loadImages(path, normalize=normalize, maskload=True, resize = doYouWantToResizeImages, size=res, test = True)), axis = 0)\n\ntestMasksInArray = np.expand_dims(testMasksInArray, axis =3)\ntestImagesInArray = np.expand_dims(testImagesInArray, axis = 3)\n\nresults = model.predict(testImagesInArray,1,verbose=1)\nsaveResult(\"U:\\\\Skripty\\\\unet_code\\\\data\\\\membrane\\\\test_jiri3\",results)\n\n\n\n", "sub_path": "unet_code/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 4512, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "numpy.random.seed", "line_number": 2, "usage_type": "call"}, {"api_name": "tensorflow.set_random_seed", "line_number": 4, "usage_type": "call"}, {"api_name": "tensorflow.ConfigProto", "line_number": 13, "usage_type": "call"}, {"api_name": "keras.backend.tensorflow_backend.set_session", "line_number": 15, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 15, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 20, "usage_type": "attribute"}, {"api_name": "operator.itemgetter", "line_number": 67, "usage_type": "call"}, {"api_name": "model.fit", "line_number": 98, "usage_type": "call"}, {"api_name": "model.save", "line_number": 100, "usage_type": "call"}, {"api_name": "model.predict", "line_number": 115, "usage_type": "call"}, {"api_name": "data.saveResult", "line_number": 116, "usage_type": "call"}]} +{"seq_id": "487760786", "text": "from .producer import produceData\nfrom rest_framework.views import APIView\nfrom api.serializers import ProductSerializer\nfrom api.models import Product, User\nfrom rest_framework import viewsets\nfrom rest_framework.response import Response\nfrom rest_framework import status\nimport random\n# Create your views here.\n\nclass ProductViewSet(viewsets.ViewSet):\n def list(self,request):\n products= Product.objects.all()\n serializer=ProductSerializer(products,many=True)\n return Response(serializer.data, status=status.HTTP_200_OK)\n \n def create(self,request):\n serializer=ProductSerializer(data=request.data)\n serializer.is_valid(raise_exception=True)\n serializer.save()\n #producer\n produceData('product_created', serializer.data)\n return Response(serializer.data, status=status.HTTP_201_CREATED)\n \n def retrieve(self,request,pk=None):\n product= Product.objects.get(id=pk)\n serializer=ProductSerializer(product)\n return Response(serializer.data, status=status.HTTP_302_FOUND)\n \n def update(self,request,pk=None):\n product= Product.objects.get(id=pk)\n serializer=ProductSerializer(data=request.data, instance=product)\n serializer.is_valid(raise_exception=True)\n serializer.save()\n #producer\n produceData('product_updated',serializer.data)\n return Response(serializer.data, status=status.HTTP_202_ACCEPTED)\n\n def destroy(self,request,pk=None):\n product= Product.objects.get(id=pk)\n product.delete()\n #producer\n produceData('product_deleted',pk)\n return Response(status=status.HTTP_204_NO_CONTENT)\n\n\nclass UserApiView(APIView):\n def get(self,_):\n users= User.objects.all()\n user=random.choice(users)\n return Response({\n 'id':user.id\n })", "sub_path": "productapp/api/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1867, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "rest_framework.viewsets.ViewSet", "line_number": 11, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 11, "usage_type": "name"}, {"api_name": "api.models.Product.objects.all", "line_number": 13, "usage_type": "call"}, {"api_name": "api.models.Product.objects", "line_number": 13, "usage_type": "attribute"}, {"api_name": "api.models.Product", "line_number": 13, "usage_type": "name"}, {"api_name": "api.serializers.ProductSerializer", "line_number": 14, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 15, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 15, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 15, "usage_type": "name"}, {"api_name": "api.serializers.ProductSerializer", "line_number": 18, "usage_type": "call"}, {"api_name": "producer.produceData", "line_number": 22, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 23, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 23, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 23, "usage_type": "name"}, {"api_name": "api.models.Product.objects.get", "line_number": 26, "usage_type": "call"}, {"api_name": "api.models.Product.objects", "line_number": 26, "usage_type": "attribute"}, {"api_name": "api.models.Product", "line_number": 26, "usage_type": "name"}, {"api_name": "api.serializers.ProductSerializer", "line_number": 27, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 28, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_302_FOUND", "line_number": 28, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 28, "usage_type": "name"}, {"api_name": "api.models.Product.objects.get", "line_number": 31, "usage_type": "call"}, {"api_name": "api.models.Product.objects", "line_number": 31, "usage_type": "attribute"}, {"api_name": "api.models.Product", "line_number": 31, "usage_type": "name"}, {"api_name": "api.serializers.ProductSerializer", "line_number": 32, "usage_type": "call"}, {"api_name": "producer.produceData", "line_number": 36, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 37, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_202_ACCEPTED", "line_number": 37, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 37, "usage_type": "name"}, {"api_name": "api.models.Product.objects.get", "line_number": 40, "usage_type": "call"}, {"api_name": "api.models.Product.objects", "line_number": 40, "usage_type": "attribute"}, {"api_name": "api.models.Product", "line_number": 40, "usage_type": "name"}, {"api_name": "producer.produceData", "line_number": 43, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 44, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_204_NO_CONTENT", "line_number": 44, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 44, "usage_type": "name"}, {"api_name": "rest_framework.views.APIView", "line_number": 47, "usage_type": "name"}, {"api_name": "api.models.User.objects.all", "line_number": 49, "usage_type": "call"}, {"api_name": "api.models.User.objects", "line_number": 49, "usage_type": "attribute"}, {"api_name": "api.models.User", "line_number": 49, "usage_type": "name"}, {"api_name": "random.choice", "line_number": 50, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 51, "usage_type": "call"}]} +{"seq_id": "616896778", "text": "from flask_swagger_ui import get_swaggerui_blueprint\nfrom flask import Flask\nfrom sqlalchemy import create_engine, MetaData\n\napp = Flask(__name__, instance_relative_config=True)\napp.config.from_pyfile('config.py')\n\n# swagger specific #\nSWAGGER_URL = '/swagger'\nAPI_URL = '/static/swagger.yaml'\nSWAGGERUI_BLUEPRINT = get_swaggerui_blueprint(\n SWAGGER_URL,\n API_URL,\n config={\n 'app_name': \"GENE-SEARCH-REST-API\"\n }\n)\napp.register_blueprint(SWAGGERUI_BLUEPRINT, url_prefix=SWAGGER_URL)\n# end swagger specific #\nconnection = \"mysql://{user}@{host}:{port}/{database}\".format(user=app.config['USER'], host=app.config['HOST'],\n port=app.config['PORT'], database=app.config['DATABASE'])\nengine = create_engine(connection)\nmetadata = MetaData(bind=engine)\nfrom app import views\n", "sub_path": "app/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 850, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "flask.Flask", "line_number": 5, "usage_type": "call"}, {"api_name": "flask_swagger_ui.get_swaggerui_blueprint", "line_number": 11, "usage_type": "call"}, {"api_name": "sqlalchemy.create_engine", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.MetaData", "line_number": 23, "usage_type": "call"}]} +{"seq_id": "343842043", "text": "from argparse import ArgumentParser\nfrom collections import OrderedDict\nfrom copy import deepcopy\nfrom pprint import pformat\nfrom string import ascii_letters\nfrom typing import Any, Callable, Dict, List, Optional, Sequence, Set, Union\nimport json\nimport sys\nimport time\n\nfrom tap.utils import get_class_variables, get_dest, get_git_root, get_git_url, has_git,has_uncommitted_changes,\\\n is_option_arg, type_to_str\n\n\nSUPPORTED_DEFAULT_BASE_TYPES = {str, int, float, bool}\nSUPPORTED_DEFAULT_OPTIONAL_TYPES = {Optional[str], Optional[int], Optional[float]}\nSUPPORTED_DEFAULT_LIST_TYPES = {List[str], List[int], List[float]}\nSUPPORTED_DEFAULT_SET_TYPES = {Set[str], Set[int], Set[float]}\nSUPPORTED_DEFAULT_COLLECTION_TYPES = SUPPORTED_DEFAULT_LIST_TYPES | SUPPORTED_DEFAULT_SET_TYPES\nSUPPORTED_DEFAULT_TYPES = set.union(SUPPORTED_DEFAULT_BASE_TYPES,\n SUPPORTED_DEFAULT_OPTIONAL_TYPES,\n SUPPORTED_DEFAULT_COLLECTION_TYPES)\n\n\nclass Tap(ArgumentParser):\n \"\"\"Tap is a typed argument parser that wraps Python's built-in ArgumentParser.\"\"\"\n\n def __init__(self, *args, **kwargs):\n \"\"\"Initializes the Tap instance.\n\n :param args: Arguments passed to the super class ArgumentParser.\n :param kwargs: Keyword arguments passed to the super class ArgumentParser.\n \"\"\"\n # Whether the arguments have been parsed (i.e. if parse_args has been called)\n self._parsed = False\n\n # Set extra arguments to empty list\n self.extra_args = []\n\n # Create argument buffer\n self.argument_buffer = OrderedDict()\n\n # Get class variables help strings from the comments\n self.class_variables = self._get_class_variables()\n\n # Get annotations from self and all super classes up through tap\n self._annotations = self._get_annotations()\n\n # Initialize the super class, i.e. ArgumentParser\n super(Tap, self).__init__(*args, **kwargs)\n\n # Add arguments to self\n self.add_arguments() # Adds user-overridden arguments to the arguments buffer\n self._add_arguments() # Adds all arguments in order to self\n\n def _add_argument(self, *name_or_flags, **kwargs) -> None:\n \"\"\"Adds an argument to self (i.e. the super class ArgumentParser).\n\n Sets the following attributes of kwargs when not explicitly provided:\n - type: Set to the type annotation of the argument.\n - default: Set to the default value of the argument (if provided).\n - required: True if a default value of the argument is not provided, False otherwise.\n - action: Set to \"store_true\" if the argument is a required bool or a bool with default value False.\n Set to \"store_false\" if the argument is a bool with default value True.\n - nargs: Set to \"*\" if the type annotation is List[str], List[int], or List[float].\n - help: Set to the argument documentation from the class docstring.\n\n :param name_or_flags: Either a name or a list of option strings, e.g. foo or -f, --foo.\n :param kwargs: Keyword arguments.\n \"\"\"\n # Get variable name\n variable = get_dest(*name_or_flags, **kwargs)\n\n # Get default if not specified\n if hasattr(self, variable):\n kwargs['default'] = kwargs.get('default', getattr(self, variable))\n\n # Set required if option arg\n if is_option_arg(*name_or_flags) and variable != 'help':\n kwargs['required'] = kwargs.get('required', not hasattr(self, variable))\n\n\n if len(name_or_flags) == 1 and name_or_flags[0][:2] == \"--\":\n # expand attributes such that (\"--attribute\",) becomes (\"--attribute\", \"-a\")\n cvar_names = {n: None for n in self.class_variables}\n for long_name in self.class_variables:\n preferred_shorts = []\n for char in long_name:\n # collect eligible characters from long_name as preferred options\n if (\n char in ascii_letters\n and char not in preferred_shorts\n and not any([char == v for v in cvar_names.values() if v])\n ):\n preferred_shorts += char\n other_shorts = [\n asc\n for asc in ascii_letters\n if asc not in long_name and asc not in preferred_shorts\n ]\n for char in preferred_shorts + other_shorts:\n if char == \"h\":\n # avoiding \"h\" because it overlaps with default behavior of ArgumentParser.add_help\n continue\n if not any([char == v for v in set(cvar_names.values()) if v]):\n short_name = char\n break\n cvar_names[long_name] = short_name\n if cvar_names and name_or_flags[0][2:] in cvar_names:\n name_or_flags = (f\"{name_or_flags[0]}\", f\"-{cvar_names[name_or_flags[0][2:]]}\")\n\n # Set help if necessary\n if 'help' not in kwargs:\n kwargs['help'] = '('\n\n # Type\n if variable in self._annotations:\n kwargs['help'] += type_to_str(self._annotations[variable]) + ', '\n\n # Required/default\n if kwargs.get('required', False):\n kwargs['help'] += 'required'\n else:\n kwargs['help'] += f'default={kwargs.get(\"default\", None)}'\n\n kwargs['help'] += ')'\n\n # Description\n if variable in self.class_variables:\n kwargs['help'] += ' ' + self.class_variables[variable]['comment']\n\n # Set other kwargs where not provided\n if variable in self._annotations:\n # Get type annotation\n var_type = self._annotations[variable]\n\n # If type is not explicitly provided, set it if it's one of our supported default types\n if 'type' not in kwargs:\n if var_type not in SUPPORTED_DEFAULT_TYPES:\n raise ValueError(\n f'Variable \"{variable}\" has type \"{var_type}\" which is not supported by default.\\n'\n f'Please explicitly add the argument to the parser by writing:\\n\\n'\n f'def add_arguments(self) -> None:\\n'\n f' self.add_argument(\"--{variable}\", type=func, {\"required=True\" if kwargs[\"required\"] else f\"default={getattr(self, variable)}\"})\\n\\n'\n f'where \"func\" maps from str to {var_type}.')\n\n # If Optional type, extract type\n if var_type in SUPPORTED_DEFAULT_OPTIONAL_TYPES:\n var_type = var_type.__args__[0]\n\n # If List type, extract type of elements in list and set nargs\n elif var_type in SUPPORTED_DEFAULT_COLLECTION_TYPES:\n var_type = var_type.__args__[0]\n kwargs['nargs'] = kwargs.get('nargs', '*')\n\n # If bool then set action, otherwise set type\n if var_type == bool:\n kwargs['action'] = kwargs.get('action', f'store_{\"true\" if kwargs[\"required\"] or not kwargs[\"default\"] else \"false\"}')\n else:\n kwargs['type'] = var_type\n\n super(Tap, self).add_argument(*name_or_flags, **kwargs)\n\n def add_argument(self, *name_or_flags, **kwargs) -> None:\n \"\"\"Adds an argument to the argument buffer, which will later be passed to _add_argument.\"\"\"\n variable = get_dest(*name_or_flags, **kwargs)\n self.argument_buffer[variable] = (name_or_flags, kwargs)\n\n def _add_arguments(self) -> None:\n \"\"\"Add arguments to self in the order they are defined as class variables (so the help string is in order).\"\"\"\n # Add class variables (in order)\n for variable in self.class_variables:\n if variable in self.argument_buffer:\n name_or_flags, kwargs = self.argument_buffer[variable]\n self._add_argument(*name_or_flags, **kwargs)\n else:\n self._add_argument(f'--{variable}')\n\n # Add any arguments that were added manually in add_arguments but aren't class variables (in order)\n for variable, (name_or_flags, kwargs) in self.argument_buffer.items():\n if variable not in self.class_variables:\n self._add_argument(*name_or_flags, **kwargs)\n\n def add_arguments(self) -> None:\n \"\"\"Explicitly add arguments to the argument buffer if not using default settings.\"\"\"\n pass\n\n def process_args(self) -> None:\n \"\"\"Perform additional argument processing and/or validation.\"\"\"\n pass\n\n @staticmethod\n def get_reproducibility_info() -> Dict[str, str]:\n \"\"\"Gets a dictionary of reproducibility information.\n\n Reproducibility information always includes:\n - command_line: The command line command used to execute the code.\n - time: The current time.\n\n If git is installed, reproducibility information also includes:\n - git_root: The root of the git repo where the command is run.\n - git_url: The url of the current hash of the git repo where the command is run.\n Ex. https://github.com/swansonk14/rationale-alignment/tree/\n - git_has_uncommitted_changes: Whether the current git repo has uncommitted changes.\n\n :return: A dictionary of reproducibility information.\n \"\"\"\n reproducibility = {\n 'command_line': f'python {\" \".join(sys.argv)}',\n 'time': time.strftime('%c')\n }\n\n if has_git():\n reproducibility['git_root'] = get_git_root()\n reproducibility['git_url'] = get_git_url(commit_hash=True)\n reproducibility['git_has_uncommitted_changes'] = has_uncommitted_changes()\n\n return reproducibility\n\n def _log_all(self) -> Dict[str, Any]:\n \"\"\"Gets all arguments along with reproducibility information.\n\n :return: A dictionary containing all arguments along with reproducibility information.\n \"\"\"\n arg_log = self.as_dict()\n arg_log['reproducibility'] = self.get_reproducibility_info()\n\n return arg_log\n\n def parse_args(self,\n args: Optional[Sequence[str]] = None,\n known_only: bool = False) -> 'Tap':\n \"\"\"Parses arguments, sets attributes of self equal to the parsed arguments, and processes arguments.\n\n :param args: List of strings to parse. The default is taken from `sys.argv`.\n :param known_only: If true, ignores extra arguments and only parses known arguments.\n Unparsed arguments are saved to self.extra_args.\n :return: self, which is a Tap instance containing all of the parsed args.\n \"\"\"\n # Parse args using super class ArgumentParser's parse_args or parse_known_args function\n if known_only:\n default_namespace, self.extra_args = super(Tap, self).parse_known_args(args)\n else:\n default_namespace = super(Tap, self).parse_args(args)\n\n # Copy parsed arguments to self\n for variable, value in vars(default_namespace).items():\n # Conversion from list to set\n if variable in self._annotations and self._annotations[variable] in SUPPORTED_DEFAULT_SET_TYPES:\n value = set(value)\n\n # Set variable in self (and deepcopy)\n setattr(self, variable, deepcopy(value))\n\n # Process args\n self.process_args()\n\n # Indicate that args have been parsed\n self._parsed = True\n\n return self\n\n @classmethod\n def _get_from_self_and_super(cls,\n extract_func: Callable[[type], dict],\n dict_type: type = dict) -> Union[Dict[str, Any], OrderedDict]:\n \"\"\"Returns a dictionary mapping variable names to values.\n\n Variables and values are extracted from classes using key starting\n with this class and traversing up the super classes up through Tap.\n\n If super class and sub class have the same key, the sub class value is used.\n\n Super classes are traversed through breadth first search.\n\n :param extract_func: A function that extracts from a class a dictionary mapping variables to values.\n :param dict_type: The type of dictionary to use (e.g. dict, OrderedDict, etc.)\n :return: A dictionary mapping variable names to values from the class dict.\n \"\"\"\n visited = set()\n super_classes = [cls]\n dictionary = dict_type()\n\n while len(super_classes) > 0:\n super_class = super_classes.pop(0)\n\n if super_class not in visited and issubclass(super_class, Tap):\n super_dictionary = extract_func(super_class)\n\n # Update only unseen variables to avoid overriding subclass values\n for variable, value in super_dictionary.items():\n if variable not in dictionary:\n dictionary[variable] = value\n for variable in super_dictionary.keys() - dictionary.keys():\n dictionary[variable] = super_dictionary[variable]\n\n super_classes += list(super_class.__bases__)\n visited.add(super_class)\n\n return dictionary\n\n def _get_class_dict(self) -> Dict[str, Any]:\n \"\"\"Returns a dictionary mapping class variable names to values from the class dict.\"\"\"\n class_dict = self._get_from_self_and_super(\n extract_func=lambda super_class: dict(getattr(super_class, '__dict__', dict()))\n )\n class_dict = {\n var: val\n for var, val in class_dict.items()\n if not (var.startswith('_') or callable(val) or isinstance(val, staticmethod))\n }\n\n return class_dict\n\n def _get_annotations(self) -> Dict[str, Any]:\n \"\"\"Returns a dictionary mapping variable names to their type annotations.\"\"\"\n return self._get_from_self_and_super(\n extract_func=lambda super_class: dict(getattr(super_class, '__annotations__', dict()))\n )\n\n def _get_class_variables(self) -> OrderedDict:\n \"\"\"Returns an OrderedDict mapping class variables names to their additional information.\"\"\"\n try:\n class_variables = self._get_from_self_and_super(\n extract_func=lambda super_class: get_class_variables(super_class),\n dict_type=OrderedDict\n )\n # Exception if inspect.getsource fails to extract the source code\n except Exception:\n class_variables = OrderedDict()\n for variable in self._get_class_dict().keys():\n class_variables[variable] = {'comment': ''}\n\n return class_variables\n\n def _get_argument_names(self) -> Set[str]:\n \"\"\"Returns a list of variable names corresponding to the arguments.\"\"\"\n return set(self._get_class_dict().keys()) | set(self._annotations.keys())\n\n def as_dict(self) -> Dict[str, Any]:\n \"\"\"Returns the member variables corresponding to the class variable arguments.\n\n :return: A dictionary mapping each argument's name to its value.\n \"\"\"\n if not self._parsed:\n raise ValueError('You should call `parse_args` before retrieving arguments.')\n\n return {var: getattr(self, var) for var in self._get_argument_names()}\n\n def save(self, path: str) -> None:\n \"\"\"Saves the arguments and reproducibility information in JSON format.\n\n :param path: Path to the JSON file where the arguments will be saved.\n \"\"\"\n with open(path, 'w') as f:\n json.dump(self._log_all(), f, indent=4, sort_keys=True)\n\n def __str__(self) -> str:\n \"\"\"Returns a string representation of self.\n\n :return: A formatted string representation of the instance's declaration\n \"\"\"\n keys = self._get_class_variables().keys()\n csv = \", \".join([f\"{k}={(vars(self)[k]).__repr__()}\" for k in keys])\n return f\"{self.__class__.__name__}({csv})\"\n", "sub_path": "tap/tap.py", "file_name": "tap.py", "file_ext": "py", "file_size_in_byte": 16228, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "typing.Optional", "line_number": 16, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 17, "usage_type": "name"}, {"api_name": "typing.Set", "line_number": 18, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 25, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 41, "usage_type": "call"}, {"api_name": "tap.utils.get_dest", "line_number": 72, "usage_type": "call"}, {"api_name": "tap.utils.is_option_arg", "line_number": 79, "usage_type": "call"}, {"api_name": "string.ascii_letters", "line_number": 91, "usage_type": "name"}, {"api_name": "string.ascii_letters", "line_number": 98, "usage_type": "name"}, {"api_name": "tap.utils.type_to_str", "line_number": 118, "usage_type": "call"}, {"api_name": "tap.utils.get_dest", "line_number": 166, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 209, "usage_type": "attribute"}, {"api_name": "time.strftime", "line_number": 210, "usage_type": "call"}, {"api_name": "tap.utils.has_git", "line_number": 213, "usage_type": "call"}, {"api_name": "tap.utils.get_git_root", "line_number": 214, "usage_type": "call"}, {"api_name": "tap.utils.get_git_url", "line_number": 215, "usage_type": "call"}, {"api_name": "tap.utils.has_uncommitted_changes", "line_number": 216, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 193, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 220, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 220, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 231, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 231, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 253, "usage_type": "call"}, {"api_name": "typing.Callable", "line_number": 265, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 266, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 266, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 266, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 266, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 302, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 302, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 315, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 315, "usage_type": "name"}, {"api_name": "tap.utils.get_class_variables", "line_number": 325, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 326, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 330, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 321, "usage_type": "name"}, {"api_name": "typing.Set", "line_number": 336, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 340, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 340, "usage_type": "name"}, {"api_name": "json.dump", "line_number": 356, "usage_type": "call"}]} +{"seq_id": "132057166", "text": "from airflow import DAG\nfrom datetime import datetime, timedelta\nfrom airflow.operators.python_operator import PythonOperator\n\nimport sys, os\nsys.path.append('../')\n\nimport data_query\nimport data_manipulation\n\ndefault_args = {\n 'owner': 'bi',\n 'depends_on_past': False,\n 'start_date': datetime(2019, 7, 16),\n 'retry_delay': timedelta(minutes=10)\n}\n\ndag = DAG(\n \"presto_data_test\",\n default_args=default_args,\n schedule_interval=\"@daily\"\n)\n\ndef query_data(**context):\n return data_query.return_df()\n\nquery_task = PythonOperator(\n task_id='query_task',\n python_callable=query_data,\n provide_context=True,\n dag=dag\n)\n\ndef show_data(**context):\n df = context[\"task_instance\"].xcom_pull(task_ids='query_task')\n data_manipulation.show_total_sum(df)\n\nprint_task = PythonOperator(\n task_id='print_task',\n python_callable=show_data,\n provide_context=True,\n dag=dag\n)\n\nquery_task >> print_task", "sub_path": "BACKUPS/python/airflow/airflow_presto_dag/dataframe_test.dag.py", "file_name": "dataframe_test.dag.py", "file_ext": "py", "file_size_in_byte": 937, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "sys.path.append", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 15, "usage_type": "call"}, {"api_name": "airflow.DAG", "line_number": 18, "usage_type": "call"}, {"api_name": "data_query.return_df", "line_number": 25, "usage_type": "call"}, {"api_name": "airflow.operators.python_operator.PythonOperator", "line_number": 27, "usage_type": "call"}, {"api_name": "data_manipulation.show_total_sum", "line_number": 36, "usage_type": "call"}, {"api_name": "airflow.operators.python_operator.PythonOperator", "line_number": 38, "usage_type": "call"}]} +{"seq_id": "24654673", "text": "#!/usr/bin/env python\n\"\"\"\nInit a diffscuss-mb code review directory.\n\"\"\"\n\nfrom optparse import OptionParser\nimport os\nfrom textwrap import dedent\n\nfrom common import get_git_root, DIFFSCUSS_MB_FILE_NAME, \\\n USERS_DIR_NAME, REVIEWS_DIR_NAME, mkdir_for_keeps\n\n\ndef main(opts, args):\n git_root = get_git_root(opts.git_exe)\n os.chdir(git_root)\n\n dmb_root_dir = args[0]\n mkdir_for_keeps(dmb_root_dir)\n\n with open(DIFFSCUSS_MB_FILE_NAME, 'wb') as fil:\n fil.write(dmb_root_dir)\n\n users_dir = os.path.join(dmb_root_dir, USERS_DIR_NAME)\n mkdir_for_keeps(users_dir)\n\n reviews_dir = os.path.join(dmb_root_dir, REVIEWS_DIR_NAME)\n mkdir_for_keeps(reviews_dir)\n\nif __name__ == '__main__':\n parser = OptionParser(usage=dedent(\"\"\"\\\n %prog [options] diffscuss-mailbox-dir\n\n Init a diffscuss mailbox system. Must be run from\n within a git checkout.\n\n diffscuss-mailbox-dir is relative to the git root,\n and will be created.\n \"\"\"))\n parser.add_option(\"-g\", \"--git-exe\", dest=\"git_exe\", default=\"git\",\n help=dedent(\"\"\"\\\n Git exe (defaults to 'git'.\n \"\"\"))\n\n (opts, args) = parser.parse_args()\n if len(args) != 1:\n parser.error(\"Must provide exactly one diffscuss mailbox directory.\")\n main(opts, args)\n\n", "sub_path": "diffscuss-mb/dmb-init.py", "file_name": "dmb-init.py", "file_ext": "py", "file_size_in_byte": 1559, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "common.get_git_root", "line_number": 15, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 16, "usage_type": "call"}, {"api_name": "common.mkdir_for_keeps", "line_number": 19, "usage_type": "call"}, {"api_name": "common.DIFFSCUSS_MB_FILE_NAME", "line_number": 21, "usage_type": "argument"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "common.USERS_DIR_NAME", "line_number": 24, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "common.mkdir_for_keeps", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "common.REVIEWS_DIR_NAME", "line_number": 27, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "common.mkdir_for_keeps", "line_number": 28, "usage_type": "call"}, {"api_name": "optparse.OptionParser", "line_number": 31, "usage_type": "call"}, {"api_name": "textwrap.dedent", "line_number": 31, "usage_type": "call"}, {"api_name": "textwrap.dedent", "line_number": 41, "usage_type": "call"}]} +{"seq_id": "453879195", "text": "#!/usr/bin/python\n#求0—7所能组成的奇数个数\n\n#计算出组成数字的个数,再用set去重,再用x%2==1过滤出奇数\n\ns= [i for i in '01234567']\nimport itertools\narr = [] #数字总个数\nfor i in range(1,9):\n a = list(itertools.permutations(s,i))\n l = list(map(lambda x:int(''.join(x)),a))\n arr+=l\n print(i,len(l)) #输出组成的数字,不同位数的都有多少\narr1 = set(arr) #去重\narr2 = list(filter(lambda x:x%2==1,arr1)) #过滤奇数\nprint(len(arr),len(arr1),len(arr2))\n", "sub_path": "jishuzuhe.py", "file_name": "jishuzuhe.py", "file_ext": "py", "file_size_in_byte": 518, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "itertools.permutations", "line_number": 10, "usage_type": "call"}]} +{"seq_id": "366771814", "text": "#############################################################################\r\n# #\r\n# Name: #\r\n# jointDuplicator.py #\r\n# #\r\n# Desc: #\r\n# Creates duplicate sets of joints with an attribute to control #\r\n# the blend. #\r\n# #\r\n# Author: #\r\n# Matt Jenkins #\r\n# #\r\n#############################################################################\r\n\r\n\r\nimport jointDuplicatorUI as customUI\r\nimport maya.OpenMayaUI as omUi\r\nimport maya.cmds as mc\r\nfrom functools import partial\r\n\r\ntry:\r\n from PySide import QtGui, QtCore\r\n import PySide.QtGui as QtWidgets\r\n from shiboken import wrapInstance\r\nexcept ImportError:\r\n from PySide2 import QtGui, QtCore, QtWidgets\r\n from shiboken2 import wrapInstance\r\n\r\n\r\ndef maya_main_window():\r\n main_window_ptr = omUi.MQtUtil.mainWindow()\r\n return wrapInstance(long(main_window_ptr), QtWidgets.QWidget)\r\n\r\ndef main():\r\n global myWindow\r\n global ctrlCrvBoxValue\r\n global suffixABoxValue\r\n global suffixBBoxValue\r\n global attrNameBoxValue\r\n global attrNiceNameBoxValue\r\n global curSuffixBoxValue\r\n global heirachyMode\r\n ctrlCrvBoxValue = None\r\n suffixABoxValue = 'IK'\r\n suffixBBoxValue = 'FK'\r\n attrNameBoxValue = 'IKFKSwitch'\r\n attrNiceNameBoxValue = 'IK / FK Switch'\r\n curSuffixBoxValue = 'result'\r\n heirachyMode = False\r\n\r\n try:\r\n myWindow.close()\r\n except: pass\r\n myWindow = myTool(parent=maya_main_window())\r\n myWindow.show()\r\n\r\n\r\nclass myTool(QtWidgets.QDialog):\r\n def __init__(self,parent = None):\r\n\r\n reload(customUI)\r\n print(\"loaded\")\r\n\r\n super(myTool, self).__init__(parent)\r\n self.setWindowFlags(QtCore.Qt.Tool)\r\n self.ui = customUI.Ui_MainWindow() #Define UI class in module\r\n\r\n self.ui.setupUi(self) # start window\r\n\r\n self.ui.ctrlCrvBox.textChanged.connect(self.updateCtrlCrvBox)\r\n self.ui.setCtrlCrvBtn.released.connect(self.updateCtrlCrvBtn)\r\n self.ui.createBtn.released.connect(self.updateCreateBtn)\r\n self.ui.modeSelRad.clicked.connect(partial(self.updateModeRad, 'modeSel'))\r\n self.ui.modeHeirRad.clicked.connect(partial(self.updateModeRad, 'modeHeir'))\r\n self.ui.suffixABox.textChanged.connect(self.updateSuffixABox)\r\n self.ui.suffixBBox.textChanged.connect(self.updateSuffixBBox)\r\n self.ui.attrNameBox.textChanged.connect(self.updateAttrNameBox)\r\n self.ui.attrNiceNameBox.textChanged.connect(self.updateAttrNiceNameBox)\r\n self.ui.curSuffixBox.textChanged.connect(self.updateCurSuffixBox)\r\n\r\n def updateSuffixABox(self, *args):\r\n global suffixABoxValue\r\n suffixABoxValue = self.ui.suffixABox.text()\r\n\r\n def updateSuffixBBox(self, *args):\r\n global suffixBBoxValue\r\n suffixBBoxValue = self.ui.suffixBBox.text()\r\n\r\n def updateAttrNameBox(self, *args):\r\n global attrNameBoxValue\r\n attrNameBoxValue = self.ui.attrNameBox.text()\r\n\r\n def updateAttrNiceNameBox(self, *args):\r\n global attrNiceNameBoxValue\r\n attrNiceNameBoxValue = self.ui.attrNiceNameBox.text()\r\n\r\n def updateCurSuffixBox(self, *args):\r\n global curSuffixBoxValue\r\n curSuffixBoxValue = self.ui.curSuffixBox.text()\r\n\r\n def updateCtrlCrvBox(self, *args):\r\n global ctrlCrvBoxValue\r\n ctrlCrvBoxValue = self.ui.ctrlCrvBox.text()\r\n if mc.objExists(ctrlCrvBoxValue):\r\n self.ui.ctrlCrvBox.setStyleSheet(\"background-color: rgb(90, 150, 50);\")\r\n else:\r\n self.ui.ctrlCrvBox.setStyleSheet(\"background-color: rgb(110, 90, 90);\")\r\n\r\n def updateCtrlCrvBtn(self, *args):\r\n crvSelection = mc.ls(sl=True)\r\n crvSelection = str(crvSelection)\r\n crvSelection = crvSelection.replace(\"[u'\",'')\r\n crvSelection = crvSelection.replace(\"']\",'')\r\n crvSelection = crvSelection.replace(\"[\",'')\r\n crvSelection = crvSelection.replace(\"]\",'')\r\n self.ui.ctrlCrvBox.setText(crvSelection)\r\n\r\n def updateModeRad(self, value, *args):\r\n global heirachyMode\r\n if value == 'modeSel':\r\n heirachyMode = False\r\n elif value == 'modeHeir':\r\n heirachyMode = True\r\n\r\n def updateCreateBtn(self, *args):\r\n self.createAttr()\r\n\r\n if heirachyMode:\r\n #-- Joint Hierachy Mode\r\n selJoints = mc.ls(sl=True, type='joint')\r\n\r\n if not selJoints:\r\n mc.confirmDialog(title='ERROR', message='Select valid Joint(s)', button=['Ok'])\r\n return\r\n\r\n for curjoint in selJoints:\r\n previousJoint=None\r\n jointTree = mc.listRelatives(curjoint, s=False, typ='joint', ad=True)\r\n jointTree.append(curjoint)\r\n jointTree.reverse()\r\n self.createJoints(jointTree)\r\n\r\n else:\r\n #-- Selected Joints Mode\r\n previousJoint=None\r\n selJoints = mc.ls(sl=True, type='joint')\r\n if not selJoints:\r\n mc.confirmDialog(title='ERROR', message='Select valid Joint(s)', button=['Ok'])\r\n return\r\n self.createJoints(selJoints)\r\n\r\n def createJoints(self, joints):\r\n #-- Set Suffix Values\r\n if not suffixABoxValue or not suffixBBoxValue:\r\n mc.confirmDialog(title='ERROR', message='Enter a valid suffix value', button=['Ok'])\r\n return\r\n suffixA = '_'+suffixABoxValue+'_'\r\n suffixB = '_'+suffixBBoxValue+'_'\r\n if not curSuffixBoxValue:\r\n mc.confirmDialog(title='ERROR', message='Enter a valid current suffix value', button=['Ok'])\r\n return\r\n curSuffix = '_'+curSuffixBoxValue+'_'\r\n\r\n #-- Main Code\r\n\r\n for x in joints:\r\n if curSuffix not in x:\r\n if '_JNT' not in x:\r\n mc.confirmDialog(title='ERROR', message='One or more incorrectly named Joints selected', button=['Ok'])\r\n return\r\n\r\n #-- Rename current Joint\r\n resultJointName = x.replace('_JNT', curSuffix+'JNT')\r\n if mc.objExists(resultJointName):\r\n mc.confirmDialog(title='ERROR', message='New Joint name already exists', button=['Ok'])\r\n return\r\n\r\n mc.rename(x,resultJointName)\r\n x=resultJointName\r\n\r\n #-- Names from current Joint\r\n blendSuffix = '_'+suffixABoxValue+suffixBBoxValue[:1].upper() + suffixBBoxValue[1:]\r\n ikJointName = x.replace(curSuffix, suffixA)\r\n fkJointName = x.replace(curSuffix, suffixB)\r\n blendRName = x.replace(curSuffix+'JNT', blendSuffix+'Rot_BLND')\r\n blendTName = x.replace(curSuffix+'JNT', blendSuffix+'Trans_BLND')\r\n blendSName = x.replace(curSuffix+'JNT', blendSuffix+'Scale_BLND')\r\n\r\n #-- Duplicate Joints\r\n if mc.objExists(ikJointName) or mc.objExists(fkJointName):\r\n mc.confirmDialog(title='ERROR', message='Joint(s) already exist', button=['Ok'])\r\n return\r\n ikJoints = mc.duplicate(x, n=ikJointName, po=True)\r\n fkJoints = mc.duplicate(x, n=fkJointName, po=True)\r\n\r\n #-- Parent Duplicated Joints\r\n resultParent=mc.listRelatives(x, p=True, c=False)\r\n if resultParent:\r\n for p in resultParent:\r\n if curSuffix in p:\r\n ikParent=p.replace(curSuffix, suffixA)\r\n fkParent=p.replace(curSuffix, suffixB)\r\n mc.parent(ikJointName, ikParent)\r\n mc.parent(fkJointName, fkParent)\r\n\r\n #-- Create and attach Blends\r\n mc.createNode('blendColors', n=blendRName)\r\n mc.createNode('blendColors', n=blendTName)\r\n mc.createNode('blendColors', n=blendSName)\r\n #--Inputs\r\n mc.connectAttr(ikJointName + '.rotate', blendRName + '.color2')\r\n mc.connectAttr(fkJointName + '.rotate', blendRName + '.color1')\r\n mc.connectAttr(ikJointName + '.translate', blendTName + '.color2')\r\n mc.connectAttr(fkJointName + '.translate', blendTName + '.color1')\r\n mc.connectAttr(ikJointName + '.scale', blendSName + '.color2')\r\n mc.connectAttr(fkJointName + '.scale', blendSName + '.color1')\r\n mc.connectAttr(ctrlCrv + '.' + attrName, blendRName + '.blender')\r\n mc.connectAttr(ctrlCrv + '.' + attrName, blendTName + '.blender')\r\n mc.connectAttr(ctrlCrv + '.' + attrName, blendSName + '.blender')\r\n #--Outputs\r\n mc.connectAttr(blendRName + '.output', x + '.rotate')\r\n mc.connectAttr(blendTName + '.output', x + '.translate')\r\n mc.connectAttr(blendSName + '.output', x + '.scale')\r\n\r\n def createAttr(self):\r\n #--Set attribute names\r\n if not attrNameBoxValue:\r\n mc.confirmDialog(title='ERROR', message='Enter a valid attribute name', button=['Ok'])\r\n return\r\n if not attrNiceNameBoxValue:\r\n mc.confirmDialog(title='ERROR', message='Enter a valid nice name', button=['Ok'])\r\n return\r\n\r\n if not ctrlCrvBoxValue or not mc.objExists(ctrlCrvBoxValue):\r\n mc.confirmDialog(title='ERROR', message='Select valid Control Curve', button=['Ok'])\r\n return\r\n\r\n global attrName\r\n global attrNiceName\r\n global ctrlCrv\r\n attrName = attrNameBoxValue\r\n attrNiceName = attrNiceNameBoxValue\r\n ctrlCrv=ctrlCrvBoxValue\r\n\r\n if not mc.attributeQuery(attrName, node=ctrlCrv, exists=True):\r\n mc.addAttr(ctrlCrv, ln=attrName, nn=attrNiceName, min=0, max=1, at='float', dv=0, k=True)\r\n\r\n\r\n\r\n\r\nmain()", "sub_path": "scripts/jointDuplicator/jointDuplicator.py", "file_name": "jointDuplicator.py", "file_ext": "py", "file_size_in_byte": 10453, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "maya.OpenMayaUI.MQtUtil.mainWindow", "line_number": 31, "usage_type": "call"}, {"api_name": "maya.OpenMayaUI.MQtUtil", "line_number": 31, "usage_type": "attribute"}, {"api_name": "maya.OpenMayaUI", "line_number": 31, "usage_type": "name"}, {"api_name": "shiboken2.wrapInstance", "line_number": 32, "usage_type": "call"}, {"api_name": "PySide2.QtWidgets.QWidget", "line_number": 32, "usage_type": "attribute"}, {"api_name": "PySide2.QtWidgets", "line_number": 32, "usage_type": "name"}, {"api_name": "PySide2.QtWidgets.QDialog", "line_number": 58, "usage_type": "attribute"}, {"api_name": "PySide2.QtWidgets", "line_number": 58, "usage_type": "name"}, {"api_name": "PySide2.QtCore.Qt", "line_number": 65, "usage_type": "attribute"}, {"api_name": "PySide2.QtCore", "line_number": 65, "usage_type": "name"}, {"api_name": "jointDuplicatorUI.Ui_MainWindow", "line_number": 66, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 73, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 74, "usage_type": "call"}, {"api_name": "maya.cmds.objExists", "line_number": 104, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 104, "usage_type": "name"}, {"api_name": "maya.cmds.ls", "line_number": 110, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 110, "usage_type": "name"}, {"api_name": "maya.cmds.ls", "line_number": 130, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 130, "usage_type": "name"}, {"api_name": "maya.cmds.confirmDialog", "line_number": 133, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 133, "usage_type": "name"}, {"api_name": "maya.cmds.listRelatives", "line_number": 138, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 138, "usage_type": "name"}, {"api_name": "maya.cmds.ls", "line_number": 146, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 146, "usage_type": "name"}, {"api_name": "maya.cmds.confirmDialog", "line_number": 148, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 148, "usage_type": "name"}, {"api_name": "maya.cmds.confirmDialog", "line_number": 155, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 155, "usage_type": "name"}, {"api_name": "maya.cmds.confirmDialog", "line_number": 160, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 160, "usage_type": "name"}, {"api_name": "maya.cmds.confirmDialog", "line_number": 169, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 169, "usage_type": "name"}, {"api_name": "maya.cmds.objExists", "line_number": 174, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 174, "usage_type": "name"}, {"api_name": "maya.cmds.confirmDialog", "line_number": 175, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 175, "usage_type": "name"}, {"api_name": "maya.cmds.rename", "line_number": 178, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 178, "usage_type": "name"}, {"api_name": "maya.cmds.objExists", "line_number": 190, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 190, "usage_type": "name"}, {"api_name": "maya.cmds.confirmDialog", "line_number": 191, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 191, "usage_type": "name"}, {"api_name": "maya.cmds.duplicate", "line_number": 193, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 193, "usage_type": "name"}, {"api_name": "maya.cmds.duplicate", "line_number": 194, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 194, "usage_type": "name"}, {"api_name": "maya.cmds.listRelatives", "line_number": 197, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 197, "usage_type": "name"}, {"api_name": "maya.cmds.parent", "line_number": 203, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 203, "usage_type": "name"}, {"api_name": "maya.cmds.parent", "line_number": 204, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 204, "usage_type": "name"}, {"api_name": "maya.cmds.createNode", "line_number": 207, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 207, "usage_type": "name"}, {"api_name": "maya.cmds.createNode", "line_number": 208, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 208, "usage_type": "name"}, {"api_name": "maya.cmds.createNode", "line_number": 209, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 209, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 211, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 211, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 212, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 212, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 213, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 213, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 214, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 214, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 215, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 215, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 216, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 216, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 217, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 217, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 218, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 218, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 219, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 219, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 221, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 221, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 222, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 222, "usage_type": "name"}, {"api_name": "maya.cmds.connectAttr", "line_number": 223, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 223, "usage_type": "name"}, {"api_name": "maya.cmds.confirmDialog", "line_number": 228, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 228, "usage_type": "name"}, {"api_name": "maya.cmds.confirmDialog", "line_number": 231, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 231, "usage_type": "name"}, {"api_name": "maya.cmds.objExists", "line_number": 234, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 234, "usage_type": "name"}, {"api_name": "maya.cmds.confirmDialog", "line_number": 235, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 235, "usage_type": "name"}, {"api_name": "maya.cmds.attributeQuery", "line_number": 245, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 245, "usage_type": "name"}, {"api_name": "maya.cmds.addAttr", "line_number": 246, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 246, "usage_type": "name"}]} +{"seq_id": "390685810", "text": "import numpy as np\nfrom ..utils import get_g\n\n\nclass DirectBias():\n \"\"\"Direct Bias Metric\"\"\"\n def __init__(self, E, c=1, g=None):\n \"\"\"Direct bias calculation\n Args:\n E (WE class object): Word embeddings object\n Kwargs:\n c (float): strictness factor\n g (np.array): gender direction\n\n \"\"\"\n if g is None:\n g = get_g(E)\n assert len(g) == E.dim \n self.g = g\n self.E = E\n self.c = c\n\n def _direct_bias(self, vec):\n \"\"\"Direct bias computation\n\n Args:\n vec (np.array): numpy array to calculate direct bias for\n \n \"\"\"\n return np.power(np.abs(vec.dot(self.g)), self.c) \n\n def compute(self, word_list):\n \"\"\"Compute direct bias\n\n Args:\n word_list (list): list of words to compute bias for. \n Returns:\n The direct bias of each word in the `word_list`.\n \n \"\"\"\n if not isinstance(word_list, list):\n word_list = [word_list]\n db = np.mean(\n [self._direct_bias(self.E.v(word)) for word in word_list]\n )\n return db\n", "sub_path": "fee/metrics/direct_bias.py", "file_name": "direct_bias.py", "file_ext": "py", "file_size_in_byte": 1174, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "utils.get_g", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 43, "usage_type": "call"}]} +{"seq_id": "255649237", "text": "from collections import defaultdict\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn.feature_extraction import DictVectorizer\nfrom sklearn.cross_validation import KFold\nfrom sklearn.metrics import accuracy_score, precision_score, recall_score, f1_score\nfrom knock73 import feature, feature_vector, train, create_feature\nfrom knock74 import fit_feature, load_dict\nfrom knock76 import predict\nimport scipy\n\ndef cv(feature_dict, feature, polarity, folds):\n kfold = KFold(len(polarity), n_folds = folds)\n count, f1, recall, precision, accuracy = 0, 0, 0, 0, 0\n for train, test in kfold:\n LR = LogisticRegression()\n count += 1\n x = [(feature[i]) for i in train]\n y = [(polarity[i])for i in train]\n LR.fit(scipy.sparse.vstack(x), (y))\n\n test_label = []\n answer_label = [(polarity[j]) for j in test]\n for j in test:\n query = feature[j]\n result = -1 if query.shape[1] != len(feature_dict) else predict(LR, query)\n test_label.append(int(result[0]))\n accuracy += accuracy_score(answer_label, test_label)\n precision += precision_score(answer_label, test_label)\n recall += recall_score(answer_label, test_label)\n f1 += f1_score(answer_label, test_label)\n print('{}_fold finished.'.format(count))\n return accuracy, precision, recall, f1\n\nif __name__ == '__main__':\n feature_dict = load_dict('knock72_file.txt')\n feature, polarity = create_feature(open('sentiment.txt'))\n threshold = 0.5\n folds = 5\n accuracy, precision, recall, f1 = cv(feature_dict, feature, polarity, folds)\n print('accuracy: {}'.format(accuracy / 5))\n print('precision: {}'.format(precision / 5))\n print('recall: {}'.format(recall / 5))\n print('F1: {}'.format(f1 / 5))\n\n\n\n\n\n", "sub_path": "arai/chapter08/knock78.py", "file_name": "knock78.py", "file_ext": "py", "file_size_in_byte": 1809, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "sklearn.cross_validation.KFold", "line_number": 12, "usage_type": "call"}, {"api_name": "knock73.train", "line_number": 14, "usage_type": "name"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 15, "usage_type": "call"}, {"api_name": "knock73.feature", "line_number": 17, "usage_type": "name"}, {"api_name": "knock73.train", "line_number": 17, "usage_type": "name"}, {"api_name": "knock73.train", "line_number": 18, "usage_type": "name"}, {"api_name": "scipy.sparse.vstack", "line_number": 19, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 19, "usage_type": "attribute"}, {"api_name": "knock73.feature", "line_number": 24, "usage_type": "name"}, {"api_name": "knock76.predict", "line_number": 25, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 27, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_score", "line_number": 28, "usage_type": "call"}, {"api_name": "sklearn.metrics.recall_score", "line_number": 29, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 30, "usage_type": "call"}, {"api_name": "knock74.load_dict", "line_number": 35, "usage_type": "call"}, {"api_name": "knock73.feature", "line_number": 36, "usage_type": "name"}, {"api_name": "knock73.create_feature", "line_number": 36, "usage_type": "call"}, {"api_name": "knock73.feature", "line_number": 39, "usage_type": "argument"}]} +{"seq_id": "533152924", "text": "from threading import Condition\nfrom datetime import datetime\nimport configparser\nimport logging\nimport sys\nimport os\n\nimport click\n\nfrom .extract import Provider\nfrom .transform import Transformer\nfrom .load import Loader\n\n\n# logging utility\nlogging.basicConfig(level=logging.INFO,\n format='%(asctime)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S')\nlog = logging.getLogger(__name__)\n\n\n@click.command()\n@click.argument(\"connect\", default='connect.ini')\n@click.argument(\"etl\", default='config.ini')\ndef etl_from_config(connect, etl):\n # config variables\n connect_config = configparser.ConfigParser()\n connect_config.read(connect)\n etl_config = configparser.ConfigParser()\n etl_config.read(etl)\n\n # thread concurrent structures\n unpr_list_cond = Condition()\n unpr_list = []\n\n pr_list_cond = Condition()\n pr_list = []\n\n extract_thread = Provider(unpr_list_cond, unpr_list, etl_config)\n extract_thread.start()\n\n transform_thread = Transformer(unpr_list_cond, pr_list_cond,\n unpr_list, pr_list, etl_config)\n transform_thread.start()\n\n load_thread = Loader(pr_list_cond, pr_list, etl_config, connect_config)\n load_thread.start()\n\n\nif __name__ == '__main__':\n etl_from_config()\n", "sub_path": "src/poe_price/data/etl/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1274, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "logging.basicConfig", "line_number": 16, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 16, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 18, "usage_type": "call"}, {"api_name": "configparser.ConfigParser", "line_number": 26, "usage_type": "call"}, {"api_name": "configparser.ConfigParser", "line_number": 28, "usage_type": "call"}, {"api_name": "threading.Condition", "line_number": 32, "usage_type": "call"}, {"api_name": "threading.Condition", "line_number": 35, "usage_type": "call"}, {"api_name": "extract.Provider", "line_number": 38, "usage_type": "call"}, {"api_name": "transform.Transformer", "line_number": 41, "usage_type": "call"}, {"api_name": "load.Loader", "line_number": 45, "usage_type": "call"}, {"api_name": "click.command", "line_number": 21, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 22, "usage_type": "call"}, {"api_name": "click.argument", "line_number": 23, "usage_type": "call"}]} +{"seq_id": "603808913", "text": "import argparse\nimport torch\nfrom rtpt import RTPT\nfrom models import BaseNet\nimport clip\nfrom PIL import Image\nfrom matplotlib.colors import LinearSegmentedColormap\nfrom captum.attr import IntegratedGradients\nfrom captum.attr import GradientShap\nfrom captum.attr import Occlusion\nfrom captum.attr import Saliency\nfrom captum.attr import NoiseTunnel\nfrom captum.attr import visualization as viz\nimport numpy as np\nfrom torchvision.transforms import Normalize\nimport matplotlib.pyplot as plt\nimport os\n\nparser = argparse.ArgumentParser(description='Crazy Stuff')\nparser.add_argument('--data_dir', default='./data',\n help='Select data path')\nparser.add_argument('--data_name', default='rt-polarity', type=str, choices=['rt-polarity', 'toxicity',\n 'toxicity_full', 'ethics', 'restaurant'],\n help='Select name of data set')\nparser.add_argument('--num_prototypes', default=10, type=int,\n help='Total number of prototypes')\nparser.add_argument('--num_classes', default=2, type=int,\n help='How many classes are to be classified?')\nparser.add_argument('--class_weights', default=[0.5,0.5],\n help='Class weight for cross entropy loss')\nparser.add_argument('-g','--gpu', type=int, default=[0], nargs='+',\n help='GPU device number(s)')\nparser.add_argument('--one_shot', type=bool, default=False,\n help='Whether to use one-shot learning or not (i.e. only a few training examples)')\nparser.add_argument('--discard', type=bool, default=False,\n help='Whether edge cases in the middle between completely toxic(1) and not toxic(0) shall be omitted')\nparser.add_argument('--proto_size', type=int, default=1,\n help='Define how many words should be used to define a prototype')\nparser.add_argument('--language_model', type=str, default='Bert', choices=['Bert','SentBert','GPT2', 'Clip_ViT-B/32','Clip_RN50x4', 'Clip_RN50'],\n help='Define which language model to use')\nparser.add_argument('--avoid_spec_token', type=bool, default=False,\n help='Whether to manually set PAD, SEP and CLS token to high value after Bert embedding computation')\nparser.add_argument('--compute_emb', type=bool, default=False,\n help='Whether to recompute (True) the embedding or just load it (False)')\nparser.add_argument('--metric', type=str, default='L2',\n help='metric')\nparser.add_argument('--input_type', type=str, required=True, choices=['text', 'img'],\n help='choose between text and image')\nparser.add_argument('--explain', type=bool, default=False,\n help='Who needs help anyway?')\n\n\nlabels = ['non toxic', 'toxic']\n\ninv_normalize = Normalize(\n mean=[-0.48145466/0.26862954, -0.4578275/0.26130258, -0.40821073/0.27577711],\n std=[1/0.26862954, 1/0.26130258, 1/0.27577711]\n)\n\ndefault_cmap = LinearSegmentedColormap.from_list('custom blue',\n [(0, '#ffffff'),\n (0.25, '#000000'),\n (1, '#000000')], N=256)\n\n\nclass ClipProbeModel(torch.nn.Module):\n def __init__(self):\n super(ClipProbeModel, self).__init__()\n self.model_probe = BaseNet(args)\n self.model_probe.load_state_dict(torch.load(args.model_path))\n self.model_probe.to(f'cuda:{args.gpu[0]}')\n self.model_probe.eval()\n self.MMM, self.preprocess = clip.load(args.language_model.split('_')[1], f'cuda:{args.gpu[0]}')\n self.MMM.to(f'cuda:{args.gpu[0]}')\n self.MMM.eval()\n\n def forward(self, x):\n if args.input_type == 'text':\n emb = self._forward_txt(x)\n else:\n emb = self._forward_img(x)\n predicted_label = self.model_probe.forward(emb, [])\n return predicted_label\n\n def _forward_txt(self, x):\n return self.MMM.encode_text(x).float()\n\n def _forward_img(self, x):\n #x = x.unsqueeze(0)\n return self.MMM.encode_image(x).float()\n\n\ndef explain_pred(model, x, y, file_name):\n #gradientshap(model, x, y, file_name)\n #saliency(model, x, y, file_name)\n #occlusion(model, x, y, file_name)\n noise_tunnel(model, x, y, file_name)\n #gradientshap(model, x, y, file_name)\n\n\n\ndef eval_model(args):\n args.model_path = './experiments/train_results/toxicity/05-19-14:07_baseline_Clip_ViT-B/best_model.pth.tar'\n #args.model_path = './experiments/train_results/toxicity/05-19-17:14_baseline_Clip_RN50/best_model.pth.tar'\n #args.model_path = './experiments/train_results/toxicity/05-19-17:30_baseline_Clip_RN50x4/best_model.pth.tar'\n model = ClipProbeModel()\n\n if args.input_type == 'text':\n print('Eval a text')\n txt = ['You are an asshole.']\n file_name = txt[0].replace(' ', '_').replace('.', '')\n x = clip.tokenize(txt)\n elif args.input_type == 'img':\n print('Eval an image')\n file_name = 'b14_p254_12'\n #file_name = 'b14_p253_4'\n #file_name = 'b13_p233_1'\n #file_name = 'b14_p253_11'\n #file_name = 'b2_p28_8'\n #file_name = 'b10_p136_15'\n x = model.preprocess(Image.open(f\"/workspace/datasets/SMID_images_400px/img/{file_name}.jpg\")).unsqueeze(0)\n else:\n raise ValueError('input type unknown')\n print(\"Running on gpu {}\".format(args.gpu))\n\n x = x.to(f'cuda:{args.gpu[0]}')\n\n logits = model(x)\n probs = logits.softmax(dim=-1)\n\n prediction_score, pred_label_idx = torch.topk(probs.float(), 1)\n\n pred_label_idx.squeeze_()\n predicted_label = labels[pred_label_idx]\n print(f'Predicted: {predicted_label} ({prediction_score.squeeze().item() * 100:.2f})')\n\n if args.explain and args.input_type == 'img':\n explain_pred(model, x, pred_label_idx, file_name)\n\n\ndef occlusion(model, x, y, file_name):\n inv_transformed_img = inv_normalize(x)\n ablator = Occlusion(model)\n attribution = ablator.attribute(x, target=y, sliding_window_shapes=(8, 8), strides=(2,2))\n fig, axis = viz.visualize_image_attr_multiple(np.transpose(attribution.cpu().detach().numpy(), (1, 2, 0)),\n np.transpose(inv_transformed_img.cpu().detach().numpy(), (1, 2, 0)),\n [\"original_image\", \"heat_map\"],\n [\"all\", \"absolute_value\"],\n cmap=default_cmap,\n show_colorbar=True)\n\n save_path = './clip_stuff/explain/toxicity/'\n os.makedirs(save_path, exist_ok=True)\n fig.savefig(os.path.join(save_path, f'occlusion_{file_name}.png'))\n\n\ndef saliency(model, x, y, file_name):\n inv_transformed_img = inv_normalize(x)\n saliency = Saliency(model)\n\n attribution = saliency.attribute(x, target=y)\n fig, axis = viz.visualize_image_attr_multiple(np.transpose(attribution.squeeze().cpu().detach().numpy(), (1, 2, 0)),\n np.transpose(inv_transformed_img.squeeze().cpu().detach().numpy(), (1, 2, 0)),\n [\"original_image\", \"heat_map\"],\n [\"all\", \"absolute_value\"],\n cmap=default_cmap,\n show_colorbar=True)\n\n save_path = './clip_stuff/explain/toxicity/'\n os.makedirs(save_path, exist_ok=True)\n fig.savefig(os.path.join(save_path, f'saliency_{file_name}.png'))\n\n\n\ndef gradientshap(model, x, y, file_name):\n inv_transformed_img = inv_normalize(x)\n gradient_shap = GradientShap(model)\n\n # Defining baseline distribution of images\n rand_img_dist = torch.cat([x * 0, x * 1])\n\n attributions_gs = gradient_shap.attribute(x,\n n_samples=50,\n stdevs=0.0001,\n baselines=rand_img_dist,\n target=y)\n fig, axis = viz.visualize_image_attr_multiple(np.transpose(attributions_gs.squeeze().cpu().detach().numpy(), (1, 2, 0)),\n np.transpose(inv_transformed_img.squeeze().cpu().detach().numpy(), (1, 2, 0)),\n [\"original_image\", \"heat_map\"],\n [\"all\", \"absolute_value\"],\n cmap=default_cmap,\n show_colorbar=True)\n\n save_path = './clip_stuff/explain/toxicity/'\n os.makedirs(save_path, exist_ok=True)\n fig.savefig(os.path.join(save_path, f'gradientshap_{file_name}.png'))\n\n\ndef noise_tunnel(model, x, y, file_name):\n inv_transformed_img = inv_normalize(x)\n integrated_gradients = IntegratedGradients(model)\n noise_tunnel_ = NoiseTunnel(integrated_gradients)\n\n attributions_ig_nt = noise_tunnel_.attribute(x, nt_samples=5, nt_type='smoothgrad_sq', target=y)\n fig, axis = viz.visualize_image_attr_multiple(np.transpose(attributions_ig_nt.squeeze().cpu().detach().numpy(), (1, 2, 0)),\n np.transpose(inv_transformed_img.squeeze().cpu().detach().numpy(), (1, 2, 0)),\n [\"original_image\", \"heat_map\"],\n [\"all\", \"positive\"],\n cmap=default_cmap,\n show_colorbar=True,\n use_pyplot=False)\n\n save_path = './clip_stuff/explain/toxicity/'\n os.makedirs(save_path, exist_ok=True)\n fig.savefig(os.path.join(save_path, f'noise_tunnel_{file_name}.png'))\n\n\nif __name__ == '__main__':\n # torch.manual_seed(0)\n # np.random.seed(0)\n torch.set_num_threads(6)\n args = parser.parse_args()\n\n # Create RTPT object and start the RTPT tracking\n rtpt = RTPT(name_initials='Kersting', experiment_name='CrazyStuff', max_iterations=1)\n rtpt.start()\n\n eval_model(args)\n", "sub_path": "main/eval.py", "file_name": "eval.py", "file_ext": "py", "file_size_in_byte": 10201, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 19, "usage_type": "call"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.colors.LinearSegmentedColormap.from_list", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.colors.LinearSegmentedColormap", "line_number": 60, "usage_type": "name"}, {"api_name": "torch.nn", "line_number": 66, "usage_type": "attribute"}, {"api_name": "models.BaseNet", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 70, "usage_type": "call"}, {"api_name": "clip.load", "line_number": 73, "usage_type": "call"}, {"api_name": "clip.tokenize", "line_number": 112, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 121, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 121, "usage_type": "name"}, {"api_name": "torch.topk", "line_number": 131, "usage_type": "call"}, {"api_name": "captum.attr.Occlusion", "line_number": 143, "usage_type": "call"}, {"api_name": "captum.attr.visualization.visualize_image_attr_multiple", "line_number": 145, "usage_type": "call"}, {"api_name": "captum.attr.visualization", "line_number": 145, "usage_type": "name"}, {"api_name": "numpy.transpose", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 146, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 153, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 154, "usage_type": "call"}, {"api_name": "os.path", "line_number": 154, "usage_type": "attribute"}, {"api_name": "captum.attr.Saliency", "line_number": 159, "usage_type": "call"}, {"api_name": "captum.attr.visualization.visualize_image_attr_multiple", "line_number": 162, "usage_type": "call"}, {"api_name": "captum.attr.visualization", "line_number": 162, "usage_type": "name"}, {"api_name": "numpy.transpose", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 163, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 170, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 171, "usage_type": "call"}, {"api_name": "os.path", "line_number": 171, "usage_type": "attribute"}, {"api_name": "captum.attr.GradientShap", "line_number": 177, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 180, "usage_type": "call"}, {"api_name": "captum.attr.visualization.visualize_image_attr_multiple", "line_number": 187, "usage_type": "call"}, {"api_name": "captum.attr.visualization", "line_number": 187, "usage_type": "name"}, {"api_name": "numpy.transpose", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 188, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 195, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 196, "usage_type": "call"}, {"api_name": "os.path", "line_number": 196, "usage_type": "attribute"}, {"api_name": "captum.attr.IntegratedGradients", "line_number": 201, "usage_type": "call"}, {"api_name": "captum.attr.NoiseTunnel", "line_number": 202, "usage_type": "call"}, {"api_name": "captum.attr.visualization.visualize_image_attr_multiple", "line_number": 205, "usage_type": "call"}, {"api_name": "captum.attr.visualization", "line_number": 205, "usage_type": "name"}, {"api_name": "numpy.transpose", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 206, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 214, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 215, "usage_type": "call"}, {"api_name": "os.path", "line_number": 215, "usage_type": "attribute"}, {"api_name": "torch.set_num_threads", "line_number": 221, "usage_type": "call"}, {"api_name": "rtpt.RTPT", "line_number": 225, "usage_type": "call"}, {"api_name": "rtpt.start", "line_number": 226, "usage_type": "call"}]} +{"seq_id": "3212806", "text": "import datetime\nimport smtplib\nimport email.MIMEMultipart\n\nfromAddress = \"DTA-TWC-Logmon@cable.comcast.com\"\ntoAddress = [\"DASTeamB@cable.comcast.com\"]\nSERVER = 'mailrelay.comcast.com'\n\nmessage = \"Test\"\nemail = \"Subject: ALERT! - DTA2 : IVR : logMon: callEndReason :%s hour\\n\\n%s\" % (datetime.datetime.now().strftime(\"%Y-%m-%d %H\"), message)\n\n\nserver = smtplib.SMTP(SERVER)\n#server.sendmail(fromAddress, toAddress, email)\nserver.quit()\n", "sub_path": "mailapp.py", "file_name": "mailapp.py", "file_ext": "py", "file_size_in_byte": 435, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "email.MIMEMultipart", "line_number": 10, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 10, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 10, "usage_type": "attribute"}, {"api_name": "smtplib.SMTP", "line_number": 13, "usage_type": "call"}]} +{"seq_id": "179508866", "text": "# Author: FFng, ffng0x15@gmail.com\n# 2017-08-18\n# Python 3.6\n\n\"\"\"\nData Structure:\nseat_info = {\n '_id': int,\n 'no': str,\n 'name': str,\n 'status': int,\n 'status_name': str\n}\n\nseat_info_dict = {\n 'meta': {\n 'updated_at': float, Unix timestamp,\n 'updated_at_human': str\n },\n seat_name: seat_info,\n ...\n}\n\nchanges_dict = {\n seat_name: {\n 'old': {\n 'status_name': str, old_status_name,\n 'updated_at': float, old_updated_at\n },\n 'new': {\n 'status_name': str, new_status_name,\n 'updated_at': float, new_updated_at\n }\n },\n ...\n}\n\nbooking_logs = {\n booking_no: { # str\n 'space': str, space,\n 'start_time': str, start_time,\n 'end_time': str, end_time,\n 'status_name': str, status_name\n },\n ...\n}\n\n\n\"\"\"\n\nimport requests\nimport time\nimport datetime\nimport json\nimport logging\nimport re\nimport os\nimport smtplib\nfrom email.mime.text import MIMEText\n\nfrom lxml import html as lxml_html\n\nimport ffc\n\ntry:\n from .trace_with_logger import TraceWithLogger\n from .exceptions import DateIsNotAvailable\n from .process_captcha import ProcessCaptcha\n from .exceptions import SeatIsNotAvailable\nexcept ImportError:\n from trace_with_logger import TraceWithLogger\n from exceptions import DateIsNotAvailable\n from process_captcha import ProcessCaptcha\n from exceptions import SeatIsNotAvailable\n\n\nAREA_MAP = {\n 46: '1层',\n 51: '301/302'\n}\n\nSEAT_STATUS_MAP = {\n 1: '空闲',\n 2: '已预约',\n 6: '使用中',\n 7: '临时离开'\n}\n\n\nLOGGER = ffc.init_logger(debug=False, name='trace_with_logger', file_path='./trace_with_logger.log',\n ch_level=logging.WARNING, ch_process_name=False, ch_thread_name=False,\n fh_level=logging.DEBUG, fh_process_name=False, fh_thread_name=False,\n fh_timed_rotating=True)\n\n\ndef trace_with_logger(record_input=True, record_output=True):\n global LOGGER\n return TraceWithLogger(logger=LOGGER, record_input=record_input, record_output=record_output)\n\n\nclass SeatMonitor:\n\n def __init__(self, debug=False, config='./private_config.json'):\n self.debug = debug\n if isinstance(config, str):\n self.config_dict = json.load(open(config, 'r', encoding='utf-8'))\n elif isinstance(config, dict):\n self.config_dict = config\n else:\n raise TypeError('Unexpected type of \"config\": {t}'.format(t=type(config)))\n log_path = self.config_dict.get('log_path') or './default_log_path.log'\n print('debug: {d}, log_path: {p}'.format(d=repr(self.debug), p=repr(log_path)))\n\n self.logger = ffc.init_logger(debug=self.debug, name=__name__, file_path=log_path,\n ch_level=logging.INFO, ch_process_name=False, ch_thread_name=False,\n fh_level=logging.DEBUG, fh_process_name=False, fh_thread_name=False,\n fh_timed_rotating=True)\n\n self.seat_info_dict = dict()\n self._seat_id_name_map = dict()\n self.session = requests.Session()\n self.login_status = False\n\n def _build_seat_id_name_map(self):\n\n for name in self.seat_info_dict:\n if name == 'meta':\n continue\n seat_id = self.seat_info_dict[name]['_id']\n self._seat_id_name_map.update({seat_id: name})\n return True\n\n @staticmethod\n def t_unix_to_human(unix_timestamp):\n \"\"\"\n :param unix_timestamp: int/float, unix timestamp\n :return: str, in format '2017-08-18 Fri 14:58:23'\n \"\"\"\n return datetime.datetime.fromtimestamp(unix_timestamp).strftime('%Y-%m-%d %a %H:%M:%S')\n\n def get_seats_status(self, area: int=46, day: str=None):\n l = self.logger\n if not isinstance(area, int):\n raise TypeError('area: {t} _{v}_'.format(t=type(area), v=repr(area)))\n if area not in AREA_MAP:\n raise ValueError('Unknown area: _{v}_'.format(v=repr(area)))\n if not isinstance(day, str):\n raise TypeError('day: {t} _{v}_'.format(t=type(day), v=repr(day)))\n if not re.match('201\\d-\\d{2}-\\d{2}', day):\n raise TypeError('Invalid day: _{v}_'.format(v=repr(day)))\n\n url = 'http://202.114.9.133/api.php/spaces_old'\n now_dt = datetime.datetime.now()\n today = now_dt.strftime('%Y-%m-%d')\n start_time = '08:30' if day != today else now_dt.strftime('%H:%M')\n payload = {\n 'area': area,\n 'day': day,\n 'startTime': start_time,\n 'endTime': '22:00'\n }\n l.debug('payload: {d}'.format(d=payload))\n\n # TODO: Separate this network access part, add more exception processes\n try:\n r = requests.get(url, params=payload, timeout=10)\n json_dict = json.loads(r.content)\n assert json_dict['status'] == 1\n except:\n raise\n\n def __process_received_content():\n global SEAT_STATUS_MAP\n nonlocal json_dict\n updated_at = time.time()\n updated_at_human = self.t_unix_to_human(updated_at)\n try:\n seat_list = json_dict['data']['list']\n assert len(seat_list) > 0\n except (KeyError, AssertionError):\n raise\n\n _seat_info_dict = {\n 'meta': {\n 'updated_at': updated_at,\n 'updated_at_human': updated_at_human\n }\n }\n for d in seat_list:\n _id = d['id'] # int, 内部id,预约座位时用于指定座位\n assert isinstance(_id, int)\n no = d['no'] # 座位的编号\n name = d['name']\n assert isinstance(no, str) and isinstance(name, str) and no == name\n\n status = d['status']\n status_name = d['status_name']\n if (status not in SEAT_STATUS_MAP) or (status_name != SEAT_STATUS_MAP[status]):\n raise ValueError('status: _{s}_, name: _{n}_'.format(s=repr(status), n=repr(status_name)))\n\n seat_info = {\n '_id': _id,\n 'no': no,\n 'name': name,\n 'status': status,\n 'status_name': status_name\n }\n if name in _seat_info_dict:\n raise ValueError('Repetitive name _{n}_. old: _{old}_, new: _{new}_'.format(\n n=repr(name), old=_seat_info_dict.get(name), new=seat_info))\n _seat_info_dict.update({\n name: seat_info # # because only str can be used as key in JSON\n })\n return _seat_info_dict\n seat_info_dict = __process_received_content()\n self.seat_info_dict.update(seat_info_dict)\n self._build_seat_id_name_map()\n l.info(f'get {len(seat_info_dict):,d} seat\\'s info')\n\n def __save_to_json_file():\n global AREA_MAP\n nonlocal payload, seat_info_dict\n try:\n _path = self.config_dict['seat_info_path']\n except:\n raise\n _json_dict = {\n 'params': {\n 'area': payload['area'],\n 'area_name': AREA_MAP[payload['area']],\n 'day': payload['day'],\n 'start_time': payload['startTime'],\n 'end_time': payload['endTime']\n },\n 'seat_info_dict': seat_info_dict\n }\n\n _json_str = json.dumps(_json_dict, ensure_ascii=False, indent=None, sort_keys=True)\n with open(_path, mode='a', encoding='utf-8') as f:\n if f.tell() != 0:\n f.write('\\n')\n f.write(_json_str)\n return _path\n json_path = __save_to_json_file()\n l.debug('Saved json(seat_info_dict) to {p}'.format(p=repr(os.path.abspath(json_path))))\n\n return seat_info_dict\n\n def diff_seat_info_dict(self, old: dict, new: dict, focus_on: list):\n \"\"\"\n \n :param old: old seat_info_dict\n :param new: \n :param focus_on: List[str] list of seat_name\n :return: dict\n \"\"\"\n l = self.logger\n\n old_updated_at = old['meta']['updated_at']\n new_updated_at = new['meta']['updated_at']\n l.debug('updated time: old: {o}, new: {n}'.format(o=self.t_unix_to_human(old_updated_at),\n n=self.t_unix_to_human(new_updated_at)))\n if old_updated_at == new_updated_at:\n return {}\n\n if not focus_on:\n for name in old.keys():\n if name == 'meta':\n continue\n focus_on.append(name)\n\n _name_status = list()\n for name in focus_on:\n _name_status.append((name, old[name]['status_name']))\n\n if len(focus_on) <= 10:\n l.info(f'Focus on {len(focus_on):,d} seats: {_name_status}')\n else:\n l.info(f'Focus on {len(focus_on):,d} seats (>10)')\n\n changes_dict = dict()\n for name in focus_on:\n old_status_name = old[name]['status_name']\n new_status_name = new[name]['status_name']\n if old_status_name == new_status_name:\n continue\n\n changes_dict[name] = {\n 'old': {\n 'status_name': old_status_name,\n 'updated_at': old_updated_at\n },\n 'new': {\n 'status_name': new_status_name,\n 'updated_at': new_updated_at\n }\n }\n l.info(f'{len(changes_dict)} seats\\' status changed.')\n\n return changes_dict\n\n def describe_changes(self, changes_dict: dict):\n l = self.logger\n\n if not changes_dict:\n l.debug('no changes')\n return []\n\n l.debug(f'changes_dict: {changes_dict}')\n\n def __convert(t_unix):\n return datetime.datetime.fromtimestamp(t_unix).strftime('%H:%M')\n\n s_list = list()\n for name in changes_dict:\n d = changes_dict[name]\n old = d['old']\n new = d['new']\n s = (f\"[{name:s}]\"\n f\"{old['status_name']}({__convert(old['updated_at'])})\"\n f\"->{new['status_name']}({__convert(new['updated_at'])})\")\n s_list.append(s)\n l.info(s)\n\n return s_list\n\n @trace_with_logger()\n def send_mail(self, subject: str, content: str):\n l = self.logger\n cfg = self.config_dict['mail']\n smtp_host = cfg['smtp_host']\n smtp_port = cfg['smtp_port']\n mail_user = cfg['username']\n mail_pass = cfg['password']\n sender = mail_user\n receivers = cfg['receivers']\n content += f'\\nthis mail was generated at {ffc.t_ff()} (timestamp={int(time.time())})'\n msg = MIMEText(content+'', 'plain', 'utf-8')\n msg['Subject'] = subject\n msg['From'] = sender\n msg['To'] = receivers[0]\n l.debug(f'Generated a mail to {receivers}, subject: {subject}')\n\n try:\n smtp_obj = smtplib.SMTP_SSL(host=smtp_host, port=smtp_port)\n smtp_obj.login(mail_user, mail_pass)\n smtp_obj.sendmail(from_addr=sender, to_addrs=receivers, msg=msg.as_string())\n smtp_obj.quit()\n l.info(f\"Sent a mail to {msg['To']}\")\n return True\n except smtplib.SMTPException as e:\n l.error(e)\n return False\n\n def login(self):\n l = self.logger\n cfg = self.config_dict['user']\n s = self.session\n\n index_url = 'http://202.114.9.133/Home/Web/Index'\n l.debug(f'Opening Index page: {index_url}')\n r = s.get(index_url, timeout=10)\n r.raise_for_status()\n\n captcha_url = 'http://202.114.9.133/api.php/check'\n l.debug(f'Loading CAPTCHA image...')\n r = s.get(captcha_url, timeout=10)\n r.raise_for_status()\n _path = './tmp.png'\n open('tmp.png', 'wb').write(r.content)\n l.info(f'Saved CAPTCHA at {repr(os.path.abspath(_path))}, please input answer: ')\n\n _start = time.time()\n process_captcha = ProcessCaptcha()\n _, answer = process_captcha.decode_captcha(file_path=_path)\n ms = int((time.time() - _start)*1000.0)\n l.info(f'Auto processed CAPTCHA, answer: {answer}, used {ms:,}ms')\n\n # answer = input('Answer: ')\n l.info(f'Received answer: _{repr(answer)}_')\n login_url = 'http://202.114.9.133/api.php/login'\n payload = {\n 'username': cfg['username'],\n 'password': cfg['password'],\n 'verify': answer\n }\n l.info(f\"Logging... username: {repr(payload['username'])}, {login_url}\")\n r = s.post(login_url, data=payload, timeout=10)\n r.raise_for_status()\n j = json.loads(r.content, encoding='utf-8')\n assert j['msg'] == '登陆成功'\n self.login_status = True\n l.info(f'Login succeed')\n return True\n\n def get_booking_history(self):\n l = self.logger\n\n s = self.session\n if s is None:\n raise ValueError('Session is invalid. Login first.')\n\n booking_history_url = 'http://202.114.9.133/User/Index/book'\n l.debug(f'Opening booking history page... {booking_history_url}')\n r = s.get(booking_history_url, timeout=10)\n r.raise_for_status()\n assert len(r.text) > 20000\n _path = './tmp.html'\n open(_path, 'w', encoding='utf-8').write(r.text)\n l.debug(f'Saved booking history page into {repr(os.path.abspath(_path))}')\n\n tree = lxml_html.fromstring(r.text)\n\n def __pick(_tr, _index):\n return _tr.xpath(f'./td[{_index}]/text()')[0].strip()\n booking_logs = dict()\n for tr in tree.xpath('//table[@id=\"menu_table\"]//tr[contains(@id, \"list\")]'):\n booking_no = __pick(tr, 1) # 预约号\n space = __pick(tr, 2) # 预约空间\n start_time = __pick(tr, 3) # 开始时间\n end_time = __pick(tr, 4) # 结束时间\n status_name = __pick(tr, 5) # 当前状态\n # print(booking_no, space, start_time, end_time, status_name)\n booking_logs.update({\n booking_no: {\n 'space': space,\n 'start_time': start_time,\n 'end_time': end_time,\n 'status_name': status_name\n }\n })\n\n _path = './booking_history.json'\n _json_str = json.dumps(booking_logs, ensure_ascii=False, indent=None, sort_keys=True)\n with open(_path, mode='a', encoding='utf-8') as f:\n if f.tell() != 0:\n f.write('\\n')\n f.write(_json_str)\n l.debug(f'Saved {len(booking_logs):,d} records into {repr(os.path.abspath(_path))}')\n\n return booking_logs\n\n @trace_with_logger()\n def _book(self, wanted_day: str='2017-08-21', area_id: int=46, seat_id: int=1638):\n l = self.logger\n s = self.session\n seat_name = self._seat_id_name_map[seat_id]\n if s is None:\n raise ValueError('Session is invalid. Login first.')\n l.info(f'wanted_day: {repr(wanted_day)}, '\n f'area_id: {repr(area_id)} -> {AREA_MAP[area_id]}, '\n f'seat_id: {repr(seat_id)} -> {seat_name}')\n\n # 获取可以预约的日期\n area_status_url = f'http://202.114.9.133/api.php/space_days/{area_id}'\n r = s.get(area_status_url, timeout=10)\n r.raise_for_status()\n j = json.loads(r.content, encoding='utf-8')\n assert j['msg'] == '获取可预约日期'\n day_list = list()\n for d in j['data']['list']:\n day_list.append(d['day'])\n l.debug(f'Available day: {day_list}')\n\n if wanted_day not in day_list:\n l.info(f'wanted_day {wanted_day} is not available')\n raise DateIsNotAvailable(f'{wanted_day} is not available.')\n\n # 获取可预约的时间段\n space_time_url = 'http://202.114.9.133/api.php/space_time_buckets'\n payload = {\n 'day': wanted_day,\n 'area': area_id\n }\n r = s.get(space_time_url, params=payload, timeout=10)\n j = json.loads(r.content, encoding='utf-8')\n assert j['msg'] == '获取可预约时间段'\n book_time_id = j['data']['list'][0]['bookTimeId']\n l.debug(f'segment/bookTimeId: {repr(book_time_id)}')\n\n # 预约\n book_url = f'http://202.114.9.133/api.php/spaces/{seat_id}/book'\n payload = {\n 'access_token': s.cookies.get('access_token'),\n 'userid': self.config_dict['user']['username'],\n 'segment': book_time_id,\n 'type': 1\n }\n r = s.post(book_url, data=payload, timeout=10)\n j = json.loads(r.content, encoding='utf-8')\n try:\n assert j['status'] == 1\n assert '预约成功' in j['msg']\n except Exception:\n l.error(f'{repr(j)}')\n raise\n\n d = j['data']['list']\n booking_no = d['no']\n seat_name = d['spaceInfo']['name']\n start_time = d['beginTime']['date']\n end_time = d['endTime']['date']\n status_name = d['statusName']\n l.info(f'Book succeed: {booking_no} {seat_name} {start_time}-{end_time} {status_name}')\n return True\n\n def book(self, wanted_day: str='2017-08-21', seat_name: str='1005'):\n l = self.logger\n\n self.get_seats_status(area=46, day=wanted_day)\n if seat_name not in self.seat_info_dict:\n raise ValueError(f'Unknown seat_name: _{repr(seat_name)}_')\n seat_id = self.seat_info_dict[seat_name]['_id']\n\n current_status_name = self.seat_info_dict[seat_name]['status_name']\n l.info(f'wanted_day: {wanted_day}, seat_name: {seat_name}, status: {current_status_name}')\n if current_status_name != '空闲':\n l.info('Seat is unavailable')\n raise SeatIsNotAvailable\n\n if self._book(wanted_day=wanted_day, area_id=46, seat_id=seat_id):\n # TODO\n return True\n else:\n return False\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "seat_monitor.py", "file_name": "seat_monitor.py", "file_ext": "py", "file_size_in_byte": 18472, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "ffc.init_logger", "line_number": 90, "usage_type": "call"}, {"api_name": "logging.WARNING", "line_number": 91, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 92, "usage_type": "attribute"}, {"api_name": "trace_with_logger.TraceWithLogger", "line_number": 98, "usage_type": "call"}, {"api_name": "json.load", "line_number": 106, "usage_type": "call"}, {"api_name": "ffc.init_logger", "line_number": 114, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 115, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 116, "usage_type": "attribute"}, {"api_name": "requests.Session", "line_number": 121, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 139, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 139, "usage_type": "attribute"}, {"api_name": "re.match", "line_number": 149, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 153, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 153, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 166, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 167, "usage_type": "call"}, {"api_name": "time.time", "line_number": 175, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 238, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 245, "usage_type": "call"}, {"api_name": "os.path", "line_number": 245, "usage_type": "attribute"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 312, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 312, "usage_type": "attribute"}, {"api_name": "ffc.t_ff", "line_number": 337, "usage_type": "call"}, {"api_name": "time.time", "line_number": 337, "usage_type": "call"}, {"api_name": "email.mime.text.MIMEText", "line_number": 338, "usage_type": "call"}, {"api_name": "smtplib.SMTP_SSL", "line_number": 345, "usage_type": "call"}, {"api_name": "smtplib.SMTPException", "line_number": 351, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 371, "usage_type": "call"}, {"api_name": "os.path", "line_number": 371, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 373, "usage_type": "call"}, {"api_name": "process_captcha.ProcessCaptcha", "line_number": 374, "usage_type": "call"}, {"api_name": "process_captcha.decode_captcha", "line_number": 375, "usage_type": "call"}, {"api_name": "time.time", "line_number": 376, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 390, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 410, "usage_type": "call"}, {"api_name": "os.path", "line_number": 410, "usage_type": "attribute"}, {"api_name": "lxml.html.fromstring", "line_number": 412, "usage_type": "call"}, {"api_name": "lxml.html", "line_number": 412, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 434, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 439, "usage_type": "call"}, {"api_name": "os.path", "line_number": 439, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 458, "usage_type": "call"}, {"api_name": "exceptions.DateIsNotAvailable", "line_number": 467, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 476, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 490, "usage_type": "call"}, {"api_name": "exceptions.SeatIsNotAvailable", "line_number": 519, "usage_type": "name"}]} +{"seq_id": "313273540", "text": "import pygame\r\nimport sys\r\nimport time\r\nimport pickle\r\nfrom network import Network\r\nfrom wordSource import wordSource\r\nfrom input import InputBox\r\n\r\npygame.font.init()\r\n\r\n# Windows size\r\n_displayWidth = 510\r\n_displayHeight = 400\r\n\r\n# Game windows\r\ngameDisplay = pygame.display.set_mode((_displayWidth, _displayHeight))\r\npygame.display.set_caption(\"Game\")\r\nclock = pygame.time.Clock()\r\n\r\n# Colors\r\n_black = (0,0,0)\r\n_white = (255,255,255)\r\n_red = (200,0,0)\r\n_brightRed = (255,0,0)\r\n_green = (0,200,0)\r\n_brightGreen = (0,255,0)\r\n_blue = (0,0,200)\r\n_brightBlue = (0,0,255)\r\n_yellow = (200,200,0)\r\n\r\n# Text size\r\n_smallText = pygame.font.SysFont(\"cambria\", 20)\r\n_mediumText = pygame.font.SysFont(\"cambria\", 30)\r\n_largeText = pygame.font.SysFont(\"cambria\", 40)\r\n\r\n# Button size\r\n_btnWidth = 120\r\n_btnHeight = 50\r\n\r\n# Text input fields\r\n_inputBoxes = []\r\n_inputBoxes.append(InputBox(149, 10, 200, 30)) # Input Word\r\n\r\n_canvas = pygame.Surface((_displayWidth, _displayHeight))\r\n_canvas.fill(_white)\r\n_input1_bckg = pygame.Rect(151,12,196,26) # Clear Input Word\r\n\r\n### Format displayed text\r\ndef text_objects(text, font, color):\r\n textSurface = font.render(text, True, color)\r\n return textSurface, textSurface.get_rect()\r\n\r\n### Create button\r\ndef button(msg, x, y, ac, ic, action=None):\r\n mouse = pygame.mouse.get_pos()\r\n click = pygame.mouse.get_pressed()\r\n\r\n pygame.draw.rect(gameDisplay, ac,(x,y,_btnWidth,_btnHeight))\r\n\r\n if x+_btnWidth > mouse[0] > x and y+_btnHeight > mouse[1] > y:\r\n if click[0] == 1 and action != None:\r\n action()\r\n time.sleep(0.5)\r\n else:\r\n pygame.draw.rect(gameDisplay, ic,(x,y,_btnWidth,_btnHeight))\r\n \r\n # Button text\r\n textSurf, textRect = text_objects(msg, _smallText, _black)\r\n textRect.center = ( (x+(_btnWidth/2)), (y+(_btnHeight/2)) )\r\n gameDisplay.blit(textSurf, textRect)\r\n\r\n## Create local game\r\ndef createLocal():\r\n guessingWordWindow()\r\n print(\"Create new local game\")\r\n\r\n# Not started \r\ndef createServerGame():\r\n # ## Nickname input\r\n # ## Choose type (Public/Private)\r\n # ## Player count\r\n\r\n onlineGame()\r\n print(\"Create new game\")\r\n\r\n### Close game window\r\ndef quitGame():\r\n print(\"Close game\")\r\n pygame.quit()\r\n quit()\r\n\r\n\r\n\r\ndef redrawWindow(win, game, p):\r\n win.fill((128,128,128))\r\n\r\n if not(game.connected()):\r\n font = pygame.font.SysFont(\"comicsans\", 80)\r\n text = font.render(\"Waiting for Player...\", 1, (255,0,0), True)\r\n win.blit(text, (width/2 - text.get_width()/2, height/2 - text.get_height()/2))\r\n else:\r\n font = pygame.font.SysFont(\"comicsans\", 60)\r\n text = font.render(\"Your Move\", 1, (0, 255,255))\r\n win.blit(text, (80, 200))\r\n\r\n text = font.render(\"Opponents\", 1, (0, 255, 255))\r\n win.blit(text, (380, 200))\r\n\r\n move1 = game.get_player_move(0)\r\n move2 = game.get_player_move(1)\r\n if game.bothWent():\r\n text1 = font.render(move1, 1, (0,0,0))\r\n text2 = font.render(move2, 1, (0, 0, 0))\r\n else:\r\n if game.p1Went and p == 0:\r\n text1 = font.render(move1, 1, (0,0,0))\r\n elif game.p1Went:\r\n text1 = font.render(\"Locked In\", 1, (0, 0, 0))\r\n else:\r\n text1 = font.render(\"Waiting...\", 1, (0, 0, 0))\r\n\r\n if game.p2Went and p == 1:\r\n text2 = font.render(move2, 1, (0,0,0))\r\n elif game.p2Went:\r\n text2 = font.render(\"Locked In\", 1, (0, 0, 0))\r\n else:\r\n text2 = font.render(\"Waiting...\", 1, (0, 0, 0))\r\n\r\n if p == 1:\r\n win.blit(text2, (100, 350))\r\n win.blit(text1, (400, 350))\r\n else:\r\n win.blit(text1, (100, 350))\r\n win.blit(text2, (400, 350))\r\n\r\n for btn in btns:\r\n btn.draw(win)\r\n\r\n pygame.display.update()\r\n\r\ndef onlineGame():\r\n run = True\r\n clock = pygame.time.Clock()\r\n n = Network()\r\n player = int(n.getP()) ## Nestrādā: TypeError: int() argument must be a string, a bytes-like object or a number, not 'NoneType'\r\n print(\"You are player\", player)\r\n\r\n ## Struct \r\n createStruct = True\r\n y = 50\r\n x = 150\r\n\r\n ## Guessing word\r\n #word = wordSource.getWord()\r\n word = \"KOKS\"\r\n inputWord = \"\"\r\n attemptCount = 0\r\n maxPoints = 0\r\n guessAttempt = False\r\n gameOver = False\r\n correctInput = False\r\n startCountdown = True\r\n\r\n gameDisplay.fill(_white)\r\n gameDisplay.blit(_canvas, (151,12), _input1_bckg)\r\n\r\n while run:\r\n clock.tick(60)\r\n try:\r\n game = n.send(\"get\")\r\n except:\r\n run = False\r\n print(\"Couldn't get game\")\r\n break\r\n\r\n if game.allEndGuessing():\r\n redrawWindow(gameDisplay, game, player)\r\n pygame.time.delay(500)\r\n try:\r\n game = n.send(\"reset\")\r\n except:\r\n run = False\r\n print(\"Couldn't get game\")\r\n break\r\n\r\n font = pygame.font.SysFont(\"comicsans\", 90)\r\n text = font.render(\"You Won!\", 1, (255,0,0))\r\n\r\n gameDisplay.blit(text, (width/2 - text.get_width()/2, height/2 - text.get_height()/2))\r\n pygame.display.update()\r\n pygame.time.delay(2000)\r\n\r\n for event in pygame.event.get():\r\n if event.type == pygame.QUIT:\r\n run = False\r\n pygame.quit()\r\n\r\n if gameOver:\r\n if player == 0:\r\n if not game.p1endGuessing:\r\n n.send(maxPoints)\r\n elif player == 1:\r\n if not game.p2endGuessing:\r\n n.send(maxPoints)\r\n else:\r\n if not game.p3endGuessing:\r\n n.send(maxPoints)\r\n\r\n redrawWindow(gameDisplay, game, player)\r\n\r\n\r\n## Local game\r\ndef guessingWordWindow():\r\n ## Struct \r\n createStruct = True\r\n y = 50\r\n x = 150\r\n\r\n ## Guessing word\r\n #word = wordSource.getWord()\r\n word = \"KOKS\"\r\n inputWord = \"\"\r\n attemptCount = 0\r\n maxPoint = 0\r\n guessAttempt = False\r\n gameOver = False\r\n correctInput = False\r\n startCountdown = True\r\n\r\n gameDisplay.fill(_white)\r\n gameDisplay.blit(_canvas, (151,12), _input1_bckg)\r\n\r\n while True:\r\n for event in pygame.event.get():\r\n if event.type == pygame.QUIT:\r\n pygame.quit()\r\n quit()\r\n # InputBox\r\n _inputBoxes[0].handle_event(event)\r\n gameDisplay.blit(_canvas, (152,12), _input1_bckg)\r\n if event.type == pygame.KEYDOWN:\r\n if _inputBoxes[0].active:\r\n if event.key == pygame.K_RETURN:\r\n inputWord = _inputBoxes[0].text\r\n _inputBoxes[0].text = ''\r\n if not gameOver:\r\n guessAttempt = True \r\n\r\n ## Generate guessing structure\r\n if createStruct:\r\n createStruct = False\r\n for n in range(5):\r\n l = 0\r\n while l < len(word):\r\n pygame.draw.rect(gameDisplay, _black,((x + 35*l)-2,(y+ 40*n)-2,34,34))\r\n pygame.draw.rect(gameDisplay, _white,((x + 35*l),(y+ 40*n),30,30))\r\n l += 1\r\n \r\n ## Attempt\r\n # Attempt input\r\n if not gameOver:\r\n _inputBoxes[0].draw(gameDisplay)\r\n \r\n # Attempt \r\n if guessAttempt:\r\n guessAttempt = False\r\n point = 0\r\n inputWord = inputWord.upper()\r\n \r\n # Input word length check\r\n if(len(inputWord) > len(word)):\r\n inputWord = inputWord[:len(word)]\r\n # Correct input\r\n if(word == inputWord):\r\n gameOver = True\r\n correctInput = True\r\n maxPoint = len(word)*2\r\n maxPoint += 5- attemptCount\r\n textSurf, textRect = text_objects(\"Your input is correct.\", _smallText, _black)\r\n textRect.center = ((_displayWidth+x)/2 , 270 )\r\n gameDisplay.blit(textSurf, textRect)\r\n\r\n l = 0\r\n for c in word:\r\n if( l < len(word)):\r\n pygame.draw.rect(gameDisplay, _green,((x + 35*l),(y + 40*attemptCount),30,30))\r\n\r\n textSurf, textRect = text_objects(c, _smallText, _black)\r\n textRect.center = ( ((x + 35*l)+(30/2)), ((y + 40*attemptCount)+(30/2)) )\r\n gameDisplay.blit(textSurf, textRect)\r\n\r\n l += 1\r\n\r\n # Incorrect input\r\n else:\r\n i = 0\r\n while i < len(inputWord):\r\n\r\n if(inputWord[i] == word[i]):\r\n pygame.draw.rect(gameDisplay, _green,((x + 35*i),(y + 40*attemptCount),30,30))\r\n point += 2\r\n elif(word.find(inputWord[i]) != -1):\r\n pygame.draw.rect(gameDisplay, _yellow,((x + 35*i),(y + 40*attemptCount),30,30))\r\n point += 1\r\n else:\r\n pygame.draw.rect(gameDisplay, _red,((x + 35*i),(y + 40*attemptCount),30,30))\r\n\r\n textSurf, textRect = text_objects(inputWord[i], _smallText, _black)\r\n textRect.center = ( ((x + 35*i)+(30/2)), ((y + 40*attemptCount)+(30/2)) )\r\n gameDisplay.blit(textSurf, textRect)\r\n\r\n i += 1\r\n \r\n empty = 0\r\n while(len(word) != len(inputWord) + empty):\r\n pygame.draw.rect(gameDisplay, _red,((x + 35*(empty + len(inputWord))),(y + 40*attemptCount),30,30))\r\n empty += 1\r\n \r\n # Point\r\n if maxPoint < point:\r\n maxPoint = point\r\n # Attempt\r\n attemptCount += 1\r\n\r\n # Dont guess after all attempt\r\n if attemptCount >= 5:\r\n gameOver = True \r\n if not correctInput:\r\n\r\n textSurf, textRect = text_objects(\"Your input isn't correct.\", _smallText, _black)\r\n textRect.center = ((_displayWidth+x)/2 , 270 )\r\n gameDisplay.blit(textSurf, textRect)\r\n\r\n i = 0\r\n while i < len(word):\r\n pygame.draw.rect(gameDisplay, _green,((x + 35*i), 360,30,30))\r\n\r\n textSurf, textRect = text_objects(word[i], _smallText, _black)\r\n textRect.center = ( ((x + 35*i)+(30/2)), (360 +(30/2)) )\r\n gameDisplay.blit(textSurf, textRect)\r\n\r\n i += 1\r\n\r\n textSurf, textRect = text_objects(\"Correct answer is:\", _smallText, _black)\r\n textRect.center = ( (_displayWidth+x)/2 , 330 )\r\n gameDisplay.blit(textSurf, textRect)\r\n\r\n if gameOver:\r\n textSurf, textRect = text_objects(\"You earned \" + str(maxPoint) + \" points\", _smallText, _black)\r\n textRect.center = ((_displayWidth+x)/2 , 300 )\r\n gameDisplay.blit(textSurf, textRect)\r\n\r\n button(\"Play Again\", 10, 10, _brightGreen, _green, guessingWordWindow)\r\n #button(\"Main\", 10, 240, _brightBlue, _blue, menuWindow)\r\n\r\n\r\n pygame.display.update()\r\n clock.tick(60)\r\n\r\n## Menu window\r\ndef menuWindow():\r\n\r\n ## Window title\r\n gameDisplay.fill(_white)\r\n TextSurf, TextRect = text_objects(\"Guess the Word\", _largeText, _black)\r\n TextRect.center = ((_displayWidth/2), (100))\r\n gameDisplay.blit(TextSurf, TextRect)\r\n\r\n while True:\r\n for event in pygame.event.get():\r\n if event.type == pygame.QUIT:\r\n quitGame()\r\n\r\n ## Create button (dp_w, dp_h, active_color, inactive_color, function)\r\n button(\"Local\", 30, 250, _brightGreen, _green, createLocal)\r\n button(\"Online\", 195, 250, _brightBlue, _blue, createServerGame)\r\n button(\"Quit\", 360, 250, _brightRed, _red, quitGame)\r\n\r\n pygame.display.update()\r\n clock.tick(60)\r\n\r\n\r\n# Programm start\r\nmenuWindow()\r\n", "sub_path": "client.py", "file_name": "client.py", "file_ext": "py", "file_size_in_byte": 12282, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "pygame.font.init", "line_number": 9, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 9, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 16, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 17, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 17, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 18, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 32, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 32, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 33, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 33, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 34, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 34, "usage_type": "attribute"}, {"api_name": "input.InputBox", "line_number": 42, "usage_type": "call"}, {"api_name": "pygame.Surface", "line_number": 44, "usage_type": "call"}, {"api_name": "pygame.Rect", "line_number": 46, "usage_type": "call"}, {"api_name": "pygame.mouse.get_pos", "line_number": 55, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 55, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pressed", "line_number": 56, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 56, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 58, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 58, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 63, "usage_type": "call"}, {"api_name": "pygame.draw.rect", "line_number": 65, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 65, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 89, "usage_type": "call"}, {"api_name": "pygame.font.SysFont", "line_number": 98, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 98, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 102, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 102, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 139, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 139, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 143, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 143, "usage_type": "attribute"}, {"api_name": "network.Network", "line_number": 144, "usage_type": "call"}, {"api_name": "pygame.time.delay", "line_number": 178, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 178, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 186, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 186, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 190, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 190, "usage_type": "attribute"}, {"api_name": "pygame.time.delay", "line_number": 191, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 191, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 193, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 193, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 194, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 196, "usage_type": "call"}, {"api_name": "pygame.event.get", "line_number": 234, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 234, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 235, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 236, "usage_type": "call"}, {"api_name": "pygame.KEYDOWN", "line_number": 241, "usage_type": "attribute"}, {"api_name": "pygame.K_RETURN", "line_number": 243, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 255, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 255, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 256, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 256, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 286, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 286, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 300, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 300, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 303, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 303, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 306, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 306, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 316, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 316, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 336, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 336, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 357, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 357, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 370, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 370, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 371, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 379, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 379, "usage_type": "attribute"}]} +{"seq_id": "270164195", "text": "#!/usr/bin/python3\n# coding=utf-8\n\n# -------------------------------------------------------------------------------\n# This file is part of Phobos, a Blender Add-On to edit robot models.\n# Copyright (C) 2020 University of Bremen & DFKI GmbH Robotics Innovation Center\n#\n# You should have received a copy of the 3-Clause BSD License in the LICENSE file.\n# If not, see .\n# -------------------------------------------------------------------------------\n\nimport json\nfrom datetime import datetime\nfrom phobos.blender.defs import version\nfrom phobos.blender.defs import repository\nfrom phobos.blender.utils import io as ioUtils\nfrom phobos.blender.utils.general import roundFloatsInDict\nfrom phobos.blender.phoboslog import log\n\n\ndef exportSMURFScene(entities, path):\n \"\"\"Exports an arranged scene into SMURFS. It will export only entities\n with a valid entity/name, and entity/type property.\n\n Args:\n selected_only(bool): If True only selected entities get exported.\n subfolder(bool): If True the models are exported into separate subfolders\n entities: \n path: \n\n Returns:\n\n \"\"\"\n log(\"Exporting scene to \" + path + '.smurfs', \"INFO\")\n with open(path + '.smurfs', 'w') as outputfile:\n sceneinfo = (\n \"# SMURF scene created at \"\n + path\n + \" \"\n + datetime.now().strftime(\"%Y%m%d_%H:%M\")\n + \"\\n\"\n )\n log(sceneinfo, \"INFO\")\n sceneinfo += \"# created with Phobos \" + version + \" - \" + repository + \"\\n\\n\"\n ioUtils.securepath(path)\n outputfile.write(sceneinfo)\n entitiesdict = roundFloatsInDict(\n {'entities': entities}, ioUtils.getExpSettings().decimalPlaces\n )\n outputfile.write(json.dumps(entitiesdict, indent=2))\n\n\n# registering import/export functions of types with Phobos\nscene_type_dict = {'smurfs': {'export': exportSMURFScene, 'extensions': ('smurfs',)}}\n", "sub_path": "phobos/blender/io/scenes/smurfs.py", "file_name": "smurfs.py", "file_ext": "py", "file_size_in_byte": 1970, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "phobos.blender.phoboslog.log", "line_number": 34, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 40, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 40, "usage_type": "name"}, {"api_name": "phobos.blender.phoboslog.log", "line_number": 43, "usage_type": "call"}, {"api_name": "phobos.blender.defs.version", "line_number": 44, "usage_type": "name"}, {"api_name": "phobos.blender.defs.repository", "line_number": 44, "usage_type": "name"}, {"api_name": "phobos.blender.utils.io.securepath", "line_number": 45, "usage_type": "call"}, {"api_name": "phobos.blender.utils.io", "line_number": 45, "usage_type": "name"}, {"api_name": "phobos.blender.utils.general.roundFloatsInDict", "line_number": 47, "usage_type": "call"}, {"api_name": "phobos.blender.utils.io.getExpSettings", "line_number": 48, "usage_type": "call"}, {"api_name": "phobos.blender.utils.io", "line_number": 48, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 50, "usage_type": "call"}]} +{"seq_id": "276097982", "text": "import matplotlib.pyplot as plt\nimport numpy as np\nfrom sklearn import linear_model\nimport math\n\n\ndata = \"\"\"2\t15.11\n4\t11.36\n6\t9.77\n8\t9.09\n10\t8.48\n15\t7.69\n20\t7.33\n25\t7.06\n30\t6.7\n40\t6.43\n50\t6.16\n60\t5.99\n70\t5.77\n80\t5.64\n90\t5.39\n110\t5.09\n130\t4.87\n150\t4.6\n160\t4.5\n170\t4.36\n180\t4.2\n\"\"\"\n\n\ndef load_data():\n v = data.splitlines()\n x = [v[i].split('\\t')[0] for i in range(len(v))]\n y = [v[i].split('\\t')[1] for i in range(len(v))]\n return x, y\n\n\n# part a\ndef a():\n regr = linear_model.LinearRegression()\n x, y = load_data()\n\n # make x, y as log x, log y\n print(x)\n x = [math.log(float(x[i])) for i in range(len(x))]\n print(x)\n y = [math.log(float(y[i])) for i in range(len(y))]\n\n # make x, y as np array\n x = np.array(x).reshape(-1, 1)\n y = np.array(y).reshape(-1, 1)\n regr.fit(x, y)\n\n plt.scatter(x.tolist(), y.tolist(), color='black')\n plt.plot(x.tolist(), regr.predict(x).tolist(), color='blue', linewidth=2)\n plt.title('Regression line in log-log coordinates')\n plt.xlabel('Log of Hours')\n plt.ylabel('Log of Sulfate')\n plt.show()\n\n\n# part b\ndef b():\n regr = linear_model.LinearRegression()\n x, y = load_data()\n\n # make x, y as log x, log y\n print(x)\n x = [math.log(float(x[i])) for i in range(len(x))]\n print(x)\n y = [math.log(float(y[i])) for i in range(len(y))]\n\n # make x, y as np array\n x = np.array(x).reshape(-1, 1)\n y = np.array(y).reshape(-1, 1)\n regr.fit(x, y)\n\n yy = regr.predict(x)\n\n # reverse log to plot original coordinate\n x = x.tolist()\n x = [math.exp(x[i][0]) for i in range(len(x))]\n\n y = y.tolist()\n y = [math.exp(y[i][0]) for i in range(len(y))]\n\n yy = yy.tolist()\n yy = [math.exp(yy[i][0]) for i in range(len(yy))]\n\n plt.scatter(x, y, color='black')\n plt.plot(x, yy, color='blue', linewidth=2)\n plt.title('Regression curve in original coordinates')\n plt.xlabel('Hours')\n plt.ylabel('Sulfate')\n plt.show()\n\n# part c\ndef c_1():\n regr = linear_model.LinearRegression()\n x, y = load_data()\n\n # make x, y as log x, log y\n print(x)\n x = [math.log(float(x[i])) for i in range(len(x))]\n print(x)\n y = [math.log(float(y[i])) for i in range(len(y))]\n\n # make x, y as np array\n x = np.array(x).reshape(-1, 1)\n y = np.array(y).reshape(-1, 1)\n regr.fit(x, y)\n yy = regr.predict(x)\n plt.title('Residual plot in log-log coordinates')\n plt.scatter(y.tolist(), (y - yy).tolist())\n plt.show()\n\ndef c_2():\n regr = linear_model.LinearRegression()\n x, y = load_data()\n\n # make x, y as log x, log y\n print(x)\n x = [math.log(float(x[i])) for i in range(len(x))]\n print(x)\n y = [math.log(float(y[i])) for i in range(len(y))]\n\n # make x, y as np array\n x = np.array(x).reshape(-1, 1)\n y = np.array(y).reshape(-1, 1)\n regr.fit(x, y)\n yy = regr.predict(x)\n\n y_original = [math.exp(y[i][0]) for i in range(len(y))]\n\n y_residuals = [math.exp(y[i][0] - yy[i][0]) for i in range(len(y))]\n\n plt.title('Residual plot in original coordinates')\n plt.scatter(y_original, y_residuals)\n plt.show()\n\n\nif __name__ == \"__main__\":\n a()\n b()\n c_1()\n c_2()\n\n\n\n", "sub_path": "CS498_AML_HM5/HW5_7-9.py", "file_name": "HW5_7-9.py", "file_ext": "py", "file_size_in_byte": 3181, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "sklearn.linear_model.LinearRegression", "line_number": 40, "usage_type": "call"}, {"api_name": "sklearn.linear_model", "line_number": 40, "usage_type": "name"}, {"api_name": "math.log", "line_number": 45, "usage_type": "call"}, {"api_name": "math.log", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 64, "usage_type": "call"}, {"api_name": "sklearn.linear_model", "line_number": 64, "usage_type": "name"}, {"api_name": "math.log", "line_number": 69, "usage_type": "call"}, {"api_name": "math.log", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 75, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 82, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 85, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 99, "usage_type": "call"}, {"api_name": "sklearn.linear_model", "line_number": 99, "usage_type": "name"}, {"api_name": "math.log", "line_number": 104, "usage_type": "call"}, {"api_name": "math.log", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 118, "usage_type": "call"}, {"api_name": "sklearn.linear_model", "line_number": 118, "usage_type": "name"}, {"api_name": "math.log", "line_number": 123, "usage_type": "call"}, {"api_name": "math.log", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 129, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 133, "usage_type": "call"}, {"api_name": "math.exp", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 137, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 139, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 139, "usage_type": "name"}]} +{"seq_id": "417816346", "text": "import numpy as np\n\n#모델 생성\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn.svm import SVC\nfrom sklearn.ensemble import VotingClassifier\n\nfrom sklearn.metrics import accuracy_score\n\n\nfrom sklearn.datasets import load_breast_cancer\nfrom sklearn.model_selection import train_test_split\n\ncancer_data = load_breast_cancer()\n\nX_data = cancer_data.data\ny_label = cancer_data.target\n\nX_training, X_testing, y_training, y_testing = train_test_split(X_data, y_label, test_size = 0.2, random_state = 0)\n\n#개별 ML 모델을 위한 Classifier 생성\nrf_clf = RandomForestClassifier(n_estimators = 100, random_state = 0)\nlr_clf = LogisticRegression(solver = 'liblinear', random_state = 0)\nsvm_clf = SVC(gamma = 'auto', probability = True, random_state = 0)\n\n# 개별 모델들을 학습\nrf_clf.fit(X_training , y_training) \nlr_clf.fit(X_training, y_training)\nsvm_clf.fit(X_training, y_training)\n\n#학습된 개별 모델들이 각자 반환하는 예측 데이터 셋을 생성하고 개별 모델의 정확도 측정\nrf_pred = rf_clf.predict(X_testing)\nlr_pred = lr_clf.predict(X_testing)\nsvm_pred = svm_clf.predict(X_testing)\n\nprint('Random Forest 정확도 : {0:.4f}'.format(accuracy_score(y_testing, rf_pred)))\nprint('Logistic Regression 정확도 : {0:.4f}'.format(accuracy_score(y_testing, lr_pred)))\nprint('SVM 정확도 : {0:.4f}'.format(accuracy_score(y_testing, svm_pred)))\n\n#투표 분류\nvoting_clf = VotingClassifier(\n\testimators = [('rf', rf_clf), ('lr', lr_clf), ('svm', svm_clf)],\n\tvoting='soft')\n\nvoting_clf.fit(X_training, y_training)\n\nvoting_pred = voting_clf.predict(X_testing)\n\nprint('CV 기반 앙상블 모델 정확도 : {0:.4f}'.format(accuracy_score(y_testing, voting_pred)))", "sub_path": "ClassFIT_AI/AI_Model/CV.py", "file_name": "CV.py", "file_ext": "py", "file_size_in_byte": 1767, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "sklearn.datasets.load_breast_cancer", "line_number": 15, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 20, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 23, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 24, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 25, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 37, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 38, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 39, "usage_type": "call"}, {"api_name": "sklearn.ensemble.VotingClassifier", "line_number": 42, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 50, "usage_type": "call"}]} +{"seq_id": "388693040", "text": "# Copyright 2019 The FastEstimator Authors. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ==============================================================================\nimport time\nfrom typing import Iterable, List, Set, Union\n\nimport numpy as np\n\nfrom fastestimator.backend.get_lr import get_lr\nfrom fastestimator.backend.to_number import to_number\nfrom fastestimator.summary.system import System\nfrom fastestimator.util.data import Data\nfrom fastestimator.util.traceability_util import traceable\nfrom fastestimator.util.util import parse_modes, to_list, to_set\n\n\n@traceable()\nclass Trace:\n \"\"\"Trace controls the training loop. Users can use the `Trace` base class to customize their own functionality.\n\n Traces are invoked by the fe.Estimator periodically as it runs. In addition to the current data dictionary, they are\n also given a pointer to the current `System` instance which allows access to more information as well as giving the\n ability to modify or even cancel training. The order of function invocations is as follows:\n\n ``` plot\n Training: Testing:\n\n on_begin on_begin\n | |\n on_epoch_begin (train) <------< on_epoch_begin (test) <------<\n | | | |\n on_batch_begin (train) <----< | on_batch_begin (test) <----< |\n | | | | | |\n on_batch_end (train) >-----^ | on_batch_end (test) >------^ |\n | ^ | |\n on_epoch_end (train) | on_epoch_end (test) >---------^\n | | |\n on_epoch_begin (eval) | on_end\n | ^\n on_batch_begin (eval) <----< |\n | | |\n on_batch_end (eval) >-----^ |\n | |\n on_epoch_end (eval) >----------^\n |\n on_end\n ```\n\n Args:\n inputs: A set of keys that this trace intends to read from the state dictionary as inputs.\n outputs: A set of keys that this trace intends to write into the system buffer.\n mode: What mode(s) to execute this Trace in. For example, \"train\", \"eval\", \"test\", or \"infer\". To execute\n regardless of mode, pass None. To execute in all modes except for a particular one, you can pass an argument\n like \"!infer\" or \"!train\".\n \"\"\"\n system: System\n inputs: List[str]\n outputs: List[str]\n mode: Set[str]\n\n def __init__(self,\n inputs: Union[None, str, Iterable[str]] = None,\n outputs: Union[None, str, Iterable[str]] = None,\n mode: Union[None, str, Iterable[str]] = None) -> None:\n self.inputs = to_list(inputs)\n self.outputs = to_list(outputs)\n self.mode = parse_modes(to_set(mode))\n\n def on_begin(self, data: Data) -> None:\n \"\"\"Runs once at the beginning of training or testing.\n\n Args:\n data: A dictionary through which traces can communicate with each other or write values for logging.\n \"\"\"\n pass\n\n def on_epoch_begin(self, data: Data) -> None:\n \"\"\"Runs at the beginning of each epoch.\n\n Args:\n data: A dictionary through which traces can communicate with each other or write values for logging.\n \"\"\"\n pass\n\n def on_batch_begin(self, data: Data) -> None:\n \"\"\"Runs at the beginning of each batch.\n\n Args:\n data: A dictionary through which traces can communicate with each other or write values for logging.\n \"\"\"\n pass\n\n def on_batch_end(self, data: Data) -> None:\n \"\"\"Runs at the end of each batch.\n\n Args:\n data: The current batch and prediction data, as well as any information written by prior `Traces`.\n \"\"\"\n pass\n\n def on_epoch_end(self, data: Data) -> None:\n \"\"\"Runs at the end of each epoch.\n\n Args:\n data: A dictionary through which traces can communicate with each other or write values for logging.\n \"\"\"\n pass\n\n def on_end(self, data: Data) -> None:\n \"\"\"Runs once at the end training.\n\n Args:\n data: A dictionary through which traces can communicate with each other or write values for logging.\n \"\"\"\n pass\n\n\n@traceable()\nclass TrainEssential(Trace):\n \"\"\"A trace to collect important information during training.\n\n Please don't add this trace into an estimator manually. FastEstimator will add it automatically.\n\n Args:\n monitor_names: Which keys from the data dictionary to monitor during training.\n \"\"\"\n def __init__(self, monitor_names: Set[str]) -> None:\n super().__init__(inputs=monitor_names, mode=\"train\", outputs=[\"steps/sec\", \"epoch_time\", \"total_time\"])\n self.elapse_times = []\n self.train_start = None\n self.epoch_start = None\n self.step_start = None\n\n def on_begin(self, data: Data) -> None:\n self.train_start = time.perf_counter()\n data.write_with_log(\"num_device\", self.system.num_devices)\n data.write_with_log(\"logging_interval\", self.system.log_steps)\n\n def on_epoch_begin(self, data: Data) -> None:\n if self.system.log_steps:\n self.epoch_start = time.perf_counter()\n self.step_start = time.perf_counter()\n\n def on_batch_end(self, data: Data) -> None:\n if self.system.log_steps and (self.system.global_step % self.system.log_steps == 0\n or self.system.global_step == 1):\n for key in self.inputs:\n if key in data:\n data.write_with_log(key, data[key])\n if self.system.global_step > 1:\n self.elapse_times.append(time.perf_counter() - self.step_start)\n data.write_with_log(\"steps/sec\", round(self.system.log_steps / np.sum(self.elapse_times), 2))\n self.elapse_times = []\n self.step_start = time.perf_counter()\n\n def on_epoch_end(self, data: Data) -> None:\n if self.system.log_steps:\n self.elapse_times.append(time.perf_counter() - self.step_start)\n data.write_with_log(\"epoch_time\", \"{} sec\".format(round(time.perf_counter() - self.epoch_start, 2)))\n\n def on_end(self, data: Data) -> None:\n data.write_with_log(\"total_time\", \"{} sec\".format(round(time.perf_counter() - self.train_start, 2)))\n for model in self.system.network.models:\n if hasattr(model, \"current_optimizer\"):\n data.write_with_log(model.model_name + \"_lr\", get_lr(model))\n\n\n@traceable()\nclass EvalEssential(Trace):\n \"\"\"A trace to collect important information during evaluation.\n\n Please don't add this trace into an estimator manually. FastEstimator will add it automatically.\n\n Args:\n monitor_names: Any keys which should be collected over the course of an eval epoch.\n \"\"\"\n def __init__(self, monitor_names: Set[str]) -> None:\n super().__init__(mode=\"eval\", inputs=monitor_names)\n self.eval_results = None\n\n def on_epoch_begin(self, data: Data) -> None:\n self.eval_results = None\n\n def on_batch_end(self, data: Data) -> None:\n if self.eval_results is None:\n self.eval_results = {key: [data[key]] for key in self.inputs if key in data}\n else:\n for key in self.inputs:\n if key in data:\n self.eval_results[key].append(data[key])\n\n def on_epoch_end(self, data: Data) -> None:\n for key, value_list in self.eval_results.items():\n data.write_with_log(key, np.mean(np.array(value_list), axis=0))\n\n\n@traceable()\nclass Logger(Trace):\n \"\"\"A Trace that prints log messages.\n\n Please don't add this trace into an estimator manually. FastEstimator will add it automatically.\n \"\"\"\n def __init__(self) -> None:\n super().__init__(inputs=\"*\")\n\n def on_begin(self, data: Data) -> None:\n if not self.system.mode == \"test\":\n start_step = 1 if not self.system.global_step else self.system.global_step\n self._print_message(\"FastEstimator-Start: step: {}; \".format(start_step), data)\n\n def on_batch_end(self, data: Data) -> None:\n if self.system.mode == \"train\" and self.system.log_steps and (\n self.system.global_step % self.system.log_steps == 0 or self.system.global_step == 1):\n self._print_message(\"FastEstimator-Train: step: {}; \".format(self.system.global_step), data)\n\n def on_epoch_end(self, data: Data) -> None:\n if self.system.mode == \"train\" and self.system.log_steps:\n self._print_message(\"FastEstimator-Train: step: {}; \".format(self.system.global_step), data, True)\n elif self.system.mode == \"eval\":\n self._print_message(\"FastEstimator-Eval: step: {}; \".format(self.system.global_step), data, True)\n elif self.system.mode == \"test\":\n self._print_message(\"FastEstimator-Test: step: {}; \".format(self.system.global_step), data, True)\n\n def on_end(self, data: Data) -> None:\n if not self.system.mode == \"test\":\n self._print_message(\"FastEstimator-Finish: step: {}; \".format(self.system.global_step), data)\n\n def _print_message(self, header: str, data: Data, log_epoch: bool = False) -> None:\n \"\"\"Print a log message to the screen, and record the `data` into the `system` summary.\n\n Args:\n header: The prefix for the log message.\n data: A collection of data to be recorded.\n log_epoch: Whether epoch information should be included in the log message.\n \"\"\"\n log_message = header\n if log_epoch:\n log_message += \"epoch: {}; \".format(self.system.epoch_idx)\n self.system.write_summary('epoch', self.system.epoch_idx)\n for key, val in data.read_logs().items():\n val = to_number(val)\n self.system.write_summary(key, val)\n if val.size > 1:\n log_message += \"\\n{}:\\n{};\".format(key, np.array2string(val, separator=','))\n else:\n log_message += \"{}: {}; \".format(key, str(val))\n print(log_message)\n", "sub_path": "fastestimator/trace/trace.py", "file_name": "trace.py", "file_ext": "py", "file_size_in_byte": 11124, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "fastestimator.summary.system.System", "line_number": 67, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 68, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 69, "usage_type": "name"}, {"api_name": "typing.Set", "line_number": 70, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 73, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 73, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 74, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 74, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 75, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 75, "usage_type": "name"}, {"api_name": "fastestimator.util.util.to_list", "line_number": 76, "usage_type": "call"}, {"api_name": "fastestimator.util.util.to_list", "line_number": 77, "usage_type": "call"}, {"api_name": "fastestimator.util.util.parse_modes", "line_number": 78, "usage_type": "call"}, {"api_name": "fastestimator.util.util.to_set", "line_number": 78, "usage_type": "call"}, {"api_name": "fastestimator.util.data.Data", "line_number": 80, "usage_type": "name"}, {"api_name": "fastestimator.util.data.Data", "line_number": 88, "usage_type": "name"}, {"api_name": "fastestimator.util.data.Data", "line_number": 96, "usage_type": "name"}, {"api_name": "fastestimator.util.data.Data", "line_number": 104, "usage_type": "name"}, {"api_name": "fastestimator.util.data.Data", "line_number": 112, "usage_type": "name"}, {"api_name": "fastestimator.util.data.Data", "line_number": 120, "usage_type": "name"}, {"api_name": "fastestimator.util.traceability_util.traceable", "line_number": 28, "usage_type": "call"}, {"api_name": "typing.Set", "line_number": 138, "usage_type": "name"}, {"api_name": "fastestimator.util.data.Data", "line_number": 145, "usage_type": "name"}, {"api_name": "time.perf_counter", "line_number": 146, "usage_type": "call"}, {"api_name": "fastestimator.util.data.Data", "line_number": 150, "usage_type": "name"}, {"api_name": "time.perf_counter", "line_number": 152, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 153, "usage_type": "call"}, {"api_name": "fastestimator.util.data.Data", "line_number": 155, "usage_type": "name"}, {"api_name": "time.perf_counter", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 163, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 165, "usage_type": "call"}, {"api_name": "fastestimator.util.data.Data", "line_number": 167, "usage_type": "name"}, {"api_name": "time.perf_counter", "line_number": 169, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 170, "usage_type": "call"}, {"api_name": "fastestimator.util.data.Data", "line_number": 172, "usage_type": "name"}, {"api_name": "time.perf_counter", "line_number": 173, "usage_type": "call"}, {"api_name": "fastestimator.backend.get_lr.get_lr", "line_number": 176, "usage_type": "call"}, {"api_name": "fastestimator.util.traceability_util.traceable", "line_number": 129, "usage_type": "call"}, {"api_name": "typing.Set", "line_number": 188, "usage_type": "name"}, {"api_name": "fastestimator.util.data.Data", "line_number": 192, "usage_type": "name"}, {"api_name": "fastestimator.util.data.Data", "line_number": 195, "usage_type": "name"}, {"api_name": "fastestimator.util.data.Data", "line_number": 203, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 205, "usage_type": "call"}, {"api_name": "fastestimator.util.traceability_util.traceable", "line_number": 179, "usage_type": "call"}, {"api_name": "fastestimator.util.data.Data", "line_number": 217, "usage_type": "name"}, {"api_name": "fastestimator.util.data.Data", "line_number": 222, "usage_type": "name"}, {"api_name": "fastestimator.util.data.Data", "line_number": 227, "usage_type": "name"}, {"api_name": "fastestimator.util.data.Data", "line_number": 235, "usage_type": "name"}, {"api_name": "fastestimator.util.data.Data", "line_number": 239, "usage_type": "name"}, {"api_name": "fastestimator.backend.to_number.to_number", "line_number": 252, "usage_type": "call"}, {"api_name": "numpy.array2string", "line_number": 255, "usage_type": "call"}, {"api_name": "fastestimator.util.traceability_util.traceable", "line_number": 208, "usage_type": "call"}]} +{"seq_id": "183583134", "text": "from movies_dataframes import movies_df, ratings_df, users_df\nimport pandas as pd\nimport numpy as np\nfrom collections import Counter\n\ndef find_liked_disliked(x):\n liked_films_id = x.query('Rating.max() == Rating').MovieID.to_numpy()\n disliked_films_id = x.query('Rating.min() == Rating').MovieID.to_numpy()\n return pd.Series([liked_films_id, disliked_films_id],\n index=['liked_films', 'disliked_films'])\n\n\ndef match_film_name(movie_id):\n return movies_df.loc[movie_id].Name\n\n\ndef take_popular_genres(reviews):\n genres_list = movies_df.loc[reviews.MovieID].Genre.str.split('|').sum()\n favourite = Counter(genres_list).most_common(1)[0]\n return '{genre} ({times})'.format(genre=favourite[0], times=favourite[1])\n\n\nif __name__ == \"__main__\":\n vectorized_match = np.vectorize(match_film_name)\n\n mostly_watched_genres = ratings_df.groupby('UserID').\\\n apply(take_popular_genres).rename('mostly watched genre (times watched)')\n\n liked_disliked_id = ratings_df.groupby('UserID').\\\n apply(find_liked_disliked)\n\n liked_disliked_films = liked_disliked_id.applymap(vectorized_match)\n\n users_favourites = users_df.merge(mostly_watched_genres, left_index=True,\n right_index=True)\n users_favourites = users_favourites.merge(liked_disliked_films, left_index=True,\n right_index=True)\n\n users_favourites.to_csv('users_favourites.csv')\n", "sub_path": "data_analysis/movies_analysis/users_favourites/create_users_favourites.py", "file_name": "create_users_favourites.py", "file_ext": "py", "file_size_in_byte": 1476, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pandas.Series", "line_number": 9, "usage_type": "call"}, {"api_name": "movies_dataframes.movies_df.loc", "line_number": 14, "usage_type": "attribute"}, {"api_name": "movies_dataframes.movies_df", "line_number": 14, "usage_type": "name"}, {"api_name": "movies_dataframes.movies_df.loc", "line_number": 18, "usage_type": "attribute"}, {"api_name": "movies_dataframes.movies_df", "line_number": 18, "usage_type": "name"}, {"api_name": "collections.Counter", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.vectorize", "line_number": 24, "usage_type": "call"}, {"api_name": "movies_dataframes.ratings_df.groupby", "line_number": 26, "usage_type": "call"}, {"api_name": "movies_dataframes.ratings_df", "line_number": 26, "usage_type": "name"}, {"api_name": "movies_dataframes.ratings_df.groupby", "line_number": 29, "usage_type": "call"}, {"api_name": "movies_dataframes.ratings_df", "line_number": 29, "usage_type": "name"}, {"api_name": "movies_dataframes.users_df.merge", "line_number": 34, "usage_type": "call"}, {"api_name": "movies_dataframes.users_df", "line_number": 34, "usage_type": "name"}]} +{"seq_id": "418519344", "text": "from flask import Flask, render_template, request\nfrom flask.views import MethodView\nfrom wtforms import Form, StringField, SubmitField\nfrom wtforms.fields.core import SelectField\nfrom calorie import Calorie\nfrom temperature import Temperature\n\napp = Flask(__name__)\n\n\"\"\"\nADD COMMENT!\n\"\"\"\n\n\nclass HomePage(MethodView):\n \n def get(self):\n return render_template('index.html')\n\n\nclass CaloriesFormPage(MethodView):\n \n \"\"\"\n ADD COMMENT!\n \"\"\"\n\n def get(self):\n calories_form = CaloriesForm()\n return render_template('calories_form_page.html',\n caloriesform=calories_form) \n \n\n def post(self):\n calories_form = CaloriesForm(request.form)\n\n temperature = Temperature(country=calories_form.country.data,\n city= calories_form.city.data)\n\n current_temp = temperature.get_temperature()\n if current_temp != 'Error':\n calorie = Calorie(weight=float(calories_form.weight.data), \n height=float(calories_form.height.data),\n age=float(calories_form.age.data), \n sex=calories_form.sex.data,\n temperature=current_temp)\n\n calories_kcal, calorie_kJ = calorie.calculate()\n\n return render_template('calories_form_page.html',\n caloriesform=calories_form,\n city=temperature.city.upper(),\n country=temperature.country.upper(),\n current_temperature=current_temp, \n calories=calories_kcal,\n calories2=calorie_kJ,\n result=True)\n else:\n return render_template('calories_form_page.html',\n caloriesform=calories_form,\n city=temperature.city.upper(),\n country=temperature.country.upper(),\n current_temperature=False, \n result=True)\n\n\nclass CaloriesForm(Form):\n\n \"\"\"\n ADD COMMENT!\n \"\"\"\n \n # Widgets\n city = StringField(\"City: \", default=\"Ljubljana\")\n country = StringField(\"Country: \", default=\"Slovenia\")\n weight = StringField(\"Weight: \", default=70)\n height = StringField(\"Height: \", default=175)\n age = StringField(\"Age: \", default=30)\n #sex = StringField(\"Gender: \", default=\"Woman\")\n sex = SelectField(\"Gender: \", choices=[\"man\", \"woman\"]) \n button = SubmitField(\"Calculate\")\n\n\n#app.add_url_rule('/', view_func=HomePage.as_view('home_page'))\n#app.add_url_rule('/calories_form', view_func=CaloriesFormPage.as_view('calories_form_page'))\napp.add_url_rule('/', view_func=CaloriesFormPage.as_view('calories_form_page'))\n\napp.run(debug=True)", "sub_path": "flask_app.py", "file_name": "flask_app.py", "file_ext": "py", "file_size_in_byte": 2985, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "flask.Flask", "line_number": 8, "usage_type": "call"}, {"api_name": "flask.views.MethodView", "line_number": 15, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.views.MethodView", "line_number": 21, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 34, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 34, "usage_type": "name"}, {"api_name": "temperature.Temperature", "line_number": 36, "usage_type": "call"}, {"api_name": "temperature.get_temperature", "line_number": 39, "usage_type": "call"}, {"api_name": "calorie.Calorie", "line_number": 41, "usage_type": "call"}, {"api_name": "calorie.calculate", "line_number": 47, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 49, "usage_type": "call"}, {"api_name": "temperature.city.upper", "line_number": 51, "usage_type": "call"}, {"api_name": "temperature.city", "line_number": 51, "usage_type": "attribute"}, {"api_name": "temperature.country.upper", "line_number": 52, "usage_type": "call"}, {"api_name": "temperature.country", "line_number": 52, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 58, "usage_type": "call"}, {"api_name": "temperature.city.upper", "line_number": 60, "usage_type": "call"}, {"api_name": "temperature.city", "line_number": 60, "usage_type": "attribute"}, {"api_name": "temperature.country.upper", "line_number": 61, "usage_type": "call"}, {"api_name": "temperature.country", "line_number": 61, "usage_type": "attribute"}, {"api_name": "wtforms.Form", "line_number": 66, "usage_type": "name"}, {"api_name": "wtforms.StringField", "line_number": 73, "usage_type": "call"}, {"api_name": "wtforms.StringField", "line_number": 74, "usage_type": "call"}, {"api_name": "wtforms.StringField", "line_number": 75, "usage_type": "call"}, {"api_name": "wtforms.StringField", "line_number": 76, "usage_type": "call"}, {"api_name": "wtforms.StringField", "line_number": 77, "usage_type": "call"}, {"api_name": "wtforms.fields.core.SelectField", "line_number": 79, "usage_type": "call"}, {"api_name": "wtforms.SubmitField", "line_number": 80, "usage_type": "call"}]} +{"seq_id": "219174317", "text": "# Copyright (c) 2015 Uber Technologies, Inc.\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\n# THE SOFTWARE.\n\n_tornado_supported = False\n_stack_context_supported = False\n\nimport opentracing\ntry:\n import tornado # noqa\n _tornado_supported = True\n from opentracing.scope_managers.tornado import TornadoScopeManager\n from opentracing.scope_managers.tornado import tracer_stack_context\n _stack_context_supported = True\nexcept ImportError:\n pass\n\n\ndef is_tornado_supported():\n return _tornado_supported\n\n\ndef is_stack_context_supported():\n return _stack_context_supported\n\n\nclass _TracerEnteredStackContext(object):\n \"\"\"\n An entered tracer_stack_context() object.\n\n Intended to have a ready-to-use context where\n Span objects can be activated before the context\n itself is returned to the user.\n \"\"\"\n\n def __init__(self, context):\n self._context = context\n self._deactivation_cb = context.__enter__()\n\n def __enter__(self):\n return self._deactivation_cb\n\n def __exit__(self, type, value, traceback):\n return self._context.__exit__(type, value, traceback)\n\n\ndef span_in_stack_context(span):\n \"\"\"\n Create Tornado's (4.x, 5.x) StackContext that stores the given span in the\n thread-local request context. This function is intended for use\n in Tornado applications based on IOLoop, although will work fine\n in single-threaded apps like Flask, albeit with more overhead.\n\n StackContext has been deprecated in Tornado 6 and higher.\n Because of asyncio nature of Tornado 6.x, consider using\n `span_in_context` with opentracing scope manager `ContextVarScopeManager`\n\n ## Usage example in Tornado application\n\n Suppose you have a method `handle_request(request)` in the http server.\n Instead of calling it directly, use a wrapper:\n\n .. code-block:: python\n\n from opentracing_instrumentation import request_context\n\n @tornado.gen.coroutine\n def handle_request_wrapper(request, actual_handler, *args, **kwargs)\n\n request_wrapper = TornadoRequestWrapper(request=request)\n span = http_server.before_request(request=request_wrapper)\n\n with request_context.span_in_stack_context(span):\n return actual_handler(*args, **kwargs)\n\n :param span:\n :return:\n Return StackContext that wraps the request context.\n \"\"\"\n if not _tornado_supported:\n raise RuntimeError('span_in_stack_context requires Tornado')\n\n if not is_stack_context_supported():\n raise RuntimeError('tornado.stack_context is not supported in '\n 'Tornado >= 6.x')\n if not isinstance(\n opentracing.tracer.scope_manager, TornadoScopeManager\n ):\n raise RuntimeError('scope_manager is not TornadoScopeManager')\n\n # Enter the newly created stack context so we have\n # storage available for Span activation.\n context = tracer_stack_context()\n entered_context = _TracerEnteredStackContext(context)\n\n if span is None:\n return entered_context\n\n opentracing.tracer.scope_manager.activate(span, False)\n assert opentracing.tracer.active_span is not None\n assert opentracing.tracer.active_span is span\n\n return entered_context\n", "sub_path": "opentracing_instrumentation/tornado_context.py", "file_name": "tornado_context.py", "file_ext": "py", "file_size_in_byte": 4251, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "opentracing.scope_managers.tornado.TornadoScopeManager", "line_number": 103, "usage_type": "argument"}, {"api_name": "opentracing.tracer", "line_number": 103, "usage_type": "attribute"}, {"api_name": "opentracing.scope_managers.tornado.tracer_stack_context", "line_number": 109, "usage_type": "call"}, {"api_name": "opentracing.tracer.scope_manager.activate", "line_number": 115, "usage_type": "call"}, {"api_name": "opentracing.tracer", "line_number": 115, "usage_type": "attribute"}, {"api_name": "opentracing.tracer", "line_number": 116, "usage_type": "attribute"}, {"api_name": "opentracing.tracer", "line_number": 117, "usage_type": "attribute"}]} +{"seq_id": "484743422", "text": "import json\r\nimport os\r\n\r\nfrom cloudshell.api.cloudshell_api import CloudShellAPISession, InputNameValue\r\n\r\n#print 'install_device_farm_app called: ' + str(os.environ)\r\n\r\nresource = json.loads(os.environ['RESOURCECONTEXT'])\r\nreservation = json.loads(os.environ['RESERVATIONCONTEXT'])\r\nconnectivity = json.loads(os.environ['QUALICONNECTIVITYCONTEXT'])\r\n\r\nresid = reservation['id']\r\n\r\n\r\n# service = resource['appData']['name']\r\n\r\n\r\napi = CloudShellAPISession(host=connectivity['serverAddress'],\r\n token_id=connectivity['adminAuthToken'],\r\n domain=reservation['domain'])\r\n\r\napi.WriteMessageToReservationOutput(resid, 'install_device_farm_app script called: ' + str(os.environ))\r\n# Temporary implementation bypassing the installation service:\r\n# try:\r\n# cp_resource = [x['value']\r\n# for x in resource['appData']['installationService']['attributes']\r\n# if x['name'] == 'AWS EC2'][0]\r\n# deploy_inputs = [InputNameValue(x['name'].lower().replace(' ', '_'), x['value'])\r\n# for x in resource['appData']['installationService']['attributes']\r\n# if x['name'] != 'AWS EC2']\r\n# except:\r\n# cp_resource = resource['attributes']['AWS EC2']\r\n# deploy_inputs = [InputNameValue(name.lower().replace(' ', '_'), value)\r\n# for name, value in resource['attributes'].iteritems()\r\n# if name != 'AWS EC2']\r\n\r\n# try:\r\n# result = api.ExecuteCommand(resid, cp_resource, \"Resource\", \"upload_app\", deploy_inputs)\r\n# except:\r\n\r\ndeployed_app_name = resource['deployedAppData']['name']\r\n\r\nresult = api.ExecuteResourceConnectedCommand(resid, deployed_app_name, \"upload_app_connected\", \"remote_app_management\", [\r\n resource['attributes']['APK URL'],\r\n resource['attributes']['APK Asset Updates'],\r\n])\r\n\r\n# Version that calls the installation service from this script\r\n# For this to work, update datamodel.xml on model \"AWS Mobile Device Installation\":\r\n# Change to SupportsConcurrentCommands=\"true\"\r\n\r\n# installation_result = api.InstallApp(reservationId=resid,\r\n# resourceName=service,\r\n# commandName='Install',\r\n# commandInputs=[],\r\n# printOutput=True)\r\n", "sub_path": "Environment/AWS Shell/Resource Scripts/install_device_farm_app.py", "file_name": "install_device_farm_app.py", "file_ext": "py", "file_size_in_byte": 2367, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "json.loads", "line_number": 8, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 8, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 9, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 9, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 10, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 10, "usage_type": "attribute"}, {"api_name": "cloudshell.api.cloudshell_api.CloudShellAPISession", "line_number": 18, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 22, "usage_type": "attribute"}]} +{"seq_id": "515638922", "text": "\"\"\"\nUtility functions to export models to the Models dataset and get information about models currently being served\nin the project.\n\"\"\"\n\nfrom hops import hdfs, constants, util, exceptions, kafka\nfrom hops.experiment_impl.util import experiment_utils\nimport os\nimport json\nimport re\n\n\ndef exists(serving_name):\n \"\"\"\n Checks if there exists a serving with the given name\n\n Example use-case:\n\n >>> from hops import serving\n >>> serving.exist(serving_name)\n\n Args:\n :serving_name: the name of the serving\n\n Returns:\n True if the serving exists, otherwise false\n \"\"\"\n try:\n return get_id(serving_name) is not None\n except ServingNotFound as e:\n print(\"No serving with name {} was found in the project {}\".format(serving_name, hdfs.project_name()))\n return False\n\n\ndef delete(serving_name):\n \"\"\"\n Deletes serving instance with a given name\n\n Example use-case:\n\n >>> from hops import serving\n >>> serving.delete(\"irisFlowerClassifier\")\n\n Args:\n :serving_name: name of the serving to delete\n\n Returns:\n None\n \"\"\"\n serving_id = get_id(serving_name)\n print(\"Deleting serving with name: {}...\".format(serving_name))\n _delete_serving_rest(serving_id)\n print(\"Serving with name: {} successfully deleted\".format(serving_name))\n\n\ndef _delete_serving_rest(serving_id):\n \"\"\"\n Makes a REST request to Hopsworks REST API for deleting a serving instance\n\n Args:\n :serving_id: id of the serving to delete\n\n Returns:\n None\n\n Raises:\n :RestAPIError: if there was an error with the REST call to Hopsworks\n \"\"\"\n method = constants.HTTP_CONFIG.HTTP_DELETE\n resource_url = (constants.DELIMITERS.SLASH_DELIMITER +\n constants.REST_CONFIG.HOPSWORKS_REST_RESOURCE + constants.DELIMITERS.SLASH_DELIMITER +\n constants.REST_CONFIG.HOPSWORKS_PROJECT_RESOURCE + constants.DELIMITERS.SLASH_DELIMITER +\n hdfs.project_id() + constants.DELIMITERS.SLASH_DELIMITER +\n constants.REST_CONFIG.HOPSWORKS_SERVING_RESOURCE + constants.DELIMITERS.SLASH_DELIMITER\n + str(serving_id))\n response = util.send_request(method, resource_url)\n\n if response.status_code != 200:\n response_object = response.json()\n error_code, error_msg, user_msg = util._parse_rest_error(response_object)\n raise exceptions.RestAPIError(\"Could not delete serving with id {} (url: {}), \"\n \"server response: \\n \"\n \"HTTP code: {}, HTTP reason: {}, error code: {}, error msg: {}, \"\n \"user msg: {}\".format(serving_id, resource_url, response.status_code,\n response.reason, error_code, error_msg, user_msg))\n\n\ndef start(serving_name):\n \"\"\"\n Starts a model serving instance with a given name\n\n Example use-case:\n\n >>> from hops import serving\n >>> serving.start(\"irisFlowerClassifier\")\n\n Args:\n :serving_name: name of the serving to start\n\n Returns:\n None\n \"\"\"\n serving_id = get_id(serving_name)\n print(\"Starting serving with name: {}...\".format(serving_name))\n _start_or_stop_serving_rest(serving_id, constants.MODEL_SERVING.SERVING_ACTION_START)\n print(\"Serving with name: {} successfully started\".format(serving_name))\n\n\ndef stop(serving_name):\n \"\"\"\n Stops a model serving instance with a given name\n\n Example use-case:\n\n >>> from hops import serving\n >>> serving.stop(\"irisFlowerClassifier\")\n\n Args:\n :serving_name: name of the serving to stop\n\n Returns:\n None\n \"\"\"\n serving_id = get_id(serving_name)\n print(\"Stopping serving with name: {}...\".format(serving_name))\n _start_or_stop_serving_rest(serving_id, constants.MODEL_SERVING.SERVING_ACTION_STOP)\n print(\"Serving with name: {} successfully stopped\".format(serving_name))\n\n\ndef _start_or_stop_serving_rest(serving_id, action):\n \"\"\"\n Makes a REST request to Hopsworks REST API for starting/stopping a serving instance\n\n Args:\n :serving_id: id of the serving to start/stop\n :action: the action to perform (start or stop)\n\n Returns:\n None\n\n Raises:\n :RestAPIError: if there was an error with the REST call to Hopsworks\n \"\"\"\n method = constants.HTTP_CONFIG.HTTP_POST\n resource_url = (constants.DELIMITERS.SLASH_DELIMITER +\n constants.REST_CONFIG.HOPSWORKS_REST_RESOURCE + constants.DELIMITERS.SLASH_DELIMITER +\n constants.REST_CONFIG.HOPSWORKS_PROJECT_RESOURCE + constants.DELIMITERS.SLASH_DELIMITER +\n hdfs.project_id() + constants.DELIMITERS.SLASH_DELIMITER +\n constants.REST_CONFIG.HOPSWORKS_SERVING_RESOURCE + constants.DELIMITERS.SLASH_DELIMITER\n + str(serving_id) + constants.MODEL_SERVING.SERVING_START_OR_STOP_PATH_PARAM + action)\n response = util.send_request(method, resource_url)\n\n if response.status_code != 200:\n response_object = response.json()\n error_code, error_msg, user_msg = util._parse_rest_error(response_object)\n raise exceptions.RestAPIError(\"Could not perform action {} on serving with id {} (url: {}), \"\n \"server response: \\n \"\n \"HTTP code: {}, HTTP reason: {}, error code: {}, error msg: {}, \"\n \"user msg: {}\".format(action, serving_id, resource_url, response.status_code,\n response.reason, error_code, error_msg, user_msg))\n\n\ndef create_or_update(artifact_path, serving_name, serving_type=\"TENSORFLOW\", model_version=1,\n batching_enabled = False, topic_name=\"CREATE\", num_partitions = 1, num_replicas = 1,\n instances = 1):\n \"\"\"\n Creates a serving in Hopsworks if it does not exist, otherwise update the existing one.\n\n Example use-case:\n\n >>> from hops import serving\n >>> serving.create_or_update(\"/Models/mnist\", \"mnist\", \"TENSORFLOW\", 1)\n\n Args:\n :artifact_path: path to the artifact to serve (tf model dir or sklearn script)\n :serving_name: name of the serving to create\n :serving_type: type of the serving, e.g \"TENSORFLOW\" or \"SKLEARN\"\n :model_version: version of the model to serve\n :batching_enabled: boolean flag whether to enable batching for the inference requests\n :instances: the number of serving instances (the more instances the more inference requests can\n be served in parallel)\n\n Returns:\n None\n \"\"\"\n serving_id = get_id(serving_name)\n artifact_path = hdfs._expand_path(artifact_path)\n _validate_user_serving_input(artifact_path, serving_name, serving_type, model_version, batching_enabled,\n num_partitions, num_replicas, instances)\n artifact_path = hdfs.get_plain_path(artifact_path)\n print(\"Creating a serving for model {} ...\".format(serving_name))\n _create_or_update_serving_rest(artifact_path, serving_name, serving_type, model_version, batching_enabled,\n topic_name, num_partitions, num_replicas, serving_id, instances)\n print(\"Serving for model {} successfully created\".format(serving_name))\n\n\ndef _validate_user_serving_input(model_path, model_name, serving_type, model_version, batching_enabled,\n num_partitions, num_replicas, instances):\n \"\"\"\n Validate user input on the client side before sending REST call to Hopsworks (additional validation will be done\n in the backend)\n\n Args:\n :model_path: path to the model or artifact being served\n :model_name: the name of the serving to create\n :serving_type: the type of serving\n :model_version: version of the serving\n :batching_enabled: boolean flag whether to enable batching for inference requests to the serving\n :num_partitions: kafka partitions\n :num_replicas: kafka replicas\n :instances: the number of serving instances (the more instances the more inference requests can\n be served in parallel)\n\n Returns:\n None\n\n Raises:\n :ValueError: if the serving input failed the validation\n \"\"\"\n name_pattern = re.compile(\"^[a-zA-Z0-9]+$\")\n if len(model_name) > 256 or model_name == \"\" or not name_pattern.match(model_name):\n raise ValueError(\"Name of serving cannot be empty, cannot exceed 256 characters and must match the regular \"\n \"expression: ^[a-zA-Z0-9]+$, the provided name: {} is not valid\".format(model_name))\n if not hdfs.exists(model_path):\n raise ValueError(\"The model/artifact path must exist in HDFS, the provided path: {} \"\n \"does not exist\".format(model_path))\n if serving_type not in constants.MODEL_SERVING.SERVING_TYPES:\n raise ValueError(\"The provided serving_type: {} is not supported, supported \"\n \"serving types are: {}\".format(serving_type, \",\".join(constants.MODEL_SERVING.SERVING_TYPES)))\n if not isinstance(model_version, int):\n raise ValueError(\"The model version must be an integer, the provided version is not: {}\".format(model_version))\n if serving_type == constants.MODEL_SERVING.SERVING_TYPE_TENSORFLOW:\n if not isinstance(num_replicas, int):\n raise ValueError(\"Number of kafka topic replicas must be an integer, the provided num replicas \"\n \"is not: {}\".format(model_version))\n if not isinstance(num_partitions, int):\n raise ValueError(\"Number of kafka topic partitions must be an integer, the provided num partitions \"\n \"is not: {}\".format(num_partitions))\n if not isinstance(batching_enabled, bool):\n raise ValueError(\"Batching enabled must be a boolean, the provided value \"\n \"is not: {}\".format(batching_enabled))\n if not isinstance(instances, int):\n raise ValueError(\"The number of serving instances must be an integer, \"\n \"the provided version is not: {}\".format(instances))\n\n\ndef _create_or_update_serving_rest(model_path, model_name, serving_type, model_version,\n batching_enabled = None, topic_name=None, num_partitions = None,\n num_replicas = None, serving_id = None, instances=1):\n \"\"\"\n Makes a REST request to Hopsworks for creating or updating a model serving instance\n\n Args:\n :model_path: path to the model or artifact being served\n :model_name: the name of the serving to create\n :serving_type: the type of serving\n :model_version: version of the serving\n :batching_enabled: boolean flag whether to enable batching for inference requests to the serving\n :topic_name: name of the kafka topic (\"CREATE\" to create a new one, or \"NONE\" to not use kafka topic)\n :num_partitions: kafka partitions\n :num_replicas: kafka replicas\n :serving_id: the id of the serving in case of UPDATE, if serving_id is None, it is a CREATE operation.\n :instances: the number of serving instances (the more instances the more inference requests can\n be served in parallel)\n\n Returns:\n None\n\n Raises:\n :RestAPIError: if there was an error with the REST call to Hopsworks\n \"\"\"\n json_contents = {\n constants.REST_CONFIG.JSON_SERVING_MODEL_VERSION: model_version,\n constants.REST_CONFIG.JSON_SERVING_ARTIFACT_PATH: model_path,\n constants.REST_CONFIG.JSON_SERVING_TYPE: serving_type,\n constants.REST_CONFIG.JSON_SERVING_NAME: model_name,\n constants.REST_CONFIG.JSON_SERVING_KAFKA_TOPIC_DTO: {\n constants.REST_CONFIG.JSON_KAFKA_TOPIC_NAME: topic_name,\n constants.REST_CONFIG.JSON_KAFKA_NUM_PARTITIONS: num_partitions,\n constants.REST_CONFIG.JSON_KAFKA_NUM_REPLICAS: num_replicas\n },\n constants.REST_CONFIG.JSON_SERVING_REQUESTED_INSTANCES: instances,\n }\n if serving_id is not None:\n json_contents[constants.REST_CONFIG.JSON_SERVING_ID] = serving_id\n if serving_type == constants.MODEL_SERVING.SERVING_TYPE_TENSORFLOW:\n json_contents[constants.REST_CONFIG.JSON_SERVING_BATCHING_ENABLED] = batching_enabled\n json_embeddable = json.dumps(json_contents)\n headers = {constants.HTTP_CONFIG.HTTP_CONTENT_TYPE: constants.HTTP_CONFIG.HTTP_APPLICATION_JSON}\n method = constants.HTTP_CONFIG.HTTP_PUT\n resource_url = (constants.DELIMITERS.SLASH_DELIMITER +\n constants.REST_CONFIG.HOPSWORKS_REST_RESOURCE + constants.DELIMITERS.SLASH_DELIMITER +\n constants.REST_CONFIG.HOPSWORKS_PROJECT_RESOURCE + constants.DELIMITERS.SLASH_DELIMITER +\n hdfs.project_id() + constants.DELIMITERS.SLASH_DELIMITER +\n constants.REST_CONFIG.HOPSWORKS_SERVING_RESOURCE + constants.DELIMITERS.SLASH_DELIMITER)\n response = util.send_request(method, resource_url, data=json_embeddable, headers=headers)\n\n if response.status_code != 201 and response.status_code != 200:\n response_object = response.json()\n error_code, error_msg, user_msg = util._parse_rest_error(response_object)\n raise exceptions.RestAPIError(\"Could not create or update serving (url: {}), server response: \\n \" \\\n \"HTTP code: {}, HTTP reason: {}, error code: {}, error msg: {}, \"\n \"user msg: {}\".format(resource_url, response.status_code, response.reason,\n error_code, error_msg, user_msg))\n\ndef get_id(serving_name):\n \"\"\"\n Gets the id of a serving with a given name\n\n Example use-case:\n\n >>> from hops import serving\n >>> serving.get_id(serving_name)\n\n Args:\n :serving_name: name of the serving to get the id for\n\n Returns:\n the id of the serving, None if Serving does not exist\n \"\"\"\n try:\n servings = get_all()\n serving = _find_serving_with_name(serving_name, servings)\n return serving.id\n except ServingNotFound:\n return None\n\n\ndef get_artifact_path(serving_name):\n \"\"\"\n Gets the artifact path of a serving with a given name\n\n Example use-case:\n\n >>> from hops import serving\n >>> serving.get_artifact_path(serving_name)\n\n Args:\n :serving_name: name of the serving to get the artifact path for\n\n Returns:\n the artifact path of the serving (model path in case of tensorflow, or python script in case of SkLearn)\n \"\"\"\n servings = get_all()\n serving = _find_serving_with_name(serving_name, servings)\n return serving.artifact_path\n\n\ndef get_type(serving_name):\n \"\"\"\n Gets the type of a serving with a given name\n\n Example use-case:\n\n >>> from hops import serving\n >>> serving.get_type(serving_name)\n\n Args:\n :serving_name: name of the serving to get the typ for\n\n Returns:\n the type of the serving (e.g Tensorflow or SkLearn)\n \"\"\"\n servings = get_all()\n serving = _find_serving_with_name(serving_name, servings)\n return serving.serving_type\n\n\ndef get_version(serving_name):\n \"\"\"\n Gets the version of a serving with a given name\n\n Example use-case:\n\n >>> from hops import serving\n >>> serving.get_version(serving_name)\n\n Args:\n :serving_name: name of the serving to get the version for\n\n Returns:\n the version of the serving\n \"\"\"\n servings = get_all()\n serving = _find_serving_with_name(serving_name, servings)\n return serving.model_version\n\n\ndef get_kafka_topic(serving_name):\n \"\"\"\n Gets the kafka topic name of a serving with a given name\n\n Example use-case:\n\n >>> from hops import serving\n >>> serving.get_kafka_topic(serving_name)\n\n Args:\n :serving_name: name of the serving to get the kafka topic name for\n\n Returns:\n the kafka topic name of the serving\n \"\"\"\n servings = get_all()\n serving = _find_serving_with_name(serving_name, servings)\n return serving.kafka_topic_dto.name\n\n\ndef get_status(serving_name):\n \"\"\"\n Gets the status of a serving with a given name\n\n Example use-case:\n\n >>> from hops import serving\n >>> serving.get_status(serving_name)\n\n Args:\n :serving_name: name of the serving to get the status for\n\n Returns:\n the status of the serving\n \"\"\"\n servings = get_all()\n serving = _find_serving_with_name(serving_name, servings)\n return serving.status\n\n\ndef get_all():\n \"\"\"\n Gets the list of servings for the current project\n\n Example:\n\n >>> from hops import serving\n >>> servings = serving.get_all()\n >>> servings[0].name\n\n Returns:\n list of servings\n \"\"\"\n return _parse_json_servings(_get_servings_rest())\n\n\ndef _find_serving_with_name(serving_name, servings):\n \"\"\"\n Finds a serving with a given name from a list of servings (O(N))\n\n Args:\n :serving_name: name of the serving to look for\n :servings: the list of servings to look through\n\n Returns:\n serving with the given name\n\n Raises:\n :ServingNotFound: if the requested serving could not be found\n \"\"\"\n serving_names = []\n for serving in servings:\n if serving.name == serving_name:\n return serving\n serving_names.append(serving.name)\n serving_names_str = \",\".join(serving_names)\n raise ServingNotFound(\"No serving with name: {} could be found among the list of \"\n \"available servings: {}\".format(serving_name, serving_names_str))\n\n\ndef _parse_json_servings(json_servings):\n \"\"\"\n Parses a list of JSON servings into Serving Objects\n\n Args:\n :json_servings: the list of JSON servings\n\n Returns:\n a list of Serving Objects\n \"\"\"\n return list(map(lambda json_serving: Serving(json_serving), json_servings))\n\n\ndef _get_servings_rest():\n \"\"\"\n Makes a REST request to Hopsworks to get a list of all servings in the current project\n\n Returns:\n JSON response parsed as a python dict\n\n Raises:\n :RestAPIError: if there was an error with the REST call to Hopsworks\n \"\"\"\n method = constants.HTTP_CONFIG.HTTP_GET\n resource_url = (constants.DELIMITERS.SLASH_DELIMITER +\n constants.REST_CONFIG.HOPSWORKS_REST_RESOURCE + constants.DELIMITERS.SLASH_DELIMITER +\n constants.REST_CONFIG.HOPSWORKS_PROJECT_RESOURCE + constants.DELIMITERS.SLASH_DELIMITER +\n hdfs.project_id() + constants.DELIMITERS.SLASH_DELIMITER +\n constants.REST_CONFIG.HOPSWORKS_SERVING_RESOURCE + constants.DELIMITERS.SLASH_DELIMITER)\n response = util.send_request(method, resource_url)\n response_object = response.json()\n if response.status_code != 200:\n error_code, error_msg, user_msg = util._parse_rest_error(response_object)\n raise exceptions.RestAPIError(\"Could not fetch list of servings from Hopsworks REST API (url: {}), \"\n \"server response: \\n \"\n \"HTTP code: {}, HTTP reason: {}, error code: {}, \"\n \"error msg: {}, user msg: {}\".format(\n resource_url, response.status_code, response.reason, error_code, error_msg, user_msg))\n return response_object\n\n\ndef make_inference_request(serving_name, data, verb=\":predict\"):\n \"\"\"\n Submit an inference request\n\n Example use-case:\n\n >>> from hops import serving\n >>> serving.make_inference_request(\"irisFlowerClassifier\", [[1,2,3,4]], \":predict\")\n\n Args:\n :serving_name: name of the model being served\n :data: data/json to send to the serving\n :verb: type of request (:predict, :classify, or :regress)\n\n Returns:\n the JSON response\n \"\"\"\n return _make_inference_request_rest(serving_name, data, verb)\n\ndef _make_inference_request_rest(serving_name, data, verb):\n \"\"\"\n Makes a REST request to Hopsworks for submitting an inference request to the serving instance\n\n Args:\n :serving_name: name of the model being served\n :data: data/json to send to the serving\n :verb: type of request (:predict, :classify, or :regress)\n\n Returns:\n the JSON response\n\n Raises:\n :RestAPIError: if there was an error with the REST call to Hopsworks\n \"\"\"\n json_embeddable = json.dumps(data)\n headers = {constants.HTTP_CONFIG.HTTP_CONTENT_TYPE: constants.HTTP_CONFIG.HTTP_APPLICATION_JSON}\n method = constants.HTTP_CONFIG.HTTP_POST\n resource_url = (constants.DELIMITERS.SLASH_DELIMITER +\n constants.REST_CONFIG.HOPSWORKS_REST_RESOURCE + constants.DELIMITERS.SLASH_DELIMITER +\n constants.REST_CONFIG.HOPSWORKS_PROJECT_RESOURCE + constants.DELIMITERS.SLASH_DELIMITER +\n hdfs.project_id() + constants.DELIMITERS.SLASH_DELIMITER +\n constants.REST_CONFIG.HOPSWORKS_INFERENCE_RESOURCE + constants.DELIMITERS.SLASH_DELIMITER +\n constants.REST_CONFIG.HOPSWORKS_MODELS_RESOURCE + constants.DELIMITERS.SLASH_DELIMITER\n + serving_name + verb)\n response = util.send_request(method, resource_url, data=json_embeddable, headers=headers)\n response_object = response.json()\n error_code, error_msg, user_msg = util._parse_rest_error(response_object)\n\n if response.status_code != 201 and response.status_code != 200:\n raise exceptions.RestAPIError(\"Could not create or update serving (url: {}), server response: \\n \" \\\n \"HTTP code: {}, HTTP reason: {}, error code: {}, error msg: {}, \"\n \"user msg: {}\".format(resource_url, response.status_code, response.reason,\n error_code, error_msg, user_msg))\n return response_object\n\nclass Serving(object):\n \"\"\"\n Represents a model being served in Hopsworks\n \"\"\"\n\n def __init__(self, serving_json):\n \"\"\"\n Initialize the serving from JSON payload returned by Hopsworks REST API\n\n Args:\n :feature_json: JSON data about the feature returned from Hopsworks REST API\n \"\"\"\n self.status = serving_json[constants.REST_CONFIG.JSON_SERVING_STATUS]\n self.artifact_path = serving_json[constants.REST_CONFIG.JSON_SERVING_ARTIFACT_PATH]\n self.name = serving_json[constants.REST_CONFIG.JSON_SERVING_NAME]\n self.creator = serving_json[constants.REST_CONFIG.JSON_SERVING_CREATOR]\n self.creator = serving_json[constants.REST_CONFIG.JSON_SERVING_CREATOR]\n self.serving_type = serving_json[constants.REST_CONFIG.JSON_SERVING_TYPE]\n self.model_version = serving_json[constants.REST_CONFIG.JSON_SERVING_MODEL_VERSION]\n self.created = serving_json[constants.REST_CONFIG.JSON_SERVING_CREATED]\n self.requested_instances = serving_json[constants.REST_CONFIG.JSON_SERVING_REQUESTED_INSTANCES]\n if constants.REST_CONFIG.JSON_SERVING_KAFKA_TOPIC_DTO in serving_json:\n self.kafka_topic_dto = kafka.KafkaTopicDTO(serving_json[constants.REST_CONFIG.JSON_SERVING_KAFKA_TOPIC_DTO])\n self.id = serving_json[constants.REST_CONFIG.JSON_SERVING_ID]\n\n\nclass ServingNotFound(Exception):\n \"\"\"This exception will be raised if the requested serving could not be found\"\"\"\n", "sub_path": "hops/serving.py", "file_name": "serving.py", "file_ext": "py", "file_size_in_byte": 23687, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "hops.hdfs.project_name", "line_number": 31, "usage_type": "call"}, {"api_name": "hops.hdfs", "line_number": 31, "usage_type": "name"}, {"api_name": "hops.constants.HTTP_CONFIG", "line_number": 69, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 69, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 70, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 70, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 71, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 71, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 71, "usage_type": "attribute"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 72, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 72, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 72, "usage_type": "attribute"}, {"api_name": "hops.hdfs.project_id", "line_number": 73, "usage_type": "call"}, {"api_name": "hops.hdfs", "line_number": 73, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 73, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 73, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 74, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 74, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 74, "usage_type": "attribute"}, {"api_name": "hops.util.send_request", "line_number": 76, "usage_type": "call"}, {"api_name": "hops.util", "line_number": 76, "usage_type": "name"}, {"api_name": "hops.util._parse_rest_error", "line_number": 80, "usage_type": "call"}, {"api_name": "hops.util", "line_number": 80, "usage_type": "name"}, {"api_name": "hops.exceptions.RestAPIError", "line_number": 81, "usage_type": "call"}, {"api_name": "hops.exceptions", "line_number": 81, "usage_type": "name"}, {"api_name": "hops.constants.MODEL_SERVING", "line_number": 105, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 105, "usage_type": "name"}, {"api_name": "hops.constants.MODEL_SERVING", "line_number": 126, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 126, "usage_type": "name"}, {"api_name": "hops.constants.HTTP_CONFIG", "line_number": 144, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 144, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 145, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 145, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 146, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 146, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 146, "usage_type": "attribute"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 147, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 147, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 147, "usage_type": "attribute"}, {"api_name": "hops.hdfs.project_id", "line_number": 148, "usage_type": "call"}, {"api_name": "hops.hdfs", "line_number": 148, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 148, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 148, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 149, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 149, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 149, "usage_type": "attribute"}, {"api_name": "hops.constants.MODEL_SERVING", "line_number": 150, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 150, "usage_type": "name"}, {"api_name": "hops.util.send_request", "line_number": 151, "usage_type": "call"}, {"api_name": "hops.util", "line_number": 151, "usage_type": "name"}, {"api_name": "hops.util._parse_rest_error", "line_number": 155, "usage_type": "call"}, {"api_name": "hops.util", "line_number": 155, "usage_type": "name"}, {"api_name": "hops.exceptions.RestAPIError", "line_number": 156, "usage_type": "call"}, {"api_name": "hops.exceptions", "line_number": 156, "usage_type": "name"}, {"api_name": "hops.hdfs._expand_path", "line_number": 187, "usage_type": "call"}, {"api_name": "hops.hdfs", "line_number": 187, "usage_type": "name"}, {"api_name": "hops.hdfs.get_plain_path", "line_number": 190, "usage_type": "call"}, {"api_name": "hops.hdfs", "line_number": 190, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 220, "usage_type": "call"}, {"api_name": "hops.hdfs.exists", "line_number": 224, "usage_type": "call"}, {"api_name": "hops.hdfs", "line_number": 224, "usage_type": "name"}, {"api_name": "hops.constants.MODEL_SERVING", "line_number": 227, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 227, "usage_type": "name"}, {"api_name": "hops.constants.MODEL_SERVING", "line_number": 229, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 229, "usage_type": "name"}, {"api_name": "hops.constants.MODEL_SERVING", "line_number": 232, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 232, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 273, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 273, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 274, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 274, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 275, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 275, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 276, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 276, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 277, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 277, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 282, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 282, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 278, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 278, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 279, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 279, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 280, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 280, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 285, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 285, "usage_type": "name"}, {"api_name": "hops.constants.MODEL_SERVING", "line_number": 286, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 286, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 287, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 287, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 288, "usage_type": "call"}, {"api_name": "hops.constants.HTTP_CONFIG", "line_number": 289, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 289, "usage_type": "name"}, {"api_name": "hops.constants.HTTP_CONFIG", "line_number": 290, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 290, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 291, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 291, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 292, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 292, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 292, "usage_type": "attribute"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 293, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 293, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 293, "usage_type": "attribute"}, {"api_name": "hops.hdfs.project_id", "line_number": 294, "usage_type": "call"}, {"api_name": "hops.hdfs", "line_number": 294, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 294, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 294, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 295, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 295, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 295, "usage_type": "attribute"}, {"api_name": "hops.util.send_request", "line_number": 296, "usage_type": "call"}, {"api_name": "hops.util", "line_number": 296, "usage_type": "name"}, {"api_name": "hops.util._parse_rest_error", "line_number": 300, "usage_type": "call"}, {"api_name": "hops.util", "line_number": 300, "usage_type": "name"}, {"api_name": "hops.exceptions.RestAPIError", "line_number": 301, "usage_type": "call"}, {"api_name": "hops.exceptions", "line_number": 301, "usage_type": "name"}, {"api_name": "hops.constants.HTTP_CONFIG", "line_number": 492, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 492, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 493, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 493, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 494, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 494, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 494, "usage_type": "attribute"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 495, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 495, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 495, "usage_type": "attribute"}, {"api_name": "hops.hdfs.project_id", "line_number": 496, "usage_type": "call"}, {"api_name": "hops.hdfs", "line_number": 496, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 496, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 496, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 497, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 497, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 497, "usage_type": "attribute"}, {"api_name": "hops.util.send_request", "line_number": 498, "usage_type": "call"}, {"api_name": "hops.util", "line_number": 498, "usage_type": "name"}, {"api_name": "hops.util._parse_rest_error", "line_number": 501, "usage_type": "call"}, {"api_name": "hops.util", "line_number": 501, "usage_type": "name"}, {"api_name": "hops.exceptions.RestAPIError", "line_number": 502, "usage_type": "call"}, {"api_name": "hops.exceptions", "line_number": 502, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 544, "usage_type": "call"}, {"api_name": "hops.constants.HTTP_CONFIG", "line_number": 545, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 545, "usage_type": "name"}, {"api_name": "hops.constants.HTTP_CONFIG", "line_number": 546, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 546, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 547, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 547, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 548, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 548, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 548, "usage_type": "attribute"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 549, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 549, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 549, "usage_type": "attribute"}, {"api_name": "hops.hdfs.project_id", "line_number": 550, "usage_type": "call"}, {"api_name": "hops.hdfs", "line_number": 550, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 550, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 550, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 551, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 551, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 551, "usage_type": "attribute"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 552, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 552, "usage_type": "name"}, {"api_name": "hops.constants.DELIMITERS", "line_number": 552, "usage_type": "attribute"}, {"api_name": "hops.util.send_request", "line_number": 554, "usage_type": "call"}, {"api_name": "hops.util", "line_number": 554, "usage_type": "name"}, {"api_name": "hops.util._parse_rest_error", "line_number": 556, "usage_type": "call"}, {"api_name": "hops.util", "line_number": 556, "usage_type": "name"}, {"api_name": "hops.exceptions.RestAPIError", "line_number": 559, "usage_type": "call"}, {"api_name": "hops.exceptions", "line_number": 559, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 577, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 577, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 578, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 578, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 579, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 579, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 580, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 580, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 581, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 581, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 582, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 582, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 583, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 583, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 584, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 584, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 585, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 585, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 586, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 586, "usage_type": "name"}, {"api_name": "hops.kafka.KafkaTopicDTO", "line_number": 587, "usage_type": "call"}, {"api_name": "hops.kafka", "line_number": 587, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 587, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 587, "usage_type": "name"}, {"api_name": "hops.constants.REST_CONFIG", "line_number": 588, "usage_type": "attribute"}, {"api_name": "hops.constants", "line_number": 588, "usage_type": "name"}]} +{"seq_id": "396694993", "text": "from PyQt5.QtWidgets import QWidget, QApplication\nfrom PyQt5.QtCore import QPoint, Qt\nfrom PyQt5.QtGui import QPixmap\nfrom Readerinfo_ui import Ui_Form\nimport sys\n\nclass ReaderinfoUi(QWidget, Ui_Form):\n def __init__(self, info, parent=None):\n super().__init__()\n self.info = info\n self.parent = parent\n self.setupUi(self)\n self._drag = False\n self.m_DragPosition = QPoint()\n self.init_ui()\n\n def init_ui(self):\n self.setFixedSize(331, 252)\n self.setWindowOpacity(0.9)\n self.setAttribute(Qt.WA_TranslucentBackground)\n self.setWindowFlag(Qt.FramelessWindowHint)\n\n self.id_label.setText(self.info[0])\n self.name_label.setText(self.info[1])\n self.sex_label.setText(self.info[2])\n self.dept_label.setText(self.info[4])\n self.grade_label.setText(self.info[3])\n\n self.setupSignal()\n\n \n self.setStyleSheet(\n '''QWidget#widget_2{\n background:gray;\n border-top:1px solid white;\n border-bottom:1px solid white;\n border-left:1px solid white;\n border-top-left-radius:10px;\n border-bottom-left-radius:10px;\n }\n QWidget#widget_3{\n color:#232C51;\n background:white;\n border-top:1px solid darkGray;\n border-bottom:1px solid darkGray;\n border-right:1px solid darkGray;\n border-top-right-radius:10px;\n border-bottom-right-radius:10px;\n }\n QPushButton{\n background-color: #4CAF50; \n border: none;\n color: white;\n border-radius:5px;\n padding: 5px 15px;\n text-align: center;\n text-decoration: none;\n display: inline-block;\n font-size: 16px;\n }\n QPushButton:hover{\n background:green;}\n ''')\n self.close_bt.setStyleSheet('''QPushButton{background:#F76677;border-radius:5px;}QPushButton:hover{background:red;}''')\n self.none_bt.setStyleSheet('''QPushButton{background:#F7D674;border-radius:5px;}QPushButton:hover{background:yellow;}''')\n self.min_bt.setStyleSheet('''QPushButton{background:#6DDF6D;border-radius:5px;}QPushButton:hover{background:green;}''')\n\n pix = QPixmap('../image/bg4.jpg')\n self.pic_label.setPixmap(pix)\n self.pic_label.setScaledContents(True)\n\n def setupSignal(self):\n self.close_bt.clicked.connect(self.close)\n self.min_bt.clicked.connect(self.showMinimized)\n\n\n def mousePressEvent(self, event):\n if event.button()== Qt.LeftButton:\n self.m_drag=True\n self.m_DragPosition=event.globalPos()-self.pos()\n event.accept()\n\n def mouseMoveEvent(self, QMouseEvent):\n if QMouseEvent.buttons() and Qt.LeftButton:\n self.move(QMouseEvent.globalPos()-self.m_DragPosition)\n QMouseEvent.accept()\n\n def mouseReleaseEvent(self, QMouseEvent):\n self.m_drag=False\n\nif __name__ == '__main__':\n app = QApplication(sys.argv)\n gui = ReaderinfoUi(None)\n gui.show()\n sys.exit(app.exec_())", "sub_path": "code/readerinfo_window.py", "file_name": "readerinfo_window.py", "file_ext": "py", "file_size_in_byte": 3247, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 7, "usage_type": "name"}, {"api_name": "Readerinfo_ui.Ui_Form", "line_number": 7, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QPoint", "line_number": 14, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt.WA_TranslucentBackground", "line_number": 20, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 20, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.FramelessWindowHint", "line_number": 21, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 21, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 68, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt.LeftButton", "line_number": 78, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 78, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.LeftButton", "line_number": 84, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 84, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 92, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 92, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 95, "usage_type": "call"}]} +{"seq_id": "471573305", "text": "# -*- coding: utf-8 -*-\n\n\n# 去掉函数中参数值是None的参数\nimport functools\n\n\ndef filter_none_param(func):\n @functools.wraps(func)\n def _wrapper(*args, **kwargs):\n params = dict()\n for key in kwargs:\n if kwargs[key] is not None:\n params[key] = kwargs[key]\n return func(*args, **params)\n return _wrapper\n", "sub_path": "common_app/decorator.py", "file_name": "decorator.py", "file_ext": "py", "file_size_in_byte": 369, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "functools.wraps", "line_number": 9, "usage_type": "call"}]} +{"seq_id": "288251408", "text": "import collections\n\nfrom django.db import migrations, transaction\n\n\ndef update_report_fields(apps, schema_editor):\n ComplianceReport = apps.get_model('api', 'compliancereport')\n for report in ComplianceReport.objects.filter(supplements__isnull=False, latest_report__isnull=True):\n with transaction.atomic():\n ancestor = report\n root = None\n latest = None\n while ancestor.supplements is not None:\n ancestor = ancestor.supplements\n\n visited = []\n id_traversal = {}\n to_visit = collections.deque([ancestor.id])\n i = 0\n\n while len(to_visit) > 0:\n current_id = to_visit.popleft()\n\n # break loops\n if current_id in visited:\n continue\n visited.append(current_id)\n\n current = ComplianceReport.objects.get(id=current_id)\n\n if current.supplements is None:\n root = current\n latest = current\n # don't count non-supplement reports (really should just be the root)\n if current.supplements is not None and \\\n not current.status.fuel_supplier_status_id == \"Deleted\":\n latest = current\n i += 1\n id_traversal[current_id] = i\n for descendant in current.supplemental_reports.order_by('create_timestamp').all():\n to_visit.append(descendant.id)\n\n for compliance_id, traversal in id_traversal.items():\n ComplianceReport.objects.filter(id=int(compliance_id)) \\\n .update(latest_report=latest, root_report=root, traversal=traversal)\n for report in ComplianceReport.objects.filter(supplements__isnull=True, latest_report__isnull=True):\n ComplianceReport.objects.filter(id=report.id) \\\n .update(latest_report=report, root_report=report)\n\n\nclass Migration(migrations.Migration):\n dependencies = [\n ('api', '0003_auto_20230526_1452'),\n ]\n\n operations = [\n migrations.RunPython(update_report_fields, reverse_code=migrations.RunPython.noop),\n ]\n", "sub_path": "backend/api/migrations/0004_migrate_compliance_report.py", "file_name": "0004_migrate_compliance_report.py", "file_ext": "py", "file_size_in_byte": 2219, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "django.db.transaction.atomic", "line_number": 9, "usage_type": "call"}, {"api_name": "django.db.transaction", "line_number": 9, "usage_type": "name"}, {"api_name": "collections.deque", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.migrations.Migration", "line_number": 51, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 51, "usage_type": "name"}, {"api_name": "django.db.migrations.RunPython", "line_number": 57, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 57, "usage_type": "name"}]} +{"seq_id": "3068377", "text": "# coding=utf-8\nimport re\nimport json\nfrom pychroner import PluginMeta, PluginType\n\n@PluginMeta(PluginType.TwitterReply, twitterAccount=\"SlashNephy\")\ndef do(pluginApi, stream):\n if stream[\"user\"][\"screen_name\"] == \"SlashNephy\":\n api = pluginApi.getTwitterAccount().getHandler()\n\n m = re.match(\"^@{} makepoll (.+?)$\".format(stream[\"user\"][\"screen_name\"]), stream[\"text\"], re.IGNORECASE)\n if m:\n try:\n t = json.loads(m.group(1))\n api.create_poll(t[\"text\"], t[\"choices\"], minutes=t.get(\"minutes\", 1440))\n\n except:\n api.statuses_update(status=f\"@{stream['user']['screen_name']} JSONが不正です\")\n", "sub_path": "plugins/SlashNephy/MakePoll.py", "file_name": "MakePoll.py", "file_ext": "py", "file_size_in_byte": 686, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "re.match", "line_number": 11, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 11, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 14, "usage_type": "call"}, {"api_name": "pychroner.PluginMeta", "line_number": 6, "usage_type": "call"}, {"api_name": "pychroner.PluginType.TwitterReply", "line_number": 6, "usage_type": "attribute"}, {"api_name": "pychroner.PluginType", "line_number": 6, "usage_type": "name"}]} +{"seq_id": "366018240", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Mon May 14 19:38:34 2018\r\n\r\nload a trained model an use it on preprocessed data\r\n\r\n@author: Wolfgang Kapferer\r\n\"\"\"\r\nfrom keras.models import model_from_json\r\nimport numpy as np\r\n\r\ndef load_model_and_apply(preprocessed):\r\n\r\n resultForDatabase=[]\r\n \r\n model=np.float32(preprocessed)\r\n \r\n input_model=model[:,1:len(model[0])-1]\r\n \r\n pk = model[:,0]\r\n resultForDatabase.append(pk)\r\n \r\n # load json and create model\r\n json_file = open('model.json', 'r')\r\n loaded_model_json = json_file.read()\r\n json_file.close()\r\n loaded_model = model_from_json(loaded_model_json)\r\n # load weights into new model\r\n loaded_model.load_weights(\"model.h5\")\r\n print(\"Loaded model from disk\")\r\n \r\n prediction = loaded_model.predict(input_model)\r\n resultForDatabase.append(prediction)\r\n \r\n return resultForDatabase\r\n", "sub_path": "keras_load_model_and_apply.py", "file_name": "keras_load_model_and_apply.py", "file_ext": "py", "file_size_in_byte": 896, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "numpy.float32", "line_number": 16, "usage_type": "call"}, {"api_name": "keras.models.model_from_json", "line_number": 27, "usage_type": "call"}]} +{"seq_id": "418306641", "text": "#!/usr/bin/python\n__author__ = 'james harrison'\n\nimport json\nimport requests\nimport xml.etree.ElementTree as ET\n\n\n\nclass Helper(object):\n def __init__(self):\n self.uni_to_home = \"2110\"\n self.home_to_uni = \"1262\"\n self.ROUTE = \"P\"\n\n self.base_url = \"http://rtt.metroinfo.org.nz/rtt/public/utility/file.aspx?\"\n self.payload = {\"contenttype\": \"SQLXML\",\n \"Name\": \"JPRoutePositionET.xml\"}\n\n def get_data(self, platform):\n ''' Fetches the xml data for a given platform'''\n self.payload[\"PlatformTag\"] = platform\n return requests.get(self.base_url, params=self.payload).text[3:]\n\n def get_bus_times(self, xml):\n ''' Get the bus times that matches self.ROUTE from the xml data '''\n times = []\n bus_line = []\n root = ET.fromstring(xml)\n\n for route in root[1]:\n if route.attrib[\"RouteNo\"] == self.ROUTE:\n bus_line = route\n\n for bus in bus_line:\n for trip in bus:\n times.append(int(trip.attrib[\"ETA\"]))\n return times\n\n def print_times(self, times):\n if len(times) > 0:\n for time in sorted(times):\n print(str(time) + \" minutes\")\n else:\n print(\"Sorry no buses going that way :(\")\n\n def main(self):\n home_to_uni_data = self.get_data(self.home_to_uni)\n home_to_uni_times = self.get_bus_times(home_to_uni_data)\n uni_to_home_data = self.get_data(self.uni_to_home)\n uni_to_home_times = self.get_bus_times(uni_to_home_data)\n print(\"------------------------------------------------------\")\n print(\"Home to Uni:\")\n self.print_times(home_to_uni_times)\n print(\"Uni to Home:\")\n self.print_times(uni_to_home_times)\n print(\"------------------------------------------------------\")\n\nif __name__ == \"__main__\":\n helper = Helper()\n helper.main()", "sub_path": "bus.py", "file_name": "bus.py", "file_ext": "py", "file_size_in_byte": 1938, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "requests.get", "line_number": 23, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.fromstring", "line_number": 29, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 29, "usage_type": "argument"}]} +{"seq_id": "635457615", "text": "# -*- coding: utf-8 -*-\nimport json\nimport logging\n\nfrom django.conf import settings\nfrom django.template.loader import render_to_string\nfrom django.utils.crypto import get_random_string\nfrom django.utils.translation import ugettext_lazy as _\nfrom rest_framework import response, viewsets\nfrom rest_framework.renderers import BrowsableAPIRenderer\n\nfrom backend.accounts import bcs_perm\nfrom backend.bcs_web.audit_log import client\nfrom backend.components import paas_cc\nfrom backend.uniapps.network.clb import constants as clb_constants\nfrom backend.uniapps.network.clb import serializers\nfrom backend.uniapps.network.clb import utils as clb_utils\nfrom backend.uniapps.network.clb.models import CloudLoadBlancer\nfrom backend.utils.error_codes import error_codes\nfrom backend.utils.renderers import BKAPIRenderer\n\n\nclass DescribeCLBNamesViewSet(viewsets.ViewSet):\n renderer_classes = (BKAPIRenderer, BrowsableAPIRenderer)\n\n def list(self, request, project_id):\n data = clb_utils.describe_clb_detail(\n request.user.token.access_token,\n request.user.username,\n request.project.cc_app_id,\n request.query_params.get('region'),\n )\n return response.Response(data.keys())\n\n\nclass CLBListCreateViewSet(viewsets.ViewSet):\n renderer_classes = (BKAPIRenderer, BrowsableAPIRenderer)\n\n def add_status_and_cluster_name(self, request, project_id, data):\n # 只有已经启动的clb,才能查询到状态\n cluster_id_list = [info['cluster_id'] for info in data]\n clb_dict = {}\n for cluster_id in cluster_id_list:\n clb_dict.update(\n clb_utils.get_deployments(\n request.user.token.access_token, project_id, request.project.kind, cluster_id\n )\n )\n # add status and cluster_name to data\n cluster_id_names = clb_utils.get_cluster_id_names_map(request.user.token.access_token, project_id)\n for info in data:\n filter_key = (info['cluster_id'], info['resource_name'])\n info['cluster_name'] = cluster_id_names.get(info['cluster_id']) or ''\n if filter_key in clb_dict:\n info.update(clb_dict[filter_key])\n return data\n\n def list(self, request, project_id):\n cluster_id = request.query_params.get(\"cluster_id\")\n data = CloudLoadBlancer.objects.get_clb_list(project_id, cluster_id=cluster_id)\n data = self.add_status_and_cluster_name(request, project_id, data)\n\n # 添加权限\n data = bcs_perm.Cluster.hook_perms(request, project_id, data)\n\n return response.Response(data)\n\n def get_vpc_id(self, request, region, clb_name):\n data = clb_utils.describe_clb_detail(\n request.user.token.access_token, request.user.username, request.project.cc_app_id, region\n )\n if clb_name not in data:\n raise error_codes.CheckFailed(f'clb:[{clb_name}] not found')\n return data[clb_name]['vpc_id']\n\n def create(self, request, project_id):\n slz = serializers.CreateCLBSLZ(data=request.data)\n slz.is_valid(raise_exception=True)\n data = slz.validated_data\n # 校验集群权限\n clb_utils.can_use_cluster(request, project_id, data['cluster_id'])\n\n # 通过clb_name渲染deployment\n # 替换clb中的'_'为'-', 后缀使用6位随机字符,并且长度限制为253以内,以满足后台的限制\n # resource_name包含: clb_name[:246] + '-' + random(6)\n replaced_name = data['clb_name'].replace('_', '-')\n data['resource_name'] = f'{replaced_name[:246]}-{get_random_string(6).lower()}'\n data['creator'] = data['updator'] = request.user.username\n data['project_id'] = request.project.project_id\n data['vpc_id'] = self.get_vpc_id(request, data['region'], data['clb_name'])\n\n # 创建并返回记录\n with client.ContextActivityLogClient(\n project_id=project_id,\n user=request.user.username,\n resource_type='lb',\n resource=data['clb_name'],\n description=_(\"集群:{}, 创建云lb controler\").format(data['cluster_id']),\n ).log_add():\n record = CloudLoadBlancer.objects.create(data)\n data = CloudLoadBlancer.objects.parse_record(record)\n\n return response.Response(data)\n\n\nclass MesosCLBOperateViewSet(viewsets.ViewSet):\n renderer_classes = (BKAPIRenderer, BrowsableAPIRenderer)\n\n def update_clb_status(self, clb_id, status):\n CloudLoadBlancer.objects.filter(id=clb_id).update(status=status)\n\n def post(self, request, project_id, clb_id):\n # 获取配置\n record = CloudLoadBlancer.objects.retrieve_record(clb_id)\n # 校验使用集群权限\n clb_utils.can_use_cluster(request, project_id, record['cluster_id'])\n # 获取 repo 地址\n repo_domain = paas_cc.get_jfrog_domain(request.user.token.access_token, project_id, record['cluster_id'])\n if not repo_domain:\n repo_domain = settings.DEFAUT_MESOS_LB_JFROG_DOMAIN\n record['repo_domain'] = repo_domain\n mesos_json = json.loads(render_to_string('mesos.json', record))\n\n with client.ContextActivityLogClient(\n project_id=project_id,\n user=request.user.username,\n resource_type='lb',\n resource=record['resource_name'],\n description=_(\"集群:{}, 创建clb关联deployment:{}\").format(record['cluster_id'], record['resource_name']),\n ).log_add():\n clb_utils.create_mesos_deployment(\n request.user.token.access_token, project_id, record['cluster_id'], record['namespace'], mesos_json\n )\n # 更改状态\n self.update_clb_status(clb_id, clb_constants.CLB_CREATED_STATUS)\n\n return response.Response()\n\n def delete(self, request, project_id, clb_id):\n record = CloudLoadBlancer.objects.retrieve_record(clb_id)\n # 校验使用集群权限\n clb_utils.can_use_cluster(request, project_id, record['cluster_id'])\n\n with client.ContextActivityLogClient(\n project_id=project_id,\n user=request.user.username,\n resource_type='lb',\n resource=record['resource_name'],\n description=_(\"集群:{}, 删除clb关联deployment:{}\").format(record['cluster_id'], record['resource_name']),\n ).log_delete():\n clb_utils.delete_mesos_deployment(\n request.user.token.access_token,\n project_id,\n record['cluster_id'],\n record['namespace'],\n record['resource_name'],\n )\n # 更新状态\n self.update_clb_status(clb_id, clb_constants.CLB_DELETED_STATUS)\n\n return response.Response()\n\n\nclass CLBRetrieveOperateViewSet(viewsets.ViewSet):\n renderer_classes = (BKAPIRenderer, BrowsableAPIRenderer)\n\n def add_status(self, request, data):\n if data['status'] != clb_constants.CLB_CREATED_STATUS:\n return data\n deployment_status = clb_utils.get_deployments(\n request.user.token.access_token,\n data['project_id'],\n request.project.kind,\n data['cluster_id'],\n name=data['resource_name'],\n )\n data.update(deployment_status)\n return data\n\n def retrieve(self, request, project_id, clb_id):\n data = CloudLoadBlancer.objects.retrieve_record(clb_id)\n data = self.add_status(request, data)\n # 添加集群名称\n data['cluster_name'] = clb_utils.get_cluster_name(\n request.user.token.access_token, project_id, data['cluster_id']\n )\n\n return response.Response(data)\n\n def delete(self, request, project_id, clb_id):\n # 获取操作对象\n record = CloudLoadBlancer.objects.retrieve(clb_id)\n if record.status not in clb_constants.ALLOW_UPDATE_DELETE_STATUS_LIST:\n raise error_codes.CheckFailed(_('当前clb状态不允许进行删除操作'))\n # 校验使用集群权限\n clb_utils.can_use_cluster(request, project_id, record.cluster_id)\n\n with client.ContextActivityLogClient(\n project_id=project_id,\n user=request.user.username,\n resource_type='lb',\n resource=record.clb_name,\n description=_(\"集群:{}, 删除clb:{}\").format(record.cluster_id, record.clb_name),\n ).log_delete():\n record.delete()\n\n return response.Response()\n\n def update(self, request, project_id, clb_id):\n slz = serializers.UpdateCLBSLZ(data=request.data)\n slz.is_valid(raise_exception=True)\n data = slz.validated_data\n # 校验使用集群权限\n clb_utils.can_use_cluster(request, project_id, data['cluster_id'])\n\n with client.ContextActivityLogClient(\n project_id=project_id,\n user=request.user.username,\n resource_type='lb',\n resource=data['clb_name'],\n description=_(\"集群:{}, 更新clb:{}\").format(data['cluster_id'], data['clb_name']),\n ).log_delete():\n CloudLoadBlancer.objects.update(clb_id, data)\n data = CloudLoadBlancer.objects.retrieve_record(clb_id)\n\n return response.Response(data)\n\n\nclass CLBStatusViewSet(viewsets.ViewSet):\n renderer_classes = (BKAPIRenderer, BrowsableAPIRenderer)\n\n def compose_data(self, listeners):\n data = []\n for info in listeners:\n item = {'name': info['Name'], 'port': info['listenPort'], 'rules': info['healthStatus']['rules']}\n data.append(item)\n return data\n\n def retrieve_status(self, request, project_id, clb_id):\n record = CloudLoadBlancer.objects.retrieve_record(clb_id)\n status_detail = clb_utils.request_clb_status(request, project_id, record)\n remote_listeners = status_detail.get('remoteListeners') or []\n data = self.compose_data(remote_listeners)\n return response.Response(data)\n\n\nclass GetCLBRegionsViewSet(viewsets.ViewSet):\n renderer_classes = (BKAPIRenderer, BrowsableAPIRenderer)\n\n def list(self, request, project_id):\n data = clb_utils.get_clb_region_list(\n request.user.token.access_token,\n )\n return response.Response(data)\n", "sub_path": "bcs-app/backend/uniapps/network/clb/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 10336, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "rest_framework.viewsets.ViewSet", "line_number": 23, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 23, "usage_type": "name"}, {"api_name": "backend.utils.renderers.BKAPIRenderer", "line_number": 24, "usage_type": "name"}, {"api_name": "rest_framework.renderers.BrowsableAPIRenderer", "line_number": 24, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.utils.describe_clb_detail", "line_number": 27, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.utils", "line_number": 27, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 33, "usage_type": "call"}, {"api_name": "rest_framework.response", "line_number": 33, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.ViewSet", "line_number": 36, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 36, "usage_type": "name"}, {"api_name": "backend.utils.renderers.BKAPIRenderer", "line_number": 37, "usage_type": "name"}, {"api_name": "rest_framework.renderers.BrowsableAPIRenderer", "line_number": 37, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.utils.get_deployments", "line_number": 45, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.utils", "line_number": 45, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.utils.get_cluster_id_names_map", "line_number": 50, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.utils", "line_number": 50, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects.get_clb_list", "line_number": 60, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects", "line_number": 60, "usage_type": "attribute"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer", "line_number": 60, "usage_type": "name"}, {"api_name": "backend.accounts.bcs_perm.Cluster.hook_perms", "line_number": 64, "usage_type": "call"}, {"api_name": "backend.accounts.bcs_perm.Cluster", "line_number": 64, "usage_type": "attribute"}, {"api_name": "backend.accounts.bcs_perm", "line_number": 64, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 66, "usage_type": "call"}, {"api_name": "rest_framework.response", "line_number": 66, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.utils.describe_clb_detail", "line_number": 69, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.utils", "line_number": 69, "usage_type": "name"}, {"api_name": "backend.utils.error_codes.error_codes.CheckFailed", "line_number": 73, "usage_type": "call"}, {"api_name": "backend.utils.error_codes.error_codes", "line_number": 73, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.serializers.CreateCLBSLZ", "line_number": 77, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.serializers", "line_number": 77, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.utils.can_use_cluster", "line_number": 81, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.utils", "line_number": 81, "usage_type": "name"}, {"api_name": "django.utils.crypto.get_random_string", "line_number": 87, "usage_type": "call"}, {"api_name": "backend.bcs_web.audit_log.client.ContextActivityLogClient", "line_number": 93, "usage_type": "call"}, {"api_name": "backend.bcs_web.audit_log.client", "line_number": 93, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 98, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects.create", "line_number": 100, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects", "line_number": 100, "usage_type": "attribute"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer", "line_number": 100, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects.parse_record", "line_number": 101, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects", "line_number": 101, "usage_type": "attribute"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer", "line_number": 101, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 103, "usage_type": "call"}, {"api_name": "rest_framework.response", "line_number": 103, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.ViewSet", "line_number": 106, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 106, "usage_type": "name"}, {"api_name": "backend.utils.renderers.BKAPIRenderer", "line_number": 107, "usage_type": "name"}, {"api_name": "rest_framework.renderers.BrowsableAPIRenderer", "line_number": 107, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects.filter", "line_number": 110, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects", "line_number": 110, "usage_type": "attribute"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer", "line_number": 110, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects.retrieve_record", "line_number": 114, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects", "line_number": 114, "usage_type": "attribute"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer", "line_number": 114, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.utils.can_use_cluster", "line_number": 116, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.utils", "line_number": 116, "usage_type": "name"}, {"api_name": "backend.components.paas_cc.get_jfrog_domain", "line_number": 118, "usage_type": "call"}, {"api_name": "backend.components.paas_cc", "line_number": 118, "usage_type": "name"}, {"api_name": "django.conf.settings.DEFAUT_MESOS_LB_JFROG_DOMAIN", "line_number": 120, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 120, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 122, "usage_type": "call"}, {"api_name": "django.template.loader.render_to_string", "line_number": 122, "usage_type": "call"}, {"api_name": "backend.bcs_web.audit_log.client.ContextActivityLogClient", "line_number": 124, "usage_type": "call"}, {"api_name": "backend.bcs_web.audit_log.client", "line_number": 124, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 129, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.utils.create_mesos_deployment", "line_number": 131, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.utils", "line_number": 131, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.constants.CLB_CREATED_STATUS", "line_number": 135, "usage_type": "attribute"}, {"api_name": "backend.uniapps.network.clb.constants", "line_number": 135, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 137, "usage_type": "call"}, {"api_name": "rest_framework.response", "line_number": 137, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects.retrieve_record", "line_number": 140, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects", "line_number": 140, "usage_type": "attribute"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer", "line_number": 140, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.utils.can_use_cluster", "line_number": 142, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.utils", "line_number": 142, "usage_type": "name"}, {"api_name": "backend.bcs_web.audit_log.client.ContextActivityLogClient", "line_number": 144, "usage_type": "call"}, {"api_name": "backend.bcs_web.audit_log.client", "line_number": 144, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 149, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.utils.delete_mesos_deployment", "line_number": 151, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.utils", "line_number": 151, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.constants.CLB_DELETED_STATUS", "line_number": 159, "usage_type": "attribute"}, {"api_name": "backend.uniapps.network.clb.constants", "line_number": 159, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 161, "usage_type": "call"}, {"api_name": "rest_framework.response", "line_number": 161, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.ViewSet", "line_number": 164, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 164, "usage_type": "name"}, {"api_name": "backend.utils.renderers.BKAPIRenderer", "line_number": 165, "usage_type": "name"}, {"api_name": "rest_framework.renderers.BrowsableAPIRenderer", "line_number": 165, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.constants.CLB_CREATED_STATUS", "line_number": 168, "usage_type": "attribute"}, {"api_name": "backend.uniapps.network.clb.constants", "line_number": 168, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.utils.get_deployments", "line_number": 170, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.utils", "line_number": 170, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects.retrieve_record", "line_number": 181, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects", "line_number": 181, "usage_type": "attribute"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer", "line_number": 181, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.utils.get_cluster_name", "line_number": 184, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.utils", "line_number": 184, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 188, "usage_type": "call"}, {"api_name": "rest_framework.response", "line_number": 188, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects.retrieve", "line_number": 192, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects", "line_number": 192, "usage_type": "attribute"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer", "line_number": 192, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.constants.ALLOW_UPDATE_DELETE_STATUS_LIST", "line_number": 193, "usage_type": "attribute"}, {"api_name": "backend.uniapps.network.clb.constants", "line_number": 193, "usage_type": "name"}, {"api_name": "backend.utils.error_codes.error_codes.CheckFailed", "line_number": 194, "usage_type": "call"}, {"api_name": "backend.utils.error_codes.error_codes", "line_number": 194, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 194, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.utils.can_use_cluster", "line_number": 196, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.utils", "line_number": 196, "usage_type": "name"}, {"api_name": "backend.bcs_web.audit_log.client.ContextActivityLogClient", "line_number": 198, "usage_type": "call"}, {"api_name": "backend.bcs_web.audit_log.client", "line_number": 198, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 203, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 207, "usage_type": "call"}, {"api_name": "rest_framework.response", "line_number": 207, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.serializers.UpdateCLBSLZ", "line_number": 210, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.serializers", "line_number": 210, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.utils.can_use_cluster", "line_number": 214, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.utils", "line_number": 214, "usage_type": "name"}, {"api_name": "backend.bcs_web.audit_log.client.ContextActivityLogClient", "line_number": 216, "usage_type": "call"}, {"api_name": "backend.bcs_web.audit_log.client", "line_number": 216, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 221, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects.update", "line_number": 223, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects", "line_number": 223, "usage_type": "attribute"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer", "line_number": 223, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects.retrieve_record", "line_number": 224, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects", "line_number": 224, "usage_type": "attribute"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer", "line_number": 224, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 226, "usage_type": "call"}, {"api_name": "rest_framework.response", "line_number": 226, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.ViewSet", "line_number": 229, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 229, "usage_type": "name"}, {"api_name": "backend.utils.renderers.BKAPIRenderer", "line_number": 230, "usage_type": "name"}, {"api_name": "rest_framework.renderers.BrowsableAPIRenderer", "line_number": 230, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects.retrieve_record", "line_number": 240, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer.objects", "line_number": 240, "usage_type": "attribute"}, {"api_name": "backend.uniapps.network.clb.models.CloudLoadBlancer", "line_number": 240, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.utils.request_clb_status", "line_number": 241, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.utils", "line_number": 241, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 244, "usage_type": "call"}, {"api_name": "rest_framework.response", "line_number": 244, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.ViewSet", "line_number": 247, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 247, "usage_type": "name"}, {"api_name": "backend.utils.renderers.BKAPIRenderer", "line_number": 248, "usage_type": "name"}, {"api_name": "rest_framework.renderers.BrowsableAPIRenderer", "line_number": 248, "usage_type": "name"}, {"api_name": "backend.uniapps.network.clb.utils.get_clb_region_list", "line_number": 251, "usage_type": "call"}, {"api_name": "backend.uniapps.network.clb.utils", "line_number": 251, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 254, "usage_type": "call"}, {"api_name": "rest_framework.response", "line_number": 254, "usage_type": "name"}]} +{"seq_id": "128024111", "text": "from odoo import api, fields, models, _\nimport xlwt, xlsxwriter\nimport base64\n\n\nclass EmployeeInfoExcelReport(models.TransientModel):\n _name = 'employee.salary.payslip.report'\n _description = 'Employee Information Excel Report'\n\n employee_id = fields.Many2one('hr.afg.payroll.batches', string='Batches', required=True)\n\n @api.multi\n def generated_excel_report(self, record):\n \n employee_obj = self.env['hr.afg.payroll.batches'].search([('name','=',self.employee_id.name)])\n \n workbook = xlwt.Workbook()\n\n # Style for Excel Report\n style0 = xlwt.easyxf('font:bold True; align: horiz left; pattern: pattern solid, fore_colour white', num_format_str='#,##0.00')\n style1 = xlwt.easyxf('font:bold True, color Yellow , height 400; borders:top double; align: horiz center; pattern: pattern solid, fore_colour blue;', num_format_str='#,##0.00')\n style2 = xlwt.easyxf('font:bold True, color White , height 440; borders:top double; align: horiz center; pattern: pattern solid, fore_colour gold;', num_format_str='#,##0.00')\n styletitle = xlwt.easyxf(\n 'font:bold True, color White, height 240; borders: top double; align: horiz center; pattern: pattern solid, fore_colour gold;',\n num_format_str='#,##0.00')\n sheet = workbook.add_sheet(\"Employee Payslip Report\")\n\n # sheet.write_merge(0, 0, 0, 10, 'GENERAL INFORMATION', style2)\n\n # sheet.write_merge(1, 1, 0, 5, 'Contact Information', style1)\n # sheet.write_merge(1, 1, 6, 10, 'Position', style1)\n\n sheet.write(0, 0, 'Employee Name', styletitle)\n sheet.write(0, 1, 'Mobile', styletitle)\n sheet.write(0, 2, 'Campus', styletitle)\n sheet.write(0, 3, 'Department', styletitle)\n sheet.write(0, 4, 'Designation', styletitle)\n sheet.write(0, 5, 'Base Salary', styletitle)\n sheet.write(0, 6, 'Loss Of Pay', styletitle)\n sheet.write(0, 7, 'Bonus', styletitle)\n sheet.write(0, 8, 'Net Pay', styletitle)\n sheet.write(0, 9, 'Tax', styletitle)\n sheet.write(0, 10, 'Advance Salary', styletitle)\n sheet.write(0, 11, 'Security Deposite', styletitle)\n sheet.write(0, 12, 'Other Deductions', styletitle)\n sheet.write(0, 13, 'Salary Payble', styletitle)\n\n sheet.col(0).width = 700 * (len('Employee Name') + 1)\n sheet.col(1).width = 700 * (len('Mobile') + 1)\n sheet.col(2).width = 700 * (len('Campus') + 1)\n sheet.col(3).width = 700 * (len('Department') + 1)\n sheet.col(4).width = 700 * (len('Designation') + 1)\n sheet.col(5).width = 700 * (len('Base Salary') + 1)\n sheet.col(6).width = 700 * (len('Loss Of Pay') + 1)\n sheet.col(7).width = 700 * (len('Bonus') + 1)\n sheet.col(8).width = 700 * (len('Net Pay') + 1)\n sheet.col(9).width = 700 * (len('Tax') + 1)\n sheet.col(10).width = 700 * (len('Advance Salary') + 1)\n sheet.col(11).width = 700 * (len('Security Deposite') + 1)\n sheet.col(12).width = 700 * (len('Other Deductions') + 1)\n sheet.col(13).width = 700 * (len('Salary Payble') + 1)\n \n sheet.row(0).height_mismatch = True\n sheet.row(0).height = 256 * 2\n # sheet.row(1).height = 256 * 2\n # sheet.row(2).height = 256 * 2\n\n row = 1\n width = 1\n\n for rec in employee_obj.slip_ids:\n sheet.row(width).height = 256 * 2\n \n sheet.write(row, 0, rec.employee_id.name)\n sheet.write(row, 1, rec.mobile)\n sheet.write(row, 2, rec.campus)\n sheet.write(row, 3, rec.department)\n sheet.write(row, 4, rec.designation)\n sheet.write(row, 5, rec.base_salary)\n sheet.write(row, 6, rec.lop)\n sheet.write(row, 7, rec.bonus)\n sheet.write(row, 8, rec.net_pay)\n sheet.write(row, 9, rec.tax)\n sheet.write(row, 10, rec.advance_salary)\n sheet.write(row, 11, rec.security_deposite)\n sheet.write(row, 12, rec.other_deductions)\n sheet.write(row, 13, rec.salary_payable)\n \n row +=1\n width += 1\n workbook.save('/tmp/employee_info_list.xls')\n result_file = open('/tmp/employee_info_list.xls', 'rb').read()\n # report_name = self.employee_id.name,'+',\n attachment_id = self.env['wizard.payslip.details.report'].create({\n 'name': self.employee_id.name +'.'+'xls',\n 'report': base64.encodestring(result_file)\n })\n\n return {\n 'name': _('Notification'),\n 'context': self.env.context,\n 'view_type': 'form',\n 'view_mode': 'form',\n 'res_model': 'wizard.payslip.details.report',\n 'res_id': attachment_id.id,\n 'data': None,\n 'type': 'ir.actions.act_window',\n 'target': 'new'\n }\n\n\nclass WizardEmployeeInformationExcelReport(models.TransientModel):\n _name = 'wizard.payslip.details.report'\n\n name = fields.Char('File Name', size=64)\n report = fields.Binary('Prepared File', filters='.xls', readonly=True)\n", "sub_path": "meli_mis/addons/afg_payroll/models/excel.py", "file_name": "excel.py", "file_ext": "py", "file_size_in_byte": 5149, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "odoo.models.TransientModel", "line_number": 6, "usage_type": "attribute"}, {"api_name": "odoo.models", "line_number": 6, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 10, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 10, "usage_type": "name"}, {"api_name": "xlwt.Workbook", "line_number": 17, "usage_type": "call"}, {"api_name": "xlwt.easyxf", "line_number": 20, "usage_type": "call"}, {"api_name": "xlwt.easyxf", "line_number": 21, "usage_type": "call"}, {"api_name": "xlwt.easyxf", "line_number": 22, "usage_type": "call"}, {"api_name": "xlwt.easyxf", "line_number": 23, "usage_type": "call"}, {"api_name": "base64.encodestring", "line_number": 96, "usage_type": "call"}, {"api_name": "odoo._", "line_number": 100, "usage_type": "call"}, {"api_name": "odoo.api.multi", "line_number": 12, "usage_type": "attribute"}, {"api_name": "odoo.api", "line_number": 12, "usage_type": "name"}, {"api_name": "odoo.models.TransientModel", "line_number": 112, "usage_type": "attribute"}, {"api_name": "odoo.models", "line_number": 112, "usage_type": "name"}, {"api_name": "odoo.fields.Char", "line_number": 115, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 115, "usage_type": "name"}, {"api_name": "odoo.fields.Binary", "line_number": 116, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 116, "usage_type": "name"}]} +{"seq_id": "80557602", "text": "# Copyright 2018 The TensorFlow Probability Authors.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ============================================================================\n\"\"\"Softfloor bijector.\"\"\"\n\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nimport tensorflow.compat.v2 as tf\n\nfrom tensorflow_probability.python import math as tfp_math\nfrom tensorflow_probability.python.bijectors import bijector\nfrom tensorflow_probability.python.internal import assert_util\nfrom tensorflow_probability.python.internal import distribution_util\n\n\n__all__ = [\n \"Softfloor\",\n]\n\n\nclass Softfloor(bijector.Bijector):\n \"\"\"Compute a differentiable approximation to `tf.math.floor`.\n\n Given `x`, compute a differentiable approximation to `tf.math.floor(x)`.\n It is parameterized by a temperature parameter `t` to control the closeness\n of the approximation at the cost of numerical stability of the inverse.\n\n This `Bijector` has the following properties:\n * This `Bijector` is a map between `R` to `R`.\n * For `t` close to `0`, this bijector mimics the identity function.\n * For `t` approaching `infinity`, this bijector converges pointwise\n to `tf.math.floor` (except at integer points).\n\n Note that for lower temperatures `t`, this bijector becomes more numerically\n unstable. In particular, the inverse for this bijector is not numerically\n stable at lower temperatures, because flooring is not a bijective function (\n and hence any pointwise limit towards the floor function will start to have a\n non-numerically stable inverse).\n\n #### Mathematical details\n\n Let `x` be in `[0.5, 1.5]`. We would like to simulate the floor function on\n this interval. We will do this via a shifted and rescaled `sigmoid`.\n\n `floor(x) = 0` for `x < 1` and `floor(x) = 1` for `x >= 1`.\n If we take `f(x) = sigmoid((x - 1.) / t)`, where `t > 0`, we can see that\n when `t` goes to zero, we get that when `x > 1`, the `f(x)` tends towards `1`\n while `f(x)` tends to `0` when `x < 1`, thus giving us a function that looks\n like the floor function. If we shift `f(x)` by `-sigmoid(-0.5 / t)` and\n rescale by `1 / (sigmoid(0.5 / t) - sigmoid(-0.5 / t))`, we preserve the\n pointwise limit, but also fix `f(0.5) = 0.` and `f(1.5) = 1.`.\n\n Thus we can define `softfloor(x, t) = a * sigmoid((x - 1.) / t) + b`\n\n where\n * `a = 1 / (sigmoid(0.5 / t) - sigmoid(-0.5 / t))`\n * `b = -sigmoid(-0.5 / t) / (sigmoid(0.5 / t) - sigmoid(-0.5 / t))`\n\n\n The implementation of the `Softfloor` bijector follows this, with the caveat\n that we extend the function to all of the real line, by appropriately shifting\n this function for each integer.\n\n #### Examples\n\n Example use:\n\n ```python\n # High temperature.\n soft_floor = Softfloor(temperature=100.)\n x = [2.1, 3.2, 5.5]\n soft_floor.forward(x)\n\n # Low temperature. This acts like a floor.\n soft_floor = Softfloor(temperature=0.01)\n soft_floor.forward(x) # Should be close to [2., 3., 5.]\n\n # Ceiling is just a shifted floor at non-integer points.\n soft_ceiling = tfb.Chain(\n [tfb.AffineScalar(1.),\n tfb.Softfloor(temperature=1.)])\n soft_ceiling.forward(x) # Should be close to [3., 5., 6.]\n ```\n \"\"\"\n\n def __init__(self,\n temperature,\n validate_args=False,\n name=\"softfloor\"):\n with tf.name_scope(name) as name:\n self._temperature = tf.convert_to_tensor(temperature, name=\"temperature\")\n if validate_args:\n self._temperature = distribution_util.with_dependencies([\n assert_util.assert_positive(\n self._temperature,\n message=\"Argument temperature was not positive\")\n ], self._temperature)\n super(Softfloor, self).__init__(\n forward_min_event_ndims=0,\n validate_args=validate_args,\n dtype=self._temperature.dtype,\n name=name)\n\n def _forward(self, x):\n # This has a well defined derivative with respect to x.\n # This is because in the range [a, a + 1.] this is just a rescaled\n # logit function and hence has a derivative. At the end points, because\n # the logit function satisfies 1 - sigma(-x) = sigma(x), we have that\n # the derivative is symmetric around the center of the interval (a + 0.5),\n # and hence is continuous at the endpoints.\n x = x - 0.5\n fractional_part = x - tf.math.floor(x)\n cyclic_part = tf.math.sigmoid((fractional_part - 0.5) / self.temperature)\n # Rescale so the left tail is 0., and the right tail is 1. This\n # will also guarantee us continuity. Differentiability comes from the\n # fact that the derivative of the sigmoid is symmetric, and hence\n # the two endpoints will have the same value for derivatives.\n rescaled_part = (\n cyclic_part / tf.math.tanh(1. / (self.temperature * 4)) -\n tf.math.exp(-0.5 / self.temperature) / (\n -tf.math.expm1(-0.5 / self.temperature)))\n return tf.math.floor(x) + rescaled_part\n\n # TODO(b/134588121): Improve the numerical stability of this function.\n def _inverse(self, y):\n fractional_part = y - tf.math.floor(y)\n # The naive thing to do is affine scale the fractional part, and apply\n # a logit function (to invert the _forward). However that has bad numerics\n # at lower temperatures, whereas this rewriting allows for lower\n # temperature scaling.\n new_fractional_part = (\n tf.math.log1p(fractional_part * -tf.math.expm1(\n -0.5 / self.temperature)) -\n tf.math.log(tf.math.exp(-0.5 / self.temperature) -\n fractional_part * tf.math.expm1(-0.5 / self.temperature)))\n new_fractional_part = self.temperature * new_fractional_part + 0.5\n return tf.math.floor(y) + new_fractional_part\n\n def _forward_log_det_jacobian(self, x):\n x = x - 0.5\n fractional_part = x - tf.math.floor(x)\n inner_part = (fractional_part - 0.5) / self.temperature\n\n offset = (tf.math.log(self.temperature) - tf.math.softplus(\n 0.5 / self.temperature) + tfp_math.softplus_inverse(\n 0.5 / self.temperature))\n\n return (-tf.math.softplus(-inner_part) -\n tf.math.softplus(inner_part) -\n offset)\n\n @property\n def temperature(self):\n return self._temperature\n", "sub_path": "tensorflow_probability/python/bijectors/softfloor.py", "file_name": "softfloor.py", "file_ext": "py", "file_size_in_byte": 6769, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "tensorflow_probability.python.bijectors.bijector.Bijector", "line_number": 34, "usage_type": "attribute"}, {"api_name": "tensorflow_probability.python.bijectors.bijector", "line_number": 34, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.name_scope", "line_number": 103, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 103, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.convert_to_tensor", "line_number": 104, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2", "line_number": 104, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.distribution_util.with_dependencies", "line_number": 106, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.distribution_util", "line_number": 106, "usage_type": "name"}, {"api_name": "tensorflow_probability.python.internal.assert_util.assert_positive", "line_number": 107, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.internal.assert_util", "line_number": 107, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.math.floor", "line_number": 125, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.math", "line_number": 125, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 125, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.math.sigmoid", "line_number": 126, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.math", "line_number": 126, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 126, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.math.tanh", "line_number": 132, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.math", "line_number": 132, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 132, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.math.exp", "line_number": 133, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.math", "line_number": 133, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 133, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.math.expm1", "line_number": 134, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.math", "line_number": 134, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 134, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.math.floor", "line_number": 135, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.math", "line_number": 135, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 135, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.math.floor", "line_number": 139, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.math", "line_number": 139, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 139, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.math.log1p", "line_number": 145, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.math", "line_number": 145, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 145, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.math.expm1", "line_number": 145, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.math.log", "line_number": 147, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.math", "line_number": 147, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 147, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.math.exp", "line_number": 147, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.math.expm1", "line_number": 148, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.math", "line_number": 148, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 148, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.math.floor", "line_number": 150, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.math", "line_number": 150, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 150, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.math.floor", "line_number": 154, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.math", "line_number": 154, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 154, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.math.log", "line_number": 157, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.math", "line_number": 157, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 157, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.math.softplus", "line_number": 157, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.math.softplus_inverse", "line_number": 158, "usage_type": "call"}, {"api_name": "tensorflow_probability.python.math", "line_number": 158, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.math.softplus", "line_number": 161, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.math", "line_number": 161, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 161, "usage_type": "name"}, {"api_name": "tensorflow.compat.v2.math.softplus", "line_number": 162, "usage_type": "call"}, {"api_name": "tensorflow.compat.v2.math", "line_number": 162, "usage_type": "attribute"}, {"api_name": "tensorflow.compat.v2", "line_number": 162, "usage_type": "name"}]} +{"seq_id": "546350413", "text": "\nfrom panda3d.core import Texture, NodePath, ShaderAttrib, Vec4, Vec3\nfrom panda3d.core import Shader\nfrom panda3d.core import Vec2, LMatrix4f, LVecBase3i, Camera, Mat4\nfrom panda3d.core import Mat4, OmniBoundingVolume, OrthographicLens\nfrom panda3d.core import PStatCollector, BoundingBox, Point3, CullFaceAttrib\nfrom panda3d.core import DepthTestAttrib, PTALVecBase3f\nfrom direct.stdpy.file import isfile, open, join\n\nfrom Globals import Globals\nfrom DebugObject import DebugObject\nfrom BetterShader import BetterShader\nfrom LightType import LightType\nfrom GUI.BufferViewerGUI import BufferViewerGUI\nfrom RenderTarget import RenderTarget\nfrom GIHelperLight import GIHelperLight\nfrom LightType import LightType\nfrom SettingsManager import SettingsManager\n\nimport time\nimport math\n\npstats_PopulateVoxelGrid = PStatCollector(\n \"App:GlobalIllumnination:PopulateVoxelGrid\")\npstats_GenerateVoxelOctree = PStatCollector(\n \"App:GlobalIllumnination:GenerateVoxelOctree\")\npstats_ClearGI = PStatCollector(\"App:GlobalIllumnination:Clear\")\npstats_GenerateMipmaps = PStatCollector(\"App:GlobalIllumnination::GenerateMipmaps\")\n\nclass GlobalIllumination(DebugObject):\n\n \"\"\" This class handles the global illumination processing. It is still\n experimental, and thus not commented. \"\"\"\n\n updateEnabled = False\n\n def __init__(self, pipeline):\n DebugObject.__init__(self, \"GlobalIllumnination\")\n self.pipeline = pipeline\n\n self.targetCamera = Globals.base.cam\n self.targetSpace = Globals.base.render\n\n self.voxelBaseResolution = 512 * 4\n self.voxelGridSizeWS = Vec3(50, 50, 20)\n self.voxelGridResolution = LVecBase3i(512, 512, 128)\n self.targetLight = None\n self.helperLight = None\n self.ptaGridPos = PTALVecBase3f.emptyArray(1)\n self.gridPos = Vec3(0)\n\n @classmethod\n def setUpdateEnabled(self, enabled):\n self.updateEnabled = enabled\n\n def setTargetLight(self, light):\n \"\"\" Sets the sun light which is the main source of GI \"\"\"\n\n if light._getLightType() != LightType.Directional:\n self.error(\"setTargetLight expects a directional light!\")\n return\n\n self.targetLight = light\n self._createHelperLight()\n\n def _prepareVoxelScene(self):\n \"\"\" Creates the internal buffer to voxelize the scene on the fly \"\"\"\n self.voxelizeScene = Globals.render\n self.voxelizeCamera = Camera(\"VoxelizeScene\")\n self.voxelizeCameraNode = self.voxelizeScene.attachNewNode(self.voxelizeCamera)\n self.voxelizeLens = OrthographicLens()\n self.voxelizeLens.setFilmSize(self.voxelGridSizeWS.x*2, self.voxelGridSizeWS.y*2)\n self.voxelizeLens.setNearFar(0.0, self.voxelGridSizeWS.x*2)\n self.voxelizeCamera.setLens(self.voxelizeLens)\n self.voxelizeCamera.setTagStateKey(\"VoxelizePassShader\")\n\n self.targetSpace.setTag(\"VoxelizePassShader\", \"Default\")\n\n self.voxelizeCameraNode.setPos(0,0,0)\n self.voxelizeCameraNode.lookAt(0,0,0)\n\n self.voxelizeTarget = RenderTarget(\"DynamicVoxelization\")\n self.voxelizeTarget.setSize(self.voxelBaseResolution) \n # self.voxelizeTarget.addDepthTexture()\n # self.voxelizeTarget.addColorTexture()\n # self.voxelizeTarget.setColorBits(16)\n self.voxelizeTarget.setSource(self.voxelizeCameraNode, Globals.base.win)\n self.voxelizeTarget.prepareSceneRender()\n\n self.voxelizeTarget.getQuad().node().removeAllChildren()\n self.voxelizeTarget.getInternalRegion().setSort(-400)\n self.voxelizeTarget.getInternalBuffer().setSort(-399)\n\n # for tex in [self.voxelizeTarget.getColorTexture()]:\n # tex.setWrapU(Texture.WMClamp)\n # tex.setWrapV(Texture.WMClamp)\n # tex.setMinfilter(Texture.FTNearest)\n # tex.setMagfilter(Texture.FTNearest)\n\n voxelSize = Vec3(\n self.voxelGridSizeWS.x * 2.0 / self.voxelGridResolution.x,\n self.voxelGridSizeWS.y * 2.0 / self.voxelGridResolution.y,\n self.voxelGridSizeWS.z * 2.0 / self.voxelGridResolution.z\n )\n\n self.targetSpace.setShaderInput(\"dv_gridSize\", self.voxelGridSizeWS * 2)\n self.targetSpace.setShaderInput(\"dv_voxelSize\", voxelSize)\n self.targetSpace.setShaderInput(\"dv_gridResolution\", self.voxelGridResolution)\n\n\n def _createVoxelizeState(self):\n \"\"\" Creates the tag state and loades the voxelizer shader \"\"\"\n self.voxelizeShader = Shader.load(Shader.SLGLSL, \n \"GI/Voxelize.vertex\",\n \"GI/Voxelize.fragment\"\n # \"GI/Voxelize.geometry\"\n )\n\n initialState = NodePath(\"VoxelizerState\")\n initialState.setShader(self.voxelizeShader, 50)\n initialState.setAttrib(CullFaceAttrib.make(CullFaceAttrib.MCullNone))\n initialState.setDepthWrite(False)\n initialState.setDepthTest(False)\n initialState.setAttrib(DepthTestAttrib.make(DepthTestAttrib.MNone))\n\n initialState.setShaderInput(\"dv_dest_tex\", self.voxelGenTex)\n\n self.voxelizeCamera.setTagState(\n \"Default\", initialState.getState())\n\n def _createHelperLight(self):\n \"\"\" Creates the helper light. We can't use the main directional light\n because it uses PSSM, so we need an extra shadow map \"\"\"\n self.helperLight = GIHelperLight()\n self.helperLight.setPos(Vec3(50,50,100))\n self.helperLight.setShadowMapResolution(512)\n self.helperLight.setFilmSize(math.sqrt( (self.voxelGridSizeWS.x**2) * 2) * 2 )\n self.helperLight.setCastsShadows(True)\n self.pipeline.addLight(self.helperLight)\n\n self.targetSpace.setShaderInput(\"dv_uv_size\", \n float(self.helperLight.shadowResolution) / self.pipeline.settings.shadowAtlasSize)\n self.targetSpace.setShaderInput(\"dv_atlas\", \n self.pipeline.getLightManager().getAtlasTex())\n\n self._updateGridPos()\n\n def setup(self):\n \"\"\" Setups everything for the GI to work \"\"\"\n\n # if self.pipeline.settings.useHardwarePCF:\n # self.fatal(\n # \"Global Illumination does not work in combination with PCF!\")\n # return\n\n self._prepareVoxelScene()\n\n # Create 3D Texture to store the voxel generation grid\n self.voxelGenTex = Texture(\"VoxelsTemp\")\n self.voxelGenTex.setup3dTexture(self.voxelGridResolution.x, self.voxelGridResolution.y, self.voxelGridResolution.z,\n Texture.TInt, Texture.FR32i)\n self.voxelGenTex.setMinfilter(Texture.FTLinearMipmapLinear)\n self.voxelGenTex.setMagfilter(Texture.FTLinear)\n\n # Create 3D Texture which is a copy of the voxel generation grid but\n # stable, as the generation grid is updated part by part\n self.voxelStableTex = Texture(\"VoxelsStable\")\n self.voxelStableTex.setup3dTexture(self.voxelGridResolution.x, self.voxelGridResolution.y, self.voxelGridResolution.z,\n Texture.TFloat, Texture.FRgba8)\n self.voxelStableTex.setMinfilter(Texture.FTLinearMipmapLinear)\n self.voxelStableTex.setMagfilter(Texture.FTLinear) \n\n for prepare in [self.voxelGenTex, self.voxelStableTex]:\n prepare.setMagfilter(Texture.FTLinear)\n prepare.setMinfilter(Texture.FTLinearMipmapLinear)\n prepare.setWrapU(Texture.WMBorderColor)\n prepare.setWrapV(Texture.WMBorderColor)\n prepare.setWrapW(Texture.WMBorderColor)\n prepare.setBorderColor(Vec4(0,0,0,0))\n\n self.voxelGenTex.setMinfilter(Texture.FTNearest)\n self.voxelGenTex.setMagfilter(Texture.FTNearest)\n self.voxelGenTex.setWrapU(Texture.WMClamp)\n self.voxelGenTex.setWrapV(Texture.WMClamp)\n self.voxelGenTex.setWrapW(Texture.WMClamp)\n\n # self.voxelStableTex.generateRamMipmapImages() \n\n self._createVoxelizeState()\n\n self.clearTextureNode = NodePath(\"ClearTexture\")\n self.copyTextureNode = NodePath(\"CopyTexture\")\n self.generateMipmapsNode = NodePath(\"GenerateMipmaps\")\n self.convertGridNode = NodePath(\"ConvertGrid\")\n\n self.reloadShader()\n\n def _generateMipmaps(self, tex):\n \"\"\" Generates all mipmaps for a 3D texture, using a gaussian function \"\"\"\n\n pstats_GenerateMipmaps.start()\n currentMipmap = 0\n computeSize = LVecBase3i(self.voxelGridResolution)\n self.generateMipmapsNode.setShaderInput(\"source\", tex)\n self.generateMipmapsNode.setShaderInput(\n \"pixelSize\", 1.0 / computeSize.x)\n\n while computeSize.z > 1:\n computeSize /= 2\n self.generateMipmapsNode.setShaderInput(\n \"sourceMipmap\", LVecBase3i(currentMipmap))\n self.generateMipmapsNode.setShaderInput(\n \"currentMipmapSize\", LVecBase3i(computeSize))\n self.generateMipmapsNode.setShaderInput(\n \"dest\", tex, False, True, -1, currentMipmap + 1)\n self._executeShader(self.generateMipmapsNode,\n (computeSize.x + 7) / 8,\n (computeSize.y + 7) / 8,\n (computeSize.z + 7) / 8)\n currentMipmap += 1\n\n pstats_GenerateMipmaps.stop()\n\n def _createCleanShader(self):\n #shader = BetterShader.loadCompute(\"GI/ClearTexture.compute\")\n shader = Shader.loadCompute(Shader.SLGLSL, \"GI/ClearTexture.compute\")\n self.clearTextureNode.setShader(shader)\n\n def _createConvertShader(self):\n #shader = BetterShader.loadCompute(\"GI/ConvertGrid.compute\")\n shader = Shader.loadCompute(Shader.SLGLSL, \"GI/ConvertGrid.compute\")\n self.convertGridNode.setShader(shader)\n\n def _createGenerateMipmapsShader(self):\n #shader = BetterShader.loadCompute(\"GI/GenerateMipmaps.compute\")\n shader = Shader.loadCompute(Shader.SLGLSL, \"GI/GenerateMipmaps.compute\")\n self.generateMipmapsNode.setShader(shader)\n\n def reloadShader(self):\n self._createCleanShader()\n self._createGenerateMipmapsShader()\n self._createConvertShader()\n self._createVoxelizeState()\n self.frameIndex = 0\n\n def _clear3DTexture(self, tex, clearVal=None):\n \"\"\" Clears a 3D Texture to \"\"\"\n if clearVal is None:\n clearVal = Vec4(0)\n\n self.clearTextureNode.setShaderInput(\"target\", tex, False, True, -1, 0)\n self.clearTextureNode.setShaderInput(\n \"clearValue\", clearVal)\n\n self._executeShader(\n self.clearTextureNode,\n (tex.getXSize() + 7) / 8,\n (tex.getYSize() + 7) / 8,\n (tex.getZSize() + 7) / 8)\n\n def _updateGridPos(self):\n\n snap = 32.0\n stepSizeX = float(self.voxelGridSizeWS.x * 2.0) / float(self.voxelGridResolution.x) * snap\n stepSizeY = float(self.voxelGridSizeWS.y * 2.0) / float(self.voxelGridResolution.y) * snap\n stepSizeZ = float(self.voxelGridSizeWS.z * 2.0) / float(self.voxelGridResolution.z) * snap\n\n self.gridPos = self.targetCamera.getPos(self.targetSpace)\n self.gridPos.x -= self.gridPos.x % stepSizeX\n self.gridPos.y -= self.gridPos.y % stepSizeY\n self.gridPos.z -= self.gridPos.z % stepSizeZ\n\n def process(self):\n if self.targetLight is None:\n self.fatal(\"The GI cannot work without a target light! Set one \"\n \"with setTargetLight() first!\")\n\n if not self.updateEnabled:\n self.voxelizeTarget.setActive(False)\n return\n\n direction = self.targetLight.getDirection()\n\n # time.sleep(0.4)\n\n if self.frameIndex == 0:\n # Find out cam pos\n \n self.targetSpace.setShaderInput(\"dv_uv_start\", \n self.helperLight.shadowSources[0].getAtlasPos())\n\n self.voxelizeTarget.setActive(True)\n # self.voxelizeTarget.setActive(False)\n\n self.voxelizeLens.setFilmSize(self.voxelGridSizeWS.y*2, self.voxelGridSizeWS.z*2)\n self.voxelizeLens.setNearFar(0.0, self.voxelGridSizeWS.x*2)\n\n self.targetSpace.setShaderInput(\"dv_mvp\", Mat4(self.helperLight.shadowSources[0].mvp))\n self.targetSpace.setShaderInput(\"dv_gridStart\", self.gridPos - self.voxelGridSizeWS)\n self.targetSpace.setShaderInput(\"dv_gridEnd\", self.gridPos + self.voxelGridSizeWS)\n self.targetSpace.setShaderInput(\"dv_lightdir\", direction)\n\n # Clear textures\n self._clear3DTexture(self.voxelGenTex, Vec4(0,0,0,0))\n\n # Voxelize from x axis\n self.voxelizeCameraNode.setPos(self.gridPos - Vec3(self.voxelGridSizeWS.x, 0, 0))\n self.voxelizeCameraNode.lookAt(self.gridPos)\n self.targetSpace.setShaderInput(\"dv_direction\", LVecBase3i(0))\n\n\n elif self.frameIndex == 1:\n # Voxelize from y axis\n\n # self.voxelizeTarget.setActive(False)\n\n self.voxelizeLens.setFilmSize(self.voxelGridSizeWS.x*2, self.voxelGridSizeWS.z*2)\n self.voxelizeLens.setNearFar(0.0, self.voxelGridSizeWS.y*2)\n\n self.voxelizeCameraNode.setPos(self.gridPos - Vec3(0, self.voxelGridSizeWS.y, 0))\n self.voxelizeCameraNode.lookAt(self.gridPos)\n self.targetSpace.setShaderInput(\"dv_direction\", LVecBase3i(1))\n\n elif self.frameIndex == 2:\n\n # self.voxelizeTarget.setActive(False)\n # Voxelize from z axis\n self.voxelizeLens.setFilmSize(self.voxelGridSizeWS.x*2, self.voxelGridSizeWS.y*2)\n self.voxelizeLens.setNearFar(0.0, self.voxelGridSizeWS.z*2)\n\n self.voxelizeCameraNode.setPos(self.gridPos + Vec3(0, 0, self.voxelGridSizeWS.z))\n self.voxelizeCameraNode.lookAt(self.gridPos)\n self.targetSpace.setShaderInput(\"dv_direction\", LVecBase3i(2))\n\n elif self.frameIndex == 3:\n\n\n self.voxelizeTarget.setActive(False)\n\n # Copy the cache to the actual texture\n self.convertGridNode.setShaderInput(\"src\", self.voxelGenTex)\n self.convertGridNode.setShaderInput(\"dest\", self.voxelStableTex)\n self._executeShader(\n self.convertGridNode, (self.voxelGridResolution.x+7) / 8, (self.voxelGridResolution.y+7) / 8, (self.voxelGridResolution.z+7) / 8)\n\n # Generate the mipmaps\n self._generateMipmaps(self.voxelStableTex)\n\n self.helperLight.setPos(self.gridPos)\n self.helperLight.setDirection(direction)\n\n # We are done now, update the inputs\n self.ptaGridPos[0] = Vec3(self.gridPos)\n self._updateGridPos()\n \n\n self.frameIndex += 1\n self.frameIndex = self.frameIndex % 5\n\n\n def bindTo(self, node, prefix):\n \"\"\" Binds all required shader inputs to a target to compute / display\n the global illumination \"\"\"\n\n normFactor = Vec3(\n 1.0,\n float(self.voxelGridResolution.y) / float(self.voxelGridResolution.x) * self.voxelGridSizeWS.y / self.voxelGridSizeWS.x,\n float(self.voxelGridResolution.z) / float(self.voxelGridResolution.x) * self.voxelGridSizeWS.z / self.voxelGridSizeWS.x\n )\n node.setShaderInput(prefix + \".gridPos\", self.ptaGridPos)\n node.setShaderInput(prefix + \".gridHalfSize\", self.voxelGridSizeWS)\n node.setShaderInput(prefix + \".gridResolution\", self.voxelGridResolution)\n node.setShaderInput(prefix + \".voxels\", self.voxelStableTex)\n node.setShaderInput(prefix + \".voxelNormFactor\", normFactor)\n node.setShaderInput(prefix + \".geometry\", self.voxelStableTex)\n\n def _executeShader(self, node, threadsX, threadsY, threadsZ=1):\n \"\"\" Executes a compute shader, fetching the shader attribute from a NodePath \"\"\"\n sattr = node.getAttrib(ShaderAttrib)\n Globals.base.graphicsEngine.dispatchCompute(\n (threadsX, threadsY, threadsZ), sattr, Globals.base.win.get_gsg())\n", "sub_path": "Code/GlobalIllumination.py", "file_name": "GlobalIllumination.py", "file_ext": "py", "file_size_in_byte": 16078, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "panda3d.core.PStatCollector", "line_number": 23, "usage_type": "call"}, {"api_name": "panda3d.core.PStatCollector", "line_number": 25, "usage_type": "call"}, {"api_name": "panda3d.core.PStatCollector", "line_number": 27, "usage_type": "call"}, {"api_name": "panda3d.core.PStatCollector", "line_number": 28, "usage_type": "call"}, {"api_name": "DebugObject.DebugObject", "line_number": 30, "usage_type": "name"}, {"api_name": "DebugObject.DebugObject.__init__", "line_number": 38, "usage_type": "call"}, {"api_name": "DebugObject.DebugObject", "line_number": 38, "usage_type": "name"}, {"api_name": "Globals.Globals.base", "line_number": 41, "usage_type": "attribute"}, {"api_name": "Globals.Globals", "line_number": 41, "usage_type": "name"}, {"api_name": "Globals.Globals.base", "line_number": 42, "usage_type": "attribute"}, {"api_name": "Globals.Globals", "line_number": 42, "usage_type": "name"}, {"api_name": "panda3d.core.Vec3", "line_number": 45, "usage_type": "call"}, {"api_name": "panda3d.core.LVecBase3i", "line_number": 46, "usage_type": "call"}, {"api_name": "panda3d.core.PTALVecBase3f.emptyArray", "line_number": 49, "usage_type": "call"}, {"api_name": "panda3d.core.PTALVecBase3f", "line_number": 49, "usage_type": "name"}, {"api_name": "panda3d.core.Vec3", "line_number": 50, "usage_type": "call"}, {"api_name": "LightType.LightType.Directional", "line_number": 59, "usage_type": "attribute"}, {"api_name": "LightType.LightType", "line_number": 59, "usage_type": "name"}, {"api_name": "Globals.Globals.render", "line_number": 68, "usage_type": "attribute"}, {"api_name": "Globals.Globals", "line_number": 68, "usage_type": "name"}, {"api_name": "panda3d.core.Camera", "line_number": 69, "usage_type": "call"}, {"api_name": "panda3d.core.OrthographicLens", "line_number": 71, "usage_type": "call"}, {"api_name": "RenderTarget.RenderTarget", "line_number": 82, "usage_type": "call"}, {"api_name": "Globals.Globals.base", "line_number": 87, "usage_type": "attribute"}, {"api_name": "Globals.Globals", "line_number": 87, "usage_type": "name"}, {"api_name": "panda3d.core.Vec3", "line_number": 100, "usage_type": "call"}, {"api_name": "panda3d.core.Shader.load", "line_number": 113, "usage_type": "call"}, {"api_name": "panda3d.core.Shader", "line_number": 113, "usage_type": "name"}, {"api_name": "panda3d.core.Shader.SLGLSL", "line_number": 113, "usage_type": "attribute"}, {"api_name": "panda3d.core.NodePath", "line_number": 119, "usage_type": "call"}, {"api_name": "panda3d.core.CullFaceAttrib.make", "line_number": 121, "usage_type": "call"}, {"api_name": "panda3d.core.CullFaceAttrib", "line_number": 121, "usage_type": "name"}, {"api_name": "panda3d.core.CullFaceAttrib.MCullNone", "line_number": 121, "usage_type": "attribute"}, {"api_name": "panda3d.core.DepthTestAttrib.make", "line_number": 124, "usage_type": "call"}, {"api_name": "panda3d.core.DepthTestAttrib", "line_number": 124, "usage_type": "name"}, {"api_name": "panda3d.core.DepthTestAttrib.MNone", "line_number": 124, "usage_type": "attribute"}, {"api_name": "GIHelperLight.GIHelperLight", "line_number": 134, "usage_type": "call"}, {"api_name": "panda3d.core.Vec3", "line_number": 135, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 137, "usage_type": "call"}, {"api_name": "panda3d.core.Texture", "line_number": 159, "usage_type": "call"}, {"api_name": "panda3d.core.Texture.TInt", "line_number": 161, "usage_type": "attribute"}, {"api_name": "panda3d.core.Texture", "line_number": 161, "usage_type": "name"}, {"api_name": "panda3d.core.Texture.FR32i", "line_number": 161, "usage_type": "attribute"}, {"api_name": "panda3d.core.Texture.FTLinearMipmapLinear", "line_number": 162, "usage_type": "attribute"}, {"api_name": "panda3d.core.Texture", "line_number": 162, "usage_type": "name"}, {"api_name": "panda3d.core.Texture.FTLinear", "line_number": 163, "usage_type": "attribute"}, {"api_name": "panda3d.core.Texture", "line_number": 163, "usage_type": "name"}, {"api_name": "panda3d.core.Texture", "line_number": 167, "usage_type": "call"}, {"api_name": "panda3d.core.Texture.TFloat", "line_number": 169, "usage_type": "attribute"}, {"api_name": "panda3d.core.Texture", "line_number": 169, "usage_type": "name"}, {"api_name": "panda3d.core.Texture.FRgba8", "line_number": 169, "usage_type": "attribute"}, {"api_name": "panda3d.core.Texture.FTLinearMipmapLinear", "line_number": 170, "usage_type": "attribute"}, {"api_name": "panda3d.core.Texture", "line_number": 170, "usage_type": "name"}, {"api_name": "panda3d.core.Texture.FTLinear", "line_number": 171, "usage_type": "attribute"}, {"api_name": "panda3d.core.Texture", "line_number": 171, "usage_type": "name"}, {"api_name": "panda3d.core.Texture.FTLinear", "line_number": 174, "usage_type": "attribute"}, {"api_name": "panda3d.core.Texture", "line_number": 174, "usage_type": "name"}, {"api_name": "panda3d.core.Texture.FTLinearMipmapLinear", "line_number": 175, "usage_type": "attribute"}, {"api_name": "panda3d.core.Texture", "line_number": 175, "usage_type": "name"}, {"api_name": "panda3d.core.Texture.WMBorderColor", "line_number": 176, "usage_type": "attribute"}, {"api_name": "panda3d.core.Texture", "line_number": 176, "usage_type": "name"}, {"api_name": "panda3d.core.Texture.WMBorderColor", "line_number": 177, "usage_type": "attribute"}, {"api_name": "panda3d.core.Texture", "line_number": 177, "usage_type": "name"}, {"api_name": "panda3d.core.Texture.WMBorderColor", "line_number": 178, "usage_type": "attribute"}, {"api_name": "panda3d.core.Texture", "line_number": 178, "usage_type": "name"}, {"api_name": "panda3d.core.Vec4", "line_number": 179, "usage_type": "call"}, {"api_name": "panda3d.core.Texture.FTNearest", "line_number": 181, "usage_type": "attribute"}, {"api_name": "panda3d.core.Texture", "line_number": 181, "usage_type": "name"}, {"api_name": "panda3d.core.Texture.FTNearest", "line_number": 182, "usage_type": "attribute"}, {"api_name": "panda3d.core.Texture", "line_number": 182, "usage_type": "name"}, {"api_name": "panda3d.core.Texture.WMClamp", "line_number": 183, "usage_type": "attribute"}, {"api_name": "panda3d.core.Texture", "line_number": 183, "usage_type": "name"}, {"api_name": "panda3d.core.Texture.WMClamp", "line_number": 184, "usage_type": "attribute"}, {"api_name": "panda3d.core.Texture", "line_number": 184, "usage_type": "name"}, {"api_name": "panda3d.core.Texture.WMClamp", "line_number": 185, "usage_type": "attribute"}, {"api_name": "panda3d.core.Texture", "line_number": 185, "usage_type": "name"}, {"api_name": "panda3d.core.NodePath", "line_number": 191, "usage_type": "call"}, {"api_name": "panda3d.core.NodePath", "line_number": 192, "usage_type": "call"}, {"api_name": "panda3d.core.NodePath", "line_number": 193, "usage_type": "call"}, {"api_name": "panda3d.core.NodePath", "line_number": 194, "usage_type": "call"}, {"api_name": "panda3d.core.LVecBase3i", "line_number": 203, "usage_type": "call"}, {"api_name": "panda3d.core.LVecBase3i", "line_number": 211, "usage_type": "call"}, {"api_name": "panda3d.core.LVecBase3i", "line_number": 213, "usage_type": "call"}, {"api_name": "panda3d.core.Shader.loadCompute", "line_number": 226, "usage_type": "call"}, {"api_name": "panda3d.core.Shader", "line_number": 226, "usage_type": "name"}, {"api_name": "panda3d.core.Shader.SLGLSL", "line_number": 226, "usage_type": "attribute"}, {"api_name": "panda3d.core.Shader.loadCompute", "line_number": 231, "usage_type": "call"}, {"api_name": "panda3d.core.Shader", "line_number": 231, "usage_type": "name"}, {"api_name": "panda3d.core.Shader.SLGLSL", "line_number": 231, "usage_type": "attribute"}, {"api_name": "panda3d.core.Shader.loadCompute", "line_number": 236, "usage_type": "call"}, {"api_name": "panda3d.core.Shader", "line_number": 236, "usage_type": "name"}, {"api_name": "panda3d.core.Shader.SLGLSL", "line_number": 236, "usage_type": "attribute"}, {"api_name": "panda3d.core.Vec4", "line_number": 249, "usage_type": "call"}, {"api_name": "panda3d.core.Mat4", "line_number": 298, "usage_type": "call"}, {"api_name": "panda3d.core.Vec4", "line_number": 304, "usage_type": "call"}, {"api_name": "panda3d.core.Vec3", "line_number": 307, "usage_type": "call"}, {"api_name": "panda3d.core.LVecBase3i", "line_number": 309, "usage_type": "call"}, {"api_name": "panda3d.core.Vec3", "line_number": 320, "usage_type": "call"}, {"api_name": "panda3d.core.LVecBase3i", "line_number": 322, "usage_type": "call"}, {"api_name": "panda3d.core.Vec3", "line_number": 331, "usage_type": "call"}, {"api_name": "panda3d.core.LVecBase3i", "line_number": 333, "usage_type": "call"}, {"api_name": "panda3d.core.Vec3", "line_number": 353, "usage_type": "call"}, {"api_name": "panda3d.core.Vec3", "line_number": 365, "usage_type": "call"}, {"api_name": "panda3d.core.ShaderAttrib", "line_number": 379, "usage_type": "argument"}, {"api_name": "Globals.Globals.base.graphicsEngine.dispatchCompute", "line_number": 380, "usage_type": "call"}, {"api_name": "Globals.Globals.base", "line_number": 380, "usage_type": "attribute"}, {"api_name": "Globals.Globals", "line_number": 380, "usage_type": "name"}, {"api_name": "Globals.Globals.base.win.get_gsg", "line_number": 381, "usage_type": "call"}, {"api_name": "Globals.Globals.base", "line_number": 381, "usage_type": "attribute"}, {"api_name": "Globals.Globals", "line_number": 381, "usage_type": "name"}]} +{"seq_id": "71667163", "text": "from jinja2 import StrictUndefined\n\nfrom flask import Flask, render_template, request, flash, redirect, session, jsonify, Response\nfrom flask_debugtoolbar import DebugToolbarExtension\nfrom model import Show, Show_Color, Brand, Color, connect_to_db, db\nfrom flask_sqlalchemy import SQLAlchemy\nimport flask_sqlalchemy\nimport flask_restless\nimport json\n\n\nfrom sqlalchemy import create_engine, Column, Integer, String, Date, Float\n\n\napp = Flask(__name__)\n# app.config['DEBUG'] = True\n# app.config['SQLALCHEMY_DATABASE_URI'] = 'postgres:///showme'\ndb = flask_sqlalchemy.SQLAlchemy(app)\n\nmanager = flask_restless.APIManager(app, flask_sqlalchemy_db=db)\nshow_blueprint = manager.create_api(Show, methods=['GET'])\nbrand_blueprint = manager.create_api(Brand, methods=['GET'])\ncolor_blueprint = manager.create_api(Color, methods=['GET'])\nshow_color_blueprint = manager.create_api(Show_Color, methods=['GET'])\n\n# Required to use Flask sessions and the debug toolbar\napp.secret_key = \"ABC\"\n\n# Normally, if you use an undefined variable in Jinja2, it fails silently.\n# This is horrible. Fix this so that, instead, it raises an error.\napp.jinja_env.undefined = StrictUndefined\napp.jinja_env.auto_reload = True\n\n\n@app.route('/')\ndef index():\n \"\"\"Homepage.\"\"\"\n shows = Show.query.all()\n show_colors = Show_Color.query.all()\n colors = Color.query.all()\n brands = Brand.query.all()\n\n return render_template(\"bleep.html\",\n shows=shows,\n show_colors=show_colors,\n colors=colors,\n brands=brands)\n\n\n# @app.route('/api/all')\n# def api_all():\n# shows = Show.query.all()\n# show_colors = Show_Color.query.all()\n# colors = Color.query.all()\n# brands = Brand.query.all()\n\n# resp = Response(response=jsonify({'key': 'value'}),\n# status=200,\n# mimetype=\"application/json\")\n# return resp\n\n\n@app.route('/_get_brands')\ndef get_brands_json():\n brands = {}\n for brand in Brand.query.all():\n brands[brand.brand_id] = {\n 'brand_name': brand.brand_name,\n }\n\n return jsonify(brands)\n\n\n@app.route('/_get_shows')\ndef get_shows_json():\n shows = {}\n for show in Show.query.all():\n shows[show.show_id] = {\n 'show_id': show.show_id,\n 'show_season': show.season,\n 'show_year': show.year,\n 'brand_name': show.brands.brand_name,\n }\n\n return jsonify(shows)\n\n\n@app.route('/_get_colors')\ndef get_colors_json():\n colors = {}\n for color in Color.query.all():\n colors[color.color_id] = {\n 'color_id': color.color_id,\n 'color': color.color_name,\n 'color_hex': color.color_hex,\n }\n\n return jsonify(colors)\n\n\n@app.route('/_get_show_colors')\ndef get_show_colors_json():\n show_colors_json = {}\n for show_color in Show_Color.query.all():\n show_colors_json[show_color.show_colors_id] = {\n 'show_color': show_color.colors.color_id,\n 'brand_name': show_color.shows.brands.brand_name,\n 'color_id': show_color.color_id,\n 'color_name': show_color.colors.color_name,\n }\n\n return jsonify(show_colors_json)\n\n# ')\n# def api_by_season(color_hex):\n# events = Events.query.filter_by(event_type=event_type).all()\n# return jsonify(json_list=[event.serialize for event in events])\n\n\n# @app.route('/', methods=['GET'])\n# def register_form():\n# \"\"\"Show form for user signup.\"\"\"\n\n# return render_template(\"base.html\")\n\n\n# @app.route('/register', methods=['POST'])\n# def register_process():\n# \"\"\"Process registration.\"\"\"\n\n# # Get form variables\n# email = request.form[\"email\"]\n# password = request.form[\"password\"]\n# age = int(request.form[\"age\"])\n# zipcode = request.form[\"zipcode\"]\n\n# new_user = User(email=email, password=password, age=age, zipcode=zipcode)\n\n# db.session.add(new_user)\n# db.session.commit()\n\n# flash(\"User %s added.\" % email)\n# return redirect(\"/\")\n\n\n# @app.route('/login', methods=['GET'])\n# def login_form():\n# \"\"\"Show login form.\"\"\"\n\n# return render_template(\"login_form.html\")\n\n\n# @app.route('/login', methods=['POST'])\n# def login_process():\n# \"\"\"Process login.\"\"\"\n\n# # Get form variables\n# email = request.form[\"email\"]\n# password = request.form[\"password\"]\n\n# user = User.query.filter_by(email=email).first()\n\n# if not user:\n# flash(\"No such user\")\n# return redirect(\"/login\")\n\n# if user.password != password:\n# flash(\"Incorrect password\")\n# return redirect(\"/login\")\n\n# session[\"user_id\"] = user.user_id\n\n# flash(\"Logged in\")\n# return redirect(\"/users/%s\" % user.user_id)\n\n\n# @app.route('/logout')\n# def logout():\n# \"\"\"Log out.\"\"\"\n\n# del session[\"user_id\"]\n# flash(\"Logged Out.\")\n# return redirect(\"/\")\n\n\n# @app.route(\"/users\")\n# def user_list():\n# \"\"\"Show list of users.\"\"\"\n\n# users = User.query.all()\n# return render_template(\"user_list.html\", users=users)\n\n\n# @app.route(\"/users/\")\n# def user_detail(user_id):\n# \"\"\"Show info about user.\"\"\"\n\n# user = User.query.get(user_id)\n# return render_template(\"user.html\", user=user)\n\n\n# @app.route(\"/movies\")\n# def movie_list():\n# \"\"\"Show list of movies.\"\"\"\n\n# movies = Movie.query.order_by('title').all()\n# return render_template(\"movie_list.html\", movies=movies)\n\n\n# @app.route(\"/movies/\", methods=['GET'])\n# def movie_detail(movie_id):\n# \"\"\"Show info about movie.\n\n# If a user is logged in, let them add/edit a rating.\n# \"\"\"\n\n# movie = Movie.query.get(movie_id)\n\n# user_id = session.get(\"user_id\")\n\n# if user_id:\n# user_rating = Rating.query.filter_by(\n# movie_id=movie_id, user_id=user_id).first()\n\n# else:\n# user_rating = None\n\n# # Get average rating of movie\n\n# rating_scores = [r.score for r in movie.ratings]\n# avg_rating = float(sum(rating_scores)) / len(rating_scores)\n\n# prediction = None\n\n# # Prediction code: only predict if the user hasn't rated it.\n\n# if (not user_rating) and user_id:\n# user = User.query.get(user_id)\n# if user:\n# prediction = user.predict_rating(movie)\n\n# # Either use the prediction or their real rating\n\n# if prediction:\n# # User hasn't scored; use our prediction if we made one\n# effective_rating = prediction\n\n# elif user_rating:\n# # User has already scored for real; use that\n# effective_rating = user_rating.score\n\n# else:\n# # User hasn't scored, and we couldn't get a prediction\n# effective_rating = None\n\n# # Get the eye's rating, either by predicting or using real rating\n\n# the_eye = (User.query.filter_by(email=\"the-eye@of-judgment.com\")\n# .one())\n# eye_rating = Rating.query.filter_by(\n# user_id=the_eye.user_id, movie_id=movie.movie_id).first()\n\n# if eye_rating is None:\n# eye_rating = the_eye.predict_rating(movie)\n\n# else:\n# eye_rating = eye_rating.score\n\n# if eye_rating and effective_rating:\n# difference = abs(eye_rating - effective_rating)\n\n# else:\n# # We couldn't get an eye rating, so we'll skip difference\n# difference = None\n\n # Depending on how different we are from the Eye, choose a\n # message\n\n# BERATEMENT_MESSAGES = [\n# \"I suppose you don't have such bad taste after all.\",\n# \"I regret every decision that I've ever made that has \" +\n# \"brought me to listen to your opinion.\",\n# \"Words fail me, as your taste in movies has clearly \" +\n# \"failed you.\",\n# \"That movie is great. For a clown to watch. Idiot.\",\n# \"Words cannot express the awfulness of your taste.\"\n# ]\n\n# if difference is not None:\n# beratement = BERATEMENT_MESSAGES[int(difference)]\n\n# else:\n# beratement = None\n\n# return render_template(\n# \"movie.html\",\n# movie=movie,\n# user_rating=user_rating,\n# average=avg_rating,\n# prediction=prediction,\n# eye_rating=eye_rating,\n# difference=difference,\n# beratement=beratement\n# )\n\n\n# @app.route(\"/movies/\", methods=['POST'])\n# def movie_detail_process(movie_id):\n# \"\"\"Add/edit a rating.\"\"\"\n\n# # Get form variables\n# score = int(request.form[\"score\"])\n\n# user_id = session.get(\"user_id\")\n# if not user_id:\n# raise Exception(\"No user logged in.\")\n\n# rating = Rating.query.filter_by(user_id=user_id, movie_id=movie_id).first()\n\n# if rating:\n# rating.score = score\n# flash(\"Rating updated.\")\n\n# else:\n# rating = Rating(user_id=user_id, movie_id=movie_id, score=score)\n# flash(\"Rating added.\")\n# db.session.add(rating)\n\n# db.session.commit()\n\n# return redirect(\"/movies/%s\" % movie_id)\n\n\nif __name__ == \"__main__\":\n # We have to set debug=True here, since it has to be True at the point\n # that we invoke the DebugToolbarExtension\n app.debug = True\n\n connect_to_db(app)\n\n # Use the DebugToolbar\n DebugToolbarExtension(app)\n\n app.run(host=\"0.0.0.0\")\n", "sub_path": "server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 12346, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "flask.Flask", "line_number": 15, "usage_type": "call"}, {"api_name": "model.db", "line_number": 18, "usage_type": "name"}, {"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 18, "usage_type": "call"}, {"api_name": "flask_restless.APIManager", "line_number": 20, "usage_type": "call"}, {"api_name": "model.db", "line_number": 20, "usage_type": "name"}, {"api_name": "model.Show", "line_number": 21, "usage_type": "argument"}, {"api_name": "model.Brand", "line_number": 22, "usage_type": "argument"}, {"api_name": "model.Color", "line_number": 23, "usage_type": "argument"}, {"api_name": "model.Show_Color", "line_number": 24, "usage_type": "argument"}, {"api_name": "jinja2.StrictUndefined", "line_number": 31, "usage_type": "name"}, {"api_name": "model.Show.query.all", "line_number": 38, "usage_type": "call"}, {"api_name": "model.Show.query", "line_number": 38, "usage_type": "attribute"}, {"api_name": "model.Show", "line_number": 38, "usage_type": "name"}, {"api_name": "model.Show_Color.query.all", "line_number": 39, "usage_type": "call"}, {"api_name": "model.Show_Color.query", "line_number": 39, "usage_type": "attribute"}, {"api_name": "model.Show_Color", "line_number": 39, "usage_type": "name"}, {"api_name": "model.Color.query.all", "line_number": 40, "usage_type": "call"}, {"api_name": "model.Color.query", "line_number": 40, "usage_type": "attribute"}, {"api_name": "model.Color", "line_number": 40, "usage_type": "name"}, {"api_name": "model.Brand.query.all", "line_number": 41, "usage_type": "call"}, {"api_name": "model.Brand.query", "line_number": 41, "usage_type": "attribute"}, {"api_name": "model.Brand", "line_number": 41, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 43, "usage_type": "call"}, {"api_name": "model.Brand.query.all", "line_number": 66, "usage_type": "call"}, {"api_name": "model.Brand.query", "line_number": 66, "usage_type": "attribute"}, {"api_name": "model.Brand", "line_number": 66, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 71, "usage_type": "call"}, {"api_name": "model.Show.query.all", "line_number": 77, "usage_type": "call"}, {"api_name": "model.Show.query", "line_number": 77, "usage_type": "attribute"}, {"api_name": "model.Show", "line_number": 77, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 85, "usage_type": "call"}, {"api_name": "model.Color.query.all", "line_number": 91, "usage_type": "call"}, {"api_name": "model.Color.query", "line_number": 91, "usage_type": "attribute"}, {"api_name": "model.Color", "line_number": 91, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 98, "usage_type": "call"}, {"api_name": "model.Show_Color.query.all", "line_number": 104, "usage_type": "call"}, {"api_name": "model.Show_Color.query", "line_number": 104, "usage_type": "attribute"}, {"api_name": "model.Show_Color", "line_number": 104, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 112, "usage_type": "call"}, {"api_name": "model.Brand.query.all", "line_number": 119, "usage_type": "call"}, {"api_name": "model.Brand.query", "line_number": 119, "usage_type": "attribute"}, {"api_name": "model.Brand", "line_number": 119, "usage_type": "name"}, {"api_name": "model.Show.query.filter_by", "line_number": 122, "usage_type": "call"}, {"api_name": "model.Show.query", "line_number": 122, "usage_type": "attribute"}, {"api_name": "model.Show", "line_number": 122, "usage_type": "name"}, {"api_name": "model.Show_Color.query.filter_by", "line_number": 126, "usage_type": "call"}, {"api_name": "model.Show_Color.query", "line_number": 126, "usage_type": "attribute"}, {"api_name": "model.Show_Color", "line_number": 126, "usage_type": "name"}, {"api_name": "model.Color.query.filter_by", "line_number": 130, "usage_type": "call"}, {"api_name": "model.Color.query", "line_number": 130, "usage_type": "attribute"}, {"api_name": "model.Color", "line_number": 130, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 138, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 143, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 143, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 143, "usage_type": "name"}, {"api_name": "model.Show.query.filter_by", "line_number": 148, "usage_type": "call"}, {"api_name": "model.Show.query", "line_number": 148, "usage_type": "attribute"}, {"api_name": "model.Show", "line_number": 148, "usage_type": "name"}, {"api_name": "model.Show_Color.query.filter_by", "line_number": 151, "usage_type": "call"}, {"api_name": "model.Show_Color.query", "line_number": 151, "usage_type": "attribute"}, {"api_name": "model.Show_Color", "line_number": 151, "usage_type": "name"}, {"api_name": "model.Color.query.filter_by", "line_number": 153, "usage_type": "call"}, {"api_name": "model.Color.query", "line_number": 153, "usage_type": "attribute"}, {"api_name": "model.Color", "line_number": 153, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 181, "usage_type": "call"}, {"api_name": "model.connect_to_db", "line_number": 441, "usage_type": "call"}, {"api_name": "flask_debugtoolbar.DebugToolbarExtension", "line_number": 444, "usage_type": "call"}]} +{"seq_id": "429650232", "text": "import requests\nimport matplotlib.pyplot as plt\nimport pandas as pd\nfrom datetime import datetime, timedelta\nimport time\nimport os\nimport queue\nimport threading\nfrom . import var\n\n\ndef floorsheets():\n \"\"\"\n Threaded Scraper For FloorSheets as we need to scrape more than 75k Data\n Returns in less than 2 seconds.\n \"\"\"\n q = queue.Queue()\n contents = []\n response = requests.get(\n 'https://newweb.nepalstock.com.np/api/' +\n 'nots/nepse-data/floorsheet?size=2000&sort=contractId,desc',\n headers=var.header)\n pages = response.json()['floorsheets']['totalPages']\n\n def scrapePage(pageNUM):\n response = requests.get(\n 'https://newweb.nepalstock.com.np/api/' +\n f'nots/nepse-data/floorsheet?page={pageNUM}&size=2000&sort=contractId,desc',\n headers=var.header)\n return response.json()['floorsheets']['content']\n\n def queGET(q):\n while True:\n task = q.get()\n contents.extend(scrapePage(task))\n q.task_done()\n\n for i in range(30):\n worker = threading.Thread(target=queGET, args=(q, ), daemon=True)\n worker.start()\n\n for j in range(pages):\n q.put(j)\n\n q.join()\n\n return pd.DataFrame(contents)\n\n\nclass Floorsheet:\n def __init__(self):\n self.fs = floorsheets()\n\n def update(self):\n self.fs = floorsheets()\n\n def volume(self, scrip):\n return self.fs[self.fs.stockSymbol == scrip].contractQuantity.sum()\n\n def matching_amt(self):\n return len(self.fs[self.fs.buyerMemberId == self.fs.sellerMemberId].\n index) / len(self.fs.index) * 100\n\n def buy_to_sell(self, bid, scrip=None):\n if not scrip:\n return self.fs[self.fs.buyerMemberId == bid].contractQuantity.sum(\n ) / self.fs[self.fs.sellerMemberId == bid].contractQuantity.sum()\n else:\n return self.fs[self.fs.buyerMemberId == bid][\n self.fs.stockSymbol == scrip].contractQuantity.sum() / self.fs[\n self.fs.sellerMemberId == bid][\n self.fs.stockSymbol == scrip].contractQuantity.sum()\n\n\nclass NEPSE:\n def __init__(self):\n self.headers = var.header\n self.sectors = var.sectors\n self.host = 'https://newweb.nepalstock.com.np/api/'\n # self.securities = requests.get(self.host +\n # 'nots/securityDailyTradeStat/58',\n # headers=self.headers).json()\n pass\n\n def dateFilter(self, working_date, data):\n \"\"\"\n Function to return next working day , if the date provided is non-working day.\n\n Returns either first or last date if the date provided is too ahead or too back.\n\n \"\"\"\n\n all_dates = [date['businessDate'] for date in data]\n if working_date in all_dates:\n return working_date\n else:\n i = 0\n while 1:\n\n date = datetime.strptime(working_date, '%Y-%m-%d')\n new_date = str(date + timedelta(days=i)).split(' ')[0]\n if new_date in all_dates:\n return new_date\n i += 1\n if i >= 7:\n month = working_date.split('-')[1]\n year = working_date.split('-')[0]\n day = working_date.split('-')[-1]\n if year > all_dates[-1].split(\n '-')[0] and month > all_dates[-1].split('-')[1]:\n return all_dates[-1]\n return all_dates[0]\n\n def isOpen(self):\n \"\"\"\n Returns True if the market is Open .\n\n \"\"\"\n response = requests.get(self.host + '/nots/nepse-data/market-open',\n headers=self.headers).json()\n if response['isOpen'] != 'CLOSE':\n return True\n return False\n\n def nonthreadedfloorsheets(self):\n content = []\n page = 0\n while 1:\n response = requests.get(\n 'https://newweb.nepalstock.com.np/api/nots/nepse-data/floorsheet?page={page}&size=2000&sort=contractId,desc',\n headers=self.headers)\n data = (response.json())['floorsheets']['content']\n isLast = response.json()['floorsheets']['last']\n content.extend(data)\n page += 1\n if isLast:\n return content\n\n def indices(self, sector='NEPSE Index', start_date=None, end_date=None):\n index = sector\n index_id = [\n id['id'] for id in self.sectors if id['indexName'] == index\n ][0]\n resp = requests.get(self.host + 'nots/graph/index/58',\n headers=self.headers).json()\n # if start_date:\n # start_date = self.dateFilter(start_date, resp)\n # start_index = next((index for (index, d) in enumerate(resp)\n # if d[\"businessDate\"] == start_date), None)\n # resp = resp[start_index:]\n # if end_date:\n\n # end_date = self.dateFilter(end_date, resp)\n # end_index = next((index for (index, d) in enumerate(resp)\n # if d[\"businessDate\"] == end_date), None) + 1\n # if start_date and end_date:\n # if end_index == start_index:\n # end_index = -1\n # resp = resp[:end_index]\n return resp\n\n def brokers(self):\n \"\"\" \n Returns all the registered brokers along with tms url and other information\n\n \"\"\"\n resp = requests.get(self.host + 'nots/member?&size=500',\n headers=self.headers).json()\n return resp\n\n def alerts(self):\n \"\"\"\n\n returns alerts and news published by \n\n \"\"\"\n resp = requests.get(self.host + 'nots/news/media/news-and-alerts',\n headers=self.headers).json()\n return resp\n\n def todayPrice(self, scrip=None):\n \"\"\"\n\n Get Live Price of All The Securities in one call or specify\n\n \"\"\"\n resp = requests.get(self.host +\n 'nots/nepse-data/today-price?&size=500',\n headers=self.headers).json()['content']\n if scrip == None:\n return resp\n return [\n script for script in resp if script['symbol'] == scrip.upper()\n ][0]\n\n def markCap(self):\n \"\"\"\n\n Get Market Caps\n\n \"\"\"\n resp = requests.get(self.host + 'nots/nepse-data/marcapbydate/?',\n headers=self.headers).json()\n return resp\n\n def getChartHistory(self, scrip, start_date=None, end_date=None):\n \"\"\"\n\n returns charts data \n raises Exception if start_date or end_date != working_days (will fix it)\n\n \"\"\"\n\n scripID = [\n security for security in self.securities\n if security['symbol'] == scrip.upper()\n ][0]['securityId']\n resp = requests.get(self.host + f'nots/market/graphdata/{scripID}',\n headers=self.headers).json()\n if start_date:\n start_date = self.dateFilter(start_date, resp)\n start_index = next((index for (index, d) in enumerate(resp)\n if d[\"businessDate\"] == start_date), None)\n resp = resp[start_index:]\n if end_date:\n\n end_date = self.dateFilter(end_date, resp)\n end_index = next((index for (index, d) in enumerate(resp)\n if d[\"businessDate\"] == end_date), None) + 1\n if start_date and end_date:\n if end_index == start_index:\n end_index = -1\n resp = resp[:end_index]\n return resp\n\n def createChart(self,\n scrip,\n theme='dark',\n start_date=None,\n end_date=None,\n close=True,\n high=True,\n low=True):\n\n symbol = scrip.upper()\n if theme.upper() == 'DARK':\n plt.style.use(['dark_background'])\n\n data = self.getChartHistory(symbol, start_date, end_date)\n open_price = [d['openPrice'] for d in data]\n x = [d['businessDate'] for d in data]\n high_data = [d['highPrice'] for d in data]\n low_data = [d['lowPrice'] for d in data]\n close_price = [d['closePrice'] for d in data]\n\n plt.plot(open_price, label='Open Price')\n if close:\n plt.plot(close_price, label=\"Close Price\")\n if high:\n plt.plot(high_data, label=\"High\")\n if low:\n plt.plot(low_data, label=\"Low\")\n\n plt.legend(loc=\"upper left\")\n\n plt.title(f'{symbol} Prices As of {x[-1]}')\n\n plt.xlabel(\n f\"Start Date : {x[0]} | END DATE : {x[-1]}\\n\\nOPEN PRICE : {open_price[-1]} | ClOSE PRICE : {close_price[-1]} | High : {high_data[-1]} | Low : {low_data[-1]}\"\n )\n ax = plt.gcf().autofmt_xdate()\n ax = plt.gca()\n ax.axes.xaxis.set_ticks([])\n filename = f'{symbol}_{str(time.time())}.png'\n data = plt.savefig(filename)\n abspath = os.path.abspath(filename)\n plt.clf()\n return {'file': abspath}\n\n def saveCSV(self, scrip, start_date=None, end_date=None, filename=None):\n scripID = [\n security for security in self.securities\n if security['symbol'] == scrip.upper()\n ][0]['securityId']\n resp = self.getChartHistory(scrip, start_date, end_date)\n if not filename:\n filename = f'{scrip.upper()}_{str(time.time())}.csv'\n pd.DataFrame(resp).to_csv(filename)\n return os.path.abspath(filename)\n\n def watch(self, watchlist):\n watchlist = [scrip.upper() for scrip in watchlist]\n priceList = self.todayPrice()\n data = [scrip for scrip in priceList if scrip[\"symbol\"] in watchlist]\n\n text = \" SCRIP PCT-CH PrevCLOSE LTP\" + \"\\n\" + \"=\" * 40 + \"\\n\"\n for datum in data:\n scrip = datum[\"symbol\"]\n closing = int(datum[\"previousDayClosePrice\"])\n ltp = int(datum[\"lastUpdatedPrice\"])\n pctchange = (ltp - closing) * 100 / closing\n text = text + (scrip.upper().center(8) +\n \"%.2f\".center(9) % pctchange +\n str(closing).center(12) + str(ltp).center(8)) + \"\\n\"\n return (text)\n\n\ndef checkIPO(scripID, boid):\n \"\"\"\n CHECK IPO RESULT\n\n \"\"\"\n\n # published = requests.get('https://iporesult.cdsc.com.np/result/companyShares/fileUploaded').json()['body']\n # print()\n\n # scripID = [\n # resp['id'] for resp in if resp['scrip'] == scrip.upper()\n # ][0]\n\n return requests.post('https://iporesult.cdsc.com.np/result/result/check',\n json={\n \"companyShareId\": scripID,\n \"boid\": boid\n }).json()[\"success\"]\n\n\nif __name__ == '__main__':\n data = NEPSE()\n print(data.indices())\n", "sub_path": "nepse/stonk.py", "file_name": "stonk.py", "file_ext": "py", "file_size_in_byte": 11190, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "queue.Queue", "line_number": 17, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 19, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 26, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 39, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 47, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 100, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 100, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 101, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 119, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 129, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 144, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 167, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 177, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 187, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 202, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 218, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 247, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 247, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 247, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 256, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 256, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 258, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 258, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 260, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 260, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 262, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 262, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 264, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 264, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 266, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 266, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 268, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 268, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 271, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 271, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 272, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 272, "usage_type": "name"}, {"api_name": "time.time", "line_number": 274, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 275, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 275, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 276, "usage_type": "call"}, {"api_name": "os.path", "line_number": 276, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 277, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 277, "usage_type": "name"}, {"api_name": "time.time", "line_number": 287, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 288, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 289, "usage_type": "call"}, {"api_name": "os.path", "line_number": 289, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 321, "usage_type": "call"}]} +{"seq_id": "478502931", "text": "\"\"\"\nCS/BIOS 112 - Program Design I\n\nFor this project, I will work with a portion of this gene, known as ALX1, from four species, 5-10 individuals each, of Darwin’s finches + 1 outgroup, L. noctis (the Lesser Antillean Bullfinch).\n Synthesizing Python knowledge: loops, conditions, strings, lists, functions, random functions, modules, plotting.\n DNA vs phenotype.\n\nFile: project_02.py\n\n@author: \nDue Date: <11/02/2020>\n\"\"\"\n\n# species named in the data file finches_2020.csv \nfinch1 = 'G.conirostris_Espanola'\nfinch2 = 'G.conirostris_Genovesa'\nfinch3 = 'G.difficilis'\nfinch4 = 'G.magnirostris'\noutgroup = 'L.noctis'\n\ndef read_input(fileName):\n ''' This function is to read the CSV file, assuming the column input format is [species, Individual ID, allele (A or B), gene sequence, beak shape score (i.e., the degree of pointedness), beak type] and creating a list variable for each individual of length 7 with the format [species, Individual ID, alleleA gene, alleleB gene, alleleA beak shape score, alleleB beak shape score, beak type]. '''\n \n import csv\n \n fileref = open(fileName, \"r\")\n data_reader = csv.reader(fileref)\n row = []\n \n for i in data_reader:\n if(i[2] == 'A'):\n x = i\n del x[2]\n else:\n x.insert(3, i[3])\n x.insert(4, i[5])\n row.append(x)\n \n fileref.close()\n \n return row\n \ndef allele_dist(gene1, gene2):\n ''' Takes 2 strings as arguments. Returns the Hamming Distance between the two arguments. Assumes both argument strings are of same length. ''' \n \n x = 0\n \n for i in range(len(gene1)):\n if (gene1[i] != gene2[i]):\n x = x + 1\n \n return x\n \ndef gene_dist(finch1, finch2):\n ''' Takes 2 arguments: list for an individual finch. Returns average Hamming Distance between the genes of the two individual finches given as arguments. '''\n \n a = allele_dist(finch1[2], finch2[2])\n b = allele_dist(finch1[2], finch2[3])\n c = allele_dist(finch1[3], finch2[2])\n d = allele_dist(finch1[3], finch2[3])\n \n avg = float((a + b + c + d) / 4)\n \n return avg\n\ndef beak_dist(finch1, finch2):\n ''' Takes 2 arguments: list for an individual finch. Returns average difference between the beak score of the two individual finches given as arguments. '''\n \n a = abs(finch1[4] - finch2[4])\n b = abs(finch1[4] - finch2[5])\n c = abs(finch1[5] - finch2[4])\n d = abs(finch1[5] - finch2[5])\n \n avg = float((a + b + c + d) / 4)\n \n return avg\n\n\ndef outgroup_distance(finches, speciesName, outgroupName):\n ''' Takes 3 arguments:\n – list of lists returned by read_input( )\n – name of Finch Species on which to collect information\n – name of Outgroup Species\n • used to standard basis for comparing the distance\n • will always be the species: L.noctis\n Returns two lists as a Tuple:\n – list of the gene differences for all individuals in that Finch Species\n – list of the beak differences for all individuals in that Finch Species '''\n \n o = [finch for finch in finches if finch[0] == outgroupName]\n s = [finch for finch in finches if finch[0] == speciesName]\n \n gd_list = []\n bd_list = []\n \n for finch1 in o:\n for finch2 in s:\n gd_list.append(gene_dist(finch1, finch2))\n bd_list.append(beak_dist(finch1, finch2))\n \n return (gd_list, bd_list)\n \n\ndef plot_data(file_name, x_column, y_column) :\n ''' Is the primary function that will call the other two functions and call the actual plotting functions. '''\n \n import matplotlib.pyplot as plt\n finch_data = read_input(file_name)\n \n x = outgroup_distance(finch_data, x_column, outgroup)\n y = outgroup_distance(finch_data, y_column, outgroup)\n plt.plot(x,y, \"ro\", label = \"G.conirostris_Espanola\")\n \n x = outgroup_distance(finch_data, x_column, outgroup)\n y = outgroup_distance(finch_data, y_column, outgroup)\n plt.plot(x,y, \"bo\", label = \"G.conirostris_Genovesa\")\n \n x = outgroup_distance(finch_data, x_column, outgroup)\n y = outgroup_distance(finch_data, y_column, outgroup)\n plt.plot(x,y, \"go\", label = \"G.difficilis\")\n \n x = outgroup_distance(finch_data, x_column, outgroup)\n y = outgroup_distance(finch_data, y_column, outgroup)\n plt.plot(x,y, \"ko\", label = \"G.magnirostris\")\n \n plt.xlabel('Gene Distance')\n plt.ylabel('Beak Distance')\n plt.legend(shadow=True, loc=\"upper right\")\n plt.title(\"Beak Distance vs. Gene Distance for Finches\")\n \nplot_data (\"finches.csv\", 0, 1)\n \n\n''' \n\n\n This dataset is perhaps the best known database in the pattern recognition literature. \n One flinch species is linearly separable from the other four, but the other three are not linearly separable from each other.\n\n G.difficilis is linearly seperate when plotting Beak Distance vs. Gene Distance for Finches. G.conirostris_Espanola, G.conirostris_Genovesa, and G.magnirostris are not linearly separable from each other when plotting Beak Distance vs. Gene Distance for Finches.\n The two species would cluster together if they are indeed different species because they have similar attributes of Finches such as the obvious from the scatterplot, Beak Distance lineraly alligning with Gene Distance.\n\n\n'''\n", "sub_path": "project_02.py", "file_name": "project_02.py", "file_ext": "py", "file_size_in_byte": 5438, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "csv.reader", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 120, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 124, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}]} +{"seq_id": "533275529", "text": "import os\nfrom http.cookiejar import MozillaCookieJar\n\nfrom macaroonbakery import httpbakery\n\n\ndef run(url, request, payload=None):\n client = httpbakery.Client(cookies=MozillaCookieJar(\".login\"))\n\n if os.path.exists(client.cookies.filename):\n client.cookies.load(ignore_discard=True)\n\n if payload:\n response = client.request(request, url=url, json=payload)\n else:\n response = client.request(request, url=url)\n\n client.cookies.save(ignore_discard=True)\n print(response, response.text)\n\n\nnew_release = {\n \"name\": \"New Version\",\n \"version\": \"30.30\",\n \"codename\": \"version\",\n \"lts\": False,\n \"development\": True,\n \"release_date\": \"2021-04-22\",\n \"esm_expires\": \"2022-01-31\",\n \"support_expires\": \"2022-01-31\",\n}\n\nhippo = {\n \"name\": \"Hirsute Hippo\",\n \"version\": \"21.04\",\n \"codename\": \"hirsute\",\n \"lts\": False,\n \"development\": True,\n \"release_date\": \"2021-04-22\",\n \"esm_expires\": \"2022-01-31\",\n \"support_expires\": \"2022-01-31\",\n}\n\n\n# RELEASE TESTS\nbase_url = \"http://0.0.0.0:8030/security\"\n\nprint(\"Test create release => 200\")\nrun(f\"{base_url}/releases.json\", \"POST\", new_release)\n\nprint(\"Test edit release => 404\")\nrun(f\"{base_url}/releases/no-exist.json\", \"PUT\", new_release)\n\nprint(\"Test edit release => 422\")\nhippo[\"name\"] = \"Bionic Beaver\"\nrun(f\"{base_url}/releases/hirsute.json\", \"PUT\", hippo)\n\nprint(\"Test edit release => 200\")\nhippo[\"name\"] = \"Hirsute Hippo\"\nhippo[\"development\"] = True\nrun(f\"{base_url}/releases/hirsute.json\", \"PUT\", hippo)\n\nprint(\"Test delete non-existing release => 404\")\nrun(f\"{base_url}/releases/no-exist.json\", \"DELETE\")\n\nprint(\"Test delete release => 200\")\nrun(f\"{base_url}/releases/version.json\", \"DELETE\")\n", "sub_path": "scripts/api-release-tests.py", "file_name": "api-release-tests.py", "file_ext": "py", "file_size_in_byte": 1718, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "macaroonbakery.httpbakery.Client", "line_number": 8, "usage_type": "call"}, {"api_name": "macaroonbakery.httpbakery", "line_number": 8, "usage_type": "name"}, {"api_name": "http.cookiejar.MozillaCookieJar", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}]} +{"seq_id": "471574668", "text": "import pandas as pd\r\nfrom matplotlib import pyplot as plt\r\nimport numpy as np\r\nfrom log_regression2 import log_regression\r\nfrom NaiveBayes3 import NaiveBayes\r\nfrom cross_validation2 import cross_validation\r\nimport separate\r\n\r\n# dictionary of categories: contains a dictionary for each feature, mapping an integer\r\n# to each category\r\ncategories = {}\r\n\r\n# Makes all the inputs of the dataset numerical\r\ndef transform(data):\r\n for j in range(len(data[0])):\r\n \r\n \r\n category_number = 0 # keeps track of the category number in the feature\r\n for i in range(len(data)):\r\n ## All features are categorical, even if some inputs are numerical \r\n # Ignore features that are already numerical\r\n #if (isinstance(data[i][j], int) or isinstance(data[i][j], float)):\r\n # break\r\n \r\n # Create a dictionary for a feature if it does not exist\r\n if j not in categories:\r\n categories[j] = {}\r\n \r\n # Add a numerical value to each category\r\n if data[i][j] not in categories[j]:\r\n categories[j][data[i][j]] = category_number\r\n category_number += 1\r\n \r\n data[i][j] = categories[j][data[i][j]]\r\n \r\n# Makes data one-hot encoded\r\ndef oneHot(data): \r\n one_hot = [] \r\n for i in range(len(data)):\r\n one_hot.append(list())\r\n \r\n for j in range(len(data[0])):\r\n # Just copy non-categorical features\r\n if j not in categories:\r\n one_hot[i].append(data[i][j])\r\n # Make categorical features one-hot encoded\r\n else:\r\n # Array of 0s of the length of the categories in that feature\r\n temp = [0]*len(categories[j])\r\n \r\n # Put 1 to the correct category\r\n temp[data[i][j]] = 1\r\n \r\n one_hot[i].extend(temp)\r\n \r\n one_hot = np.array(one_hot) \r\n return one_hot \r\n \r\n \r\n\r\n\r\ndata = pd.read_csv(\"car.csv\", header=None)\r\n\r\n# Transform data in numpy arrays\r\ndata = data.values\r\n\r\ndata_temp = np.zeros(data.shape, dtype = 'O')\r\n \r\nrow = 0\r\n# Detect oddities\r\nfor i in range(len(data)):\r\n \r\n # Test for missing data (row length)\r\n if len(data[0]) != len(data[i]):\r\n print(\"values on row \",i, \" are missing\")\r\n \r\n # Test for malformed features\r\n include = True # Only include rows that don't have missing features (?)\r\n for j in range(len(data[0])):\r\n # remove leading/trailing spaces\r\n if isinstance(data[i][j], str):\r\n data[i][j] = data[i][j].strip()\r\n \r\n # remove instances with missing data\r\n if data[i][j] == \"?\":\r\n include = False\r\n break\r\n \r\n # Only include rows that don't have missing features (?)\r\n if include == True:\r\n data_temp[row,:] = data[i,:]\r\n row += 1 \r\n\r\ndata = data_temp[0:row] \r\ndata_raw = data.copy()\r\n\r\n# Transform data into numerical values\r\ntransform(data)\r\n\r\n\r\nres = np.zeros(len(data))\r\n\r\n# Transform targets in a binary representation\r\n# combine unacc into bad and acc, good and vgood into good\r\n \r\nfor i in range(len(data)):\r\n #if data[i][len(data[0])-1] == 0 or data[i][len(data[0])-1] == 1:\r\n if data[i][len(data[0])-1] != 0:\r\n res[i] = 1\r\n else:\r\n res[i] = 0\r\n \r\n# if data[i][len(data[0])-1] == \">50K\":\r\n# res[i]=1\r\n# elif data[i][len(data[0])-1] == \"<=50K\":\r\n# res[i]=0\r\n\r\n# delete target from data (last column from data)\r\ndata = np.delete(data, len(data[0])-1, 1)\r\n\r\n# Make the input categories one_hot encoded\r\none_hot = oneHot(data)\r\n\r\n#print(data)\r\n\r\n# Histogram of the targets\r\nplt.figure(1)\r\n#plt.hist(res) \r\nplt.hist([res[np.argwhere(res == 0)], res[np.argwhere(res == 1)]], label=['bad', 'good'])\r\nplt.legend(loc='upper right') \r\nplt.title(\"Distribution of the positive vs negative classes\") \r\nplt.show()\r\n\r\n# Note: there are no numerical features in this dataset\r\n# Distributions of some categorical features (feature columns 0,1,2,3 were considered)\r\nf = (0,1,2,3)\r\n\r\npos = np.argwhere(res == 1)\r\nneg = np.argwhere(res == 0)\r\n\r\n# matrices (feature, data point) - separation between positive and negative features\r\npos_features = np.zeros((4,len(pos))) \r\nneg_features = np.zeros((4,len(neg)))\r\nfor i in range(4):\r\n neg_features[i,:] = np.squeeze(data[neg, f[i]])\r\n pos_features[i,:] = np.squeeze(data[pos, f[i]])\r\n\r\n\r\nplt.figure(2)\r\n\r\nfor i in range(4):\r\n plt.subplot(2,2,i+1)\r\n \r\n # Set bin boundaries by the minimum and maximum values of the features\r\n bins = np.linspace(min(min(neg_features[i,:]), min(pos_features[i,:])),\r\n max(max(neg_features[i,:]), max(pos_features[i,:])), 30)\r\n \r\n # Plot the histogram of the positive and negative features\r\n plt.hist([neg_features[i,:], pos_features[i,:]], bins, label=['neg', 'pos'])\r\n plt.legend(loc='upper right') \r\n \r\n plt.title(\"Distribution of feature #\" + str(f[i]))\r\n\r\nplt.show()\r\n\r\n# This dataset does not have numerical features, so correlation between features is not helpful\r\n# Correlation between some numerical features (feature columns 2,3,4,5 were considered)\r\n\r\n#plt.figure(3)\r\n\r\n#for i in range(4):\r\n #plt.subplot(2,2,i+1)\r\n \r\n ## Correlation coefficients\r\n #r_neg = np.corrcoef(neg_features[i,:], neg_features[(i+1)%4,:])\r\n #r_pos = np.corrcoef(pos_features[i,:], pos_features[(i+1)%4,:])\r\n \r\n ## Labels for the legend\r\n #lbl_neg = \"r_neg = \" + str(round(r_neg[0,1],4))\r\n #lbl_pos = \"r_pos = \" + str(round(r_pos[0,1],4))\r\n \r\n #plt.scatter(neg_features[i,:], neg_features[(i+1)%4,:], label=lbl_neg)\r\n #plt.scatter(pos_features[i,:], pos_features[(i+1)%4,:], label=lbl_pos)\r\n \r\n #plt.legend(loc='upper right') \r\n \r\n #plt.title(\"Correlation between feature #\" + str(f[i]) + \" and #\" + str(f[(i+1)%4])) \r\n \r\n\r\n#plt.show()\r\n\r\n# Final data variables X and target variables Y\r\nX = np.array(one_hot)\r\nY = np.array(res)\r\n\r\n##fit log model\r\n#log_model = log_regression(0.01, 20000)\r\n#X = log_model.bias(X) # add bias column\r\n\r\n## Separate training and testing sets \r\n#X_train, Y_train, X_test, Y_test = separate.separate(X,Y)\r\n\r\n## train the data\r\n#fit_iono = log_model.fit(X_train,Y_train) \r\n\r\n## test data\r\n#pre = log_model.predict(X_test,fit_iono) \r\n#acc = log_model.evaluate_acc(pre,Y_test)\r\n#print(acc)\r\n\r\n## Cross validation\r\n#validation = cross_validation(5)\r\n#score = validation.evaluate_log(X_train,Y_train)\r\n#print(score)\r\n\r\n#print(\"Naive Bayes:\")\r\n## fit naive bayes\r\n#bayes_model = NaiveBayes()\r\n#fit_bayes = bayes_model.fit(X_train,Y_train)\r\n#pre = bayes_model.predict(X_test)\r\n##acc = log_model.evaluate_acc(pre,Y_test)\r\n#acc = bayes_model.evaluate_acc(pre,Y_test)\r\n#print(acc)\r\n\r\n## Cross validation\r\n#score = bayes_model.cross_validation(X_train,Y_train, 5)\r\n#print(score)\r\n\r\n### Compare accuracy of naive Bayes and logistic regression\r\n\r\n## All datasets will use for logistic regression the same learning rate = 0.01 and # iterations = 500\r\n#rate = 0.01\r\n#iterations = 500\r\n\r\n#log_model = log_regression(rate, iterations)\r\n#X = log_model.bias(X) # add bias column\r\n\r\n## Separate training and testing sets \r\n#X_train, Y_train, X_test, Y_test = separate.separate(X,Y)\r\n\r\n\r\n\r\n## Compare accuracy of naive Bayes and logistic regression\r\n\r\n# All datasets will use for logistic regression the same learning rate = 0.01 and # iterations = 500\r\nrate = 0.01\r\niterations = 500\r\n\r\nlog_model = log_regression(rate, iterations)\r\nX = log_model.bias(X) # add bias column\r\n\r\n# Separate training and testing sets \r\nX_train, Y_train, X_test, Y_test = separate.separate(X,Y)\r\n\r\n## Logistic regression\r\n\r\n# train the data\r\nfit_iono = log_model.fit(X_train,Y_train) \r\n\r\n# Cross validation\r\nvalidation = cross_validation(rate, max_iterations = 500)\r\nscore = validation.evaluate_log(X_train,Y_train)\r\nprint(\"Averaged training accuracy for Logistic Regression: \", score)\r\n\r\n# Test data\r\npre = log_model.predict(X_test,fit_iono) \r\nacc = log_model.evaluate_acc(pre,Y_test)\r\nprint(\"Accuracy on testing data for Logistic Regression: \", acc)\r\n\r\n## Naive Bayes\r\n\r\n# train the data\r\nbayes_model = NaiveBayes()\r\nfit_bayes = bayes_model.fit(X_train,Y_train)\r\n\r\n\r\n# Cross validation\r\nscore = bayes_model.cross_validation(X_train,Y_train)\r\nprint(\"Averaged training accuracy for Naive Bayes: \", score)\r\n\r\n\r\n# Test data\r\npre = bayes_model.predict(X_test)\r\nacc = bayes_model.evaluate_acc(pre,Y_test)\r\nprint(\"Accuracy on testing data for Naive Bayes: \", acc)\r\n\r\nprint()\r\n## Test different learning rates for gradient descent\r\n\r\n# Loss function threshold = 5*10^-5; maximum number of iterations = 1000\r\niters = []\r\nacc=[]\r\nrate = 10**-15\r\nfor i in range(20):\r\n \r\n # Cross validation\r\n validation = cross_validation(rate, threshold = True)\r\n score, iterations = validation.evaluate_log(X_train,Y_train)\r\n #print(\"Averaged training accuracy for Logistic Regression: \", score)\r\n print(\"rate = \", rate, \"; iterations = \", iterations, \"; accuracy = \", score)\r\n acc.append(score)\r\n iters.append(iterations)\r\n rate *= 10\r\n\r\n\r\n\r\nrate = 1\r\n\r\nplt.scatter(iters, acc)\r\nplt.xlabel(\"iterations\")\r\nplt.ylabel(\"accurary\")\r\nplt.title(\"the accuracy on train set as a function of iterations of gradient descent\")\r\n\r\nplt.show()\r\n#size of x and accuracy\r\nacc = []\r\nsize = []\r\nsplit_size = 0.1\r\n\r\nfor i in range(9):\r\n X_train, Y_train, X_test, Y_test = separate.separate(X,Y, split=split_size)\r\n # Cross validation\r\n validation = cross_validation(rate, threshold = True)\r\n score, iterations = validation.evaluate_log(X_train,Y_train)\r\n #print(\"Averaged training accuracy for Logistic Regression: \", score)\r\n print(\"size of X = \", X_train.shape[0], \"; iterations = \", iterations, \"; accuracy = \", score)\r\n size.append(X_train.shape[0])\r\n acc.append(score)\r\n split_size += 0.1\r\n \r\nplt.scatter(size, acc)\r\nplt.xlabel(\"size of X_train\")\r\nplt.ylabel(\"accurary\")\r\nplt.title(\"the accuracy on train set as a function of size of X on logistic model\")\r\n\r\nplt.show()\r\n#bayes model\r\nacc = []\r\nsize = []\r\nsplit_size = 0.1\r\n\r\nfor i in range(9):\r\n X_train, Y_train, X_test, Y_test = separate.separate(X,Y, split=split_size)\r\n # Cross validation\r\n score = bayes_model.cross_validation(X_train,Y_train)\r\n #print(\"Averaged training accuracy for Logistic Regression: \", score)\r\n print(\"size of X = \", X_train.shape[0], \"; accuracy = \", score)\r\n size.append(X_train.shape[0])\r\n acc.append(score)\r\n split_size += 0.1\r\n \r\nplt.xlabel(\"size of X_train\")\r\nplt.ylabel(\"accurary\")\r\nplt.title(\"the accuracy on train set as a function of size od X on Naive Bayes model\")\r\nplt.scatter(size, acc)\r\nplt.show()", "sub_path": "car.py", "file_name": "car.py", "file_ext": "py", "file_size_in_byte": 10839, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "numpy.array", "line_number": 56, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}, {"api_name": "numpy.argwhere", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "numpy.argwhere", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 139, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 149, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 152, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 159, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 159, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 160, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 162, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 162, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 164, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 193, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 194, "usage_type": "call"}, {"api_name": "log_regression2.log_regression", "line_number": 249, "usage_type": "call"}, {"api_name": "separate.separate", "line_number": 253, "usage_type": "call"}, {"api_name": "cross_validation2.cross_validation", "line_number": 261, "usage_type": "call"}, {"api_name": "NaiveBayes3.NaiveBayes", "line_number": 273, "usage_type": "call"}, {"api_name": "cross_validation2.cross_validation", "line_number": 297, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 309, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 309, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 310, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 310, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 311, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 311, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 312, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 312, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 314, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 314, "usage_type": "name"}, {"api_name": "separate.separate", "line_number": 321, "usage_type": "call"}, {"api_name": "cross_validation2.cross_validation", "line_number": 323, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 331, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 331, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 332, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 332, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 333, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 333, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 334, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 334, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 336, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 336, "usage_type": "name"}, {"api_name": "separate.separate", "line_number": 343, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 352, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 352, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 353, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 353, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 354, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 354, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 355, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 355, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 356, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 356, "usage_type": "name"}]} +{"seq_id": "423177826", "text": "import pandas as pd\nimport numpy as np\nimport pickle\n\nfrom sklearn.pipeline import Pipeline\nfrom sklearn.compose import ColumnTransformer\nfrom sklearn.impute import SimpleImputer\n\nfrom sklearn.preprocessing import StandardScaler\nfrom sklearn.ensemble import RandomForestClassifier\n\nCAT_FEATS = ['Gender',\n 'Married',\n 'Dependents',\n 'Education',\n 'Self_Employed',\n 'Property_Area']\n\nNUM_FEATS = ['ApplicantIncome',\n 'CoapplicantIncome',\n 'LoanAmount',\n 'Loan_Amount_Term',\n 'Credit_History']\n\n###########################\n# DATA PROCESSING FUNCTIONS\n###########################\n\n\nclass Transformer():\n def __init__(self, func):\n self.func = func\n\n def transform(self, input_df, **transform_params):\n return self.func(input_df)\n\n def fit(self, X, y=None, **fit_params):\n return self\n\n\ndef outliers(data):\n \"\"\"\n Removes ApplicantIncome and Loan Amount rows less than or equal to three standard deviations from mean\n \"\"\"\n data = data[np.abs(data.ApplicantIncome-data.ApplicantIncome.mean()) <= (3*data.ApplicantIncome.std())]\n data = data[np.abs(data.LoanAmount-data.LoanAmount.mean()) <= (3*data.LoanAmount.std())]\n return data\n\n\ndef impute_credit(data):\n \"\"\"\n Imputes credit history binary value.\n Probability for this random choice is hardcoded and based on the distribution of the population.\n \"\"\"\n data['Credit_History'] = data['Credit_History'].fillna(pd.Series(np.random.choice([1.0,0.0],\n p=[0.842199, 0.157801],\n size=len(data))))\n return data\n\n\ndef impute_gender(data):\n \"\"\"\n Imputes gender binary value. Non-Binary gender was not included in dataset.\n Probability for this random choice is hardcoded and based on the distribution of the population.\n \"\"\"\n data['Gender'] = data['Gender'].fillna(pd.Series(np.random.choice(['Male','Female'],\n p=[0.81, 0.19],\n size=len(data))))\n return data\n\n\ndef impute_marriage(data):\n data['Married'] = data['Married'].fillna(pd.Series(np.random.choice(['Yes','No'], \n p=[0.65, 0.35],\n size=len(data))))\n data['Dependents'] = data['Dependents'].replace('3+', 3)\n data['Dependents'] = data[['Dependents']].fillna(0).astype('int16')\n return data\n\n\ndef impute_employment(data):\n data['Self_Employed'] = data['Self_Employed'].fillna(pd.Series(np.random.choice(['Yes','No'],\n p=[0.86, 0.14],\n size=len(data))))\n return data\n\n\ndef binarizer(data):\n \"\"\"\n Manually label encodes binary features\n \"\"\"\n data['Male'] = np.where(data['Gender'] == 'Male', 1, 0)\n data['Graduated'] = np.where(data['Education'] == 'Graduate', 1, 0)\n data['Married'] = np.where(data['Married'] == 'Yes', 1, 0)\n data['Self_Employed'] = np.where(data['Self_Employed'] == 'Yes', 1, 0)\n return data\n\n\ndef dummy(data):\n data = pd.get_dummies(data, columns=['Property_Area'])\n return data\n\n\ndef dummies(data):\n \"\"\"\n Manually creates labels for property type. sklearn label encoder was breaking the model shape.\n \"\"\"\n data['Property_Area'] = data['Property_Area'].replace('Rural', 0, regex=True)\n data['Property_Area'] = data['Property_Area'].replace('Semiurban', 1, regex=True)\n data['Property_Area'] = data['Property_Area'].replace('Urban', 2, regex=True)\n return data\n\n\ndef shed(data):\n \"\"\"\n Drops columns that have been encoded\n \"\"\"\n data = data.drop(['Gender','Education','Married','Self_Employed'],axis=1)\n return data\n\n\ndef impute_loan_term(data):\n \"\"\"\n Imputes loan term value with the mean loan term.\n \"\"\"\n data['Loan_Amount_Term'] = data['Loan_Amount_Term'].fillna(data['Loan_Amount_Term'].mean())\n data['Loan_Amount_Term'] = data['Loan_Amount_Term']/12\n return data\n\n#######################\n# PIPELINE\n#######################\n#\n# Numerical and categorical data will have separate pipelines.\n# Both pieplines feed into the prediction pipeline.\n\n\nnumeric_transformer = Pipeline(steps=[\n ('credit', Transformer(impute_credit)),\n ('term', Transformer(impute_loan_term)),\n ('imputer', SimpleImputer(strategy='mean', fill_value='missing')),\n ('scaler', StandardScaler())\n ])\n\ncategorical_transformer = Pipeline(steps=[\n ('gender', Transformer(impute_gender)),\n ('marriage', Transformer(impute_marriage)),\n ('employment', Transformer(impute_employment)),\n ('dummies', Transformer(dummies)),\n ('shed', Transformer(shed)),\n ])\n\npreprocessor = ColumnTransformer(\n transformers=[\n ('num', numeric_transformer, NUM_FEATS),\n ('cat', categorical_transformer, CAT_FEATS)])\n\nrf_clf = Pipeline(steps=[('preprocessor', preprocessor),\n ('rf_clf', RandomForestClassifier())])", "sub_path": "functions.py", "file_name": "functions.py", "file_ext": "py", "file_size_in_byte": 5352, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "numpy.abs", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 46, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 55, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 66, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 73, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 82, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 95, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 100, "usage_type": "call"}, {"api_name": "sklearn.pipeline.Pipeline", "line_number": 138, "usage_type": "call"}, {"api_name": "sklearn.impute.SimpleImputer", "line_number": 141, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 142, "usage_type": "call"}, {"api_name": "sklearn.pipeline.Pipeline", "line_number": 145, "usage_type": "call"}, {"api_name": "sklearn.compose.ColumnTransformer", "line_number": 153, "usage_type": "call"}, {"api_name": "sklearn.pipeline.Pipeline", "line_number": 158, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 159, "usage_type": "call"}]} +{"seq_id": "166149396", "text": "# -*- coding: utf-8 -*-\n\"\"\"\n flask.ext.fragment\n ------------------\n\n Flask extension to implement fragment caching.\n\n :copyright: (c) 2013 by Alexey Poryadin.\n :license: MIT, see LICENSE for more details.\n\"\"\"\nimport flask\nimport jinja2\nimport inspect\nfrom functools import partial\nfrom flask import Flask, Blueprint\nfrom flask import _app_ctx_stack as stack\n\n\nclass Fragment(object):\n\n def __init__(self, app=None):\n self.mod = None\n self.endpoint_url = None\n self.resethandler = None\n self.app = app\n if app is not None:\n self.init_app(app)\n\n def __call__(self, mod, endpoint_url=None, resethandler=None):\n \"\"\"Decorator to define function as fragment cached view\n\n Args:\n mod: Flask app or blueprint\n endpoint_url: Access point\n \"\"\"\n self.mod = mod\n self.endpoint_url = endpoint_url\n self.resethandler = resethandler\n\n def decorator(fragment_view):\n fragment_view_name = fragment_view.__name__\n fragment_view.cache_endpoint_url = self.endpoint_url if self.endpoint_url else fragment_view_name\n fragment_view.cache_resethandler = resethandler\n if self.endpoint_url:\n rule = '/{0}'.format(self.endpoint_url)\n elif isinstance(mod, Blueprint):\n rule = '/{0}.{1}'.format(mod.name, fragment_view_name)\n else:\n rule = '/{0}'.format(fragment_view_name)\n fragment_view_signature = inspect.signature(fragment_view)\n fragment_view.args_names = fragment_view_signature.parameters\n for arg_name in fragment_view.args_names:\n rule += '/<{0}>'.format(arg_name)\n mod.add_url_rule(rule, fragment_view_name, fragment_view)\n return fragment_view\n return decorator\n\n def init_app(self, app):\n self.app = app\n self.app.context_processor(lambda: {'fragment': self._fragment_tmpl_func})\n \n def resethandler(self, fragment_view):\n \"\"\"Decorator sets reset fragment cache handler for `fragment_view` function.\"\"\"\n def decorator(handler):\n fragment_view.cache_resethandler = handler\n return handler\n return decorator \n\n def reset(self, target, *args, **kwargs):\n \"\"\"Resets cache for fragment cached view\n \n Args:\n target: Endpoint or the view itself.\n \"\"\"\n if isinstance(target, str):\n fragment_view = flask.current_app.view_functions.get(target)\n if fragment_view is None:\n raise ValueError('Not found view for endpoint \"{0}\"'.format(target))\n else:\n fragment_view = target\n if fragment_view.cache_resethandler is None:\n # Tries default resethandler handler\n try:\n for N in range(0, len(args)):\n kwargs[fragment_view.args_names[N]] = args[N]\n url = flask.url_for(fragment_view.cache_endpoint_url, **kwargs)\n except Exception as exc:\n raise RuntimeError('Cannot reset cache for \"{0}\",'\n ' resethandler is not set and default handler canot'\n ' build URL. Detail: \"{1}\"'.format(fragment_view, exc))\n self.reset_url(url)\n else:\n fragment_view.cache_resethandler(*args, **kwargs)\n \n \n def reset_url(self, url):\n \"\"\"Resets cache for URL\n \n Args:\n url: URL value\n \"\"\"\n raise NotImplementedError('Need to look around ngx_cache_purge')\n\n def _fragment_tmpl_func(self, endpoint, *args, **kwargs):\n \"\"\"Template context function that renders fragment cached view.\n \n Accepts `*args`, `**kwargs` that must match by the number and by the\n order of parameters from function that defined with 'endpoint'.\n \n Args:\n endpoint: The endpoint name.\n \"\"\"\n func = flask.current_app.view_functions.get(endpoint)\n if func is not None:\n for N in range(0, len(args)):\n kwargs[func.args_names[N]] = args[N]\n url = flask.url_for(endpoint, **kwargs)\n return self._render(url, partial(func, **kwargs))\n raise ValueError('Not found view for endpoint \"{0}\"'.format(endpoint))\n\n def _render(self, url, deferred_view):\n if self.app.config.get('FRAGMENT_SSI', False):\n content = ''.format(url)\n else:\n response = deferred_view()\n content = response.get_data().decode(response.mimetype_params['charset'])\n return jinja2.Markup(content)\n", "sub_path": "flask_ssi/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 4729, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "flask.Blueprint", "line_number": 46, "usage_type": "argument"}, {"api_name": "inspect.signature", "line_number": 50, "usage_type": "call"}, {"api_name": "flask.current_app.view_functions.get", "line_number": 76, "usage_type": "call"}, {"api_name": "flask.current_app", "line_number": 76, "usage_type": "attribute"}, {"api_name": "flask.url_for", "line_number": 86, "usage_type": "call"}, {"api_name": "flask.current_app.view_functions.get", "line_number": 113, "usage_type": "call"}, {"api_name": "flask.current_app", "line_number": 113, "usage_type": "attribute"}, {"api_name": "flask.url_for", "line_number": 117, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 118, "usage_type": "call"}, {"api_name": "jinja2.Markup", "line_number": 127, "usage_type": "call"}]} +{"seq_id": "595059139", "text": "# coding=utf-8\n# --------------------------------------------------------------------------\n# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License. See License.txt in the project root for\n# license information.\n#\n# Code generated by Microsoft (R) AutoRest Code Generator.\n# Changes may cause incorrect behavior and will be lost if the code is\n# regenerated.\n# --------------------------------------------------------------------------\n\nfrom msrest.serialization import Model\n\n\nclass ContainerExecRequestTerminalSize(Model):\n \"\"\"The size of the terminal.\n\n :param rows: The row size of the terminal\n :type rows: int\n :param cols: The column size of the terminal\n :type cols: int\n \"\"\"\n\n _attribute_map = {\n 'rows': {'key': 'rows', 'type': 'int'},\n 'cols': {'key': 'cols', 'type': 'int'},\n }\n\n def __init__(self, rows=None, cols=None):\n super(ContainerExecRequestTerminalSize, self).__init__()\n self.rows = rows\n self.cols = cols\n", "sub_path": "azure-mgmt-containerinstance/azure/mgmt/containerinstance/models/container_exec_request_terminal_size.py", "file_name": "container_exec_request_terminal_size.py", "file_ext": "py", "file_size_in_byte": 1027, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "msrest.serialization.Model", "line_number": 15, "usage_type": "name"}]} +{"seq_id": "30590007", "text": "#!/usr/bin/env python3\n# coding=utf-8\nimport os\nimport sys\nfrom time import sleep\nimport re\nimport socket\nimport shutil\nimport subprocess\nimport logging\nimport traceback\nfrom urllib.parse import urljoin\n\n__AUTHOR__ = 'Aploium '\n__VERSION__ = '0.7.0'\n__ZMIRROR_PROJECT_URL__ = 'https://github.com/aploium/zmirror/'\n__ZMIRROR_GIT_URL__ = 'https://github.com/aploium/zmirror.git'\n__ONKEY_PROJECT_URL__ = 'https://github.com/aploium/zmirror-onekey/'\n__ONKEY_PROJECT_URL_CONTENT__ = 'https://raw.githubusercontent.com/aploium/zmirror-onekey/master/'\n__REPORT_URLS__ = {\n \"error\": \"https://report.zmirror.org/onekey/log/error\",\n \"success\": \"https://report.zmirror.org/onekey/log/success\",\n}\n\nDEBUG = '--debug' in sys.argv\n\n\ndef cmd(command, cwd=None, **kwargs):\n \"\"\"运行shell命令\"\"\"\n return subprocess.check_call(command, shell=True, cwd=cwd or os.getcwd(), **kwargs)\n\n\ncmd('export LC_ALL=C.UTF-8') # 设置bash环境为utf-8\n\ncmd('apt-get update && apt-get install python3 python3-pip -y')\n\n# for some old version Linux, pip has bugs, causing:\n# ImportError: cannot import name 'IncompleteRead'\n# so we need to upgrade pip first\ncmd('easy_install3 -U pip')\n\n# 安装本脚本必须的python包\ncmd('python3 -m pip install -U requests')\ncmd('python3 -m pip install -U distro')\n\nimport distro\n\n\ndef onekey_report(report_type=\"success\", installing_mirror=None, traceback_str=None):\n \"\"\"\n 发送报告到服务器\n \"\"\"\n import json\n import distro\n import re\n\n dist = json.dumps(distro.info(best=True))\n data = {\"linux_dist\": dist}\n\n if installing_mirror is not None:\n if isinstance(installing_mirror, (list, tuple)):\n installing_mirror = ','.join(installing_mirror)\n data['installing_mirror'] = installing_mirror\n\n if traceback_str is not None:\n data['traceback'] = traceback_str\n\n try:\n meminfo = open('/proc/meminfo').read()\n matched = re.search(r'^MemTotal:\\s+(\\d+)', meminfo)\n if matched:\n mem_total_KB = int(matched.groups()[0])\n data['memory'] = mem_total_KB\n except:\n pass\n\n if DEBUG:\n print(__REPORT_URLS__[report_type], data)\n\n try:\n r = requests.post(__REPORT_URLS__[report_type], data=data)\n except:\n if DEBUG:\n traceback.print_exc()\n else:\n if DEBUG:\n print(r.text, r.headers, r.request.body)\n\n\ntry:\n import requests\nexcept:\n print('Could not install requests, program exit')\n exit(1)\n\nif DEBUG:\n logging.basicConfig(\n level=logging.NOTSET,\n format='[%(levelname)s %(asctime)s %(funcName)s] %(message)s',\n )\n\nif sys.platform != 'linux':\n print('This program can ONLY be used in debian-like Linux (debian, ubuntu and some others)')\n exit(1)\nif os.geteuid() != 0:\n print('Root privilege is required for this program. Please use `sudo python3 deploy.py`')\n exit(2)\n\nserver_configs = {\n \"apache\": {\n \"config_root\": \"/etc/apache2/\",\n \"htdoc\": \"/var/www/\",\n\n \"common_configs\": [\"http_generic\", \"apache_boilerplate\"],\n \"site_unique_configs\": [\"https\"],\n\n \"pre_delete_files\": [\n \"{config_root}/sites-enabled/000-default.conf\",\n \"{config_root}/conf-enabled/apache2-doc.conf\",\n \"{config_root}/conf-enabled/security.conf\",\n ],\n\n \"configs\": {\n \"apache_boilerplate\": {\n \"url\": urljoin(__ONKEY_PROJECT_URL_CONTENT__, \"configs/apache2-boilerplate.conf\"),\n \"file_path\": \"conf-enabled/zmirror-apache-boilerplate.conf\",\n },\n\n \"http_generic\": {\n \"url\": urljoin(__ONKEY_PROJECT_URL_CONTENT__, \"configs/apache2-http.conf\"),\n \"file_path\": \"sites-enabled/zmirror-http-redirection.conf\",\n },\n\n \"https\": {\n \"url\": urljoin(__ONKEY_PROJECT_URL_CONTENT__, \"configs/apache2-https.conf\"),\n \"file_path\": \"sites-enabled/zmirror-{mirror_name}-https.conf\",\n },\n }\n\n }\n}\n\nmirrors_settings = {\n 'google': {\n 'domain': None,\n 'cfg': [('more_configs/config_google_and_zhwikipedia.py', 'config.py'), ],\n },\n\n 'youtubePC': {\n 'domain': None,\n 'cfg': [('more_configs/config_youtube.py', 'config.py'),\n ('more_configs/custom_func_youtube.py', 'custom_func.py')],\n },\n\n 'twitterPC': {\n 'domain': None,\n 'cfg': [('more_configs/config_twitter_pc.py', 'config.py'),\n ('more_configs/custom_func_twitter.py', 'custom_func.py'), ],\n },\n\n 'twitterMobile': {\n 'domain': None,\n 'cfg': [('more_configs/config_twitter_mobile.py', 'config.py'),\n ('more_configs/custom_func_twitter.py', 'custom_func.py'), ],\n },\n\n 'instagram': {\n 'domain': None,\n 'cfg': [('more_configs/config_instagram.py', 'config.py'), ],\n },\n}\n\nprint('OneKey deploy script for zmirror. version', __VERSION__)\nprint('This script will automatically deploy mirror(s) using zmirror in your ubuntu')\nprint('You could cancel this script in the config stage by precessing Ctrl-C')\nprint('Installation will start after 1 second')\nprint()\nsleep(1)\n\n# ################# 安装一些依赖包 ####################\nprint('[zmirror] Installing some necessarily packages')\n\ntry:\n # 设置本地时间为北京时间\n cmd('cp /usr/share/zoneinfo/Asia/Shanghai /etc/localtime')\n # 告诉apt-get要安静\n cmd('export DEBIAN_FRONTEND=noninteractive')\n # 更新apt-get\n cmd('apt-get update')\n # 安装必须的包\n cmd('apt-get install git python3 python3-pip wget curl -y')\n # 安装非必须的包\n try:\n # 更新一下openssl\n cmd('apt-get install openssl -y')\n # 如果安装了, 则可以启用http2\n cmd('apt-get install software-properties-common python-software-properties -y')\n except:\n pass\n\n if distro.id() == 'ubuntu':\n # 安装高版本的Apache2(支持http2), 仅限ubuntu\n cmd(\"\"\"LC_ALL=C.UTF-8 add-apt-repository -y ppa:ondrej/apache2 &&\n apt-key update &&\n apt-get update &&\n apt-get install apache2 -y\"\"\")\n elif distro.id() == 'debian':\n # debian 只有低版本的可以用\n cmd(\"apt-get install apache2 -y\")\n\n cmd(\"\"\"a2enmod rewrite mime include headers filter expires deflate autoindex setenvif ssl http2\"\"\")\n\n # (可选) 更新一下各种包\n if not (distro.id() == 'ubuntu' and distro.version() == '14.04'): # 系统不是ubuntu 14.04\n # Ubuntu 14.04 执行本命令的时候会弹一个postfix的交互, 所以不执行\n cmd('apt-get upgrade -y')\n\n cmd(\"\"\"apt-get install libapache2-mod-wsgi-py3 -y && a2enmod wsgi\"\"\")\n\n # 安装和更新必须的python包\n cmd('python3 -m pip install -U requests flask')\n # 安装和更新非必须, 但是有好处的python包\n try:\n cmd('python3 -m pip install -U chardet fastcache cchardet')\n except:\n pass # 允许安装失败\n\n print('[zmirror] Dependency packages install completed')\n print('[zmirror] Installing letsencrypt...')\n sleep(1)\n\n if not os.path.exists('/etc/certbot/'):\n # certbot 不存在, 则安装\n cmd('git clone https://github.com/certbot/certbot.git', cwd='/etc/')\n cmd('chmod a+x /etc/certbot/certbot-auto', cwd='/etc/certbot/')\n cmd('service apache2 stop')\n cmd('./certbot-auto renew --agree-tos -n --standalone '\n '--pre-hook \"service apache2 stop\" '\n '--post-hook \"service apache2 start\"',\n cwd='/etc/certbot/')\n else:\n # 否则升级一下\n cmd('git pull', cwd='/etc/certbot/')\n\n print(\"[zmirror] let's encrypt Installation Completed\")\n sleep(1)\n\n print('\\n\\n\\n-------------------------------\\n'\n '[zmirror] Now we need some information:')\nexcept:\n onekey_report('error', traceback_str=traceback.format_exc())\n raise\n\nmirrors_to_deploy = []\ntry:\n _input = -1\n while _input: # 不断循环输入, 因为用户可能想要安装多个镜像\n print('----------------------')\n _input = input(\n \"\"\"Please select mirror you want to deploy?\n select one mirror a time, you could select zero or more mirror(s)\n\n 1. Google (include scholar, image, zh_wikipedia) {google}\n 2. twitter (PC) {twitterPC}\n 3. twitter (Mobile) {twitterMobile}\n 4. youtube (pc) {youtubePC}\n 5. instagram {instagram}\n 0. Go to next steps. (OK, I have selected all mirror(s) I want to deploy)\n\n input 0-5: \"\"\".format(\n google='[SELECTED]' if 'google' in mirrors_to_deploy else '',\n twitterPC='[SELECTED]' if 'twitterPC' in mirrors_to_deploy else '',\n twitterMobile='[SELECTED]' if 'twitterMobile' in mirrors_to_deploy else '',\n youtubePC='[SELECTED]' if 'youtubePC' in mirrors_to_deploy else '',\n instagram='[SELECTED]' if 'instagram' in mirrors_to_deploy else '',\n )\n\n )\n\n if not _input:\n break\n\n logging.debug(\"input:\" + _input)\n\n try:\n _input = int(_input)\n except:\n print(\"Please input correct number\")\n _input = -1\n\n if _input == 0:\n break\n if not (0 <= _input <= 5):\n print('[ERROR] please input correct number (0-5), only select one mirror a time\\n'\n '-------------------------\\n\\n')\n continue\n\n # 将数字选项转化为字符串\n mirror_type = {\n 1: \"google\",\n 2: \"twitterPC\",\n 3: \"twitterMobile\",\n 4: \"youtubePC\",\n 5: \"instagram\",\n }[_input]\n\n # 在选项里, 镜像已存在, 则删去, 并且跳过下面的步骤\n if mirror_type in mirrors_to_deploy:\n mirrors_to_deploy.remove(mirror_type)\n print(\"Mirror:{mirror_type} unchecked.\".format(mirror_type=mirror_type))\n continue\n\n # 输入镜像对应的域名, 要求已经在DNS设置中用一个A记录指向了本服务器\n while True: # 这里面会检查输入的是否是三级域名\n domain = input(\"Please input your domain for this mirror: \")\n domain = domain.strip(' /.\\t').replace('https://', '').replace('http://', '') # 修剪\n if domain.count('.') != 2:\n if input((\"Your domain [{domain}] is not an third-level domain, \"\n \"which contains three parts and two dots. \\n\"\n \"eg1: lovelucia.zmirrordemo.com eg2: g.mymirror.com\\n\"\n \"zmirror officially only support third-level domain\\n\"\n \"a none third-level domain MAY work, but may cause potential errors\\n\"\n \"Continue anyway(y/N)?\"\n ).format(domain=domain)) in ('y', 'yes', 'Yes', 'YES'):\n break\n # 如果选择的是 N, 则重新输入\n else: # 输入的是三级域名\n break\n\n # 初步检验域名是否已经被正确设置\n try:\n domain_ip = socket.gethostbyname(domain)\n local_ip = socket.gethostbyname(socket.gethostname())\n except Exception as e: # 查询目标域名的IP失败\n print(\"Sorry, your domain [{domain}] is not setting correctly. {exc}\".format(domain=domain, exc=str(e)))\n continue_anyway = input(\"Continue anyway? (y/N): \")\n if continue_anyway not in ('y', 'yes', 'Yes', 'YES'):\n continue # 重新来\n else:\n # 仍然继续的话, 把domain_ip当做local_ip\n domain_ip = local_ip\n\n # 域名检验--目标域名的IP不等于本地机器的IP\n if domain_ip != local_ip:\n print(\"\"\"Sorry, your domain({domain})'s ip does not equals to this machine's ip.\n domain's ip is: {domain_ip}\n this machine's ip is: {local_ip}\n \"\"\".format(domain=domain, domain_ip=domain_ip, local_ip=local_ip)\n )\n continue_anyway = input(\"Continue anyway? (y/N): \")\n if continue_anyway not in ('y', 'yes', 'Yes', 'YES'):\n continue # 重新来\n\n # 域名检验--域名是否重复\n _dup_flag = False\n for mirror in mirrors_to_deploy:\n if mirrors_settings[mirror_type]['domain'] == domain:\n print(\"Duplicated domain! conflict with mirror: \" + mirror)\n _dup_flag = True\n break\n if _dup_flag:\n continue\n\n # 将镜像添加到待安装列表中\n mirrors_to_deploy.append(mirror_type)\n mirrors_settings[mirror_type]['domain'] = domain\n print(\"Mirror:{mirror_type} Domain:{domain} checked\".format(mirror_type=mirror_type, domain=domain))\n\n logging.debug(mirrors_to_deploy)\n\n if not mirrors_to_deploy:\n print('[ERROR] you didn\\'t select any mirror.\\nAbort installation')\n exit(4)\n\n email = input('Please input your email (because letsencrypt requires an email for certification)\\n')\n\n print('Your email:', email)\n\n # 最后确认一遍设置\n print('----------------------')\n print('Now, we are going to install, please check your settings here:')\n print(\"Email: \" + email)\n print()\n for mirror in mirrors_to_deploy:\n print(\"Mirror: {mirror} Domain: {domain}\".format(mirror=mirror, domain=mirrors_settings[mirror]['domain']))\n\n if input('Are these settings correct (Y/n)? ') in ('N', 'No', 'n', 'no', 'not', 'none'):\n print('installation aborted.')\n exit(5)\n\n # ############### Really Install ###################\n\n # 通过 letsencrypt 获取HTTPS证书\n print(\"Fetching HTTPS certifications\")\n cmd(\"service apache2 stop\") # 先关掉apache\n for mirror in mirrors_to_deploy:\n domain = mirrors_settings[mirror]['domain']\n\n if os.path.exists('/etc/letsencrypt/live/{domain}'.format(domain=domain)):\n # 如果证书已存在, 则跳过\n print(\"Certification for {domain} already exists, skipping\".format(domain=domain))\n continue\n\n print(\"Obtaining: {domain}\".format(domain=domain))\n cmd(\n ('./certbot-auto certonly -n --agree-tos -t -m \"{email}\" --standalone -d \"{domain}\" '\n '--pre-hook \"/usr/sbin/service apache2 stop\" '\n '--post-hook \"/usr/sbin/service apache2 start\"'\n ).format(email=email, domain=domain),\n cwd='/etc/certbot/'\n )\n\n # 检查是否成功获取证书\n if not os.path.exists('/etc/letsencrypt/live/{domain}'.format(domain=domain)):\n print('[ERROR] Could NOT obtain an ssl cert, '\n 'please check your DNS record, '\n 'and then run again.\\n'\n 'Installation abort')\n exit(3)\n print(\"Succeed: {domain}\".format(domain=domain))\n cmd(\"service apache2 start\") # 重新启动apache\n\n # ####### 安装zmirror自身 #############\n print('[zmirror] Successfully obtain SSL cert, now installing zmirror itself...')\n sleep(1)\n\n this_server = server_configs['apache']\n htdoc = this_server['htdoc']\n config_root = this_server['config_root']\n assert isinstance(htdoc, str)\n assert isinstance(config_root, str)\n os.chdir(htdoc)\n cmd('git clone %s zmirror' % __ZMIRROR_GIT_URL__, cwd=htdoc)\n zmirror_source_folder = os.path.join(htdoc, 'zmirror')\n\n # 预删除文件\n for pre_delete_file in this_server['pre_delete_files']:\n abs_path = pre_delete_file.format(\n config_root=config_root, htdoc=htdoc\n )\n print(\"deleting: \" + abs_path)\n try:\n os.remove(abs_path)\n except:\n logging.debug(\"Unable to remove file:\" + abs_path + \"\\n\" + traceback.format_exc())\n\n # 拷贝并设置各个镜像\n for mirror in mirrors_to_deploy:\n domain = mirrors_settings[mirror]['domain']\n this_mirror_folder = os.path.join(htdoc, mirror)\n\n # 如果文件夹已存在, 则报错\n if os.path.exists(this_mirror_folder):\n print(\n (\"Folder {folder} already exists.\"\n \"If you want to override, please delete that folder manually and run this script again\"\n ).format(folder=this_mirror_folder)\n )\n raise FileExistsError(\"Folder {folder} for mirror [{mirror_name}] already exists.\".format(\n folder=this_mirror_folder, mirror_name=mirror))\n\n # 将 zmirror 文件夹复制一份\n shutil.copytree(zmirror_source_folder, this_mirror_folder)\n # 更改文件夹所有者为 www-data (apache的用户)\n shutil.chown(this_mirror_folder, \"www-data\", \"www-data\")\n\n this_mirror = mirrors_settings[mirror]\n\n for file_from, file_to in this_mirror['cfg']:\n shutil.copy(os.path.join(this_mirror_folder, file_from),\n os.path.join(this_mirror_folder, file_to))\n\n with open(os.path.join(this_mirror_folder, 'config.py'), 'r+', encoding='utf-8') as fp:\n # noinspection PyRedeclaration\n content = fp.read()\n\n # 将 my_host_name 修改为对应的域名\n content = re.sub(r\"\"\"my_host_name *= *(['\"])[-.\\w]+\\1\"\"\",\n \"my_host_name = '{domain}' # Modified by zmirror-onekey\".format(domain=domain),\n content, count=1)\n # 将 my_host_scheme 修改为 https://\n content = re.sub(r\"\"\"my_host_scheme *= *(['\"])https?://\\1\"\"\",\n \"my_host_scheme = 'https://' # Modified by zmirror-onekey\",\n content, count=1)\n # 在文件末尾添加 verbose_level = 2\n content += '\\nverbose_level = 2 # Added by zmirror-onekey\\n'\n\n fp.seek(0) # 指针返回文件头\n fp.write(content) # 回写\n\n shutil.rmtree(zmirror_source_folder) # 删除无用的 zmirror 文件夹\n\n print(\"[zmirror] zmirror program folders deploy completed\")\n\n # ############# 配置Apache ###############\n print(\"[zmirror] Now installing apache configs...\")\n sleep(0.5)\n\n os.chdir(config_root)\n\n # 下载通用配置文件\n for conf_name in this_server['common_configs']:\n assert isinstance(config_root, str)\n url = this_server['configs'][conf_name]['url']\n file_path = os.path.join(config_root, this_server['configs'][conf_name]['file_path'])\n\n if os.path.exists(file_path): # 若配置文件已存在则跳过\n print(\"Config {path} already exists, skipping\".format(path=file_path))\n continue\n\n with open(file_path, 'w', encoding='utf-8') as fp:\n print(\"downloading: \", conf_name)\n fp.write(requests.get(url).text)\n\n # 下载并设置各个镜像的Apache配置文件\n for mirror in mirrors_to_deploy:\n domain = mirrors_settings[mirror]['domain']\n this_mirror_folder = os.path.join(htdoc, mirror)\n\n for conf_name in this_server['site_unique_configs']:\n url = this_server['configs'][conf_name]['url']\n file_path = os.path.join(config_root, this_server['configs'][conf_name]['file_path'])\n file_path = file_path.format(mirror_name=mirror, conf_name=conf_name)\n\n if os.path.exists(file_path): # 若配置文件已存在则跳过\n print(\"Config {path} already exists, skipping\".format(path=file_path))\n continue\n\n print(\"downloading: \", mirror, conf_name)\n\n conf = requests.get(url).text\n\n # 因为Apache conf里面有 {Ascii字符} 这种结构, 与python的string format冲突\n # 这边只能手动format\n for key, value in [\n ('domain', domain),\n ('mirror_name', mirror),\n ('path_to_wsgi_py', os.path.join(this_mirror_folder, 'wsgi.py')),\n ]:\n conf = conf.replace(\"{{%s}}\" % key, value)\n\n with open(file_path, 'w', encoding='utf-8') as fp:\n fp.write(conf)\n\n # ##### Add linux cron script for letsencrypt auto renewal ######\n if not os.path.exists(\"/etc/cron.weekly/zmirror-letsencrypt-renew.sh\"): # 若脚本已存在则跳过\n print(\"Adding cert auto renew script to `/etc/cron.weekly/zmirror-letsencrypt-renew.sh`\")\n cron_script = \"\"\"#!/bin/bash\n cd /etc/certbot\n ./certbot-auto renew -n --agree-tos --standalone --pre-hook \"/usr/sbin/service apache2 stop\" --post-hook \"/usr/sbin/service apache2 start\"\n exit 0\n \"\"\"\n with open(\"/etc/cron.weekly/zmirror-letsencrypt-renew.sh\", \"w\", encoding='utf-8') as fp:\n fp.write(cron_script)\n\n cmd('chmod +x /etc/cron.weekly/zmirror-letsencrypt-renew.sh')\n cmd('/etc/cron.weekly/zmirror-letsencrypt-renew.sh')\n\n # 重启一下apache\n print(\"Restarting apache2\")\n cmd('service apache2 restart')\n\nexcept:\n onekey_report('error', traceback_str=traceback.format_exc(), installing_mirror=mirrors_to_deploy)\n raise\n\nonekey_report('success', installing_mirror=mirrors_to_deploy)\n\n# ####### 完成 ########\nprint(\"Completed.\")\n# 最后打印一遍配置\nprint(\"------------ mirrors ------------\")\nfor mirror in mirrors_to_deploy:\n print(\"Mirror: {mirror} URL: https://{domain}/\".format(mirror=mirror, domain=mirrors_settings[mirror]['domain']))\n\nprint(\"\\nFor more information, please view zmirror's github: \", __ZMIRROR_PROJECT_URL__)\nprint(\"Contribution and Issues are more than welcomed.\")\nprint(\"btw, if you feeling good, I'll be grateful for your Star in github :)\")\n", "sub_path": "deploy.py", "file_name": "deploy.py", "file_ext": "py", "file_size_in_byte": 21716, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "sys.argv", "line_number": 25, "usage_type": "attribute"}, {"api_name": "subprocess.check_call", "line_number": 30, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 30, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 57, "usage_type": "call"}, {"api_name": "distro.info", "line_number": 57, "usage_type": "call"}, {"api_name": "re.search", "line_number": 70, "usage_type": "call"}, {"api_name": "traceback.print_exc", "line_number": 84, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 97, "usage_type": "call"}, {"api_name": "logging.NOTSET", "line_number": 98, "usage_type": "attribute"}, {"api_name": "sys.platform", "line_number": 102, "usage_type": "attribute"}, {"api_name": "os.geteuid", "line_number": 105, "usage_type": "call"}, {"api_name": "urllib.parse.urljoin", "line_number": 125, "usage_type": "call"}, {"api_name": "urllib.parse.urljoin", "line_number": 130, "usage_type": "call"}, {"api_name": "urllib.parse.urljoin", "line_number": 135, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 178, "usage_type": "call"}, {"api_name": "distro.id", "line_number": 201, "usage_type": "call"}, {"api_name": "distro.id", "line_number": 207, "usage_type": "call"}, {"api_name": "distro.id", "line_number": 214, "usage_type": "call"}, {"api_name": "distro.version", "line_number": 214, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 230, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 232, "usage_type": "call"}, {"api_name": "os.path", "line_number": 232, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 246, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 251, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 283, "usage_type": "call"}, {"api_name": "socket.gethostbyname", "line_number": 332, "usage_type": "call"}, {"api_name": "socket.gethostbyname", "line_number": 333, "usage_type": "call"}, {"api_name": "socket.gethostname", "line_number": 333, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 369, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 399, "usage_type": "call"}, {"api_name": "os.path", "line_number": 399, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 414, "usage_type": "call"}, {"api_name": "os.path", "line_number": 414, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 425, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 432, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 434, "usage_type": "call"}, {"api_name": "os.path", "line_number": 434, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 443, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 445, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 445, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 450, "usage_type": "call"}, {"api_name": "os.path", "line_number": 450, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 453, "usage_type": "call"}, {"api_name": "os.path", "line_number": 453, "usage_type": "attribute"}, {"api_name": "shutil.copytree", "line_number": 463, "usage_type": "call"}, {"api_name": "shutil.chown", "line_number": 465, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 470, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 470, "usage_type": "call"}, {"api_name": "os.path", "line_number": 470, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 471, "usage_type": "call"}, {"api_name": "os.path", "line_number": 471, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 473, "usage_type": "call"}, {"api_name": "os.path", "line_number": 473, "usage_type": "attribute"}, {"api_name": "re.sub", "line_number": 478, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 482, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 491, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 497, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 499, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 505, "usage_type": "call"}, {"api_name": "os.path", "line_number": 505, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 507, "usage_type": "call"}, {"api_name": "os.path", "line_number": 507, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 513, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 518, "usage_type": "call"}, {"api_name": "os.path", "line_number": 518, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 522, "usage_type": "call"}, {"api_name": "os.path", "line_number": 522, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 525, "usage_type": "call"}, {"api_name": "os.path", "line_number": 525, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 531, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 538, "usage_type": "call"}, {"api_name": "os.path", "line_number": 538, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 546, "usage_type": "call"}, {"api_name": "os.path", "line_number": 546, "usage_type": "attribute"}, {"api_name": "traceback.format_exc", "line_number": 564, "usage_type": "call"}]} +{"seq_id": "496885302", "text": "#\n# License: BSD\n# https://raw.github.com/robotics-in-concert/rocon_concert/license/LICENSE\n#\n##############################################################################\n# Imports\n##############################################################################\n\nimport yaml\nimport zlib # crc32\n\nimport genpy\nimport rospkg\nimport rocon_console.console as console\nimport rocon_python_utils\nimport rocon_interaction_msgs.msg as interaction_msgs\n\nfrom .exceptions import InvalidInteraction, MalformedInteractionsYaml, YamlResourceNotFoundException\n\n##############################################################################\n# Utility Methods\n##############################################################################\n\n\ndef generate_hash(name, role, namespace):\n '''\n Compute a unique hash for this interaction corresponding to the\n name-role-namespace triple. We use zlib's crc32 here instead of unique_id because\n of it's brevity which is important when trying to id a remocon app by its hash\n from an nfc tag.\n\n Might be worth checking here http://docs.python.org/2.7/library/zlib.html#zlib.crc32 if\n his doesn't produce the same hash on all platforms.\n\n :param name: the executable name of the interaction\n :type name: str\n :param role: the role the interaction is embedded in\n :type role: str\n :param namespace: the namespace in which to embed this interaction\n :type namespace: str\n '''\n return zlib.crc32(name + \"-\" + role + \"-\" + namespace)\n\n\ndef load_msgs_from_yaml_resource(resource_name):\n \"\"\"\n Load interactions from a yaml resource.\n\n :param resource_name: pkg/filename of a yaml formatted interactions file (ext=.interactions).\n :type resource_name: str\n\n :returns: a list of ros msg interaction specifications\n :rtype: concert_msgs.Interaction[]\n\n :raises: :exc:`rocon_interactions.YamlResourceNotFoundException,` if yaml is not found.\n :raises: :exc:`rocon_interactions.MalformedInteractionsYaml,` if yaml is malformed.\n \"\"\"\n interactions = []\n try:\n yaml_filename = rocon_python_utils.ros.find_resource_from_string(resource_name, extension='interactions')\n except rospkg.ResourceNotFound as e: # resource not found.\n raise YamlResourceNotFoundException(str(e))\n with open(yaml_filename) as f:\n # load the interactions from yaml into a python object\n interaction_yaml_objects = yaml.load(f)\n # now drop it into message format\n for interaction_yaml_object in interaction_yaml_objects:\n # convert the parameters from a freeform yaml variable to a yaml string suitable for\n # shipping off in ros msgs (where parameters is a string variable)\n if 'parameters' in interaction_yaml_object: # it's an optional key\n # chomp trailing newlines\n interaction_yaml_object['parameters'] = yaml.dump(interaction_yaml_object['parameters']).rstrip()\n interaction = interaction_msgs.Interaction()\n try:\n genpy.message.fill_message_args(interaction, interaction_yaml_object)\n except genpy.MessageException as e:\n raise MalformedInteractionsYaml(\n \"malformed yaml preventing converting of yaml to interaction msg type [%s]\" % str(e))\n interactions.append(interaction)\n return interactions\n\n##############################################################################\n# Classes\n##############################################################################\n\n\nclass Interaction(object):\n '''\n Defines an interaction. This wraps the base ros msg data structure with\n a few convenient management handles.\n '''\n __slots__ = [\n 'msg', # rocon_interaction_msgs.Interaction\n # aliases\n 'name',\n 'compatibility',\n 'namespace',\n 'display_name',\n 'role',\n 'hash',\n 'max',\n ]\n\n def __init__(self, msg):\n '''\n Validate the incoming fields and populate remaining fields with sane defaults.\n Also compute a unique hash for this object based on the incoming\n name-role-namespace triple.\n\n @param msg\n @type rocon_interaction_msgs.Interaction\n\n @raise exceptions.InvalidInteraction\n '''\n self.msg = msg\n if self.msg.max < -1:\n raise InvalidInteraction(\"maximum instance configuration cannot be negative [%s]\" % self.msg.display_name)\n if self.msg.max == 0:\n self.msg.max = 1\n if self.msg.role == '':\n raise InvalidInteraction(\"role not configured [%s]\" % self.msg.display_name)\n if self.msg.icon.resource_name == \"\":\n self.msg.icon.resource_name = 'rocon_bubble_icons/rocon.png'\n if not self.msg.icon.data:\n self.msg.icon = rocon_python_utils.ros.icon_resource_to_msg(self.msg.icon.resource_name)\n if self.msg.namespace == '':\n self.msg.namespace = '/'\n self.msg.hash = generate_hash(self.msg.name, self.msg.role, self.msg.namespace)\n # some convenient aliases\n self.role = self.msg.role\n self.name = self.msg.name\n self.namespace = self.msg.namespace\n self.display_name = self.msg.display_name\n self.hash = self.msg.hash\n self.compatibility = self.msg.compatibility\n self.max = self.msg.max\n\n def _eq__(self, other):\n if type(other) is type(self):\n return self.msg.hash == other.msg.hash\n else:\n return False\n\n def __ne__(self, other):\n return not self.__eq__(other)\n\n def __str__(self):\n '''\n Format the interaction into a human-readable string.\n '''\n s = ''\n s += console.green + \"%s\" % self.msg.display_name + console.reset + '\\n'\n s += console.cyan + \" Name\" + console.reset + \" : \" + console.yellow + \"%s\" % self.msg.name + console.reset + '\\n' # noqa\n s += console.cyan + \" Description\" + console.reset + \" : \" + console.yellow + \"%s\" % self.msg.description + console.reset + '\\n' # noqa\n s += console.cyan + \" Icon\" + console.reset + \" : \" + console.yellow + \"%s\" % str(self.msg.icon.resource_name) + console.reset + '\\n' # noqa\n s += console.cyan + \" Rocon URI\" + console.reset + \" : \" + console.yellow + \"%s\" % self.msg.compatibility + console.reset + '\\n' # noqa\n s += console.cyan + \" Namespace\" + console.reset + \" : \" + console.yellow + \"%s\" % self.msg.namespace + console.reset + '\\n' # noqa\n if self.msg.max == -1:\n s += console.cyan + \" Max\" + console.reset + \" : \" + console.yellow + \"infinity\" + console.reset + '\\n' # noqa\n else:\n s += console.cyan + \" Max\" + console.reset + \" : \" + console.yellow + \"%s\" % self.msg.max + console.reset + '\\n' # noqa\n already_prefixed = False\n for remapping in self.msg.remappings:\n if not already_prefixed:\n s += console.cyan + \" Remapping\" + console.reset + \" : \" + console.yellow + \"%s->%s\" % (remapping.remap_from, remapping.remap_to) + console.reset + '\\n' # noqa\n already_prefixed = True\n else:\n s += \" : \" + console.yellow + \"%s->%s\" % (remapping.remap_from, remapping.remap_to) + console.reset + '\\n' # noqa\n if self.msg.parameters != '':\n s += console.cyan + \" Parameters\" + console.reset + \" : \" + console.yellow + \"%s\" % self.msg.parameters + console.reset + '\\n' # noqa\n s += console.cyan + \" Hash\" + console.reset + \" : \" + console.yellow + \"%s\" % str(self.msg.hash) + console.reset + '\\n' # noqa\n return s\n", "sub_path": "rocon_interactions/src/rocon_interactions/interactions.py", "file_name": "interactions.py", "file_ext": "py", "file_size_in_byte": 7792, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "zlib.crc32", "line_number": 42, "usage_type": "call"}, {"api_name": "rocon_python_utils.ros.find_resource_from_string", "line_number": 60, "usage_type": "call"}, {"api_name": "rocon_python_utils.ros", "line_number": 60, "usage_type": "attribute"}, {"api_name": "rospkg.ResourceNotFound", "line_number": 61, "usage_type": "attribute"}, {"api_name": "exceptions.YamlResourceNotFoundException", "line_number": 62, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 65, "usage_type": "call"}, {"api_name": "yaml.dump", "line_number": 72, "usage_type": "call"}, {"api_name": "rocon_interaction_msgs.msg.Interaction", "line_number": 73, "usage_type": "call"}, {"api_name": "rocon_interaction_msgs.msg", "line_number": 73, "usage_type": "name"}, {"api_name": "genpy.message.fill_message_args", "line_number": 75, "usage_type": "call"}, {"api_name": "genpy.message", "line_number": 75, "usage_type": "attribute"}, {"api_name": "genpy.MessageException", "line_number": 76, "usage_type": "attribute"}, {"api_name": "exceptions.MalformedInteractionsYaml", "line_number": 77, "usage_type": "call"}, {"api_name": "exceptions.InvalidInteraction", "line_number": 117, "usage_type": "call"}, {"api_name": "exceptions.InvalidInteraction", "line_number": 121, "usage_type": "call"}, {"api_name": "rocon_python_utils.ros.icon_resource_to_msg", "line_number": 125, "usage_type": "call"}, {"api_name": "rocon_python_utils.ros", "line_number": 125, "usage_type": "attribute"}, {"api_name": "rocon_console.console.green", "line_number": 152, "usage_type": "attribute"}, {"api_name": "rocon_console.console", "line_number": 152, "usage_type": "name"}, {"api_name": "rocon_console.console.reset", "line_number": 152, "usage_type": "attribute"}, {"api_name": "rocon_console.console.cyan", "line_number": 153, "usage_type": "attribute"}, {"api_name": "rocon_console.console", "line_number": 153, "usage_type": "name"}, {"api_name": "rocon_console.console.reset", "line_number": 153, "usage_type": "attribute"}, {"api_name": "rocon_console.console.yellow", "line_number": 153, "usage_type": "attribute"}, {"api_name": "rocon_console.console.cyan", "line_number": 154, "usage_type": "attribute"}, {"api_name": "rocon_console.console", "line_number": 154, "usage_type": "name"}, {"api_name": "rocon_console.console.reset", "line_number": 154, "usage_type": "attribute"}, {"api_name": "rocon_console.console.yellow", "line_number": 154, "usage_type": "attribute"}, {"api_name": "rocon_console.console.cyan", "line_number": 155, "usage_type": "attribute"}, {"api_name": "rocon_console.console", "line_number": 155, "usage_type": "name"}, {"api_name": "rocon_console.console.reset", "line_number": 155, "usage_type": "attribute"}, {"api_name": "rocon_console.console.yellow", "line_number": 155, "usage_type": "attribute"}, {"api_name": "rocon_console.console.cyan", "line_number": 156, "usage_type": "attribute"}, {"api_name": "rocon_console.console", "line_number": 156, "usage_type": "name"}, {"api_name": "rocon_console.console.reset", "line_number": 156, "usage_type": "attribute"}, {"api_name": "rocon_console.console.yellow", "line_number": 156, "usage_type": "attribute"}, {"api_name": "rocon_console.console.cyan", "line_number": 157, "usage_type": "attribute"}, {"api_name": "rocon_console.console", "line_number": 157, "usage_type": "name"}, {"api_name": "rocon_console.console.reset", "line_number": 157, "usage_type": "attribute"}, {"api_name": "rocon_console.console.yellow", "line_number": 157, "usage_type": "attribute"}, {"api_name": "rocon_console.console.cyan", "line_number": 159, "usage_type": "attribute"}, {"api_name": "rocon_console.console", "line_number": 159, "usage_type": "name"}, {"api_name": "rocon_console.console.reset", "line_number": 159, "usage_type": "attribute"}, {"api_name": "rocon_console.console.yellow", "line_number": 159, "usage_type": "attribute"}, {"api_name": "rocon_console.console.cyan", "line_number": 161, "usage_type": "attribute"}, {"api_name": "rocon_console.console", "line_number": 161, "usage_type": "name"}, {"api_name": "rocon_console.console.reset", "line_number": 161, "usage_type": "attribute"}, {"api_name": "rocon_console.console.yellow", "line_number": 161, "usage_type": "attribute"}, {"api_name": "rocon_console.console.cyan", "line_number": 165, "usage_type": "attribute"}, {"api_name": "rocon_console.console", "line_number": 165, "usage_type": "name"}, {"api_name": "rocon_console.console.reset", "line_number": 165, "usage_type": "attribute"}, {"api_name": "rocon_console.console.yellow", "line_number": 165, "usage_type": "attribute"}, {"api_name": "rocon_console.console.yellow", "line_number": 168, "usage_type": "attribute"}, {"api_name": "rocon_console.console", "line_number": 168, "usage_type": "name"}, {"api_name": "rocon_console.console.reset", "line_number": 168, "usage_type": "attribute"}, {"api_name": "rocon_console.console.cyan", "line_number": 170, "usage_type": "attribute"}, {"api_name": "rocon_console.console", "line_number": 170, "usage_type": "name"}, {"api_name": "rocon_console.console.reset", "line_number": 170, "usage_type": "attribute"}, {"api_name": "rocon_console.console.yellow", "line_number": 170, "usage_type": "attribute"}, {"api_name": "rocon_console.console.cyan", "line_number": 171, "usage_type": "attribute"}, {"api_name": "rocon_console.console", "line_number": 171, "usage_type": "name"}, {"api_name": "rocon_console.console.reset", "line_number": 171, "usage_type": "attribute"}, {"api_name": "rocon_console.console.yellow", "line_number": 171, "usage_type": "attribute"}]} +{"seq_id": "575721209", "text": "\"\"\"\n 14강. 이미지 Gradient를 이용한 경계 찾기\n\n 이미지 gradient(이미지 경사도 혹은 수학적인 용어로 이미지의 변화율)를 이용한 에지(경계선)를 찾는 방법에 대해 알아 보겠습니다.\n OpenCV는 Sobel, Scharr, Laplacian 이 세가지 타입의 Gradient 필터(High-pass filters;HPF)를 제공합니다.\n Sobel, Scharr 미분 (Sobel and Sharr Derivatives)\n Sobel 오퍼레이터는 가우스 스무딩(Gaussian Smoothing)과 미분연산을 결합한 형태의 연산을 수행함으로써 노이즈에 보다 강력한 저항성을 제공합니다.\n Sobel 오퍼레이션은 세로 방향 또는 가로 방향으로 연산 수행이 가능합니다. cv2.Sobel() 함수는 이미지에 sobel 연산을 수행하는 함수입니다.\n\n cv2.Sobel(src, ddepth, dx, dy, ksize)\n src: Sobel 미분을 적용할 원본 이미지\n ddepth: 결과 이미지 데이터 타입\n CV_8U: 이미지 픽셀값을 uint8 로 설정\n CV_16U: 이미지 픽셀값을 uint16으로 설정\n CV_32F: 이미지 픽셀값을 float32로 설정\n CV_64F: 이미지 픽셀값ㅇ르 float64로 설정\n dx, dy: 각각 x방향, y방향으로 미분 차수 (eg. 1, 0 이면, x 방향으로 1차 미분 수행, y 방향으로 그대로 두라는 의미)\n ksize: 확장 Sobel 커널의 크기. 1, 3, 5, 7 중 하나의 값으로 설정. -1로 설정되면 3x3 Soble 필터 대신 3x3 Scharr 필터를 적용하게 됨\n\n\"\"\"\n\nimport numpy as np\nimport cv2 as cv2\nimport matplotlib.pyplot as plt\nimport default_import as selImg\n#from default_import import select_img as selImg\n#img = cv2.imread(selImg.select_img(6), cv2.IMREAD_GRAYSCALE)\n\n\ndef grad():\n img = cv2.imread(selImg.select_img(6), cv2.IMREAD_GRAYSCALE)\n #img = cv2.imread(select_img('2'), cv2.IMREAD_GRAYSCALE)\n\n laplacian = cv2.Laplacian(img, cv2.CV_64F)\n sobelx = cv2.Sobel(img, cv2.CV_64F, 1, 0, ksize = 3)\n sobely = cv2.Sobel(img, cv2.CV_64F, 0, 1, ksize = 3)\n\n plt.subplot(2, 2, 1), plt.imshow(img, cmap = 'gray')\n plt.title('orignal'), plt.xticks([]), plt.yticks([])\n\n plt.subplot(2, 2, 2), plt.imshow(laplacian, cmap = 'gray')\n plt.title('Laplacian'), plt.xticks([]), plt.yticks([])\n\n plt.subplot(2, 2, 3), plt.imshow(sobelx, cmap = 'gray')\n plt.title('sobel X'), plt.xticks([]), plt.yticks([])\n\n plt.subplot(2, 2, 4), plt.imshow(sobely, cmap = 'gray')\n plt.title('sobel Y'), plt.xticks([]), plt.yticks([])\n\n plt.show()\n\ngrad()\n\n\n\n", "sub_path": "OpenCV/gradient.py", "file_name": "gradient.py", "file_ext": "py", "file_size_in_byte": 2656, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "cv2.imread", "line_number": 31, "usage_type": "call"}, {"api_name": "default_import.select_img", "line_number": 31, "usage_type": "call"}, {"api_name": "cv2.IMREAD_GRAYSCALE", "line_number": 31, "usage_type": "attribute"}, {"api_name": "cv2.Laplacian", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.CV_64F", "line_number": 34, "usage_type": "attribute"}, {"api_name": "cv2.Sobel", "line_number": 35, "usage_type": "call"}, {"api_name": "cv2.CV_64F", "line_number": 35, "usage_type": "attribute"}, {"api_name": "cv2.Sobel", "line_number": 36, "usage_type": "call"}, {"api_name": "cv2.CV_64F", "line_number": 36, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}]} +{"seq_id": "266825334", "text": "# Copyright 2013 - Mirantis, Inc.\n# Copyright 2016 - Brocade Communications Systems, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom oslo_log import log as logging\nfrom osprofiler import profiler\n\nfrom mistral.actions import action_factory as a_f\nfrom mistral.engine import base\nfrom mistral.engine.rpc_backend import rpc\nfrom mistral import exceptions as exc\nfrom mistral.utils import inspect_utils as i_u\nfrom mistral.workflow import utils as wf_utils\n\n\nLOG = logging.getLogger(__name__)\n\n\nclass DefaultExecutor(base.Executor):\n def __init__(self):\n self._engine_client = rpc.get_engine_client()\n\n @profiler.trace('executor-run-action', hide_args=True)\n def run_action(self, action_ex_id, action_class_str, attributes,\n action_params, safe_rerun, redelivered=False):\n \"\"\"Runs action.\n\n :param action_ex_id: Action execution id.\n :param action_class_str: Path to action class in dot notation.\n :param attributes: Attributes of action class which will be set to.\n :param action_params: Action parameters.\n :param safe_rerun: Tells if given action can be safely rerun.\n :param redelivered: Tells if given action was run before on another\n executor.\n \"\"\"\n\n def send_error_back(error_msg):\n error_result = wf_utils.Result(error=error_msg)\n\n if action_ex_id:\n self._engine_client.on_action_complete(\n action_ex_id,\n error_result\n )\n\n return None\n\n return error_result\n\n if redelivered and not safe_rerun:\n msg = (\n \"Request to run action %s was redelivered, but action %s\"\n \" cannot be re-run safely. The only safe thing to do is fail\"\n \" action.\"\n % (action_class_str, action_class_str)\n )\n\n return send_error_back(msg)\n\n action_cls = a_f.construct_action_class(action_class_str, attributes)\n\n # Instantiate action.\n\n try:\n action = action_cls(**action_params)\n except Exception as e:\n msg = (\"Failed to initialize action %s. Action init params = %s.\"\n \" Actual init params = %s. More info: %s\"\n % (action_class_str, i_u.get_arg_list(action_cls.__init__),\n action_params.keys(), e))\n LOG.warning(msg)\n\n return send_error_back(msg)\n\n # Run action.\n\n try:\n result = action.run()\n\n # Note: it's made for backwards compatibility with already\n # existing Mistral actions which don't return result as\n # instance of workflow.utils.Result.\n if not isinstance(result, wf_utils.Result):\n result = wf_utils.Result(data=result)\n\n except Exception as e:\n msg = (\"Failed to run action [action_ex_id=%s, action_cls='%s',\"\n \" attributes='%s', params='%s']\\n %s\"\n % (action_ex_id, action_cls, attributes, action_params, e))\n LOG.exception(msg)\n\n return send_error_back(msg)\n\n # Send action result.\n\n try:\n if action_ex_id and (action.is_sync() or result.is_error()):\n self._engine_client.on_action_complete(\n action_ex_id,\n result,\n async_=True\n )\n\n except exc.MistralException as e:\n # In case of a Mistral exception we can try to send error info to\n # engine because most likely it's not related to the infrastructure\n # such as message bus or network. One known case is when the action\n # returns a bad result (e.g. invalid unicode) which can't be\n # serialized.\n msg = (\"Failed to call engine's on_action_complete() method due\"\n \" to a Mistral exception\"\n \" [action_ex_id=%s, action_cls='%s',\"\n \" attributes='%s', params='%s']\\n %s\"\n % (action_ex_id, action_cls, attributes, action_params, e))\n LOG.exception(msg)\n\n return send_error_back(msg)\n except Exception as e:\n # If it's not a Mistral exception all we can do is only\n # log the error.\n msg = (\"Failed to call engine's on_action_complete() method due\"\n \" to an unexpected exception\"\n \" [action_ex_id=%s, action_cls='%s',\"\n \" attributes='%s', params='%s']\\n %s\"\n % (action_ex_id, action_cls, attributes, action_params, e))\n LOG.exception(msg)\n\n return result\n", "sub_path": "mistral-4.0.2/mistral/engine/default_executor.py", "file_name": "default_executor.py", "file_ext": "py", "file_size_in_byte": 5258, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "oslo_log.log.getLogger", "line_number": 27, "usage_type": "call"}, {"api_name": "oslo_log.log", "line_number": 27, "usage_type": "name"}, {"api_name": "mistral.engine.base.Executor", "line_number": 30, "usage_type": "attribute"}, {"api_name": "mistral.engine.base", "line_number": 30, "usage_type": "name"}, {"api_name": "mistral.engine.rpc_backend.rpc.get_engine_client", "line_number": 32, "usage_type": "call"}, {"api_name": "mistral.engine.rpc_backend.rpc", "line_number": 32, "usage_type": "name"}, {"api_name": "mistral.workflow.utils.Result", "line_number": 49, "usage_type": "call"}, {"api_name": "mistral.workflow.utils", "line_number": 49, "usage_type": "name"}, {"api_name": "mistral.actions.action_factory.construct_action_class", "line_number": 71, "usage_type": "call"}, {"api_name": "mistral.actions.action_factory", "line_number": 71, "usage_type": "name"}, {"api_name": "mistral.utils.inspect_utils.get_arg_list", "line_number": 80, "usage_type": "call"}, {"api_name": "mistral.utils.inspect_utils", "line_number": 80, "usage_type": "name"}, {"api_name": "mistral.workflow.utils.Result", "line_number": 94, "usage_type": "attribute"}, {"api_name": "mistral.workflow.utils", "line_number": 94, "usage_type": "name"}, {"api_name": "mistral.workflow.utils.Result", "line_number": 95, "usage_type": "call"}, {"api_name": "mistral.workflow.utils", "line_number": 95, "usage_type": "name"}, {"api_name": "mistral.exceptions.MistralException", "line_number": 115, "usage_type": "attribute"}, {"api_name": "mistral.exceptions", "line_number": 115, "usage_type": "name"}, {"api_name": "osprofiler.profiler.trace", "line_number": 34, "usage_type": "call"}, {"api_name": "osprofiler.profiler", "line_number": 34, "usage_type": "name"}]} +{"seq_id": "213680437", "text": "# Q14: write a Python program to find the factorial of a number provided by the user\nfrom functools import reduce\n\n\ndef input_positive_integer(msg):\n try:\n num = int(input(msg).strip())\n except ValueError:\n print(\"Please input positive integers only!\")\n num = int(input(msg))\n if num < 0:\n print(\"Please input positive integers only!\")\n num = int(input(msg))\n return num\n\n\ndef factorial():\n num = input_positive_integer(\"Please enter a positive integer: \")\n print(\"%d! = \"% num, int(reduce(lambda x, y: x * y, range(1, num + 1))))\n\n\nfactorial()\n", "sub_path": "Python/Assignment4/question14.py", "file_name": "question14.py", "file_ext": "py", "file_size_in_byte": 598, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "functools.reduce", "line_number": 19, "usage_type": "call"}]} +{"seq_id": "56624966", "text": "import os\nfrom itertools import product\nimport numpy as np\nimport math\nfrom pymatgen.io.cif import CifWriter\nfrom ase.build import general_surface\nfrom ase.spacegroup import crystal\nfrom ase.visualize import view\nfrom ase.lattice.surface import *\nfrom ase.io import *\nimport pymatgen as mg\nfrom pymatgen.io.vasp.inputs import Poscar\nimport argparse\nimport pymatgen as mg\nfrom pymatgen.core.structure import Structure\nfrom pymatgen.io.vasp.inputs import Poscar\nfrom pymatgen.analysis.structure_matcher import StructureMatcher\nfrom pymatgen.core.surface import Slab, SlabGenerator, ReconstructionGenerator\nfrom pymatgen.analysis.substrate_analyzer import SubstrateAnalyzer, ZSLGenerator\nfrom core.utils.utils import *\n\ndef get_equiv_transformations_Sam(self, transformation_sets, film_vectors,\n substrate_vectors):\n # Monkey-patching the original function of pymatgen to generate the transformation matrices\n\n text_file = open(self.working_dir + \"film_sub_sets\", \"w\")\n film_sub_sets = []\n for (film_transformations, substrate_transformations) in \\\n transformation_sets:\n\n # Apply transformations and reduce using Zur reduce methodology\n films = [reduce_vectors(*np.dot(f, film_vectors))\n for f in film_transformations]\n Sam_films = []\n for i in films:\n Sam_films.append(mat_clean(i))\n\n substrates = [reduce_vectors(*np.dot(s, substrate_vectors))\n for s in substrate_transformations]\n\n sam_substrates = []\n for i in substrates:\n sam_substrates.append(mat_clean(i))\n # Check if equivelant super lattices\n\n for f, s in product(films, substrates):\n if self.is_same_vectors(f, s):\n f_index = Sam_films.index(mat_clean(f))\n s_index = sam_substrates.index(mat_clean(s))\n\n print([film_transformations[f_index].tolist(), substrate_transformations[s_index].tolist()],\n file=text_file)\n\n film_sub_sets.append(\n [film_transformations[f_index].tolist(), substrate_transformations[s_index].tolist()])\n yield [f, s]\n\n text_file.close()\n\n\ndef Interface_generator(Ini_sub_slab, Ini_film_slab, sub_tr_mat, film_tr_mat, distance, fparam):\n\n raw_ini_sub_slab_mat = np.array(Ini_sub_slab.lattice.matrix)\n raw_ini_film_slab_mat = np.array(Ini_film_slab.lattice.matrix)\n sub_reduction = reduce_vectors(\n raw_ini_sub_slab_mat[0], raw_ini_sub_slab_mat[1])\n film_reduction = reduce_vectors(\n raw_ini_film_slab_mat[0], raw_ini_film_slab_mat[1])\n reduced_sub_mat = np.array(\n [sub_reduction[0], sub_reduction[1], raw_ini_sub_slab_mat[2]])\n reduced_film_mat = np.array(\n [film_reduction[0], film_reduction[1], raw_ini_film_slab_mat[2]])\n red_Ini_sub_slab = Structure(mg.Lattice(reduced_sub_mat), Ini_sub_slab.species, Ini_sub_slab.cart_coords,\n coords_are_cartesian=True)\n red_Ini_film_slab = Structure(mg.Lattice(reduced_film_mat), Ini_film_slab.species, Ini_film_slab.cart_coords,\n coords_are_cartesian=True)\n red_Ini_sub_slab.make_supercell(scaling_matrix=scale_mat(sub_tr_mat))\n red_Ini_film_slab.make_supercell(scaling_matrix=scale_mat(film_tr_mat))\n Ini_sub_mat = red_Ini_sub_slab.lattice.matrix\n Ini_film_mat = red_Ini_film_slab.lattice.matrix\n sub_r_vecs = reduce_vectors(Ini_sub_mat[0], Ini_sub_mat[1])\n film_r_vecs = reduce_vectors(Ini_film_mat[0], Ini_film_mat[1])\n sub_mat = np.array([sub_r_vecs[0], sub_r_vecs[1], Ini_sub_mat[2]])\n film_mat = np.array([film_r_vecs[0], film_r_vecs[1], Ini_film_mat[2]])\n modif_sub_struc = mg.Structure(mg.Lattice(sub_mat), red_Ini_sub_slab.species, red_Ini_sub_slab.cart_coords,\n coords_are_cartesian=True)\n modif_film_struc = mg.Structure(mg.Lattice(film_mat), red_Ini_film_slab.species, red_Ini_film_slab.cart_coords,\n coords_are_cartesian=True)\n sub_sl_vecs = [modif_sub_struc.lattice.matrix[0],\n modif_sub_struc.lattice.matrix[1]]\n film_sl_vecs = [modif_film_struc.lattice.matrix[0],\n modif_film_struc.lattice.matrix[1]]\n film_angel = angle(film_sl_vecs[0], film_sl_vecs[1])\n sub_angel = angle(sub_sl_vecs[0], sub_sl_vecs[1])\n u_size = fparam * \\\n (np.linalg.norm(sub_sl_vecs[0])) + (1 -\n fparam) * (np.linalg.norm(film_sl_vecs[0]))\n v_size = fparam * \\\n (np.linalg.norm(sub_sl_vecs[1])) + (1 -\n fparam) * (np.linalg.norm(film_sl_vecs[1]))\n mean_angle = fparam * sub_angel + (1 - fparam) * film_angel\n sub_rot_mat = [[u_size, 0, 0], [v_size * math.cos(mean_angle), v_size * math.sin(mean_angle), 0],\n [0, 0, np.linalg.norm(modif_sub_struc.lattice.matrix[2])]]\n film_rot_mat = [[u_size, 0, 0], [v_size * math.cos(mean_angle), v_size * math.sin(mean_angle), 0],\n [0, 0, -np.linalg.norm(modif_film_struc.lattice.matrix[2])]]\n film_normal = np.cross(film_sl_vecs[0], film_sl_vecs[1])\n sub_normal = np.cross(sub_sl_vecs[0], sub_sl_vecs[1])\n film_un = film_normal / np.linalg.norm(film_normal)\n sub_un = sub_normal / np.linalg.norm(sub_normal)\n film_sl_vecs.append(film_un)\n L1_mat = np.transpose(film_sl_vecs)\n L1_res = [[u_size, v_size * math.cos(mean_angle), 0],\n [0, v_size * math.sin(mean_angle), 0], [0, 0, 1]]\n L1_mat_inv = np.linalg.inv(L1_mat)\n L1 = np.matmul(L1_res, L1_mat_inv)\n sub_sl_vecs.append(sub_un)\n L2_mat = np.transpose(sub_sl_vecs)\n L2_res = [[u_size, v_size * math.cos(mean_angle), 0],\n [0, v_size * math.sin(mean_angle), 0], [0, 0, -1]]\n L2_mat_inv = np.linalg.inv(L2_mat)\n L2 = np.matmul(L2_res, L2_mat_inv)\n sub_rot_lattice = mg.Lattice(sub_rot_mat)\n film_rot_lattice = mg.Lattice(film_rot_mat)\n r_sub_coords = np.array(modif_sub_struc.cart_coords)\n r_film_coords = np.array(modif_film_struc.cart_coords)\n\n for ii in range(len(r_sub_coords)):\n r_sub_coords[ii] = np.matmul(L2, r_sub_coords[ii])\n for ii in range(len(r_film_coords)):\n r_film_coords[ii] = np.matmul(L1, r_film_coords[ii])\n\n sub_slab = mg.Structure(\n sub_rot_lattice, modif_sub_struc.species, r_sub_coords, coords_are_cartesian=True)\n film_slab = mg.Structure(\n film_rot_lattice, modif_film_struc.species, r_film_coords, coords_are_cartesian=True)\n # SP_text = open(working_dir+ \"SP_typ\" , \"w\")\n sub_sp_num = len(sub_slab.types_of_specie)\n film_sp_num = len(film_slab.types_of_specie)\n # SP_text.close()\n sub_slab_mat = np.array(sub_slab.lattice.matrix)\n film_slab_mat = np.array(film_slab.lattice.matrix)\n sub_slab_coords = sub_slab.cart_coords\n film_slab_coords = film_slab.cart_coords\n sub_slab_zmat = sub_slab_coords[:, [2]]\n film_slab_zmat = film_slab_coords[:, [2]]\n sub_slab_zmat = sub_slab_zmat - min(sub_slab_zmat)\n film_slab_zmat = film_slab_zmat - min(film_slab_zmat)\n sub_max_z = max(sub_slab_zmat)\n sub_min_z = min(sub_slab_zmat)\n modif_film_slab_zmat = film_slab_zmat + sub_max_z - sub_min_z + distance\n film_slab_coords[:, [2]] = modif_film_slab_zmat\n sub_slab_coords[:, [2]] = sub_slab_zmat\n\n sub_max_z = max(sub_slab_zmat)\n film_min_z = min(modif_film_slab_zmat)\n sub_max_list = coords_sperator(sub_slab_zmat, sub_sp_num, True)\n film_min_list = coords_sperator(modif_film_slab_zmat, film_sp_num, False)\n\n interface_coords = np.concatenate(\n (sub_slab_coords, film_slab_coords), axis=0)\n interface_species = sub_slab.species + film_slab.species\n interface_latt = sub_slab_mat\n interface_latt[2][2] = abs(sub_slab_mat[2][2]) + \\\n abs(film_slab_mat[2][2]) + distance\n\n Adding_val = 0.5 * (interface_latt[2][2] - max(interface_coords[:, [2]]))\n sub_max_list += Adding_val\n film_min_list += Adding_val\n sub_max_z += Adding_val\n film_min_z += Adding_val\n\n interface_coords[:, [2]] += 0.5 * \\\n (interface_latt[2][2] - max(interface_coords[:, [2]]))\n sub_slab_coords[:, [2]] += Adding_val\n film_slab_coords[:, [2]] += Adding_val\n # sub_slab_coords[:, [2]] += 0.5 * \\\n # (interface_latt[2][2] - max(sub_slab_coords[:, [2]]))\n # film_slab_coords[:, [2]] += 0.5 * \\\n # (interface_latt[2][2] - max(film_slab_coords[:, [2]]))\n interface_lattice = mg.Lattice(interface_latt)\n interface_struc = mg.Structure(\n interface_lattice, interface_species, interface_coords, coords_are_cartesian=True)\n interface_struc = interface_struc.get_reduced_structure()\n # Poscar(interface_struc.get_reduced_structure()).write_file(working_dir + \"POSCAR_interface\", direct=False )\n\n ###################\n # Seperate the first two layers\n\n surf_int_species = []\n surf_int_coords = []\n\n for k in range(len(interface_coords)):\n for k1 in sub_max_list:\n if (round(interface_coords[k, 2], 8) == round(k1[0], 8)):\n surf_int_coords.append(interface_coords[k, :])\n surf_int_species.append(interface_species[k])\n for k2 in film_min_list:\n if (round(interface_coords[k, 2], 8) == round(k2[0], 8)):\n surf_int_coords.append(interface_coords[k, :])\n surf_int_species.append(interface_species[k])\n\n surf_int_coords = np.array(surf_int_coords)\n surf_int_coords[:, 2] -= min(surf_int_coords[:, 2])\n surf_int_lat = np.array([interface_latt[0], interface_latt[1], [\n 0, 0, 15*max(surf_int_coords[:, 2])]])\n surf_int_coords[:, 2] += 7 * max(surf_int_coords[:, 2])\n surf_struc = Structure(surf_int_lat, surf_int_species,\n surf_int_coords, coords_are_cartesian=True)\n # surf_cif = CifWriter(surf_struc)\n # surf_cif.write_file(working_dir+\"surf.cif\")\n surf_struc = surf_struc.get_reduced_structure()\n # Poscar(surf_struc.get_reduced_structure()).write_file(working_dir + \"POSCAR_Surf_int\", direct = False)\n\n return [interface_struc, surf_struc, sub_slab_coords, film_slab_coords]", "sub_path": "core/interface_generator.py", "file_name": "interface_generator.py", "file_ext": "py", "file_size_in_byte": 10240, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "numpy.dot", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 38, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 71, "usage_type": "call"}, {"api_name": "pymatgen.core.structure.Structure", "line_number": 73, "usage_type": "call"}, {"api_name": "pymatgen.Lattice", "line_number": 73, "usage_type": "call"}, {"api_name": "pymatgen.core.structure.Structure", "line_number": 75, "usage_type": "call"}, {"api_name": "pymatgen.Lattice", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 84, "usage_type": "call"}, {"api_name": "pymatgen.Structure", "line_number": 85, "usage_type": "call"}, {"api_name": "pymatgen.Lattice", "line_number": 85, "usage_type": "call"}, {"api_name": "pymatgen.Structure", "line_number": 87, "usage_type": "call"}, {"api_name": "pymatgen.Lattice", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 96, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 97, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 99, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 100, "usage_type": "attribute"}, {"api_name": "math.cos", "line_number": 102, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 103, "usage_type": "attribute"}, {"api_name": "math.cos", "line_number": 104, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 105, "usage_type": "attribute"}, {"api_name": "numpy.cross", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.cross", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 108, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 109, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 111, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 112, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 114, "usage_type": "attribute"}, {"api_name": "numpy.matmul", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 117, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 118, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 120, "usage_type": "attribute"}, {"api_name": "numpy.matmul", "line_number": 121, "usage_type": "call"}, {"api_name": "pymatgen.Lattice", "line_number": 122, "usage_type": "call"}, {"api_name": "pymatgen.Lattice", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.matmul", "line_number": 130, "usage_type": "call"}, {"api_name": "pymatgen.Structure", "line_number": 132, "usage_type": "call"}, {"api_name": "pymatgen.Structure", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 159, "usage_type": "call"}, {"api_name": "pymatgen.Lattice", "line_number": 180, "usage_type": "call"}, {"api_name": "pymatgen.Structure", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 202, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 204, "usage_type": "call"}, {"api_name": "pymatgen.core.structure.Structure", "line_number": 207, "usage_type": "call"}]} +{"seq_id": "384162084", "text": "import select, socket, queue as Queue\nimport chat_utils as chat_utils\nfrom colorama import Fore, Style\n\n#This class implements the downgrade attack which Trudy does by blocking the chat_STARTTLS messages between Alice and Bob.\n\nclass Downgrade_Server:\n #Setting up the downgrade server with IP address and port 8000.\n def __init__(self, self_ip, server_ip, client_ip):\n self.server = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n self.server.bind((self_ip, 8000))\n self.server.listen(5)\n print(Fore.GREEN + Style.BRIGHT + 'Server up and running! Waiting for connections...\\n')\n\n #Intercepting connections from the client to server.\n connection, client_address = self.server.accept()\n new_socket = socket.socket(socket.AF_INET,socket.SOCK_STREAM)\n #Connecting to the server to now establish an indirect link between a client and server\n new_socket.connect((server_ip, 8000))\n print(Fore.CYAN + Style.BRIGHT + \"Intercepting messages...\\n\")\n # Setting the connection to non blocking, so that the Trudy can send and recieve messages whenever.\n self.server.setblocking(0)\n new_socket.setblocking(0)\n self.start_downgrade(connection, new_socket)\n\n def start_downgrade(self, client_side, server_side):\n #Setting up the client side and server side of the downgrade server\n self.client_side = client_side\n self.server_side = server_side\n # The list of input streams present.\n self.inputs = [client_side, server_side]\n # The list of entities with pending outgoing messages.\n self.outputs = []\n # message queues for the pending outgoing messages.\n self.message_queues = {}\n # A message queue for the pending outgoing messages to the client.\n self.message_queues[client_side] = Queue.Queue()\n # A message queue for the pending outgoing messages to the server.\n self.message_queues[server_side] = Queue.Queue()\n # Lists to contain the received fragments of a message untill all of its fragments have arrived.\n self.fragment_lists = {}\n # A list to contain the received fragments of a message untill all of its fragments have arrived on the client side.\n self.fragment_lists[client_side] = []\n # A list to contain the received fragments of a message untill all of its fragments have arrived on the server side.\n self.fragment_lists[server_side] = []\n # The number of messages sent so far.\n self.received_message_numbers = {}\n self.received_message_numbers[client_side] = 0\n self.received_message_numbers[server_side] = 0\n self.lastline_type = client_side\n # Now running a loop for as long as the connection between the server and client exists.\n while len(self.inputs) > 1:\n # Using the select library to filter the inputs and outputs list and obtain from them the entities that are ready for IO.\n readable, writable, exceptional = select.select(self.inputs, self.outputs, self.inputs)\n # Iterating over all the entities that have pending messages to be read\n for s in readable:\n incoming_msg = s.recv(4096).decode('UTF-8')\n # If the entity is the client, then there is an incoming message from the client to the server.\n if s is client_side:\n #If the incoming message from client is to establish TLS connection,downgrade server responds with CHAT_STARTTLS_NOT_SUPPORTED.\n #This message is not sent to the server so server assumes that the client doesn't wish to use TLS protocol.\n if incoming_msg == chat_utils.CHAT_STARTTLS:\n # Adding the server to the outputs list, implying that there are pending messages to be sent to the server.\n if client_side not in self.outputs:\n self.outputs.append(client_side)\n response = chat_utils.CHAT_STARTTLS_NOT_SUPPORTED\n self.message_queues[client_side].put(response)\n # Other incoming messages thus read are enqueued into the message queue, implying that it has to be sent to the server.\n else:\n self.message_queues[server_side].put(incoming_msg)\n if server_side not in self.outputs:\n # Adding the server to the outputs list, implying that there are pending messages to be sent to the server.\n self.outputs.append(server_side)\n if chat_utils.CHAT_MESSAGE in incoming_msg:\n self.handle_new_message(s, incoming_msg)\n # If the entity is the server, then there is an incoming message from the server to the client.\n else:\n # The message thus read is enqueued into the message queue, implying that it has to be sent to the client.\n self.message_queues[client_side].put(incoming_msg)\n # Adding the client to the outputs list, implying that there are pending messages to be sent to the client.\n if client_side not in self.outputs:\n self.outputs.append(client_side)\n if chat_utils.CHAT_MESSAGE in incoming_msg:\n self.handle_new_message(s, incoming_msg)\n\n # Now iterating over the list of entities that have pending messages to be written to.\n for s in writable:\n try:\n next_msg = self.message_queues[s].get_nowait()\n except Queue.Empty:\n self.outputs.remove(s)\n else:\n # If the user types CHAT_CLOSE, it means that he intends to close the connection.\n if next_msg == chat_utils.CHAT_CLOSE:\n person = 'Bob'\n if s is server_side:\n person = 'Alice'\n print(Fore.RED + Style.BRIGHT + '\\n' + person +' closed the connection!', Style.RESET_ALL+'\\n')\n s.send(next_msg.encode('UTF-8'))\n self.close_connection(s)\n break\n else:\n # If the message is not CHAT_CLOSE, then in accordance with the protocol,the message is sent.\n s.send(next_msg.encode('UTF-8'))\n # Iterating over the list of entities that have thrown an exception.\n for s in exceptional:\n self.close_connection(s)\n\n # This function handles the messages received from the user.\n def handle_new_message(self, s, data):\n # First the details of the message like the message number, number of fragments and the fragment number are obtained.\n msg_num, num_fragments, fragment_num = chat_utils.get_message_details(data)\n # if the message number is not equal to the number of messages received yet, it implies that this is a new message.\n if self.received_message_numbers[s] != msg_num:\n self.received_message_numbers[s] = msg_num\n # If the new message has only one fragment, then we just print it.\n if num_fragments == 1:\n self.print_message(s, data[28:])\n # If it has more than one fragment then we append it into the fragment list.\n else:\n self.fragment_lists[s].append(data)\n # If the message received is not an entirely new message but a fragment of a message,\n else:\n # If the fragment received is indeed the last fragment, then we can parse all the received fragments, reconstruct the message and display it.\n if num_fragments == fragment_num:\n self.fragment_lists[s].append(data)\n received_msg = chat_utils.parse(self.fragment_lists[s])\n self.print_message(s, received_msg)\n self.fragment_lists[s].clear()\n # If the received fragment is not the last fragment, then we simply append it into the fragment list.\n else:\n self.fragment_lists[s].append(data)\n\n # This is a handy function to print the messages.\n def print_message(self, s, message):\n if self.lastline_type != s:\n print(\"\")\n self.lastline_type = s\n if s is self.client_side:\n print(Fore.YELLOW + Style.BRIGHT +'Alice says: ', Fore.BLUE + Style.BRIGHT + message, Fore.CYAN + Style.BRIGHT)\n else:\n print(Fore.MAGENTA + Style.BRIGHT +'Bob says: ', Fore.GREEN + Style.BRIGHT + message, Fore.CYAN + Style.BRIGHT)\n \n\n # This is a handy function to help with the closing of the connection.\n def close_connection(self, s):\n if s in self.outputs:\n self.outputs.remove(s)\n self.inputs.remove(s)\n s.close()\n del self.message_queues[s]\n del self.received_message_numbers[s]\n del self.fragment_lists[s]\n", "sub_path": "trudy/downgrade.py", "file_name": "downgrade.py", "file_ext": "py", "file_size_in_byte": 9130, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "socket.socket", "line_number": 10, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 10, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 10, "usage_type": "attribute"}, {"api_name": "colorama.Fore.GREEN", "line_number": 13, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 13, "usage_type": "name"}, {"api_name": "colorama.Style.BRIGHT", "line_number": 13, "usage_type": "attribute"}, {"api_name": "colorama.Style", "line_number": 13, "usage_type": "name"}, {"api_name": "socket.socket", "line_number": 17, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 17, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 17, "usage_type": "attribute"}, {"api_name": "colorama.Fore.CYAN", "line_number": 20, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 20, "usage_type": "name"}, {"api_name": "colorama.Style.BRIGHT", "line_number": 20, "usage_type": "attribute"}, {"api_name": "colorama.Style", "line_number": 20, "usage_type": "name"}, {"api_name": "queue.Queue", "line_number": 37, "usage_type": "call"}, {"api_name": "queue.Queue", "line_number": 39, "usage_type": "call"}, {"api_name": "select.select", "line_number": 54, "usage_type": "call"}, {"api_name": "chat_utils.CHAT_STARTTLS", "line_number": 62, "usage_type": "attribute"}, {"api_name": "chat_utils.CHAT_STARTTLS_NOT_SUPPORTED", "line_number": 66, "usage_type": "attribute"}, {"api_name": "chat_utils.CHAT_MESSAGE", "line_number": 74, "usage_type": "attribute"}, {"api_name": "chat_utils.CHAT_MESSAGE", "line_number": 83, "usage_type": "attribute"}, {"api_name": "queue.Empty", "line_number": 90, "usage_type": "attribute"}, {"api_name": "chat_utils.CHAT_CLOSE", "line_number": 94, "usage_type": "attribute"}, {"api_name": "colorama.Fore.RED", "line_number": 98, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 98, "usage_type": "name"}, {"api_name": "colorama.Style.BRIGHT", "line_number": 98, "usage_type": "attribute"}, {"api_name": "colorama.Style", "line_number": 98, "usage_type": "name"}, {"api_name": "colorama.Style.RESET_ALL", "line_number": 98, "usage_type": "attribute"}, {"api_name": "chat_utils.get_message_details", "line_number": 112, "usage_type": "call"}, {"api_name": "chat_utils.parse", "line_number": 127, "usage_type": "call"}, {"api_name": "colorama.Fore.YELLOW", "line_number": 140, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 140, "usage_type": "name"}, {"api_name": "colorama.Style.BRIGHT", "line_number": 140, "usage_type": "attribute"}, {"api_name": "colorama.Style", "line_number": 140, "usage_type": "name"}, {"api_name": "colorama.Fore.BLUE", "line_number": 140, "usage_type": "attribute"}, {"api_name": "colorama.Fore.CYAN", "line_number": 140, "usage_type": "attribute"}, {"api_name": "colorama.Fore.MAGENTA", "line_number": 142, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 142, "usage_type": "name"}, {"api_name": "colorama.Style.BRIGHT", "line_number": 142, "usage_type": "attribute"}, {"api_name": "colorama.Style", "line_number": 142, "usage_type": "name"}, {"api_name": "colorama.Fore.GREEN", "line_number": 142, "usage_type": "attribute"}, {"api_name": "colorama.Fore.CYAN", "line_number": 142, "usage_type": "attribute"}]} +{"seq_id": "495463671", "text": "from django.db import models\r\nfrom consulta.models_login import User\r\n\r\n# Create your models here.\r\n\r\nclass Pacientes(User):\r\n role = models.CharField(max_length=9, default='paciente')\r\n birth = models.DateField(verbose_name='Data de Nascimento') # formato: 1991-11-15\r\n cpf = models.CharField(max_length=11, null=True, verbose_name='CPF')\r\n\r\n def __str__(self):\r\n return self.get_full_name()\r\n\r\n class Meta:\r\n verbose_name = 'Paciente'\r\n verbose_name_plural = 'Pacientes'\r\n\r\n\r\nclass Medicos(User):\r\n role = models.CharField(max_length=9, default='medico')\r\n birth = models.DateField(verbose_name='Data de Nascimento') # formato: 1991-11-15\r\n cpf = models.CharField(max_length=11, null=True, verbose_name='CPF')\r\n crm = models.CharField(max_length=20, null=True)\r\n foto = models.ImageField(null=True, blank=True)\r\n\r\n def __str__(self):\r\n return self.get_full_name()\r\n\r\n @property\r\n def imageURL(self):\r\n try:\r\n url = self.foto.url\r\n except:\r\n url = ''\r\n return url\r\n\r\n class Meta:\r\n verbose_name = 'Médico'\r\n verbose_name_plural = 'Médicos'\r\n\r\n\r\nclass Especialidades(models.Model):\r\n tipos = [('225105 - Acupuntura','Médico Acumputurista'),\r\n ('225110 - Alergia e Imunologia','Médico Alergista e Imunologista'),\r\n ('225151 - Anestesiologia','Médico anestesiologista'),\r\n ('225115 - Angiologia','Médico angiologista'),\r\n ('225120 - Cardiologia','Médico Cardiologista'),\r\n ('225124A - Cardiologia Pediátrica','Médico Cardiologista'),\r\n ('225210 - Cirurgia Cardiovascular','Médico Cirurgião Cardiovascular'),\r\n ('225215 - Cirurgia de Cabeça e Pescoço','Médico cirurgião de cabeça e pescoço'),\r\n ('225295 - Cirurgia de Mão','Médico cirurgião da mão'),\r\n ('225225 - Cirurgia Geral ','Médico cirurgião geral '),\r\n ('225230 - Cirurgia Pediátrica','Médico cirurgião pediátrico'),\r\n ('225235 - Cirurgia Plástica','Médico cirurgião plástico '),\r\n ('225305 - Citopatologia','Médico citopatologista'),\r\n ('225125 - Clínica Médica','Médico clínico'),\r\n ('225280 - Coloproctologia','Médico proctologista'),\r\n ('225135 - Dermatologia','Médico dermatologista'),\r\n ('225155 - Endocrinologia e Metabologia','Médico endocrinologista e metabologista'),\r\n \r\n ('225310A - Endoscopia digestiva','Médico em endoscopia'),\r\n ('225310B - Endoscopia respiratoria','Médico em endoscopia'),\r\n ('225160 - Fisiatria','Médico fisiatra'),\r\n ('223605 - Fisioterapia','Fisioterapeuta geral'),\r\n ('223810 - Fonoaudiologia','Fonoaudiólogo'),\r\n ('225165 - Gastroenterologia','Médico gastroenterologista'),\r\n ('225175 - Genetica Medica','Médico geneticista'),\r\n ('225180 - Geriatria','Médico geriatra'),\r\n ('225250 - Ginecologia e obstetricia','Médico ginecologista e obstetra'),\r\n ('225185 - Hematologia e hemoterapia','Médico Hematologista'),\r\n ('225210A - Hemodinamica e Cardiologia Intervencionista','Hemodinamica e Cardiologia Intervencionista'),\r\n ('225165A - Hepatologia','Hepatologia'),\r\n ('225195 - Homeopatia','Médico Homeopata'),\r\n ('225103 - Infectologia','Médico infectologista'),\r\n ('225255 - Mastologia','Médico Mastologista'),\r\n ('225315 - Medicina nuclear','Médico em medicina nuclear'),\r\n ('225170 - Medico Clinico Geral','Médico Generalista'),\r\n ('225109 - Nefrologia','Médico Nefrologista'),\r\n ('225124H - Neonatologia','Neonatologia'),\r\n ('225124I - Neurologia pediatrica','Médico neurologista'),\r\n ('225260 - Neurocirurgia ','Médico neurocirurgião'),\r\n ('225112 - Neurologia','Médico neurologista'),\r\n\r\n ('225118 - Nutrologia','Médico nutrologista'),\r\n ('225265 - Oftalmologia','Médico oftalmologista'),\r\n ('225121 - Oncologia','Médico oncologista clínico'),\r\n ('225270 - Ortopedia e traumatologia','Médico ortopedista e traumatologista'),\r\n ('225275 - Otorrinolaringologia','Médico otorrinolaringologista'),\r\n ('225335 - Patologia clínica/Medicina laboratorial','Médico patologista clínico /medicina laboratorial'),\r\n ('225124 - Pediatria','Médico pediatra'),\r\n ('225127 - Pneumologia','Médico pneumologista'),\r\n ('251510 - Psicologia','Psicólogo clínico'),\r\n ('225133 - Psiquiatria','Médico psiquiatra'),\r\n ('225320 - Radiologia e diagnóstico por imagem','Médico em radiologia e diagnóstico por imagem'),\r\n ('225330 - Radioterapia','Médico radioterapeuta'),\r\n ('225136 - Reumatologia','Médico reumatologista'),\r\n ('223905 - Terapia ocupacional','Terapeuta ocupacional'),\r\n ('225285 - Urologia','Médico Médico urologista'),\r\n ('223268 - Cirurgia e traumatologia buco-maxilo-facial','Cirurgião dentista - traumatologista bucomaxilofacial'),\r\n ('223208 - Dentista clinico geral','Cirurgião dentista - clínico geral'),\r\n ('223212 - Endodontia','Cirurgião dentista - endodontista'),\r\n ('223220 - Estomatologia','Cirurgião dentista - estomatologista'),\r\n ('223224 - Implantodontia','Cirurgião dentista - implantodontista'),\r\n ('223236 - Odontopediatria','Cirurgião dentista - odontopediatra'),\r\n ('223240 - Ortodontia','Cirurgião dentista - ortopedista e ortodontista'),\r\n\r\n ('223248 - Periodontia','Cirurgião dentista - periodontista'),\r\n ('223256 - Protese dentaria','Cirurgião dentista - protesista'),\r\n ('223260 - Imaginologia odontologica','Imaginologia odontologica'),\r\n ('225220 - Cirurgia do aparelho digestivo','Médico cirurgião do aparelho digestivo'),\r\n ('225240 - Cirurgia toracica','Médico cirurgião torácico'),\r\n ('225124B - Endocrinologia pediatrica','Médico endocrinologista e metabologista'),\r\n ('223845 - Foniatria','Fonoaudiólogo'),\r\n ('225124C - Gastroenterologia pediatrica','Médico gastroenterologista'),\r\n ('225124D - Hematologia e hemoterapia pediatrica','Médico Hematologista'),\r\n ('225124E - Medicina do adolescente','Medicina do adolescente'),\r\n ('225150 - Medicina intensiva','Médico em medicina intensiva'),\r\n ('225124F - Medicina intensiva pediatrica','Médico em medicina intensiva'),\r\n ('225124G - Nefrologia pediatrica','Médico Nefrologista'),\r\n ('223710 - Nutricionista','Nutricionista'),\r\n ('225290 - Oncologia cirurgica','Médico oncologista clínico'),\r\n ('225122 - Oncologia pediatrica','Médico oncologista clínico'),\r\n ('225124K - Pneumologia pediatrica','Médico pneumologista'),\r\n ('225133A - Psiquiatria da infancia eda adolescencia','Psicopedagogo'),\r\n ('225124L - Reumatologia pediatrica','Médico reumatologista'),\r\n ('223280 - Dentistica','Dentistica'),\r\n ('223284 - Disfuncao temporomandibular e dor orafacial.','Cirurgião dentista - disfunção temporomandibular e dor orofacial'),\r\n ('223228 - Odontogeriatria','Cirurgião dentista - odontogeriatra'),\r\n \r\n ('223288 - Odontologia para pacientes com necessidades especiais','Cirurgião dentista - odontologia para pacientes com necessidades especiais'),\r\n ('223284A - Ortopedia funcional dos maxilares','Ortopedia funcional dos maxilares'),\r\n ('999999 - Medicamentos Hospitalares','Medicamentos Hospitalares'),\r\n ('999999A - Opme','Opme'),\r\n ('225151A - Dor','Dor'),\r\n ('225112A - Medicina do Sono','Medicina do Sono'),\r\n ('225203 - Cirurgia Vascular','Médico em cirurgia vascular'),\r\n ('225310 - Endoscopia','Médico em endoscopia'),\r\n ('225160A - Medicina Fisica e Reabilitação','Medicina Fisica e Reabilitação'),\r\n ('225325 - patologia','Médico patologista'),\r\n ('251605 - Assistente social ','Assistente social '),\r\n ('223505 - Enfermagem','Técnico de enfermagem'),\r\n ('999999B - Não Informado','CBO desconhecido ou não informado pelo solicitante'),\r\n\r\n ]\r\n especialidade = models.CharField(max_length=100, choices=tipos)\r\n \r\n\r\n def __str__(self):\r\n return self.especialidade\r\n \r\n class Meta:\r\n verbose_name = 'Especialidade'\r\n verbose_name_plural = 'Especialidades'\r\n \r\n\r\nclass Medicos_especialidade(models.Model):\r\n id_medico = models.ForeignKey(Medicos, on_delete=models.SET_NULL, null=True)\r\n id_especialidade = models.ForeignKey(Especialidades, on_delete=models.SET_NULL, null=True)\r\n preco = models.DecimalField(max_digits=7, decimal_places=2)\r\n certificado_especialidade = models.ImageField(null=True, blank=True)\r\n\r\n def __str__(self):\r\n return str(self.id)\r\n\r\n class Meta:\r\n verbose_name = 'Médico Por Especialidade'\r\n verbose_name_plural = 'Médicos Por Especialidade'\r\n\r\nclass Localidades(models.Model):\r\n user = models.ForeignKey(User, on_delete=models.SET_NULL, null=True)\r\n cep = models.CharField(max_length=11, null=True)\r\n rua = models.CharField(max_length=50, null=True)\r\n bairro = models.CharField(max_length=50, null=True)\r\n cidade = models.CharField(max_length=100, null=True)\r\n estado = models.CharField(max_length=2, null=True)\r\n complemento = models.CharField(max_length=30, null=True)\r\n\r\n\r\n def __str__(self):\r\n return str(self.id)\r\n\r\n class Meta:\r\n verbose_name = 'Localidade'\r\n verbose_name_plural = 'Localidades'\r\n\r\nclass Agendas(models.Model):\r\n id_medico = models.ForeignKey(Medicos, on_delete=models.SET_NULL, null=True)\r\n id_especialidade = models.ForeignKey(Especialidades, on_delete=models.SET_NULL, null=True)\r\n tipos = [('Presencial', 'Presencial'), ('Digital', 'Digital')]\r\n tipos_consulta = models.CharField(max_length=50, choices=tipos)\r\n data = models.DateField()\r\n hora = models.TimeField()\r\n\r\n def __str__(self):\r\n return str(self.id)\r\n\r\n class Meta:\r\n verbose_name = 'Agenda'\r\n verbose_name_plural = 'Agendas'\r\n\r\nclass Compras(models.Model):\r\n id_paciente = models.ForeignKey(Pacientes, on_delete=models.SET_NULL, null=True)\r\n data_emissao = models.DateTimeField(auto_now_add=True)\r\n complete = models.BooleanField(default=False, null=True, blank=True)\r\n transaction_id = models.CharField(max_length=200, null=True)\r\n\r\n\r\n def __str__(self):\r\n return str(self.id)\r\n \r\n\r\n def get_cart_total(self):\r\n comprasitems = self.compras_consulta_set.all()\r\n total = sum([item.get_total for item in comprasitems])\r\n return total\r\n\r\n @property\r\n def get_cart_items(self):\r\n comprasitems = self.compras_consulta_set.all()\r\n total = sum([item.quantity for item in comprasitems])\r\n return total\r\n\r\n class Meta:\r\n verbose_name = 'Compra'\r\n verbose_name_plural = 'Compras'\r\n\r\n\r\nclass Compras_consulta(models.Model):\r\n id_compra = models.ForeignKey(Compras, on_delete=models.SET_NULL, null=True)\r\n id_medicos_especialidade = models.ForeignKey(Medicos_especialidade, on_delete=models.SET_NULL, null=True)\r\n quantity = models.IntegerField(default=0, null=True, blank=True)\r\n\r\n\r\n def __str__(self):\r\n return str(self.id)\r\n\r\n\r\n @property\r\n def get_total(self):\r\n total = self.id_medicos_especialidade.preco * self.quantity\r\n return total\r\n\r\n class Meta:\r\n verbose_name = 'Consulta Por Compra'\r\n verbose_name_plural = 'Consultas Por Compra'\r\n\r\n\r\nclass Contato_paciente(models.Model):\r\n id_paciente = models.ForeignKey(Pacientes, on_delete=models.SET_NULL, blank=True, null=True)\r\n id_compra = models.ForeignKey(Compras, on_delete=models.SET_NULL, blank=True, null=True)\r\n nome_completo = models.CharField(max_length=200, null=True)\r\n email = models.EmailField(max_length=200, null=True)\r\n cep = models.CharField(max_length=200, null=True)\r\n rua = models.CharField(max_length=200, null=True)\r\n bairro = models.CharField(max_length=200, null=True)\r\n cidade = models.CharField(max_length=200, null=True)\r\n estado = models.CharField(max_length=200, null=True)\r\n items_pedido = models.CharField(max_length=200, null=True)\r\n total = models.CharField(max_length=200, null=True)\r\n data_compra = models.DateTimeField(auto_now_add=True)\r\n\r\n def __str__(self):\r\n return str(self.id)", "sub_path": "medline/consulta/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 12953, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "consulta.models_login.User", "line_number": 6, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 7, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 8, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 9, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 9, "usage_type": "name"}, {"api_name": "consulta.models_login.User", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.models.ImageField", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 42, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 42, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 145, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 145, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 156, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 156, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 157, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 157, "usage_type": "name"}, {"api_name": "django.db.models.SET_NULL", "line_number": 157, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 158, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 158, "usage_type": "name"}, {"api_name": "django.db.models.SET_NULL", "line_number": 158, "usage_type": "attribute"}, {"api_name": "django.db.models.DecimalField", "line_number": 159, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 159, "usage_type": "name"}, {"api_name": "django.db.models.ImageField", "line_number": 160, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 160, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 169, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 169, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 170, "usage_type": "call"}, {"api_name": "consulta.models_login.User", "line_number": 170, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 170, "usage_type": "name"}, {"api_name": "django.db.models.SET_NULL", "line_number": 170, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 171, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 171, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 172, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 172, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 173, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 173, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 174, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 174, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 175, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 175, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 176, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 176, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 186, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 186, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 187, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 187, "usage_type": "name"}, {"api_name": "django.db.models.SET_NULL", "line_number": 187, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 188, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 188, "usage_type": "name"}, {"api_name": "django.db.models.SET_NULL", "line_number": 188, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 190, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 190, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 191, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 191, "usage_type": "name"}, {"api_name": "django.db.models.TimeField", "line_number": 192, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 192, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 201, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 201, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 202, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 202, "usage_type": "name"}, {"api_name": "django.db.models.SET_NULL", "line_number": 202, "usage_type": "attribute"}, {"api_name": "django.db.models.DateTimeField", "line_number": 203, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 203, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 204, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 204, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 205, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 205, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 228, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 228, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 229, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 229, "usage_type": "name"}, {"api_name": "django.db.models.SET_NULL", "line_number": 229, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 230, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 230, "usage_type": "name"}, {"api_name": "django.db.models.SET_NULL", "line_number": 230, "usage_type": "attribute"}, {"api_name": "django.db.models.IntegerField", "line_number": 231, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 231, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 248, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 248, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 249, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 249, "usage_type": "name"}, {"api_name": "django.db.models.SET_NULL", "line_number": 249, "usage_type": "attribute"}, {"api_name": "django.db.models.ForeignKey", "line_number": 250, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 250, "usage_type": "name"}, {"api_name": "django.db.models.SET_NULL", "line_number": 250, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 251, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 251, "usage_type": "name"}, {"api_name": "django.db.models.EmailField", "line_number": 252, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 252, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 253, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 253, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 254, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 254, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 255, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 255, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 256, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 256, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 257, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 257, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 258, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 258, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 259, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 259, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 260, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 260, "usage_type": "name"}]} +{"seq_id": "489583457", "text": "import re\nimport logging\n\nLOGGER = logging.getLogger()\n\n\nIMAGE_FILE_RE = re.compile( \n \"(?P\\\\d{8})/\" +\n \"CS(?P\\\\d{1})/\" +\n \"(?P.+)/\" +\n \"p(?P\\\\d{1,2}|XX)/\" +\n \"ch(?P\\\\d{1}|XX)/\" +\n \"z(?P\\\\d{1,2}|XX)\"+\n \"(?P.*)\" +\n \"\\.\" + \n \"(?P.+)\"\n )\n\nclass LSMImageFilename:\n @classmethod\n def parse(cls, image_filename_str):\n match = IMAGE_FILE_RE.match(image_filename_str)\n if not match:\n raise Exception(\"invalid image filename: %s\" % image_filename_str)\n return cls(\n date=match[\"date\"],\n position=match[\"position\"],\n group=match[\"group\"],\n f=(None if match[\"f\"] == \"XXX\" else int(match[\"f\"])),\n z=(None if match[\"z\"] == \"XX\" else int(match[\"z\"])),\n c=(None if match[\"c\"] == \"XX\" else int(match[\"c\"])),\n suffix=match[\"suffix\"],\n extension=match[\"extension\"],\n )\n\n def __init__(self, date, position, group, f, z, c, suffix, extension):\n self.date = date\n self.position = position\n self.group = group\n self.f = f\n self.z = z\n self.c = c\n self.suffix = suffix\n self.extension = extension\n\n def __str__(self):\n return \"%s/CS%s/%s/p%s/ch%s/z%s%s.%s\" % (\n self.date,\n self.position,\n self.group,\n self.f_str,\n self.c_str,\n self.z_str,\n self.suffix,\n self.extension\n )\n\n def __copy__(self):\n return LSMImageFilename(\n date=self.date,\n position=self.position,\n group=self.group,\n f=self.f,\n z=self.z,\n c=self.c,\n suffix=self.suffix,\n extension=self.extension\n )\n\n @property\n def f_str(self):\n return \"XXX\" if not self.f else (\"%i\" % self.f)\n\n @property\n def z_str(self):\n return \"XX\" if not self.z else (\"%i\" % self.z)\n\n @property\n def c_str(self):\n return \"XX\" if not self.c else (\"%i\" % self.c)\n", "sub_path": "models/image_name_dictionaries/image_filename_LSM.py", "file_name": "image_filename_LSM.py", "file_ext": "py", "file_size_in_byte": 1861, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "logging.getLogger", "line_number": 4, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 7, "usage_type": "call"}]} +{"seq_id": "23854682", "text": "from tkinter import *\r\nfrom tkinter import messagebox\r\nimport pymysql\r\nimport sqlite3\r\n\r\n\r\ndef ingresodenotas():\r\n\r\n conn=sqlite3.connect(\"notas\")\r\n \r\n\r\n cur=conn.cursor()\r\n cur.execute(\"INSERT INTO calificaciones1 VALUES (null,'\" + AF1.get() +\r\n \"','\" + so1.get() +\r\n \"','\" + PA1.get() +\r\n \"','\" + Comu1.get() + \"')\")\r\n\r\n conn.commit()\r\n \r\n \r\n messagebox.showinfo(\"DB\",\"Registro ingresado con exito\")\r\ndef leer():\r\n\r\n conn=sqlite3.connect(\"notas\")\r\n \r\n\r\n cur=conn.cursor()\r\n cur.execute(\"SELECT * FROM calificaciones1 WHERE ID=\" + ID1.get())\r\n\r\n elUsuario=cur.fetchall()\r\n\r\n for usuario in elUsuario:\r\n\r\n ID1.set(usuario[0])\r\n AF1.set(usuario[1])\r\n so1.set(usuario[2])\r\n PA1.set(usuario[3])\r\n Comu1.set(usuario[4])\r\n conn.commit()\r\ndef Limpiar ():\r\n\tID1.set(\" \")\r\n\tAF1.set(\" \")\r\n\tso1.set(\" \")\r\n\tPA1.set(\" \")\r\n\tComu1.set(\" \")\r\ndef salir():\r\n\tvalor=messagebox.askquestion(\"salir\",\"deseas salir de la aplicacion\")\r\n\tif valor==\"yes\":\r\n\t\ttk.destroy()\r\n\r\n\r\ndef actualizar():\r\n conn=sqlite3.connect(\"notas\")\r\n cur=conn.cursor()\r\n cur.execute(\"UPDATE calificaciones1 SET Algoritmos='\" + AF1.get() +\r\n \"', SistemasOp='\" + so1.get() +\r\n \"', Programacion='\" + PA1.get() +\r\n \"', Comunicaciones='\" + Comu1.get() +\r\n \"' WHERE ID=\" + ID1.get())\r\n \r\n conn.commit()\r\n messagebox.showinfo(\"DB\",\"Registro Actualizado con exito\")\r\ndef eliminar():\r\n\r\n conn=sqlite3.connect(\"notas\")\r\n cur=conn.cursor()\r\n cur.execute(\"DELETE FROM calificaciones1 WHERE ID=\" + ID1.get())\r\n conn.commit()\r\n messagebox.showinfo(\"BD\",\"Registro borrado con exito\")\r\n\r\ntk=Tk()\r\ntk.title(\"Ingreso de datos\")\r\nventana=Frame (height=400,width=700)\r\nventana.pack(padx=5,pady=5)\r\nventana.configure(background=\"MediumPurple2\")\r\ntema=Label(ventana,font=('Times New Roman',12,'bold'),text=\"Registro de notas\",padx=80,pady=3,bd=5,background=\"gray77\").place(x=170,y=0)\r\n#nota1\r\nnum_ced=Label(ventana,font=('Times New Roman',12,'bold'),text=\"Ingrese la ID del alumno que desea averiguar\",background=\"#91F467\").place(x=195,y=37)\r\nID1=StringVar()\r\ntxt=Entry(ventana,font=('Times New Roman',15,'bold'),textvariable=ID1,width=10,bg=\"cyan\",bd=5).place(x=400,y=70)\r\n#nota2\r\nAF=Label(ventana,font=('Times New Roman',12,'bold'),text=\"Algoritmos funtamentales\",background=\"#91F467\").place(x=0,y=80)\r\nAF1=StringVar()\r\ntxt1=Entry(ventana,font=('Times New Roman',15,'bold'),textvariable=AF1,width=10,bg=\"powder blue\",bd=5).place(x=210,y=80)\r\n#notas3\r\nso=Label(ventana,font=('Times New Roman',12,'bold'),text=\"Sistemas Operativos \",background=\"#91F467\").place(x=0,y=123)\r\nso1=StringVar()\r\ntxt2=Entry(ventana,font=('Times New Roman',15,'bold'),textvariable=so1,width=10,bg=\"powder blue\",bd=5).place(x=210,y=123)\r\n#notas4\r\nPA=Label(ventana,font=('Times New Roman',12,'bold'),text=\"Programacion avanzada \",background=\"#91F467\").place(x=0,y=166)\r\nPA1=StringVar()\r\ntxt3=Entry(ventana,font=('Times New Roman',15,'bold'),textvariable=PA1,width=10,bg=\"powder blue\",bd=5).place(x=210,y=166)\r\n#notas5\r\nComu=Label(ventana,font=('Times New Roman',12,'bold'),text=\"Comunicaciones\",background=\"#91F467\").place(x=0,y=209)\r\nComu1=StringVar()\r\ntxt4=Entry(ventana,font=('Times New Roman',15,'bold'),textvariable=Comu1,width=10,bg=\"powder blue\",bd=5).place(x=210,y=209)\r\n\r\n#botones\r\n\r\n#conectar=Button(ventana,font=('Times New Roman',12,'bold'),bd=15,command=conectar,text=\"conectar\",padx=20,pady=5,background=\"#91F467\").place(x=20,y=340)\r\ningresar=Button(ventana,font=('Times New Roman',12,'bold'),bd=15,command=ingresodenotas,text=\"Ingresar\",padx=20,pady=5,background=\"#91F467\").place(x=0,y=300)\r\nlimpiar=Button(ventana,font=('Times New Roman',12,'bold'),bd=15,command=Limpiar,text=\"Limpiar\",padx=20,pady=5,background=\"#FF60F3\").place(x=130,y=300)\r\nsalir=Button(ventana,font=('Times New Roman',12,'bold'),bd=15,command=salir,text=\"salir\",padx=20,pady=5,background=\"#FF60F3\").place(x=268,y=300)\r\nactualizar=Button(ventana,font=('Times New Roman',12,'bold'),bd=15,command=actualizar,text=\"Actualizar\",padx=20,pady=5,background=\"#91F467\").place(x=400,y=300)\r\nLeer=Button(ventana,font=('Times New Roman',12,'bold'),bd=15,command=leer,text=\"Leer\",padx=20,pady=5,background=\"#91F467\").place(x=580,y=300)\r\neliminar=Button(ventana,font=('Times New Roman',12,'bold'),bd=15,command=eliminar,text=\"borrar\",padx=20,pady=5,background=\"#91F467\").place(x=580,y=220)\r\ncanvas = Canvas(tk,width=300, height=300)\r\ncanvas.pack()\r\n\r\nesfera = PhotoImage(file='buo.png')\r\ncanvas.create_image(0, 0, anchor=NW, image=esfera)\r\n\r\ntk.mainloop()", "sub_path": "Juego_Retro_P_A-master/programacion avanzada base de datos/notas_saew.py", "file_name": "notas_saew.py", "file_ext": "py", "file_size_in_byte": 4591, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "sqlite3.connect", "line_number": 9, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 21, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 21, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 24, "usage_type": "call"}, {"api_name": "tkinter.messagebox.askquestion", "line_number": 47, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 47, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 53, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 62, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 62, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 65, "usage_type": "call"}, {"api_name": "tkinter.messagebox.showinfo", "line_number": 69, "usage_type": "call"}, {"api_name": "tkinter.messagebox", "line_number": 69, "usage_type": "name"}]} +{"seq_id": "116729729", "text": "\"\"\"empty message\n\nRevision ID: 43c36ca0314d\nRevises: ab3b53b230a\nCreate Date: 2015-12-16 21:23:57.583027\n\n\"\"\"\n\n# revision identifiers, used by Alembic.\nrevision = '43c36ca0314d'\ndown_revision = 'ab3b53b230a'\n\nfrom alembic import op\nimport sqlalchemy as sa\n\n\ndef upgrade():\n ### commands auto generated by Alembic - please adjust! ###\n op.create_table('feedback',\n sa.Column('id', sa.Integer(), nullable=False),\n sa.Column('value', sa.Boolean(), nullable=True),\n sa.Column('timestamp', sa.DateTime(), nullable=True),\n sa.Column('user_id', sa.Integer(), nullable=True),\n sa.Column('answer_id', sa.Integer(), nullable=True),\n sa.ForeignKeyConstraint(['answer_id'], ['answer.id'], ),\n sa.ForeignKeyConstraint(['user_id'], ['user.id'], ),\n sa.PrimaryKeyConstraint('id')\n )\n op.drop_column(u'answer', 'feedback')\n op.drop_column(u'user', 'credit')\n ### end Alembic commands ###\n\n\ndef downgrade():\n ### commands auto generated by Alembic - please adjust! ###\n op.add_column(u'user', sa.Column('credit', sa.INTEGER(), autoincrement=False, nullable=True))\n op.add_column(u'answer', sa.Column('feedback', sa.VARCHAR(), autoincrement=False, nullable=True))\n op.drop_table('feedback')\n ### end Alembic commands ###\n", "sub_path": "migrations/versions/43c36ca0314d_.py", "file_name": "43c36ca0314d_.py", "file_ext": "py", "file_size_in_byte": 1260, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "alembic.op.create_table", "line_number": 19, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 19, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 20, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 20, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlalchemy.Boolean", "line_number": 21, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.DateTime", "line_number": 22, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 23, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.Integer", "line_number": 24, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 25, "usage_type": "call"}, {"api_name": "sqlalchemy.ForeignKeyConstraint", "line_number": 26, "usage_type": "call"}, {"api_name": "sqlalchemy.PrimaryKeyConstraint", "line_number": 27, "usage_type": "call"}, {"api_name": "alembic.op.drop_column", "line_number": 29, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 29, "usage_type": "name"}, {"api_name": "alembic.op.drop_column", "line_number": 30, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 30, "usage_type": "name"}, {"api_name": "alembic.op.add_column", "line_number": 36, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 36, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 36, "usage_type": "call"}, {"api_name": "sqlalchemy.INTEGER", "line_number": 36, "usage_type": "call"}, {"api_name": "alembic.op.add_column", "line_number": 37, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 37, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 37, "usage_type": "call"}, {"api_name": "sqlalchemy.VARCHAR", "line_number": 37, "usage_type": "call"}, {"api_name": "alembic.op.drop_table", "line_number": 38, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 38, "usage_type": "name"}]} +{"seq_id": "399828948", "text": "import cv2\nimport sys\nfrom moviepy.editor import VideoFileClip\n\n(major_ver, minor_ver, subminor_ver) = (cv2.cv2.__version__).split('.')\n\nfileToProcess = \"test.mp4\"\nfileToSave = \"out.mp4\"\nsecondsToSkip = 0\n\nif __name__ == '__main__' :\n # first we cut the seconds off of the video\n if secondsToSkip > 0:\n clip = VideoFileClip(fileToProcess).subclip(secondsToSkip)\n clip.write_videofile(fileToSave, codec=\"mpeg4\")\n fileToProcess=fileToSave\n\n # the trackers included in OpenCV version 4.5.1\n tracker_types = ['MIL','KCF', 'GOTURN', 'CSRT']\n tracker_type = tracker_types[3]\n if int(minor_ver) < 3:\n tracker = cv2.cv2.Tracker_create(tracker_type)\n else:\n if tracker_type == 'MIL':\n tracker = cv2.cv2.TrackerMIL_create()\n if tracker_type == 'KCF':\n tracker = cv2.cv2.TrackerKCF_create()\n if tracker_type == 'GOTURN':\n # not working try to remove opencv-contrib-python\n tracker = cv2.cv2.TrackerGOTURN_create()\n if tracker_type == \"CSRT\":\n tracker = cv2.cv2.TrackerCSRT_create()\n\n # Read video\n video = cv2.cv2.VideoCapture(fileToProcess)\n\n # Exit if video not opened.\n if not video.isOpened():\n print(\"Could not open video\")\n sys.exit()\n\n # Read first frame.\n ok, frame = video.read()\n if not ok:\n print('Cannot read video file')\n sys.exit()\n \n video.release()\n\n # Define an initial bounding box if you know it\n# bbox = (488*2, 266*2, 629*2, 730*2)\n\n # Uncomment the line below to select a different bounding box\n bbox = cv2.cv2.selectROI('Select Area', frame, False)\n cv2.cv2.destroyAllWindows()\n\n # Initialize tracker with first frame and bounding box\n ok = tracker.init(frame, bbox)\n\n def blur(image):\n frame = image.copy()\n try:\n ok, bbox = tracker.update(frame)\n if ok:\n blured = cv2.cv2.blur(frame,(int(frame.shape[0]*.05),int(frame.shape[0]*.05)))\n bluredroi = blured[int(bbox[1]):int(bbox[1]+bbox[3]), int(bbox[0]):int(bbox[0]+bbox[2])] \n frame[int(bbox[1]):int(bbox[1]+bbox[3]), int(bbox[0]):int(bbox[0]+bbox[2])] = bluredroi\n # in case you want a rectangle around the object\n # p1 = (int(bbox[0]), int(bbox[1]))\n # p2 = (int(bbox[0] + bbox[2]), int(bbox[1] + bbox[3]))\n # cv2.cv2.rectangle(frame, p1, p2, (255,0,0), 2, 1)\n else:\n return frame\n # print(\"Error\")\n # sys.exit()\n finally:\n return frame\n \n\n clip = VideoFileClip(fileToProcess)\n clip_blurred = clip.fl_image(blur)\n clip_blurred.write_videofile(fileToSave)", "sub_path": "mosaic.py", "file_name": "mosaic.py", "file_ext": "py", "file_size_in_byte": 2774, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "cv2.cv2.__version__.split", "line_number": 5, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 5, "usage_type": "attribute"}, {"api_name": "moviepy.editor.VideoFileClip", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.cv2.Tracker_create", "line_number": 22, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 22, "usage_type": "attribute"}, {"api_name": "cv2.cv2.TrackerMIL_create", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 25, "usage_type": "attribute"}, {"api_name": "cv2.cv2.TrackerKCF_create", "line_number": 27, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 27, "usage_type": "attribute"}, {"api_name": "cv2.cv2.TrackerGOTURN_create", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 30, "usage_type": "attribute"}, {"api_name": "cv2.cv2.TrackerCSRT_create", "line_number": 32, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 32, "usage_type": "attribute"}, {"api_name": "cv2.cv2.VideoCapture", "line_number": 35, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 35, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 40, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 46, "usage_type": "call"}, {"api_name": "cv2.cv2.selectROI", "line_number": 54, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 54, "usage_type": "attribute"}, {"api_name": "cv2.cv2.destroyAllWindows", "line_number": 55, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 55, "usage_type": "attribute"}, {"api_name": "cv2.cv2.blur", "line_number": 65, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 65, "usage_type": "attribute"}, {"api_name": "moviepy.editor.VideoFileClip", "line_number": 80, "usage_type": "call"}]} +{"seq_id": "557094202", "text": "import logging\n\nimport pytest\nimport tensorflow as tf\n\nfrom ludwig.combiners.combiners import (\n ConcatCombiner,\n SequenceConcatCombiner,\n SequenceCombiner,\n TabNetCombiner,\n ComparatorCombiner,\n TransformerCombiner,\n TabTransformerCombiner,\n sequence_encoder_registry,\n)\n\nlogger = logging.getLogger(__name__)\nlogger.setLevel(logging.INFO)\nlogging.getLogger(\"ludwig\").setLevel(logging.INFO)\n\nBATCH_SIZE = 16\nSEQ_SIZE = 12\nHIDDEN_SIZE = 128\nOTHER_HIDDEN_SIZE = 32\nFC_SIZE = 64\nBASE_FC_SIZE = 256\n\n\n# set up simulated encoder outputs\n@pytest.fixture\ndef encoder_outputs():\n # generates simulated encoder outputs dictionary:\n # feature_1: shape [b, h1] tensor\n # feature_2: shape [b, h2] tensor\n # feature_3: shape [b, s, h1] tensor\n # feature_4: shape [b, sh, h2] tensor\n\n encoder_outputs = {}\n shapes_list = [\n [BATCH_SIZE, HIDDEN_SIZE],\n [BATCH_SIZE, OTHER_HIDDEN_SIZE],\n [BATCH_SIZE, SEQ_SIZE, HIDDEN_SIZE],\n [BATCH_SIZE, SEQ_SIZE, OTHER_HIDDEN_SIZE],\n ]\n feature_names = [\"feature_\" + str(i + 1) for i in range(len(shapes_list))]\n\n for feature_name, batch_shape in zip(feature_names, shapes_list):\n encoder_outputs[feature_name] = {\n \"encoder_output\": tf.random.normal(batch_shape, dtype=tf.float32)\n }\n if len(batch_shape) > 2:\n encoder_outputs[feature_name][\n \"encoder_output_state\"] = tf.random.normal(\n [batch_shape[0], batch_shape[2]], dtype=tf.float32\n )\n\n return encoder_outputs\n\n\n# setup encoder outputs for ComparatorCombiner\n@pytest.fixture\ndef encoder_comparator_outputs():\n # generates simulated encoder outputs dictionary:\n # feature_1: shape [b, h1] tensor\n # feature_2: shape [b, h2] tensor\n # feature_3: shape [b, s, h1] tensor\n # feature_4: shape [b, sh, h2] tensor\n\n encoder_outputs = {}\n shapes_list = [\n [BATCH_SIZE, HIDDEN_SIZE],\n [BATCH_SIZE, OTHER_HIDDEN_SIZE],\n [BATCH_SIZE, SEQ_SIZE, HIDDEN_SIZE],\n [BATCH_SIZE, SEQ_SIZE, OTHER_HIDDEN_SIZE],\n ]\n text_feature_names = [\"text_feature_\" + str(i + 1) for i in\n range(len(shapes_list))]\n image_feature_names = [\n \"image_feature_\" + str(i + 1) for i in range(len(shapes_list))\n ]\n for i, (feature_name, batch_shape) in enumerate(\n zip(text_feature_names, shapes_list)\n ):\n # is there a better way to do this?\n if i == 0 or i == 3:\n dot_product_shape = [batch_shape[0], BASE_FC_SIZE]\n encoder_outputs[feature_name] = {\n \"encoder_output\": tf.random.normal(dot_product_shape,\n dtype=tf.float32)\n }\n else:\n encoder_outputs[feature_name] = {\n \"encoder_output\": tf.random.normal(batch_shape,\n dtype=tf.float32)\n }\n\n for i, (feature_name, batch_shape) in enumerate(\n zip(image_feature_names, shapes_list)\n ):\n if i == 0 or i == 3:\n dot_product_shape = [batch_shape[0], BASE_FC_SIZE]\n encoder_outputs[feature_name] = {\n \"encoder_output\": tf.random.normal(dot_product_shape,\n dtype=tf.float32)\n }\n else:\n encoder_outputs[feature_name] = {\n \"encoder_output\": tf.random.normal(batch_shape,\n dtype=tf.float32)\n }\n\n return encoder_outputs\n\n\n# test for simple concatenation combiner\n@pytest.mark.parametrize(\"fc_layer\",\n [None, [{\"fc_size\": 64}, {\"fc_size\": 64}]])\ndef test_concat_combiner(encoder_outputs, fc_layer):\n # clean out unneeded encoder outputs\n del encoder_outputs[\"feature_3\"]\n del encoder_outputs[\"feature_4\"]\n\n # setup combiner to test\n combiner = ConcatCombiner(fc_layers=fc_layer)\n\n # concatenate encoder outputs\n results = combiner(encoder_outputs)\n\n # required key present\n assert \"combiner_output\" in results\n\n # confirm correct output shapes\n if fc_layer:\n assert results[\"combiner_output\"].shape.as_list() == [BATCH_SIZE,\n FC_SIZE]\n else:\n # calculate expected hidden size for concatenated tensors\n hidden_size = 0\n for k in encoder_outputs:\n hidden_size += encoder_outputs[k][\"encoder_output\"].shape[1]\n\n assert results[\"combiner_output\"].shape.as_list() == [BATCH_SIZE,\n hidden_size]\n\n\n# test for sequence concatenation combiner\n@pytest.mark.parametrize(\"reduce_output\", [None, \"sum\"])\n@pytest.mark.parametrize(\"main_sequence_feature\", [None, \"feature_3\"])\ndef test_sequence_concat_combiner(\n encoder_outputs, main_sequence_feature, reduce_output\n):\n combiner = SequenceConcatCombiner(\n main_sequence_feature=main_sequence_feature,\n reduce_output=reduce_output\n )\n\n # calculate expected hidden size for concatenated tensors\n hidden_size = 0\n for k in encoder_outputs:\n hidden_size += encoder_outputs[k][\"encoder_output\"].shape[-1]\n\n # concatenate encoder outputs\n results = combiner(encoder_outputs)\n\n # required key present\n assert \"combiner_output\" in results\n\n # confirm correct shape\n if reduce_output is None:\n assert results[\"combiner_output\"].shape.as_list() == [\n BATCH_SIZE,\n SEQ_SIZE,\n hidden_size,\n ]\n else:\n assert results[\"combiner_output\"].shape.as_list() == [BATCH_SIZE,\n hidden_size]\n\n\n# test for sequence combiner\n@pytest.mark.parametrize(\"reduce_output\", [None, \"sum\"])\n@pytest.mark.parametrize(\"encoder\", sequence_encoder_registry)\n@pytest.mark.parametrize(\"main_sequence_feature\", [None, \"feature_3\"])\ndef test_sequence_combiner(\n encoder_outputs, main_sequence_feature, encoder, reduce_output\n):\n combiner = SequenceCombiner(\n main_sequence_feature=main_sequence_feature,\n encoder=encoder,\n reduce_output=reduce_output,\n )\n\n # calculate expected hidden size for concatenated tensors\n hidden_size = 0\n for k in encoder_outputs:\n hidden_size += encoder_outputs[k][\"encoder_output\"].shape[-1]\n\n # concatenate encoder outputs\n results = combiner(encoder_outputs)\n\n # required key present\n assert \"combiner_output\" in results\n\n combiner_shape = results[\"combiner_output\"].shape\n # test for correct dimension\n if reduce_output:\n assert len(combiner_shape) == 2\n else:\n assert len(combiner_shape) == 3\n\n # Shape test assumes on Ludwig sequence encoder defaults\n # parallel encoders: # layers = 4, fc_size=256\n # non-parallel encoders: fc_size=256\n # if defaults change, then this test has to be updated\n default_layer = 4\n default_fc_size = 256\n\n if \"parallel\" in encoder:\n combiner_shape[-1] == default_layer * default_fc_size\n else:\n combiner_shape[-1] == default_fc_size\n\n\ndef tabnet_encoder_outputs():\n # Need to do this in a function, otherwise TF will try to initialize\n # too early\n return {\n 'batch_128': {\n 'feature_1': {\n 'encoder_output': tf.random.normal(\n [128, 1],\n dtype=tf.float32\n )\n },\n 'feature_2': {\n 'encoder_output': tf.random.normal(\n [128, 1],\n dtype=tf.float32\n )\n },\n },\n 'inputs': {\n 'feature_1': {\n 'encoder_output': tf.keras.Input(\n (),\n dtype=tf.float32,\n name='feature_1',\n )\n },\n 'feature_2': {\n 'encoder_output': tf.keras.Input(\n (),\n dtype=tf.float32,\n name='feature_2',\n )\n },\n }\n }\n\n\n@pytest.mark.parametrize(\"encoder_outputs_key\", ['batch_128', 'inputs'])\ndef test_tabnet_combiner(encoder_outputs_key):\n encoder_outputs = tabnet_encoder_outputs()[encoder_outputs_key]\n\n # setup combiner to test\n combiner = TabNetCombiner(\n size=2,\n output_size=2,\n num_steps=3,\n num_total_blocks=4,\n num_shared_blocks=2,\n dropout=0.1\n )\n\n # concatenate encoder outputs\n results = combiner(encoder_outputs)\n\n # required key present\n assert 'combiner_output' in results\n assert 'attention_masks' in results\n\n\n@pytest.mark.parametrize(\"fc_layer\",\n [None, [{\"fc_size\": 64}, {\"fc_size\": 64}]])\n@pytest.mark.parametrize(\"entity_1\", [[\"text_feature_1\", \"text_feature_2\"]])\n@pytest.mark.parametrize(\"entity_2\", [[\"image_feature_1\", \"image_feature_2\"]])\ndef test_comparator_combiner(encoder_comparator_outputs, fc_layer, entity_1,\n entity_2):\n # clean out unneeded encoder outputs since we only have 2 layers\n del encoder_comparator_outputs[\"text_feature_3\"]\n del encoder_comparator_outputs[\"image_feature_3\"]\n del encoder_comparator_outputs[\"text_feature_4\"]\n del encoder_comparator_outputs[\"image_feature_4\"]\n\n # setup combiner to test set to 256 for case when none as it's the default size\n fc_size = fc_layer[0][\"fc_size\"] if fc_layer else 256\n combiner = ComparatorCombiner(\n entity_1, entity_2, fc_layers=fc_layer, fc_size=fc_size\n )\n\n # concatenate encoder outputs\n results = combiner(encoder_comparator_outputs)\n\n # required key present\n assert \"combiner_output\" in results\n\n # confirm correct output shapes\n # concat on axis=1\n # because of dot products, 2 of the shapes added will be the fc_size\n # other 2 will be of shape BATCH_SIZE\n # this assumes dimensionality = 2\n size = BATCH_SIZE * 2 + fc_size * 2\n assert results[\"combiner_output\"].shape.as_list() == [BATCH_SIZE, size]\n\n\ndef test_transformer_combiner(encoder_outputs):\n # clean out unneeded encoder outputs\n encoder_outputs = {}\n encoder_outputs['feature_1'] = {\n 'encoder_output': tf.random.normal(\n [128, 1],\n dtype=tf.float32\n )\n }\n encoder_outputs['feature_2'] = {\n 'encoder_output': tf.random.normal(\n [128, 1],\n dtype=tf.float32\n )\n }\n\n input_features_def = [\n {'name': 'feature_1', 'type': 'numerical'},\n {'name': 'feature_2', 'type': 'numerical'}\n ]\n\n # setup combiner to test\n combiner = TransformerCombiner(\n input_features=input_features_def\n )\n\n # concatenate encoder outputs\n results = combiner(encoder_outputs)\n\n # required key present\n assert 'combiner_output' in results\n\n\ndef test_tabtransformer_combiner(encoder_outputs):\n # clean out unneeded encoder outputs\n encoder_outputs = {}\n encoder_outputs['feature_1'] = {\n 'encoder_output': tf.random.normal(\n [128, 1],\n dtype=tf.float32\n )\n }\n encoder_outputs['feature_2'] = {\n 'encoder_output': tf.random.normal(\n [128, 16],\n dtype=tf.float32\n )\n }\n\n input_features_def = [\n {'name': 'feature_1', 'type': 'numerical'},\n {'name': 'feature_2', 'type': 'category'}\n ]\n\n # setup combiner to test\n combiner = TabTransformerCombiner(\n input_features=input_features_def\n )\n\n # concatenate encoder outputs\n results = combiner(encoder_outputs)\n\n # required key present\n assert 'combiner_output' in results\n\n # setup combiner to test\n combiner = TabTransformerCombiner(\n input_features=input_features_def,\n embed_input_feature_name=56\n )\n\n # concatenate encoder outputs\n results = combiner(encoder_outputs)\n\n # required key present\n assert 'combiner_output' in results\n\n # setup combiner to test\n combiner = TabTransformerCombiner(\n input_features=input_features_def,\n embed_input_feature_name='add'\n )\n\n # concatenate encoder outputs\n results = combiner(encoder_outputs)\n\n # required key present\n assert 'combiner_output' in results\n", "sub_path": "tests/integration_tests/test_combiners.py", "file_name": "test_combiners.py", "file_ext": "py", "file_size_in_byte": 12397, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "logging.getLogger", "line_number": 17, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 18, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 19, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 19, "usage_type": "attribute"}, {"api_name": "tensorflow.random.normal", "line_number": 49, "usage_type": "call"}, {"api_name": "tensorflow.random", "line_number": 49, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 49, "usage_type": "attribute"}, {"api_name": "tensorflow.random.normal", "line_number": 53, "usage_type": "call"}, {"api_name": "tensorflow.random", "line_number": 53, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 54, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 30, "usage_type": "attribute"}, {"api_name": "tensorflow.random.normal", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.random", "line_number": 88, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 89, "usage_type": "attribute"}, {"api_name": "tensorflow.random.normal", "line_number": 93, "usage_type": "call"}, {"api_name": "tensorflow.random", "line_number": 93, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 94, "usage_type": "attribute"}, {"api_name": "tensorflow.random.normal", "line_number": 103, "usage_type": "call"}, {"api_name": "tensorflow.random", "line_number": 103, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 104, "usage_type": "attribute"}, {"api_name": "tensorflow.random.normal", "line_number": 108, "usage_type": "call"}, {"api_name": "tensorflow.random", "line_number": 108, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 109, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 61, "usage_type": "attribute"}, {"api_name": "ludwig.combiners.combiners.ConcatCombiner", "line_number": 124, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 116, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 116, "usage_type": "attribute"}, {"api_name": "ludwig.combiners.combiners.SequenceConcatCombiner", "line_number": 152, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 147, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 147, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 148, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 148, "usage_type": "attribute"}, {"api_name": "ludwig.combiners.combiners.SequenceCombiner", "line_number": 187, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 181, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 181, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 182, "usage_type": "call"}, {"api_name": "ludwig.combiners.combiners.sequence_encoder_registry", "line_number": 182, "usage_type": "argument"}, {"api_name": "pytest.mark", "line_number": 182, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 183, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 183, "usage_type": "attribute"}, {"api_name": "tensorflow.random.normal", "line_number": 230, "usage_type": "call"}, {"api_name": "tensorflow.random", "line_number": 230, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 232, "usage_type": "attribute"}, {"api_name": "tensorflow.random.normal", "line_number": 236, "usage_type": "call"}, {"api_name": "tensorflow.random", "line_number": 236, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 238, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.Input", "line_number": 244, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 244, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 246, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.Input", "line_number": 251, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 251, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 253, "usage_type": "attribute"}, {"api_name": "ludwig.combiners.combiners.TabNetCombiner", "line_number": 266, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 261, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 261, "usage_type": "attribute"}, {"api_name": "ludwig.combiners.combiners.ComparatorCombiner", "line_number": 297, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 283, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 283, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 285, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 285, "usage_type": "attribute"}, {"api_name": "pytest.mark.parametrize", "line_number": 286, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 286, "usage_type": "attribute"}, {"api_name": "tensorflow.random.normal", "line_number": 320, "usage_type": "call"}, {"api_name": "tensorflow.random", "line_number": 320, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 322, "usage_type": "attribute"}, {"api_name": "tensorflow.random.normal", "line_number": 326, "usage_type": "call"}, {"api_name": "tensorflow.random", "line_number": 326, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 328, "usage_type": "attribute"}, {"api_name": "ludwig.combiners.combiners.TransformerCombiner", "line_number": 338, "usage_type": "call"}, {"api_name": "tensorflow.random.normal", "line_number": 353, "usage_type": "call"}, {"api_name": "tensorflow.random", "line_number": 353, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 355, "usage_type": "attribute"}, {"api_name": "tensorflow.random.normal", "line_number": 359, "usage_type": "call"}, {"api_name": "tensorflow.random", "line_number": 359, "usage_type": "attribute"}, {"api_name": "tensorflow.float32", "line_number": 361, "usage_type": "attribute"}, {"api_name": "ludwig.combiners.combiners.TabTransformerCombiner", "line_number": 371, "usage_type": "call"}, {"api_name": "ludwig.combiners.combiners.TabTransformerCombiner", "line_number": 382, "usage_type": "call"}, {"api_name": "ludwig.combiners.combiners.TabTransformerCombiner", "line_number": 394, "usage_type": "call"}]} +{"seq_id": "123126141", "text": "from django.conf.urls import include, url\nfrom django.contrib import admin\n\n\nurlpatterns = [\n # uvodne kecy, kontakt, demo, faq,...\n url(r'^', include('about.urls', namespace=\"about\")),\n # zadania, riesenia uloh; sady\n url(r'^tasks/', include('tasks.urls', namespace=\"tasks\")),\n # submity, statistiky\n url(r'^submits/', include('submit.urls', namespace=\"submit\")),\n # django admin\n url(r'^admin/', include(admin.site.urls)),\n # login\n url(r'^account/', include('ksp_login.urls')),\n]\n", "sub_path": "page/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 513, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 13, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 13, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 13, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 15, "usage_type": "call"}, {"api_name": "django.conf.urls.include", "line_number": 15, "usage_type": "call"}]} +{"seq_id": "180261169", "text": "import sys\nfrom collections import defaultdict\n\n\nclass MockRandomName:\n def __init__(self):\n self.generator = defaultdict(lambda: {}, {\n 'a': {\n 'get': lambda: 'codeA',\n 'filename': 'a.a',\n 'language_name': 'A'\n },\n 'b': {\n 'get': lambda: 'codeB',\n 'filename': 'b.b',\n 'language_name': 'B'\n }\n })\n\n\nsys.modules['app.helpers.code_generator'] = MockRandomName()\n", "sub_path": "tests/mocks/mock_code_generator.py", "file_name": "mock_code_generator.py", "file_ext": "py", "file_size_in_byte": 511, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "collections.defaultdict", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.modules", "line_number": 21, "usage_type": "attribute"}]} +{"seq_id": "300497462", "text": "#Code Used From:\n#https://www.101computing.net/creating-sprites-using-pygame/\n\nimport pygame\nfrom pygame.math import Vector2\nimport math\n\nWHITE = (255, 255, 255)\nYELLOW = (255, 255, 63)\n\nclass Boat(pygame.sprite.Sprite):\n #This class represents a car. It derives from the \"Sprite\" class in Pygame.\n\n def __init__(self, pos):\n # Call the parent class (Sprite) constructor\n super().__init__()\n\n # Pass in the color of the car, and its x and y position, width and height.\n # Set the background color and set it to be transparent\n\n self.image = pygame.Surface((80,202), pygame.SRCALPHA)\n # self.image.fill(WHITE)\n # self.image.set_colorkey(WHITE)\n\n picture = pygame.image.load(\"images/yellow_boat.png\")\n boat_image = pygame.transform.scale(picture, (80, 200))\n self.image = boat_image\n self.rotated_image = self.image\n\n self.rect = self.image.get_rect(center=(pos))\n self.pos = Vector2(pos)\n self.offset = Vector2(0, 0)\n self.angle = 0\n\n #pygame.draw.rect(picture, WHITE, 0, width = 0)\n\n\n def rotate(self):\n self.image = pygame.transform.rotozoom(self.rotated_image, -self.angle, 1)\n offset_rotated = self.offset.rotate(self.angle)\n self.rect = self.image.get_rect(center=self.pos+offset_rotated)\n\n def rotate_right(self):\n self.angle += 1\n self.rotate()\n\n def rotate_left(self):\n self.angle -= 1\n self.rotate()\n\n def moveRight(self, pixels):\n self.rect.x += pixels\n self.pos.x += pixels\n\n def moveLeft(self, pixels):\n self.rect.x -= pixels\n self.pos.x -= pixels\n\n def moveUp(self, pixels):\n #self.rect.y -= pixels\n self.pos.y -= math.cos(math.radians(self.angle))*pixels\n self.pos.x += math.sin(math.radians(self.angle))*pixels\n \n print(self.pos.x, self.pos.y)\n self.rect.y -= math.cos(math.radians(self.angle))*pixels\n self.rect.x += math.sin(math.radians(self.angle))*pixels\n self.rect = self.image.get_rect(center=self.pos)\n \n def moveDown(self, pixels):\n self.rect.y += pixels\n self.pos.y += pixels\n #def distance(self):\n ", "sub_path": "Final_Blind_Sailing/Previous_Work/boat1.py", "file_name": "boat1.py", "file_ext": "py", "file_size_in_byte": 2226, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "pygame.sprite", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pygame.Surface", "line_number": 21, "usage_type": "call"}, {"api_name": "pygame.SRCALPHA", "line_number": 21, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 25, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 25, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 26, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pygame.math.Vector2", "line_number": 31, "usage_type": "call"}, {"api_name": "pygame.math.Vector2", "line_number": 32, "usage_type": "call"}, {"api_name": "pygame.transform.rotozoom", "line_number": 39, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 39, "usage_type": "attribute"}, {"api_name": "math.cos", "line_number": 61, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 61, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 62, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 62, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 65, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 65, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 66, "usage_type": "call"}, {"api_name": "math.radians", "line_number": 66, "usage_type": "call"}]} +{"seq_id": "272244445", "text": "import sqlite3\n\nimport click\nfrom flask import g\nfrom flask import current_app\nfrom flask.cli import with_appcontext\n\ndef get_db():\n if \"db\" not in g:\n g.db = sqlite3.connect(current_app.config['DATABASE'])\n\n return g.db\n\ndef close_db(_):\n db = g.get(\"pop\", None)\n if db:\n db.close()\n\ndef init_db():\n db = get_db()\n with current_app.open_resource(\"schema.sql\") as f:\n db.executescript(f.read().decode(\"utf8\"))\n\nclass Poll(object):\n @classmethod\n def get_poll_by_id(cls, id):\n db = get_db()\n raw_poll = get_db().execute(\"SELECT id, name, date FROM polls WHERE id = ?\", [id]).fetchone()\n\n if raw_poll:\n return cls(\n id=raw_poll[0],\n name=raw_poll[1],\n date=raw_poll[2],\n )\n\n return None\n\n @classmethod\n def get_polls(cls):\n db = get_db()\n raw_polls = get_db().execute(\"SELECT id, name, date FROM polls ORDER BY id ASC\").fetchall()\n polls = []\n for raw_poll in raw_polls:\n polls.append(cls(\n id=raw_poll[0],\n name=raw_poll[1],\n date=raw_poll[2]\n ))\n\n return polls\n\n def __init__(self, id, name, date):\n self.id = id\n self.name = name\n self.date = date\n\n def save(self):\n db = get_db()\n if self.id == None:\n db.execute(\"INSERT INTO polls VALUES (null, ?, ?)\", [self.name, self.date])\n self.id = db.execute(\"SELECT last_insert_rowid()\").fetchone()[0]\n else:\n db.execute(\"UPDATE polls SET name = ?, date = ? WHERE id = ?\", [self.name, self.date, self.id])\n\n db.commit()\n\n def number_of_votes(self):\n return get_db().execute(\"SELECT count(*) FROM vote_casts WHERE poll_id = ?\", [self.id]).fetchone()[0]\n\n def __str__(self):\n return self.name\n\n def __repr__(self):\n return \" {name}\".format(id=self.id, name=self.name)\n\n\nclass Choice(object):\n @classmethod\n def get_choices_for_poll(cls, poll):\n db = get_db()\n raw_choices = get_db().execute(\"SELECT id, choice FROM choices WHERE poll_id = ?\", [poll.id]).fetchall()\n\n choices = []\n for raw_choice in raw_choices:\n choices.append(cls(\n id=raw_choice[0],\n choice=raw_choice[1],\n poll=poll,\n ))\n\n return choices\n\n @classmethod\n def get_by_id_for_poll(cls, poll, choice_id):\n db = get_db()\n raw_poll = get_db().execute(\"SELECT id, choice FROM choices WHERE id = ? and poll_id= ? \", [choice_id, poll.id]).fetchone()\n\n if raw_poll:\n return cls(\n id=raw_poll[0],\n choice=raw_poll[1],\n poll=poll,\n )\n\n return None\n\n def __init__(self, id, choice, poll):\n self.id = id\n self.choice = choice\n self.poll = poll\n\n def save(self):\n db = get_db()\n if self.id == None:\n db.execute(\"INSERT INTO choices VALUES (null, ?, ?)\", [self.choice, self.poll.id])\n self.id = db.execute(\"SELECT last_insert_rowid()\").fetchone()[0]\n else:\n db.execute(\"UPDATE choices SET choice = ? WHERE id = ?\", [self.choice, self.id])\n\n db.commit()\n\n def cast_vote(self):\n db = get_db()\n db.execute(\"INSERT INTO vote_casts VALUES (null, ?, ?)\", [self.poll.id, self.id])\n db.commit()\n\n def number_of_votes(self):\n return get_db().execute(\"SELECT count(*) FROM vote_casts WHERE choice_id = ?\", [self.id]).fetchone()[0]\n\n def __str__(self):\n return self.choice\n\n def __repr__(self):\n return \" {choice}\".format(id=self.id, choice=self.choice)\n\n\n\n@click.command(\"init-db\")\n@with_appcontext\ndef init_db_command():\n init_db()\n click.echo(\"Initialized the database.\")\n\n\ndef init_app(app):\n app.teardown_appcontext(close_db)\n app.cli.add_command(init_db_command)\n", "sub_path": "cours/03-sept-18/exemples/station-de-vote/poll/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 4000, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "flask.g", "line_number": 9, "usage_type": "name"}, {"api_name": "flask.g.db", "line_number": 10, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 10, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 10, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 10, "usage_type": "name"}, {"api_name": "flask.g.db", "line_number": 12, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 12, "usage_type": "name"}, {"api_name": "flask.g.get", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.g", "line_number": 15, "usage_type": "name"}, {"api_name": "flask.current_app.open_resource", "line_number": 21, "usage_type": "call"}, {"api_name": "flask.current_app", "line_number": 21, "usage_type": "name"}, {"api_name": "click.echo", "line_number": 143, "usage_type": "call"}, {"api_name": "click.command", "line_number": 139, "usage_type": "call"}, {"api_name": "flask.cli.with_appcontext", "line_number": 140, "usage_type": "name"}]} +{"seq_id": "99750552", "text": "#!/usr/bin/env python\n\nfrom ase.io import read\nfrom ase import Atom\n\nslabO = read('POSCAR')\n# get index of first O atom\nindex_O = next(a.index for a in slabO if a.symbol == 'O')\nO = slabO[index_O]\npos_H = O.position + (0, 0, 0.987)\nH = Atom('H', pos_H)\nslabOH = slabO + H\nslabOH.write('POSCAR', format='vasp', vasp5=True, sort=True, direct=True)\n\n", "sub_path": "bin/add_H_on_O.py", "file_name": "add_H_on_O.py", "file_ext": "py", "file_size_in_byte": 347, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "ase.io.read", "line_number": 6, "usage_type": "call"}, {"api_name": "ase.Atom", "line_number": 11, "usage_type": "call"}]} +{"seq_id": "483109238", "text": "from django.http import HttpResponseRedirect\nfrom django.utils.deprecation import MiddlewareMixin\nfrom users.models import UserModel\n\n\nclass AuthMiddleware(MiddlewareMixin):\n def process_request(self, request):\n ticket = request.COOKIES.get('ticket')\n if request.path == '/users/cart/':\n ticket = request.COOKIES.get('ticket')\n if not ticket:\n return HttpResponseRedirect('/users/login/')\n users = UserModel.objects.filter(t_ticket=ticket)\n if users:\n request.user = users[0]\n else:\n return HttpResponseRedirect('/users/login/')\n else:\n if not ticket:\n return None\n users = UserModel.objects.filter(t_ticket=ticket)\n if users:\n request.user = users[0]", "sub_path": "aixianfeng3.0/utils/UserAuthMiddleware.py", "file_name": "UserAuthMiddleware.py", "file_ext": "py", "file_size_in_byte": 838, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "django.utils.deprecation.MiddlewareMixin", "line_number": 6, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 12, "usage_type": "call"}, {"api_name": "users.models", "line_number": 13, "usage_type": "name"}, {"api_name": "users.models.UserModel.objects.filter", "line_number": 13, "usage_type": "call"}, {"api_name": "users.models.UserModel.objects", "line_number": 13, "usage_type": "attribute"}, {"api_name": "users.models.UserModel", "line_number": 13, "usage_type": "name"}, {"api_name": "users.models", "line_number": 14, "usage_type": "name"}, {"api_name": "users.models", "line_number": 15, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 17, "usage_type": "call"}, {"api_name": "users.models", "line_number": 21, "usage_type": "name"}, {"api_name": "users.models.UserModel.objects.filter", "line_number": 21, "usage_type": "call"}, {"api_name": "users.models.UserModel.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "users.models.UserModel", "line_number": 21, "usage_type": "name"}, {"api_name": "users.models", "line_number": 22, "usage_type": "name"}, {"api_name": "users.models", "line_number": 23, "usage_type": "name"}]} +{"seq_id": "120021978", "text": "# Copyright 2017 Google LLC All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport collections\nimport unittest\n\nimport mock\n\n\nclass TestDocumentReference(unittest.TestCase):\n @staticmethod\n def _get_target_class():\n from google.cloud.firestore_v1.document import DocumentReference\n\n return DocumentReference\n\n def _make_one(self, *args, **kwargs):\n klass = self._get_target_class()\n return klass(*args, **kwargs)\n\n def test_constructor(self):\n collection_id1 = \"users\"\n document_id1 = \"alovelace\"\n collection_id2 = \"platform\"\n document_id2 = \"*nix\"\n client = mock.MagicMock()\n client.__hash__.return_value = 1234\n\n document = self._make_one(\n collection_id1, document_id1, collection_id2, document_id2, client=client\n )\n self.assertIs(document._client, client)\n expected_path = \"/\".join(\n (collection_id1, document_id1, collection_id2, document_id2)\n )\n self.assertEqual(document.path, expected_path)\n\n def test_constructor_invalid_path(self):\n with self.assertRaises(ValueError):\n self._make_one()\n with self.assertRaises(ValueError):\n self._make_one(None, \"before\", \"bad-collection-id\", \"fifteen\")\n with self.assertRaises(ValueError):\n self._make_one(\"bad-document-ID\", None)\n with self.assertRaises(ValueError):\n self._make_one(\"Just\", \"A-Collection\", \"Sub\")\n\n def test_constructor_invalid_kwarg(self):\n with self.assertRaises(TypeError):\n self._make_one(\"Coh-lek-shun\", \"Dahk-yu-mehnt\", burger=18.75)\n\n def test___copy__(self):\n client = _make_client(\"rain\")\n document = self._make_one(\"a\", \"b\", client=client)\n # Access the document path so it is copied.\n doc_path = document._document_path\n self.assertEqual(doc_path, document._document_path_internal)\n\n new_document = document.__copy__()\n self.assertIsNot(new_document, document)\n self.assertIs(new_document._client, document._client)\n self.assertEqual(new_document._path, document._path)\n self.assertEqual(\n new_document._document_path_internal, document._document_path_internal\n )\n\n def test___deepcopy__calls_copy(self):\n client = mock.sentinel.client\n document = self._make_one(\"a\", \"b\", client=client)\n document.__copy__ = mock.Mock(return_value=mock.sentinel.new_doc, spec=[])\n\n unused_memo = {}\n new_document = document.__deepcopy__(unused_memo)\n self.assertIs(new_document, mock.sentinel.new_doc)\n document.__copy__.assert_called_once_with()\n\n def test__eq__same_type(self):\n document1 = self._make_one(\"X\", \"YY\", client=mock.sentinel.client)\n document2 = self._make_one(\"X\", \"ZZ\", client=mock.sentinel.client)\n document3 = self._make_one(\"X\", \"YY\", client=mock.sentinel.client2)\n document4 = self._make_one(\"X\", \"YY\", client=mock.sentinel.client)\n\n pairs = ((document1, document2), (document1, document3), (document2, document3))\n for candidate1, candidate2 in pairs:\n # We use == explicitly since assertNotEqual would use !=.\n equality_val = candidate1 == candidate2\n self.assertFalse(equality_val)\n\n # Check the only equal one.\n self.assertEqual(document1, document4)\n self.assertIsNot(document1, document4)\n\n def test__eq__other_type(self):\n document = self._make_one(\"X\", \"YY\", client=mock.sentinel.client)\n other = object()\n equality_val = document == other\n self.assertFalse(equality_val)\n self.assertIs(document.__eq__(other), NotImplemented)\n\n def test___hash__(self):\n client = mock.MagicMock()\n client.__hash__.return_value = 234566789\n document = self._make_one(\"X\", \"YY\", client=client)\n self.assertEqual(hash(document), hash((\"X\", \"YY\")) + hash(client))\n\n def test__ne__same_type(self):\n document1 = self._make_one(\"X\", \"YY\", client=mock.sentinel.client)\n document2 = self._make_one(\"X\", \"ZZ\", client=mock.sentinel.client)\n document3 = self._make_one(\"X\", \"YY\", client=mock.sentinel.client2)\n document4 = self._make_one(\"X\", \"YY\", client=mock.sentinel.client)\n\n self.assertNotEqual(document1, document2)\n self.assertNotEqual(document1, document3)\n self.assertNotEqual(document2, document3)\n\n # We use != explicitly since assertEqual would use ==.\n inequality_val = document1 != document4\n self.assertFalse(inequality_val)\n self.assertIsNot(document1, document4)\n\n def test__ne__other_type(self):\n document = self._make_one(\"X\", \"YY\", client=mock.sentinel.client)\n other = object()\n self.assertNotEqual(document, other)\n self.assertIs(document.__ne__(other), NotImplemented)\n\n def test__document_path_property(self):\n project = \"hi-its-me-ok-bye\"\n client = _make_client(project=project)\n\n collection_id = \"then\"\n document_id = \"090909iii\"\n document = self._make_one(collection_id, document_id, client=client)\n doc_path = document._document_path\n expected = \"projects/{}/databases/{}/documents/{}/{}\".format(\n project, client._database, collection_id, document_id\n )\n self.assertEqual(doc_path, expected)\n self.assertIs(document._document_path_internal, doc_path)\n\n # Make sure value is cached.\n document._document_path_internal = mock.sentinel.cached\n self.assertIs(document._document_path, mock.sentinel.cached)\n\n def test__document_path_property_no_client(self):\n document = self._make_one(\"hi\", \"bye\")\n self.assertIsNone(document._client)\n with self.assertRaises(ValueError):\n getattr(document, \"_document_path\")\n\n self.assertIsNone(document._document_path_internal)\n\n def test_id_property(self):\n document_id = \"867-5309\"\n document = self._make_one(\"Co-lek-shun\", document_id)\n self.assertEqual(document.id, document_id)\n\n def test_parent_property(self):\n from google.cloud.firestore_v1.collection import CollectionReference\n\n collection_id = \"grocery-store\"\n document_id = \"market\"\n client = _make_client()\n document = self._make_one(collection_id, document_id, client=client)\n\n parent = document.parent\n self.assertIsInstance(parent, CollectionReference)\n self.assertIs(parent._client, client)\n self.assertEqual(parent._path, (collection_id,))\n\n def test_collection_factory(self):\n from google.cloud.firestore_v1.collection import CollectionReference\n\n collection_id = \"grocery-store\"\n document_id = \"market\"\n new_collection = \"fruits\"\n client = _make_client()\n document = self._make_one(collection_id, document_id, client=client)\n\n child = document.collection(new_collection)\n self.assertIsInstance(child, CollectionReference)\n self.assertIs(child._client, client)\n self.assertEqual(child._path, (collection_id, document_id, new_collection))\n\n @staticmethod\n def _write_pb_for_create(document_path, document_data):\n from google.cloud.firestore_v1.proto import common_pb2\n from google.cloud.firestore_v1.proto import document_pb2\n from google.cloud.firestore_v1.proto import write_pb2\n from google.cloud.firestore_v1 import _helpers\n\n return write_pb2.Write(\n update=document_pb2.Document(\n name=document_path, fields=_helpers.encode_dict(document_data)\n ),\n current_document=common_pb2.Precondition(exists=False),\n )\n\n @staticmethod\n def _make_commit_repsonse(write_results=None):\n from google.cloud.firestore_v1.proto import firestore_pb2\n\n response = mock.create_autospec(firestore_pb2.CommitResponse)\n response.write_results = write_results or [mock.sentinel.write_result]\n response.commit_time = mock.sentinel.commit_time\n return response\n\n def test_create(self):\n # Create a minimal fake GAPIC with a dummy response.\n firestore_api = mock.Mock(spec=[\"commit\"])\n firestore_api.commit.return_value = self._make_commit_repsonse()\n\n # Attach the fake GAPIC to a real client.\n client = _make_client(\"dignity\")\n client._firestore_api_internal = firestore_api\n\n # Actually make a document and call create().\n document = self._make_one(\"foo\", \"twelve\", client=client)\n document_data = {\"hello\": \"goodbye\", \"count\": 99}\n write_result = document.create(document_data)\n\n # Verify the response and the mocks.\n self.assertIs(write_result, mock.sentinel.write_result)\n write_pb = self._write_pb_for_create(document._document_path, document_data)\n firestore_api.commit.assert_called_once_with(\n client._database_string,\n [write_pb],\n transaction=None,\n metadata=client._rpc_metadata,\n )\n\n def test_create_empty(self):\n # Create a minimal fake GAPIC with a dummy response.\n from google.cloud.firestore_v1.document import DocumentReference\n from google.cloud.firestore_v1.document import DocumentSnapshot\n\n firestore_api = mock.Mock(spec=[\"commit\"])\n document_reference = mock.create_autospec(DocumentReference)\n snapshot = mock.create_autospec(DocumentSnapshot)\n snapshot.exists = True\n document_reference.get.return_value = snapshot\n firestore_api.commit.return_value = self._make_commit_repsonse(\n write_results=[document_reference]\n )\n\n # Attach the fake GAPIC to a real client.\n client = _make_client(\"dignity\")\n client._firestore_api_internal = firestore_api\n client.get_all = mock.MagicMock()\n client.get_all.exists.return_value = True\n\n # Actually make a document and call create().\n document = self._make_one(\"foo\", \"twelve\", client=client)\n document_data = {}\n write_result = document.create(document_data)\n self.assertTrue(write_result.get().exists)\n\n @staticmethod\n def _write_pb_for_set(document_path, document_data, merge):\n from google.cloud.firestore_v1.proto import common_pb2\n from google.cloud.firestore_v1.proto import document_pb2\n from google.cloud.firestore_v1.proto import write_pb2\n from google.cloud.firestore_v1 import _helpers\n\n write_pbs = write_pb2.Write(\n update=document_pb2.Document(\n name=document_path, fields=_helpers.encode_dict(document_data)\n )\n )\n if merge:\n field_paths = [\n field_path\n for field_path, value in _helpers.extract_fields(\n document_data, _helpers.FieldPath()\n )\n ]\n field_paths = [\n field_path.to_api_repr() for field_path in sorted(field_paths)\n ]\n mask = common_pb2.DocumentMask(field_paths=sorted(field_paths))\n write_pbs.update_mask.CopyFrom(mask)\n return write_pbs\n\n def _set_helper(self, merge=False, **option_kwargs):\n # Create a minimal fake GAPIC with a dummy response.\n firestore_api = mock.Mock(spec=[\"commit\"])\n firestore_api.commit.return_value = self._make_commit_repsonse()\n\n # Attach the fake GAPIC to a real client.\n client = _make_client(\"db-dee-bee\")\n client._firestore_api_internal = firestore_api\n\n # Actually make a document and call create().\n document = self._make_one(\"User\", \"Interface\", client=client)\n document_data = {\"And\": 500, \"Now\": b\"\\xba\\xaa\\xaa \\xba\\xaa\\xaa\"}\n write_result = document.set(document_data, merge)\n\n # Verify the response and the mocks.\n self.assertIs(write_result, mock.sentinel.write_result)\n write_pb = self._write_pb_for_set(document._document_path, document_data, merge)\n\n firestore_api.commit.assert_called_once_with(\n client._database_string,\n [write_pb],\n transaction=None,\n metadata=client._rpc_metadata,\n )\n\n def test_set(self):\n self._set_helper()\n\n def test_set_merge(self):\n self._set_helper(merge=True)\n\n @staticmethod\n def _write_pb_for_update(document_path, update_values, field_paths):\n from google.cloud.firestore_v1.proto import common_pb2\n from google.cloud.firestore_v1.proto import document_pb2\n from google.cloud.firestore_v1.proto import write_pb2\n from google.cloud.firestore_v1 import _helpers\n\n return write_pb2.Write(\n update=document_pb2.Document(\n name=document_path, fields=_helpers.encode_dict(update_values)\n ),\n update_mask=common_pb2.DocumentMask(field_paths=field_paths),\n current_document=common_pb2.Precondition(exists=True),\n )\n\n def _update_helper(self, **option_kwargs):\n from google.cloud.firestore_v1.transforms import DELETE_FIELD\n\n # Create a minimal fake GAPIC with a dummy response.\n firestore_api = mock.Mock(spec=[\"commit\"])\n firestore_api.commit.return_value = self._make_commit_repsonse()\n\n # Attach the fake GAPIC to a real client.\n client = _make_client(\"potato-chip\")\n client._firestore_api_internal = firestore_api\n\n # Actually make a document and call create().\n document = self._make_one(\"baked\", \"Alaska\", client=client)\n # \"Cheat\" and use OrderedDict-s so that iteritems() is deterministic.\n field_updates = collections.OrderedDict(\n ((\"hello\", 1), (\"then.do\", False), (\"goodbye\", DELETE_FIELD))\n )\n if option_kwargs:\n option = client.write_option(**option_kwargs)\n write_result = document.update(field_updates, option=option)\n else:\n option = None\n write_result = document.update(field_updates)\n\n # Verify the response and the mocks.\n self.assertIs(write_result, mock.sentinel.write_result)\n update_values = {\n \"hello\": field_updates[\"hello\"],\n \"then\": {\"do\": field_updates[\"then.do\"]},\n }\n field_paths = list(field_updates.keys())\n write_pb = self._write_pb_for_update(\n document._document_path, update_values, sorted(field_paths)\n )\n if option is not None:\n option.modify_write(write_pb)\n firestore_api.commit.assert_called_once_with(\n client._database_string,\n [write_pb],\n transaction=None,\n metadata=client._rpc_metadata,\n )\n\n def test_update_with_exists(self):\n with self.assertRaises(ValueError):\n self._update_helper(exists=True)\n\n def test_update(self):\n self._update_helper()\n\n def test_update_with_precondition(self):\n from google.protobuf import timestamp_pb2\n\n timestamp = timestamp_pb2.Timestamp(seconds=1058655101, nanos=100022244)\n self._update_helper(last_update_time=timestamp)\n\n def test_empty_update(self):\n # Create a minimal fake GAPIC with a dummy response.\n firestore_api = mock.Mock(spec=[\"commit\"])\n firestore_api.commit.return_value = self._make_commit_repsonse()\n\n # Attach the fake GAPIC to a real client.\n client = _make_client(\"potato-chip\")\n client._firestore_api_internal = firestore_api\n\n # Actually make a document and call create().\n document = self._make_one(\"baked\", \"Alaska\", client=client)\n # \"Cheat\" and use OrderedDict-s so that iteritems() is deterministic.\n field_updates = {}\n with self.assertRaises(ValueError):\n document.update(field_updates)\n\n def _delete_helper(self, **option_kwargs):\n from google.cloud.firestore_v1.proto import write_pb2\n\n # Create a minimal fake GAPIC with a dummy response.\n firestore_api = mock.Mock(spec=[\"commit\"])\n firestore_api.commit.return_value = self._make_commit_repsonse()\n\n # Attach the fake GAPIC to a real client.\n client = _make_client(\"donut-base\")\n client._firestore_api_internal = firestore_api\n\n # Actually make a document and call delete().\n document = self._make_one(\"where\", \"we-are\", client=client)\n if option_kwargs:\n option = client.write_option(**option_kwargs)\n delete_time = document.delete(option=option)\n else:\n option = None\n delete_time = document.delete()\n\n # Verify the response and the mocks.\n self.assertIs(delete_time, mock.sentinel.commit_time)\n write_pb = write_pb2.Write(delete=document._document_path)\n if option is not None:\n option.modify_write(write_pb)\n firestore_api.commit.assert_called_once_with(\n client._database_string,\n [write_pb],\n transaction=None,\n metadata=client._rpc_metadata,\n )\n\n def test_delete(self):\n self._delete_helper()\n\n def test_delete_with_option(self):\n from google.protobuf import timestamp_pb2\n\n timestamp_pb = timestamp_pb2.Timestamp(seconds=1058655101, nanos=100022244)\n self._delete_helper(last_update_time=timestamp_pb)\n\n def _get_helper(self, field_paths=None, use_transaction=False, not_found=False):\n from google.api_core.exceptions import NotFound\n from google.cloud.firestore_v1.proto import common_pb2\n from google.cloud.firestore_v1.proto import document_pb2\n from google.cloud.firestore_v1.transaction import Transaction\n\n # Create a minimal fake GAPIC with a dummy response.\n create_time = 123\n update_time = 234\n firestore_api = mock.Mock(spec=[\"get_document\"])\n response = mock.create_autospec(document_pb2.Document)\n response.fields = {}\n response.create_time = create_time\n response.update_time = update_time\n\n if not_found:\n firestore_api.get_document.side_effect = NotFound(\"testing\")\n else:\n firestore_api.get_document.return_value = response\n\n client = _make_client(\"donut-base\")\n client._firestore_api_internal = firestore_api\n\n document = self._make_one(\"where\", \"we-are\", client=client)\n\n if use_transaction:\n transaction = Transaction(client)\n transaction_id = transaction._id = b\"asking-me-2\"\n else:\n transaction = None\n\n snapshot = document.get(field_paths=field_paths, transaction=transaction)\n\n self.assertIs(snapshot.reference, document)\n if not_found:\n self.assertIsNone(snapshot._data)\n self.assertFalse(snapshot.exists)\n self.assertIsNone(snapshot.read_time)\n self.assertIsNone(snapshot.create_time)\n self.assertIsNone(snapshot.update_time)\n else:\n self.assertEqual(snapshot.to_dict(), {})\n self.assertTrue(snapshot.exists)\n self.assertIsNone(snapshot.read_time)\n self.assertIs(snapshot.create_time, create_time)\n self.assertIs(snapshot.update_time, update_time)\n\n # Verify the request made to the API\n if field_paths is not None:\n mask = common_pb2.DocumentMask(field_paths=sorted(field_paths))\n else:\n mask = None\n\n if use_transaction:\n expected_transaction_id = transaction_id\n else:\n expected_transaction_id = None\n\n firestore_api.get_document.assert_called_once_with(\n document._document_path,\n mask=mask,\n transaction=expected_transaction_id,\n metadata=client._rpc_metadata,\n )\n\n def test_get_not_found(self):\n self._get_helper(not_found=True)\n\n def test_get_default(self):\n self._get_helper()\n\n def test_get_w_string_field_path(self):\n with self.assertRaises(ValueError):\n self._get_helper(field_paths=\"foo\")\n\n def test_get_with_field_path(self):\n self._get_helper(field_paths=[\"foo\"])\n\n def test_get_with_multiple_field_paths(self):\n self._get_helper(field_paths=[\"foo\", \"bar.baz\"])\n\n def test_get_with_transaction(self):\n self._get_helper(use_transaction=True)\n\n def _collections_helper(self, page_size=None):\n from google.api_core.page_iterator import Iterator\n from google.api_core.page_iterator import Page\n from google.cloud.firestore_v1.collection import CollectionReference\n from google.cloud.firestore_v1.gapic.firestore_client import FirestoreClient\n\n class _Iterator(Iterator):\n def __init__(self, pages):\n super(_Iterator, self).__init__(client=None)\n self._pages = pages\n\n def _next_page(self):\n if self._pages:\n page, self._pages = self._pages[0], self._pages[1:]\n return Page(self, page, self.item_to_value)\n\n collection_ids = [\"coll-1\", \"coll-2\"]\n iterator = _Iterator(pages=[collection_ids])\n api_client = mock.create_autospec(FirestoreClient)\n api_client.list_collection_ids.return_value = iterator\n\n client = _make_client()\n client._firestore_api_internal = api_client\n\n # Actually make a document and call delete().\n document = self._make_one(\"where\", \"we-are\", client=client)\n if page_size is not None:\n collections = list(document.collections(page_size=page_size))\n else:\n collections = list(document.collections())\n\n # Verify the response and the mocks.\n self.assertEqual(len(collections), len(collection_ids))\n for collection, collection_id in zip(collections, collection_ids):\n self.assertIsInstance(collection, CollectionReference)\n self.assertEqual(collection.parent, document)\n self.assertEqual(collection.id, collection_id)\n\n api_client.list_collection_ids.assert_called_once_with(\n document._document_path, page_size=page_size, metadata=client._rpc_metadata\n )\n\n def test_collections_wo_page_size(self):\n self._collections_helper()\n\n def test_collections_w_page_size(self):\n self._collections_helper(page_size=10)\n\n @mock.patch(\"google.cloud.firestore_v1.document.Watch\", autospec=True)\n def test_on_snapshot(self, watch):\n client = mock.Mock(_database_string=\"sprinklez\", spec=[\"_database_string\"])\n document = self._make_one(\"yellow\", \"mellow\", client=client)\n document.on_snapshot(None)\n watch.for_document.assert_called_once()\n\n\nclass TestDocumentSnapshot(unittest.TestCase):\n @staticmethod\n def _get_target_class():\n from google.cloud.firestore_v1.document import DocumentSnapshot\n\n return DocumentSnapshot\n\n def _make_one(self, *args, **kwargs):\n klass = self._get_target_class()\n return klass(*args, **kwargs)\n\n def _make_reference(self, *args, **kwargs):\n from google.cloud.firestore_v1.document import DocumentReference\n\n return DocumentReference(*args, **kwargs)\n\n def _make_w_ref(self, ref_path=(\"a\", \"b\"), data={}, exists=True):\n client = mock.sentinel.client\n reference = self._make_reference(*ref_path, client=client)\n return self._make_one(\n reference,\n data,\n exists,\n mock.sentinel.read_time,\n mock.sentinel.create_time,\n mock.sentinel.update_time,\n )\n\n def test_constructor(self):\n client = mock.sentinel.client\n reference = self._make_reference(\"hi\", \"bye\", client=client)\n data = {\"zoop\": 83}\n snapshot = self._make_one(\n reference,\n data,\n True,\n mock.sentinel.read_time,\n mock.sentinel.create_time,\n mock.sentinel.update_time,\n )\n self.assertIs(snapshot._reference, reference)\n self.assertEqual(snapshot._data, data)\n self.assertIsNot(snapshot._data, data) # Make sure copied.\n self.assertTrue(snapshot._exists)\n self.assertIs(snapshot.read_time, mock.sentinel.read_time)\n self.assertIs(snapshot.create_time, mock.sentinel.create_time)\n self.assertIs(snapshot.update_time, mock.sentinel.update_time)\n\n def test___eq___other_type(self):\n snapshot = self._make_w_ref()\n other = object()\n self.assertFalse(snapshot == other)\n\n def test___eq___different_reference_same_data(self):\n snapshot = self._make_w_ref((\"a\", \"b\"))\n other = self._make_w_ref((\"c\", \"d\"))\n self.assertFalse(snapshot == other)\n\n def test___eq___same_reference_different_data(self):\n snapshot = self._make_w_ref((\"a\", \"b\"))\n other = self._make_w_ref((\"a\", \"b\"), {\"foo\": \"bar\"})\n self.assertFalse(snapshot == other)\n\n def test___eq___same_reference_same_data(self):\n snapshot = self._make_w_ref((\"a\", \"b\"), {\"foo\": \"bar\"})\n other = self._make_w_ref((\"a\", \"b\"), {\"foo\": \"bar\"})\n self.assertTrue(snapshot == other)\n\n def test___hash__(self):\n from google.protobuf import timestamp_pb2\n\n client = mock.MagicMock()\n client.__hash__.return_value = 234566789\n reference = self._make_reference(\"hi\", \"bye\", client=client)\n data = {\"zoop\": 83}\n update_time = timestamp_pb2.Timestamp(seconds=123456, nanos=123456789)\n snapshot = self._make_one(\n reference, data, True, None, mock.sentinel.create_time, update_time\n )\n self.assertEqual(\n hash(snapshot), hash(reference) + hash(123456) + hash(123456789)\n )\n\n def test__client_property(self):\n reference = self._make_reference(\n \"ok\", \"fine\", \"now\", \"fore\", client=mock.sentinel.client\n )\n snapshot = self._make_one(reference, {}, False, None, None, None)\n self.assertIs(snapshot._client, mock.sentinel.client)\n\n def test_exists_property(self):\n reference = mock.sentinel.reference\n\n snapshot1 = self._make_one(reference, {}, False, None, None, None)\n self.assertFalse(snapshot1.exists)\n snapshot2 = self._make_one(reference, {}, True, None, None, None)\n self.assertTrue(snapshot2.exists)\n\n def test_id_property(self):\n document_id = \"around\"\n reference = self._make_reference(\n \"look\", document_id, client=mock.sentinel.client\n )\n snapshot = self._make_one(reference, {}, True, None, None, None)\n self.assertEqual(snapshot.id, document_id)\n self.assertEqual(reference.id, document_id)\n\n def test_reference_property(self):\n snapshot = self._make_one(mock.sentinel.reference, {}, True, None, None, None)\n self.assertIs(snapshot.reference, mock.sentinel.reference)\n\n def test_get(self):\n data = {\"one\": {\"bold\": \"move\"}}\n snapshot = self._make_one(None, data, True, None, None, None)\n\n first_read = snapshot.get(\"one\")\n second_read = snapshot.get(\"one\")\n self.assertEqual(first_read, data.get(\"one\"))\n self.assertIsNot(first_read, data.get(\"one\"))\n self.assertEqual(first_read, second_read)\n self.assertIsNot(first_read, second_read)\n\n with self.assertRaises(KeyError):\n snapshot.get(\"two\")\n\n def test_nonexistent_snapshot(self):\n snapshot = self._make_one(None, None, False, None, None, None)\n self.assertIsNone(snapshot.get(\"one\"))\n\n def test_to_dict(self):\n data = {\"a\": 10, \"b\": [\"definitely\", \"mutable\"], \"c\": {\"45\": 50}}\n snapshot = self._make_one(None, data, True, None, None, None)\n as_dict = snapshot.to_dict()\n self.assertEqual(as_dict, data)\n self.assertIsNot(as_dict, data)\n # Check that the data remains unchanged.\n as_dict[\"b\"].append(\"hi\")\n self.assertEqual(data, snapshot.to_dict())\n self.assertNotEqual(data, as_dict)\n\n def test_non_existent(self):\n snapshot = self._make_one(None, None, False, None, None, None)\n as_dict = snapshot.to_dict()\n self.assertIsNone(as_dict)\n\n\nclass Test__get_document_path(unittest.TestCase):\n @staticmethod\n def _call_fut(client, path):\n from google.cloud.firestore_v1.document import _get_document_path\n\n return _get_document_path(client, path)\n\n def test_it(self):\n project = \"prah-jekt\"\n client = _make_client(project=project)\n path = (\"Some\", \"Document\", \"Child\", \"Shockument\")\n document_path = self._call_fut(client, path)\n\n expected = \"projects/{}/databases/{}/documents/{}\".format(\n project, client._database, \"/\".join(path)\n )\n self.assertEqual(document_path, expected)\n\n\nclass Test__consume_single_get(unittest.TestCase):\n @staticmethod\n def _call_fut(response_iterator):\n from google.cloud.firestore_v1.document import _consume_single_get\n\n return _consume_single_get(response_iterator)\n\n def test_success(self):\n response_iterator = iter([mock.sentinel.result])\n result = self._call_fut(response_iterator)\n self.assertIs(result, mock.sentinel.result)\n\n def test_failure_not_enough(self):\n response_iterator = iter([])\n with self.assertRaises(ValueError):\n self._call_fut(response_iterator)\n\n def test_failure_too_many(self):\n response_iterator = iter([None, None])\n with self.assertRaises(ValueError):\n self._call_fut(response_iterator)\n\n\nclass Test__first_write_result(unittest.TestCase):\n @staticmethod\n def _call_fut(write_results):\n from google.cloud.firestore_v1.document import _first_write_result\n\n return _first_write_result(write_results)\n\n def test_success(self):\n from google.protobuf import timestamp_pb2\n from google.cloud.firestore_v1.proto import write_pb2\n\n single_result = write_pb2.WriteResult(\n update_time=timestamp_pb2.Timestamp(seconds=1368767504, nanos=458000123)\n )\n write_results = [single_result]\n result = self._call_fut(write_results)\n self.assertIs(result, single_result)\n\n def test_failure_not_enough(self):\n write_results = []\n with self.assertRaises(ValueError):\n self._call_fut(write_results)\n\n def test_more_than_one(self):\n from google.cloud.firestore_v1.proto import write_pb2\n\n result1 = write_pb2.WriteResult()\n result2 = write_pb2.WriteResult()\n write_results = [result1, result2]\n result = self._call_fut(write_results)\n self.assertIs(result, result1)\n\n\ndef _make_credentials():\n import google.auth.credentials\n\n return mock.Mock(spec=google.auth.credentials.Credentials)\n\n\ndef _make_client(project=\"project-project\"):\n from google.cloud.firestore_v1.client import Client\n\n credentials = _make_credentials()\n return Client(project=project, credentials=credentials)\n", "sub_path": "firestore/tests/unit/v1/test_document.py", "file_name": "test_document.py", "file_ext": "py", "file_size_in_byte": 31704, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "unittest.TestCase", "line_number": 21, "usage_type": "attribute"}, {"api_name": "google.cloud.firestore_v1.document.DocumentReference", "line_number": 26, "usage_type": "name"}, {"api_name": "mock.MagicMock", "line_number": 37, "usage_type": "call"}, {"api_name": "mock.sentinel", "line_number": 79, "usage_type": "attribute"}, {"api_name": "mock.Mock", "line_number": 81, "usage_type": "call"}, {"api_name": "mock.sentinel", "line_number": 81, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 85, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 89, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 90, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 91, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 92, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 105, "usage_type": "attribute"}, {"api_name": "mock.MagicMock", "line_number": 112, "usage_type": "call"}, {"api_name": "mock.sentinel", "line_number": 118, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 119, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 120, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 121, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 133, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 153, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 154, "usage_type": "attribute"}, {"api_name": "google.cloud.firestore_v1.collection.CollectionReference", "line_number": 178, "usage_type": "name"}, {"api_name": "google.cloud.firestore_v1.collection.CollectionReference", "line_number": 192, "usage_type": "name"}, {"api_name": "google.cloud.firestore_v1.proto.write_pb2.Write", "line_number": 203, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.write_pb2", "line_number": 203, "usage_type": "name"}, {"api_name": "google.cloud.firestore_v1.proto.document_pb2.Document", "line_number": 204, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.document_pb2", "line_number": 204, "usage_type": "name"}, {"api_name": "google.cloud.firestore_v1._helpers.encode_dict", "line_number": 205, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1._helpers", "line_number": 205, "usage_type": "name"}, {"api_name": "google.cloud.firestore_v1.proto.common_pb2.Precondition", "line_number": 207, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.common_pb2", "line_number": 207, "usage_type": "name"}, {"api_name": "mock.create_autospec", "line_number": 214, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.firestore_pb2.CommitResponse", "line_number": 214, "usage_type": "attribute"}, {"api_name": "google.cloud.firestore_v1.proto.firestore_pb2", "line_number": 214, "usage_type": "name"}, {"api_name": "mock.sentinel", "line_number": 215, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 216, "usage_type": "attribute"}, {"api_name": "mock.Mock", "line_number": 221, "usage_type": "call"}, {"api_name": "mock.sentinel", "line_number": 234, "usage_type": "attribute"}, {"api_name": "mock.Mock", "line_number": 248, "usage_type": "call"}, {"api_name": "mock.create_autospec", "line_number": 249, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.document.DocumentReference", "line_number": 249, "usage_type": "name"}, {"api_name": "mock.create_autospec", "line_number": 250, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.document.DocumentSnapshot", "line_number": 250, "usage_type": "name"}, {"api_name": "mock.MagicMock", "line_number": 260, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.write_pb2.Write", "line_number": 276, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.write_pb2", "line_number": 276, "usage_type": "name"}, {"api_name": "google.cloud.firestore_v1.proto.document_pb2.Document", "line_number": 277, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.document_pb2", "line_number": 277, "usage_type": "name"}, {"api_name": "google.cloud.firestore_v1._helpers.encode_dict", "line_number": 278, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1._helpers", "line_number": 278, "usage_type": "name"}, {"api_name": "google.cloud.firestore_v1._helpers.extract_fields", "line_number": 284, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1._helpers", "line_number": 284, "usage_type": "name"}, {"api_name": "google.cloud.firestore_v1._helpers.FieldPath", "line_number": 285, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1._helpers", "line_number": 285, "usage_type": "name"}, {"api_name": "google.cloud.firestore_v1.proto.common_pb2.DocumentMask", "line_number": 291, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.common_pb2", "line_number": 291, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 297, "usage_type": "call"}, {"api_name": "mock.sentinel", "line_number": 310, "usage_type": "attribute"}, {"api_name": "google.cloud.firestore_v1.proto.write_pb2.Write", "line_number": 333, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.write_pb2", "line_number": 333, "usage_type": "name"}, {"api_name": "google.cloud.firestore_v1.proto.document_pb2.Document", "line_number": 334, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.document_pb2", "line_number": 334, "usage_type": "name"}, {"api_name": "google.cloud.firestore_v1._helpers.encode_dict", "line_number": 335, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1._helpers", "line_number": 335, "usage_type": "name"}, {"api_name": "google.cloud.firestore_v1.proto.common_pb2.DocumentMask", "line_number": 337, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.common_pb2", "line_number": 337, "usage_type": "name"}, {"api_name": "google.cloud.firestore_v1.proto.common_pb2.Precondition", "line_number": 338, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.common_pb2", "line_number": 338, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 345, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 355, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.transforms.DELETE_FIELD", "line_number": 356, "usage_type": "name"}, {"api_name": "mock.sentinel", "line_number": 366, "usage_type": "attribute"}, {"api_name": "google.protobuf.timestamp_pb2.Timestamp", "line_number": 394, "usage_type": "call"}, {"api_name": "google.protobuf.timestamp_pb2", "line_number": 394, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 399, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 417, "usage_type": "call"}, {"api_name": "mock.sentinel", "line_number": 434, "usage_type": "attribute"}, {"api_name": "google.cloud.firestore_v1.proto.write_pb2.Write", "line_number": 435, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.write_pb2", "line_number": 435, "usage_type": "name"}, {"api_name": "google.protobuf.timestamp_pb2.Timestamp", "line_number": 451, "usage_type": "call"}, {"api_name": "google.protobuf.timestamp_pb2", "line_number": 451, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 463, "usage_type": "call"}, {"api_name": "mock.create_autospec", "line_number": 464, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.document_pb2.Document", "line_number": 464, "usage_type": "attribute"}, {"api_name": "google.cloud.firestore_v1.proto.document_pb2", "line_number": 464, "usage_type": "name"}, {"api_name": "google.api_core.exceptions.NotFound", "line_number": 470, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.transaction.Transaction", "line_number": 480, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.common_pb2.DocumentMask", "line_number": 503, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.common_pb2", "line_number": 503, "usage_type": "name"}, {"api_name": "mock.create_autospec", "line_number": 556, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.gapic.firestore_client.FirestoreClient", "line_number": 556, "usage_type": "name"}, {"api_name": "google.cloud.firestore_v1.collection.CollectionReference", "line_number": 572, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 588, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 586, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 594, "usage_type": "attribute"}, {"api_name": "google.cloud.firestore_v1.document.DocumentSnapshot", "line_number": 599, "usage_type": "name"}, {"api_name": "google.cloud.firestore_v1.document.DocumentReference", "line_number": 608, "usage_type": "call"}, {"api_name": "mock.sentinel", "line_number": 611, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 617, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 618, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 619, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 623, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 630, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 631, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 632, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 638, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 639, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 640, "usage_type": "attribute"}, {"api_name": "mock.MagicMock", "line_number": 665, "usage_type": "call"}, {"api_name": "google.protobuf.timestamp_pb2.Timestamp", "line_number": 669, "usage_type": "call"}, {"api_name": "google.protobuf.timestamp_pb2", "line_number": 669, "usage_type": "name"}, {"api_name": "mock.sentinel", "line_number": 671, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 679, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 682, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 685, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 695, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 702, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 703, "usage_type": "attribute"}, {"api_name": "unittest.TestCase", "line_number": 740, "usage_type": "attribute"}, {"api_name": "google.cloud.firestore_v1.document._get_document_path", "line_number": 745, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 759, "usage_type": "attribute"}, {"api_name": "google.cloud.firestore_v1.document._consume_single_get", "line_number": 764, "usage_type": "call"}, {"api_name": "mock.sentinel", "line_number": 767, "usage_type": "attribute"}, {"api_name": "mock.sentinel", "line_number": 769, "usage_type": "attribute"}, {"api_name": "unittest.TestCase", "line_number": 782, "usage_type": "attribute"}, {"api_name": "google.cloud.firestore_v1.document._first_write_result", "line_number": 787, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.write_pb2.WriteResult", "line_number": 793, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.write_pb2", "line_number": 793, "usage_type": "name"}, {"api_name": "google.protobuf.timestamp_pb2.Timestamp", "line_number": 794, "usage_type": "call"}, {"api_name": "google.protobuf.timestamp_pb2", "line_number": 794, "usage_type": "name"}, {"api_name": "google.cloud.firestore_v1.proto.write_pb2.WriteResult", "line_number": 808, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.write_pb2", "line_number": 808, "usage_type": "name"}, {"api_name": "google.cloud.firestore_v1.proto.write_pb2.WriteResult", "line_number": 809, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.proto.write_pb2", "line_number": 809, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 818, "usage_type": "call"}, {"api_name": "google.cloud.firestore_v1.document.auth", "line_number": 818, "usage_type": "attribute"}, {"api_name": "google.cloud.firestore_v1.document", "line_number": 818, "usage_type": "name"}, {"api_name": "google.cloud.firestore_v1.client.Client", "line_number": 825, "usage_type": "call"}]} +{"seq_id": "398947161", "text": "'''\nThis file is the starting point of this flash server.\n\nApenx:\n @@ means todo\n\n'''\n\nfrom flask import Flask, request, jsonify, after_this_request, render_template\nimport tldextract # to extract the domain name from url\nimport whois\nimport geocoder\nimport pycountry\nfrom geopy.geocoders import Nominatim\nimport requests\nimport json\nimport time\nfrom datetime import datetime\n\nimport controller.alexa as alexa\nimport controller.blazeGraph as blazegraph\nimport controller.dbpedia as dbpedia\nimport controller.privacyMetrics as privacyMetrics\n\n# from OpenSSL import SSL\n# context = SSL.Context(SSL.PROTOCOL_TLSv1_2)\n# context.use_privatekey_file('server.key')\n# context.use_certificate_file('server.crt')\n\napp = Flask(__name__)\n#app.run('127.0.0.1', debug=True, port=5000, ssl_context=context)\n\n@app.route('/')\ndef index():\n blazegraph.baseRules()\n return \"privary matters :) \\n The basic rules are set in your RDF Triple store.\"\n\n@app.route('/privacyMetric', methods=['GET', 'POST'])\ndef privacyMetric():\n \n @after_this_request\n def add_header(response):\n response.headers.add('Access-Control-Allow-Origin', '*')\n return response\n\n # variable initialise\n userInfo = {}\n\n # get domain name from url ======= \n url = request.form.get(\"url\")\n protocol = url.split(':')[0]\n urlInfo = tldextract.extract(url)\n domain = urlInfo.domain +'.' + urlInfo.suffix\n\n if protocol == \"chrome-extension\":\n return jsonify({'privacyScore': 0, 'reasonForPrivacyScore': \"This webpage is completely safe.\", \"websiteType\":\"privacyProtection\"})\n\n print(\"protocol: \", protocol)\n\n # get data from request\n userInfo['domainVisitCount'] = int(request.form.get(\"domainVisitCount\"))\n\n # get user profile\n userProfile = request.form.get(\"userProfile\")\n userInfo['userProfile'] = json.loads(userProfile)\n print(\"UserProfile: \", request.form.get(\"userProfile\"))\n\n # initialising privacyScore Variable\n privacyScore = 0\n\n # flags\n calledWhois = False\n\n if domain not in ['localhost.']:\n # get user location from userLocation ======\n userLocationLat = request.form.get(\"userLocationLat\")\n userLocationLong = request.form.get(\"userLocationLong\")\n print(userLocationLat)\n print(userLocationLong)\n\n g = geocoder.osm([userLocationLat, userLocationLong], method='reverse')\n geocoderTriedNum = 0\n while g is None or geocoderTriedNum < 5:\n time.sleep(2)\n g = geocoder.osm([userLocationLat, userLocationLong], method='reverse')\n geocoderTriedNum += 1\n print(g.json['country'])\n userInfo['websitevisitedcountry'] = g.json['country']\n # check user's country is present in the dbpedia\n if dbpedia.IsInfoInDBPedia(userInfo['websitevisitedcountry']):\n userInfo['websitevisitedcountry'] = userInfo['websitevisitedcountry']\n\n print(domain)\n\n # check domain is present in the our graph\n objectIsPresent = blazegraph.checkForSubject(domain)\n comp_info_score = []\n\n # if not present, add info to that graph\n isPresentInDBPedia = False\n # get domain information using alexa api\n if objectIsPresent == False:\n comp_info = alexa.alexa(domain)\n \n comp_info = list(comp_info)\n # check if the website creation date is present\n if comp_info[7] == 'NaN':\n # get expiration date using whois\n websiteInfoFromWhoIs = whois.whois(domain)\n print(\"websiteInfoFromWhoIs:@: \",websiteInfoFromWhoIs)\n calledWhois = True\n if isinstance(websiteInfoFromWhoIs.creation_date, list):\n print(\"websiteDate1:\")\n comp_info[7] = datetime.strftime(websiteInfoFromWhoIs.creation_date[1], \"%Y-%m-%d %H:%M:%S\")\n else:\n print(\"websiteDate2:\")\n comp_info[7] = datetime.strftime(websiteInfoFromWhoIs.creation_date, \"%Y-%m-%d %H:%M:%S\")\n \n # to add create info into rdf.\n blazegraph.add_companyInfo(comp_info)\n\n # delete if NaN is present\n blazegraph.deleteNaN()\n \n # get complete URL and connect with DBPedia\n # check info is present in DBPedia\n comp_info[1] = comp_info[1].replace('/', '')\n comp_info[1] = comp_info[1].replace(' ', '_')\n isPresentInDBPedia = dbpedia.IsInfoInDBPedia(comp_info[1]) \n print(\"isPresentInDBPedia:\", isPresentInDBPedia)\n \n if isPresentInDBPedia:\n print(\"same\")\n # get company name \n #companyTitle = blazegraph.getCompanyName(domain)\n blazegraph.sameAs(domain, comp_info[1])\n\n # get company location information from dbpedia\n companyLoc = dbpedia.getCompanyLocation(comp_info[1])\n \n if companyLoc != None:\n # convert company location into country\n geoLocator = Nominatim(user_agent=\"privacyProtection\")\n companyLocForGeoCoder = companyLoc.split('/')[-1]\n location = geoLocator.geocode(companyLocForGeoCoder)\n geocoderTriedNum2 = 0\n while location is None or geocoderTriedNum2 < 5:\n time.sleep(2)\n location = geoLocator.geocode(companyLocForGeoCoder)\n geocoderTriedNum2 += 1\n companyLoc = location.raw['display_name'].split(\" \")[-1]\n print(\"location country 222\", companyLoc)\n \n if isPresentInDBPedia == False or companyLoc == None:\n # get website domain reg. location using whois\n if calledWhois == False:\n # get expiration location using whois\n websiteInfoFromWhoIs = whois.whois(domain)\n \n # websiteDomainCity = websiteInfoFromWhoIs.city\n # if websiteDomainCity != None: \n # print(\"Company location in app @1@: \", websiteInfoFromWhoIs)\n # companyLoc = websiteDomainCity.replace(\" \", \"_\")\n # else:\n websiteDomainCountry = websiteInfoFromWhoIs.country\n companyLoc = pycountry.countries.get(alpha_2=websiteDomainCountry)\n if companyLoc == None:\n companyLoc = \"NaN\"\n else:\n companyLoc = companyLoc.name\n companyLoc = companyLoc.replace(\" \", \"_\")\n\n # get company information from dbpedia\n print(\"Company location in app @@: \", companyLoc)\n comp_info.append(companyLoc)\n blazegraph.addCompanyLocation(domain, comp_info[8])\n print(\"companyLoc: \", comp_info[8])\n\n # add website protocol info to calculate privacy score\n comp_info.append(protocol)\n comp_info_score = comp_info \n # --------\n else:\n # get company information from our triple store\n comp_info = blazegraph.getCompanyInfoInFormat(subject_m=domain)\n print(\"Company's information: \",comp_info)\n comp_info.append(protocol)\n comp_info_score = comp_info\n\n # get privacy score based on company Info @@to-do send this data to the client\n privacyScore, reasonForPrivacyScore = privacyMetrics.calculatePrivacyScore(comp_info, userInfo)\n print(\"comp_info[4]\", comp_info[4])\n if comp_info[4] is not None and comp_info[4] is not \"NaN\":\n websiteType = comp_info[4].split('/')[0]\n else:\n websiteType = \"others\"\n\n print(\"privacyRiskScore :\", privacyScore)\n print(\"reasonForPrivacyScore :\", reasonForPrivacyScore)\n print(\"websiteType :\", websiteType)\n\n return jsonify({'privacyRiskScore': privacyScore, 'reasonForPrivacyScore': reasonForPrivacyScore, \"websiteType\":websiteType })\n\n################################################################################################\n@app.route('/getRDF', methods=['GET','POST'])\ndef getRDF():\n print(\"RDF\")\n @after_this_request\n def add_header(response):\n response.headers.add('Access-Control-Allow-Origin', '*')\n return response\n\n res = blazegraph.select_all()\n return jsonify(res)\n\n\n################################################################################################\n@app.route('/userProfile', methods=['GET','POST'])\ndef getUserProfile():\n @after_this_request\n def add_header(response):\n response.headers.add('Access-Control-Allow-Origin', '*')\n return response\n \n LinkedInAUTHCode = request.args.get('code')\n\n print(LinkedInAUTHCode)\n\n res = requests.post(\"https://www.linkedin.com/oauth/v2/accessToken?grant_type=authorization_code&code=\"+ LinkedInAUTHCode + \"&redirect_uri=http%3A%2F%2Flocalhost%3A5000%2FuserProfile&client_id=77mpeeyrvnkjaa&client_secret=loIraKqPvMjE9fOe\",\n headers = {\n \"content-type\": \"x-www-form-urlencoded\"\n })\n \n\n if res.ok:\n print('Access_token', res.json())\n linkedInAccessToken = res.json()['access_token']\n linkedInAccessTokenExpiresIn = res.json()['expires_in'] # @@todo time + expiresIn\n else:\n print(res.json())\n\n res1 = requests.get(\"https://api.linkedin.com/v2/me\",\n headers = {\n \"Authorization\": \"Bearer \" + linkedInAccessToken,\n \"connection\" : \"Keep-Alive\"\n })\n\n if res1.ok:\n print('Access_token', res1.json())\n else:\n print(res1.json())\n\n #return render_template('userProfile.html')\n return 'OK'\n\nif __name__ == \"__main__\": \n app.run(debug=True)", "sub_path": "backend/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 9891, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "flask.Flask", "line_number": 30, "usage_type": "call"}, {"api_name": "controller.blazeGraph.baseRules", "line_number": 35, "usage_type": "call"}, {"api_name": "controller.blazeGraph", "line_number": 35, "usage_type": "name"}, {"api_name": "flask.after_this_request", "line_number": 41, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 50, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 50, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 50, "usage_type": "name"}, {"api_name": "tldextract.extract", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 56, "usage_type": "call"}, {"api_name": "flask.request.form.get", "line_number": 61, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 61, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 61, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 64, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 64, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 64, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 65, "usage_type": "call"}, {"api_name": "flask.request.form.get", "line_number": 66, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 66, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 66, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 76, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 76, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 76, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 77, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 77, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 77, "usage_type": "name"}, {"api_name": "geocoder.osm", "line_number": 81, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 84, "usage_type": "call"}, {"api_name": "geocoder.osm", "line_number": 85, "usage_type": "call"}, {"api_name": "controller.dbpedia.IsInfoInDBPedia", "line_number": 90, "usage_type": "call"}, {"api_name": "controller.dbpedia", "line_number": 90, "usage_type": "name"}, {"api_name": "controller.blazeGraph.checkForSubject", "line_number": 96, "usage_type": "call"}, {"api_name": "controller.blazeGraph", "line_number": 96, "usage_type": "name"}, {"api_name": "controller.alexa.alexa", "line_number": 103, "usage_type": "call"}, {"api_name": "controller.alexa", "line_number": 103, "usage_type": "name"}, {"api_name": "whois.whois", "line_number": 109, "usage_type": "call"}, {"api_name": "datetime.datetime.strftime", "line_number": 114, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 114, "usage_type": "name"}, {"api_name": "datetime.datetime.strftime", "line_number": 117, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 117, "usage_type": "name"}, {"api_name": "controller.blazeGraph.add_companyInfo", "line_number": 120, "usage_type": "call"}, {"api_name": "controller.blazeGraph", "line_number": 120, "usage_type": "name"}, {"api_name": "controller.blazeGraph.deleteNaN", "line_number": 123, "usage_type": "call"}, {"api_name": "controller.blazeGraph", "line_number": 123, "usage_type": "name"}, {"api_name": "controller.dbpedia.IsInfoInDBPedia", "line_number": 129, "usage_type": "call"}, {"api_name": "controller.dbpedia", "line_number": 129, "usage_type": "name"}, {"api_name": "controller.blazeGraph.sameAs", "line_number": 136, "usage_type": "call"}, {"api_name": "controller.blazeGraph", "line_number": 136, "usage_type": "name"}, {"api_name": "controller.dbpedia.getCompanyLocation", "line_number": 139, "usage_type": "call"}, {"api_name": "controller.dbpedia", "line_number": 139, "usage_type": "name"}, {"api_name": "geopy.geocoders.Nominatim", "line_number": 143, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 148, "usage_type": "call"}, {"api_name": "whois.whois", "line_number": 158, "usage_type": "call"}, {"api_name": "pycountry.countries.get", "line_number": 166, "usage_type": "call"}, {"api_name": "pycountry.countries", "line_number": 166, "usage_type": "attribute"}, {"api_name": "controller.blazeGraph.addCompanyLocation", "line_number": 176, "usage_type": "call"}, {"api_name": "controller.blazeGraph", "line_number": 176, "usage_type": "name"}, {"api_name": "controller.blazeGraph.getCompanyInfoInFormat", "line_number": 185, "usage_type": "call"}, {"api_name": "controller.blazeGraph", "line_number": 185, "usage_type": "name"}, {"api_name": "controller.privacyMetrics.calculatePrivacyScore", "line_number": 191, "usage_type": "call"}, {"api_name": "controller.privacyMetrics", "line_number": 191, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 202, "usage_type": "call"}, {"api_name": "flask.after_this_request", "line_number": 208, "usage_type": "name"}, {"api_name": "controller.blazeGraph.select_all", "line_number": 213, "usage_type": "call"}, {"api_name": "controller.blazeGraph", "line_number": 213, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 214, "usage_type": "call"}, {"api_name": "flask.after_this_request", "line_number": 220, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 225, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 225, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 225, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 229, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 242, "usage_type": "call"}]} +{"seq_id": "281391442", "text": "import io\nimport copy\nimport _pickle as cPickle\nimport logging\nfrom pathlib import Path\nfrom typing import List, Dict, Set, Optional, Iterator, Union, TextIO\nfrom collections import OrderedDict\n\nfrom pyknp import BList, Bunsetsu, Tag, Morpheme, Rel\nimport jaconv\n\nfrom kyoto_reader.pas import Pas, Predicate, BaseArgument, Argument, SpecialArgument\nfrom kyoto_reader.coreference import Mention, Entity\nfrom kyoto_reader.ne import NamedEntity\nfrom kyoto_reader.constants import ALL_CASES, CORE_CASES, ALL_EXOPHORS, ALL_COREFS, CORE_COREFS, NE_CATEGORIES\nfrom kyoto_reader.base_phrase import BasePhrase\n\nlogger = logging.getLogger(__name__)\nlogger.setLevel(logging.WARNING)\n\n\nclass KyotoReader:\n \"\"\" KWDLC(または Kyoto Corpus)の文書集合を扱うクラス\n\n Args:\n source (Union[Path, str]): 入力ソース.Path オブジェクトを指定するとその場所のファイルを読む\n target_cases (Optional[List[str]]): 抽出の対象とする格\n target_corefs (Optional[List[str]]): 抽出の対象とする共参照関係(=など)\n relax_cases (bool): ガ≒格などをガ格として扱うか\n extract_nes (bool): 固有表現をコーパスから抽出するかどうか\n knp_ext (str): KWDLC または KC ファイルの拡張子\n pickle_ext (str): Document を pickle 形式で読む場合の拡張子\n use_pas_tag (bool): タグからではなく、<述語項構造:>タグから PAS を読むかどうか\n \"\"\"\n def __init__(self,\n source: Union[Path, str],\n target_cases: Optional[List[str]],\n target_corefs: Optional[List[str]],\n relax_cases: bool = False,\n extract_nes: bool = True,\n use_pas_tag: bool = False,\n knp_ext: str = '.knp',\n pickle_ext: str = '.pkl',\n ) -> None:\n if not (isinstance(source, Path) or isinstance(source, str)):\n raise TypeError(f'source must be Path or str type, but got {type(source)}')\n if isinstance(source, Path):\n if source.is_dir():\n logger.info(f'got directory path, files in the directory is treated as source files')\n file_paths: List[Path] = \\\n sorted(source.glob(f'**/*{knp_ext}')) + sorted(source.glob(f'**/*{pickle_ext}'))\n self.did2source: Dict[str, Union[Path, str]] = OrderedDict((path.stem, path) for path in file_paths)\n else:\n logger.info(f'got file path, this file is treated as a source knp file')\n self.did2source: Dict[str, Union[Path, str]] = {source.stem: source}\n else:\n logger.info(f'got string, this string is treated as a source knp string')\n self.did2source: Dict[str, Union[Path, str]] = {'doc': source}\n\n self.target_cases: List[str] = self._get_target(target_cases, ALL_CASES, CORE_CASES, 'case')\n self.target_corefs: List[str] = self._get_target(target_corefs, ALL_COREFS, CORE_COREFS, 'coref')\n self.relax_cases: bool = relax_cases\n self.extract_nes: bool = extract_nes\n self.use_pas_tag: bool = use_pas_tag\n self.knp_ext: str = knp_ext\n self.pickle_ext: str = pickle_ext\n\n @staticmethod\n def _get_target(input_: Optional[list],\n all_: list,\n default: list,\n type_: str,\n ) -> list:\n if input_ is None:\n return default\n target = []\n for item in input_:\n if item not in all_:\n logger.warning(f'unknown target {type_}: {item}')\n continue\n target.append(item)\n\n return target\n\n def get_doc_ids(self) -> List[str]:\n return list(self.did2source.keys())\n\n def process_document(self, doc_id: str) -> Optional['Document']:\n if doc_id not in self.did2source:\n logger.error(f'unknown document id: {doc_id}')\n return None\n if isinstance(self.did2source[doc_id], Path):\n if self.did2source[doc_id].suffix == self.pickle_ext:\n with self.did2source[doc_id].open(mode='rb') as f:\n return cPickle.load(f)\n elif self.did2source[doc_id].suffix == self.knp_ext:\n with self.did2source[doc_id].open() as f:\n input_string = f.read()\n else:\n return None\n else:\n input_string = self.did2source[doc_id]\n return Document(input_string,\n doc_id,\n self.target_cases,\n self.target_corefs,\n self.relax_cases,\n self.extract_nes,\n self.use_pas_tag)\n\n def process_documents(self, doc_ids: List[str]) -> Iterator[Optional['Document']]:\n for doc_id in doc_ids:\n yield self.process_document(doc_id)\n\n def process_all_documents(self) -> Iterator['Document']:\n for doc_id in self.did2source.keys():\n yield self.process_document(doc_id)\n\n\nclass Document:\n \"\"\" KWDLC(または Kyoto Corpus)の1文書を扱うクラス\n\n Args:\n knp_string (str): 文書ファイルの内容(knp形式)\n doc_id (str): 文書ID\n cases (List[str]): 抽出の対象とする格\n corefs (List[str]): 抽出の対象とする共参照関係(=など)\n relax_cases (bool): ガ≒格などをガ格として扱うか\n extract_nes (bool): 固有表現をコーパスから抽出するかどうか\n use_pas_tag (bool): タグからではなく、<述語項構造:>タグから PAS を読むかどうか\n\n Attributes:\n knp_string (str): 文書ファイルの内容(knp形式)\n doc_id (str): 文書ID(ファイル名から拡張子を除いたもの)\n cases (List[str]): 抽出の対象とする格\n corefs (List[str]): 抽出の対象とする共参照関係(=など)\n extract_nes (bool): 固有表現をコーパスから抽出するかどうか\n sid2sentence (dict): 文IDと文を紐付ける辞書\n bnst2dbid (dict): 文節IDと文書レベルの文節IDを紐付ける辞書\n tag2dtid (dict): 基本句IDと文書レベルの基本句IDを紐付ける辞書\n mrph2dmid (dict): 形態素IDと文書レベルの形態素IDを紐付ける辞書\n mentions (dict): dtid を key とする mention の辞書\n entities (dict): entity id を key として entity オブジェクトが格納されている\n named_entities (list): 抽出した固有表現\n \"\"\"\n def __init__(self,\n knp_string: str,\n doc_id: str,\n cases: List[str],\n corefs: List[str],\n relax_cases: bool,\n extract_nes: bool,\n use_pas_tag: bool,\n ) -> None:\n self.knp_string: str = knp_string\n self.doc_id: str = doc_id\n self.cases: List[str] = cases\n self.corefs: List[str] = corefs\n self.relax_cases: bool = relax_cases\n self.extract_nes: bool = extract_nes\n self.use_pas_tag: bool = use_pas_tag\n\n self.sid2sentence: Dict[str, BList] = OrderedDict()\n buff = []\n for line in knp_string.strip().split('\\n'):\n buff.append(line)\n if line.strip() == 'EOS':\n sentence = BList('\\n'.join(buff) + '\\n')\n if sentence.sid in self.sid2sentence:\n logger.warning(f'{sentence.sid:24}duplicated sid found')\n self.sid2sentence[sentence.sid] = sentence\n buff = []\n\n self.bnst2dbid = {}\n self.tag2dtid = {}\n self.mrph2dmid = {}\n self._assign_document_wide_id()\n\n self._pas: Dict[int, Pas] = OrderedDict()\n self.mentions: Dict[int, Mention] = OrderedDict()\n self.entities: Dict[int, Entity] = OrderedDict()\n if use_pas_tag:\n self._analyze_pas()\n else:\n self._analyze_rel()\n\n if extract_nes:\n self.named_entities: List[NamedEntity] = []\n self._extract_nes()\n\n def _assign_document_wide_id(self) -> None:\n \"\"\"文節・基本句・形態素に文書全体に渡る通し番号を振る\"\"\"\n dbid, dtid, dmid = 0, 0, 0\n for sentence in self.sentences:\n for bnst in sentence.bnst_list():\n for tag in bnst.tag_list():\n for mrph in tag.mrph_list():\n self.mrph2dmid[mrph] = dmid\n dmid += 1\n self.tag2dtid[tag] = dtid\n dtid += 1\n self.bnst2dbid[bnst] = dbid\n dbid += 1\n\n def _analyze_pas(self) -> None:\n \"\"\"extract predicate argument structure from <述語項構造:> tag in knp string\"\"\"\n sid2idx = {sid: idx for idx, sid in enumerate(self.sid2sentence.keys())}\n for tag in self.tag_list():\n if tag.pas is None:\n continue\n pas = Pas(BasePhrase(tag, self.tag2dtid[tag], tag.pas.sid, self.mrph2dmid), self.mrph2dmid)\n for case, arguments in tag.pas.arguments.items():\n if self.relax_cases:\n if case in ALL_CASES and case.endswith('≒'):\n case = case.rstrip('≒') # ガ≒ -> ガ\n for arg in arguments:\n arg.midasi = jaconv.h2z(arg.midasi, digit=True) # 不特定:人1 -> 不特定:人1\n # exophor\n if arg.flag == 'E':\n entity = self._create_entity(exophor=arg.midasi, eid=arg.eid)\n pas.add_special_argument(case, arg.midasi, entity.eid, '')\n else:\n sid = self.sentences[sid2idx[arg.sid] - arg.sdist].sid\n arg_bp = self._get_bp(sid, arg.tid)\n mention = self._create_mention(arg_bp)\n pas.add_argument(case, mention, '', self.mrph2dmid)\n if pas.arguments:\n self._pas[pas.dtid] = pas\n\n def _analyze_rel(self) -> None:\n \"\"\"extract predicate argument structure and coreference relation from tag in knp string\"\"\"\n tag2sid = {tag: sentence.sid for sentence in self.sentences for tag in sentence.tag_list()}\n for tag in self.tag_list():\n rels = []\n for rel in self._extract_rel_tags(tag):\n if self.relax_cases:\n if rel.atype in ALL_CASES and rel.atype.endswith('≒'):\n rel.atype = rel.atype.rstrip('≒') # ガ≒ -> ガ\n valid = True\n if rel.sid is not None and rel.sid not in self.sid2sentence:\n logger.warning(f'{tag2sid[tag]:24}sentence: {rel.sid} not found in {self.doc_id}')\n valid = False\n if rel.atype in (ALL_CASES + ALL_COREFS):\n if rel.atype not in (self.cases + self.corefs):\n logger.info(f'{tag2sid[tag]:24}relation type: {rel.atype} is ignored')\n valid = False\n else:\n logger.warning(f'{tag2sid[tag]:24}unknown relation: {rel.atype}')\n if valid:\n rels.append(rel)\n src_bp = BasePhrase(tag, self.tag2dtid[tag], tag2sid[tag], self.mrph2dmid)\n # extract PAS\n pas = Pas(src_bp, self.mrph2dmid)\n for rel in rels:\n if rel.atype in self.cases:\n if rel.sid is not None:\n assert rel.tid is not None\n arg_bp = self._get_bp(rel.sid, rel.tid)\n if arg_bp is None:\n continue\n # 項を発見したら同時に mention と entity を作成\n mention = self._create_mention(arg_bp)\n pas.add_argument(rel.atype, mention, rel.mode, self.mrph2dmid)\n # exophora\n else:\n if rel.target == 'なし':\n pas.set_arguments_optional(rel.atype)\n continue\n if rel.target not in ALL_EXOPHORS:\n logger.warning(f'{pas.sid:24}unknown exophor: {rel.target}')\n continue\n entity = self._create_entity(rel.target)\n pas.add_special_argument(rel.atype, rel.target, entity.eid, rel.mode)\n if pas.arguments:\n self._pas[pas.dtid] = pas\n\n # extract coreference\n for rel in rels:\n if rel.atype in self.corefs:\n if rel.mode in ('', 'AND'): # ignore \"OR\" and \"?\"\n self._add_corefs(src_bp, rel)\n\n # to extract rels with mode: '?', rewrite initializer of pyknp Futures class\n @staticmethod\n def _extract_rel_tags(tag: Tag) -> List[Rel]:\n \"\"\"parse tag.fstring to extract tags\"\"\"\n splitter = \"><\"\n rels = []\n spec = tag.fstring\n\n tag_start = 1\n tag_end = None\n while tag_end != -1:\n tag_end = spec.find(splitter, tag_start)\n if spec[tag_start:].startswith('rel '):\n rel = Rel(spec[tag_start:tag_end])\n if rel.target:\n rel.target = jaconv.h2z(rel.target, digit=True) # 不特定:人1 -> 不特定:人1\n if rel.atype is not None:\n rels.append(rel)\n\n tag_start = tag_end + len(splitter)\n return rels\n\n def _add_corefs(self,\n source_bp: BasePhrase,\n rel: Rel,\n ) -> None:\n if rel.sid is not None:\n target_bp = self._get_bp(rel.sid, rel.tid)\n if target_bp is None:\n return\n if target_bp.dtid == source_bp.dtid:\n logger.warning(f'{source_bp.sid:24}coreference with self found: {source_bp.midasi}')\n return\n else:\n target_bp = None\n if rel.target not in ALL_EXOPHORS:\n logger.warning(f'{source_bp.sid:24}unknown exophor: {rel.target}')\n return\n\n uncertain: bool = rel.atype.endswith('≒')\n source_mention = self._create_mention(source_bp)\n for eid in source_mention.all_eids:\n # _merge_entities によって source_mention の eid が削除されているかもしれない\n if eid not in self.entities:\n continue\n source_entity = self.entities[eid]\n if rel.sid is not None:\n target_mention = self._create_mention(target_bp)\n for target_eid in target_mention.all_eids:\n target_entity = self.entities[target_eid]\n self._merge_entities(source_mention, target_mention, source_entity, target_entity, uncertain)\n else:\n target_entity = self._create_entity(exophor=rel.target)\n self._merge_entities(source_mention, None, source_entity, target_entity, uncertain)\n\n def _create_mention(self, bp: BasePhrase) -> Mention:\n \"\"\"メンションを作成\n bp がまだ mention として登録されていなければ新しく entity と共に作成.\n 登録されていればその mention を返す.\n\n Args:\n bp (BasePhrase): 基本句\n\n Returns:\n Mention: メンション\n \"\"\"\n if bp.dtid not in self.mentions:\n # new coreference cluster is made\n mention = Mention(bp, self.mrph2dmid)\n self.mentions[bp.dtid] = mention\n entity = self._create_entity()\n entity.add_mention(mention, uncertain=False)\n else:\n mention = self.mentions[bp.dtid]\n return mention\n\n def _create_entity(self,\n exophor: Optional[str] = None,\n eid: Optional[int] = None,\n ) -> Entity:\n \"\"\"エンティティを作成\n\n exophor が singleton entity だった場合を除き、新しく Entity のインスタンスを作成して返す\n singleton entity とは、「著者」や「不特定:人1」などの必ず一つしか存在しないような entity\n 一方で、「不特定:人」や「不特定:物」は複数存在しうるので singleton entity ではない\n eid を指定しない場合、最後に作成した entity の次の eid を選択\n\n Args:\n exophor (Optional[str]): 外界照応詞(optional)\n eid (Optional[int]): エンティティID(省略推奨)\n\n Returns:\n Entity: エンティティ\n \"\"\"\n if exophor:\n if exophor not in ('不特定:人', '不特定:物', '不特定:状況'): # exophor が singleton entity だった時\n entities = [e for e in self.entities.values() if exophor == e.exophor]\n # すでに singleton entity が存在した場合、新しい entity は作らずにその entity を返す\n if entities:\n assert len(entities) == 1 # singleton entity が1つしかないことを保証\n return entities[0]\n eids: List[int] = [e.eid for e in self.entities.values()]\n if eid in eids:\n eid_ = eid\n eid: int = max(eids) + 1\n logger.warning(f'{self.doc_id:24}eid: {eid_} is already used. use eid: {eid} instead.')\n elif eid is None or eid < 0:\n eid: int = max(eids) + 1 if eids else 0\n entity = Entity(eid, exophor=exophor)\n self.entities[eid] = entity\n return entity\n\n def _merge_entities(self,\n source_mention: Mention,\n target_mention: Optional[Mention],\n se: Entity,\n te: Entity,\n uncertain: bool,\n ) -> None:\n \"\"\"2つのエンティティをマージする\n\n source_mention と se, target_mention と te の間には mention が張られているが、\n source と target 間には張られていないので、add_mention する\n se と te が同一のエンティティであり、exophor も同じか片方が None ならば te の方を削除する\n\n Args:\n source_mention (Mention): 参照元メンション\n target_mention (Mention?): 参照先メンション\n se (Entity): 参照元エンティティ\n te (Entity): 参照先エンティティ\n uncertain (bool): source_mention と target_mention のアノテーションが ≒ かどうか\n \"\"\"\n uncertain_tgt = (target_mention is not None) and target_mention.is_uncertain_to(te)\n uncertain_src = source_mention.is_uncertain_to(se)\n if se is te:\n if not uncertain:\n # se(te), source_mention, target_mention の三角形のうち2辺が certain ならもう1辺も certain\n if (not uncertain_src) and uncertain_tgt:\n se.add_mention(target_mention, uncertain=False)\n if uncertain_src and (not uncertain_tgt):\n se.add_mention(source_mention, uncertain=False)\n return\n if target_mention is not None:\n se.add_mention(target_mention, uncertain=(uncertain or uncertain_src))\n te.add_mention(source_mention, uncertain=(uncertain or uncertain_tgt))\n # se と te が同一でない可能性が拭えない場合、te は削除しない\n if uncertain_src or uncertain or uncertain_tgt:\n return\n # se と te が同一でも exophor が異なれば te は削除しない\n if se.exophor is not None and te.exophor is not None and se.exophor != te.exophor:\n return\n # 以下 te を削除する準備\n if se.exophor is None:\n se.exophor = te.exophor\n for tm in te.all_mentions:\n se.add_mention(tm, uncertain=tm.is_uncertain_to(te))\n # argument も eid を持っているので eid が変わった場合はこちらも更新\n for arg in [arg for pas in self._pas.values() for args in pas.arguments.values() for arg in args]:\n if isinstance(arg, SpecialArgument) and arg.eid == te.eid:\n arg.eid = se.eid\n self._delete_entity(te.eid, source_mention.sid) # delete target entity\n\n def _delete_entity(self,\n eid: int,\n sid: str\n ) -> None:\n \"\"\"entity を削除する\n\n 対象の entity を entities から削除すると共に、\n その entity を参照する全ての mention からも削除\n eid に欠番ができる\n\n Args:\n eid (int): 削除対象の entity の EID\n sid (int): 削除された時解析されていた文の文ID\n \"\"\"\n if eid not in self.entities:\n return\n entity = self.entities[eid]\n logger.info(f'{sid:24}delete entity: {eid} ({entity.midasi})')\n for mention in entity.all_mentions:\n entity.remove_mention(mention)\n self.entities.pop(eid)\n\n def _get_bp(self,\n sid: str,\n tid: int,\n ) -> Optional[BasePhrase]:\n \"\"\"文IDと基本句IDから基本句を得る\n\n Args:\n sid (str): 文ID\n tid (int): 基本句ID\n\n Returns:\n Optional[BasePhrase]: 対応する基本句\n \"\"\"\n tag_list = self.sid2sentence[sid].tag_list()\n if not (0 <= tid < len(tag_list)):\n logger.warning(f'{sid:24}tag id: {tid} out of range')\n return None\n tag = tag_list[tid]\n return BasePhrase(tag, self.tag2dtid[tag], sid, self.mrph2dmid)\n\n def _extract_nes(self) -> None:\n \"\"\"KNP の tag を参照して文書中から固有表現を抽出する\"\"\"\n for sentence in self.sentences:\n tag_list = sentence.tag_list()\n # tag.features = {'NE': 'LOCATION:ダーマ神殿'}\n for tag in tag_list:\n if 'NE' not in tag.features:\n continue\n category, midasi = tag.features['NE'].split(':', maxsplit=1)\n if category not in NE_CATEGORIES:\n logger.warning(f'{sentence.sid:24}unknown NE category: {category}')\n continue\n mrph_list = [m for t in tag_list[:tag.tag_id + 1] for m in t.mrph_list()]\n mrph_span = self._find_mrph_span(midasi, mrph_list, tag)\n if mrph_span is None:\n logger.warning(f'{sentence.sid:24}mrph span of \"{midasi}\" not found')\n continue\n ne = NamedEntity(category, midasi, sentence, mrph_span, self.mrph2dmid)\n self.named_entities.append(ne)\n\n @staticmethod\n def _find_mrph_span(midasi: str,\n mrph_list: List[Morpheme],\n tag: Tag\n ) -> Optional[range]:\n \"\"\"midasiにマッチする形態素の範囲を返す\"\"\"\n for i in range(len(tag.mrph_list())):\n end_mid = len(mrph_list) - i\n mrph_span = ''\n for mrph in reversed(mrph_list[:end_mid]):\n mrph_span = mrph.midasi + mrph_span\n if mrph_span == midasi:\n return range(mrph.mrph_id, end_mid)\n return None\n\n @property\n def sentences(self) -> List[BList]:\n \"\"\"文を構成する全文節列オブジェクト\n\n Returns:\n List[BList]\n \"\"\"\n return list(self.sid2sentence.values())\n\n def bnst_list(self) -> List[Bunsetsu]:\n return [bnst for sentence in self.sentences for bnst in sentence.bnst_list()]\n\n def tag_list(self) -> List[Tag]:\n return [tag for sentence in self.sentences for tag in sentence.tag_list()]\n\n def mrph_list(self) -> List[Morpheme]:\n return [mrph for sentence in self.sentences for mrph in sentence.mrph_list()]\n\n def get_entities(self, tag: Tag) -> List[Entity]:\n return [e for e in self.entities.values() if any(m.dtid == self.tag2dtid[tag] for m in e.mentions)]\n\n def pas_list(self) -> List[Pas]:\n return list(self._pas.values())\n\n def get_predicates(self) -> List[Predicate]:\n return [pas.predicate for pas in self._pas.values()]\n\n def get_arguments(self,\n predicate: Predicate,\n relax: bool = False,\n include_optional: bool = False,\n ) -> Dict[str, List[BaseArgument]]:\n \"\"\"述語 predicate が持つ全ての項を返す\n\n Args:\n predicate (Predicate): 述語\n relax (bool): coreference chain によってより多くの項を返すかどうか\n include_optional (bool): 「すぐに」などの修飾的な項も返すかどうか\n\n Returns:\n Dict[str, List[BaseArgument]]: 格を key とする述語の項の辞書\n \"\"\"\n if predicate.dtid not in self._pas:\n return {}\n pas = copy.copy(self._pas[predicate.dtid])\n pas.arguments = cPickle.loads(cPickle.dumps(pas.arguments, -1))\n if include_optional is False:\n for case in self.cases:\n pas.arguments[case] = list(filter(lambda a: a.optional is False, pas.arguments[case]))\n\n if relax is True:\n for case, args in self._pas[predicate.dtid].arguments.items():\n for arg in args:\n for eid in (arg.all_eids if isinstance(arg, Argument) else arg.eids):\n entity = self.entities[eid]\n if entity.is_special and entity.exophor != arg.midasi:\n pas.add_special_argument(case, entity.exophor, entity.eid, 'AND')\n for mention in entity.all_mentions:\n if isinstance(arg, Argument) and mention.dtid == arg.dtid:\n continue\n pas.add_argument(case, mention, 'AND', self.mrph2dmid)\n\n return pas.arguments\n\n def get_siblings(self, mention: Mention, relax: bool = False) -> Set[Mention]:\n \"\"\"mention と共参照関係にある他の全ての mention を返す\"\"\"\n mentions = set()\n for eid in mention.eids:\n entity = self.entities[eid]\n mentions.update(entity.mentions)\n if relax is True:\n for eid in mention.eids_unc:\n entity = self.entities[eid]\n mentions.update(entity.all_mentions)\n if mention in mentions:\n mentions.remove(mention)\n return mentions\n\n def draw_tree(self,\n sid: str,\n coreference: bool,\n fh: Optional[TextIO] = None,\n ) -> None:\n \"\"\"sid で指定された文の述語項構造・共参照関係をツリー形式で fh に書き出す\n\n Args:\n sid (str): 出力対象の文ID\n coreference (bool): 共参照関係も出力するかどうか\n fh (Optional[TextIO]): 出力ストリーム\n \"\"\"\n sentence: BList = self[sid]\n with io.StringIO() as string:\n sentence.draw_tag_tree(fh=string)\n tree_strings = string.getvalue().rstrip('\\n').split('\\n')\n assert len(tree_strings) == len(sentence.tag_list())\n all_midasis = [m.midasi for m in self.mentions.values()]\n for predicate in filter(lambda p: p.sid == sid, self.get_predicates()):\n idx = predicate.tid\n tree_strings[idx] += ' '\n arguments = self.get_arguments(predicate)\n for case in self.cases:\n args = arguments[case]\n targets = set()\n for arg in args:\n target = arg.midasi\n if all_midasis.count(arg.midasi) > 1 and isinstance(arg, Argument):\n target += str(arg.dtid)\n targets.add(target)\n tree_strings[idx] += f'{\",\".join(targets)}:{case} '\n if coreference:\n for src_mention in filter(lambda m: m.sid == sid, self.mentions.values()):\n tgt_mentions = [tgt for tgt in self.get_siblings(src_mention) if tgt.dtid < src_mention.dtid]\n targets = set()\n for tgt_mention in tgt_mentions:\n target = tgt_mention.midasi\n if all_midasis.count(target) > 1:\n target += str(tgt_mention.dtid)\n targets.add(target)\n for eid in src_mention.eids:\n entity = self.entities[eid]\n if entity.is_special:\n targets.add(entity.exophor)\n if not targets:\n continue\n idx = src_mention.tid\n tree_strings[idx] += ' =:'\n tree_strings[idx] += ','.join(targets)\n\n print('\\n'.join(tree_strings), file=fh)\n\n def stat(self) -> dict:\n \"\"\"calculate document statistics\"\"\"\n ret = dict()\n ret['num_sents'] = len(self)\n ret['num_tags'] = len(self.tag_list())\n ret['num_mrphs'] = len(self.mrph_list())\n ret['num_taigen'] = sum(1 for tag in self.tag_list() if '体言' in tag.features)\n ret['num_yougen'] = sum(1 for tag in self.tag_list() if '用言' in tag.features)\n ret['num_entities'] = len(self.entities)\n ret['num_special_entities'] = sum(1 for ent in self.entities.values() if ent.is_special)\n\n num_mention = num_taigen = num_yougen = 0\n for src_mention in self.mentions.values():\n tgt_mentions: Set[Mention] = self.get_siblings(src_mention)\n if tgt_mentions:\n num_mention += 1\n for tgt_mention in tgt_mentions:\n if '体言' in tgt_mention.tag.features:\n num_taigen += 1\n if '用言' in tgt_mention.tag.features:\n num_yougen += 1\n ret['num_mentions'] = num_mention\n ret['num_taigen_mentions'] = num_taigen\n ret['num_yougen_mentions'] = num_yougen\n\n return ret\n\n def __len__(self):\n return len(self.sid2sentence)\n\n def __getitem__(self, sid: str):\n if sid in self.sid2sentence:\n return self.sid2sentence[sid]\n else:\n logger.error(f'sentence: {sid} is not in this document')\n return None\n\n def __iter__(self):\n return iter(self.sid2sentence.values())\n\n def __str__(self):\n return '\\n'.join(''.join(tag.midasi for tag in sent.tag_list()) for sent in self)\n", "sub_path": "src/kyoto_reader/reader.py", "file_name": "reader.py", "file_ext": "py", "file_size_in_byte": 31117, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "logging.getLogger", "line_number": 18, "usage_type": "call"}, {"api_name": "logging.WARNING", "line_number": 19, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 36, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 36, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 37, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 37, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 38, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 38, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 45, "usage_type": "argument"}, {"api_name": "pathlib.Path", "line_number": 47, "usage_type": "argument"}, {"api_name": "typing.List", "line_number": 50, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 50, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 52, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 52, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 52, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 52, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 55, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 55, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 55, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 58, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 58, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 58, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 60, "usage_type": "name"}, {"api_name": "kyoto_reader.constants.ALL_CASES", "line_number": 60, "usage_type": "argument"}, {"api_name": "kyoto_reader.constants.CORE_CASES", "line_number": 60, "usage_type": "argument"}, {"api_name": "typing.List", "line_number": 61, "usage_type": "name"}, {"api_name": "kyoto_reader.constants.ALL_COREFS", "line_number": 61, "usage_type": "argument"}, {"api_name": "kyoto_reader.constants.CORE_COREFS", "line_number": 61, "usage_type": "argument"}, {"api_name": "typing.Optional", "line_number": 69, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 85, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 92, "usage_type": "argument"}, {"api_name": "_pickle.load", "line_number": 95, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 88, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 111, "usage_type": "name"}, {"api_name": "typing.Iterator", "line_number": 111, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 111, "usage_type": "name"}, {"api_name": "typing.Iterator", "line_number": 115, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 149, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 150, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 157, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 158, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 163, "usage_type": "name"}, {"api_name": "pyknp.BList", "line_number": 163, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 163, "usage_type": "call"}, {"api_name": "pyknp.BList", "line_number": 168, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 179, "usage_type": "name"}, {"api_name": "kyoto_reader.pas.Pas", "line_number": 179, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 179, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 180, "usage_type": "name"}, {"api_name": "kyoto_reader.coreference.Mention", "line_number": 180, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 180, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 181, "usage_type": "name"}, {"api_name": "kyoto_reader.coreference.Entity", "line_number": 181, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 181, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 188, "usage_type": "name"}, {"api_name": "kyoto_reader.ne.NamedEntity", "line_number": 188, "usage_type": "name"}, {"api_name": "kyoto_reader.pas.Pas", "line_number": 211, "usage_type": "call"}, {"api_name": "kyoto_reader.base_phrase.BasePhrase", "line_number": 211, "usage_type": "call"}, {"api_name": "kyoto_reader.constants.ALL_CASES", "line_number": 214, "usage_type": "name"}, {"api_name": "jaconv.h2z", "line_number": 217, "usage_type": "call"}, {"api_name": "kyoto_reader.constants.ALL_CASES", "line_number": 237, "usage_type": "name"}, {"api_name": "kyoto_reader.constants.ALL_CASES", "line_number": 243, "usage_type": "name"}, {"api_name": "kyoto_reader.constants.ALL_COREFS", "line_number": 243, "usage_type": "name"}, {"api_name": "kyoto_reader.base_phrase.BasePhrase", "line_number": 251, "usage_type": "call"}, {"api_name": "kyoto_reader.pas.Pas", "line_number": 253, "usage_type": "call"}, {"api_name": "kyoto_reader.constants.ALL_EXOPHORS", "line_number": 269, "usage_type": "name"}, {"api_name": "pyknp.Tag", "line_number": 285, "usage_type": "name"}, {"api_name": "pyknp.Rel", "line_number": 296, "usage_type": "call"}, {"api_name": "jaconv.h2z", "line_number": 298, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 285, "usage_type": "name"}, {"api_name": "pyknp.Rel", "line_number": 285, "usage_type": "name"}, {"api_name": "kyoto_reader.base_phrase.BasePhrase", "line_number": 306, "usage_type": "name"}, {"api_name": "pyknp.Rel", "line_number": 307, "usage_type": "name"}, {"api_name": "kyoto_reader.constants.ALL_EXOPHORS", "line_number": 318, "usage_type": "name"}, {"api_name": "kyoto_reader.base_phrase.BasePhrase", "line_number": 338, "usage_type": "name"}, {"api_name": "kyoto_reader.coreference.Mention", "line_number": 351, "usage_type": "call"}, {"api_name": "kyoto_reader.coreference.Mention", "line_number": 338, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 360, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 361, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 384, "usage_type": "name"}, {"api_name": "kyoto_reader.coreference.Entity", "line_number": 391, "usage_type": "call"}, {"api_name": "kyoto_reader.coreference.Entity", "line_number": 362, "usage_type": "name"}, {"api_name": "kyoto_reader.coreference.Mention", "line_number": 396, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 397, "usage_type": "name"}, {"api_name": "kyoto_reader.coreference.Mention", "line_number": 397, "usage_type": "name"}, {"api_name": "kyoto_reader.coreference.Entity", "line_number": 398, "usage_type": "name"}, {"api_name": "kyoto_reader.coreference.Entity", "line_number": 399, "usage_type": "name"}, {"api_name": "kyoto_reader.pas.SpecialArgument", "line_number": 441, "usage_type": "argument"}, {"api_name": "kyoto_reader.base_phrase.BasePhrase", "line_number": 485, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 470, "usage_type": "name"}, {"api_name": "kyoto_reader.base_phrase.BasePhrase", "line_number": 470, "usage_type": "name"}, {"api_name": "kyoto_reader.constants.NE_CATEGORIES", "line_number": 496, "usage_type": "name"}, {"api_name": "kyoto_reader.ne.NamedEntity", "line_number": 504, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 509, "usage_type": "name"}, {"api_name": "pyknp.Morpheme", "line_number": 509, "usage_type": "name"}, {"api_name": "pyknp.Tag", "line_number": 510, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 511, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 523, "usage_type": "name"}, {"api_name": "pyknp.BList", "line_number": 523, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 531, "usage_type": "name"}, {"api_name": "pyknp.Bunsetsu", "line_number": 531, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 534, "usage_type": "name"}, {"api_name": "pyknp.Tag", "line_number": 534, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 537, "usage_type": "name"}, {"api_name": "pyknp.Morpheme", "line_number": 537, "usage_type": "name"}, {"api_name": "pyknp.Tag", "line_number": 540, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 540, "usage_type": "name"}, {"api_name": "kyoto_reader.coreference.Entity", "line_number": 540, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 543, "usage_type": "name"}, {"api_name": "kyoto_reader.pas.Pas", "line_number": 543, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 546, "usage_type": "name"}, {"api_name": "kyoto_reader.pas.Predicate", "line_number": 546, "usage_type": "name"}, {"api_name": "kyoto_reader.pas.Predicate", "line_number": 550, "usage_type": "name"}, {"api_name": "copy.copy", "line_number": 566, "usage_type": "call"}, {"api_name": "_pickle.loads", "line_number": 567, "usage_type": "call"}, {"api_name": "_pickle.dumps", "line_number": 567, "usage_type": "call"}, {"api_name": "kyoto_reader.pas.Argument", "line_number": 575, "usage_type": "argument"}, {"api_name": "kyoto_reader.pas.Argument", "line_number": 580, "usage_type": "argument"}, {"api_name": "typing.Dict", "line_number": 553, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 553, "usage_type": "name"}, {"api_name": "kyoto_reader.pas.BaseArgument", "line_number": 553, "usage_type": "name"}, {"api_name": "kyoto_reader.coreference.Mention", "line_number": 586, "usage_type": "name"}, {"api_name": "typing.Set", "line_number": 586, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 603, "usage_type": "name"}, {"api_name": "typing.TextIO", "line_number": 603, "usage_type": "name"}, {"api_name": "pyknp.BList", "line_number": 612, "usage_type": "name"}, {"api_name": "io.StringIO", "line_number": 613, "usage_type": "call"}, {"api_name": "kyoto_reader.pas.Argument", "line_number": 627, "usage_type": "argument"}, {"api_name": "typing.Set", "line_number": 665, "usage_type": "name"}, {"api_name": "kyoto_reader.coreference.Mention", "line_number": 665, "usage_type": "name"}]} +{"seq_id": "260999875", "text": "import pytest\n\nfrom conftest import quoted_object, run_setup_sql\nfrom pgbedrock import ownerships as own\nfrom pgbedrock import attributes, privileges\nfrom pgbedrock.context import ObjectInfo\n\nQ_CREATE_SEQUENCE = 'SET ROLE {}; CREATE SEQUENCE {}.{}; RESET ROLE;'\nQ_SCHEMA_EXISTS = \"SELECT schema_name FROM information_schema.schemata WHERE schema_name='{}';\"\n\nROLES = tuple('role{}'.format(i) for i in range(2))\nSCHEMAS = tuple('schema{}'.format(i) for i in range(3))\nTABLES = tuple('table{}'.format(i) for i in range(4))\nSEQUENCES = tuple('seq{}'.format(i) for i in range(4))\nDUMMY = 'foo'\n\n\n@run_setup_sql([\n 'DROP SCHEMA public',\n 'CREATE SCHEMA {}'.format(SCHEMAS[0]),\n ])\ndef test_analyze_schemas_with_undocumented_items(capsys, cursor):\n spec = {'postgres': {'has_personal_schema': False}}\n\n with pytest.raises(SystemExit):\n own.analyze_schemas(spec, cursor, verbose=False)\n\n # Undocumented schemas will come back double-quoted\n missing_schemas = '\"information_schema\", \"pg_catalog\", \"schema0\"'\n assert capsys.readouterr()[0] == own.UNDOCUMENTED_SCHEMAS_MSG.format(missing_schemas) + \"\\n\"\n\n\n@run_setup_sql([\n 'DROP SCHEMA public',\n attributes.Q_CREATE_ROLE.format(ROLES[0]),\n attributes.Q_CREATE_ROLE.format(ROLES[1]),\n ])\ndef test_analyze_schemas_create_schemas(cursor):\n spec = {\n ROLES[0]: {\n 'has_personal_schema': True,\n 'owns': {\n 'schemas': [SCHEMAS[0]]\n },\n },\n ROLES[1]: {\n 'owns': {\n 'schemas': [SCHEMAS[1]],\n },\n },\n 'postgres': {\n 'owns': {\n 'schemas': [\n 'information_schema',\n 'pg_catalog',\n ]\n },\n },\n }\n actual = own.analyze_schemas(spec, cursor, verbose=False)\n\n expected = set([\n own.Q_CREATE_SCHEMA.format(ROLES[0], ROLES[0]),\n own.Q_CREATE_SCHEMA.format(SCHEMAS[0], ROLES[0]),\n own.Q_CREATE_SCHEMA.format(SCHEMAS[1], ROLES[1]),\n ])\n assert set(actual) == expected\n\n\ndef test_get_spec_schemas():\n spec = {\n ROLES[0]: {\n 'has_personal_schema': True,\n 'owns': {\n 'schemas': [SCHEMAS[0]]\n },\n },\n ROLES[1]: {\n 'owns': {\n 'schemas': [SCHEMAS[1]]\n },\n }\n }\n\n assert own.get_spec_schemas(spec) == set([ROLES[0], SCHEMAS[0], SCHEMAS[1]])\n\n\ndef test_init(mockdbcontext):\n mockdbcontext.get_schema_owner = lambda x: 'foo'\n mockdbcontext.get_schema_objects = lambda x: 'bar'\n schemaconf = own.SchemaAnalyzer(rolename=ROLES[0], schema=SCHEMAS[0], dbcontext=mockdbcontext)\n\n assert schemaconf.rolename == ROLES[0]\n assert schemaconf.schema == SCHEMAS[0]\n assert schemaconf.current_owner == 'foo'\n assert schemaconf.exists is True\n assert schemaconf.schema_objects == 'bar'\n\n\ndef test_analyze_create_schema(mockdbcontext):\n schemaconf = own.SchemaAnalyzer(ROLES[0], schema=SCHEMAS[0], dbcontext=mockdbcontext)\n actual = schemaconf.analyze()\n expected = [own.Q_CREATE_SCHEMA.format(SCHEMAS[0], ROLES[0])]\n assert actual == expected\n\n\ndef test_analyze_existing_schema_owner_change(mockdbcontext):\n mockdbcontext.get_schema_owner = lambda x: ROLES[1]\n schemaconf = own.SchemaAnalyzer(ROLES[0], schema=SCHEMAS[0], dbcontext=mockdbcontext)\n changes = schemaconf.analyze()\n assert changes == [own.Q_SET_SCHEMA_OWNER.format(SCHEMAS[0], ROLES[0], ROLES[1])]\n\n\ndef test_analyze_existing_schema_same_owner(mockdbcontext):\n mockdbcontext.get_schema_owner = lambda x: ROLES[0]\n schemaconf = own.SchemaAnalyzer(ROLES[0], schema=SCHEMAS[0], dbcontext=mockdbcontext)\n changes = schemaconf.analyze()\n assert changes == []\n\n\ndef test_analyze_existing_personal_schema_change_object_owners(mockdbcontext):\n mockdbcontext.get_schema_owner = lambda x: ROLES[0]\n mockdbcontext.get_schema_objects = lambda x: [\n ObjectInfo('tables', quoted_object(ROLES[0], TABLES[0]), ROLES[0], False),\n ObjectInfo('sequences', quoted_object(ROLES[0], SEQUENCES[0]), ROLES[0], False),\n ObjectInfo('tables', quoted_object(ROLES[0], TABLES[1]), ROLES[1], False),\n ObjectInfo('sequences', quoted_object(ROLES[0], SEQUENCES[1]), ROLES[1], False),\n ]\n schema = ROLES[0]\n\n schemaconf = own.SchemaAnalyzer(ROLES[0], schema=schema, dbcontext=mockdbcontext,\n is_personal_schema=True)\n actual = schemaconf.analyze()\n expected = [\n own.Q_SET_OBJECT_OWNER.format('TABLE', quoted_object(ROLES[0], TABLES[1]), ROLES[0], ROLES[1]),\n own.Q_SET_OBJECT_OWNER.format('SEQUENCE', quoted_object(ROLES[0], SEQUENCES[1]), ROLES[0], ROLES[1]),\n ]\n assert actual == expected\n\n\ndef test_create_schema(mockdbcontext):\n schemaconf = own.SchemaAnalyzer(ROLES[0], schema=SCHEMAS[0], dbcontext=mockdbcontext)\n schemaconf.create_schema()\n\n assert schemaconf.sql_to_run == [own.Q_CREATE_SCHEMA.format(SCHEMAS[0], ROLES[0])]\n\n\ndef test_set_owner(mockdbcontext):\n previous_owner = ROLES[1]\n mockdbcontext.get_schema_owner = lambda x: previous_owner\n\n schemaconf = own.SchemaAnalyzer(ROLES[0], schema=SCHEMAS[0], dbcontext=mockdbcontext)\n schemaconf.set_owner()\n\n expected = [own.Q_SET_SCHEMA_OWNER.format(SCHEMAS[0], ROLES[0], previous_owner)]\n assert schemaconf.sql_to_run == expected\n\n\ndef test_alter_object_owner(mockdbcontext):\n previous_owner = ROLES[1]\n owner = ROLES[0]\n schema = SCHEMAS[0]\n table_name = quoted_object(schema, TABLES[0])\n mockdbcontext.get_schema_owner = lambda x: owner\n\n schemaconf = own.SchemaAnalyzer(owner, schema=schema, dbcontext=mockdbcontext)\n schemaconf.alter_object_owner('tables', table_name, previous_owner)\n assert schemaconf.sql_to_run == [own.Q_SET_OBJECT_OWNER.format('TABLE', table_name, owner, previous_owner)]\n\n\ndef test_get_improperly_owned_objects(mockdbcontext):\n mockdbcontext.get_schema_owner = lambda x: ROLES[0]\n mockdbcontext.get_schema_objects = lambda x: [\n # Properly owned\n ObjectInfo('tables', quoted_object(ROLES[0], TABLES[0]), ROLES[0], False),\n ObjectInfo('sequences', quoted_object(ROLES[0], SEQUENCES[0]), ROLES[0], False),\n\n # Improperly owned\n ObjectInfo('tables', quoted_object(ROLES[0], TABLES[1]), ROLES[1], False),\n ObjectInfo('sequences', quoted_object(ROLES[0], SEQUENCES[1]), ROLES[1], False),\n\n # Improperly owned but dependent (i.e. should be skipped)\n ObjectInfo('sequences', quoted_object(ROLES[0], SEQUENCES[2]), ROLES[1], True),\n ]\n schema = ROLES[0]\n\n schemaconf = own.SchemaAnalyzer(rolename=ROLES[0], schema=schema, dbcontext=mockdbcontext,\n is_personal_schema=True)\n\n actual = schemaconf.get_improperly_owned_objects()\n expected = [('tables', quoted_object(schema, TABLES[1]), ROLES[1]),\n ('sequences', quoted_object(schema, SEQUENCES[1]), ROLES[1])]\n assert set(actual) == set(expected)\n", "sub_path": "tests/test_ownerships.py", "file_name": "test_ownerships.py", "file_ext": "py", "file_size_in_byte": 7080, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "pytest.raises", "line_number": 25, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships.analyze_schemas", "line_number": 26, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships", "line_number": 26, "usage_type": "name"}, {"api_name": "pgbedrock.ownerships.UNDOCUMENTED_SCHEMAS_MSG.format", "line_number": 30, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships.UNDOCUMENTED_SCHEMAS_MSG", "line_number": 30, "usage_type": "attribute"}, {"api_name": "pgbedrock.ownerships", "line_number": 30, "usage_type": "name"}, {"api_name": "conftest.run_setup_sql", "line_number": 18, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships.analyze_schemas", "line_number": 60, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships", "line_number": 60, "usage_type": "name"}, {"api_name": "pgbedrock.ownerships.Q_CREATE_SCHEMA.format", "line_number": 63, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships.Q_CREATE_SCHEMA", "line_number": 63, "usage_type": "attribute"}, {"api_name": "pgbedrock.ownerships", "line_number": 63, "usage_type": "name"}, {"api_name": "pgbedrock.ownerships.Q_CREATE_SCHEMA.format", "line_number": 64, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships.Q_CREATE_SCHEMA", "line_number": 64, "usage_type": "attribute"}, {"api_name": "pgbedrock.ownerships", "line_number": 64, "usage_type": "name"}, {"api_name": "pgbedrock.ownerships.Q_CREATE_SCHEMA.format", "line_number": 65, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships.Q_CREATE_SCHEMA", "line_number": 65, "usage_type": "attribute"}, {"api_name": "pgbedrock.ownerships", "line_number": 65, "usage_type": "name"}, {"api_name": "conftest.run_setup_sql", "line_number": 33, "usage_type": "call"}, {"api_name": "pgbedrock.attributes.Q_CREATE_ROLE.format", "line_number": 35, "usage_type": "call"}, {"api_name": "pgbedrock.attributes.Q_CREATE_ROLE", "line_number": 35, "usage_type": "attribute"}, {"api_name": "pgbedrock.attributes", "line_number": 35, "usage_type": "name"}, {"api_name": "pgbedrock.attributes.Q_CREATE_ROLE.format", "line_number": 36, "usage_type": "call"}, {"api_name": "pgbedrock.attributes.Q_CREATE_ROLE", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pgbedrock.attributes", "line_number": 36, "usage_type": "name"}, {"api_name": "pgbedrock.ownerships.get_spec_schemas", "line_number": 85, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships", "line_number": 85, "usage_type": "name"}, {"api_name": "pgbedrock.ownerships.SchemaAnalyzer", "line_number": 91, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships", "line_number": 91, "usage_type": "name"}, {"api_name": "pgbedrock.ownerships.SchemaAnalyzer", "line_number": 101, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships", "line_number": 101, "usage_type": "name"}, {"api_name": "pgbedrock.ownerships.Q_CREATE_SCHEMA.format", "line_number": 103, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships.Q_CREATE_SCHEMA", "line_number": 103, "usage_type": "attribute"}, {"api_name": "pgbedrock.ownerships", "line_number": 103, "usage_type": "name"}, {"api_name": "pgbedrock.ownerships.SchemaAnalyzer", "line_number": 109, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships", "line_number": 109, "usage_type": "name"}, {"api_name": "pgbedrock.ownerships.Q_SET_SCHEMA_OWNER.format", "line_number": 111, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships.Q_SET_SCHEMA_OWNER", "line_number": 111, "usage_type": "attribute"}, {"api_name": "pgbedrock.ownerships", "line_number": 111, "usage_type": "name"}, {"api_name": "pgbedrock.ownerships.SchemaAnalyzer", "line_number": 116, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships", "line_number": 116, "usage_type": "name"}, {"api_name": "pgbedrock.context.ObjectInfo", "line_number": 124, "usage_type": "call"}, {"api_name": "conftest.quoted_object", "line_number": 124, "usage_type": "call"}, {"api_name": "pgbedrock.context.ObjectInfo", "line_number": 125, "usage_type": "call"}, {"api_name": "conftest.quoted_object", "line_number": 125, "usage_type": "call"}, {"api_name": "pgbedrock.context.ObjectInfo", "line_number": 126, "usage_type": "call"}, {"api_name": "conftest.quoted_object", "line_number": 126, "usage_type": "call"}, {"api_name": "pgbedrock.context.ObjectInfo", "line_number": 127, "usage_type": "call"}, {"api_name": "conftest.quoted_object", "line_number": 127, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships.SchemaAnalyzer", "line_number": 131, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships", "line_number": 131, "usage_type": "name"}, {"api_name": "pgbedrock.ownerships.Q_SET_OBJECT_OWNER.format", "line_number": 135, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships.Q_SET_OBJECT_OWNER", "line_number": 135, "usage_type": "attribute"}, {"api_name": "pgbedrock.ownerships", "line_number": 135, "usage_type": "name"}, {"api_name": "conftest.quoted_object", "line_number": 135, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships.Q_SET_OBJECT_OWNER.format", "line_number": 136, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships.Q_SET_OBJECT_OWNER", "line_number": 136, "usage_type": "attribute"}, {"api_name": "pgbedrock.ownerships", "line_number": 136, "usage_type": "name"}, {"api_name": "conftest.quoted_object", "line_number": 136, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships.SchemaAnalyzer", "line_number": 142, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships", "line_number": 142, "usage_type": "name"}, {"api_name": "pgbedrock.ownerships.Q_CREATE_SCHEMA.format", "line_number": 145, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships.Q_CREATE_SCHEMA", "line_number": 145, "usage_type": "attribute"}, {"api_name": "pgbedrock.ownerships", "line_number": 145, "usage_type": "name"}, {"api_name": "pgbedrock.ownerships.SchemaAnalyzer", "line_number": 152, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships", "line_number": 152, "usage_type": "name"}, {"api_name": "pgbedrock.ownerships.Q_SET_SCHEMA_OWNER.format", "line_number": 155, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships.Q_SET_SCHEMA_OWNER", "line_number": 155, "usage_type": "attribute"}, {"api_name": "pgbedrock.ownerships", "line_number": 155, "usage_type": "name"}, {"api_name": "conftest.quoted_object", "line_number": 163, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships.SchemaAnalyzer", "line_number": 166, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships", "line_number": 166, "usage_type": "name"}, {"api_name": "pgbedrock.ownerships.Q_SET_OBJECT_OWNER.format", "line_number": 168, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships.Q_SET_OBJECT_OWNER", "line_number": 168, "usage_type": "attribute"}, {"api_name": "pgbedrock.ownerships", "line_number": 168, "usage_type": "name"}, {"api_name": "pgbedrock.context.ObjectInfo", "line_number": 175, "usage_type": "call"}, {"api_name": "conftest.quoted_object", "line_number": 175, "usage_type": "call"}, {"api_name": "pgbedrock.context.ObjectInfo", "line_number": 176, "usage_type": "call"}, {"api_name": "conftest.quoted_object", "line_number": 176, "usage_type": "call"}, {"api_name": "pgbedrock.context.ObjectInfo", "line_number": 179, "usage_type": "call"}, {"api_name": "conftest.quoted_object", "line_number": 179, "usage_type": "call"}, {"api_name": "pgbedrock.context.ObjectInfo", "line_number": 180, "usage_type": "call"}, {"api_name": "conftest.quoted_object", "line_number": 180, "usage_type": "call"}, {"api_name": "pgbedrock.context.ObjectInfo", "line_number": 183, "usage_type": "call"}, {"api_name": "conftest.quoted_object", "line_number": 183, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships.SchemaAnalyzer", "line_number": 187, "usage_type": "call"}, {"api_name": "pgbedrock.ownerships", "line_number": 187, "usage_type": "name"}, {"api_name": "conftest.quoted_object", "line_number": 191, "usage_type": "call"}, {"api_name": "conftest.quoted_object", "line_number": 192, "usage_type": "call"}]} +{"seq_id": "455395491", "text": "#!/usr/bin/python3\n\"\"\"\nAgents are the basis for (nearly) everything. An agent has a set of controls, some of which are in all agents, others can\nbe defined by classes inheriting from one of the agent classes. \n\nAgents can also have children which are themselves agents, allowing hierarchic structures to be built up. Once an agent is created\nall activities revolve around the controls which the agent has. Higher agents (all the way up to the end user via a web interface)\ncan request that particular controls are updated, and controls can themselves declare changed values which will be passed 'up' the\nhierarchy.\n\nThere are 3 primary types of agent:\n A synchronous agent is 'owned' by a higher agent, and runs in the same process as the higher agent. Most actions will be performed\n synchronously, although this will not necessarily be obvious to the higher agent.\n \n A subprocess agent can be the top of the tree, or a subordinate agent. It runs in a separate process = even in a separate machine.\n All communication is through a pipe or socket type connection. Subprocess agents will normally initiate connection to the 'higher'\n agent. The underlying implementation allows the connections to fail and be recovered automatically.\n \n A stub agent is a special type of synchronous agent that is used to manage the connection with subprocess agent. This allows the \n interface to an agent to appear identical irrespective of whether the agent is in process or out of process.\n \nNormally synchronous agents are instantiated by the agent that owns them, and subprocess agents are initiated independently, either from\nthe command line or run as a daemon.\n\nAgents can set up listeners that will respond to incoming connect requests by informing the agent \n- typically this will spawn a new stub agent.\n\nThe interface provided by an agent is defined by a few methods, which are primarily about the controls that represent the agent's state\nand action capabilities.\n\n setControls - provides a set of control names and the new values that are being requested by an owning agent to a subordinate agent\n \n notifyControls - informs an owning agent that 1 or more controls of a subordinate agent have changed\n \n\"\"\"\n\nimport procman\nimport controls\nimport camlib\nimport socket, time, logging, sys, traceback\n\nreadonlyflagset = {controls.ctrlFlags.readOnly}\n\nclass AgentBase():\n \"\"\"A base class for all agents\n \n Each Agent has a set of controls which define its externally visible state. Methods are provided\n to access and manipulate these controls.\n \n Every agent is expected to have at least these controls:\n aType - the type of the agent\n aLoglevel - the current logging level used for this agent\n and optionally:\n aInstance - the instance name of this agent - if not present (for single instance agents) then aType is used\n aDesc - description of this Agent\n \"\"\"\n def __init__(self, agentType, agentInstance = None, agentDesc=None, parentinfo = None, agentLogLevel = logging.DEBUG):\n \"\"\"creates the base for an agent of any type.\n \"\"\"\n tnow =time.time()\n self.controls = {\n 'aType': controls.genControl('aType', controls.strExtras(),'agent type', agentType, tnow, \"\", readonlyflagset, None)}\n if not agentInstance is None:\n self.controls['aInst'] = controls.genControl(\n 'aInst', controls.strExtras(),'agent instance', agentInstance, tnow, \"\", readonlyflagset, None)\n if not agentDesc is None:\n self.controls['aDesc'] = controls.genControl(\n 'aDesc', controls.strExtras(),'agent description', agentDesc, tnow, \"\", readonlyflagset, None)\n\n if agentLogLevel is None:\n self.lgr = self.parent.pcl\n self.controls['aLogLvl'] = controls.genControl('aLogLvl', controls.intExtras(None, None, None)\n , 'log level', 0, tnow, 0, {controls.ctrlFlags.unavailable}, None)\n else:\n self.lgr = logging.getLogger(\"%s%s\" % (agentType, '' if agentInstance is None else '(%s)' % agentInstance))\n self.controls['aLogLvl'] = controls.genControl('aLogLvl', controls.intExtras(None, None, None)\n , 'log level', agentLogLevel, tnow, agentLogLevel, {}, None)\n \n self.controls.update(self.makeControlset())\n self.pushpends = []\n self.pushtimer = None\n for cid, ctrl in self.controls.items():\n ctrl.setPushAgent(self.pushctrlchange)\n if parentinfo is None:\n self.master = None\n else:\n self.master = procman.ProcHostConnection(self, parentinfo, self.bghmessage,\"baseconn\")\n self.lgr.info(\"%s%s setup complete using AgentBase\" % (agentType, '' if agentInstance is None else '(%s)' % agentInstance))\n \n def makeControlset(self):\n \"\"\"override this method and return a dict of additional controls\"\"\"\n return {}\n self.lgr.warn(\"makeControlset: no additional controls defined\")\n\n def agentType(self):\n return self.controls['aType'].getValue()\n\n def agentName(self):\n return self.controls['aInst'].getValue() if 'aInst' in self.controls else self.controls['aType'].getValue()\n\n def setControls(self, upinf):\n \"\"\"uses a list of lists to update the local controls\"\"\"\n updateresponses = {'updresp':{}}\n allgood = True\n withfails = False\n for c in self.ctrlProcessOrder():\n if c in msg['upd']:\n uresp = self.controls[c].update(msg['upd'][c])\n self.lgr.debug(\"setControls: update %s result %s\" , c, repr(uresp) )\n allgood = allgood and uresp[2]\n if not uresp[0]:\n withfails=True\n updateresponses['updresp'][c] = uresp\n rmsg = \"all updates applied OK\" if allgood else \"some updates failed\" if withfails else \"adjustments to values made\"\n if not allgood:\n self.lgr.warn(\"setControls: %s\" % rmsg)\n return (allgood, withfails, rmsg, updateresponses)\n\n def ctrlProcessOrder(self):\n \"\"\"override this and return the order in which a set of controls should be processed\"\"\"\n self.lgr.warn(\"ctrlProcessOrder: no process control order defined - no updates will be made\")\n return list()\n\n def getControls(self):\n return self.controls\n\n def getControlsAsDicts(self):\n \"\"\"Standard function to get all the controls in handy dict format \"\"\"\n return controls.getControlsetInfo(self.controls)\n\n def addChildAgent(self,newchild):\n \"\"\"for now we'll assume that this isn't a dynamic thing, so no need to report changes up.\n Of course the new agent can be a stub or a local (synchronous) one.\n \"\"\"\n childid = newchild.agentName()\n print(\"XXXXXXXXXXXXXXXXXXXXX \" + str(type(newchild)))\n if not 'children' in self.controls:\n self.controls['children'] = controls.childControl('children', 'child agents', {childid: newchild}, time.time()\n , readonlyflagset)\n else:\n self.controls['children'].getValue()[childid] = newchild\n \n# def goneAgent(self,agentstub):\n# if agentstub.agenttype in self.agents:\n# del self.agents[agentstub.agenttype]\n# self.notify(\"lifecycle\", 'info', agentstub.agenttype\n# , \"agent %s disconnected from %s\" % (agentstub.agenttype, self.agentname))\n\n def childAgent(self,cname):\n if 'children' in self.controls and cname in self.controls['children'].getValue():\n return self.controls['children'].getValue()[cname]\n else:\n return None\n\n def bghmessage(self,msg):\n if 'ping' in msg:\n self.linktester.pingback(msg)\n return\n if msg['cmd'] == 'getControls':\n camlib.makeResponse(True, msg=self.agentname, cmd=msg, extras=self.getControlsAsDicts())\n self.master.sendOb(msg)\n return\n if msg['cmd'] == 'ctrlupdates':\n if '140' in msg['upd']:\n uresp = self.controlset['140'].update(msg['upd']['140'])\n camlib.makeResponse(True, msg = 'link test runnung', cmd=msg\n , extras={'updresp': {'140': uresp}})\n self.master.sendOb(msg)\n return\n else:\n if self.isGood():\n self.pcl.debug('updates: ' + msg.__repr__())\n if 'subagent' in msg:\n children = self.controlset\n else:\n updateresponses = {'updresp':{}}\n allgood = True\n withfails = False\n for c in self.ctrlProcessOrder():\n if c in msg['upd']:\n uresp = self.controlset[c].update(msg['upd'][c])\n self.pcl.debug(\"inmessage: update %s result %s\" , c, repr(uresp) )\n allgood = allgood and uresp[2]\n if not uresp[0]:\n withfails=True\n updateresponses['updresp'][c] = uresp\n rmsg = \"all updates applied OK\" if allgood else \"some updates failed\" if withfails else \"adjustments to values made\"\n updateresponses['state'] = self.getstate()\n camlib.makeResponse(not withfails, msg = rmsg, cmd=msg, extras=updateresponses)\n else:\n camlib.makeResponse(False, msg=\"device unavailable\", cmd=msg)\n self.master.sendOb(msg)\n return\n\n self.oncmd(msg)\n\n\n def pushctrlchange(self, acontrol):\n self.pcl.debug (\"pushy bit for %s to %s\" % (acontrol.id, acontrol.getValue()) )\n pushdelay = 0 if acontrol.flagIsSet(controls.ctrlFlags.autopushimm) \\\n else 0.1 if acontrol.flagIsSet(controls.ctrlFlags.autopushfast) \\\n else 1 if acontrol.flagIsSet(controls.ctrlFlags.autopushmed) \\\n else 10 if acontrol.flagIsSet(controls.ctrlFlags.autopushslow) \\\n else 100\n tnow = time.time()\n if acontrol in self.pushpends:\n self.pcl.debug(\"already in queue\")\n else:\n self.pushpends.append(acontrol)\n self.pcl.debug(\"added to queue\")\n if pushdelay == 0 or pushdelay > (tnow - acontrol.lastpush):\n self.pushbundle()\n else:\n if self.pushtimer is None or self.pushtimer.timeTillTick() > pushdelay:\n if not self.pushtimer is None:\n self.motor.owningclient.callmeback(self.pushtimer, False)\n self.pushtimer = procman.ProcCallback(time.time()+pushdelay, 1000, self.pushtick, None)\n self.callmeback(self.pushtimer, True)\n self.pcl.debug(\"new timer setup for %f\" % pushdelay)\n\n \nclass AgentSync(AgentBase):\n \"\"\"The base class for synchronous agents that are created within an AgentProc. The can be the direct descendant of either\n type of agent.\n \"\"\"\n\n def __init__(self, parent, agentType, agentInstance = None, agentDesc = None, agentLogLevel = logging.DEBUG):\n AgentBase.__init__(self, agentType, agentInstance, agentDesc, None, agentLogLevel)\n self.parent = parent\n self.lgr.info(\"AgentSync setup complete\") \n\nclass AgentProc(AgentBase, procman.ProcClient):\n \"\"\"A base class for subprocess agents. There must only be 1 of these in any subprocess and it is the\n top of the tree within the subprocess.\n \n It can support an outward connect to a higher agent (or any other socket based thing), and receive incoming\n connect requests from subordinate agents running in their own processes either in the same or in other machines.\n \n \"\"\"\n def __init__(self, agentType, agentInstance, agentDesc = None, parentInfo = None, agentLogLevel = logging.DEBUG, logfile=None):\n self.hostname = socket.gethostname()\n lfmt = logging.Formatter(fmt='%(asctime)s %(levelname)s %(name)s: %(message)s', datefmt='%I:%M:%S')\n agname = agentType if agentInstance is None else agentInstance\n if logfile:\n lhdlr = logging.FileHandler(logfile,'w')\n lhdlr.setFormatter(lfmt)\n lhdlr.setLevel(agentLogLevel)\n lgr = logging.getLogger('BasicAgent(%s)' % agname)\n lgr.addHandler(lhdlr)\n lgr.propagate = False\n else:\n lgr = logging.getLogger('ProcClient(%s)' % agname)\n lgr.setLevel(agentLogLevel)\n self.agentname = agname\n \n procman.ProcClient.__init__(self, '%s(%s):' % (agname,self.hostname),lgr)\n AgentBase.__init__(self, agentType, agentInstance, agentDesc, parentInfo, agentLogLevel)\n self.pushpends = []\n self.pushtimer = None\n\n def addListener(self, listenport):\n self.listener = procman.ProcAccepter(self,False\n ,listenport,self.agentConnects, self.agentname + \" listen\")\n\n def agentConnects(self,clientsock):\n AgentStub(self, clientsock)\n self.pcl.info(\"agentConnects\")\n\nclass linktester(controls.genControl):\n def __init__(self, cid, client, pstore):\n super().__init__(cid, controls.clickButtonExtras(), \"link test\", 0, time.time(), 25\n , controls.ctrlFlags.writeOnly, pstore)\n self.owningClient = client\n self.ticker = None\n\n def update(self,upvalue):\n logging.warn(\"XXXXXXXXXX-link test running.\")\n self.writeValue(time.time())\n self.ticker = procman.ProcCallback(time.time(),.15, self.tickme, None)\n self.owningClient.callmeback(self.ticker, True)\n self.ticksleft = 101\n self.ticklog=[]\n return (True, 1, False, \"link performance test\")\n \n def tickme(self, dummy):\n self.ticksleft -= 1\n if self.ticksleft <= 0:\n self.owningClient.callmeback(self.ticker, False)\n self.ticker = None\n for i in range(len(self.ticklog)):\n if i == 0:\n print('0.000, %f, %f' %(self.ticklog[i]['tb']-self.ticklog[i]['ts']\n , self.ticklog[i]['te']-self.ticklog[i]['ts']))\n else:\n print('%f, %f, %f' %(self.ticklog[i]['ts']-self.ticklog[i-1]['ts']\n , self.ticklog[i]['tb']-self.ticklog[i]['ts']\n , self.ticklog[i]['te']-self.ticklog[i]['ts']))\n\n print(self.ticklog)\n \n else:\n self.owningClient.master.sendOb({'ping': self.ticksleft, 'ts': time.time(), 'tb':0, 'te':0})\n \n def pingback(self, pmsg):\n pmsg['te'] = time.time()\n self.ticklog.append(pmsg)\n \n#class BasicAgent(procman.ProcClient):\n# \"\"\" basic agent framework - establishes connection back to base and provides basic separation\n# of messages arriving into:\n# cmd messages - for the main agent\n# log messages - to control the logging being done by the agent\n# stat messages - to control the status report levels that are returned to base\n# \"\"\"\n# def __init__(self, agname, maddress, mport, loglevel, statlevel, oncmd, logfile = None):\n# self.hostname = socket.gethostname()\n# self.agentname = agname\n# lfmt = logging.Formatter(fmt='%(asctime)s %(levelname)s %(name)s: %(message)s', datefmt='%I:%M:%S')\n# if logfile:\n# lhdlr = logging.FileHandler(logfile,'w')\n# lhdlr.setFormatter(lfmt)\n# lhdlr.setLevel(loglevel)\n# lgr = logging.getLogger('BasicAgeng(%s)' % agname)\n# lgr.addHandler(lhdlr)\n# lgr.propagate = False\n## root_logger = logging.getLogger()\n## root_logger.disabled = True\n# else:\n# lgr = logging.getLogger('ProcClient(%s)' % agname)\n# lgr.setLevel(loglevel)\n# super().__init__('%s(%s):' % (agname,self.hostname),lgr)\n# self.setLogLevels(loglevel, loglevel)\n# self.statlevel = statlevel\n# self.controlset = self.makeControlSet()\n# self.linktester = linktester('140', self, None)\n# self.controlset['140'] = self.linktester\n# self.oncmd = oncmd\n# self.agents={} # for any agents below this one in the hierarchy\n# if maddress is None:\n# self.master = None\n# else:\n# self.master = procman.ProcHostConnection(self, (maddress,mport), self.bghmessage,\"baseconn\")\n# self.pushpends = []\n# self.pushtimer = None\n \n# def addListener(self, listenport):\n# self.listener = procman.ProcAccepter(self,False\n# ,listenport,self.agentConnects, self.agentname + \" listen\")\n\n# def agentConnects(self,clientsock):\n# unknownagent(self, clientsock)\n# self.pcl.info(\"agentConnects\")\n \n# def addAgent(self,agentstub):\n# self.agents[agentstub.agenttype] = agentstub\n# self.notify(\"lifecycle\", 'info', agentstub.agenttype\n# , \"agent %s connected to %s\" % (agentstub.agenttype, self.agentname))\n\n# def goneAgent(self,agentstub):\n# if agentstub.agenttype in self.agents:\n# del self.agents[agentstub.agenttype]\n# self.notify(\"lifecycle\", 'info', agentstub.agenttype\n# , \"agent %s disconnected from %s\" % (agentstub.agenttype, self.agentname))\n\n# def addAgentControlsSub(self, agentstub, subid):\n# \"\"\"Adds this agent stub to the 'children' control using the key subid.\n# \n# The children control is created if it does not exist.\n# \n# The special control class xx is used for this control (defined below)\n# \n# The superior agent this procClient is conencted to is notified of the new child if there is\n# an existing connection.\n# \"\"\"\n# if not 'children' in self.controlset:\n# self.controlset['children'] = {agentstub.agenttype if subid is None else subid : agentstub}\n\n# def bghmessage(self,msg):\n# if 'ping' in msg:\n# self.linktester.pingback(msg)\n# return\n# if msg['cmd'] == 'getControls':\n# camlib.makeResponse(True, msg=self.agentname, cmd=msg, extras=self.getControlsAsDicts())\n# self.master.sendOb(msg)\n# return\n# if msg['cmd'] == 'ctrlupdates':\n# if '140' in msg['upd']:\n# uresp = self.controlset['140'].update(msg['upd']['140'])\n# camlib.makeResponse(True, msg = 'link test runnung', cmd=msg\n# , extras={'updresp': {'140': uresp}})\n# self.master.sendOb(msg)\n# return\n# else:\n# if self.isGood():\n# self.pcl.debug('updates: ' + msg.__repr__())\n# if 'subagent' in msg:\n# children = self.controlset\n# else:\n# updateresponses = {'updresp':{}}\n# allgood = True\n# withfails = False\n# for c in self.ctrlProcessOrder():\n# if c in msg['upd']:\n# uresp = self.controlset[c].update(msg['upd'][c])\n# self.pcl.debug(\"inmessage: update %s result %s\" , c, repr(uresp) )\n# allgood = allgood and uresp[2]\n# if not uresp[0]:\n # withfails=True\n # updateresponses['updresp'][c] = uresp\n # rmsg = \"all updates applied OK\" if allgood else \"some updates failed\" if withfails else \"adjustments to values made\"\n # updateresponses['state'] = self.getstate()\n# camlib.makeResponse(not withfails, msg = rmsg, cmd=msg, extras=updateresponses)\n# else:\n# camlib.makeResponse(False, msg=\"device unavailable\", cmd=msg)\n# self.master.sendOb(msg)\n# return\n\n# self.oncmd(msg)\n\n# def notify(self, mtype, mlevel, morigin, message):\n# \"\"\"incoming status from any lower level agents arrive here, and are by default forwarded up the chain.\n# override this to do any local actions based on the message, call super to forward them up as well.\n# \"\"\"\n# self.pcl.debug(\"status message %s (%s), (%s), (%s)\" ,message, mlevel, morigin, message)\n# self.master.sendOb({'statusmsg': {\n# 'mtype': mtype\n# , 'mlevel': mlevel\n# , 'morigin': self.agentname + '.' + morigin\n# ,'message': message}})\n\n# def enablepush(self, acontrol):\n# acontrol.setPushAgent(self.pushctrlchange)\n \n# def pushctrlchange(self, acontrol):\n# self.pcl.debug (\"pushy bit for %s to %s\" % (acontrol.id, acontrol.getValue()) )\n# pushdelay = 0 if acontrol.flagIsSet(controls.ctrlFlags.autopushimm) \\\n# else 0.1 if acontrol.flagIsSet(controls.ctrlFlags.autopushfast) \\\n# else 1 if acontrol.flagIsSet(controls.ctrlFlags.autopushmed) \\\n# else 10 if acontrol.flagIsSet(controls.ctrlFlags.autopushslow) \\\n# else 100\n# tnow = time.time()\n# if acontrol in self.pushpends:\n# self.pcl.debug(\"already in queue\")\n# else:\n# self.pushpends.append(acontrol)\n# self.pcl.debug(\"added to queue\")\n# if pushdelay == 0 or pushdelay > (tnow - acontrol.lastpush):\n# self.pushbundle()\n# else:\n# if self.pushtimer is None or self.pushtimer.timeTillTick() > pushdelay:\n# if not self.pushtimer is None:\n# self.motor.owningclient.callmeback(self.pushtimer, False)\n# self.pushtimer = procman.ProcCallback(time.time()+pushdelay, 1000, self.pushtick, None)\n# self.callmeback(self.pushtimer, True)\n# self.pcl.debug(\"new timer setup for %f\" % pushdelay)\n\n# def pushtick(self,dummy):\n# self.callmeback(self.pushtimer, False)\n# self.pushtimer = None\n# self.pushbundle()\n \n# def pushbundle(self):\n# if self.pushpends:\n# self.pcl.debug(\"pushing outstanding queue ---------\")\n# uset = {k.id: k.getUpdateValue() for k in self.pushpends}\n# self.pcl.debug(uset)\n# self.master.sendOb({'update': uset})\n# self.pushpends.clear()\n \n# def sendStatus(self, stype, level, sttext):\n# \"\"\"sends a standard status message 'up' the chain, originating from here\n# \"\"\"\n# self.master.sendOb({'statusmsg': {'mtype': stype, 'mlevel': level, 'morigin': self.agentname, 'message': sttext}})\n\n# def getControlsAsDicts(self):\n## resu = ({k: v.getControlInfo(None) for k, v in self.controlset.items()})\n## return resu\n# return controls.getControlsetInfo(self.controlset)\n \nclass AgentStub(procman.ProcServerConnection):\n \"\"\"A basic stub agent which manages the connection to the 'real' agent through a socket.\n \n It fires off a 'getControls' command initially and runs initially as a free-standing (unconnected) entity.\n When the response to the getControls arrives it identifies the agent type and calls it's\n owner's addChildAgent Method.\n \"\"\"\n def __init__(self, master, clientsock):\n \"\"\"sets up an initial agent stub\"\"\"\n super().__init__(master, clientsock, self.msgin, \"AgentStub\")\n self.controls = None\n self.sendOb({'cmd':'getControls'})\n self.changedcontrols=[]\n self.agenttype = ''\n self.onUpdateIn = None\n\n def agentName(self):\n return self.controls['aInst']['value'] if 'aInst' in self.controls else self.controls['aType']['value']\n\n def setOnUpdateIn(self, func):\n self.onUpdateIn = func\n\n def getControls(self):\n return self.controls\n \n def getChangedControls(self):\n self.lgr.info(\"getChangedControls....\")\n if self.changedcontrols:\n resp = ((k, self.controls[k]['mlist'][self.controls[k]['value']][1]\n if self.controls[k]['type'] == controls.ctrlTypes.cmenu.value else self.controls[k]['value'])\n for k in self.changedcontrols)\n self.changedcontrols = []\n return resp\n return None\n \n def msgin(self,msg):\n if 'ping' in msg:\n msg['tb'] = time.time()\n self.sendOb(msg)\n elif 'cmd' in msg and msg['cmd'] == 'getControls':\n self.controls = msg['response']['extras']\n self.agentType = self.controls['aInst']['value'] if 'aInst' in self.controls else self.controls['aType']['value']\n self.lgr.info(\"msgin agenttype is %s\" % self.agenttype)\n self.parent.addChildAgent(self)\n elif 'update' in msg:\n self.lgr.debug(\"msgin is update\")\n if not self.controls is None:\n for k,ctrlupd in msg['update'].items():\n actrl = self.controls[k]\n actrl['value'] = ctrlupd['v']\n actrl['valueat'] = ctrlupd['t']\n self.lgr.debug(\"ctrl %s set to %s\", actrl['name'], str(ctrlupd['v']))\n if not k in self.changedcontrols:\n self.changedcontrols.append(k)\n if not self.onUpdateIn is None:\n self.wrappedCall(self.onUpdateIn, msg)\n else:\n self.lgr.info(\"no controls available to update in %s\" % self.agenttype)\n# elif 'statusmsg' in msg:\n# nargs = msg['statusmsg']\n# self.parent.notify(morigin = self.agenttype, **nargs)\n else:\n self.lgr.info(\"WHAT? msgin %s\", msg)\n\n def controlUpdates(self,msg):\n msg['to'] = time.time()\n self.sendOb(msg)\n for ctrlid, newval in msg['upd'].items():\n self.controls[ctrlid]['value'] = newval\n\n def close(self):\n self.parent.goneAgent(self)\n super().close()\n\n def wrappedCall(self, fcall, param):\n try: # wrap the call so we can report errors, and then carry on\n \tfcall(param)\n except KeyboardInterrupt:\n \traise\n except:\n \texc_type, exc_value, exc_traceback = sys.exc_info()\n \tself.parent.pcl.warning(\"wrappedCall: exception in called code - \\n%s\\n%s\" % (\n \t\t\tstr(exc_value), ''.join(traceback.format_tb(exc_traceback))))\n ", "sub_path": "basicagent.py", "file_name": "basicagent.py", "file_ext": "py", "file_size_in_byte": 26639, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "controls.ctrlFlags", "line_number": 42, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 57, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 60, "usage_type": "call"}, {"api_name": "controls.genControl", "line_number": 62, "usage_type": "call"}, {"api_name": "controls.strExtras", "line_number": 62, "usage_type": "call"}, {"api_name": "controls.genControl", "line_number": 64, "usage_type": "call"}, {"api_name": "controls.strExtras", "line_number": 65, "usage_type": "call"}, {"api_name": "controls.genControl", "line_number": 67, "usage_type": "call"}, {"api_name": "controls.strExtras", "line_number": 68, "usage_type": "call"}, {"api_name": "controls.genControl", "line_number": 72, "usage_type": "call"}, {"api_name": "controls.intExtras", "line_number": 72, "usage_type": "call"}, {"api_name": "controls.ctrlFlags", "line_number": 73, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 75, "usage_type": "call"}, {"api_name": "controls.genControl", "line_number": 76, "usage_type": "call"}, {"api_name": "controls.intExtras", "line_number": 76, "usage_type": "call"}, {"api_name": "procman.ProcHostConnection", "line_number": 87, "usage_type": "call"}, {"api_name": "controls.getControlsetInfo", "line_number": 129, "usage_type": "call"}, {"api_name": "controls.childControl", "line_number": 138, "usage_type": "call"}, {"api_name": "time.time", "line_number": 138, "usage_type": "call"}, {"api_name": "camlib.makeResponse", "line_number": 160, "usage_type": "call"}, {"api_name": "camlib.makeResponse", "line_number": 166, "usage_type": "call"}, {"api_name": "camlib.makeResponse", "line_number": 189, "usage_type": "call"}, {"api_name": "camlib.makeResponse", "line_number": 191, "usage_type": "call"}, {"api_name": "controls.ctrlFlags", "line_number": 200, "usage_type": "attribute"}, {"api_name": "controls.ctrlFlags", "line_number": 201, "usage_type": "attribute"}, {"api_name": "controls.ctrlFlags", "line_number": 202, "usage_type": "attribute"}, {"api_name": "controls.ctrlFlags", "line_number": 203, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 205, "usage_type": "call"}, {"api_name": "procman.ProcCallback", "line_number": 217, "usage_type": "call"}, {"api_name": "time.time", "line_number": 217, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 227, "usage_type": "attribute"}, {"api_name": "procman.ProcClient", "line_number": 232, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 240, "usage_type": "attribute"}, {"api_name": "socket.gethostname", "line_number": 241, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 242, "usage_type": "call"}, {"api_name": "logging.FileHandler", "line_number": 245, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 248, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 252, "usage_type": "call"}, {"api_name": "procman.ProcClient.__init__", "line_number": 256, "usage_type": "call"}, {"api_name": "procman.ProcClient", "line_number": 256, "usage_type": "attribute"}, {"api_name": "procman.ProcAccepter", "line_number": 262, "usage_type": "call"}, {"api_name": "controls.genControl", "line_number": 269, "usage_type": "attribute"}, {"api_name": "controls.clickButtonExtras", "line_number": 271, "usage_type": "call"}, {"api_name": "time.time", "line_number": 271, "usage_type": "call"}, {"api_name": "controls.ctrlFlags", "line_number": 272, "usage_type": "attribute"}, {"api_name": "logging.warn", "line_number": 277, "usage_type": "call"}, {"api_name": "time.time", "line_number": 278, "usage_type": "call"}, {"api_name": "procman.ProcCallback", "line_number": 279, "usage_type": "call"}, {"api_name": "time.time", "line_number": 279, "usage_type": "call"}, {"api_name": "time.time", "line_number": 302, "usage_type": "call"}, {"api_name": "time.time", "line_number": 305, "usage_type": "call"}, {"api_name": "procman.ProcServerConnection", "line_number": 480, "usage_type": "attribute"}, {"api_name": "controls.ctrlTypes", "line_number": 509, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 517, "usage_type": "call"}, {"api_name": "time.time", "line_number": 545, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 560, "usage_type": "call"}, {"api_name": "traceback.format_tb", "line_number": 562, "usage_type": "call"}]} +{"seq_id": "86960237", "text": "#!/usr/bin/python\nimport requests\nfrom bs4 import BeautifulSoup\nimport lxml.html\nimport pymysql\nimport time\n\ndef main():\n\n conn = pymysql.connect(host='localhost', user='scraper', password='tiger', db='scraping', charset='utf8')\n\n try:\n with conn.cursor() as curs:\n sql = \"\"\"insert into naver_article(title, url, dt, category, rank, pv) values (%s, %s, %s, %s, %s, %s)\"\"\"\n\n session = requests.session()\n response = session.get('http://cis.kbs.co.kr/cis/favorite_naver_mobile-3.html')\n root = lxml.html.fromstring(response.content)\n root.make_links_absolute(response.url)\n\n time.sleep(2)\n\n dt = root.cssselect('input')[0].get('value')\n\n time.sleep(2)\n \n for row in root.cssselect('#ntabs-1 tr'):\n for cell in row.cssselect('td:nth-child(3)'):\n if cell.text_content() == 'KBS':\n curs.execute(sql,(\n row.cssselect('td:nth-child(4)')[0].text_content(), #title\n row.cssselect('a')[1].get('href'), #url\n dt, #date\n row.cssselect('td:nth-child(2)')[0].text_content(), #category\n int(row.cssselect('td:nth-child(1)')[0].text_content()), #rank\n int(row.cssselect('td:nth-child(5)')[0].text_content().replace(',','')) #pv\n #row.cssselect('td:nth-child(5)')[0].text_content() #pv\n ))\n conn.commit()\n\n with conn.cursor() as curs:\n sql = \"select * from naver_article order by dt desc\"\n curs.execute(sql)\n rs = curs.fetchall()\n for row in rs:\n print(row)\n\n finally:\n conn.close()\n\nif __name__ == '__main__':\n main()\n", "sub_path": "naverScraping.py", "file_name": "naverScraping.py", "file_ext": "py", "file_size_in_byte": 2366, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pymysql.connect", "line_number": 10, "usage_type": "call"}, {"api_name": "requests.session", "line_number": 16, "usage_type": "call"}, {"api_name": "lxml.html.html.fromstring", "line_number": 18, "usage_type": "call"}, {"api_name": "lxml.html.html", "line_number": 18, "usage_type": "attribute"}, {"api_name": "lxml.html", "line_number": 18, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 21, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 25, "usage_type": "call"}]} +{"seq_id": "303817825", "text": "\nimport copy\nimport time \nimport logging\nimport collections\n\nfrom PyQt5 import QtCore, QtGui, QtWidgets\n\n\nclass Item(QtGui.QStandardItem):\n \"\"\"\n The Item class used for all filter tree steps. \n The item can take any of the following types:\n - Generic Item (unused)\n - Input Item\n - Filter Item\n - Group Item\n - Modifier Item\n\n All except the Generic Item type can be automatically created using the\n respective factory methods. \n\n All Item data is stored in the data structure of `QStandardItem` and\n can be accessed directly as attributes of the `Item` instance. \n \"\"\"\n\n #TYPE constants: use Item.type() to return item type\n FILTER_TYPE = QtGui.QStandardItem.UserType + 1\n MODIFIER_TYPE = QtGui.QStandardItem.UserType + 2\n GROUP_TYPE = QtGui.QStandardItem.UserType + 3\n INPUT_TYPE = QtGui.QStandardItem.UserType + 10\n\n #DATA ROLE constants: use with Item.setData(value, role) or Item.data(role)\n TYPE = QtCore.Qt.UserRole + 100\n NAME = QtCore.Qt.UserRole + 200\n IS_PROCESSED = QtCore.Qt.UserRole + 500\n HAS_PROCESSING_ERROR = QtCore.Qt.UserRole + 501\n STATUS_MESSAGE = QtCore.Qt.UserRole + 502\n OUTPUT = QtCore.Qt.UserRole + 600\n FN = QtCore.Qt.UserRole + 700\n PARAMS = QtCore.Qt.UserRole + 900\n ID = QtCore.Qt.UserRole + 1000\n\n def __init__(self):\n super().__init__()\n\n self.name = \"\"\n self.full_name = \"\"\n self.description = \"\"\n self.fn = None\n self.params = None\n self.is_active = True\n self.is_processed = False\n self.has_processing_error = False\n self.status_message = \"Not processed\"\n self.output = None\n self.id = str(time.time()) #Item id is the current time, converted to string. This ensures uniqueness\n\n def __getattribute__(self, name):\n if name == 'name':\n return self.data(self.NAME)\n elif name == 'full_name':\n return self.data(QtCore.Qt.DisplayRole)\n elif name == 'description':\n return self.data(QtCore.Qt.ToolTipRole)\n elif name == 'fn':\n return self.data(self.FN)\n elif name == 'params':\n return self.data(self.PARAMS)\n elif name == 'is_active':\n value = self.data(QtCore.Qt.CheckStateRole)\n return True if value == QtCore.Qt.Checked else False\n elif name == 'is_processed':\n return self.data(self.IS_PROCESSED)\n elif name == 'has_processing_error':\n return self.data(self.HAS_PROCESSING_ERROR)\n elif name == 'status_message':\n return self.data(self.STATUS_MESSAGE)\n elif name == 'output':\n return self.data(self.OUTPUT)\n elif name == 'id':\n return self.data(self.ID)\n elif name == 'icon':\n return Item._getIcon(self)\n else:\n return super().__getattribute__(name)\n\n def __setattr__(self, name, value):\n if name == 'name':\n self.setData(value, self.NAME)\n elif name == 'full_name':\n self.setData(value, QtCore.Qt.DisplayRole)\n self.setData(value, QtCore.Qt.EditRole)\n elif name == 'description':\n self.setData(value, QtCore.Qt.ToolTipRole)\n elif name == 'fn':\n self.setData(value, self.FN)\n elif name == 'params': \n self.setData(value, self.PARAMS)\n elif name == 'is_active':\n if type(value) == bool:\n value = QtCore.Qt.Checked if value else QtCore.Qt.Unchecked\n self.setData(value, QtCore.Qt.CheckStateRole)\n elif name == 'is_processed':\n self.setData(value, self.IS_PROCESSED)\n elif name == 'has_processing_error':\n self.setData(value, self.HAS_PROCESSING_ERROR)\n elif name == 'status_message':\n self.setData(value, self.STATUS_MESSAGE)\n elif name == 'output':\n self.setData(value, self.OUTPUT)\n elif name == 'id':\n self.setData(value, self.ID)\n else:\n super().__setattr__(name, value) \n\n def updateParam(self, name, value):\n \"\"\" Update the value of a given parameter. \"\"\"\n params = self.params\n params[name]['value'] = value\n self.params = params\n\n def resetId(self):\n \"\"\" Reset the item's id \"\"\"\n self.id = str(time.time())\n\n def clone(self, keep_id=False, keep_output=False, keep_children_mode='all', keep_children_output=False):\n \"\"\"\n Clone the item.\n\n Parameters\n ----------\n keep_id : bool\n If True, maintain the original item id\n keep_output : bool\n If True, maintain the original item's output. Otherwise output will be None. \n keep_children_mode : str\n Must be one of the following:\n - all: recursively copy all the item's children\n - first: only copy the item's immediate children\n - none: don't copy children\n keep_children_output : bool\n If True, maintain the children's output. \n Only has an effect if `keep_children_mode` is not `none`. \n\n Returns\n -------\n item : Item\n The cloned Item instance\n \"\"\"\n\n item = Item()\n item.name = self.name\n item.full_name = self.full_name\n item.description = self.description\n item.fn = self.fn\n item.params = copy.deepcopy(self.params)\n item.is_active = self.is_active\n\n item.setData(self.type(), self.TYPE)\n item.setData(Item._getIcon(item), QtCore.Qt.DecorationRole)\n item.setFlags(Item._getFlags(item))\n\n if keep_id:\n item.id = self.id\n\n if keep_output:\n item.output = self.output\n \n if keep_children_mode.lower() == 'all':\n for child in self.children():\n child_clone = child.clone(keep_id=keep_id, \n keep_output=keep_children_output, keep_children_mode='all')\n item.appendRow(child_clone)\n\n elif keep_children_mode.lower() == 'first':\n for child in self.children():\n child_clone = child.clone(keep_id=keep_id, \n keep_output=keep_children_output, keep_children_mode='none')\n item.appendRow(child_clone)\n \n else:\n pass\n\n logging.debug(\"Cloned item {} (keep_id={}, keep_output={}, keep_children_mode={}, keep_children_output={})\".format(item.full_name, keep_id, keep_output, keep_children_mode, keep_children_output))\n \n return item\n\n def setParamValueDict(self, params): \n \"\"\" \n Replace all parameter's values with the given values \n \n Parameters\n ----------\n params : dict\n Dictionary with name-value pairs for each parameter\n \"\"\"\n\n for name, value in params.items():\n self.updateParam(name, value)\n\n def getParamValueDict(self):\n \"\"\"\n Return the parameter values dictionary. \n\n Returns\n -------\n params : dict\n Dictionary with name-value pairs for each parameter\n \"\"\"\n\n params = collections.OrderedDict()\n for name, param in self.params.items():\n params[name] = param['value']\n return params\n\n def children(self):\n \"\"\"\n Return iterable of all the item's children. \n\n Returns\n -------\n children : iterable\n All the item's immediate children. \n \"\"\"\n\n if self.hasChildren():\n child_count = self.rowCount()\n for child_i in range(child_count):\n yield self.child(child_i)\n\n def type(self):\n \"\"\" Return the item's type role \"\"\"\n return self.data(self.TYPE)\n\n @classmethod\n def createFilterItem(cls, name, full_name, description=\"\", fn=None, params=collections.OrderedDict()):\n \n item = cls()\n\n item.name = name\n item.full_name = full_name\n item.description = description\n \n item.fn = fn\n if params: \n try: \n item.params = cls._initializeFilterParams(params)\n except ParameterError as e:\n print(\"Parameter error in item {}: \".format(name), e)\n item.params = collections.OrderedDict()\n else:\n item.params = collections.OrderedDict()\n\n item.params['modifier_coefficient'] = {\n 'full_name': 'Modifier Coefficient',\n 'wtype':'double_spinbox',\n 'dtype': float,\n 'description': 'Coefficient to use for this item if it is contained within a modifier',\n 'optional': True,\n 'value': 1.0,\n 'default': 1.0,\n 'minimum': -9999.0,\n 'maximum': 9999.0,\n 'single_step': 0.1,\n 'decimals': 2}\n\n item.setData(cls.FILTER_TYPE, cls.TYPE)\n item.setData(cls._getIcon(item), QtCore.Qt.DecorationRole)\n item.setFlags(cls._getFlags(item))\n \n return item\n\n @classmethod\n def createGroupItem(cls):\n\n item = cls()\n\n item.name = 'group'\n item.full_name = 'Group'\n item.description = \"Processes multiple items in sequence\"\n\n item.params = collections.OrderedDict()\n\n item.params['modifier_coefficient'] = {\n 'full_name': 'Modifier Coefficient',\n 'wtype':'double_spinbox',\n 'dtype': float,\n 'description': 'Coefficient to use for this item if it is contained within a modifier',\n 'optional': True,\n 'value': 1.0,\n 'default': 1.0,\n 'minimum': -9999.0,\n 'maximum': 9999.0,\n 'single_step': 0.1,\n 'decimals': 2}\n\n item.setData(cls.GROUP_TYPE, cls.TYPE)\n item.setData(cls._getIcon(item), QtCore.Qt.DecorationRole)\n item.setFlags(cls._getFlags(item))\n\n return item\n\n @classmethod\n def createModifierItem(cls):\n \n item = cls()\n\n item.name = 'modifier'\n item.full_name = 'Modifier'\n item.description = 'Processes multiple items individually and then combines their results'\n\n item.params = collections.OrderedDict({\n 'mode': {\n 'full_name': 'Modifier Mode',\n 'wtype': 'combobox',\n 'dtype': str,\n 'description': \"Determines whether children of modifier are added or multiplied\",\n 'optional': False,\n 'value': 'add',\n 'default': 'add',\n 'options': ['add', 'multiply'],\n 'options_description': ['Add components', 'Multiply components']},\n 'clip': {\n 'full_name': 'clip',\n 'wtype': 'checkbox',\n 'dtype': bool, \n 'description': \"Clip result to maximum for given dtype, else range will be stretched to dtype range\", \n 'optional': False,\n 'default': True,\n 'value': True},\n 'modifier_coefficient': {\n 'full_name': 'Modifier Coefficient',\n 'wtype': 'double_spinbox',\n 'dtype': float,\n 'description': 'Coefficient to use for this item if it is contained within a modifier',\n 'optional': True,\n 'value': 1.0,\n 'default': 1.0,\n 'minimum': -9999.0,\n 'maximum': 9999.0,\n 'single_step': 0.1,\n 'decimals': 2}})\n\n item.setData(cls.MODIFIER_TYPE, cls.TYPE) \n item.setData(cls._getIcon(item), QtCore.Qt.DecorationRole)\n item.setFlags(cls._getFlags(item))\n\n return item \n\n @classmethod\n def createInputItem(cls,input_path=None):\n\n item = cls()\n \n item.name = 'input_item'\n item.full_name = 'Input Image'\n item.description = \"The tree's input image. Cannot be moved!\"\n\n item.params = collections.OrderedDict(\n {'input_path': {\n 'full_name': 'Input Path',\n 'wtype':'lineedit',\n 'dtype': str,\n 'description': \"Path of the input image\",\n 'optional': False,\n 'value': \"\",\n 'default': \"\",\n },\n 'modifier_coefficient': {\n 'full_name': 'Modifier Coefficient',\n 'wtype':'double_spinbox',\n 'dtype': float,\n 'description': 'Coefficient to use for this item if it is contained within a modifier',\n 'optional': True,\n 'value': 1.0,\n 'default': 1.0,\n 'minimum': -9999.0,\n 'maximum': 9999.0,\n 'single_step': 0.1,\n 'decimals': 2}})\n\n item.setData(cls.INPUT_TYPE, cls.TYPE)\n item.setData(cls._getIcon(item), QtCore.Qt.DecorationRole)\n item.setFlags(cls._getFlags(item))\n\n if input_path:\n item.updateParam('input_path',input_path)\n\n return item\n\n @staticmethod\n def _initializeFilterParams(params):\n \n if not params: \n return collections.OrderedDict()\n else:\n params = collections.OrderedDict(params)\n\n for name, p in params.items():\n if not 'full_name' in p.keys(): p['full_name'] = name\n if not 'optional' in p.keys(): p['options'] = False\n if not 'description' in p.keys(): p['description'] = \"\"\n\n if not 'dtype' in p.keys(): raise ParameterError(\"No dtype provided for parameter {}\".format(name))\n if not p['dtype'] in [float, int, str, bool]: raise ParameterError(\"Invalid dtype for paramter {}\".format(name))\n\n if not 'wtype' in p.keys(): raise ParameterError(\"No wtype provided for parameter {}\".format(name))\n if not p['wtype'] in ['lineedit', 'combobox', 'spinbox', 'double_spinbox', 'checkbox']: raise AttributeError(\"Invalid wtype for parameter {}\".format(name))\n \n if p['wtype'] == 'spinbox':\n if p['dtype'] != int: raise ParameterError(\"dtype for wtype=spinbox must be int in parameter {}\".format(name))\n elif p['wtype'] == 'double_spinbox':\n if p['dtype'] != float: raise ParameterError(\"dtype for wtype=double_spinbox must be float in parameter {}\".format(name))\n\n if not 'default' in p.keys():\n if p['dtype'] == float:\n p['default'] = 0.0\n elif p['dtype'] == int: \n p['default'] = 0\n elif p['dtype'] == str:\n p['default'] = \"\"\n elif p['dtype'] == bool:\n p['default'] = False\n \n if not type(p['default']) == p['dtype']: raise ParameterError(\"Type of default value doesn't match dtype in parameter {}\".format(name))\n \n if p['wtype'] == 'spinbox' or p['wtype'] == 'double_spinbox':\n if 'minimum' in p.keys(): \n if type(p['minimum']) != p['dtype']: raise ParameterError(\"Type of spinbox minimum value doesn't match dtype in parameter {}\".format(name))\n if 'maximum' in p.keys(): \n if type(p['maximum']) != p['dtype']: raise ParameterError(\"Type of spinbox maximum value doesn't match dtype in parameter {}\".format(name))\n if 'single_step' in p.keys(): \n if type(p['single_step']) != p['dtype']: raise ParameterError(\"Type of spinbox single step value doesn't match dtype in parameter {}\".format(name))\n \n if p['wtype'] == 'double_spinbox':\n if not 'decimals' in p.keys():\n p['decimals'] = 2\n\n if p['wtype'] == 'combobox':\n if not 'options' in p.keys(): raise ParameterError(\"No options provided for wtype=combobox in parameter {}\".format(name))\n if 'options_description' in p.keys():\n if len(p['options']) != len(p['options_description']):\n raise ParameterError(\"Length of options list doesn't match length of options_description in parameter {}\".format(name))\n \n if not 'value' in p.keys(): \n p['value'] = p['default']\n\n return params\n \n @staticmethod\n def _getIcon(item):\n\n if item.type() == item.FILTER_TYPE:\n return QtGui.QIcon('resources/filter.png')\n elif item.type() == item.MODIFIER_TYPE:\n return QtGui.QIcon('resources/modifier.png')\n elif item.type() == item.GROUP_TYPE:\n return QtGui.QIcon('resources/folder.png')\n elif item.type() == item.INPUT_TYPE:\n return QtGui.QIcon('resources/input.png')\n else:\n return QtGui.QIcon()\n\n @staticmethod\n def _getFlags(item):\n \n flags = QtCore.Qt.ItemIsSelectable|QtCore.Qt.ItemIsDragEnabled|QtCore.Qt.ItemIsUserCheckable|QtCore.Qt.ItemIsEnabled\n if item.type() == item.FILTER_TYPE:\n flags = flags|QtCore.Qt.ItemNeverHasChildren\n elif item.type() == item.GROUP_TYPE:\n flags = flags|QtCore.Qt.ItemIsDropEnabled|QtCore.Qt.ItemIsEditable\n elif item.type() == item.MODIFIER_TYPE:\n flags = flags|QtCore.Qt.ItemIsDropEnabled\n elif item.type() == item.INPUT_TYPE:\n flags &= ~(QtCore.Qt.ItemIsDragEnabled|QtCore.Qt.ItemIsUserCheckable)\n else: \n return super().flags()\n\n return flags \n\n\nclass ParameterError(Exception):\n pass\n \n", "sub_path": "tree/item_old.py", "file_name": "item_old.py", "file_ext": "py", "file_size_in_byte": 17658, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "PyQt5.QtGui.QStandardItem", "line_number": 10, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui", "line_number": 10, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QStandardItem", "line_number": 28, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui", "line_number": 28, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QStandardItem", "line_number": 29, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui", "line_number": 29, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QStandardItem", "line_number": 30, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui", "line_number": 30, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QStandardItem", "line_number": 31, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui", "line_number": 31, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 34, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 34, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 35, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 35, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 36, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 36, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 37, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 37, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 38, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 38, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 39, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 39, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 40, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 40, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 41, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 41, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 42, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 42, "usage_type": "name"}, {"api_name": "time.time", "line_number": 57, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 63, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 63, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 65, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 65, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 71, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 71, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 72, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 72, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 92, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 92, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 93, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 93, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 95, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 95, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 102, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 102, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 103, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 103, "usage_type": "name"}, {"api_name": "time.time", "line_number": 125, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 157, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 161, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 161, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 185, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 212, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 237, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 251, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 253, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 269, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 269, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 283, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 299, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 299, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 313, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 346, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 346, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 360, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 384, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 384, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 396, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 398, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 455, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 455, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 457, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 457, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 459, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 459, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 461, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 461, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 463, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 463, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 468, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 468, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 470, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 470, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 472, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 472, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 474, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 474, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 476, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 476, "usage_type": "name"}]} +{"seq_id": "537172855", "text": "import copy\nimport pickle\nimport pytz\nimport sys\nimport unittest\n\nfrom datetime import date, datetime, time, timedelta\n\nfrom backports.datetime_fromisoformat import MonkeyPatch\nMonkeyPatch.patch_fromisoformat()\n\n\nclass TestFromIsoFormat(unittest.TestCase):\n def test_basic_naive_parse(self):\n expected = datetime(2014, 2, 5, 23, 45)\n self.assertEqual(expected, datetime.fromisoformat(expected.isoformat()))\n\n\nclass TestsFromCPython(unittest.TestCase):\n # These test cases are taken from cPython's `Lib/test/datetimetester.py`\n\n def test_fromisoformat(self):\n # Test that isoformat() is reversible\n base_dates = [\n (1, 1, 1),\n (1000, 2, 14),\n (1900, 1, 1),\n (2000, 2, 29),\n (2004, 11, 12),\n (2004, 4, 3),\n (2017, 5, 30)\n ]\n\n for date_cls in [datetime, date]:\n for dt_tuple in base_dates:\n dt = date_cls(*dt_tuple)\n dt_str = dt.isoformat()\n with self.subTest(dt_str=dt_str):\n dt_rt = date_cls.fromisoformat(dt.isoformat())\n\n self.assertEqual(dt, dt_rt)\n self.assertIsInstance(dt_rt, date_cls)\n\n def test_fromisoformat_fails(self):\n # Test that fromisoformat() fails on invalid values\n bad_strs = [\n '', # Empty string\n '009-03-04', # Not 10 characters\n '123456789', # Not a date\n '200a-12-04', # Invalid character in year\n '2009-1a-04', # Invalid character in month\n '2009-12-0a', # Invalid character in day\n '2009-01-32', # Invalid day\n '2009-02-29', # Invalid leap day\n '20090228', # Valid ISO8601 output not from isoformat()\n ]\n\n for bad_str in bad_strs:\n with self.assertRaises(ValueError, msg=\"Did not fail on '{0}'\".format(bad_str)):\n datetime.fromisoformat(bad_str)\n\n with self.assertRaises(ValueError, msg=\"Did not fail on '{0}'\".format(bad_str)):\n date.fromisoformat(bad_str)\n\n def test_fromisoformat_fails_typeerror(self):\n # Test that fromisoformat fails when passed the wrong type\n import io\n\n bad_types = [b'2009-03-01', None, io.StringIO('2009-03-01')]\n for bad_type in bad_types:\n with self.assertRaises(TypeError, msg=\"Did not fail on '{0}'\".format(bad_type)):\n datetime.fromisoformat(bad_type)\n\n with self.assertRaises(TypeError, msg=\"Did not fail on '{0}'\".format(bad_type)):\n date.fromisoformat(bad_type)\n\n def test_fromisoformat_datetime(self):\n # Test that isoformat() is reversible\n base_dates = [\n (1, 1, 1),\n (1900, 1, 1),\n (2004, 11, 12),\n (2017, 5, 30)\n ]\n\n base_times = [\n (0, 0, 0, 0),\n (0, 0, 0, 241000),\n (0, 0, 0, 234567),\n (12, 30, 45, 234567)\n ]\n\n separators = [' ', 'T']\n\n tzinfos = [None, pytz.utc,\n pytz.FixedOffset(-5 * 60),\n pytz.FixedOffset(2 * 60)]\n\n dts = [datetime(*(date_tuple + time_tuple), tzinfo=tzi)\n for date_tuple in base_dates\n for time_tuple in base_times\n for tzi in tzinfos]\n\n for dt in dts:\n for sep in separators:\n dtstr = dt.isoformat(sep=sep)\n\n with self.subTest(dtstr=dtstr):\n dt_rt = datetime.fromisoformat(dtstr)\n self.assertEqual(dt, dt_rt)\n\n def test_fromisoformat_timezone(self):\n base_dt = datetime(2014, 12, 30, 12, 30, 45, 217456)\n\n tzoffsets = [\n timedelta(hours=5),\n timedelta(hours=2),\n timedelta(hours=6, minutes=27),\n\n # Our timezone implementation doesn't handle sub-minute offsets.\n # timedelta(hours=12, minutes=32, seconds=30),\n # timedelta(hours=2, minutes=4, seconds=9, microseconds=123456)\n ]\n\n tzoffsets += [-1 * td for td in tzoffsets]\n\n tzinfos = [None, pytz.utc,\n pytz.FixedOffset(0)]\n\n tzinfos += [pytz.FixedOffset(td.total_seconds() / 60) for td in tzoffsets]\n\n for tzi in tzinfos:\n dt = base_dt.replace(tzinfo=tzi)\n dtstr = dt.isoformat()\n\n with self.subTest(tstr=dtstr):\n dt_rt = datetime.fromisoformat(dtstr)\n assert dt == dt_rt, dt_rt\n\n def test_fromisoformat_separators(self):\n separators = [\n ' ', 'T', '\\u007f', # 1-bit widths\n '\\u0080', 'ʁ', # 2-bit widths\n 'ᛇ', '時', # 3-bit widths\n '🐍' # 4-bit widths\n ]\n\n for sep in separators:\n dt = datetime(2018, 1, 31, 23, 59, 47, 124789)\n dtstr = dt.isoformat(sep=sep)\n\n with self.subTest(dtstr=dtstr):\n dt_rt = datetime.fromisoformat(dtstr)\n self.assertEqual(dt, dt_rt)\n\n def test_fromisoformat_ambiguous(self):\n # Test strings like 2018-01-31+12:15 (where +12:15 is not a time zone)\n separators = ['+', '-']\n for sep in separators:\n dt = datetime(2018, 1, 31, 12, 15)\n dtstr = dt.isoformat(sep=sep)\n\n with self.subTest(dtstr=dtstr):\n dt_rt = datetime.fromisoformat(dtstr)\n self.assertEqual(dt, dt_rt)\n\n def test_fromisoformat_timespecs(self):\n if sys.version_info >= (3, 6):\n datetime_bases = [\n (2009, 12, 4, 8, 17, 45, 123456),\n (2009, 12, 4, 8, 17, 45, 0)]\n\n tzinfos = [None, pytz.utc,\n pytz.FixedOffset(-5 * 60),\n pytz.FixedOffset(2 * 60),\n pytz.FixedOffset(6 * 60 + 27)]\n\n timespecs = ['hours', 'minutes', 'seconds', 'milliseconds', 'microseconds']\n\n for ip, ts in enumerate(timespecs):\n for tzi in tzinfos:\n for dt_tuple in datetime_bases:\n if ts == 'milliseconds':\n new_microseconds = 1000 * (dt_tuple[6] // 1000)\n dt_tuple = dt_tuple[0:6] + (new_microseconds,)\n\n dt = datetime(*(dt_tuple[0:(4 + ip)]), tzinfo=tzi)\n dtstr = dt.isoformat(timespec=ts)\n with self.subTest(dtstr=dtstr):\n dt_rt = datetime.fromisoformat(dtstr)\n self.assertEqual(dt, dt_rt)\n\n def test_fromisoformat_fails_datetime(self):\n # Test that fromisoformat() fails on invalid values\n bad_strs = [\n '', # Empty string\n '2009.04-19T03', # Wrong first separator\n '2009-04.19T03', # Wrong second separator\n '2009-04-19T0a', # Invalid hours\n '2009-04-19T03:1a:45', # Invalid minutes\n '2009-04-19T03:15:4a', # Invalid seconds\n '2009-04-19T03;15:45', # Bad first time separator\n '2009-04-19T03:15;45', # Bad second time separator\n '2009-04-19T03:15:4500:00', # Bad time zone separator\n '2009-04-19T03:15:45.2345', # Too many digits for milliseconds\n '2009-04-19T03:15:45.1234567', # Too many digits for microseconds\n\n # Our timezone implementation doesn't mind > 24h offsets\n # '2009-04-19T03:15:45.123456+24:30', # Invalid time zone offset\n # '2009-04-19T03:15:45.123456-24:30', # Invalid negative offset\n\n '2009-04-10ᛇᛇᛇᛇᛇ12:15', # Too many unicode separators\n '2009-04-19T1', # Incomplete hours\n '2009-04-19T12:3', # Incomplete minutes\n '2009-04-19T12:30:4', # Incomplete seconds\n '2009-04-19T12:', # Ends with time separator\n '2009-04-19T12:30:', # Ends with time separator\n '2009-04-19T12:30:45.', # Ends with time separator\n '2009-04-19T12:30:45.123456+', # Ends with timzone separator\n '2009-04-19T12:30:45.123456-', # Ends with timzone separator\n '2009-04-19T12:30:45.123456-05:00a', # Extra text\n '2009-04-19T12:30:45.123-05:00a', # Extra text\n '2009-04-19T12:30:45-05:00a', # Extra text\n ]\n\n for bad_str in bad_strs:\n with self.subTest(bad_str=bad_str):\n with self.assertRaises(ValueError, msg=\"Did not fail on '{0}'\".format(bad_str)):\n datetime.fromisoformat(bad_str)\n\n # Our timezone implementation doesn't have a concept of UTC being special\n # def test_fromisoformat_utc(self):\n # dt_str = '2014-04-19T13:21:13+00:00'\n # dt = datetime.fromisoformat(dt_str)\n # self.assertIs(dt.tzinfo, pytz.utc)\n\n def test_time_fromisoformat(self):\n tsc = time(12, 14, 45, 203745, tzinfo=pytz.utc)\n tsc_rt = time.fromisoformat(tsc.isoformat())\n\n self.assertEqual(tsc, tsc_rt)\n self.assertIsInstance(tsc_rt, time)\n\n\nclass TestCopy(unittest.TestCase):\n def test_basic_pickle_and_copy(self):\n dt = datetime.fromisoformat('2018-11-01 20:42:09.058000')\n dt2 = pickle.loads(pickle.dumps(dt))\n self.assertEqual(dt, dt2)\n dt3 = copy.deepcopy(dt)\n self.assertEqual(dt, dt3)\n\n # FixedOffset\n dt = datetime.fromisoformat('2018-11-01 20:42:09.058000+01:30')\n dt2 = pickle.loads(pickle.dumps(dt))\n self.assertEqual(dt, dt2)\n dt3 = copy.deepcopy(dt)\n self.assertEqual(dt, dt3)\n\n\nif __name__ == '__main__':\n unittest.main()\n", "sub_path": "tests/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 9986, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "backports.datetime_fromisoformat.MonkeyPatch.patch_fromisoformat", "line_number": 10, "usage_type": "call"}, {"api_name": "backports.datetime_fromisoformat.MonkeyPatch", "line_number": 10, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 13, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 15, "usage_type": "call"}, {"api_name": "datetime.datetime.fromisoformat", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 16, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 19, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 34, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 34, "usage_type": "name"}, {"api_name": "datetime.datetime.fromisoformat", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 60, "usage_type": "name"}, {"api_name": "datetime.date.fromisoformat", "line_number": 63, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 63, "usage_type": "name"}, {"api_name": "io.StringIO", "line_number": 69, "usage_type": "call"}, {"api_name": "datetime.datetime.fromisoformat", "line_number": 72, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 72, "usage_type": "name"}, {"api_name": "datetime.date.fromisoformat", "line_number": 75, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 75, "usage_type": "name"}, {"api_name": "pytz.utc", "line_number": 95, "usage_type": "attribute"}, {"api_name": "pytz.FixedOffset", "line_number": 96, "usage_type": "call"}, {"api_name": "pytz.FixedOffset", "line_number": 97, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 99, "usage_type": "call"}, {"api_name": "datetime.datetime.fromisoformat", "line_number": 109, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 109, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 113, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 116, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 117, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 118, "usage_type": "call"}, {"api_name": "pytz.utc", "line_number": 127, "usage_type": "attribute"}, {"api_name": "pytz.FixedOffset", "line_number": 128, "usage_type": "call"}, {"api_name": "pytz.FixedOffset", "line_number": 130, "usage_type": "call"}, {"api_name": "datetime.datetime.fromisoformat", "line_number": 137, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 137, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 149, "usage_type": "call"}, {"api_name": "datetime.datetime.fromisoformat", "line_number": 153, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 153, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 160, "usage_type": "call"}, {"api_name": "datetime.datetime.fromisoformat", "line_number": 164, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 164, "usage_type": "name"}, {"api_name": "sys.version_info", "line_number": 168, "usage_type": "attribute"}, {"api_name": "pytz.utc", "line_number": 173, "usage_type": "attribute"}, {"api_name": "pytz.FixedOffset", "line_number": 174, "usage_type": "call"}, {"api_name": "pytz.FixedOffset", "line_number": 175, "usage_type": "call"}, {"api_name": "pytz.FixedOffset", "line_number": 176, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 187, "usage_type": "call"}, {"api_name": "datetime.datetime.fromisoformat", "line_number": 190, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 190, "usage_type": "name"}, {"api_name": "datetime.datetime.fromisoformat", "line_number": 229, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 229, "usage_type": "name"}, {"api_name": "datetime.time", "line_number": 238, "usage_type": "call"}, {"api_name": "pytz.utc", "line_number": 238, "usage_type": "attribute"}, {"api_name": "datetime.time.fromisoformat", "line_number": 239, "usage_type": "call"}, {"api_name": "datetime.time", "line_number": 239, "usage_type": "name"}, {"api_name": "datetime.time", "line_number": 242, "usage_type": "argument"}, {"api_name": "unittest.TestCase", "line_number": 245, "usage_type": "attribute"}, {"api_name": "datetime.datetime.fromisoformat", "line_number": 247, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 247, "usage_type": "name"}, {"api_name": "pickle.loads", "line_number": 248, "usage_type": "call"}, {"api_name": "pickle.dumps", "line_number": 248, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 250, "usage_type": "call"}, {"api_name": "datetime.datetime.fromisoformat", "line_number": 254, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 254, "usage_type": "name"}, {"api_name": "pickle.loads", "line_number": 255, "usage_type": "call"}, {"api_name": "pickle.dumps", "line_number": 255, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 257, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 262, "usage_type": "call"}]} +{"seq_id": "648779513", "text": "import cv2\nimport numpy as np\nfrom matplotlib import pyplot as plt\nimport IPython.display as ipd\nimport scipy.signal\nfrom scipy import signal, linalg\nfrom scipy.io import wavfile\nimport math\nfrom find_path import solve\n\n\ndef Hist(img):\n row, col = img.shape \n y = np.zeros(256)\n for i in range(0,row):\n for j in range(0,col):\n y[img[i,j]] += 1\n x = np.arange(0,256)\n # plt.bar(x, y, color='b', width=5, align='center', alpha=0.25)\n # plt.show()\n return y\n\ndef variance(h, s, e):\n v = 0\n m = mean(h, s, e)\n w = wieght(h, s, e)\n for i in range(s, e):\n v += ((i - m) **2) * h[i]\n v /= w\n return v\n\ndef wieght(h, s, e):\n w = 0\n for i in range(s, e):\n w += h[i]\n return w\n\n\ndef mean(h, s, e):\n m = 0\n w = wieght(h, s, e)\n for i in range(s, e):\n m += h[i] * i\n \n return m/float(w)\n\ndef countPixel(h):\n cnt = 0\n for i in range(0, len(h)):\n if h[i]>0:\n cnt += h[i]\n return cnt\n\ndef threshold(h, threshold_values):\n cnt = countPixel(h)\n for i in range(1, len(h)):\n vb = variance(h, 0, i)\n wb = wieght(h, 0, i) / float(cnt)\n mb = mean(h, 0, i)\n \n vf = variance(h, i, len(h))\n wf = wieght(h, i, len(h)) / float(cnt)\n mf = mean(h, i, len(h))\n \n V2w = wb * (vb) + wf * (vf)\n V2b = wb * wf * (mb - mf)**2\n \n # fw = open(\"trace.txt\", \"a\")\n # fw.write('T='+ str(i) + \"\\n\")\n\n # fw.write('Wb='+ str(wb) + \"\\n\")\n # fw.write('Mb='+ str(mb) + \"\\n\")\n # fw.write('Vb='+ str(vb) + \"\\n\")\n \n # fw.write('Wf='+ str(wf) + \"\\n\")\n # fw.write('Mf='+ str(mf) + \"\\n\")\n # fw.write('Vf='+ str(vf) + \"\\n\")\n\n # fw.write('within class variance='+ str(V2w) + \"\\n\")\n # fw.write('between class variance=' + str(V2b) + \"\\n\")\n # fw.write(\"\\n\")\n \n if not math.isnan(V2w):\n threshold_values[i] = V2w\n\ndef regenerate_img(img, threshold):\n row, col = img.shape \n y = np.zeros((row, col))\n for i in range(0,row):\n for j in range(0,col):\n if img[i,j] >= threshold:\n y[i,j] = 255\n else:\n y[i,j] = 0\n return y\n\ndef binarize(img, offest = 0):\n threshold_values = {}\n h = [1]\n h = Hist(img.astype(int))\n threshold(h, threshold_values)\n op_thres = get_optimal_threshold(threshold_values)\n res = regenerate_img(img, op_thres + offest)\n return res\n\ndef get_optimal_threshold(threshold_values):\n min_V2w = min(threshold_values.values())\n optimal_threshold = [k for k, v in threshold_values.items() if v == min_V2w]\n return optimal_threshold[0]\n\ndef gkern(size=5, sigma=1.0):\n \"\"\"\n Returns a gaussian kernel with zero mean.\n \n Adapted from:\n https://stackoverflow.com/questions/29731726/how-to-calculate-a-gaussian-kernel-matrix-efficiently-in-numpy\n \n Parameters:\n size - Sidelength of gaussian kernel\n sigma - Value of standard deviation\n \"\"\"\n ax = np.arange(-size // 2 + 1.0, size // 2 + 1.0)\n xx, yy = np.meshgrid(ax, ax)\n kernel = np.exp(-(xx**2 + yy**2) / (2.0 * sigma**2))\n return kernel / np.sum(kernel)\n\ndef rgb2gray(rgb):\n\n r, g, b = rgb[:,:,0], rgb[:,:,1], rgb[:,:,2]\n gray = 0.2989 * r + 0.5870 * g + 0.1140 * b\n \n gray = gray / np.max(gray) * 255\n \n return gray\n\ndef post_porcess(img, whiten_radius = 1, whiten_thres = 3, blacken_radius = 4):\n h, w = img.shape\n ref = img.copy()\n for i in range(h):\n for j in range(w):\n if img[i,j] == 0:\n count = 0\n for i0 in range(max(0, i-whiten_radius), min(h, i + whiten_radius)):\n for j0 in range(max(0, j-whiten_radius), min(w, j + whiten_radius)):\n if not img[i0, j0]:\n count += 1\n if count <= whiten_thres:\n ref[i,j] = 255\n img = ref.copy()\n\n for i in range(h):\n for j in range(w):\n if ref[i,j]:\n flag = False\n for i0 in range(max(0, i-blacken_radius), min(h, i + blacken_radius)):\n if flag:\n break\n for j0 in range(max(0, j-blacken_radius), min(w, j + blacken_radius)):\n if not ref[i0, j0]:\n flag = True\n break\n if flag:\n img[i,j] = 0\n return img\n\n\ndef auto_canny(image, sigma=0.33):\n # compute the median of the single channel pixel intensities\n v = np.median(image)\n \n # apply automatic Canny edge detection using the computed median\n lower = int(max(0, (1.0 - sigma) * v))\n upper = int(min(255, (1.0 + sigma) * v))\n# lower = 100\n# upper = 200\n edged = cv2.Canny(image, lower, upper)\n \n # return the edged image\n return edged\n\ndef find(img, screen_corner, ideal, start, goal):\n img = rgb2gray(img)\n\n\n cv2.imwrite('/home/cc/ee106a/fa19/class/ee106a-adt/ros_workspaces/proj/src/vision/imgs/img.png', img)\n height_0, width_0 = img.shape\n height, width = ideal[3][0], ideal[3][1]\n\n ref = np.ones(img.shape)*255\n for i, j, in screen_corner[0:1]:\n for i0 in range(-20, 20):\n for j0 in range(-20, 20):\n if height_0 > i+i0 >= 0 and width_0 > j + j0 >= 0:\n ref[i+i0,j+j0] = 0\n cv2.imwrite('/home/cc/ee106a/fa19/class/ee106a-adt/ros_workspaces/proj/src/vision/imgs/ref1.png', ref)\n\n ref = np.ones(img.shape)*255\n for i, j, in screen_corner[1:2]:\n for i0 in range(-20, 20):\n for j0 in range(-20, 20):\n if height_0 > i+i0 >= 0 and width_0 > j + j0 >= 0:\n ref[i+i0,j+j0] = 0\n cv2.imwrite('/home/cc/ee106a/fa19/class/ee106a-adt/ros_workspaces/proj/src/vision/imgs/ref2.png', ref)\n\n ref = np.ones(img.shape)*255\n for i, j, in screen_corner[2:3]:\n for i0 in range(-20, 20):\n for j0 in range(-20, 20):\n if height_0 > i+i0 >= 0 and width_0 > j + j0 >= 0:\n ref[i+i0,j+j0] = 0\n cv2.imwrite('/home/cc/ee106a/fa19/class/ee106a-adt/ros_workspaces/proj/src/vision/imgs/ref3.png', ref)\n\n ref = np.ones(img.shape)*255\n for i, j, in screen_corner[3:4]:\n for i0 in range(-20, 20):\n for j0 in range(-20, 20):\n if height_0 > i+i0 >= 0 and width_0 > j + j0 >= 0:\n ref[i+i0,j+j0] = 0\n cv2.imwrite('/home/cc/ee106a/fa19/class/ee106a-adt/ros_workspaces/proj/src/vision/imgs/ref4.png', ref)\n\n\n def average(im, n, x, y):\n count = 0\n total = 0\n for i in range(n):\n for j in range(n):\n if (height > i+x>=0 and width >y+j>=0):\n total += im[i+x, y+j]\n count += 1\n return total/count\n\n M, mask = cv2.findHomography(screen_corner.reshape(-1, 1, 2), ideal.reshape(-1, 1, 2), method = 0)\n\n out = np.zeros((height, width))\n ref = np.zeros((height, width))\n for i in range(height_0):\n for j in range(width_0):\n a = np.dot(M,np.array([i,j,1]))\n a /= a[2]\n a = a.astype(int)\n if (height > a[0]>=0 and width >a[1]>=0):\n out[a[0], a[1]] += img[i,j]\n ref[a[0], a[1]] = 255\n cv2.imwrite('/home/cc/ee106a/fa19/class/ee106a-adt/ros_workspaces/proj/src/vision/imgs/out1.png', out)\n out2 = out.copy()\n for i in range(height):\n for j in range(width):\n if not ref[i,j]:\n out2[i,j] = average(out2, 5, i, j)\n\n cv2.imwrite('/home/cc/ee106a/fa19/class/ee106a-adt/ros_workspaces/proj/src/vision/imgs/out2.png', out2)\n\n kern = gkern(5, 5)\n\n\n blurred = scipy.signal.convolve2d(out2, kern, mode=\"same\")\n blurred = blurred/ np.max(blurred) * 255\n\n\n ref = binarize(blurred, 40)\n print(start, goal)\n i,j = start\n for i0 in range(-5, 5):\n for j0 in range(-5, 5):\n if height > i+i0 >= 0 and width > j + j0 >= 0:\n ref[i+i0,j+j0] = 0\n cv2.imwrite('/home/cc/ee106a/fa19/class/ee106a-adt/ros_workspaces/proj/src/vision/imgs/start2.png', ref)\n\n ref = np.ones(blurred.shape)*255\n print(start, goal)\n i,j = start\n for i0 in range(-5, 5):\n for j0 in range(-5, 5):\n if height > i+i0 >= 0 and width > j + j0 >= 0:\n ref[i+i0,j+j0] = 0\n cv2.imwrite('/home/cc/ee106a/fa19/class/ee106a-adt/ros_workspaces/proj/src/vision/imgs/start.png', ref)\n\n ref = binarize(blurred, 40)\n i, j = goal\n print(goal)\n for i0 in range(-5, 5):\n for j0 in range(-5, 5):\n if height > i+i0 >= 0 and width > j + j0 >= 0:\n ref[i+i0,j+j0] = 0\n cv2.imwrite('/home/cc/ee106a/fa19/class/ee106a-adt/ros_workspaces/proj/src/vision/imgs/goal2.png', ref)\n\n ref = np.ones(blurred.shape)*255\n i, j = goal\n print(goal)\n for i0 in range(-5, 5):\n for j0 in range(-5, 5):\n if height > i+i0 >= 0 and width > j + j0 >= 0:\n ref[i+i0,j+j0] = 0\n cv2.imwrite('/home/cc/ee106a/fa19/class/ee106a-adt/ros_workspaces/proj/src/vision/imgs/goal.png', ref)\n\n j = 0\n path = None\n for i in range(60, -20, -5):\n binary = binarize(blurred, i)\n # cv2.imwrite(str(j)+'binary_' + str(i) +'.png', binary)\n binary = post_porcess(binary)\n # i,j = start\n # for i0 in range(-5, 5):\n # for j0 in range(-5, 5):\n # if height > i+i0 >= 0 and width > j + j0 >= 0:\n # binary[i+i0,j+j0] = 0\n # i,j = goal\n # for i0 in range(-5, 5):\n # for j0 in range(-5, 5):\n # if height > i+i0 >= 0 and width > j + j0 >= 0:\n # binary[i+i0,j+j0] = 0\n\n\n j += 1\n cv2.imwrite('/home/cc/ee106a/fa19/class/ee106a-adt/ros_workspaces/proj/src/vision/imgs/' + str(j)+'binary_post' + str(i) +'.png', binary)\n binary = binary <= 0\n flag = False\n for i in range(-5,5):\n for j in range(-5,5):\n if not binary[start[0] + i][start[1] + j]:\n start0 = np.array([start[0] + i, start[1] + j])\n flag = True\n if not flag:\n print('start not found')\n continue\n flag = False \n for i in range(-10, 10):\n for j in range(-10,10):\n if not binary[goal[0] + i][goal[1] + j]:\n goal0 = np.array([goal[0] + i, goal[1] + j])\n flag = True\n if not flag:\n print('goal not found')\n continue\n \n print(start0, goal0) \n path = solve(binary, start0, goal0)\n if path:\n return path\n\n\n\n\n # blurred = scipy.signal.convolve2d(out2, kern, mode=\"same\")\n # blurred = blurred/ np.max(blurred) * 255\n # binary = binarize(blurred)\n # cv2.imwrite('binary_2.png', binary)\n # paths.append(find_path_bfs(binary))\n\n # binary2 = post_porcess(binary)\n # cv2.imwrite('binary_2_post.png', binary2)\n # paths.append(find_path_bfs(binary2))\n\n # binary = post_porcess(binary, blacken_radius = 2)\n # cv2.imwrite('binary_2_post_2.png', binary)\n # paths.append(find_path_bfs(binary))\n\n\n # blurred = scipy.signal.convolve2d(out2, kern, mode=\"same\")\n # blurred = blurred/ np.max(blurred) * 255\n # binary = binarize(blurred)\n # cv2.imwrite('binary_3.png', binary)\n # paths.append(find_path_bfs(binary))\n\n # binary2 = post_porcess(binary)\n # cv2.imwrite('binary_3_post.png', binary2)\n # paths.append(find_path_bfs(binary2))\n\n # binary = post_porcess(binary, blacken_radius = 2)\n # cv2.imwrite('binary_3_post_2.png', binary)\n # paths.append(find_path_bfs(binary))\n\n # blurred = scipy.signal.convolve2d(out2, kern, mode=\"same\")\n # blurred = blurred/ np.max(blurred) * 255\n # binary = binarize(blurred)\n # cv2.imwrite('binary_4.png', binary)\n # paths.append(find_path_bfs(binary))\n\n # binary2 = post_porcess(binary)\n # cv2.imwrite('binary_4_post.png', binary2)\n # paths.append(find_path_bfs(binary2))\n\n # binary = post_porcess(binary, blacken_radius = 2)\n # cv2.imwrite('binary_4_post_2.png', binary)\n # paths.append(find_path_bfs(binary))\n\n\n\n# img = cv2.imread('Screenshot at 2019-12-10 11_18_43.png')\n# screen_corner = np.array([[610, 650], [800, 600], [600, 950], [750, 1000]])\n# ideal = np.array([[0,0], [100,0],[0,140], [100,140]])\n# find(img, screen_corner, ideal)", "sub_path": "vision/src/image_processing.py", "file_name": "image_processing.py", "file_ext": "py", "file_size_in_byte": 12564, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "numpy.zeros", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 18, "usage_type": "call"}, {"api_name": "math.isnan", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 169, "usage_type": "call"}, {"api_name": "cv2.Canny", "line_number": 176, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 185, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 189, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 195, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 197, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 205, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 211, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 213, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 219, "usage_type": "call"}, {"api_name": "cv2.findHomography", "line_number": 232, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 234, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 235, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 238, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 238, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 244, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 251, "usage_type": "call"}, {"api_name": "scipy.signal.signal.convolve2d", "line_number": 256, "usage_type": "call"}, {"api_name": "scipy.signal.signal", "line_number": 256, "usage_type": "attribute"}, {"api_name": "scipy.signal", "line_number": 256, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 257, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 267, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 269, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 276, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 285, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 287, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 294, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 315, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 321, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 330, "usage_type": "call"}, {"api_name": "find_path.solve", "line_number": 337, "usage_type": "call"}]} +{"seq_id": "144211308", "text": "\"\"\" User Interface (UI) module \"\"\"\nimport common\n\n\ndef get_width_columns(table, title_list):\n \"\"\"\n Find the longest width in table.\n :param table: table to display - text file where are included some information.\n :return: List with width of columns.\n \"\"\"\n number_of_columns = len(table[0])\n file_columns_width = [max([len(data[index]) for data in table])\n for index in range(number_of_columns)]\n\n titles_width = (list(len(title) for title in title_list))\n width_columns = [file_columns_width[index] if file_columns_width[index] >\n titles_width[index] else titles_width[index]\n for index in range(number_of_columns)]\n\n return width_columns\n\n\n\ndef get_position_value_dictionary(table, title_list):\n \"\"\"\n Create a dictionary with position and column width. Is need to **kwargs\n in print table function.\n :return: Dictionary with position:column width\n \"\"\"\n width_columns = get_width_columns(table, title_list)\n number_of_columns = len(width_columns)\n string_positions = [\"pos\" + str(index) for index in range(number_of_columns)]\n position_value = dict(zip(string_positions, width_columns))\n\n return position_value\n\n\ndef get_total_sum_of_width_columns(table, title_list):\n \"\"\"\n Calcualte total sum of width in each column.\n :param table: table: table to display - text file where are included some information.\n :param title_list: title_list: list containing table headers\n :return: Sum of width\n \"\"\"\n width_columns = get_width_columns(table, title_list)\n total_column_lenght = common.sum_values(width_columns) + 1 # due to end in var:string \"|\"\n number_of_columns = len(width_columns)\n PADDINGS = 3\n\n total_width_sum = total_column_lenght + (number_of_columns * PADDINGS)\n return total_width_sum\n\n\ndef print_table(table, title_list):\n \"\"\"\n Prints table with data.\n :param table: table to display - text file where are included some information.\n :param title_list: list containing table headers\n \"\"\"\n dict_pos_value = get_position_value_dictionary(table, title_list)\n total_width_sum = get_total_sum_of_width_columns(table, title_list)\n string = ''.join(['| {:^{' + pos + '}} ' for pos in dict_pos_value.keys()]) + \"|\"\n\n print(\"-\" * total_width_sum)\n print(string.format(*title_list, **dict_pos_value))\n\n print(\"-\" * total_width_sum)\n for record in table:\n print(string.format(*record, **dict_pos_value))\n print(\"-\" * total_width_sum)\n\n\ndef print_result(result, label):\n \"\"\"\n Displays results of the special functions.\n\n Args:\n result: result of the special function (string, list or dict)\n label (str): label of the result\n\n Returns:\n None: This function doesn't return anything it only prints to console.\n \"\"\"\n\n # your code\n\n\ndef print_menu(title, list_options, exit_message):\n \"\"\"\n Displays a menu.\n :param title: menu title\n :param list_options: list of strings - options that will be shown in menu\n :param exit_message: option for back to main menu\n \"\"\"\n print(\"{}:\" .format(title))\n i = 1\n for option in list_options:\n print(\"({}) {}\" .format(i, option))\n i += 1\n print(\"(0) {}\" .format(exit_message))\n\n\ndef get_inputs(list_labels, title):\n \"\"\"\n Gets list of inputs from the user.\n :param list_labels: labels of inputs\n :param title: title of the \"input section\"\n :return: list of data given by the user\n \"\"\"\n print(\"{}:\" .format(title))\n inputs = []\n for label in list_labels:\n answers = input(\"{}: \" .format(label))\n inputs.append(answers)\n\n return inputs\n\n\ndef print_error_message(message):\n \"\"\"\n Displays an error message\n :param message: error message to be displayed\n \"\"\"\n print(\"{}\" .format(message))\n", "sub_path": "ui.py", "file_name": "ui.py", "file_ext": "py", "file_size_in_byte": 3866, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "common.sum_values", "line_number": 46, "usage_type": "call"}]} +{"seq_id": "321176917", "text": "# Copyright The PyTorch Lightning team.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"\nAzure Machine Learning\n----------------------\n\"\"\"\n\n\nimport uuid\nfrom typing import Optional\n\nfrom pytorch_lightning.loggers import MLFlowLogger\n\ntry:\n from azureml.core import Run as AzureMlRun\nexcept ImportError: # pragma: no-cover\n AzureMlOfflineRun = None\n _AZURE_ML_AVAILABLE = False\nelse:\n _AZURE_ML_AVAILABLE = True\n\n\nclass AzureMlLogger(MLFlowLogger):\n r\"\"\"\n Log using `Azure Machine Learning `_.\n Install it with pip:\n\n .. code-block:: bash\n\n pip install mlflow azureml-mlflow\n\n The Azure Machine Learning logger will log to standard output if running in\n offline mode, or to\n `Azure Machine Learning metrics `_\n via\n `MLFlow `_\n if running remotely.\n\n **Online and offline mode**\n\n Example::\n\n from azureml.core import Run\n from pytorch_lightning.loggers import AzureMlLogger\n\n # Optional: this is the default value if no run argument is provided.\n run = Run.get_context()\n azureml_logger = AzureMlLogger(run)\n trainer = Trainer(max_epochs=10, logger=azureml_logger)\n\n Args:\n run: Optionally inject Azure Machine Learning `Run` object directly.\n If this is not provided, default to `Run.get_context()`.\n \"\"\"\n\n def __init__(self, run: Optional[AzureMlRun] = None):\n\n if not _AZURE_ML_AVAILABLE:\n raise ImportError(\n \"You want to use `azureml-sdk` logger which is not installed yet,\"\n \" install it with `pip install azureml-sdk`.\"\n )\n\n if run is None:\n run = AzureMlRun.get_context(allow_offline=True)\n\n try:\n experiment = run.experiment\n tracking_uri = experiment.workspace.get_mlflow_tracking_uri()\n experiment_name = experiment.name\n except AttributeError:\n tracking_uri = None\n experiment_name = str(uuid.uuid4())\n\n super().__init__(experiment_name, tracking_uri)\n", "sub_path": "pl_bolts/loggers/azureml.py", "file_name": "azureml.py", "file_ext": "py", "file_size_in_byte": 2759, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "pytorch_lightning.loggers.MLFlowLogger", "line_number": 35, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 68, "usage_type": "name"}, {"api_name": "azureml.core.Run", "line_number": 68, "usage_type": "name"}, {"api_name": "azureml.core.Run.get_context", "line_number": 77, "usage_type": "call"}, {"api_name": "azureml.core.Run", "line_number": 77, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 85, "usage_type": "call"}]} +{"seq_id": "375723516", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nhttps://colab.research.google.com/github/pytorch/pytorch.github.io/blob/master/assets/hub/facebookresearch_pytorch-gan-zoo_pgan.ipynb\nhttps://github.com/pytorch/pytorch/blob/98362d11ffe81ca48748f6b0e1e417cb81ba5998/torch/hub.py#L330\n the following models only: alexnet, densenet121, densenet169, densenet201,\\\n densenet161, inception_v3, resnet18, resnet34, resnet50, resnet101, resnet152,\\\n resnext50_32x4d, resnext101_32x8d, wide_resnet50_2, wide_resnet101_2, squeezenet1_0,\\\n squeezenet1_1, vgg11, vgg13, vgg16, vgg19, vgg11_bn, vgg13_bn, vgg16_bn, vgg19_bn,\\\n googlenet, shufflenet_v2_x0_5, shufflenet_v2_x1_0, mobilenet_v2\"\n assert _model in ['alexnet', 'densenet121', 'densenet169', 'densenet201', 'densenet161', \n 'inception_v3', 'resnet18', 'resnet34', 'resnet50', 'resnet101', 'resnet152', \n 'resnext50_32x4d', 'resnext101_32x8d', 'wide_resnet50_2', 'wide_resnet101_2',\n 'squeezenet1_0', 'squeezenet1_1', 'vgg11', 'vgg13', 'vgg16', 'vgg19', 'vgg11_bn',\n 'vgg13_bn', 'vgg16_bn', 'vgg19_bn', 'googlenet', 'shufflenet_v2_x0_5', \n 'shufflenet_v2_x1_0', 'mobilenet_v2'],\\\n \"Pretrained models are available for \\\n the following models only: alexnet, densenet121, densenet169, densenet201,\\\n densenet161, inception_v3, resnet18, resnet34, resnet50, resnet101, resnet152,\\\n resnext50_32x4d, resnext101_32x8d, wide_resnet50_2, wide_resnet101_2, squeezenet1_0,\\\n squeezenet1_1, vgg11, vgg13, vgg16, vgg19, vgg11_bn, vgg13_bn, vgg16_bn, vgg19_bn,\\\n googlenet, shufflenet_v2_x0_5, shufflenet_v2_x1_0, mobilenet_v2\"\n\"\"\"\nimport os, json\nimport copy\nfrom pathlib import Path\n\nimport torch\nimport torch.optim as optim\nimport torch.nn as nn\nimport torch.nn.functional as F\nfrom torchvision import datasets, transforms\nfrom torch import hub \n\nfrom mlmodels.util import os_package_root_path, log, path_norm, get_model_uri, path_norm_dict\nMODEL_URI = get_model_uri(__file__)\n\n# CV datasets come in various formats, we should write a dataloader for each dataset\n# I assume that the dataloader (itrator) will be ready and imported from another file\n\n###########################################################################################################\n###########################################################################################################\ndef _train(m, device, train_itr, criterion, optimizer, epoch, max_epoch, imax=1):\n m.train()\n corrects, train_loss = 0.0,0.0\n\n for i,batch in enumerate(train_itr):\n if i >= imax: break\n\n image, target = batch[0], batch[1]\n image, target = image.to(device), target.to(device)\n optimizer.zero_grad()\n logit = m(image)\n \n loss = criterion(logit, target)\n loss.backward()\n optimizer.step()\n \n train_loss += loss.item()\n result = torch.max(logit,1)[1]\n corrects += (result.view(target.size()).data == target.data).sum()\n \n size = len(train_itr)\n train_loss /= size \n accuracy = 100.0 * corrects/size\n \n return train_loss, accuracy\n \ndef _valid(m, device, test_itr, criterion, imax=1):\n m.eval()\n corrects, test_loss = 0.0,0.0\n for i,batch in enumerate(test_itr):\n if i >= imax: break\n \n image, target = batch[0], batch[1]\n image, target = image.to(device), target.to(device)\n \n logit = m(image)\n loss = criterion(logit, target)\n\n \n test_loss += loss.item()\n result = torch.max(logit,1)[1]\n corrects += (result.view(target.size()).data == target.data).sum()\n \n size = len(test_itr)\n test_loss /= size \n accuracy = 100.0 * corrects/size\n \n return test_loss, accuracy\n\ndef _get_device():\n # use GPU if it is available\n device = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n return device\n\ndef get_config_file():\n return path_norm('config/model_tch/Imagecnn.json')\n\n\n\n\n\n###########################################################################################################\n###########################################################################################################\nclass Model:\n def __init__(self, model_pars=None, data_pars=None, compute_pars=None, out_pars=None):\n self.model_pars = copy.deepcopy(model_pars)\n self.compute_pars = copy.deepcopy(compute_pars)\n self.data_pars = copy.deepcopy(data_pars)\n m = model_pars \n\n ### Model Structure ################################\n if model_pars is None :\n self.model = None\n return None\n\n #### Progressive GAN ################################\n if m['repo_uri'] == 'facebookresearch/pytorch_GAN_zoo:hub' :\n #'DCGAN',\n self.model = torch.hub.load(m['repo_uri'],\n m.get('model', 'PGAN'), \n model_name = m.get('model_name', 'celebAHQ-512'),\n pretrained = bool( m.get('pretrained', True)), \n useGPU = compute_pars.get('use_gpu', _get_device()) )\n return None\n \n\n #### Other CNN models ################################\n num_classes = m['num_classes']\n _model = m['model']\n self.model = hub.load( m['repo_uri'], _model, \n # model_name = m.get(\"model_name\", m['model']),\n pretrained = bool( m.get('pretrained', True)),\n # useGPU = m.get('use_gpu',False)\n ) \n\n if num_classes != 1000:\n fc_in_features = self.model.fc.in_features\n self.model.fc = nn.Linear(fc_in_features, num_classes)\n\n\n\n\ndef get_params(param_pars=None, **kw):\n pp = param_pars\n choice = pp['choice']\n config_mode = pp['config_mode']\n data_path = pp['data_path']\n\n if choice == \"json\":\n data_path = path_norm(data_path)\n cf = json.load(open(data_path, mode='r'))\n cf = cf[config_mode]\n\n ####Normalize path : add /models/dataset/\n cf['data_pars'] = path_norm_dict(cf['data_pars'])\n cf['out_pars'] = path_norm_dict(cf['out_pars'])\n\n return cf['model_pars'], cf['data_pars'], cf['compute_pars'], cf['out_pars']\n\n else:\n raise Exception(f\"Not support choice {choice} yet\")\n\n\n\n\n# def get_dataset(data_pars=None, **kw):\n\n# #if data_pars['dataset'] == 'MNIST':\n# # train_loader, valid_loader = get_dataset_mnist_torch(data_pars)\n# # return train_loader, valid_loader \n# from mlmodels.preprocess.generic import get_dataset_torch\n\n# if data_pars['dataset'] :\n# train_loader, valid_loader = get_dataset_torch(data_pars)\n# return train_loader, valid_loader \n\n# else:\n# raise Exception(\"dataset not provided \")\n# return 0\n\n\ndef get_dataset(data_pars=None, **kw):\n\n #if data_pars['dataset'] == 'MNIST':\n # train_loader, valid_loader = get_dataset_mnist_torch(data_pars)\n # return train_loader, valid_loader \n from mlmodels.dataloader import DataLoader\n\n loader = DataLoader(data_pars)\n\n if data_pars['data_info']['dataset'] :\n loader.compute()\n try:\n (train_loader, valid_loader), internal_states = loader.get_data()\n except:\n raise Exception(\"the last Preprocessor have to return (train_loader, valid_loader), internal_states.\")\n \n return train_loader, valid_loader \n\n else:\n raise Exception(\"dataset not provided \")\n return 0\n\n\n\n\ndef fit(model, data_pars=None, compute_pars=None, out_pars=None, **kwargs):\n model0 = model.model\n lr = compute_pars['learning_rate']\n epochs = compute_pars[\"epochs\"]\n criterion = nn.CrossEntropyLoss()\n device = _get_device()\n model0.to(device)\n train_loss = []\n train_acc = []\n test_loss = []\n test_acc = []\n best_test_acc = -1\n \n optimizer = optim.Adam(model0.parameters(), lr=lr)\n train_iter, valid_iter = get_dataset(data_pars)\n\n imax_train = compute_pars.get('max_batch_sample', len(train_iter) )\n imax_valid = compute_pars.get('max_batch_sample', len(valid_iter) )\n\n os.makedirs(out_pars[\"checkpointdir\"], exist_ok=True)\n \n for epoch in range(1, epochs + 1):\n #train loss\n tr_loss, tr_acc = _train(model0, device, train_iter, criterion, optimizer, epoch, epochs, imax=imax_train)\n print( f'Train Epoch: {epoch} \\t Loss: {tr_loss} \\t Accuracy: {tr_acc}')\n\n\n ts_loss, ts_acc = _valid(model0, device, valid_iter, criterion, imax=imax_valid)\n print( f'Train Epoch: {epoch} \\t Loss: {ts_loss} \\t Accuracy: {ts_acc}')\n\n if ts_acc > best_test_acc:\n best_test_acc = ts_acc\n #save paras(snapshot)\n print( f\"model saves at {best_test_acc} accuracy\")\n torch.save(model0.state_dict(), os.path.join(out_pars[\"checkpointdir\"], \"best_accuracy\"))\n\n train_loss.append(tr_loss)\n train_acc.append(tr_acc)\n test_loss.append(ts_loss)\n test_acc.append(ts_acc)\n\n model.model = model0\n return model, None\n\n\ndef predict(model, session=None, data_pars=None, compute_pars=None, out_pars=None, imax = 1, return_ytrue=1):\n ###### Progressive GAN\n if model.model_pars['repo_uri'] == 'facebookresearch/pytorch_GAN_zoo:hub' :\n model0 = model.model \n num_images = compute_pars.get('num_images', 4)\n noise, _ = model0.buildNoiseData(num_images)\n with torch.no_grad():\n generated_images = model0.test(noise)\n\n # let's plot these images using torchvision and matplotlib\n import matplotlib.pyplot as plt\n import torchvision\n grid = torchvision.utils.make_grid(generated_images.clamp(min=-1, max=1), scale_each=True, normalize=True)\n plt.imshow(grid.permute(1, 2, 0).cpu().numpy())\n # plt.show()\n\n os.makedirs(out_pars['path'], exist_ok=True)\n plt.savefig(out_pars['path'] + \"/img_01.png\")\n os.system(\"ls \" + out_pars['path'])\n return 0\n \n\n ###### CNN models\n import numpy as np\n from mlmodels.metrics import metrics_eval\n device = _get_device()\n model = model.model\n _, test_iter = get_dataset(data_pars=data_pars)\n\n # test_iter = get_dataset(data_pars, out_pars)\n y_pred = []\n y_true = []\n for i,batch in enumerate(test_iter):\n if i >= imax: break\n image, target = batch[0], batch[1]\n image = image.to(device)\n logit = model(image)\n predictions = torch.max(logit,1)[1].cpu().numpy()\n y_pred.append(predictions)\n y_true.append(target)\n y_pred = np.vstack(y_pred)[0]\n y_true = np.vstack(y_true)[0]\n\n return y_pred, y_true if return_ytrue else y_pred\n\n\ndef evaluate(model, data_pars=None, compute_pars=None, out_pars=None):\n pass\n\n\ndef save(model, session=None, save_pars=None):\n import pickle\n from mlmodels.util import save_tch\n save2 = copy.deepcopy(save_pars)\n path = path_norm( save_pars['path'] + \"/torch_model/\")\n os.makedirs(Path(path), exist_ok = True)\n\n\n ### Specialized part\n save2['path'] = path\n save_tch(model=model, save_pars=save2)\n\n\n ### Setup Model\n d = {\"model_pars\" : model.model_pars, \n \"compute_pars\": model.compute_pars,\n \"data_pars\" : model.data_pars\n }\n pickle.dump(d, open(path + \"/torch_model_pars.pkl\", mode=\"wb\"))\n log(path, os.listdir(path))\n\n\ndef load(load_pars):\n from mlmodels.util import load_tch\n import pickle\n load_pars2 = copy.deepcopy(load_pars)\n path = path_norm( load_pars['path'] + \"/torch_model/\" )\n\n ### Setup Model\n d = pickle.load( open(path + \"/torch_model_pars.pkl\", mode=\"rb\") )\n model = Model(model_pars= d['model_pars'], compute_pars= d['compute_pars'],\n data_pars= d['data_pars']) \n\n ### Specialized part\n load_pars2['path'] = path\n model2 = load_tch(load_pars2)\n model.model = model2.model\n\n return model\n\n\n\n###########################################################################################################\n###########################################################################################################\ndef test(data_path=\"dataset/\", pars_choice=\"json\", config_mode=\"test\"):\n ### Local test\n\n log(\"#### Loading params ##############################################\")\n param_pars = {\"choice\":pars_choice, \"data_path\":data_path, \"config_mode\": config_mode}\n model_pars, data_pars, compute_pars, out_pars = get_params(param_pars)\n log( data_pars, out_pars )\n\n log(\"#### Loading dataset #############################################\")\n xtuple = get_dataset(data_pars)\n\n\n log(\"#### Model init, fit #############################################\")\n session = None\n model = Model(model_pars, data_pars, compute_pars)\n model, session = fit(model, data_pars, compute_pars, out_pars)\n\n\n log(\"#### Predict #####################################################\")\n ypred = predict(model, session, data_pars, compute_pars, out_pars)\n\n\n log(\"#### metrics #####################################################\")\n metrics_val = evaluate(model, data_pars, compute_pars, out_pars)\n print(metrics_val)\n\n\n log(\"#### Plot ########################################################\")\n\n\n log(\"#### Save #########################################################\")\n save_pars = { \"path\": out_pars[\"path\"] }\n save(model=model, save_pars=save_pars)\n\n\n log(\"#### Load ########################################################\")\n model2 = load( save_pars )\n # ypred = predict(model2, data_pars=data_pars, compute_pars=compute_pars, out_pars=out_pars)\n print(model2)\n\n\n\ndef test2(data_path=\"dataset/\", pars_choice=\"json\", config_mode=\"test\"):\n ### Local test\n\n log(\"#### Loading params ##############################################\")\n param_pars = {\"choice\":pars_choice, \"data_path\":data_path, \"config_mode\": config_mode}\n model_pars, data_pars, compute_pars, out_pars = get_params(param_pars)\n log( data_pars, out_pars )\n\n log(\"#### Loading dataset #############################################\")\n #xtuple = get_dataset(data_pars)\n\n\n log(\"#### Model init, fit #############################################\")\n session = None\n model = Model(model_pars, data_pars, compute_pars)\n #model, session = fit(model, data_pars, compute_pars, out_pars)\n\n\n log(\"#### Predict #####################################################\")\n predict(model, session, data_pars, compute_pars, out_pars)\n\n\n log(\"#### metrics #####################################################\")\n #metrics_val = evaluate(model, data_pars, compute_pars, out_pars)\n #print(metrics_val)\n\n\n log(\"#### Plot ########################################################\")\n\n\n log(\"#### Save/Load ###################################################\")\n save_pars = { \"path\": out_pars[\"path\"] }\n save(model=model, save_pars=save_pars)\n model2 = load( save_pars )\n ypred = predict(model2, data_pars=data_pars, compute_pars=compute_pars, out_pars=out_pars)\n print(model2)\n\n\nif __name__ == \"__main__\":\n\n #### CNN Type\n # test(data_path=\"model_tch/torchhub_cnn_list.json\", pars_choice=\"json\", config_mode=\"resnet18\")\n test(data_path=\"dataset/json/refactor/resnet18_benchmark_mnist.json\", pars_choice=\"json\", config_mode=\"test\")\n\n\n\n #### GAN Type\n # test2(data_path=\"model_tch/torchhub_gan_list.json\", pars_choice=\"json\", config_mode=\"PGAN\")\n test2(data_path=\"dataset/json/refactor/torchhub_cnn_dataloader.json\", pars_choice=\"json\", config_mode=\"test\")\n\n\n\n\n\n\n\n\"\"\"\ndef get_dataset2(data_pars=None, **kw):\n import importlib\n \n from torchvision import datasets, transforms\n data_path = data_pars['data_path']\n train_batch_size = data_pars['train_batch_size']\n test_batch_size = data_pars['test_batch_size']\n try:\n transform=transforms.Compose([\n transforms.Grayscale(num_output_channels=3),\n transforms.ToTensor(),\n transforms.Normalize((0.1307,), (0.3081,))\n ])\n dset = getattr(importlib.import_module(\"torchvision.datasets\"), data_pars[\"dataset\"])\n train_loader = torch.utils.data.DataLoader( dset(data_pars['data_path'], train=True, download=True, transform= transform),\n batch_size=train_batch_size, shuffle=True)\n\n valid_loader = torch.utils.data.DataLoader( dset(data_pars['data_path'], train=False, download=True, transform= transform),\n batch_size=test_batch_size, shuffle=True)\n return train_loader, valid_loader \n except :\n raise Exception(\"Dataset doesn't exist\")\n\"\"\"\n\n\n\"\"\"\ndef get_dataset_mnist_torch(data_pars):\n train_loader = torch.utils.data.DataLoader( datasets.MNIST(data_pars['data_path'], train=True, download=True,\n transform=transforms.Compose([\n transforms.Grayscale(num_output_channels=3),\n transforms.ToTensor(),\n transforms.Normalize((0.1307,), (0.3081,))\n ])),\n batch_size=data_pars['train_batch_size'], shuffle=True)\n\n\n valid_loader = torch.utils.data.DataLoader( datasets.MNIST(data_pars['data_path'], train=False,\n transform=transforms.Compose([\n transforms.Grayscale(num_output_channels=3),\n transforms.ToTensor(),\n transforms.Normalize((0.1307,), (0.3081,))\n ])),\n batch_size=data_pars['test_batch_size'], shuffle=True)\n return train_loader, valid_loader \n\"\"\"\n\n\n\n\"\"\"\ndef load_function(uri_name=\"path_norm\"):\n # Can load remote part \n import importlib\n pkg = uri_name.split(\":\")\n package, name = pkg[0], pkg[1]\n return getattr(importlib.import_module(package), name)\n\n\n\ndef get_dataset_torch(data_pars):\n\n transform = None\n if data_pars.get(\"transform_uri\") :\n transform = load_function( data_pars.get(\"transform_uri\", \"mlmodels.preprocess.image:torch_transform_mnist\" ))()\n \n\n dset = load_function(data_pars.get(\"dataset\", \"torchvision.datasets:MNIST\") )\n\n train_loader = torch.utils.data.DataLoader( dset(data_pars['data_path'], train=True, download=True, transform= transform),\n batch_size=data_pars['train_batch_size'], shuffle=True)\n \n valid_loader = torch.utils.data.DataLoader( dset(data_pars['data_path'], train=False, download=True, transform= transform),\n batch_size=data_pars['train_batch_size'], shuffle=True)\n\n return train_loader, valid_loader \n\"\"\"\n\n", "sub_path": "source/models/atorch/torchhub.py", "file_name": "torchhub.py", "file_ext": "py", "file_size_in_byte": 19067, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "mlmodels.util.get_model_uri", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 82, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 93, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 93, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 93, "usage_type": "attribute"}, {"api_name": "mlmodels.util.path_norm", "line_number": 97, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 107, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 108, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 109, "usage_type": "call"}, {"api_name": "torch.hub.load", "line_number": 120, "usage_type": "call"}, {"api_name": "torch.hub", "line_number": 120, "usage_type": "attribute"}, {"api_name": "torch.hub.load", "line_number": 131, "usage_type": "call"}, {"api_name": "torch.hub", "line_number": 131, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 139, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 139, "usage_type": "name"}, {"api_name": "mlmodels.util.path_norm", "line_number": 151, "usage_type": "call"}, {"api_name": "json.load", "line_number": 152, "usage_type": "call"}, {"api_name": "mlmodels.util.path_norm_dict", "line_number": 156, "usage_type": "call"}, {"api_name": "mlmodels.util.path_norm_dict", "line_number": 157, "usage_type": "call"}, {"api_name": "mlmodels.dataloader.DataLoader", "line_number": 190, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 212, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 212, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 221, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 221, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 227, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 242, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 242, "usage_type": "call"}, {"api_name": "os.path", "line_number": 242, "usage_type": "attribute"}, {"api_name": "torch.no_grad", "line_number": 259, "usage_type": "call"}, {"api_name": "torchvision.utils.make_grid", "line_number": 265, "usage_type": "call"}, {"api_name": "torchvision.utils", "line_number": 265, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 266, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 266, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 269, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 270, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 270, "usage_type": "name"}, {"api_name": "os.system", "line_number": 271, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 290, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 293, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 294, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 306, "usage_type": "call"}, {"api_name": "mlmodels.util.path_norm", "line_number": 307, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 308, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 308, "usage_type": "call"}, {"api_name": "mlmodels.util.save_tch", "line_number": 313, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 321, "usage_type": "call"}, {"api_name": "mlmodels.util.log", "line_number": 322, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 322, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 328, "usage_type": "call"}, {"api_name": "mlmodels.util.path_norm", "line_number": 329, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 332, "usage_type": "call"}, {"api_name": "mlmodels.util.load_tch", "line_number": 338, "usage_type": "call"}, {"api_name": "mlmodels.util.log", "line_number": 350, "usage_type": "call"}, {"api_name": "mlmodels.util.log", "line_number": 353, "usage_type": "call"}, {"api_name": "mlmodels.util.log", "line_number": 355, "usage_type": "call"}, {"api_name": "mlmodels.util.log", "line_number": 359, "usage_type": "call"}, {"api_name": "mlmodels.util.log", "line_number": 365, "usage_type": "call"}, {"api_name": "mlmodels.util.log", "line_number": 369, "usage_type": "call"}, {"api_name": "mlmodels.util.log", "line_number": 374, "usage_type": "call"}, {"api_name": "mlmodels.util.log", "line_number": 377, "usage_type": "call"}, {"api_name": "mlmodels.util.log", "line_number": 382, "usage_type": "call"}, {"api_name": "mlmodels.util.log", "line_number": 392, "usage_type": "call"}, {"api_name": "mlmodels.util.log", "line_number": 395, "usage_type": "call"}, {"api_name": "mlmodels.util.log", "line_number": 397, "usage_type": "call"}, {"api_name": "mlmodels.util.log", "line_number": 401, "usage_type": "call"}, {"api_name": "mlmodels.util.log", "line_number": 407, "usage_type": "call"}, {"api_name": "mlmodels.util.log", "line_number": 411, "usage_type": "call"}, {"api_name": "mlmodels.util.log", "line_number": 416, "usage_type": "call"}, {"api_name": "mlmodels.util.log", "line_number": 419, "usage_type": "call"}]} +{"seq_id": "285703208", "text": "from mpl_toolkits.basemap import Basemap\nimport seaborn as sns\nimport matplotlib as mpl\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom netCDF4 import Dataset \n\n\nplt.figure(figsize=(40,40))\nmap = Basemap(llcrnrlon=95, llcrnrlat=-20, urcrnrlon=165, urcrnrlat=25, ellps='WGS84', resolution='i')\nmap.drawcoastlines(linewidth=0.25)\nmap.drawcountries(linewidth=0.25)\nmap.fillcontinents(color='gray')\nmap.drawmapboundary()\n\nlats = np.load(\"../Data/NPZ/Data_lat.npy\")\nlons = np.load(\"../Data/NPZ/Data_lon.npy\")\nArea = np.load(\"../Data/NPZ/Data_Area.npy\")\nArea_log = np.log(Area)\n\nx,y = map(lons,lats)\nfigure = map.scatter(x, y, s=150, marker='o', alpha=0.5, c=Area_log, edgecolors='none', cmap='jet')\ncaxis = plt.axes([0.83,0.3,0.02,0.4])#manual positioning of the colorbar[x-position, y-position, dx, dy]\ncbar = plt.colorbar(figure, cax=caxis, orientation=\"vertical\") #plot the colorbar\ncbar.ax.tick_params(labelsize=20) #set the size of colorbar labels \nplt.savefig(\"CT_areas.eps\")\nplt.close(\"all\")", "sub_path": "Python/BK/V4/Figs/plot_areas.py", "file_name": "plot_areas.py", "file_ext": "py", "file_size_in_byte": 1001, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 9, "usage_type": "name"}, {"api_name": "mpl_toolkits.basemap.Basemap", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}]} +{"seq_id": "566493556", "text": "# https://leetcode-cn.com/problems/sliding-window-maximum/\n# 双端队列 滑动窗口最大值\n# - 只保留当前滑动窗口中有的元素的索引。\n# - 移除比当前元素小的所有元素,它们不可能是最大的。\nimport collections\n\nnums = [1,3,-1,-3,5,3,6,7]\nk = 3\n\n# https://leetcode-cn.com/problems/sliding-window-maximum/solution/python-deque-by-dangerusswilson/\nq, res = collections.deque(), [] \nres.append(max(nums[:k]))\nfor i in range(len(nums)):\n while q and nums[i] > q[-1]: # 移除队列尾部比 nums[i] 小的\n q.pop()\n q.append(nums[i]) # 加入队列\n if i-k >= 0:\n if nums[i-k] == q[0]:\n q.popleft() # 移除上个窗口的\n res.append(q[0]) # 队列头始终是最大的\nprint (res)\n", "sub_path": "others/sliding_window/239_maxSlidingWindow.py", "file_name": "239_maxSlidingWindow.py", "file_ext": "py", "file_size_in_byte": 803, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "collections.deque", "line_number": 11, "usage_type": "call"}]} +{"seq_id": "554822386", "text": "import os\nimport copy\nimport src.imageclassify as ic\nfrom flask import Flask, jsonify, send_from_directory, render_template, request, session,json\nfrom flask_dropzone import Dropzone\nfrom flask_uploads import UploadSet, configure_uploads, IMAGES, patch_request_class\nfrom werkzeug import secure_filename, FileStorage\n\n# import flask_uploads\n# import flask_dropzone\n# from flask import CORS, cross_origin\n\nUPLOAD_DIRECTORY = \"kaplptreeimages\"\n\napi = Flask(__name__)\napi.secret_key = (\"DG123\")\ndropzone = Dropzone(api)\n# Dropzone settingsapp.config['DROPZONE_UPLOAD_MULTIPLE'] = True\napi.config['DROPZONE_ALLOWED_FILE_CUSTOM'] = True\napi.config['DROPZONE_ALLOWED_FILE_TYPE'] = 'image/*'\napi.config['DROPZONE_REDIRECT_VIEW'] = 'results'\ndestination = ''\n\n# Uploads settings\napi.config['UPLOADED_PHOTOS_DEST'] = os.getcwd() + '/kaplptreeimages/uploads'\nphotos = UploadSet('photos', IMAGES)\nconfigure_uploads(api, photos)\npatch_request_class(api) # set maximum file size, default is 16MB\nfile_srces = []\nclass_output=[None] * 20\nclass_output_2d=[[None] *2]*10\n# CORS(api)\nSWAGGER_URL = '/api/docs'\nAPI_URL = '/static/swagger.json'\napi_result_str = ''\n\n# Call factory function to create our blueprint\n@api.route('/')\n@api.route('/', methods=['GET', 'POST'])\ndef index():\n # list to hold our uploaded image urls\n # set session for image results\n\n if \"file_srces\" not in session:\n session['file_srces'] = []\n # list to hold our uploaded image urls\n # file_srces = session['file_srces']\n if request.method == 'POST':\n file_obj = request.files\n\n for f in file_obj:\n file = request.files.get(f)\n # save the file with to our photos folder\n target = os.path.join(os.getcwd(), \"kaplptreeimages\")\n target = os.path.join(target, \"uploads\")\n print(target)\n if not os.path.isdir(target):\n os.mkdir(target)\n filename = photos.save(file, name=file.filename)\n #filename = photos.save(file)\n\n print(photos.url(filename))\n #file_srces.append(photos.url(filename))\n destination = os.path.join(target, filename)\n\n print(file.filename)\n print(destination)\n #file.save(destination)\n # Load API keys\n api_keys={}\n print('before json- api_keys' )\n with open('./api_keys.json') as data_file:\n api_keys= json.loads(data_file.read())\n\n print(type(api_keys),'- ',api_keys)\n api_result = ic.call_vision_api(destination, api_keys)\n # api_result = ic.call_vision_api(photos.url(filename), api_keys)\n file_srces.append(api_result)\n os.remove(destination)\n api_result_str = json.dumps(api_result, sort_keys=True, indent=4, separators=(',', ': '))\n print(api_result)\n print(api_result[\"images\"])\n classifiers=api_result[\"images\"]\n for clx in classifiers:\n list_classfier=clx[\"classifiers\"]\n print(len(list_classfier))\n for i in range(len(clx[\"classifiers\"])):\n\n classes=list_classfier[0].get(\"classes\")\n print('XXXXDDGGG')\n\n class_output_2d=copy.copy(classes)\n\n for i in range(len(classes)):\n class_output[i]= classes[i].get('class') +' | '+str(classes[i].get('score')*100)+'%'\n\n #for image_class in jsonData:\n # print (image_class.get(\"class\"))\n\n # session['file_srces'] = file_srces\n # return \"uploading...\"\n # return dropzone template on GET request\n\n return render_template('index.html')\n\n\n@api.route('/results')\ndef results():\n print(\" in results\")\n\n # redirect to home if no images to display\n # if \"file_srces\" not in session or session['file_srces'] == []:\n # return render_template('index.html')\n\n # set the file_urls and remove the session variable\n # file_srces = session['file_srces']\n # session.pop('file_srces', None)\n print(*class_output_2d)\n print(*class_output)\n return render_template('success.html', result=file_srces,class_output=[x for x in class_output if x is not None])\n #return render_template('success.html', result=file_srces, class_output=[x for x in class_output_2d if x[0] is not None])\n\n\n@api.route('/info')\ndef info():\n print(\" in info\")\n\nif __name__ == \"__main__\":\n #api.run(host=\"0.0.0.0\")\n api.run(debug=True, port=5020)\n", "sub_path": "src/FileUploadDownloadGooglesearch.py", "file_name": "FileUploadDownloadGooglesearch.py", "file_ext": "py", "file_size_in_byte": 4570, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "flask.Flask", "line_number": 15, "usage_type": "call"}, {"api_name": "flask_dropzone.Dropzone", "line_number": 17, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 25, "usage_type": "call"}, {"api_name": "flask_uploads.UploadSet", "line_number": 26, "usage_type": "call"}, {"api_name": "flask_uploads.IMAGES", "line_number": 26, "usage_type": "argument"}, {"api_name": "flask_uploads.configure_uploads", "line_number": 27, "usage_type": "call"}, {"api_name": "flask_uploads.patch_request_class", "line_number": 28, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 44, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 45, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 48, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 48, "usage_type": "name"}, {"api_name": "flask.request.files", "line_number": 49, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 49, "usage_type": "name"}, {"api_name": "flask.request.files.get", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.request.files", "line_number": 52, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 52, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "flask.json.loads", "line_number": 73, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 73, "usage_type": "name"}, {"api_name": "src.imageclassify.call_vision_api", "line_number": 76, "usage_type": "call"}, {"api_name": "src.imageclassify", "line_number": 76, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 79, "usage_type": "call"}, {"api_name": "flask.json.dumps", "line_number": 80, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 80, "usage_type": "name"}, {"api_name": "copy.copy", "line_number": 92, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 104, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 120, "usage_type": "call"}]} +{"seq_id": "280650204", "text": "import jieba\nfrom wordcloud import WordCloud, ImageColorGenerator\nfrom scipy.misc import imread\n\n# 读取文本\nlyric = ''\nwith open('page.txt', 'r', encoding='utf-8') as f:\n lyric = f.readlines()\n# 分词\nresult = jieba.lcut(str(lyric))\n# 已空格拼接形成一个字符串\nkeywords = ' '.join(result)\nprint(keywords)\n# 读取图片\nimage = imread('timg.jpg')\n# 实例化词云\nwc = WordCloud(background_color='White', max_words=50, mask=image)\n# 绘制词云\nwc.generate(keywords)\n# 收集原图色彩\nimage_color = ImageColorGenerator(image)\n# 使用原图色彩着色词云\nwc.recolor(color_func=image_color)\n# 保存图像\nwc.to_file('output.png')\n", "sub_path": "wdcloud.py", "file_name": "wdcloud.py", "file_ext": "py", "file_size_in_byte": 658, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "jieba.lcut", "line_number": 10, "usage_type": "call"}, {"api_name": "scipy.misc.imread", "line_number": 15, "usage_type": "call"}, {"api_name": "wordcloud.WordCloud", "line_number": 17, "usage_type": "call"}, {"api_name": "wordcloud.ImageColorGenerator", "line_number": 21, "usage_type": "call"}]} +{"seq_id": "243076942", "text": "import torch\r\nfrom torch import nn\r\nfrom torch.nn import functional as F\r\n\r\n\r\n# 사용 채널들 리스트 \r\n# 추후 제거 고민중 굳이 파이썬에서도 제거할 필요가 있을까\r\n# list 0 1 2 3 4 5 6 7 8 9 \r\nCHANNELS = [512,512,512,512,512,256,128, 64, 32, 16]\r\nPIXELS = [ 0, 4, 8, 16, 32, 64,128,256,512,1024]\r\nNOISE_PROB = [0,0.1,0.1,0.1,0.1,0.1,0.1,0.1,0.1, 0.1]\r\nEPSILON = 1e-8\r\nZ_SIZE = 100\r\nMAPPING_UINT = 512\r\n\r\n\r\n\r\n# 생성기 정의\r\nclass Generator(nn.Module) :\r\n def __init__(self, batch_size, block_count = 9) :\r\n super(Generator, self).__init__()\r\n self.mapping = MappingNet()\r\n self.block = nn.ModuleDict()\r\n self.to_RGB = nn.ModuleDict()\r\n\r\n self.base = nn.Parameter(torch.randn(batch_size, CHANNELS[1], PIXELS[1], PIXELS[1]))\r\n\r\n for i in range(1, block_count+1) :\r\n\r\n self.block[str(i)] = GBlock(i)\r\n self.to_RGB[str(i)] = ToRGB(i)\r\n\r\n\r\n def forward(self, z, step, alpha=0):\r\n\r\n x = self.base\r\n w = self.mapping(z)\r\n\r\n # 스텝 수만큼 레이어 반복\r\n for i in range(1, step+1) :\r\n x = self.block[str(i)](x, w, NOISE_PROB[i])\r\n\r\n x = self.to_RGB[str(i)](x)\r\n \r\n # 3채널로 변경\r\n return x \r\n\r\n #######################\r\n # 스무스한 변화를 위한 알파 적용 구현 필요 - 보류\r\n #######################\r\n\r\n #for i in range(1, step) :\r\n # x = self.block[str(i)](x, w, NOISE_PROB[i+1])\r\n #ori = nn.Upsample(scale_factor=2, mode='bilinear')(x)\r\n #new = self.block[str(step)](x, w, NOISE_PROB[i+1])\r\n\r\n #ori = self.to_RGB[str(step)](ori) * alpha\r\n #new = self.to_RGB[str(step)](new) * (1-alpha)\r\n #x = ori + new\r\n\r\n \r\n \r\n \r\n# 생성기 내부 반복블럭 정의, step별 생성가능\r\nclass GBlock(nn.Module) :\r\n def __init__(self, step) :\r\n super(GBlock, self).__init__()\r\n\r\n self.step = step\r\n self.pixel = PIXELS[self.step]\r\n self.prev_channel = CHANNELS[self.step - 1]\r\n self.channel = CHANNELS[self.step]\r\n\r\n self.conv0 = nn.Conv2d(self.prev_channel, self.channel, 3, padding = 1)\r\n self.conv1 = nn.Conv2d(self.channel, self.channel, 3, padding = 1)\r\n\r\n # 현재 스텝에서 사용할 레이어의 크기 미리 저장, 계산 최소화\r\n self.layer_shape = [-1, 2, self.channel, 1, 1]\r\n self.noise_shape = [1, self.channel, self.pixel, self.pixel]\r\n\r\n # StyleMixing을 위해 W 에서 a, b 로 맵핑\r\n layer_size = 2 * self.channel\r\n self.style1 = nn.Linear(MAPPING_UINT, layer_size)\r\n self.style2 = nn.Linear(MAPPING_UINT, layer_size)\r\n\r\n # noise 미리지정 \r\n self.noise1 = nn.Parameter(torch.randn(self.noise_shape))\r\n self.noise2 = nn.Parameter(torch.randn(self.noise_shape))\r\n\r\n # 그냥 업샘플\r\n self.upsample = nn.Upsample(scale_factor=2, mode='bilinear')\r\n\r\n # 리키렐루\r\n self.leaky1 = nn.LeakyReLU(0.2)\r\n self.leaky2 = nn.LeakyReLU(0.2)\r\n \r\n\r\n\r\n def forward(self, x, w, noise_prob) :\r\n\r\n # 업샘플 및 컨볼류션, 첫블록에서는 사용하지않음\r\n if self.step != 1 :\r\n x = self.upsample(x)\r\n x = self.conv0(x)\r\n x = self.leaky1(x)\r\n\r\n ################\r\n # 노이즈 추가 - 추후 방식 변경\r\n ################\r\n noise = self.noise1\r\n x = x + noise * noise_prob\r\n\r\n # 피쳐당 노말라이즈 실행, 배치당이 아님\r\n x = x - torch.mean(x, dim=(2,3), keepdim=True)\r\n p = torch.rsqrt(torch.mean(x**2, dim=(2,3), keepdim=True) + EPSILON) \r\n x = torch.mul(p,x)\r\n\r\n # 생성된 스타일로 ax + b Pixelwise 연산\r\n style = self.style1(w)\r\n style = style.view(self.layer_shape)\r\n x = x * style[:,0] + style[:,1]\r\n\r\n # 위 과정 반복, 모듈화 시킬까 고민중\r\n x = self.conv1(x)\r\n x = self.leaky2(x)\r\n\r\n noise = self.noise2\r\n x = x + noise * noise_prob\r\n\r\n x = x - torch.mean(x, dim=(2,3), keepdim=True)\r\n p = torch.rsqrt(torch.mean(x**2, dim=(2,3), keepdim=True) + EPSILON) \r\n x = torch.mul(p,x)\r\n\r\n style = self.style2(w)\r\n x = x * style[0] + style[1]\r\n\r\n return x\r\n\r\n# 다채널 데이터를 3채널로 변경\r\nclass ToRGB(nn.Module) :\r\n def __init__(self, step) :\r\n super(ToRGB, self).__init__()\r\n self.conv = nn.Conv2d(CHANNELS[step] ,3, 1)\r\n\r\n def forward(self, x):\r\n return self.conv(x)\r\n\r\n# 3채널 데이터를 레이어에 필요한 채널수로 변경\r\nclass FromRGB(nn.Module) :\r\n def __init__(self, step) :\r\n super(FromRGB, self).__init__()\r\n self.conv = nn.Conv2d(3, CHANNELS[step], 1)\r\n\r\n def forward(self, x) :\r\n return self.conv(x)\r\n \r\n# Discriminator 정의\r\nclass Discriminator(nn.Module) :\r\n def __init__(self, block_count = 9) :\r\n super(Discriminator, self).__init__()\r\n self.block = nn.ModuleDict()\r\n self.from_RGB = nn.ModuleDict()\r\n\r\n for i in range(block_count, 0, -1) :\r\n self.from_RGB[str(i)] = FromRGB(i)\r\n self.block[str(i)] = DBlock(i)\r\n\r\n\r\n def forward(self, x, step) :\r\n\r\n #######################\r\n # 스무스한 변화를 위한 알파 적용 구현 필요\r\n #######################\r\n\r\n # 다채널 데이터로 변경, 전체 스텝에서 1회만 필요\r\n x = self.from_RGB[str(step)](x)\r\n\r\n # 블록 반복 실행\r\n for i in range(step, 0, -1) :\r\n x = self.block[str(i)](x)\r\n\r\n return x\r\n\r\n\r\nclass DBlock(nn.Module):\r\n def __init__(self, step):\r\n super(DBlock, self).__init__()\r\n\r\n self.step = step\r\n self.pixel = PIXELS[self.step]\r\n self.channel = CHANNELS[self.step]\r\n self.next_channel = CHANNELS[self.step - 1]\r\n\r\n self.leaky1 = nn.LeakyReLU(0.2)\r\n self.leaky2 = nn.LeakyReLU(0.2)\r\n \r\n\r\n self.stddev = MinibatchStandardDeviation()\r\n\r\n\r\n if self.step != 1 :\r\n self.conv1 = nn.Conv2d(self.channel, self.channel, 3, padding=1)\r\n self.conv2 = nn.Conv2d(self.channel, self.next_channel, 3, padding=1) \r\n self.avgpool = nn.AvgPool2d(2)\r\n\r\n else :\r\n self.conv1 = nn.Conv2d(self.channel+1, self.channel, 3, padding=1)\r\n self.conv2 = nn.Conv2d(self.channel, self.channel, 4, padding=0)\r\n self.fc = nn.Linear(self.next_channel, 1)\r\n\r\n \r\n \r\n def forward(self, x) :\r\n\r\n if self.step == 1 :\r\n # minibatch standard deviation 구현\r\n x = self.stddev(x)\r\n\r\n x = self.conv1(x)\r\n x = self.leaky1(x)\r\n x = self.conv2(x)\r\n x = self.leaky2(x)\r\n\r\n if self.step != 1 :\r\n x = self.avgpool(x)\r\n\r\n else :\r\n\r\n \r\n\r\n x = x.view(x.shape[0], -1)\r\n x = self.fc(x)\r\n\r\n return x\r\n\r\nclass MinibatchStandardDeviation(nn.Module) :\r\n def __init__(self) :\r\n super(MinibatchStandardDeviation, self).__init__()\r\n\r\n def forward(self, x) :\r\n s = x.shape\r\n y = x - x.mean(dim=0, keepdim=True)\r\n y = (y**2).mean(0)\r\n y = torch.sqrt(y + EPSILON)\r\n y = y.mean()\r\n y = y.expand((s[0],1,s[2],s[3]))\r\n x = torch.cat([x, y], 1)\r\n return x\r\n\r\n\r\n# Latent space 맵핑 네트워크 z > w\r\nclass MappingNet(nn.Module) :\r\n def __init__(self) :\r\n super(MappingNet, self).__init__()\r\n\r\n self.dense = nn.ModuleList([nn.Linear(Z_SIZE, MAPPING_UINT)])\r\n\r\n for i in range(7) :\r\n self.dense.append(nn.Linear(MAPPING_UINT, MAPPING_UINT))\r\n\r\n def forward(self, x) :\r\n # 추후 조절을 위해 입력을 따로 받음\r\n # 입력은 (100,1) 의 Unifrom-Dist 벡터\r\n for i in range(8) :\r\n x = self.dense[i](x)\r\n x = nn.ReLU()(x)\r\n\r\n return x\r\n\r\n# 테스트\r\nif __name__ == \"__main__\" :\r\n z = torch.rand(100).cuda()\r\n \r\n g = Generator(9)\r\n g = g.cuda()\r\n d = Discriminator(9)\r\n d = d.cuda()\r\n step = 6\r\n\r\n\r\n y = g(z, step)\r\n print(y.shape)\r\n\r\n z = d(y, step)\r\n print(z.shape)\r\n\r\n\r\n\r\n", "sub_path": "model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 8376, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "torch.nn.Module", "line_number": 19, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.ModuleDict", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 23, "usage_type": "name"}, {"api_name": "torch.nn.ModuleDict", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 24, "usage_type": "name"}, {"api_name": "torch.nn.Parameter", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.randn", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 65, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 65, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 74, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 75, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 75, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 83, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 84, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 84, "usage_type": "name"}, {"api_name": "torch.nn.Parameter", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 87, "usage_type": "name"}, {"api_name": "torch.randn", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.nn.Parameter", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 88, "usage_type": "name"}, {"api_name": "torch.randn", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.nn.Upsample", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 91, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 94, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 95, "usage_type": "name"}, {"api_name": "torch.mean", "line_number": 114, "usage_type": "call"}, {"api_name": "torch.rsqrt", "line_number": 115, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 115, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 116, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 130, "usage_type": "call"}, {"api_name": "torch.rsqrt", "line_number": 131, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 131, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 132, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 140, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 140, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 143, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 143, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 149, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 149, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 152, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 152, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 158, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 158, "usage_type": "name"}, {"api_name": "torch.nn.ModuleDict", "line_number": 161, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 161, "usage_type": "name"}, {"api_name": "torch.nn.ModuleDict", "line_number": 162, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 162, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 185, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 185, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 194, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 194, "usage_type": "name"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 195, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 195, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 202, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 202, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 203, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 203, "usage_type": "name"}, {"api_name": "torch.nn.AvgPool2d", "line_number": 204, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 204, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 207, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 207, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 208, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 208, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 209, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 209, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 236, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 236, "usage_type": "name"}, {"api_name": "torch.sqrt", "line_number": 244, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 247, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 252, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 252, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 256, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 256, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 256, "usage_type": "call"}, {"api_name": "torch.nn.Linear", "line_number": 259, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 259, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 266, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 266, "usage_type": "name"}, {"api_name": "torch.rand", "line_number": 272, "usage_type": "call"}]} +{"seq_id": "321888724", "text": "from minio import Minio\nimport os\nimport logbook\n\n# from ..utils.log_helper import g_log_helper\n# g_log = g_log_helper.make_logger(logbook.INFO)\n\nSERVER_ADDR = os.environ.get('MINIO_SERVER_ADDR', '192.168.135.15:9000')\nACCESS_KEY = os.environ.get('MINIO_ACCESS_KEY', 'minioadmin')\nSECRET_KEY = os.environ.get('MINIO_SECRET_KEY', 'minioadmin@pcl')\nHTTP_PROTOCOL = os.environ.get('MINIO_HTTP_PROTOCOL', 'http')\n\n\nclass MinioHelper(object):\n def __init__(self, access_key=ACCESS_KEY, secret_key=SECRET_KEY, server_addr=SERVER_ADDR):\n self.server_addr = server_addr\n # g_log.info('server_addr: %s' % SERVER_ADDR)\n self.minio_client = Minio(server_addr,\n access_key=access_key,\n secret_key=secret_key,\n secure=False)\n\n def upload_file(self, object_name, file_path=None, file=None, bucket_name='panos', content_type=''):\n ret_dict = {\n 'code': 0,\n 'file_url': '',\n 'code_msg': '',\n }\n try:\n if file is not None:\n etag = self.minio_client.put_object(bucket_name=bucket_name, object_name=object_name, data=file, length=len(file.getvalue()) )\n \n if file_path is not None:\n etag = self.minio_client.fput_object(bucket_name=bucket_name,\n file_path=file_path,\n object_name=object_name,\n content_type=content_type)\n\n ret_dict['etag'] = etag\n ret_dict['public_url'] = HTTP_PROTOCOL + '://' + self.server_addr + '/' + bucket_name + '/' + object_name\n ret_dict['presigned_url'] = self.minio_client.presigned_get_object(bucket_name=bucket_name,\n object_name=object_name)\n return ret_dict\n except Exception as e:\n ret_dict['code'] = 1\n ret_dict['code_msg'] = str(e)\n return ret_dict\n\n def file_exist(self, object_name, bucket_name='panos'):\n try:\n data = self.minio_client.get_object(bucket_name, object_name)\n return True\n except Exception as err:\n # print(err)\n return False\n\n\ndef get_minio_file(url, folder=\"./download\"):\n import requests\n\n r = requests.get(url)\n fn = os.path.join( folder, url.split(\"?\")[0].split(\"/\")[-1] )\n \n with open( fn, \"wb\") as f:\n f.write(r.content)\n \n return fn\n\n\nif __name__ == '__main__':\n minio_helper = MinioHelper()\n # ret_dict = minio_helper.upload_file(file_path='./09005700121902131735110055A.jpg', object_name='09005700121902131735110055A.jpg')\n ret_dict = minio_helper.upload_file(file_path='./9e7a5c-a72d-1625-bed6-b74eb6_15_01005700001312021447154435T_181.jpg', \n object_name='9e7a5c-a72d-1625-bed6-b74eb6_15_01005700001312021447154435T_181.jpg')\n url = ret_dict['public_url']\n get_minio_file(url, '../download')\n\n fn = '9e7a5c-a72d-1625-bed6-b74eb6_15_01005700001312021447154435T_181.jpg'\n minio_helper.file_exist( fn )\n \n# import zipfile\n\n# zip_file = zipfile.ZipFile(\"./download/LSTMAC_input.zip\")\n# zip_file.extractall()\n\n\n", "sub_path": "src/utils/minio_helper.py", "file_name": "minio_helper.py", "file_ext": "py", "file_size_in_byte": 3373, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "os.environ.get", "line_number": 8, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 9, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 10, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 11, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 11, "usage_type": "attribute"}, {"api_name": "minio.Minio", "line_number": 18, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path", "line_number": 62, "usage_type": "attribute"}]} +{"seq_id": "302057378", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nimport glob\nimport time\nfrom mpl_toolkits.mplot3d import Axes3D\n#from keras.models import load_model\n\npath = \"/hdd1/work1/neural/\"\n#factor = np.array([256.,1024.,256.,1024.,256.,1024.,256.,1024.])\n#ans1 = np.empty((0,8))\n#ans2 = np.empty((0,5))\n#pred = np.empty((0,8))\n#model = load_model(path+\"/keras/exp/indirect_norm-7.h5\")\n#\n#p = glob.glob(path+\"data/result/*.npy\")\n#filename = []\n#for i in range(len(p)):\n# temp = p[i].split(\"/\")[-1].split(\"_\")[0]\n# if (temp == \"run0178\" or\n# temp == \"run0179\" or\n# temp == \"run0180\" or\n# temp == \"run0181\" or\n# temp == \"run0182\" or\n# temp == \"run0183\" or\n# temp == \"run0184\" or\n# temp == \"run0185\" or\n# temp == \"run0186\" or\n# temp == \"run0187\" or\n# temp == \"run0188\" or\n# temp == \"run0189\"):\n# continue\n# print(temp)\n# filename.append(temp)\n#\n#start = time.time()\n#for i in range(len(filename)):\n# cell = np.load(path+\"data/track/\"+filename[i]+\"_track.npy\")\n# ans = np.load(path+\"data/result/\"+filename[i]+\"_result.npy\")\n# ans1 = np.append(ans1,ans[:,5:13],axis=0)\n# ans2 = np.append(ans2,ans[:,:5],axis=0)\n# temp = model.predict([cell[:,0:1],cell[:,1:2]])\n# pred = np.append(pred,temp*factor,axis=0)\n# print(filename[i],len(pred))\n#end = time.time()\n#\n#print((end-start)/60.,\" min.\")\n#np.savetxt(\"ans_1_all.dat\",ans1,header=\"avs avc aes aec cvs csc ces cec [pixel]\")\n#np.savetxt(\"ans_2_all.dat\",ans2,header=\"dx theta phi [degree]\")\n#np.savetxt(\"pred_all.dat\",pred,header=\"avs avc aes aec cvs csc ces cec [pixel]\")\n\nans = np.loadtxt(\"ans_1_all.dat\")\npred = np.loadtxt(\"pred_all.dat\")\n\ndiff = ans-pred\ndiff = diff[np.all(diff<1e8,axis=1),:]\nfactor = np.array([.4,.174,.4,.174,.4,.174,.4,.174])\ndiff = diff*factor\n\ndiff_v_x = diff[:,0]\ndiff_v_y = diff[:,4]\ndiff_v_z = (diff[:,1]+diff[:,5])/2.\ndiff_e_x = diff[:,2]\ndiff_e_y = diff[:,6]\ndiff_e_z = (diff[:,3]+diff[:,7])/2.\n\nnp.savetxt(\"diff_v_all.dat\",np.concatenate((diff_v_x.reshape((-1,1)),diff_v_y.reshape((-1,1)),diff_v_z.reshape((-1,1))),axis=1),header=\"x y z [mm]\")\nnp.savetxt(\"diff_e_all.dat\",np.concatenate((diff_e_x.reshape((-1,1)),diff_e_y.reshape((-1,1)),diff_e_z.reshape((-1,1))),axis=1),header=\"x y z [mm]\")\n\nfig = plt.figure()\nax = Axes3D(fig)\n\nax.set_xlabel(\"x (mm)\")\nax.set_ylabel(\"y (mm)\")\nax.set_zlabel(\"z (mm)\")\n\n#ax.set_xlim(-1,1)\n#ax.set_ylim(-1,1)\n#ax.set_zlim(-1,1)\n\nax.plot(diff_v_x,diff_v_y,diff_v_z,\"ro\",ms=4,mew=0.5)\nax.plot(diff_e_x,diff_e_y,diff_e_z,\"bo\",ms=4,mew=0.5)\n\nv = np.sqrt(diff_v_x*diff_v_x+diff_v_y*diff_v_y+diff_v_z*diff_v_z)\nfig2 = plt.figure()\nax2 = fig2.add_subplot(1,1,1)\nax2.hist(v,bins=100,range=(0,20))\n#ax2.hist(v,bins=100,range=(v.min(),v.max()))\n\ne = np.sqrt(diff_e_x*diff_e_x+diff_e_y*diff_e_y+diff_e_z*diff_e_z)\nfig3 = plt.figure()\nax3 = fig3.add_subplot(1,1,1)\nax3.hist(e,bins=100,range=(0,20))\n#ax3.hist(e,bins=100,range=(e.min(),e.max()))\n\nplt.show()\n", "sub_path": "evaluation/point_detection/diff_plot.py", "file_name": "diff_plot.py", "file_ext": "py", "file_size_in_byte": 2961, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "numpy.loadtxt", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "mpl_toolkits.mplot3d.Axes3D", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}]} +{"seq_id": "443740685", "text": "\"\"\"\r\nSam Lambert - sam.gervase.lambert@gmail.com\r\n\r\nThis script looks in a directory for docx files\r\n\r\n\r\n\"\"\"\r\n\r\nimport os\r\nimport docx2txt\r\n\r\nos.chdir('c:/users/Sam Lambert/desktop/projectx')\r\n\r\npath = ('c:/users/Sam Lambert/desktop/projectx')\r\n\r\nfiles = []\r\n\r\nx = str(input(\"search: \"))\r\n\r\nfor file in os.listdir(path):\r\n if file.endswith('.docx'):\r\n files.append(file)\r\n\r\nfor i in range(len(files)):\r\n text = docx2txt.process(files[i])\r\n if x.upper() in text.upper() or x.lower() in text.lower():\r\n print (files[i])\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n", "sub_path": "word_search.py", "file_name": "word_search.py", "file_ext": "py", "file_size_in_byte": 558, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "os.chdir", "line_number": 12, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 20, "usage_type": "call"}, {"api_name": "docx2txt.process", "line_number": 25, "usage_type": "call"}]} +{"seq_id": "180156841", "text": "#!/usr/bin/python3\nfrom collections import namedtuple\nimport math\nimport sys\nimport re\nimport subprocess\nimport time\nfrom matplotlib import pyplot as plt\nimport numpy as np\n\ndocorrect = False\n\ndef process( name ):\n #f = open(\"cosy_optics.twiss\")\n f = open( name )\n for line in f:\n if line[0] == \"@\":\n l = line.split()\n if l[1] == \"LENGTH\":\n lenght = float(l[3])\n if l[1] == \"GAMMA\":\n gamma = float(l[3])\n beta = math.sqrt(1-1./(gamma*gamma))\n if l[1] == \"ALFA\":\n alfa = float(l[3])\n if l[1] == \"PC\":\n p = float(l[3])\n if l[1] == \"CHARGE\":\n q = float(l[3])\n elif line[0] == \"*\":\n column_names = line.split()[1:]\n break\n c = 299792458\n rigidity = p * 1e9/(c * q)\n twiss = namedtuple(\"twiss\", column_names)\n twiss_data = []\n for line in f:\n if line[0] != \"$\":\n data = [x.strip('\"').lstrip(\"M\") for x in line.split()]\n if data[0].startswith(\"BLW\"):\n data[0] = data[0][:3] + \"0\" + data[0][3:]\n if data[0][-2] == \"D\":\n dl = list(data[0])\n dl[-3] = \"-\"\n data[0] = \"\".join(dl)\n if data[0].startswith(\"DPOS\"):\n m = re.match(r\"DPOS(\\d\\d)(H|V)\", data[0])\n if m:\n data[0] = (\"bpmx\" if m.group(2) == \"H\" else \"bpmy\") + m.group(1)\n else:\n m = re.match(r\"DPOSEC1(\\d\\d)(V|H)\", data[0])\n if m:\n data[0] = (\"ecbpmx1\" if m.group(2) == \"H\" else \"ecbpmy1\") + m.group(1)\n else:\n m = re.match(r\"DPOSANKE(\\d)(V|H)\", data[0])\n if m:\n data[0] = (\"banx0\" if m.group(2) == \"H\" else \"bany0\") + m.group(1)\n else:\n m = re.match(r\"DPOSEC(\\d)(V|H)\", data[0])\n if m:\n data[0] = (\"becx0\" if m.group(2) == \"H\" else \"becy0\") + m.group(1)\n\n twiss_data.append(twiss(*data))\n hs, hv = [], []\n for d in twiss_data:\n if d.KEYWORD == \"HKICKER\" and (d.NAME.startswith(\"SH\") or d.NAME.startswith(\"BLW\")) :\n hs.append(d)\n for d in twiss_data:\n if d.KEYWORD == \"VKICKER\" and (d.NAME.startswith(\"SV\") or d.NAME.startswith(\"BLW\")):\n hv.append(d)\n if d.NAME == \"COSY$END\":\n mu = (d.MUX, d.MUY)\n bpmh, bpmv = [], []\n for d in twiss_data:\n if d.KEYWORD == \"VMONITOR\":\n bpmv.append(d)\n if d.KEYWORD == \"HMONITOR\":\n bpmh.append(d)\n\n print(\"Len(hs) = \", len(hs))\n print(\"Len(hv) = \", len(hv))\n print(\"Len bpm = \", len(bpmv) + len(bpmh))\n f.close()\n return hs, hv, bpmh, bpmv, mu, beta, lenght, alfa, gamma, rigidity\n\ndef process_cf():\n files = [\"cosmoBLW.txt\", \"cosmoSH.txt\", \"cosmoSV.txt\"]\n fd = [open(f, \"r\") for f in files]\n correction_data = []\n for f in fd:\n for line in f:\n l = line.split(\":\")\n if l[2] == '%/mrad/brho' and not l[0].startswith(\"SHblw\") and not l[0].startswith(\"SVblw\"):\n correction_data.append([l[0], float(l[6]), float(l[7])])\n return correction_data\n\ndef correction_factors(cf, steerer_name):\n names = [n[0] for n in cf]\n i = names.index(steerer_name)\n return cf[i][2]\n\ndef response_matrix_x(steerers, bpms, mu, beta, lenght, alfa, gamma):\n response_matrix = [[0]*len(steerers) for i in range(len(bpms))]\n eta = alfa - 1./(gamma * gamma)\n for ib, b in enumerate(bpms):\n for ic, c in enumerate(steerers):\n response_matrix[ ib ][ ic ] = \\\n math.sqrt(float(b.BETX) * float(c.BETX)) / math.sin(math.pi * float(mu)) *\\\n math.cos(2. * math.pi * abs(float(b.MUX) - float(c.MUX)) - math.pi * float(mu)) * 0.5-\\\n float(b.DX) * beta * float(c.DX) * beta / (eta * lenght)\n return response_matrix\n\ndef response_matrix_y(steerers, bpms, mu):\n response_matrix = [[0]*len(steerers) for i in range(len(bpms))]\n for ib, b in enumerate(bpms):\n for ic, c in enumerate(steerers):\n response_matrix[ ib ][ ic ] = \\\n math.sqrt(float(b.BETY) * float(c.BETY)) / math.sin(math.pi * float(mu)) *\\\n math.cos(2. * math.pi * abs(float(b.MUY) - float(c.MUY)) - math.pi * float(mu)) * 0.5\n return response_matrix\n\ndef matrix_100_units(response_matrix, rigidity, cf, steerers, bpms):\n response_matrix_u = [[0]*len(response_matrix[0]) for i in range(len(response_matrix))]\n for ib, b in enumerate(bpms):\n for ic, c in enumerate(steerers):\n response_matrix_u[ib][ic] = response_matrix[ib][ic] / (correction_factors(cf, c.NAME) * rigidity)\n return response_matrix_u\n\n\ndef index_from_name(data, name):\n names = [d.NAME for d in data]\n return names.index(name)\n\ndef create_matrix_fabian_format(matrix, matriy, hs, vs, bpmh, bpmv):\n bfile = open(\"BPMS.txt\", \"r\")\n sfile = open(\"STEERERS.data\", \"r\")\n outfile = open(\"normal_matrix.data\", \"w\")\n\n lenx = len(matrix[0]) + len(matriy[0])\n leny = len(matrix) + len(matriy)\n\n\n bpm_names, steerer_names = [], []\n for line in bfile:\n bpm_names.append(line.split()[0])\n for line in sfile:\n if line[0] != \"#\":\n steerer_names.append(line.split(\":\")[0])\n\n out_matrix = [[0] * len(steerer_names) for i in range(len(bpm_names))]\n\n for indb, bpmn in enumerate(bpm_names[:32]):\n for inds, sn in enumerate(steerer_names[:22]):\n idb = index_from_name(bpmh, bpmn)\n ids = index_from_name(hs, sn)\n out_matrix[indb][inds] = matrix[idb][ids]\n #print(lenx, leny)\n #print(len(bpm_names), len(steerer_names))\n for indb, bpmn in enumerate(bpm_names[32:]):\n for inds, sn in enumerate(steerer_names[22:]):\n idb = index_from_name(bpmv, bpmn)\n ids = index_from_name(vs, sn)\n out_matrix[32 + indb][22 + inds] = matriy[idb][ids]\n for row in out_matrix:\n for c in row:\n outfile.write(str(c) + \" \")\n outfile.write(\"\\n\")\n bfile.close()\n sfile.close()\n outfile.close()\n return out_matrix\n\n\ndef makeresponse( twiss, cnames, bnames, positions, plane, printRM=False ):\n endTwiss = twiss[ 'ende' ]\n muTotx = endTwiss.mux\n muToty = endTwiss.muy\n print( \"mu_x, mu_y:\", muTotx, muToty )\n selected = dict(zip(bnames, positions))\n # print selected\n\n # [len(bnames), len(cnames)] x [len(cnames), 1] = [len(bnames), 1]\n ResponseMatrix = np.zeros((len(bnames), len(cnames)))\n OrbitVector = np.zeros((len(bnames)))\n for ib, b in enumerate(bnames):\n if plane == 'v':\n for ic, c in enumerate(cnames):\n ResponseMatrix[ ib, ic ] = \\\n math.sqrt(twiss[ b ].betay * twiss[ c ].betay) / math.sin(math.pi * muToty) *\\\n math.cos(2. * math.pi * abs(twiss[ b ].muy - twiss[ c ].muy) - math.pi * muToty) * 0.5\n else:\n for ic, c in enumerate(cnames):\n ResponseMatrix[ ib, ic ] = \\\n math.sqrt(twiss[ b ].betax * twiss[ c ].betax) / math.sin(math.pi * muTotx) *\\\n math.cos(2. * math.pi * abs(twiss[ b ].mux - twiss[ c ].mux) - math.pi * muTotx) * 0.5\n if abs(ResponseMatrix[ ib, ic ]) > 2e3:\n print( \"FAIL: ResponseMatrix[\", ib, \",\", ic, \"] =\", ResponseMatrix[ ib, ic ] )\n if b in selected:\n OrbitVector[ ib ] = selected[ b ]\n\n if printRM:\n print( 'ResponseMatrix size =', ResponseMatrix.shape )\n print( ResponseMatrix )\n print( 'OrbitVector size =', OrbitVector.shape )\n print( OrbitVector )\n return ResponseMatrix, OrbitVector\n\ndef threeBump( twiss, name1, name2, name3 ):\n index1 = index_from_name( twiss, name1 )\n index2 = index_from_name( twiss, name2 )\n index3 = index_from_name( twiss, name3 )\n\n bhc = float( twiss[ index2 ].BETY )\n fihc = float( twiss[ index2 ].MUY ) * 2. * math.pi\n\n # Check the phase relations between correctors\n bhcl = float( twiss[ index1 ].BETY )\n fihcl = float( twiss[ index1 ].MUY ) * 2. * math.pi\n # if(i0 > i1) fihcl -= mu_tot // if the bump overlap the ring initial point\n\n bhcr = float( twiss[ index3 ].BETY )\n fihcr = float( twiss[ index3 ].MUY ) * 2. * math.pi\n # if(i1 > i2) fihcr += mu_tot // if the bump overlap the ring initial point\n\n correctorResponse = [0,0,0]\n correctorResponse[0] = (1.0 / math.sqrt(bhcl * bhc)) * (1.0 / math.sin(fihc - fihcl))\n correctorResponse[1] = (1.0 / bhc) * (math.sin(fihcl - fihcr) / (math.sin(fihcr - fihc) * math.sin(fihc - fihcl)))\n correctorResponse[2] = (1.0 / math.sqrt(bhcr * bhc)) * (1.0 / math.sin(fihcr - fihc))\n strengths = [0.] * len(twiss)\n strengths[ index1 ] = correctorResponse[0]\n strengths[ index2 ] = correctorResponse[1]\n strengths[ index3 ] = correctorResponse[2]\n print( name1, name2, name3, correctorResponse[0], correctorResponse[1], correctorResponse[2] )\n return correctorResponse, strengths\n\ndef makeb( args ):\n twname = args[0]\n hs, vs, bpmh, bpmv, mu, beta, lenght, alfa, gamma, rigidity = process(twname)\n cf = process_cf()\n rmy = response_matrix_y(vs, bpmv, mu[1])\n u,svalues,vh = np.linalg.svd( rmy )\n plt.plot(svalues, marker=2)\n plt.show()\n if True:\n rmyu = matrix_100_units(rmy, rigidity, cf, vs, bpmv)\n u,svalues,vh = np.linalg.svd( rmyu )\n plt.plot(svalues, marker=2)\n plt.show()\n if True:\n rmx = response_matrix_x(hs, bpmh, mu[0], beta, lenght, alfa, gamma)\n if True:\n u,svalues,vh = np.linalg.svd( rmx )\n plt.plot(svalues, marker=2)\n plt.show()\n rmxu = matrix_100_units(rmx, rigidity, cf, hs, bpmh)\n rmf = create_matrix_fabian_format(rmxu, rmyu, hs, vs, bpmh, bpmv)\n u,svalues,vh = np.linalg.svd( rmf )\n plt.plot(svalues, marker=2)\n plt.show()\n for s in vs:\n c = \"caput \" + s.NAME + \" 0\"\n print(c)\n if docorrect:\n subprocess.call(c, shell=True)\n scord = [x.S for x in bpmv]\n #time.sleep(5)\n junk= input('--> start ')\n for i in range(len(vs) - 2):\n name = [x.NAME for x in vs[i : i + 3]]\n print( [x.NAME for x in vs[i : i + 3]] )\n if any(['BLW' in x for x in name]): continue\n \n #name = [\"\"] * 3;\n t = [\"\"] * 3\n '''\n name[0] = 'SV10'; t[0] = 'SV08/10/12/14:SDI2'\n name[1] = 'SV12'; t[1] = 'SV08/10/12/14:SDI3'\n name[2] = 'SV14'; t[2] = 'SV08/10/12/14:SDI4'\n '''\n t[0] = 'SV08/10/12/14:SDI2'\n t[1] = 'SV08/10/12/14:SDI3'\n t[2] = 'SV08/10/12/14:SDI4'\n\n s, sv = threeBump( vs, name[0], name[1], name[2] )\n calcResponse( rmy, sv, scord )\n a = 2.\n p = [0.] * 3; c = [\"\"] * 3\n for i in range(3):\n p[i] = a * s[i] / (correction_factors(cf, name[i]) * rigidity) * 2048\n c[i] = \"caput \" + name[i] + \" \" + str(p[i])\n print(c[i])\n if docorrect:\n subprocess.call(c[i], shell=True)\n #junk = input('--> 5 ')\n #time.sleep(5)\n calcResponse( rmy, [-x for x in sv], scord )\n for i in range(3):\n p[i] = -a * s[i] / (correction_factors(cf, name[i]) * rigidity) * 2048\n c[i] = \"caput \" + name[i] + \" \" + str(p[i])\n print(c[i])\n if docorrect:\n subprocess.call(c[i], shell=True)\n #junk = input('--> -5 ')\n #time.sleep(5)\n calcResponse( rmy, [0. for x in sv], scord )\n for i in range(3):\n c[i] = \"caput \" + name[i] + \" 0\"\n print(c[i])\n if docorrect:\n subprocess.call(c[i], shell=True)\n #junk = input('--> 0 ')\n #time.sleep(5)\n\ndef calcResponse( rm, strengths, scord ):\n # print( len(strengths), len(rm), len(rm[0]) )\n pos = [0.] * len(rm)\n for i in range( len(rm) ):\n for j in range( len(rm[0]) ):\n pos[i] += strengths[j] * rm[i][j]\n plt.plot(scord, pos)\n plt.show()\n'''\ndef name_to_target():\n file_crateh s = open(\"crate_steererv.data\", \"r\")\n for line in s:\n l = line.split(line)\n print(l[0], l[4:])\n'''\n\ndef makematrix():\n hs, vs, bpmh, bpmv, mu, beta, lenght, alfa, gamma, rigidity = process()\n cd = process_cf()\n rmx = response_matrix_x(hs, bpmh, mu[0], beta, lenght, alfa, gamma)\n rmy = response_matrix_y(vs, bpmv, mu[1])\n rmxu = matrix_100_units(rmx, rigidity, cd, hs, bpmh)\n rmyu = matrix_100_units(rmy, rigidity, cd, vs, bpmv)\n create_matrix_fabian_format(rmxu, rmyu, hs, vs, bpmh, bpmv)\n\n\nif __name__ == \"__main__\":\n #name_to_target()\n sys.exit( makeb( sys.argv[1:] ) )\n\n'''\n \"MSV10\" \"VKICKER\" 49.771 0 0 14.96869392 6.648205105 14.01694053 0 1.20580545 1.153532505\n \"DPOS11H\" \"HMONITOR\" 59.99 0 0 7.02140788 21.04016513 11.00821148 0 1.328283583 1.371387996\n \"DPOS11V\" \"VMONITOR\" 60.15 0 0 6.648254238 21.85627509 10.83037948 0 1.33201105 1.372575491\n \"MSV12\" \"VKICKER\" 61.1534 0 0 5.592221629 23.52779521 10.55459841 0 1.3589585 1.379452598\n \"DPOS12H\" \"HMONITOR\" 66.318 0 0 12.80546554 11.28969619 14.95153491 0 1.481987687 1.421787535\n \"DPOS12V\" \"VMONITOR\" 66.483 0 0 13.40127951 10.59549147 15.26582399 0 1.483992365 1.424188683\n \"MSV14\" \"VKICKER\" 67.1092 0 0 14.41949265 9.077319218 15.70792023 0 1.491011324 1.434577182\n'''\n", "sub_path": "EPICS_IOCs/orbitCorrectionIOC/Scripts/makebumps.py", "file_name": "makebumps.py", "file_ext": "py", "file_size_in_byte": 14531, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "math.sqrt", "line_number": 23, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 35, "usage_type": "call"}, {"api_name": "re.match", "line_number": 47, "usage_type": "call"}, {"api_name": "re.match", "line_number": 51, "usage_type": "call"}, {"api_name": "re.match", "line_number": 55, "usage_type": "call"}, {"api_name": "re.match", "line_number": 59, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 108, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 108, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 108, "usage_type": "attribute"}, {"api_name": "math.cos", "line_number": 109, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 109, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 118, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 118, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 118, "usage_type": "attribute"}, {"api_name": "math.cos", "line_number": 119, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 119, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 184, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 189, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 189, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 189, "usage_type": "attribute"}, {"api_name": "math.cos", "line_number": 190, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 190, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 194, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 194, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 194, "usage_type": "attribute"}, {"api_name": "math.cos", "line_number": 195, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 195, "usage_type": "attribute"}, {"api_name": "math.pi", "line_number": 214, "usage_type": "attribute"}, {"api_name": "math.pi", "line_number": 218, "usage_type": "attribute"}, {"api_name": "math.pi", "line_number": 222, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 226, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 226, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 227, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 228, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 228, "usage_type": "call"}, {"api_name": "numpy.linalg.svd", "line_number": 241, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 241, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 242, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 242, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 243, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 243, "usage_type": "name"}, {"api_name": "numpy.linalg.svd", "line_number": 246, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 246, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 247, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 247, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 248, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 248, "usage_type": "name"}, {"api_name": "numpy.linalg.svd", "line_number": 252, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 252, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 253, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 253, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 254, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 254, "usage_type": "name"}, {"api_name": "numpy.linalg.svd", "line_number": 257, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 257, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 258, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 258, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 259, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 259, "usage_type": "name"}, {"api_name": "subprocess.call", "line_number": 264, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 293, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 302, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 310, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 320, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 320, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 321, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 321, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 342, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 342, "usage_type": "attribute"}]} +{"seq_id": "640799924", "text": "import pytest\n\nimport bloop.tables\nimport bloop.util\n\nfrom test_models import SimpleModel, ComplexModel, User\n\n\nstatuses = [\n (\"ACTIVE\", \"ACTIVE\", \"ACTIVE\"),\n (\"ACTIVE\", None, \"ACTIVE\"),\n (\"ACTIVE\", \"BUSY\", \"BLOOP_NOT_ACTIVE\"),\n (\"BUSY\", \"ACTIVE\", \"BLOOP_NOT_ACTIVE\"),\n (\"BUSY\", \"BUSY\", \"BLOOP_NOT_ACTIVE\")\n]\n\n\ndef assert_unordered(obj, other):\n assert bloop.util.ordered(obj) == bloop.util.ordered(other)\n\n\ndef test_create_simple():\n expected = {\n 'AttributeDefinitions': [\n {'AttributeName': 'id', 'AttributeType': 'S'}],\n 'KeySchema': [{'AttributeName': 'id', 'KeyType': 'HASH'}],\n 'ProvisionedThroughput': {\n 'ReadCapacityUnits': 1,\n 'WriteCapacityUnits': 1},\n 'TableName': 'Simple'}\n assert_unordered(bloop.tables.create_request(SimpleModel), expected)\n\n\ndef test_create_complex():\n expected = {\n 'AttributeDefinitions': [\n {'AttributeType': 'S', 'AttributeName': 'date'},\n {'AttributeType': 'S', 'AttributeName': 'email'},\n {'AttributeType': 'S', 'AttributeName': 'joined'},\n {'AttributeType': 'S', 'AttributeName': 'name'}],\n 'GlobalSecondaryIndexes': [{\n 'IndexName': 'by_email',\n 'KeySchema': [{'KeyType': 'HASH', 'AttributeName': 'email'}],\n 'Projection': {'ProjectionType': 'ALL'},\n 'ProvisionedThroughput': {\n 'ReadCapacityUnits': 4, 'WriteCapacityUnits': 5}}],\n 'KeySchema': [{'KeyType': 'HASH', 'AttributeName': 'name'},\n {'KeyType': 'RANGE', 'AttributeName': 'date'}],\n 'LocalSecondaryIndexes': [{\n 'IndexName': 'by_joined',\n 'KeySchema': [\n {'KeyType': 'HASH', 'AttributeName': 'name'},\n {'KeyType': 'RANGE', 'AttributeName': 'joined'}],\n 'Projection': {\n 'NonKeyAttributes': ['joined', 'email', 'date', 'name'],\n 'ProjectionType': 'INCLUDE'}}],\n 'ProvisionedThroughput': {\n 'ReadCapacityUnits': 3, 'WriteCapacityUnits': 2},\n 'TableName': 'CustomTableName'}\n assert_unordered(bloop.tables.create_request(ComplexModel), expected)\n\n\ndef test_expected_description():\n # Eventually expected_description will probably diverge from create_table\n # This will guard against (or coverage should show) if there's drift\n create = bloop.tables.create_request(ComplexModel)\n expected = bloop.tables.expected_description(ComplexModel)\n assert_unordered(create, expected)\n\n\ndef test_sanitize_drop_empty_lists():\n expected = bloop.tables.expected_description(ComplexModel)\n # Start from the same base, but inject an unnecessary NonKeyAttributes\n description = bloop.tables.expected_description(ComplexModel)\n index = description[\"GlobalSecondaryIndexes\"][0]\n index[\"Projection\"][\"NonKeyAttributes\"] = []\n\n assert_unordered(expected, bloop.tables.sanitized_description(description))\n\n\ndef test_sanitize_drop_empty_indexes():\n expected = bloop.tables.expected_description(SimpleModel)\n # Start from the same base, but inject an unnecessary NonKeyAttributes\n description = bloop.tables.expected_description(SimpleModel)\n description[\"GlobalSecondaryIndexes\"] = []\n\n assert_unordered(expected, bloop.tables.sanitized_description(description))\n\n\ndef test_sanitize_expected():\n expected = bloop.tables.expected_description(User)\n # Add some extra fields\n description = {\n 'AttributeDefinitions': [\n {'AttributeType': 'S', 'AttributeName': 'email'},\n {'AttributeType': 'S', 'AttributeName': 'id'}],\n 'CreationDateTime': 'EXTRA_FIELD',\n 'ItemCount': 'EXTRA_FIELD',\n 'KeySchema': [{'AttributeName': 'id', 'KeyType': 'HASH'}],\n 'GlobalSecondaryIndexes': [{\n 'IndexArn': 'EXTRA_FIELD',\n 'IndexName': 'by_email',\n 'IndexSizeBytes': 'EXTRA_FIELD',\n 'IndexStatus': 'EXTRA_FIELD',\n 'KeySchema': [{'AttributeName': 'email', 'KeyType': 'HASH'}],\n 'Projection': {'ProjectionType': 'ALL'},\n 'ProvisionedThroughput': {\n 'NumberOfDecreasesToday': 'EXTRA_FIELD',\n 'ReadCapacityUnits': 1,\n 'WriteCapacityUnits': 1}}],\n 'ProvisionedThroughput': {\n 'LastDecreaseDateTime': 'EXTRA_FIELD',\n 'LastIncreaseDateTime': 'EXTRA_FIELD',\n 'NumberOfDecreasesToday': 'EXTRA_FIELD',\n 'ReadCapacityUnits': 1,\n 'WriteCapacityUnits': 1},\n 'TableArn': 'EXTRA_FIELD',\n 'TableName': 'User',\n 'TableSizeBytes': 'EXTRA_FIELD',\n 'TableStatus': 'EXTRA_FIELD'}\n sanitized = bloop.tables.sanitized_description(description)\n assert_unordered(expected, sanitized)\n\n\n@pytest.mark.parametrize(\"table_status, gsi_status, expected_status\", statuses)\ndef test_simple_status(table_status, gsi_status, expected_status):\n \"\"\"Status is busy because table isn't ACTIVE, no GSIs\"\"\"\n description = {\"TableStatus\": table_status}\n if gsi_status is not None:\n description[\"GlobalSecondaryIndexes\"] = [{\"IndexStatus\": gsi_status}]\n assert bloop.tables.simple_status(description) == expected_status\n", "sub_path": "tests/test_tables.py", "file_name": "test_tables.py", "file_ext": "py", "file_size_in_byte": 5242, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "bloop.tables.util.ordered", "line_number": 19, "usage_type": "call"}, {"api_name": "bloop.tables.util", "line_number": 19, "usage_type": "attribute"}, {"api_name": "bloop.tables", "line_number": 19, "usage_type": "name"}, {"api_name": "bloop.tables.tables.create_request", "line_number": 31, "usage_type": "call"}, {"api_name": "test_models.SimpleModel", "line_number": 31, "usage_type": "argument"}, {"api_name": "bloop.tables.tables", "line_number": 31, "usage_type": "attribute"}, {"api_name": "bloop.tables", "line_number": 31, "usage_type": "name"}, {"api_name": "bloop.tables.tables.create_request", "line_number": 60, "usage_type": "call"}, {"api_name": "test_models.ComplexModel", "line_number": 60, "usage_type": "argument"}, {"api_name": "bloop.tables.tables", "line_number": 60, "usage_type": "attribute"}, {"api_name": "bloop.tables", "line_number": 60, "usage_type": "name"}, {"api_name": "bloop.tables.tables.create_request", "line_number": 66, "usage_type": "call"}, {"api_name": "test_models.ComplexModel", "line_number": 66, "usage_type": "argument"}, {"api_name": "bloop.tables.tables", "line_number": 66, "usage_type": "attribute"}, {"api_name": "bloop.tables", "line_number": 66, "usage_type": "name"}, {"api_name": "bloop.tables.tables.expected_description", "line_number": 67, "usage_type": "call"}, {"api_name": "test_models.ComplexModel", "line_number": 67, "usage_type": "argument"}, {"api_name": "bloop.tables.tables", "line_number": 67, "usage_type": "attribute"}, {"api_name": "bloop.tables", "line_number": 67, "usage_type": "name"}, {"api_name": "bloop.tables.tables.expected_description", "line_number": 72, "usage_type": "call"}, {"api_name": "test_models.ComplexModel", "line_number": 72, "usage_type": "argument"}, {"api_name": "bloop.tables.tables", "line_number": 72, "usage_type": "attribute"}, {"api_name": "bloop.tables", "line_number": 72, "usage_type": "name"}, {"api_name": "bloop.tables.tables.expected_description", "line_number": 74, "usage_type": "call"}, {"api_name": "test_models.ComplexModel", "line_number": 74, "usage_type": "argument"}, {"api_name": "bloop.tables.tables", "line_number": 74, "usage_type": "attribute"}, {"api_name": "bloop.tables", "line_number": 74, "usage_type": "name"}, {"api_name": "bloop.tables.tables.sanitized_description", "line_number": 78, "usage_type": "call"}, {"api_name": "bloop.tables.tables", "line_number": 78, "usage_type": "attribute"}, {"api_name": "bloop.tables", "line_number": 78, "usage_type": "name"}, {"api_name": "bloop.tables.tables.expected_description", "line_number": 82, "usage_type": "call"}, {"api_name": "test_models.SimpleModel", "line_number": 82, "usage_type": "argument"}, {"api_name": "bloop.tables.tables", "line_number": 82, "usage_type": "attribute"}, {"api_name": "bloop.tables", "line_number": 82, "usage_type": "name"}, {"api_name": "bloop.tables.tables.expected_description", "line_number": 84, "usage_type": "call"}, {"api_name": "test_models.SimpleModel", "line_number": 84, "usage_type": "argument"}, {"api_name": "bloop.tables.tables", "line_number": 84, "usage_type": "attribute"}, {"api_name": "bloop.tables", "line_number": 84, "usage_type": "name"}, {"api_name": "bloop.tables.tables.sanitized_description", "line_number": 87, "usage_type": "call"}, {"api_name": "bloop.tables.tables", "line_number": 87, "usage_type": "attribute"}, {"api_name": "bloop.tables", "line_number": 87, "usage_type": "name"}, {"api_name": "bloop.tables.tables.expected_description", "line_number": 91, "usage_type": "call"}, {"api_name": "test_models.User", "line_number": 91, "usage_type": "argument"}, {"api_name": "bloop.tables.tables", "line_number": 91, "usage_type": "attribute"}, {"api_name": "bloop.tables", "line_number": 91, "usage_type": "name"}, {"api_name": "bloop.tables.tables.sanitized_description", "line_number": 121, "usage_type": "call"}, {"api_name": "bloop.tables.tables", "line_number": 121, "usage_type": "attribute"}, {"api_name": "bloop.tables", "line_number": 121, "usage_type": "name"}, {"api_name": "bloop.tables.tables.simple_status", "line_number": 131, "usage_type": "call"}, {"api_name": "bloop.tables.tables", "line_number": 131, "usage_type": "attribute"}, {"api_name": "bloop.tables", "line_number": 131, "usage_type": "name"}, {"api_name": "pytest.mark.parametrize", "line_number": 125, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 125, "usage_type": "attribute"}]} +{"seq_id": "560697620", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations, models\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('voterguide', '0008_auto_20160907_1439'),\n ]\n\n operations = [\n migrations.AlterField(\n model_name='district',\n name='chamber',\n field=models.IntegerField(verbose_name='Chamber', choices=[(1, 'State Senate'), (2, 'State House'), (3, 'County'), (8, 'Mayor'), (4, 'City Council'), (5, 'US Senate'), (6, 'US House'), (7, \"Governor's Council\")]),\n ),\n migrations.AlterField(\n model_name='office',\n name='chamber',\n field=models.IntegerField(blank=True, help_text='Optional', null=True, verbose_name='Chamber', choices=[(1, 'State Senate'), (2, 'State House'), (3, 'County'), (8, 'Mayor'), (4, 'City Council'), (5, 'US Senate'), (6, 'US House'), (7, \"Governor's Council\")]),\n ),\n ]\n", "sub_path": "voterguide/migrations/0009_auto_20170709_1938.py", "file_name": "0009_auto_20170709_1938.py", "file_ext": "py", "file_size_in_byte": 957, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}]} +{"seq_id": "453763003", "text": "# SPDX-License-Identifier: BSD-3-Clause\n# Copyright (c) 2021 Scipp contributors (https://github.com/scipp)\n# @file\n# @author Neil Vaytet\n\nimport numpy as np\nimport scipp as sc\nfrom ..factory import make_dense_data_array, make_binned_data_array\nfrom .plot_helper import plot\nimport matplotlib\nmatplotlib.use('Agg')\n\n\ndef _with_fake_pos(*args, **kwargs):\n da = make_dense_data_array(*args, **kwargs)\n da.coords['pos'] = sc.geometry.position(da.coords['xx'], da.coords['yy'],\n da.coords['zz']).transpose(da.dims[:3])\n return da\n\n\ndef make_data_array_with_position_vectors():\n N = 1000\n M = 100\n theta = np.random.random(N) * np.pi\n phi = np.random.random(N) * 2.0 * np.pi\n r = 10.0 + (np.random.random(N) - 0.5)\n x = r * np.sin(theta) * np.sin(phi)\n y = r * np.sin(theta) * np.cos(phi)\n z = r * np.cos(theta)\n time = np.arange(M, dtype=np.float64)\n a = np.arange(M * N).reshape([M, N]) * np.sin(y)\n da = sc.DataArray(data=sc.Variable(dims=['time', 'xyz'], values=a),\n coords={\n 'xyz':\n sc.vectors(dims=['xyz'], values=np.array([x, y, z]).T),\n 'pos':\n sc.vectors(dims=['xyz'], values=np.array([x, y, z]).T + 20.0),\n 'time':\n sc.Variable(dims=['time'], values=time)\n })\n return da\n\n\ndef test_plot_projection_3d():\n da = _with_fake_pos(ndim=3)\n plot(da, positions='pos', projection=\"3d\")\n plot(da, positions='pos', projection=\"3d\", resampling_mode='sum')\n plot(da, positions='pos', projection=\"3d\", resampling_mode='mean')\n\n\ndef test_plot_projection_3d_log_norm():\n plot(_with_fake_pos(ndim=3), positions='pos', projection=\"3d\", norm='log')\n\n\ndef test_plot_projection_3d_dataset():\n plot(_with_fake_pos(ndim=3), positions='pos', projection=\"3d\")\n\n\ndef test_plot_projection_3d_with_labels():\n plot(_with_fake_pos(ndim=3, labels=True),\n positions='pos',\n projection=\"3d\",\n labels={'x': \"lab\"})\n\n\ndef test_plot_projection_3d_with_masks():\n plot(_with_fake_pos(ndim=3, masks=True), positions='pos', projection=\"3d\")\n\n\ndef test_plot_projection_3d_with_vectors():\n plot(make_data_array_with_position_vectors(), projection=\"3d\", positions=\"xyz\")\n\n\ndef test_plot_projection_3d_with_vectors_non_dim_coord():\n plot(make_data_array_with_position_vectors(), projection=\"3d\", positions=\"pos\")\n\n\ndef test_plot_variable_3d():\n N = 50\n v3d = sc.Variable(dims=['time', 'y', 'x'],\n values=np.random.rand(N, N, N),\n unit=sc.units.m)\n positions = sc.vectors(dims=v3d.dims, values=np.random.rand(N, N, N, 3))\n plot(v3d, positions=positions, projection=\"3d\")\n\n\ndef test_plot_4d_with_masks_projection_3d():\n da = sc.DataArray(data=sc.Variable(dims=['pack', 'tube', 'straw', 'pixel'],\n values=np.random.rand(2, 8, 7, 256)),\n coords={})\n a = np.sin(np.linspace(0, 3.14, num=256))\n da += sc.Variable(dims=['pixel'], values=a)\n da.masks['tube_ends'] = sc.Variable(dims=['pixel'],\n values=np.where(a > 0.5, True, False))\n da.coords['pos'] = sc.geometry.position(sc.arange(dim='pack', start=0., stop=2),\n sc.arange(dim='tube', start=0., stop=8),\n sc.arange(dim='straw', start=0., stop=7))\n plot(da, positions='pos', projection=\"3d\")\n\n\ndef test_plot_customized_axes():\n da = _with_fake_pos(ndim=3)\n plot(da,\n positions='pos',\n projection=\"3d\",\n xlabel=\"MyXlabel\",\n ylabel=\"MyYlabel\",\n zlabel=\"MyZlabel\")\n\n\ndef test_plot_3d_with_2d_position_coordinate():\n nx = 50\n ny = 40\n nt = 10\n\n xx, yy = np.meshgrid(np.arange(nx, dtype=np.float64), np.arange(ny,\n dtype=np.float64))\n da = sc.DataArray(\n data=sc.Variable(dims=['x', 'y', 't'],\n values=np.arange(nx * ny * nt).reshape(nx, ny, nt)),\n coords={\n 'pos':\n sc.vectors(dims=['x', 'y'],\n values=np.array([xx, yy,\n np.zeros_like(xx)]).T.reshape(nx, ny, 3)),\n 't':\n sc.arange('t', nt + 1, dtype=np.float64)\n })\n\n plot(da, projection=\"3d\", positions=\"pos\")\n\n\ndef test_plot_3d_binned_data():\n da = make_binned_data_array(ndim=1)\n pos = sc.vectors(dims=da.dims, values=np.random.rand(da.sizes[da.dims[0]], 3))\n plot(da, projection='3d', positions=pos)\n plot(da, projection='3d', positions=pos, resampling_mode='sum')\n plot(da, projection='3d', positions=pos, resampling_mode='mean')\n\n\ndef test_plot_redraw():\n da = _with_fake_pos(ndim=3, unit='K')\n p = sc.plot(da, positions='pos', projection=\"3d\")\n before = p.view.figure.points_geometry.attributes[\"rgba_color\"].array\n da *= 5.0\n p.redraw()\n after = p.view.figure.points_geometry.attributes[\"rgba_color\"].array\n assert np.any(before != after)\n", "sub_path": "tests/plotting/plot_3d_test.py", "file_name": "plot_3d_test.py", "file_ext": "py", "file_size_in_byte": 5206, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "matplotlib.use", "line_number": 11, "usage_type": "call"}, {"api_name": "factory.make_dense_data_array", "line_number": 15, "usage_type": "call"}, {"api_name": "scipp.geometry.position", "line_number": 16, "usage_type": "call"}, {"api_name": "scipp.geometry", "line_number": 16, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 24, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 24, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 25, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 25, "usage_type": "attribute"}, {"api_name": "numpy.random.random", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 26, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 30, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 31, "usage_type": "call"}, {"api_name": "scipp.DataArray", "line_number": 32, "usage_type": "call"}, {"api_name": "scipp.Variable", "line_number": 32, "usage_type": "call"}, {"api_name": "scipp.vectors", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 35, "usage_type": "call"}, {"api_name": "scipp.vectors", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 37, "usage_type": "call"}, {"api_name": "scipp.Variable", "line_number": 39, "usage_type": "call"}, {"api_name": "plot_helper.plot", "line_number": 46, "usage_type": "call"}, {"api_name": "plot_helper.plot", "line_number": 47, "usage_type": "call"}, {"api_name": "plot_helper.plot", "line_number": 48, "usage_type": "call"}, {"api_name": "plot_helper.plot", "line_number": 52, "usage_type": "call"}, {"api_name": "plot_helper.plot", "line_number": 56, "usage_type": "call"}, {"api_name": "plot_helper.plot", "line_number": 60, "usage_type": "call"}, {"api_name": "plot_helper.plot", "line_number": 67, "usage_type": "call"}, {"api_name": "plot_helper.plot", "line_number": 71, "usage_type": "call"}, {"api_name": "plot_helper.plot", "line_number": 75, "usage_type": "call"}, {"api_name": "scipp.Variable", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 81, "usage_type": "attribute"}, {"api_name": "scipp.units", "line_number": 82, "usage_type": "attribute"}, {"api_name": "scipp.vectors", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 83, "usage_type": "attribute"}, {"api_name": "plot_helper.plot", "line_number": 84, "usage_type": "call"}, {"api_name": "scipp.DataArray", "line_number": 88, "usage_type": "call"}, {"api_name": "scipp.Variable", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 89, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 91, "usage_type": "call"}, {"api_name": "scipp.Variable", "line_number": 92, "usage_type": "call"}, {"api_name": "scipp.Variable", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 94, "usage_type": "call"}, {"api_name": "scipp.geometry.position", "line_number": 95, "usage_type": "call"}, {"api_name": "scipp.geometry", "line_number": 95, "usage_type": "attribute"}, {"api_name": "scipp.arange", "line_number": 95, "usage_type": "call"}, {"api_name": "scipp.arange", "line_number": 96, "usage_type": "call"}, {"api_name": "scipp.arange", "line_number": 97, "usage_type": "call"}, {"api_name": "plot_helper.plot", "line_number": 98, "usage_type": "call"}, {"api_name": "plot_helper.plot", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 116, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 117, "usage_type": "attribute"}, {"api_name": "scipp.DataArray", "line_number": 118, "usage_type": "call"}, {"api_name": "scipp.Variable", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 120, "usage_type": "call"}, {"api_name": "scipp.vectors", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 125, "usage_type": "call"}, {"api_name": "scipp.arange", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 127, "usage_type": "attribute"}, {"api_name": "plot_helper.plot", "line_number": 130, "usage_type": "call"}, {"api_name": "factory.make_binned_data_array", "line_number": 134, "usage_type": "call"}, {"api_name": "scipp.vectors", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 135, "usage_type": "attribute"}, {"api_name": "plot_helper.plot", "line_number": 136, "usage_type": "call"}, {"api_name": "plot_helper.plot", "line_number": 137, "usage_type": "call"}, {"api_name": "plot_helper.plot", "line_number": 138, "usage_type": "call"}, {"api_name": "scipp.plot", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.any", "line_number": 148, "usage_type": "call"}]} +{"seq_id": "592201682", "text": "# -*- coding:utf-8 -*-\n# @Author: Phoebe\n# @File: test_touchactions.py\nfrom selenium import webdriver\nfrom selenium.webdriver import TouchActions\n\n\nclass TestTouchActions:\n def setup(self):\n option = webdriver.ChromeOptions()\n option.add_experimental_option('w3', False)\n self.driver = webdriver.Chrome(options=option)\n self.driver.implicitly_wait(3)\n self.driver.maximize_window()\n\n def teardown(self):\n self.driver.quit()\n\n def test_touchactions_scrollbottom(self):\n self.driver.get('https://www.baidu.com/')\n el = self.driver.find_element_by_id('kw')\n el_search = self.driver.find_element_by_id('su')\n el.send_keys('selenium测试')\n action = TouchActions(self.driver)\n action.tap(el_search)\n action.perform()\n\n action.scroll_from_element(el, 0, 1000).perform()\n", "sub_path": "selenium_event/test_touchactions.py", "file_name": "test_touchactions.py", "file_ext": "py", "file_size_in_byte": 874, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "selenium.webdriver.ChromeOptions", "line_number": 10, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 10, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 12, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 12, "usage_type": "name"}, {"api_name": "selenium.webdriver.TouchActions", "line_number": 24, "usage_type": "call"}]} +{"seq_id": "591446899", "text": "\r\nfrom typing import List\r\n\r\n# Definition for a Node.\r\nclass Node:\r\n def __init__(self, val=None, children=None):\r\n self.val = val\r\n self.children = children\r\n\r\nclass Solution:\r\n def postorder(self, root: 'Node') -> List[int]:\r\n a = []\r\n if not root:\r\n return a\r\n b = [root]\r\n while len(b) != 0:\r\n c = b.pop()\r\n if c.children:\r\n b += c.children\r\n if c:\r\n a.append(c.val)\r\n return a[::-1]\r\n\r\nn1 = Node(1)\r\nn2 = Node(2)\r\nn3 = Node(3)\r\nn4 = Node(4)\r\nn5 = Node(5)\r\nn6 = Node(6)\r\n\r\nn1.children = [n3, n2, n4]\r\nn3.children = [n5, n6]\r\n\r\ns = Solution()\r\nprint(s.postorder(n1))\r\n", "sub_path": "leetcode/t000590_2.py", "file_name": "t000590_2.py", "file_ext": "py", "file_size_in_byte": 700, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "typing.List", "line_number": 11, "usage_type": "name"}]} +{"seq_id": "255609090", "text": "import math\nimport tensorflow as tf\nfrom termcolor import colored as c, cprint\nimport numpy as np\nfrom tensorflow.examples.tutorials.mnist import input_data\n\nmnist = input_data.read_data_sets(\"MNIST_data/\", one_hot=True)\n\nfrom . import helpers\n\n### helper functions\nfrom functools import reduce\n\n\ndef fc_layer(x, weight_shape, bias_shape, layer_name):\n with tf.name_scope(layer_name):\n # initializing at 0 is no-good.\n norm = math.sqrt(float(\n reduce(lambda v, e: v * e, weight_shape)\n ))\n weight = tf.Variable(\n tf.truncated_normal(weight_shape,\n mean=0.5,\n stddev=1.0 / norm),\n name='weight')\n bias = tf.Variable(tf.zeros(bias_shape), name='bias')\n activation = tf.matmul(x, weight) + bias\n return weight, bias, activation\n\n\n# main network build stages\ndef inference():\n x = tf.placeholder(tf.float32, shape=[None, 784], name='input')\n image = tf.reshape(x, [-1, 28, 28, 1])\n\n with tf.name_scope('conv_layer_1'):\n W_conv1 = helpers.weight_variable([5, 5, 1, 32], 'W_conv1')\n b_conv1 = helpers.bias_variable([32], 'bias_conv1')\n # alphas_conv1 = helpers.bias_variable([32], 'alpha_conv1')\n layer_conv_1 = tf.nn.softplus(helpers.conv2d(image, W_conv1) + b_conv1)\n stage_1_pool = helpers.max_pool_2x2(layer_conv_1)\n\n with tf.name_scope('conv_layer_2'):\n W_conv2 = helpers.weight_variable([5, 5, 32, 64], \"W_conv2\")\n b_conv2 = helpers.bias_variable([64], 'bias_conv2')\n # alphas_conv3 = helpers.bias_variable([64], 'alpha_conv3')\n layer_conv_2 = tf.nn.softplus(helpers.conv2d(stage_1_pool, W_conv2) + b_conv2)\n stage_2_pool = helpers.max_pool_2x2(layer_conv_2)\n stage_2_pool_flat = tf.reshape(stage_2_pool, [-1, 7 * 7 * 64])\n\n with tf.name_scope('conv_layer_3'):\n W_conv3 = helpers.weight_variable([5, 5, 64, 128], \"W_conv3\")\n b_conv3 = helpers.bias_variable([128], 'bias_conv3')\n # alphas_conv3 = helpers.bias_variable([64], 'alpha_conv3')\n layer_conv_3 = tf.nn.softplus(helpers.conv2d(stage_2_pool, W_conv3) + b_conv3)\n stage_3_pool = helpers.max_pool_2x2(layer_conv_3)\n\n stage_3_pool_flat = tf.reshape(stage_3_pool, [-1, 4 * 4 * 128])\n\n with tf.name_scope('fc_layer_1'):\n W_fc1 = helpers.weight_variable([4 * 4 * 128, 2], \"W_fc1\")\n b_fc1 = helpers.bias_variable([2], 'bias_fc1')\n output = tf.nn.softplus(tf.matmul(stage_3_pool_flat, W_fc1) + b_fc1)\n\n # with tf.name_scope('fc_output'):\n # W_output = helpers.weight_variable([500, 10], \"W_putput\")\n # b_output = helpers.bias_variable([10], 'bias_output')\n # output = tf.nn.softplus(tf.matmul(h_fc1, W_output) + b_output)\n\n # with tf.name_scope('output'):\n # W_output = helpers.weight_variable([2, 10], \"W_output\")\n # b_output = helpers.bias_variable([10])\n # output = tf.nn.softplus(tf.matmul(h_fc2, W_output) + b_output)\n\n return x, output\n\n\ndef loss(deep_features):\n with tf.name_scope('softmax_loss'):\n batch_labels = tf.placeholder(tf.float32, name='labels')\n W_loss = helpers.weight_variable([2, 10], \"W_loss\")\n bias_loss = tf.Variable(\n tf.truncated_normal(shape=[10], stddev=1e-2, mean=1e-1), 'bias_loss')\n # Note: we don't use the bias here because it does not affect things. removing the\n # bias also makes the analysis simpler.\n logits = tf.matmul(deep_features, W_loss) + bias_loss\n cross_entropy = - tf.reduce_mean(\n tf.mul(batch_labels, tf.nn.log_softmax(logits)),\n reduction_indices=[1]\n )\n xentropy_mean = tf.reduce_mean(cross_entropy, name=\"xentropy_mean\")\n tf.scalar_summary(xentropy_mean.op.name, xentropy_mean)\n\n return batch_labels, logits, xentropy_mean\n\n\ndef training(loss, learning_rate):\n with tf.name_scope('training'):\n global_step = tf.Variable(0, name='global_step', trainable=False)\n # optimizer = tf.train.GradientDescentOptimizer(learning_rate)\n optimizer = tf.train.AdamOptimizer(learning_rate)\n train_op = optimizer.minimize(loss, global_step=global_step)\n # optimizer = tf.train.GradientDescentOptimizer(learning_rate)\n # grads_and_vars = optimizer.compute_gradients(loss, tf.trainable_variables())\n # capped_grads_and_vars = [(tf.clip_by_value(grads, 1e-10, 1e10), vars) for grads, vars in grads_and_vars]\n # train_op = optimizer.apply_gradients(capped_grads_and_vars)\n return train_op, global_step\n\n\ndef evaluation(logits, labels):\n correct = tf.nn.in_top_k(logits, tf.cast(tf.argmax(labels, dimension=1), dtype=tf.int32), 1)\n accuracy = tf.reduce_mean(tf.cast(correct, tf.float64), name='accuracy')\n tf.scalar_summary(accuracy.op.name, accuracy)\n # Return the number of true entries.\n return accuracy\n", "sub_path": "Proj_Centroid_Loss_LeNet/convnet_2_hidden/network.py", "file_name": "network.py", "file_ext": "py", "file_size_in_byte": 4936, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "tensorflow.examples.tutorials.mnist.input_data.read_data_sets", "line_number": 7, "usage_type": "call"}, {"api_name": "tensorflow.examples.tutorials.mnist.input_data", "line_number": 7, "usage_type": "name"}, {"api_name": "tensorflow.name_scope", "line_number": 16, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 18, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 19, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 21, "usage_type": "call"}, {"api_name": "tensorflow.truncated_normal", "line_number": 22, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 26, "usage_type": "call"}, {"api_name": "tensorflow.zeros", "line_number": 26, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 27, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 33, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 33, "usage_type": "attribute"}, {"api_name": "tensorflow.reshape", "line_number": 34, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.nn.softplus", "line_number": 40, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 43, "usage_type": "call"}, {"api_name": "tensorflow.nn.softplus", "line_number": 47, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 47, "usage_type": "attribute"}, {"api_name": "tensorflow.reshape", "line_number": 49, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.nn.softplus", "line_number": 55, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 55, "usage_type": "attribute"}, {"api_name": "tensorflow.reshape", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 60, "usage_type": "call"}, {"api_name": "tensorflow.nn.softplus", "line_number": 63, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 63, "usage_type": "attribute"}, {"api_name": "tensorflow.matmul", "line_number": 63, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 79, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 80, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 80, "usage_type": "attribute"}, {"api_name": "tensorflow.Variable", "line_number": 82, "usage_type": "call"}, {"api_name": "tensorflow.truncated_normal", "line_number": 83, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 87, "usage_type": "call"}, {"api_name": "tensorflow.mul", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.nn.log_softmax", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 88, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_mean", "line_number": 91, "usage_type": "call"}, {"api_name": "tensorflow.scalar_summary", "line_number": 92, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 98, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 99, "usage_type": "call"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 101, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 101, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.in_top_k", "line_number": 111, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 111, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 111, "usage_type": "call"}, {"api_name": "tensorflow.argmax", "line_number": 111, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 111, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_mean", "line_number": 112, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 112, "usage_type": "call"}, {"api_name": "tensorflow.float64", "line_number": 112, "usage_type": "attribute"}, {"api_name": "tensorflow.scalar_summary", "line_number": 113, "usage_type": "call"}]} +{"seq_id": "514650363", "text": "\"\"\"\r\nDeveloper: Jc\r\nGoal: To Create a Simple Game Engine /Game FrameWork To help me make simple 2d games using pygame \r\n\r\nTodo: \r\n\t\t- Colors \t\t\t\t********DONE********\r\n\r\n\t\t- State class \t\t\t********DONE********\r\n\r\n\t\t- Entity Class \t\t\t********DONE********\r\n\r\n\t\t- Button Class \t\t\t********DONE********\r\n\r\n\t \t- DisplayText \t\t\t********DONE********\r\n\r\n\t\t- Text box \t\t\t\t\r\n\r\n\t\t- Animation \r\n\r\n\t\t- Collision Detection \tAABB Detection\r\n\r\n\t\t- Game Class \r\n\r\n\t\t- Create a Simple Website in Royalcraft.co/Jc (That Documents Your Simple Game Engine) (learn a little HTML/ Javascript)\r\n\r\n\r\n\"\"\"\r\n\r\nimport pygame\r\nimport random \r\nfrom enum import Enum \r\nfrom time import sleep\r\n\r\n#Initializing pygame \r\npygame.init()\r\n\r\n\r\n#pygame Colors \r\nBLACK = (0,0,0)\r\nWHITE = (255,255,255)\r\nLIGHTGRAY = (200,200,200)\r\nGRAY = (150,150,150)\r\nDARKGRAY = (75,75,75)\r\nRED = (200,0,0)\r\nGREEN = (0,200,0)\r\nBLUE = (0,0,200)\r\nPURPLE = (100,0,100)\r\nLIGHTBLUE = (0,200,255)\r\nPINK = (230,0,230)\r\n\r\n\r\n\r\n#Base Class For All States \r\nclass State:\r\n\r\n\tdef __init__(self):\r\n\t\tself.isActive = False #initially False \r\n\r\n\tdef update(self,mouseState):\r\n\t\tpass\r\n\r\n\tdef draw(self, screen):\r\n\t\tpass\r\n\r\n\r\n#Entity Class (all Entities inherit from this class)\r\nclass Entity(pygame.Surface):\r\n\r\n\tdef __init__(self,pos, size = (100,100), fillColor = PINK):\r\n\t\tsuper().__init__(size) #Initializing the size of the entity \r\n\t\tself.pos = pos #Saving the Position of the entity \r\n\t\tself.fill(fillColor)\r\n\t\tself.size = size #Saving the Size of the Entity\r\n\r\n\t#Update Place Holder \r\n\tdef update(self,mouseState):\r\n\t\tpass\t\t\r\n\r\n\t#Draws the Entity to the Screen \r\n\tdef draw(self,screen):\r\n\t\tscreen.blit(self,self.pos)\r\n\r\n\r\n#Allows the User to Easily display text on the screen \r\nclass DisplayText:\r\n\tdef __init__(self, message, pos, textColor = BLACK, sizeOfText = 30, fontFile = \"BebasNeue-Regular.ttf\"):\r\n\t\tself.font = pygame.font.Font(fontFile,sizeOfText)\r\n\t\t#Saving the textColor\r\n\t\tself.textColor = textColor\r\n\t\t#Creating a text surface object on which text is drawn on to \r\n\t\tself.text = self.font.render(message, True,textColor)\r\n\t\t#Saving the position of the text\r\n\t\tself.pos = pos \r\n\t\t#Saving the message \r\n\t\tself.message = message\r\n\r\n\t#Changes the Message of the DisplayText Object \r\n\tdef update(self, newMessage):\r\n\t\t#Ensures it only changes the message when it has to \r\n\t\tif self.message != newMessage:\r\n\t\t\tself.text = self.font.render(newMessage, True, self.textColor)\r\n\t\t\tself.message = newMessage\r\n\r\n\t#Draws the DisplayText Object To the Screen \r\n\tdef draw(self,screen):\r\n\t\tscreen.blit(self.text, self.pos)\r\n\r\n#Compacts all the colors of the button into a simple data structure \r\nclass ButtonColor:\r\n\tdef __init__(self, idleColor = WHITE, hoverColor = GRAY, pressedColor = DARKGRAY):\r\n\t\tself.idleColor = idleColor\r\n\t\tself.hoverColor = hoverColor\r\n\t\tself.pressedColor = pressedColor\r\n\r\n#Keeps Track of the Button State \r\nclass ButtonState(Enum):\r\n\tIDLE = 1\r\n\tHOVER = 2\r\n\tPRESSED = 3\r\n#Button Class \r\nclass Button(Entity):\r\n\r\n\tdef __init__(self, message, pos, size = (200,50), buttonColor = ButtonColor()):\r\n\t\tsuper().__init__(pos,size)\r\n\r\n\t\tself.message = message\r\n\t\tself.buttonColor = buttonColor \r\n\t\tself.buttonState = ButtonState.IDLE #Initially The Button is Idle \r\n\r\n\t\t#message, pos, textColor = BLACK, sizeOfText = 30, fontFile = \"BebasNeue-Regular.ttf\"\r\n\t\ttextPos = (pos[0] + 50, pos[1] )\r\n\t\tself.text = DisplayText(self.message,textPos, BLACK, 50)\r\n\r\n\t\tself.isPressed = False #Initially the Button is Not Pressed \r\n\r\n\tdef update(self, mouseState):\r\n\r\n\t\tself.buttonState = ButtonState.IDLE #If Nothing Happening then its IDLE \r\n\r\n\t\t#Getting the Position of the Mouse \r\n\t\tmousePos = pygame.mouse.get_pos()\r\n\r\n\t\t#Checking if the Mouse is Hovering over the Button \r\n\t\tif mousePos[0] >= self.pos[0] and mousePos[0] <= (self.pos[0] + self.get_width()):\r\n\t\t\tif mousePos[1] > self.pos[1] and mousePos[1] <= (self.pos[1] + self.get_height()):\r\n\t\t\t\t#Mouse is Currently Hovering over the Button \r\n\t\t\t\tself.buttonState = ButtonState.HOVER\r\n\r\n\t\t\t\t#Check if the button is being pressed \r\n\t\t\t\tif mouseState[0]:\r\n\t\t\t\t\tself.buttonState = ButtonState.PRESSED #Switching the Button State \r\n\t\t\t\t\tself.isPressed = True # \r\n\r\n\tdef draw(self, screen, outline = True):\r\n\r\n\t\t#Changing the Button Color Based off its State \r\n\t\tif self.buttonState == ButtonState.IDLE:\r\n\t\t\tfillColor = self.buttonColor.idleColor\r\n\t\tif self.buttonState == ButtonState.HOVER:\r\n\t\t\tfillColor = self.buttonColor.hoverColor\r\n\t\tif self.buttonState == ButtonState.PRESSED:\r\n\t\t\tfillColor = self.buttonColor.pressedColor\r\n\r\n\t\tself.fill(fillColor) #Changing the Color Based off the Button State \r\n\t\t\r\n\t\t#Draws the Button Outline \r\n\t\tif outline == True:\r\n\t\t\tpygame.draw.rect(screen, (0,0,0), (self.pos[0] - 2, self.pos[1] -2, self.get_width() + 4, self.get_height() + 4), 0)\r\n\t\t\tpass\r\n\t\t#Drawing the Button (Using the Entity draw Function)\r\n\t\tsuper().draw(screen)\r\n\t\t#Drawing the Text Of the button \r\n\t\tself.text.draw(screen)\r\n\r\n\r\ndef main():\r\n\r\n\tscreen = pygame.display.set_mode((800,800)) #Main Screen \r\n\tscreen.fill(RED)\r\n\r\n\ttestEntity = Entity((200,20), (10,300))\r\n\r\n\r\n\t#self, message, pos, textColor = BLACK, sizeOfText = 30, fontFile = \"BebasNeue-Regular.ttf\"\r\n\ttext = DisplayText(\"TEST\" , (400,400), LIGHTBLUE, 100)\r\n\r\n\r\n\t#Testing the Button Class \r\n\t#self, message, pos, size = (100,50), buttonColor = ButtonColor()\r\n\ttestButton = Button(\"Play\", (400,400))\r\n\r\n\r\n\t#Temp Variable Used to Count the DisplayText\r\n\tx = 0\r\n\r\n\r\n\t#Game Loop\r\n\tisOver = False \r\n\twhile(not isOver):\r\n\r\n\t\t#Event Loop Handler \r\n\t\tfor event in pygame.event.get():\r\n\t\t\tif event.type == pygame.QUIT:\r\n\t\t\t\tisOver = True #Stop The game loop \r\n\r\n\t\tmouseState = pygame.mouse.get_pressed()\r\n\r\n\t\tkeys = pygame.key.get_pressed() #Getting a list of booleans of all the keys in the game \r\n\r\n\t\tscreen.fill(RED)\r\n\r\n\r\n\t\t#Testing the Button \r\n\t\ttestButton.update(mouseState)\r\n\t\ttestButton.draw(screen)\r\n\r\n\r\n\t\tpygame.display.update() #Updating the display module \r\n\r\n\tpygame.quit() #Quitting pygame \r\n\r\nif __name__ == \"__main__\":\r\n\tmain()", "sub_path": "gameEngine/engine_1.py", "file_name": "engine_1.py", "file_ext": "py", "file_size_in_byte": 5999, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "pygame.init", "line_number": 35, "usage_type": "call"}, {"api_name": "pygame.Surface", "line_number": 67, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 87, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 87, "usage_type": "attribute"}, {"api_name": "enum.Enum", "line_number": 116, "usage_type": "name"}, {"api_name": "pygame.mouse.get_pos", "line_number": 141, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 141, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 168, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 168, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 178, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 178, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 202, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 202, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 203, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pressed", "line_number": 206, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 206, "usage_type": "attribute"}, {"api_name": "pygame.key.get_pressed", "line_number": 208, "usage_type": "call"}, {"api_name": "pygame.key", "line_number": 208, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 218, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 218, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 220, "usage_type": "call"}]} +{"seq_id": "464243898", "text": "\" Import of modules \"\r\nimport numpy.random as np\r\nimport xlsxwriter\r\n\r\n\" Initialization \"\r\nprob = [0] * 5 #Range of mutational probabilities\r\na = len ( prob )\r\nfor i in range ( 0, a ):\r\n prob [i] = 10 ** -(i+4)\r\n\r\ngrowth_rate = [0] * 20 #Range of growth rates\r\nb = len ( growth_rate )\r\nfor i in range ( 0, b ):\r\n growth_rate [i] = 0.25 * (i+1)\r\n\r\nmut_pop = [ [0 for i in range ( 0, b )] for j in range ( 0, a ) ] #Threshold values for mutant population sizes\r\nproduct = [ [0 for i in range ( 0, b )] for j in range ( 0, a ) ] #Product of p and mut_pop will be stored here\r\nmutant_prev = [ [0 for i in range ( 0, b )] for j in range ( 0, a ) ] #Size of mutant population at the step preceding transition\r\n\r\niter_num = 100 #Number of iterations for the entire simuation\r\nn = 8 * 10 ** 9 #Stem cell carrying capacity for the tissue\r\n\r\n\" Main simulation \"\r\n\r\nfor i in range ( 0,a ):\r\n p = prob [i]\r\n \r\n for j in range ( 0, b ):\r\n g = growth_rate [j]\r\n time = 100 #Duration of each simulation\r\n \r\n for x in range ( 0, iter_num ):\r\n t = 0 #Index to track time\r\n n_mut = [0] * time\r\n m = 0 #Initial mutant population\r\n g_total = 0 #Total number of mutant cells with a particular set of mutations\r\n m_prev = 0 #Number of cells in the previous step\r\n m_hold = 0 #Holding variable for previous population size\r\n avg = 0 #Average mutant population size per mutation\r\n p_mut = 1 - ( (1-p) ** n ) #Initial probabiltiy of first mutation arising in the population\r\n\r\n if p_mut > np.random_sample ( ): #At t = 0\r\n n_mut [0] += 1\r\n m = 1\r\n else:\r\n m = 0\r\n\r\n for t in range ( 1, time ): #From t = 1 to end of time\r\n n_mut [t] = n_mut [t-1]\r\n m_hold = m\r\n m += ( ( m*g ) * ( 1 - ( m/n ) ) )\r\n p_mut = 1 - ( (1-p) ** m )\r\n\r\n if p_mut > np.random_sample ( ):\r\n n_mut [t] += 1\r\n m_prev += m_hold\r\n g_total += m\r\n m = 1\r\n \r\n den = n_mut [time-1] - 1 #Number of mutations - 1 = number of transitions between subsequent mutant populations\r\n avg = g_total / den #Average mutant population size per iteration\r\n m_prev = m_prev / den\r\n mut_pop [i][j] += avg / iter_num #Average mutant population size over all iterations\r\n mutant_prev [i][j] += m_prev / iter_num\r\n\r\n\" Calculation of the product, p * mut_pop \"\r\nfor i in range ( 0, a ):\r\n for j in range ( 0, b ):\r\n product [i][j] = prob [i] * mutant_prev [i][j]\r\n\r\n\"\"\" Export to Excel \"\"\"\r\nworkbook = xlsxwriter.Workbook ( 'Model.xlsx' )\r\nworksheet = workbook.add_worksheet ( )\r\nrow = 0\r\ncol = 0\r\n\r\nfor i in range ( 0, a ):\r\n for j in range ( 0, b ):\r\n worksheet.write ( row, col, mut_pop [i][j] )\r\n worksheet.write ( row + 7, col, mutant_prev [i][j] )\r\n col += 1\r\n row += 1\r\n col = 0\r\n\r\nworkbook.close ( )\r\n \r\n\r\n \r\n \r\n \r\n \r\n", "sub_path": "MGW/codes/old-codes/cancer_incidence_model05_v2 (2017_06_10 00_46_26 UTC).py", "file_name": "cancer_incidence_model05_v2 (2017_06_10 00_46_26 UTC).py", "file_ext": "py", "file_size_in_byte": 3154, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "numpy.random.random_sample", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 42, "usage_type": "name"}, {"api_name": "numpy.random.random_sample", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 54, "usage_type": "name"}, {"api_name": "xlsxwriter.Workbook", "line_number": 72, "usage_type": "call"}]} +{"seq_id": "86249924", "text": "import streamlit as st\n\ndef main():\n uploaded_files = st.file_uploader(\"Choose a CSV file\", accept_multiple_files=True)\n for uploaded_file in uploaded_files:\n bytes_data = uploaded_file.read()\n st.write(\"filename:\", uploaded_file.name)\n st.write(bytes_data)\n\nif __name__ == \"__main__\":\n #application.run()\n main()\n", "sub_path": "app/upload.py", "file_name": "upload.py", "file_ext": "py", "file_size_in_byte": 347, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "streamlit.file_uploader", "line_number": 4, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 7, "usage_type": "call"}, {"api_name": "streamlit.write", "line_number": 8, "usage_type": "call"}]} +{"seq_id": "403726527", "text": "import numpy as np\nfrom typing import List\nfrom classifier import Classifier\n\nclass DecisionTree(Classifier):\n\tdef __init__(self):\n\t\tself.clf_name = \"DecisionTree\"\n\t\tself.root_node = None\n\n\tdef train(self, features: List[List[float]], labels: List[int]):\n\t\t# init.\n\t\tassert(len(features) > 0)\n\t\tself.feautre_dim = len(features[0])\n\t\tnum_cls = np.max(labels)+1\n\n\t\t# build the tree\n\t\tself.root_node = TreeNode(features, labels, num_cls)\n\t\tif self.root_node.splittable:\n\t\t\tself.root_node.split()\n\n\t\treturn\n\t\t\n\tdef predict(self, features: List[List[float]]) -> List[int]:\n\t\ty_pred = []\n\t\tfor feature in features:\n\t\t\ty_pred.append(self.root_node.predict(feature))\n\t\treturn y_pred\n\n\tdef print_tree(self, node=None, name='node 0', indent=''):\n\t\tif node is None:\n\t\t\tnode = self.root_node\n\t\tprint(name + '{')\n\t\t\n\t\tstring = ''\n\t\tfor idx_cls in range(node.num_cls):\n\t\t\tstring += str(node.labels.count(idx_cls)) + ' '\n\t\tprint(indent + ' num of sample / cls: ' + string)\n\n\t\tif node.splittable:\n\t\t\tprint(indent + ' split by dim {:d}'.format(node.dim_split))\n\t\t\tfor idx_child, child in enumerate(node.children):\n\t\t\t\tself.print_tree(node=child, name= ' '+name+'/'+str(idx_child), indent=indent+' ')\n\t\telse:\n\t\t\tprint(indent + ' cls', node.cls_max)\n\t\tprint(indent+'}')\n\n\nclass TreeNode(object):\n\tdef __init__(self, features: List[List[float]], labels: List[int], num_cls: int):\n\t\tself.features = features\n\t\tself.labels = labels\n\t\tself.children = []\n\t\tself.num_cls = num_cls\n\n\t\tcount_max = 0\n\t\tfor label in np.unique(labels):\n\t\t\tif self.labels.count(label) > count_max:\n\t\t\t\tcount_max = labels.count(label)\n\t\t\t\tself.cls_max = label # majority of current node\n\n\t\tif len(np.unique(labels)) < 2:\n\t\t\tself.splittable = False\n\t\telse:\n\t\t\tself.splittable = True\n\n\t\tself.dim_split = None # the index of the feature to be split\n\n\t\tself.feature_uniq_split = None # the possible unique values of the feature to be split\n\n\n\tdef split(self):\n\t\tdef conditional_entropy(branches: List[List[int]]) -> float:\n\t\t\t'''\n\t\t\tbranches: C x B array, \n\t\t\t\t\t C is the number of classes,\n\t\t\t\t\t B is the number of branches\n\t\t\t\t\t it stores the number of \n\t\t\t\t\t corresponding training samples \n\t\t\t\t\t e.g.\n\t\t\t\t\t ○ ○ ○ ○\n\t\t\t\t\t ● ● ● ●\n\t\t\t\t\t ┏━━━━┻━━━━┓\n\t\t\t\t ○ ○ ○ ○\n\t\t\t\t ● ● ● ●\n\t\t\t\t \n\t\t\t\t branches = [[2,2], [4,0]]\n\t\t\t'''\n\t\t\t########################################################\n\t\t\t# TODO: compute the conditional entropy\n\t\t\t########################################################\n\t\t\tbranch_arr = np.array(branches)\n\t\t\tbranch_T = np.transpose(branch_arr).tolist()\n\t\t\ttotal = float(np.sum(branch_arr)) # 8\n\t\t\tcd_entropy = 0.0\n\n\t\t\tfor b in branch_T:\n\t\t\t br_total = float(sum(b)) # 6 and then for = 2\n\t\t\t weight = br_total / total\n\t\t\t if br_total == 0:\t# can't divide by zero so go to next iteration\n\t\t\t continue\n\t\t\t for cls_val in b:\n\t\t\t \tif cls_val > 0:\n\t\t\t \t prob = float(cls_val) / br_total # 2/6 and then 2/2\n\t\t\t \t cd_entropy -= prob * np.log(prob) * weight\n\t\t\treturn cd_entropy\n\n\t\tif not self.splittable:\n\t\t\treturn\n\n\t\tfeatures = np.array(self.features)\n\t\tentr = []\n\n\t\tfor idx_dim in range(len(self.features[0])):\n\t\t############################################################\n\t\t# TODO: compare each split using conditional entropy\n\t\t# find the best split\n\t\t############################################################\n\t\t\tfeat = features[:, idx_dim]\n\t\t\tdivn = []\n\n\t\t\tfor t in np.unique(feat):\n\t\t\t\tt_features = feat[np.where(feat == t)]\n\t\t\t\tt_labels = np.array(self.labels)[np.where(feat == t)]\n\t\t\t\tbran = []\n\n\t\t\t\tfor i in range(self.num_cls):\n\t\t\t\t\tbran.append(np.sum(t_labels == i))\n\t\t\t\tdivn.append(bran)\n\n\t\t\tentr.append(conditional_entropy(np.array(divn).T.tolist()))\n\n\t\tself.dim_split = np.argmin(entr)\n\t\tfeat = features[:,self.dim_split]\n\t\tself.feature_uniq_split = np.unique(feat).tolist()\n\t\t#print('self.feature unique = ',self.feature_uniq_split)\n\n\t\t############################################################\n\t\t# TODO: split the node, add child nodes\n\t\t############################################################\n\t\tif len(np.unique(feat)) > 1:\n\t\t\tfor t in np.unique(feat):\n\t\t\t\tt_features = features[np.where(feat == t)].tolist()\n\t\t\t\tt_labels = np.array(self.labels)[np.where(feat == t)].tolist()\n\t\t\t\tself.children.append(TreeNode(t_features, t_labels, self.num_cls))\n\t\telse:\n\t\t\tself.splittable = False\n\n\t\t# split the child nodes\n\t\tfor child in self.children:\n\t\t\tif child.splittable:\n\t\t\t\tchild.split()\n\n\t\treturn\n\n\tdef predict(self, feature: List[int]) -> int:\n\t\tif self.splittable:\n\t\t\t# print(feature)\n\t\t\tidx_child = self.feature_uniq_split.index(feature[self.dim_split])\n\t\t\treturn self.children[idx_child].predict(feature)\n\t\telse:\n\t\t\treturn self.cls_max\n\n\n\n", "sub_path": "machine_learning_assignments/P3/decision_tree.py", "file_name": "decision_tree.py", "file_ext": "py", "file_size_in_byte": 4813, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "classifier.Classifier", "line_number": 5, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 10, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 14, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 23, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 49, "usage_type": "name"}, {"api_name": "numpy.unique", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 61, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 72, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 143, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 155, "usage_type": "name"}]} +{"seq_id": "506787415", "text": "import base64\nimport hmac\nimport time\nimport traceback\nimport typing\nimport re\nimport jwt\nimport datetime\nfrom jwt.exceptions import ExpiredSignatureError, DecodeError\nfrom starlette.requests import Request\nfrom starlette.responses import JSONResponse\nfrom app.core.utils.exceptions import APIException\nfrom starlette.status import HTTP_500_INTERNAL_SERVER_ERROR\nfrom app.core.middlewares.api_logger import api_logger\n\nEXCEPT_PATH_REGEX = \"^(/docs|/redoc|/api/auth)\"\nEXCEPT_PATH_LIST = [\"/\", \"/openapi.json\"]\n\ndef to_dict(data):\n return data.__dict__['__data__']\n\nasync def access_control(request: Request, call_next):\n request.state.req_time = datetime.date.today()\n request.state.start = time.time()\n request.state.inspect = None\n request.state.user = None\n request.state.service = None\n\n ip = request.headers[\"x-forwarded-for\"] if \"x-forwarded-for\" in request.headers.keys() else request.client.host\n request.state.ip = ip.split(\",\")[0] if \",\" in ip else ip\n headers = request.headers\n cookies = request.cookies\n\n url = request.url.path\n if await url_pattern_check(url, EXCEPT_PATH_REGEX) or url in EXCEPT_PATH_LIST:\n response = await call_next(request)\n if url != \"/\":\n await api_logger(request=request, response=response)\n return response\n\n try:\n if url.startswith(\"/test\"):\n # test 인경우 헤더로 토큰 검사 => 너가 원하는대로 고쳐서 쓰면 됨\n if url.startswith(\"/api/services\"):\n qs = str(request.query_params)\n qs_list = qs.split(\"&\")\n\n response = await call_next(request)\n return response\n\n else:\n # 템플릿 렌더링인 경우 쿠키에서 토큰 검사\n cookies[\"Authorization\"] = \"Bearer\"\n\n response = await call_next(request)\n await api_logger(request=request, response=response)\n except APIException as e:\n response = await test_exception_handler(e)\n await api_logger(request=request, error=e)\n except Exception as e:\n\n error = await exception_handler(e)\n\n # error_dict = dict(status=error.status_code, msg=error.msg, detail=error.detail, code=error.code)\n # response = JSONResponse(status_code=error.status_code, content=error_dict)\n print(e.args)\n print(traceback.format_exc())\n error.status_code, content = HTTP_500_INTERNAL_SERVER_ERROR, {\n \"statusCode\": 500,\n \"error\": \"Bad Request\",\n \"message\": \"잠시 후 다시 시도해 주시길 바랍니다.\"\n }\n response = JSONResponse(status_code=error.status_code, content=content)\n try:\n request.errorMessage = e.args\n except:\n pass\n await api_logger(request=request, error=error)\n\n return response\n\n\nasync def url_pattern_check(path, pattern):\n result = re.match(pattern, path)\n if result:\n return True\n return False\n\nasync def exception_handler(error: Exception):\n print(error)\n return error\n\nasync def test_exception_handler(error: APIException):\n error_dict = dict(status=error.status_code, msg=error.msg, detail=error.detail, code=error.code)\n res = JSONResponse(status_code=error.status_code,content=error_dict)\n return res", "sub_path": "app/core/middlewares/token_validator.py", "file_name": "token_validator.py", "file_ext": "py", "file_size_in_byte": 3304, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "starlette.requests.Request", "line_number": 22, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 23, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 24, "usage_type": "call"}, {"api_name": "app.core.middlewares.api_logger.api_logger", "line_number": 38, "usage_type": "call"}, {"api_name": "app.core.middlewares.api_logger.api_logger", "line_number": 56, "usage_type": "call"}, {"api_name": "app.core.utils.exceptions.APIException", "line_number": 57, "usage_type": "name"}, {"api_name": "app.core.middlewares.api_logger.api_logger", "line_number": 59, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 67, "usage_type": "call"}, {"api_name": "starlette.status.HTTP_500_INTERNAL_SERVER_ERROR", "line_number": 68, "usage_type": "name"}, {"api_name": "starlette.responses.JSONResponse", "line_number": 73, "usage_type": "call"}, {"api_name": "app.core.middlewares.api_logger.api_logger", "line_number": 78, "usage_type": "call"}, {"api_name": "re.match", "line_number": 84, "usage_type": "call"}, {"api_name": "app.core.utils.exceptions.APIException", "line_number": 93, "usage_type": "name"}, {"api_name": "starlette.responses.JSONResponse", "line_number": 95, "usage_type": "call"}]} +{"seq_id": "178375027", "text": "import sys, traceback\n\nimport discord\nfrom discord.ext import commands\n\nimport settings\n\ninitial_extensions = [\n \"bot-gen\",\n \"bot-help\",\n \"bot-errors\",\n\n \"bot-roles\",\n \"bot-money\",\n]\n\nbot = commands.Bot(command_prefix = \"-\", case_insensitive = True)\nbot.remove_command(\"help\")\n\nif __name__ == \"__main__\":\n for extension in initial_extensions:\n try:\n bot.load_extension(extension)\n except Exception as e:\n #print(f\"Failed to load extension {extension}.\", file = sys.stderr)\n traceback.print_exc()\n\n@bot.event\nasync def on_ready():\n print(bot.user.name)\n print(f\"\\nLogged in as: {bot.user.name} - {bot.user.id}\\ndiscord.py version: {discord.__version__}\\n\")\n await bot.change_presence(activity = discord.Game(name=\"with the matrix\"))\n print(\"Successfully logged in and booted...!\")\n\nbot.run(\"NjE3MjQwNTMxNjQ2NDE0ODQ4.XZ6-RA.25BAY_eiS1fNInIwTDaxiC_JIBw\", bot=True, reconnect=True)\n", "sub_path": "bot-DESKTOP-VJFKRT9.py", "file_name": "bot-DESKTOP-VJFKRT9.py", "file_ext": "py", "file_size_in_byte": 955, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "discord.ext.commands.Bot", "line_number": 17, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 17, "usage_type": "name"}, {"api_name": "traceback.print_exc", "line_number": 26, "usage_type": "call"}, {"api_name": "discord.__version__", "line_number": 31, "usage_type": "attribute"}, {"api_name": "discord.Game", "line_number": 32, "usage_type": "call"}]} +{"seq_id": "505136630", "text": "from binascii import hexlify, unhexlify\n\nimport structlog\nfrom eth_utils import (\n is_binary_address,\n to_checksum_address,\n to_normalized_address,\n to_canonical_address,\n)\nfrom web3.utils.filters import Filter\nfrom web3.exceptions import BadFunctionCallOutput\n\nfrom raiden.blockchain.abi import (\n CONTRACT_MANAGER,\n CONTRACT_REGISTRY,\n EVENT_TOKEN_ADDED,\n)\nfrom raiden.exceptions import (\n NoTokenManager,\n TransactionThrew,\n)\nfrom raiden.network.proxies.channel_manager import ChannelManager\nfrom raiden.network.rpc.client import check_address_has_code\nfrom raiden.network.rpc.smartcontract_proxy import ContractProxy\nfrom raiden.network.rpc.transactions import check_transaction_threw\nfrom raiden.utils import (\n pex,\n privatekey_to_address,\n typing,\n)\n\nlog = structlog.get_logger(__name__) # pylint: disable=invalid-name\n\ntry:\n from eth_tester.exceptions import TransactionFailed\nexcept ModuleNotFoundError:\n TransactionFailed = Exception()\n\n\nclass Registry:\n def __init__(\n self,\n jsonrpc_client,\n registry_address,\n ):\n # pylint: disable=too-many-arguments\n contract = jsonrpc_client.new_contract(\n CONTRACT_MANAGER.get_contract_abi(CONTRACT_REGISTRY),\n to_normalized_address(registry_address),\n )\n proxy = ContractProxy(jsonrpc_client, contract)\n\n if not is_binary_address(registry_address):\n raise ValueError('registry_address must be a valid address')\n\n check_address_has_code(jsonrpc_client, registry_address, 'Registry')\n\n CONTRACT_MANAGER.check_contract_version(\n proxy.contract.functions.contract_version().call(),\n CONTRACT_REGISTRY,\n )\n\n self.address = registry_address\n self.client = jsonrpc_client\n self.node_address = privatekey_to_address(self.client.privkey)\n self.proxy = proxy\n\n self.address_to_channelmanager = dict()\n self.token_to_channelmanager = dict()\n\n def manager_address_by_token(self, token_address):\n \"\"\" Return the channel manager address for the given token or None if\n there is no correspoding address.\n \"\"\"\n try:\n address = self.proxy.contract.functions.channelManagerByToken(\n to_checksum_address(token_address),\n ).call()\n except (BadFunctionCallOutput, TransactionFailed):\n check_address_has_code(self.client, self.address)\n return None\n\n return to_canonical_address(address)\n\n def add_token(self, token_address):\n if not is_binary_address(token_address):\n raise ValueError('token_address must be a valid address')\n\n log.info(\n 'add_token called',\n node=pex(self.node_address),\n token_address=pex(token_address),\n registry_address=pex(self.address),\n )\n\n transaction_hash = self.proxy.transact(\n 'addToken',\n self.address,\n token_address,\n )\n\n self.client.poll(unhexlify(transaction_hash))\n receipt_or_none = check_transaction_threw(self.client, transaction_hash)\n if receipt_or_none:\n log.info(\n 'add_token failed',\n node=pex(self.node_address),\n token_address=pex(token_address),\n registry_address=pex(self.address),\n )\n raise TransactionThrew('AddToken', receipt_or_none)\n\n manager_address = self.manager_address_by_token(token_address)\n\n if manager_address is None:\n log.info(\n 'add_token failed and check_transaction_threw didnt detect it',\n node=pex(self.node_address),\n token_address=pex(token_address),\n registry_address=pex(self.address),\n )\n\n raise RuntimeError('channelManagerByToken failed')\n\n log.info(\n 'add_token sucessful',\n node=pex(self.node_address),\n token_address=pex(token_address),\n registry_address=pex(self.address),\n manager_address=pex(manager_address),\n )\n\n return manager_address\n\n def token_addresses(self):\n addresses = self.proxy.contract.functions.tokenAddresses().call()\n return [\n to_canonical_address(address)\n for address in addresses\n ]\n\n def manager_addresses(self):\n addresses = self.proxy.contract.functions.channelManagerAddresses().call()\n return [\n to_canonical_address(address)\n for address in addresses\n ]\n\n def tokenadded_filter(\n self,\n from_block: typing.BlockSpecification = 0,\n to_block: typing.BlockSpecification = 'latest',\n ) -> Filter:\n topics = [CONTRACT_MANAGER.get_event_id(EVENT_TOKEN_ADDED)]\n\n registry_address_bin = self.proxy.contract_address\n return self.client.new_filter(\n registry_address_bin,\n topics=topics,\n from_block=from_block,\n to_block=to_block,\n )\n\n def manager(self, manager_address):\n \"\"\" Return a proxy to interact with a ChannelManagerContract. \"\"\"\n if not is_binary_address(manager_address):\n raise ValueError('manager_address must be a valid address')\n\n if manager_address not in self.address_to_channelmanager:\n manager = ChannelManager(\n self.client,\n manager_address,\n )\n\n token_address = manager.token_address()\n\n self.token_to_channelmanager[token_address] = manager\n self.address_to_channelmanager[manager_address] = manager\n\n return self.address_to_channelmanager[manager_address]\n\n def manager_by_token(self, token_address):\n \"\"\" Find the channel manager for `token_address` and return a proxy to\n interact with it.\n\n If the token is not already registered it raises `EthNodeCommunicationError`,\n since we try to instantiate a Channel manager with an empty address.\n \"\"\"\n if not is_binary_address(token_address):\n raise ValueError('token_address must be a valid address')\n\n if token_address not in self.token_to_channelmanager:\n check_address_has_code(self.client, token_address) # check that the token exists\n manager_address = self.manager_address_by_token(token_address)\n\n if manager_address is None:\n raise NoTokenManager(\n 'Manager for token 0x{} does not exist'.format(hexlify(token_address)),\n )\n\n manager = ChannelManager(\n self.client,\n manager_address,\n )\n\n self.token_to_channelmanager[token_address] = manager\n self.address_to_channelmanager[manager_address] = manager\n\n return self.token_to_channelmanager[token_address]\n", "sub_path": "raiden/network/proxies/registry.py", "file_name": "registry.py", "file_ext": "py", "file_size_in_byte": 6983, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "structlog.get_logger", "line_number": 32, "usage_type": "call"}, {"api_name": "eth_tester.exceptions.TransactionFailed", "line_number": 37, "usage_type": "name"}, {"api_name": "raiden.blockchain.abi.CONTRACT_MANAGER.get_contract_abi", "line_number": 48, "usage_type": "call"}, {"api_name": "raiden.blockchain.abi.CONTRACT_REGISTRY", "line_number": 48, "usage_type": "argument"}, {"api_name": "raiden.blockchain.abi.CONTRACT_MANAGER", "line_number": 48, "usage_type": "name"}, {"api_name": "eth_utils.to_normalized_address", "line_number": 49, "usage_type": "call"}, {"api_name": "raiden.network.rpc.smartcontract_proxy.ContractProxy", "line_number": 51, "usage_type": "call"}, {"api_name": "eth_utils.is_binary_address", "line_number": 53, "usage_type": "call"}, {"api_name": "raiden.network.rpc.client.check_address_has_code", "line_number": 56, "usage_type": "call"}, {"api_name": "raiden.blockchain.abi.CONTRACT_MANAGER.check_contract_version", "line_number": 58, "usage_type": "call"}, {"api_name": "raiden.blockchain.abi.CONTRACT_REGISTRY", "line_number": 60, "usage_type": "argument"}, {"api_name": "raiden.blockchain.abi.CONTRACT_MANAGER", "line_number": 58, "usage_type": "name"}, {"api_name": "raiden.utils.privatekey_to_address", "line_number": 65, "usage_type": "call"}, {"api_name": "eth_utils.to_checksum_address", "line_number": 77, "usage_type": "call"}, {"api_name": "web3.exceptions.BadFunctionCallOutput", "line_number": 79, "usage_type": "name"}, {"api_name": "eth_tester.exceptions.TransactionFailed", "line_number": 79, "usage_type": "name"}, {"api_name": "raiden.network.rpc.client.check_address_has_code", "line_number": 80, "usage_type": "call"}, {"api_name": "eth_utils.to_canonical_address", "line_number": 83, "usage_type": "call"}, {"api_name": "eth_utils.is_binary_address", "line_number": 86, "usage_type": "call"}, {"api_name": "raiden.utils.pex", "line_number": 91, "usage_type": "call"}, {"api_name": "raiden.utils.pex", "line_number": 92, "usage_type": "call"}, {"api_name": "raiden.utils.pex", "line_number": 93, "usage_type": "call"}, {"api_name": "binascii.unhexlify", "line_number": 102, "usage_type": "call"}, {"api_name": "raiden.network.rpc.transactions.check_transaction_threw", "line_number": 103, "usage_type": "call"}, {"api_name": "raiden.utils.pex", "line_number": 107, "usage_type": "call"}, {"api_name": "raiden.utils.pex", "line_number": 108, "usage_type": "call"}, {"api_name": "raiden.utils.pex", "line_number": 109, "usage_type": "call"}, {"api_name": "raiden.exceptions.TransactionThrew", "line_number": 111, "usage_type": "call"}, {"api_name": "raiden.utils.pex", "line_number": 118, "usage_type": "call"}, {"api_name": "raiden.utils.pex", "line_number": 119, "usage_type": "call"}, {"api_name": "raiden.utils.pex", "line_number": 120, "usage_type": "call"}, {"api_name": "raiden.utils.pex", "line_number": 127, "usage_type": "call"}, {"api_name": "raiden.utils.pex", "line_number": 128, "usage_type": "call"}, {"api_name": "raiden.utils.pex", "line_number": 129, "usage_type": "call"}, {"api_name": "raiden.utils.pex", "line_number": 130, "usage_type": "call"}, {"api_name": "eth_utils.to_canonical_address", "line_number": 138, "usage_type": "call"}, {"api_name": "eth_utils.to_canonical_address", "line_number": 145, "usage_type": "call"}, {"api_name": "raiden.utils.typing.BlockSpecification", "line_number": 151, "usage_type": "attribute"}, {"api_name": "raiden.utils.typing", "line_number": 151, "usage_type": "name"}, {"api_name": "raiden.utils.typing.BlockSpecification", "line_number": 152, "usage_type": "attribute"}, {"api_name": "raiden.utils.typing", "line_number": 152, "usage_type": "name"}, {"api_name": "raiden.blockchain.abi.CONTRACT_MANAGER.get_event_id", "line_number": 154, "usage_type": "call"}, {"api_name": "raiden.blockchain.abi.EVENT_TOKEN_ADDED", "line_number": 154, "usage_type": "argument"}, {"api_name": "raiden.blockchain.abi.CONTRACT_MANAGER", "line_number": 154, "usage_type": "name"}, {"api_name": "web3.utils.filters.Filter", "line_number": 153, "usage_type": "name"}, {"api_name": "eth_utils.is_binary_address", "line_number": 166, "usage_type": "call"}, {"api_name": "raiden.network.proxies.channel_manager.ChannelManager", "line_number": 170, "usage_type": "call"}, {"api_name": "eth_utils.is_binary_address", "line_number": 189, "usage_type": "call"}, {"api_name": "raiden.network.rpc.client.check_address_has_code", "line_number": 193, "usage_type": "call"}, {"api_name": "raiden.exceptions.NoTokenManager", "line_number": 197, "usage_type": "call"}, {"api_name": "binascii.hexlify", "line_number": 198, "usage_type": "call"}, {"api_name": "raiden.network.proxies.channel_manager.ChannelManager", "line_number": 201, "usage_type": "call"}]} +{"seq_id": "475649473", "text": "from django.db import models\n\n# Create your models here.\n\nfrom bases.models import ClassModelo\n\n\nclass Category(ClassModelo):\n description = models.CharField(\n max_length=100,\n help_text='Category Description',\n unique=True\n )\n\n def __srt__(self):\n return '{}'.format(self.description)\n\n def save(self):\n self.description = self.description.upper()\n\n super(Category, self).save()\n \n class Meta:\n verbose_name_plural = 'Categories'\n\n\nclass SubCategory(ClassModelo):\n category = models.ForeignKey(Category, on_delete=models.CASCADE)\n description = models.CharField(\n max_length=100,\n help_text='Description Category',\n unique=True\n )\n\n def __str__(self):\n return '{}:{}'.format(self.category.description, self.description)\n\n def save(self):\n self.description = self.description.upper()\n\n super(SubCategory, self).save()\n\n class Meta:\n verbose_name_plural = 'Sub Categories'\n unique_together = ('category', 'description')\n", "sub_path": "app/inv/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 1060, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "bases.models.ClassModelo", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 9, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 9, "usage_type": "name"}, {"api_name": "bases.models.ClassModelo", "line_number": 27, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 28, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 28, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 28, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 29, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 29, "usage_type": "name"}]} +{"seq_id": "13608599", "text": "from bs4 import BeautifulSoup\nfrom splinter import Browser\nfrom splinter.exceptions import ElementDoesNotExist\nfrom selenium import webdriver\nimport time\nimport requests\nimport pandas as pd\n\ndef init_browser():\n # @NOTE: Replace the path with your actual path to the chromedriver\n executable_path = {\"executable_path\": \"chromedriver\"}\n return Browser(\"chrome\", **executable_path, headless=False)\n\ndef scrape():\n mars = {}\n\n #-----------------------------------------------------------------------\n #Part 1: NASA Mars News\n url_news = 'https://mars.nasa.gov/news/?page=0&per_page=40&order=publish_date+desc%2Ccreated_at+desc&search=&category=19%2C165%2C184%2C204&blank_scope=Latest'\n\n browser_sele = webdriver.Chrome()\n browser_sele.get(url_news)\n soup_selenium = BeautifulSoup(browser_sele.page_source, \"html.parser\")\n\n mars[\"news_title\"] = soup_selenium.find('div', class_=\"content_title\").text.strip()\n time.sleep(2)\n mars[\"news_p\"] = soup_selenium.find('div', class_=\"article_teaser_body\").text.strip()\n\n print(\"---------------------------------------------------------------------------------------\")\n print(\"Part 1: NASA Mars News Complete! Next: Part 3: Mars Weather \")\n\n browser_sele.quit()\n\n #-----------------------------------------------------------------------\n #Part 3: Mars Weather\n url_twit = 'https://twitter.com/marswxreport?lang=en'\n response_tw = requests.get(url_twit)\n soup_tw = BeautifulSoup(response_tw.text, 'html.parser')\n last_twit = (soup_tw.find('div', class_=\"js-tweet-text-container\")).find('p')\n split = last_twit.get_text().split(last_twit.find('a').text.strip()) \n mars[\"mars_weather\"] = split[0].replace('\\n', ', ')\n print(\"---------------------------------------------------------------------------------------\")\n print(\"Part 3: Mars Weather Complete! Next: Part 4: Mars Facts\")\n \n #-----------------------------------------------------------------------\n #Part 4: Mars Facts\n url_facts = 'https://space-facts.com/mars/'\n tables = pd.read_html(url_facts)\n df = tables[0]\n df.columns = ['Description', 'Value']\n df.set_index('Description', inplace=True)\n html_table = df.to_html()\n html_table.replace('\\n', '')\n mars[\"mars_facts\"]= html_table\n print(\"---------------------------------------------------------------------------------------\")\n print(\"Part 4: Mars Facts Complete! Next: Part 2: JPL Mars Space Images - Featured Image\")\n\n #-----------------------------------------------------------------------\n #Part 2: JPL Mars Space Images - Featured Image\n\n browser = init_browser()\n url_pic = 'https://www.jpl.nasa.gov/spaceimages/?search=&category=Mars'\n browser.visit(url_pic)\n\n browser.links.find_by_partial_text('FULL IMAGE')\n browser.click_link_by_partial_text('FULL IMAGE')\n time.sleep(3)\n\n html = browser.html\n soup_pic = BeautifulSoup(html, 'html.parser')\n\n featured_image_url = \"http://www.jpl.nasa.gov\" + soup_pic.find('img', class_='fancybox-image')['src']\n mars[\"featured_image_url\"] = featured_image_url\n print(\"---------------------------------------------------------------------------------------\")\n print(\"Part 2: JPL Mars Space Images - Featured Image Complete! Next: Part 5: Mars Hemispheres\")\n\n #-----------------------------------------------------------------------\n #Part 5 Mars Hemispheres\n url_hemisphere = 'https://astrogeology.usgs.gov/search/results?q=hemisphere+enhanced&k1=target&v1=Mars'\n browser.visit(url_hemisphere)\n\n html = browser.html\n soup = BeautifulSoup(html, 'html.parser')\n\n print(\"---------------------------------------------------------------------------------------\")\n print(\"Compiling Hemisphere List\")\n\n results_hemisphere = soup.find('div', class_='full-content')\n titles_hemisphere = results_hemisphere.find_all('h3')\n\n title_list = []\n for title in titles_hemisphere:\n clean_title = str.strip(title.text)\n title_list.append(clean_title)\n \n print(\"Hemisphere List Complete! Next: Scraping Hemisphere Pic Url\")\n \n hemisphere_image_urls = []\n for x in range(len(title_list)):\n \n try:\n url_hemisphere = 'https://astrogeology.usgs.gov/search/results?q=hemisphere+enhanced&k1=target&v1=Mars'\n browser.visit(url_hemisphere)\n time.sleep(1)\n \n browser.links.find_by_partial_text(title_list[x])\n browser.click_link_by_partial_text(title_list[x])\n \n html = browser.html\n soup_pic = BeautifulSoup(html, 'html.parser')\n \n eachpic_hemisphere = soup_pic.find('div', class_='downloads')\n eachpic_link = eachpic_hemisphere.find('a')['href']\n \n pics_dict = {}\n pics_dict[\"title\"] = title_list[x]\n pics_dict[\"img_url\"] = eachpic_link\n hemisphere_image_urls.append(pics_dict)\n \n time.sleep(2)\n \n except ElementDoesNotExist:\n print(\"Scraping Complete\")\n\n mars[\"hemisphere_image_urls\"] = hemisphere_image_urls \n print(\"Part 5: Mars Hemispheres Complete! Next: Compiling into one dictionary\")\n browser.quit()\n\n #-----------------------------------------------------------------------\n print(\"---------------------------------------------------------------------------------------\")\n print(\"Complete Mars Data\")\n print(mars)\n return mars\n", "sub_path": "HW12 WebScraping-Challenge/Flask - Scrape Render/scrape_mars.py", "file_name": "scrape_mars.py", "file_ext": "py", "file_size_in_byte": 5518, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "splinter.Browser", "line_number": 12, "usage_type": "call"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 21, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 21, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 23, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 26, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 37, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 38, "usage_type": "call"}, {"api_name": "pandas.read_html", "line_number": 48, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 67, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 70, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 83, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 104, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 110, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 120, "usage_type": "call"}, {"api_name": "splinter.exceptions.ElementDoesNotExist", "line_number": 122, "usage_type": "name"}]} +{"seq_id": "618561631", "text": "import json\n\nfrom collections import defaultdict, OrderedDict\n\nfrom bricklink import read_json\nfrom operator import itemgetter\n\n\n# Una query ad-hoc para ver como de costoso es generar este codigo vs tenerlo en mysql.\ndef count_set_pieces():\n filtered_sets = filter_sets(read_json('data/BrickLink/Sets.json'))\n\n inventoried_parts = defaultdict(lambda: 0)\n unknown_qty_count = 0\n non_parts_count = 0\n missing_inventory_count = 0\n\n for the_set in filtered_sets:\n set_id = the_set['ITEMID']\n try:\n with open('data/BrickLink/Inventories/{}.json'.format(set_id)) as set_inventory_json_file:\n set_inventory_json = json.load(set_inventory_json_file)\n\n for part in set_inventory_json:\n if not part['QTY'].isdigit():\n unknown_qty_count += 1\n elif part['ITEMTYPE'] != 'P':\n non_parts_count += 1\n else:\n qty = int(part['QTY'])\n part_id = part['ITEMID']\n inventoried_parts[part_id] += qty\n except FileNotFoundError as exc:\n missing_inventory_count += 1\n\n print(\"Number of part types: {}\".format(len(inventoried_parts)))\n print(\"Number of non-parts (mini-figures...): {}, Missing inventories: {}, Unknown quantities for {} parts\"\n .format(non_parts_count, missing_inventory_count, unknown_qty_count))\n\n inventoried_sorted_parts = OrderedDict(sorted(inventoried_parts.items(), key=itemgetter(1), reverse=True))\n parts = read_json('data/BrickLink/Parts.json')\n\n for key, value in inventoried_sorted_parts.items():\n\n if value <= 100:\n break\n\n part = next(part for part in parts if part['ITEMID'] == key)\n from html import unescape\n print(\"Part {}, Qty: {}, Name: {}\".format(key, value, unescape(part['ITEMNAME'])))\n\n\ndef filter_sets(the_sets):\n modern_sets = [the_set for the_set in the_sets\n if the_set['ITEMYEAR'].isdigit() and int(the_set['ITEMYEAR']) >= 1995]\n\n print(\"Number of modern sets found: {}\".format(len(modern_sets)))\n\n categories = read_json('data/BrickLink/Categories.json')\n\n def get_category_name(category_id):\n return next(category['CATEGORYNAME'] for category in categories if category['CATEGORY'] == category_id)\n\n return [the_set for the_set in modern_sets\n if \"duplo\" not in get_category_name(the_set['CATEGORY']).lower()]\n", "sub_path": "query_experiment.py", "file_name": "query_experiment.py", "file_ext": "py", "file_size_in_byte": 2506, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "bricklink.read_json", "line_number": 11, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 13, "usage_type": "call"}, {"api_name": "json.load", "line_number": 22, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 40, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 40, "usage_type": "call"}, {"api_name": "bricklink.read_json", "line_number": 41, "usage_type": "call"}, {"api_name": "html.unescape", "line_number": 50, "usage_type": "call"}, {"api_name": "bricklink.read_json", "line_number": 59, "usage_type": "call"}]} +{"seq_id": "542764533", "text": "# -*- coding: utf-8 -*-\nimport scrapy\nfrom scrapy.shell import inspect_response\nimport json\nfrom urllib.parse import urlencode\nfrom urllib.parse import urljoin\nimport csv\nimport io\nimport logging\n\nclass CodeSpider(scrapy.Spider):\n name = 'code'\n \n allowed_domains = ['mercadopublico.cl']\n # Enter Login and Password\n rut = '22.087.581-4'\n passw= 'Licita.1234'\n \n \n \n \n \n def __init__(self, gs_id='', **kwargs):\n \n super().__init__(**kwargs)\n #Check if google id is provided, else use predefined\n if not gs_id:\n gs_id = '1AVcpYlYTBqepqycOUPp01Aub9NcT2xGfBbdep3x4-W4'\n \n logging.info(gs_id) \n # logging.info(gs_id)\n # Base url for downloading data from google sheets as csv\n self.input_url = 'https://docs.google.com/spreadsheets/d/{}/export?format=csv'.format(gs_id)\n # Decalre variables for later use\n self.detail_keys = []\n # Active product ( being edited at current time)\n self.active = []\n # List of all products to edit\n self.details = []\n self.headers = {\n 'accept': \"application/json, text/javascript, */*; q=0.01\",\n 'origin': \"https://www.mercadopublico.cl\",\n 'x-requested-with': \"XMLHttpRequest\",\n 'user-agent': \"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/72.0.3626.121 Safari/537.36\",\n 'dnt': \"1\",\n 'content-type': \"application/x-www-form-urlencoded\",\n 'cache-control': \"no-cache\", \n }\n \n\n \n def start_requests(self):\n #Download Google sheets with input data\n url = self.input_url\n yield scrapy.Request(url)\n \n \n \n def parse(self, response):\n #Read Google sheet and create details list from its data.\n sio = io.StringIO( response.text, newline=None) #.encode('ascii','ignore').decode('ascii')\n reader = csv.reader(sio, dialect=csv.excel)\n \n \n for row in reader:\n self.details.append(row)\n self.detail_keys = self.details[0] \n self.details = self.details[1:] \n self.active = self.details.pop()\n \n # login here. Then select organization that is supposed to be used\n url = 'https://www.mercadopublico.cl/Home/Autenticacion/NuevoLogin'\n data = {'Rut': self.rut,\n 'contraseña': self.passw,\n 'tipoUsuario': 'nacional',\n 'idPais': '0'}\n body = urlencode(data) \n yield scrapy.Request(url, method = 'POST', body = body, headers = self.headers, callback = self.parsed)\n \n \n \n def parsed(self, response):\n # inspect_response(response,self)\n data = json.loads(response.text)\n id = data['sessionID']\n \n # Get organization\n url ='https://www.mercadopublico.cl/Home/Autenticacion/ObtenerOrganizaciones'\n \n data = {'rut': self.rut,\n 'pass': self.passw,\n 'session': id,\n 'tipo': 'login',}\n \n yield scrapy.Request(url, method = 'POST', body = urlencode(data), headers = self.headers, callback = self.select_entity)\n \n def select_entity(self, response):\n # inspect_response(response, self)\n # Getting entity value by entity name\n headers = {\n 'accept': \"*/*\",\n 'origin': \"https://www.mercadopublico.cl\",\n 'x-requested-with': \"XMLHttpRequest\",\n 'user-agent': \"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/73.0.3683.86 Safari/537.36\",\n 'dnt': \"1\",\n 'content-type': \"application/json; charset=UTF-8\",\n 'cache-control': \"no-cache\",\n \n }\n \n entity = self.active[0]\n entity_value = response.xpath(\"//span[text()[ contains(.,'{}')]]/input/@value\".format(entity)).extract_first()\n url = 'https://www.mercadopublico.cl/Home/Autenticacion/LoginPorOrganizacion'\n id = str(response.xpath(\"//input[@id ='hdSession']/@value\").extract_first())\n body = json.dumps({\"CodigoEmpresa\":entity_value,\"sessionId\": id})\n yield scrapy.Request(url, method=\"POST\", body = body, callback = self.middle_menu, headers = headers)\n \n \n \n \n def middle_menu(self, response): \n \n # should click on middle menu\n url = 'http://www.mercadopublico.cl/CMII/Communication/comm.aspx?mode=seller'\n yield scrapy.Request(url, callback = self.convenio_marco)\n \n def convenio_marco(self, response):\n # inspect_response(response, self)\n # Click Administrator on specified agreement\n # Check if agreement description is available\n try:\n agreement_desc = self.active[7]\n url = response.xpath(\"//div[contains(.//h5,'{}')]\".format(agreement_desc))[0].xpath(\".//button/@onclick\").extract_first()\n # If not agreement description use agreement \n except:\n agreement = self.active[1]\n url = response.xpath(\"//div[contains(.//a,'{}')]\".format(agreement))[0].xpath(\".//button/@onclick\").extract_first()\n \n url = url.split(\"('\")[1][:-2]\n url = urljoin(response.url, url)\n yield scrapy.Request(url, callback = self.administrar)\n \n \n def administrar(self, response):\n # inspect_response(response, self)\n #Click on APlicat Uno Oferta button for OfertaEspecial to get to search bar\n url = response.xpath(\"//a[contains(@href, 'frmOfertaEspecial')]/@href\").extract_first()\n url = urljoin(response.url, url)\n \n i = []\n yield scrapy.Request(url, callback = self.search_bar, dont_filter = True)\n \n def search_bar(self, response):\n # inspect_response(response, self)\n \n i = self.active\n body = {'__EVENTARGUMENT': '',\n '__EVENTTARGET': '',\n '__SCROLLPOSITIONX': '0',\n '__SCROLLPOSITIONY': '0',\n \n 'ddlOfertaEspecial': '0',\n 'imgBotonBusca.x': '47',\n 'imgBotonBusca.y': '12',\n }\n # iterating through the list of items to be changed.\n \n body[ 'txtTextoBuscado'] = i[3]\n \n \n yield scrapy.FormRequest.from_response(response, formdata=body, callback=self.item_edited , dont_filter = True )\n \n def item_edited(self, response): \n \n i = self.active\n # inspect_response(response, self) \n \n body = {'CTRL_ListadoOfertasEspeciales1:dgrOfertasEspeciales:_ctl3:imgbtn_Editar.y' : '11',\n 'CTRL_ListadoOfertasEspeciales1:dgrOfertasEspeciales:_ctl3:imgbtn_Editar.x' : '12',\n '__EVENTARGUMENT': '',\n '__EVENTTARGET': '',\n '__SCROLLPOSITIONX': '0',\n '__SCROLLPOSITIONY': '0',\n 'ddlOfertaEspecial': '0',\n }\n body['txtTextoBuscado'] = i[3] \n yield scrapy.FormRequest.from_response(response, formdata=body, callback=self.item_found, dont_filter = True)\n \n def item_found(self, response):\n # inspect_response(response, self) \n cant_modify = response.xpath(\"//input[contains(@title,'no puede modificarla')]\")\n \n i = self.active\n description =i[3]\n # price = str(float(i[1])+1)\n price = i[4]\n startDate = i[5]\n endDate = i[6]\n \n body = {\n 'CTRL_ListadoOfertasEspeciales1:dgrOfertasEspeciales:_ctl3:imgbtn_Guardar.x': '1',\n 'CTRL_ListadoOfertasEspeciales1:dgrOfertasEspeciales:_ctl3:imgbtn_Guardar.y': '1',\n 'CTRL_ListadoOfertasEspeciales1:dgrOfertasEspeciales:_ctl3:txtPrecio_OfEsp': '',\n '__EVENTARGUMENT': '',\n '__EVENTTARGET': '',\n '__SCROLLPOSITIONX': '0',\n '__SCROLLPOSITIONY': '0',\n 'ddlOfertaEspecial': '0',\n }\n body['CTRL_ListadoOfertasEspeciales1:dgrOfertasEspeciales:_ctl3:txtPrecio_OfEsp'] = price\n body['CTRL_ListadoOfertasEspeciales1:dgrOfertasEspeciales:_ctl3:CalendarioPopUp1:txtNombre'] = startDate\n body[ 'CTRL_ListadoOfertasEspeciales1:dgrOfertasEspeciales:_ctl3:CalendarioPopUp1:valorID'] = startDate\n body['CTRL_ListadoOfertasEspeciales1:dgrOfertasEspeciales:_ctl3:CalendarioPopUp2:txtNombre'] = endDate\n body['CTRL_ListadoOfertasEspeciales1:dgrOfertasEspeciales:_ctl3:CalendarioPopUp2:valorID'] = endDate\n body['txtTextoBuscado'] =description \n \n yield scrapy.FormRequest.from_response(response, formdata=body, callback=self.edit_oferta, dont_filter = True)\n pass\n if cant_modify:\n print('Item cant not be modified')\n \n \n \n \n def edit_oferta(self, response):\n # inspect_response(response, self)\n \n alert = response.xpath(\"//script[contains(text(), 'alert')]/text()\").extract()\n active_old = self.active\n value_to_set = active_old[4]\n key = self.detail_keys\n # logging.info(key)\n logging.info(active_old)\n new_ofert = response.xpath(\"//td[@class ='ofertasInterior']/span[contains(@id,'Precio')]/text()\").extract_first()\n if not alert:\n # try:\n inspect_response(response, self)\n logging.info(active_old[4])\n logging.info(type(active_old[4]))\n logging.info( new_ofert)\n logging.info( type(new_ofert))\n if new_ofert:\n # remove spaces dots and comas from ofert values to compare them\n value_to_set = value_to_set.replace('.','').replace(',','').replace(' ','')\n new_ofert = new_ofert.replace('.','').replace(',','').replace(' ','')\n if value_to_set in new_ofert:\n alert = 'OK'\n elif value_to_set == new_ofert:\n alert = 'OK'\n elif value_to_set in new_ofert:\n alert = 'OK'\n else: alert = 'ERROR. New Ofert not equal to Expected value'\n else: alert = 'ERROR. No New Offert'\n \n yield { \n key[0] : active_old[0],\n key[1] : active_old[1],\n key[2] : active_old[2],\n key[3] : active_old[3],\n 'new_desc' : response.xpath(\"//a[contains(@href, 'NombreProducto')]/text()\").extract_first(), \n key[4] : active_old[4],\n 'new_ofert' : new_ofert,\n key[5] : active_old[5],\n 'new_date_start' : response.xpath(\"//td[@class ='ofertasInterior']/span[contains(@id,'Inicio')]/text()\").extract_first(),\n key[6] : active_old[6],\n 'new_date_end' : response.xpath(\"//td[@class ='ofertasInterior']/span[contains(@id,'Termino')]/text()\").extract_first(),\n 'alert' : alert,\n }\n \n # Check if any products left to edit\n if self.details:\n \n # Assign new product to active\n self.active = self.details.pop()\n active = self.active\n # Check if entity and agreement are the same for previous and next item.\n if active_old[0] == active[0] and active_old[1] == active[1]:\n \n body = {'__EVENTARGUMENT': '',\n '__EVENTTARGET': '',\n '__SCROLLPOSITIONX': '0',\n '__SCROLLPOSITIONY': '0',\n \n 'ddlOfertaEspecial': '0',\n 'imgBotonBusca.x': '47',\n 'imgBotonBusca.y': '12',\n }\n # iterating through the list of items to be changed.\n \n body[ 'txtTextoBuscado'] = active[3]\n \n \n yield scrapy.FormRequest.from_response(response, formdata=body, callback=self.item_edited , dont_filter = True )\n # Check if entity is different but agreement is the same for previous and next item.\n elif active_old[0] != active[0]:\n \n cookie = str(response.request.headers['Cookie']).split('; ')\n id = [i.split('=')[-1] for i in cookie if 'ASP.NET_SessionId=' in i][0]\n \n # Get organization\n url ='https://www.mercadopublico.cl/Home/Autenticacion/ObtenerOrganizaciones'\n \n data = {'rut': self.rut,\n 'pass': self.passw,\n 'session': id,\n 'tipo': 'login',}\n \n yield scrapy.Request(url, method = 'POST', body = urlencode(data), headers = self.headers, callback = self.select_entity)\n else: \n url = 'http://www.mercadopublico.cl/CMII/Communication/comm.aspx?mode=seller'\n yield scrapy.Request(url, callback = self.convenio_marco)", "sub_path": "mercado/spiders/code.py", "file_name": "code.py", "file_ext": "py", "file_size_in_byte": 13855, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "scrapy.Spider", "line_number": 11, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 30, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 55, "usage_type": "call"}, {"api_name": "io.StringIO", "line_number": 61, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 62, "usage_type": "call"}, {"api_name": "csv.excel", "line_number": 62, "usage_type": "attribute"}, {"api_name": "urllib.parse.urlencode", "line_number": 77, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 78, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 84, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 95, "usage_type": "call"}, {"api_name": "urllib.parse.urlencode", "line_number": 95, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 115, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 116, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 125, "usage_type": "call"}, {"api_name": "urllib.parse.urljoin", "line_number": 140, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 141, "usage_type": "call"}, {"api_name": "urllib.parse.urljoin", "line_number": 148, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 151, "usage_type": "call"}, {"api_name": "scrapy.FormRequest.from_response", "line_number": 171, "usage_type": "call"}, {"api_name": "scrapy.FormRequest", "line_number": 171, "usage_type": "attribute"}, {"api_name": "scrapy.FormRequest.from_response", "line_number": 187, "usage_type": "call"}, {"api_name": "scrapy.FormRequest", "line_number": 187, "usage_type": "attribute"}, {"api_name": "scrapy.FormRequest.from_response", "line_number": 217, "usage_type": "call"}, {"api_name": "scrapy.FormRequest", "line_number": 217, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 233, "usage_type": "call"}, {"api_name": "scrapy.shell.inspect_response", "line_number": 237, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 238, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 239, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 240, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 241, "usage_type": "call"}, {"api_name": "scrapy.FormRequest.from_response", "line_number": 293, "usage_type": "call"}, {"api_name": "scrapy.FormRequest", "line_number": 293, "usage_type": "attribute"}, {"api_name": "scrapy.Request", "line_number": 308, "usage_type": "call"}, {"api_name": "urllib.parse.urlencode", "line_number": 308, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 311, "usage_type": "call"}]} +{"seq_id": "362651292", "text": "\"\"\" Module for handling the serialization of Cast- and CastCollection-like\nobjects to persistent files. \"\"\"\n\nimport six\nimport json\nimport copy\nimport datetime\nimport dateutil.parser\nimport numpy\nimport pandas\nfrom karta import Point, geojson\nfrom . import units\n\ndef castasdict(cast):\n scalars = [key for key in cast.properties]\n vectors = list(cast.data.keys())\n dscalar, dvector = {}, {}\n for key in scalars:\n if isinstance(cast.properties[key], datetime.datetime):\n dscalar[key] = cast.properties[key].isoformat(sep=\" \")\n else:\n dscalar[key] = cast.properties[key]\n for key in vectors:\n if isinstance(cast[key], numpy.ndarray):\n dvector[key] = cast[key].tolist()\n elif isinstance(cast[key], pandas.Series):\n dvector[key] = cast[key].values.tolist()\n else:\n dvector[key] = list(cast[key])\n d = dict(type=cast._type, scalars=dscalar, vectors=dvector,\n coords=cast.coords, zunits=str(cast.zunits), zname=str(cast.zname))\n return d\n\ndef findunit(unitname):\n for name in units.__dict__:\n if str(units.__dict__[name]) == unitname:\n return units.__dict__[name]\n raise NameError(\"'{0}' not recognized as a unit\".format(unitname))\n\ndef dictascast(d, obj):\n \"\"\" Read a file-like stream and construct an object with a Cast-like\n interface. \"\"\"\n d_ = d.copy()\n _ = d_.pop(\"type\")\n coords = d_[\"scalars\"].pop(\"coordinates\")\n zunits = findunit(d_.pop(\"zunits\", \"meter\"))\n zname = d_.pop(\"zname\", \"z\")\n z = d_[\"vectors\"].pop(zname)\n prop = d[\"scalars\"]\n for (key, value) in prop.items():\n if \"date\" in key or \"time\" in key and isinstance(prop[key], str):\n try:\n prop[key] = dateutil.parser.parse(value)\n except (TypeError, ValueError):\n pass\n prop.update(d_[\"vectors\"])\n cast = obj(z, coords=coords, zunits=zunits, zname=zname, **prop)\n return cast\n\ndef dictascastcollection(d, castobj):\n \"\"\" Read a file-like stream and return a list of Cast-like objects.\n \"\"\"\n casts = [dictascast(cast, castobj) for cast in d[\"casts\"]]\n return casts\n\ndef writecast(f, cast, binary=True):\n \"\"\" Write Cast data to a file-like stream. \"\"\"\n d = castasdict(cast)\n if binary:\n s = json.dumps(d, indent=2)\n # f.write(bytes(s, \"utf-8\"))\n f.write(six.b(s))\n else:\n json.dump(d, f, indent=2)\n return\n\ndef writecastcollection(f, cc, binary=True):\n \"\"\" Write CastCollection to a file-like stream. \"\"\"\n casts = [castasdict(cast) for cast in cc]\n d = dict(type=\"castcollection\", casts=casts)\n if binary:\n s = json.dumps(d, indent=2)\n # f.write(bytes(s, \"utf-8\"))\n f.write(six.b(s))\n else:\n json.dump(d, f, indent=2)\n return\n\ndef castcollection_as_geojson(cc):\n castpoints = (Point(c.coords, properties={\"id\":i})\n for i, c in enumerate(cc))\n geojsonstring = geojson.printFeatureCollection(castpoints)\n return geojsonstring\n\n", "sub_path": "narwhal/fileio.py", "file_name": "fileio.py", "file_ext": "py", "file_size_in_byte": 3055, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "datetime.datetime", "line_number": 19, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 26, "usage_type": "attribute"}, {"api_name": "dateutil.parser.parser.parse", "line_number": 53, "usage_type": "call"}, {"api_name": "dateutil.parser.parser", "line_number": 53, "usage_type": "attribute"}, {"api_name": "dateutil.parser", "line_number": 53, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 70, "usage_type": "call"}, {"api_name": "six.b", "line_number": 72, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 74, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 82, "usage_type": "call"}, {"api_name": "six.b", "line_number": 84, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 86, "usage_type": "call"}, {"api_name": "karta.Point", "line_number": 90, "usage_type": "call"}, {"api_name": "karta.geojson.printFeatureCollection", "line_number": 92, "usage_type": "call"}, {"api_name": "karta.geojson", "line_number": 92, "usage_type": "name"}]} +{"seq_id": "172381681", "text": "from django.shortcuts import render\r\n\r\nfrom rest_framework import status\r\nfrom rest_framework.decorators import api_view, permission_classes\r\nfrom rest_framework.permissions import AllowAny\r\nfrom rest_framework.response import Response\r\n#from django.views.decorators.csrf import csrf_exempt\r\n\r\nfrom .models import AnnoteModel\r\n\r\nimport json\r\n\r\n#@csrf_exempt\r\n@api_view(['GET', 'PATCH'])\r\n@permission_classes((AllowAny, ))\r\ndef image_metadata(request):\r\n \"\"\"\r\n Return image metadata.\r\n \"\"\"\r\n if request.method == 'GET':\r\n try:\r\n image_metadata = AnnoteModel.objects.get(pk=1)\r\n return Response(image_metadata.content)\r\n except AnnoteModel.DoesNotExist:\r\n return Response({})\r\n\r\n elif request.method == 'PATCH':\r\n print(type(request.data['content']))\r\n try:\r\n image_metadata = AnnoteModel.objects.get(pk=1)\r\n except AnnoteModel.DoesNotExist:\r\n image_metadata = AnnoteModel()\r\n image_metadata.content = request.data['content']\r\n image_metadata.save()\r\n return Response(image_metadata.content, status=status.HTTP_201_CREATED)\r\n\r\n#@csrf_exempt\r\n@api_view(['POST'])\r\n@permission_classes((AllowAny, ))\r\ndef handle_annotation(request):\r\n #print ('------------ fsdfa: {}'.format(request.body))\r\n if request.method == 'POST':\r\n #print(request.FILES)\r\n dic = json.loads(request.data['annotate'])\r\n print(dic)\r\n for i in dic.keys():\r\n filename = i\r\n count=len(dic[filename]['regions'])\r\n for i in range(count):\r\n i = str(i)\r\n attempt = AnnoteModel()\r\n #attempt.shape = dic[filename]['regions'][i]['shape_attributes']['name']\r\n attempt.width = dic[filename]['regions'][i]['shape_attributes']['width']\r\n attempt.height = dic[filename]['regions'][i]['shape_attributes']['height']\r\n attempt.xcoordinate = dic[filename]['regions'][i]['shape_attributes']['x']\r\n attempt.ycoordinate = dic[filename]['regions'][i]['shape_attributes']['y']\r\n attempt.filename = filename\r\n attempt.image = request.FILES['image']\r\n attempt.attribute=request.POST.get('region')\r\n attempt.garbage = request.POST.get('garbage')\r\n #attempt.attribute = dic[filename]['regions'][i]['region_attributes']['name']\r\n attempt.save()\r\n\r\n #print ('---- JSON DATA ----\\nDATA: {}',dic)\r\n #print(json_data)\r\n\r\n return Response({})\r\n\r\ndef index(request):\r\n return render(request, 'index.html')\r\n\r\ndef SaveAnnotate(request):\r\n return HttpResponse(\":)\")\r\n\r\n# def parseJSON(dic,request):\r\n# for i in dic.keys():\r\n# filename = i\r\n# count=len(dic[filename]['regions'])\r\n# for i in range(count):\r\n# i = str(i)\r\n# attempt = AnnoteModel()\r\n# attempt.shape = dic[filename]['regions'][i]['shape_attributes']['name']\r\n# attempt.width = dic[filename]['regions'][i]['shape_attributes']['width']\r\n# attempt.height = dic[filename]['regions'][i]['shape_attributes']['height']\r\n# attempt.xcoordinate = dic[filename]['regions'][i]['shape_attributes']['x']\r\n# attempt.ycoordinate = dic[filename]['regions'][i]['shape_attributes']['y']\r\n# attempt.filename = filename\r\n# attempt.image = request.data['imagefile']\r\n# attempt.save()\r\n\r\n", "sub_path": "annote/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3407, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "models.AnnoteModel.objects.get", "line_number": 22, "usage_type": "call"}, {"api_name": "models.AnnoteModel.objects", "line_number": 22, "usage_type": "attribute"}, {"api_name": "models.AnnoteModel", "line_number": 22, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 23, "usage_type": "call"}, {"api_name": "models.AnnoteModel.DoesNotExist", "line_number": 24, "usage_type": "attribute"}, {"api_name": "models.AnnoteModel", "line_number": 24, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 25, "usage_type": "call"}, {"api_name": "models.AnnoteModel.objects.get", "line_number": 30, "usage_type": "call"}, {"api_name": "models.AnnoteModel.objects", "line_number": 30, "usage_type": "attribute"}, {"api_name": "models.AnnoteModel", "line_number": 30, "usage_type": "name"}, {"api_name": "models.AnnoteModel.DoesNotExist", "line_number": 31, "usage_type": "attribute"}, {"api_name": "models.AnnoteModel", "line_number": 31, "usage_type": "name"}, {"api_name": "models.AnnoteModel", "line_number": 32, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 35, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 35, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 35, "usage_type": "name"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 14, "usage_type": "call"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 15, "usage_type": "call"}, {"api_name": "rest_framework.permissions.AllowAny", "line_number": 15, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 44, "usage_type": "call"}, {"api_name": "models.AnnoteModel", "line_number": 51, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 67, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 38, "usage_type": "call"}, {"api_name": "rest_framework.decorators.permission_classes", "line_number": 39, "usage_type": "call"}, {"api_name": "rest_framework.permissions.AllowAny", "line_number": 39, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 70, "usage_type": "call"}]} +{"seq_id": "559718345", "text": "'''\nThis file does the GUI stuff, and syncxml.py does the work.\nI don't have a full rig set up for 'compiling' PyQt and Anki, so I only debug some non-GUI behavior,\nby using local_launch instead. For debugging within Anki itself, try Ctrl+Shift+; as shown here:\nhttp://ankisrs.net/docs/addons.html#debugging \n\n@author: Jonathan Coombs, copyright sil.org 2012-2014\n'''\n\n# Force Python 3 syntax\nfrom __future__ import print_function, absolute_import, division , unicode_literals\nimport os\nimport shutil\n\nfrom . import logg as L\nfrom . import syncxml as SX\nfrom . import anki_util as A\nfrom . import xml_util as X\nfrom . import flex_util as F\n\n# 'Constants'\nCFGMSG = \"Sync from XML is not yet configured. You can do so from the Tools menu, or by manually creating a custom config file.\\n\"\nIMPMSG = \"Please do NOT use Import for data you wish to sync. Instead, sync from the Tools menu.\\n\"\nTEST_PATH = \"C:\\\\Users\\\\user57\\\\Documents\\\\dict4anki 2\\\\cgg.lift\" # what to use for testing if nothing else is found\nTEST_FWDATA = \"\" # making this non-empty makes the non-UI test run answer Yes and pick this as the FLEx project\n\nif A.IN_ANKI:\n from aqt import mw\n from aqt.qt import *\n from aqt.utils import showInfo, askUserDialog, getFile\n from anki.notes import Note\n # from anki.utils import isMac #obsolete now, I think\n # from PyQt4.QtGui import QMessageBox #done for us already?\n QUESTION = QMessageBox.Question\n CRITICAL = QMessageBox.Critical\nelse:\n # crash-prevention dummies\n QUESTION = 0\n CRITICAL = 0\n\ndef msgbox(m):\n text = '{}: \\n\\n{}'.format(SX.ADDON_NAME2, m)\n L.debug(\"\\n_____ Messagebox _____\\n\" + text + \"\\n----------\")\n if A.IN_ANKI:\n no_hourglass()\n showInfo(text)\n\ndef dialogbox(text, buttons, icon=4, log=True):\n if log:\n L.debug(\"\\n_____ Dialog _____\\n\" + text + \"\\n----------\")\n if A.IN_ANKI: \n no_hourglass()\n b = askUserDialog(text, buttons)\n if icon: b.setIcon(icon)\n x = b.run()\n else:\n x = buttons[0] # first button is default when testing\n return x\n\ndef hourglass():\n if A.IN_ANKI:\n mw.app.setOverrideCursor(QCursor(Qt.WaitCursor)) # display an hourglass cursor\ndef no_hourglass():\n if A.IN_ANKI:\n mw.app.restoreOverrideCursor()\n \n\ndef try_sync(cfpath=SX.CONFIG_FILE):\n L.debug('Preparing to launch sync...')\n L.debug(\"Using config file: {}\".format(cfpath))\n mm = ''\n if A.IN_ANKI:\n L.debug('===== Launching sync from within Anki... =====')\n hourglass()\n mm = SX.sync(cfpath)\n else:\n L.debug('===== Launching sync locally... =====')\n (all_src_records, num_src) = SX.sync(cfpath)\n e = L.error_count()\n w = L.warn_count()\n mm = 'errors: {}; warnings: {}'.format(e, w)\n mm += '\\nDone reading from source file ({} records). That is all we can do for now (cannot access Anki from here), so quitting here.'.format(num_src)\n\n msgbox(mm)\n # TODO: Prompt to auto-delete now-superfluous records. Check...\n # mmm = \"\"\"The target deck in Anki ({}) has existing data and/or media files in it. Delete or leave these?\"\"\"\n # x = dialogbox(mmm, ['Yes', 'No', 'Cancel'], QUESTION)\n # if (x == 'Cancel'): return\n\n\ndef on_sync_clicked():\n if A.IN_ANKI: A.anki_user_profile = mw.pm.profileFolder() \n L.init(SX.LOG_FP, L.VERBOSITY)\n try:\n cfpath = SX.get_config_path() \n if not os.path.exists(cfpath):\n # must create config before syncing\n L.w(\"Can't sync yet. Launching wizard to create the necessary config file: {}\".format(cfpath))\n if wizard():\n try_sync()\n else:\n # A standard config file already exists. Sync.\n try_sync()\n finally:\n no_hourglass()\n launch_paths_maybe()\n # L.close_file() # Otherwise, you can't read the log until you've close Anki's error window.\n \n \ndef on_reconfigure_clicked():\n A.anki_user_profile = mw.pm.profileFolder() \n L.init(SX.LOG_FP, L.VERBOSITY)\n try:\n if reconfigure():\n try_sync()\n else:\n L.w(\"Reconfiguration cancelled.\")\n # msgbox(\"Reconfiguration cancelled.\", close_log=True);\n finally:\n no_hourglass()\n launch_paths_maybe()\n # L.close_file()\n\ndef reconfigure(target=\"\"):\n \"\"\"Tries to reconfigure. Returns True if a sync should follow immediately.\"\"\"\n L.debug('Preparing to reconfigure...')\n cfmainpath = SX.get_config_path()\n cfnewpath = SX.rename_config_to()\n if cfnewpath:\n msg = \"Your existing configuration will be renamed to {}; Continue?\".format(cfnewpath)\n x = dialogbox(msg, ['Ok', 'Cancel'], CRITICAL)\n if (x == 'Cancel'): \n return False\n os.rename(cfmainpath, cfnewpath)\n result = wizard(target)\n return result\n\ndef ensure_models(models):\n \"\"\"If we're in Anki and the given models (i.e. note types) don't all exist,\n create them by importing the sample APKG file, then deleting all\n records in the resulting deck. \n ASSUMPTION: the sample APKG contains all needed models.\n WARNING: this could blow away user data in one unlikely scenario: user creates a lift_dictionary deck\n manually, or imports the APKG, and then manually enters data.\"\"\"\n def all_models_ok(ms):\n for m in ms:\n if (not A.anki_model_exists(m)):\n L.w(\"Target data model '{}' does not yet exist. Will attempt to get it by importing the default APKG file.\".format(m))\n return False\n return True\n \n if A.IN_ANKI:\n if not all_models_ok(models):\n no_hourglass()\n try:\n # Import the APKG.\n fp = A.get_filepath(A.APKG_PATH)\n delete_failure = A.import_apkg_model(fp, True)\n L.w(\"Done importing APKG file.\")\n mw.col.models.flush()\n mw.reset(True)\n mw.reset()\n if delete_failure:\n L.warn(delete_failure) # not a big deal\n msgbox(delete_failure)\n else:\n L.w(\"Successfully deleted existing records.\")\n except Exception as e:\n # We were unable to automatically import the APKG\n L.error(e)\n L.warn(A.NO_MODEL_MSG)\n msgbox(A.NO_MODEL_MSG + \"\\nError message: {}\".format(e))\n return False\n return all_models_ok(models) # verify\n return True\n\n\ndef wizard(target=\"\"):\n \"\"\"Assumption: The main config file does not exist. Auto-configure. \n Returns True if successful and ready to sync.\n target: this parameter is for testing; it overrides the logic for finding a LIFT file.\"\"\"\n L.debug('Launching the auto-config wizard')\n\n cfmainpath = SX.get_config_path()\n cfdefpath = SX.get_config_path(SX.CONFIG_DEFAULT_FILE)\n\n if not os.path.exists(cfdefpath): # if no default config exists either...\n msg = \"Cannot find and copy the default config file:\\n {}\\nCannot continue.\".format(cfdefpath)\n x = dialogbox(msg, ['ok'], CRITICAL)\n return False\n \n src_dir = SX.get_home_dir_plus(os.path.join(SX.get_docs_dir_name(), SX.SRC_DIR_LIFT), False) # use True if creating dict4anki\n\n flex_dir = F.flex_dir()\n flex_msg = ''\n if flex_dir:\n flex_msg = \" For quickest setup, give it the same name as one of the projects here:\\n {}\\n\".format(flex_dir)\n\n \n msg = \"Would you like to bring in your own* LIFT data? If so, either...\\n\" \\\n \"A) FLEx users, export a LIFT file here (or to a direct subfolder of it):\\n\" \\\n \" {} \\n{}\" \\\n \"B) WeSay (or FLEx) users can just click LIFT and choose a LIFT file.\\n\\n\" \\\n \"A copy of the default configuration file will be auto-configured for you,\\n\" \\\n \" which may take a few seconds. After configuration, the LIFT file to be synced\\n\" \\\n \" must always be located in that same place.\\n\\n\" \\\n \"Or, click Sample to sync from the sample file instead.\\n\\n\" \\\n \"*Audio will only be auto-detected if your main writing systems are 2- or 3-letter codes.\"\n msg = msg.format(src_dir, flex_msg)\n # L.debug(\"Dialog: {}\\n\".format(msg))\n x = dialogbox(msg, ['LIFT', 'Sample', 'Cancel'], QUESTION)\n if (x == 'Cancel'): return False\n\n hourglass()\n \n # Make sure Anki has the default deck and models already there; else import the APKG file.\n if not ensure_models([X.MODEL1, X.MODEL2]):\n return False\n if (x == 'Sample'):\n # TODO idea: preprocess = (('vern', 'klw'),('other', 'id')) , hard-coding here.\n # After that, the default config file wouldn't need to be hard-coded to specific languages anymore.\n # Note that klw should cover klw-Zxxx-x-audio too, etc.\n try:\n msg = SX.sync(SX.CONFIG_DEFAULT_FILE) # , preprocess)\n msgbox(msg)\n # launch_paths_maybe()\n return False\n except:\n # launch_paths(suppressExceptions=True)\n raise\n finally: \n no_hourglass()\n\n hourglass()\n\n # prepare to copy default config to make a new config (via a temp file first)\n shutil.copy(cfdefpath, cfmainpath + '.tmp') # will overwrite silently if need be\n \n lift = '' # was: lift = SX.get_first_lift_file(src_dir) # check the dict4anki folder\n if not lift: # fall back to anything in a direct subfolder of Documents\\WeSay (Linux: ~/WeSay)\n tmp = SX.get_home_dir_plus(os.path.join(SX.get_docs_dir_name(), SX.SRC_DIR_WESAY))\n lift = SX.get_first_lift_file(tmp)\n if not lift: # fall back to anything in a direct subfolder of Documents (Windows: %USERPROFILE%\\My Documents; Linux: ~/)\n # src_dir = os.path.split(src_dir)[0] # remove \"/dict4anki/\"\n lift = SX.get_first_lift_file(SX.get_home_dir_plus(SX.get_docs_dir_name()))\n if lift:\n src_dir = os.path.split(lift)[0]\n if A.IN_ANKI:\n no_hourglass()\n # pop up a File Open dialog using ankilib's convenience method\n lift = getFile(mw, \"Open LIFT file\", None, filter=\"*.lift\", dir=src_dir, key=\"\") # \"*.lift\"\n L.debug(\"User chose this LIFT file: {}\".format(lift))\n elif (not lift) and os.path.exists(TEST_PATH): \n lift = TEST_PATH # hard-coded test\n if target:\n lift = target # for testing, a passed parameter overrides all of the above\n\n L.debug(\"Using this LIFT file: {}\".format(lift))\n if not lift: \n # Still no LIFT file. Fail.\n msg = \"No file chosen. Auto-configuration cancelled for now.\" \n x = dialogbox(msg, ['ok'], CRITICAL)\n return False\n \n m = \"LIFT file: \\n {}\\n\".format(lift)\n flex_audio, flex_image = None, None\n\n # Check for WeSay. E.g. if Catalan.lift has a Catalan.WeSayConfig next to it, assume it's a WeSay project\n # Would it be better to make sure it's in the official WeSay directory?\n p, f = os.path.split(lift)\n f = os.path.splitext(f)[0]\n is_wesay = os.path.exists(os.path.join (p, f + \".WeSayConfig\")) \n\n if (not is_wesay) and flex_dir:\n L.debug(\"Checking for projects in this flex_dir: {}\".format(flex_dir))\n tmp = F.flex_media(lift, flex_dir)\n L.debug(\"Found tmp: {}\".format(tmp))\n if tmp: \n flex_audio, flex_image = tmp\n\n if flex_audio:\n msg = \"{}Also found a FLEx project with the same name as your LIFT file and it probably has these media folders:\\n\" \\\n \" {}\\n {}\\n\" \\\n \"Shall we sync media files directly from there, so that before each \\n\" \\\n \"sync the only thing you'll have to export from FLEx will be the LIFT data?\\n\" \\\n \"(If No, the 'audio' and 'pictures' folders in the LIFT file's location will be used.)\".format(m, flex_audio, flex_image)\n answer = dialogbox(msg, ['Yes', 'No'], QUESTION)\n if not A.IN_ANKI:\n # answer = 'No' #Or, put the No button first, then delete this\n pass\n if answer != 'Yes':\n flex_audio, flex_image = None, None # dump them\n elif not is_wesay:\n msg = \"{}Could not find a FLEx project with the same name as your LIFT file.\\n\" \\\n \"Do you wish to select a FLEx project that does/will contain your media files? \\n\" \\\n \"WeSay users: choose No. FLEx users: choose Yes unless you want to export \\n\" \\\n \" the media files along with the LIFT before each sync.\".format(m)\n answer = dialogbox(msg, ['Yes', 'No', 'Cancel'], QUESTION)\n if (answer == 'Cancel'): return False\n fwdata = TEST_FWDATA\n if (answer == 'Yes') and (A.IN_ANKI):\n # pop up a File Open dialog using ankilib's convenience method\n fwdata = getFile(mw, \"Select FLEx project\", None, filter=\"*.fwdata\", key=\"\", dir=flex_dir)\n if fwdata:\n tmp = F.flex_media(fwdata)\n if tmp: flex_audio, flex_image = tmp\n \n # Note: working with a temp copy of config, so as to not create an official config until we're sure we've succeeded.\n try: \n xset = X.XmlSettings(cfmainpath + '.tmp', lift, flex_audio, flex_image)\n except:\n launch_paths(suppressExceptions=True)\n raise\n\n if os.path.getsize(lift) > 1000000:\n msg = m + \"Your file is large, so analyzing it may take a while. Please click Ok and then wait.\"\n answer = dialogbox(msg, ['Ok', 'Cancel'], QUESTION)\n if answer == 'Cancel' : return\n\n # status bar (no longer supported by Anki?)\n# from aqt.main import setStatus as set_status\n# mw.setStatus(\"Analyzing the LIFT file...\", timeout=3000) \n\n # Find and Replace WS's in the new config file\n hourglass()\n to_replace = xset.find_vern_nat()\n L.w(\"For LIFT file\\n ({}) we will now find/replace writing systems in our settings as follows: \\n{}\".format(lift, X.lang_table(to_replace)))\n\n # Using regex (on cfmainpath + '.tmp') to replace WSes in our new config file ...\n xset.save() \n tmp = xset.file_path\n X.replace_all(tmp, to_replace)\n xset = None # since it is now outdated \n # TODO: xset.dispose() # this would help with safety if implemented\n\n # do a dry run: use the new config file to load the LIFT file...\n try:\n xset = X.XmlSettings(tmp, lift, flex_audio, flex_image) # we need those last two parameters or we'll lose any FLEx path we had\n except:\n launch_paths(suppressExceptions=True)\n raise\n \n xdl = X.XmlDataLoader() # No try block, since presumably the user knows where the data file is.\n _recs, empties = xdl.load_src_file(xset.get_attr(), xset.entry, sync_media=False, dry_run=True)\n _recs, empties2 = xdl.load_src_file(xset.get_attr(), xset.example, sync_media=False, dry_run=True)\n # empties.append(empties2)\n\n # ... so we can disable any xpaths that don't match any data.\n if empties:\n L.w(\"The following entry fields yielded no data and will now be disabled so as to not generate warnings: {}\".format(empties))\n xset.entry.disable_fields(empties)\n if empties2:\n L.w(\"The following example fields yielded no data and will now be disabled so as to not generate warnings: {}\".format(empties))\n xset.example.disable_fields(empties2)\n\n # If no example sentences, we already disabled auto-disabled that section.\n if xset.example.get_attr()['enabled'] == 'true':\n # It's still enabled, which means it has data. \n msg = m + \"Found dictionary Examples containing data; when imported, each will show up on the main entry's flashcard.\\n\" \\\n \"Will you also need a separate flashcard for each Example?\"\n x = dialogbox(msg, ['No', 'Yes', 'Cancel'], QUESTION)\n if (x == \"Cancel\"):\n return False\n if (x != 'Yes'): \n xset.example.disable()\n\n hourglass()\n xset.save()\n\n # rename the default config file (remove the .tmp)\n shutil.move(cfmainpath + '.tmp', cfmainpath) # will overwrite silently if need be, but see our initial assumption\n L.debug(\"Configuration file saved.\")\n m2 = \"\\nReplaced writing systems in our new configuration as follows: \\n{}\".format(X.lang_table(to_replace))\n m3, m5 = '', ''\n if flex_audio:\n m3 = \"\\nConfigured to copy media files from these locations: \\n {}\\n {}\\n\".format(flex_audio, flex_image)\n m4 = \"\\nConfiguration file saved. Click Yes to sync now, or No if you wish to review/tweak the configuration first.\\n\"\n if L.error_count() or L.warn_count():\n m5 = \"\\nThere were errors or warnings during auto-config. Please review the log.\"\n msg = m + m2 + m3 + m4 + m5\n # TODO: \" or want to run a Sync Preview.\"\n L.w(msg)\n # msgbox(msg)\n x = dialogbox(msg, ['Yes', 'No'], QUESTION)\n if (x == 'No'): \n return False # successful config, but don't sync right now\n return True\n \ndef launch_paths(suppressExceptions=False):\n \"\"\" Open the addon folder (ignoring any errors), \n and open the log file (not ignoring errors, unless suppress).\n Typically called when there's a problem, to show the user where to go fix it.\n \"\"\"\n\n tmp = os.path.abspath(os.curdir)\n folder = os.path.split(SX.LOG_FP)[0]\n try:\n launch_file(folder) # Launch the folder containing the log file\n except: \n pass # ignore\n \n try:\n L.close_file()\n launch_file(SX.LOG_FP) # Launch the log file (e.g. in Notepad)\n except:\n if not suppress:\n raise\n \ndef launch_paths_maybe():\n \"\"\" If there are errors/warnings, open the addon folder and the log file. \"\"\"\n if L.error_count() or L.warn_count():\n launch_paths()\n else:\n L.close_file()\n\ndef launch_file(filepath):\n \"\"\"See: http://stackoverflow.com/questions/434597/open-document-with-default-application-in-python\n Also note this alternative. Quote:\n I tried this code and it worked fine in Windows 7 and Ubuntu Natty:\n import webbrowser\n webbrowser.open(\"path_to_file\")\n \"\"\"\n if SX.WINDOWS:\n os.startfile(filepath)\n elif SX.MAC:\n import subprocess\n subprocess.call(('open', filepath))\n elif SX.LINUX:\n import subprocess\n subprocess.call(('xdg-open', filepath)) \n\nif A.IN_ANKI:\n mw.form.menuTools.addSeparator()\n # create a new menu item in Anki\n action = QAction(SX.ADDON_NAME, mw)\n # set it to call our function when it's clicked\n mw.connect(action, SIGNAL(\"triggered()\"), on_sync_clicked)\n # and add it to the tools menu\n mw.form.menuTools.addAction(action)\n \n action = QAction('(re)configure ' + SX.ADDON_SHORT_NAME, mw)\n mw.connect(action, SIGNAL(\"triggered()\"), on_reconfigure_clicked)\n mw.form.menuTools.addAction(action)\n\ncpath = SX.get_config_path()\nif A.IN_ANKI and (not os.path.exists(cpath)):\n showInfo(CFGMSG + \"\\n\" + IMPMSG)\n\n# TODO: have the config file indicate whether to sync automatically whenever Anki starts (e.g. for WeSay)\n# But, there's a problem: mw.col is None at this point in Anki's loading process, so we \n# need a different way to get the collection object if we have to get it NOW. \n", "sub_path": "syncxml/SyncFromXML.py", "file_name": "SyncFromXML.py", "file_ext": "py", "file_size_in_byte": 19166, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "aqt.utils.showInfo", "line_number": 46, "usage_type": "call"}, {"api_name": "aqt.utils.askUserDialog", "line_number": 53, "usage_type": "call"}, {"api_name": "aqt.mw.app.setOverrideCursor", "line_number": 62, "usage_type": "call"}, {"api_name": "aqt.mw.app", "line_number": 62, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 62, "usage_type": "name"}, {"api_name": "aqt.mw.app.restoreOverrideCursor", "line_number": 65, "usage_type": "call"}, {"api_name": "aqt.mw.app", "line_number": 65, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 65, "usage_type": "name"}, {"api_name": "aqt.mw.pm.profileFolder", "line_number": 92, "usage_type": "call"}, {"api_name": "aqt.mw.pm", "line_number": 92, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 92, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path", "line_number": 96, "usage_type": "attribute"}, {"api_name": "aqt.mw.pm.profileFolder", "line_number": 111, "usage_type": "call"}, {"api_name": "aqt.mw.pm", "line_number": 111, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 111, "usage_type": "name"}, {"api_name": "os.rename", "line_number": 134, "usage_type": "call"}, {"api_name": "aqt.mw.col.models.flush", "line_number": 160, "usage_type": "call"}, {"api_name": "aqt.mw.col", "line_number": 160, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 160, "usage_type": "name"}, {"api_name": "aqt.mw.reset", "line_number": 161, "usage_type": "call"}, {"api_name": "aqt.mw", "line_number": 161, "usage_type": "name"}, {"api_name": "aqt.mw.reset", "line_number": 162, "usage_type": "call"}, {"api_name": "aqt.mw", "line_number": 162, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 187, "usage_type": "call"}, {"api_name": "os.path", "line_number": 187, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 192, "usage_type": "call"}, {"api_name": "os.path", "line_number": 192, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 237, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 241, "usage_type": "call"}, {"api_name": "os.path", "line_number": 241, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 247, "usage_type": "call"}, {"api_name": "os.path", "line_number": 247, "usage_type": "attribute"}, {"api_name": "aqt.utils.getFile", "line_number": 251, "usage_type": "call"}, {"api_name": "aqt.mw", "line_number": 251, "usage_type": "argument"}, {"api_name": "os.path.exists", "line_number": 253, "usage_type": "call"}, {"api_name": "os.path", "line_number": 253, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 270, "usage_type": "call"}, {"api_name": "os.path", "line_number": 270, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 271, "usage_type": "call"}, {"api_name": "os.path", "line_number": 271, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 272, "usage_type": "call"}, {"api_name": "os.path", "line_number": 272, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 272, "usage_type": "call"}, {"api_name": "aqt.utils.getFile", "line_number": 303, "usage_type": "call"}, {"api_name": "aqt.mw", "line_number": 303, "usage_type": "argument"}, {"api_name": "os.path.getsize", "line_number": 315, "usage_type": "call"}, {"api_name": "os.path", "line_number": 315, "usage_type": "attribute"}, {"api_name": "shutil.move", "line_number": 371, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 395, "usage_type": "call"}, {"api_name": "os.path", "line_number": 395, "usage_type": "attribute"}, {"api_name": "os.curdir", "line_number": 395, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 396, "usage_type": "call"}, {"api_name": "os.path", "line_number": 396, "usage_type": "attribute"}, {"api_name": "os.startfile", "line_number": 424, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 427, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 430, "usage_type": "call"}, {"api_name": "aqt.mw.form.menuTools.addSeparator", "line_number": 433, "usage_type": "call"}, {"api_name": "aqt.mw.form", "line_number": 433, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 433, "usage_type": "name"}, {"api_name": "aqt.mw", "line_number": 435, "usage_type": "argument"}, {"api_name": "aqt.mw.connect", "line_number": 437, "usage_type": "call"}, {"api_name": "aqt.mw", "line_number": 437, "usage_type": "name"}, {"api_name": "aqt.mw.form.menuTools.addAction", "line_number": 439, "usage_type": "call"}, {"api_name": "aqt.mw.form", "line_number": 439, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 439, "usage_type": "name"}, {"api_name": "aqt.mw", "line_number": 441, "usage_type": "argument"}, {"api_name": "aqt.mw.connect", "line_number": 442, "usage_type": "call"}, {"api_name": "aqt.mw", "line_number": 442, "usage_type": "name"}, {"api_name": "aqt.mw.form.menuTools.addAction", "line_number": 443, "usage_type": "call"}, {"api_name": "aqt.mw.form", "line_number": 443, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 443, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 446, "usage_type": "call"}, {"api_name": "os.path", "line_number": 446, "usage_type": "attribute"}, {"api_name": "aqt.utils.showInfo", "line_number": 447, "usage_type": "call"}]} +{"seq_id": "363664983", "text": "# Magic statements.\n\nimport warnings\nwarnings.filterwarnings('ignore')\nwarnings.simplefilter('ignore')\n\nfrom pprint import pprint\nimport time\n\n# Import graph libraries.\nimport matplotlib.pyplot as plt\nfrom matplotlib.ticker import MultipleLocator, AutoMinorLocator\n\n# Import main modules, packages, and third party libraries.\nimport numpy as np; from numpy import nan\nimport pandas as pd\nimport seaborn as sns; sns.set()\n\n# Import scikit-learn classes: datasets.\nfrom sklearn.datasets import fetch_20newsgroups\nfrom sklearn.datasets import load_digits\nfrom sklearn.datasets import load_iris\n\n# Import scikit-learn classes: preprocessing step utility functions.\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.decomposition import TruncatedSVD\nfrom sklearn.decomposition import PCA # Unsupervised Machine Learning tasks: feature reduction, dimensionality reduction\nfrom sklearn.mixture import GaussianMixture # Unsupervised Machine Learning tasks: clustering\nfrom sklearn.manifold import Isomap # Unsupervised Machine Learning tasks: feature reduction, dimensionality reduction\n\n# Import scikit-learn classes: models (Estimators).\nfrom sklearn.naive_bayes import GaussianNB\nfrom sklearn.naive_bayes import MultinomialNB\nfrom sklearn.linear_model import LinearRegression\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn.linear_model import Ridge\n\n# Import scikit-learn classes: preprocessing.\nfrom sklearn.preprocessing import PolynomialFeatures\n# from sklearn.preprocessing import Imputer\nfrom sklearn.pipeline import make_pipeline\nfrom sklearn.utils import shuffle\n\nfrom sklearn.neighbors import KNeighborsClassifier\n\n# Import scikit-learn classes: Hyperparameters Validation utility functions.\nfrom sklearn.model_selection import cross_val_score\nfrom sklearn.model_selection import LeavePOut\nfrom sklearn.model_selection import LeaveOneOut\n\nfrom sklearn.model_selection import StratifiedKFold\nfrom sklearn.model_selection import StratifiedShuffleSplit\nfrom sklearn.model_selection import GridSearchCV\nfrom sklearn.model_selection import KFold\n\n# Import scikit-learn classes: model's evaluation step utility functions.\nfrom sklearn.metrics import accuracy_score \nfrom sklearn.metrics import confusion_matrix\n\n\ndef decorator_run_pipeline_cv(\n a_func\n ):\n \n def wrapper_decorator(\n model=None, train=None,\n test=None, random_state=0, cv=None,\n ):\n assert model is not None, 'model is None'\n assert train is not None, 'train is None'\n assert test is not None, 'test is None'\n \n print('[*] Running pipeline...')\n\n a_func(model, train, test, random_state, cv)\n \n print('[*] Pipeline done.')\n pass\n\n return wrapper_decorator\n\n@decorator_run_pipeline_cv\ndef run_pipeline(model, train, test, random_state=0, cv=None):\n \n # Shuffle data used for training.\n Xtrain, ytrain = train.data, train.target\n Xtrain, ytrain = shuffle(Xtrain, ytrain, random_state=random_state)\n\n # Shuffle data used for test phase, if any.\n Xtest, ytest = test.data, test.target\n Xtest, ytest = shuffle(Xtest, ytest, random_state=random_state)\n\n # check whether to perform cv, having provided as input argument\n # passed to this function a quantity representing:\n # - either, number of folds in which training set will be splitted\n # - or, a cross-validation scheme, techique, pattern, represented\n # by means of a Scikit-Learn class object.\n if cv is not None:\n print('[*] CV running...')\n scores = cross_val_score(model, Xtrain, ytrain , cv=cv)\n print(scores)\n print(scores.mean())\n print('[*] CV done.')\n \n # Fit the model to training data\n model.fit(Xtrain, ytrain)\n labels = model.predict(Xtest)\n \n if test is not None:\n mat = confusion_matrix(ytest, labels)\n sns.heatmap(mat.T, square=True,\n annot=True, fmt='d',\n cbar=False,\n xticklabels=train.target_names, yticklabels=train.target_names, )\n plt.xlabel('true label')\n plt.ylabel('predicted label')\n\n print('K-Neighbors Classifier accuracy score:', accuracy_score(ytest, labels))\n print(f\"K-Neighbors Classifier accuracy score (percentage): {accuracy_score(ytest, labels)*100:.2f}%\")\n \n def predict_category(s, train=train, model=model):\n pred = model.predict([s])\n \n return ', '.join([ str(pred), str(train.target_names[pred[0]]) ])\n \n print(predict_category('sending a payload to the ISS'))\n print(predict_category('discussing islam versus atheism'))\n print(predict_category('determinig screen resolution and size'))\n return model\n\ndef decorator_inner_vs_outer_cv(\n a_func\n ):\n def wrapper_decorator(\n estimator,\n Xtrain, ytrain,\n param_grid,\n Xtest, ytest,\n num_trials=30,\n random_state=0,\n n_splits_i=4,\n n_splits_o=4,\n verbose=0,\n ):\n\n assert estimator is not None, 'cls is None'\n assert param_grid is not None, 'param_grid is None'\n\n print('[*] inner_vs_outer_cv running...')\n a_func(\n estimator,\n Xtrain, ytrain,\n param_grid,\n Xtest, ytest,\n test=None,\n num_trials=30,\n random_state=0,\n n_splits_i=n_splits_i,\n n_splits_o=n_splits_o,\n verbose=0\n )\n print('[*] inner_vs_outer_cv done.')\n pass\n return wrapper_decorator\n\n@decorator_inner_vs_outer_cv\ndef inner_vs_outer_cv(\n estimator,\n Xtrain, ytrain,\n param_grid,\n Xtest, ytest,\n test=None,\n num_trials=30,\n random_state=0,\n n_splits_i=4,\n n_splits_o=4,\n verbose=0):\n \n Xtrain, ytrain = shuffle(Xtrain, ytrain, random_state=random_state)\n\n if Xtest is not None and ytest is not None:\n Xtest, ytest = shuffle(Xtest, ytest, random_state=random_state)\n\n # Arrays to store scores\n non_nested_scores = np.zeros(num_trials)\n nested_scores = np.zeros(num_trials)\n \n best_models = list()\n cv_results = list()\n best_params = list()\n\n # Loop for each trial\n for i in range(num_trials):\n # Choose cross-validation techniques for the inner and outer loops,\n # independently of the dataset.\n # E.g \"GroupKFold\", \"LeaveOneOut\", \"LeaveOneGroupOut\", etc.\n inner_cv = StratifiedKFold(n_splits=n_splits_i, random_state=i, shuffle=False) # KFold(n_splits=4, shuffle=True, random_state=i)\n outer_cv = StratifiedKFold(n_splits=n_splits_o, random_state=i, shuffle=False) # KFold(n_splits=4, shuffle=True, random_state=i)\n\n # Non_nested parameter search and scoring\n clf = GridSearchCV(estimator=estimator, param_grid=param_grid, cv=inner_cv)\n clf.fit(Xtrain, ytrain)\n non_nested_scores[i] = clf.best_score_\n \n best_models.append(clf.best_estimator_)\n cv_results.append(clf.cv_results_)\n best_params.append(clf.best_params_)\n\n # Nested CV with parameter optimization\n nested_score = cross_val_score(clf, X=Xtrain, y=ytrain, cv=outer_cv)\n nested_scores[i] = nested_score.mean()\n\n score_difference = non_nested_scores - nested_scores\n\n print(\"Average difference of {:6f} with std. dev. of {:6f}.\"\n .format(score_difference.mean(), score_difference.std()))\n \n \n plot_scores_nested_vs_non_nested(\n non_nested_scores,\n nested_scores,\n score_difference,\n num_trials=num_trials)\n return best_models, cv_results, best_params\n\ndef plot_scores_nested_vs_non_nested(non_nested_scores, nested_scores, score_difference, num_trials):\n # Plot scores on each trial for nested and non-nested CV\n plt.figure()\n plt.subplot(211)\n non_nested_scores_line, = plt.plot(non_nested_scores, color='r')\n nested_line, = plt.plot(nested_scores, color='b')\n plt.ylabel(\"score\", fontsize=\"14\")\n plt.legend([non_nested_scores_line, nested_line],\n [\"Non-Nested CV\", \"Nested CV\"],\n bbox_to_anchor=(0, .4, .5, 0))\n plt.title(\"Non-Nested and Nested Cross Validation on Iris Dataset\",\n x=.5, y=1.1, fontsize=\"15\")\n\n # Plot bar chart of the difference.\n plt.subplot(212)\n difference_plot = plt.bar(range(num_trials), score_difference)\n plt.xlabel(\"Individual Trial #\")\n plt.legend([difference_plot],\n [\"Non-Nested CV - Nested CV Score\"],\n bbox_to_anchor=(0, 1, .8, 0))\n plt.ylabel(\"score difference\", fontsize=\"14\")\n\n plt.show()\n pass\n\ndef plot_cm(ytest, y_model):\n mat = confusion_matrix(ytest, y_model)\n\n sns.heatmap(mat, square=True, annot=True, cbar=False)\n plt.xlabel('predicted value')\n plt.ylabel('true value')\n plt.show()\n pass\n ", "sub_path": "Python_Data_Science_Handbook/chapter_5/Pittsburgh_Bridges_Dataset/utils/grid_search_cv_utils.py", "file_name": "grid_search_cv_utils.py", "file_ext": "py", "file_size_in_byte": 8780, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "warnings.filterwarnings", "line_number": 4, "usage_type": "call"}, {"api_name": "warnings.simplefilter", "line_number": 5, "usage_type": "call"}, {"api_name": "seaborn.set", "line_number": 17, "usage_type": "call"}, {"api_name": "sklearn.utils.shuffle", "line_number": 87, "usage_type": "call"}, {"api_name": "sklearn.utils.shuffle", "line_number": 91, "usage_type": "call"}, {"api_name": "sklearn.model_selection.cross_val_score", "line_number": 100, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 110, "usage_type": "call"}, {"api_name": "seaborn.heatmap", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 118, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 119, "usage_type": "call"}, {"api_name": "sklearn.utils.shuffle", "line_number": 179, "usage_type": "call"}, {"api_name": "sklearn.utils.shuffle", "line_number": 182, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 185, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 186, "usage_type": "call"}, {"api_name": "sklearn.model_selection.StratifiedKFold", "line_number": 197, "usage_type": "call"}, {"api_name": "sklearn.model_selection.StratifiedKFold", "line_number": 198, "usage_type": "call"}, {"api_name": "sklearn.model_selection.GridSearchCV", "line_number": 201, "usage_type": "call"}, {"api_name": "sklearn.model_selection.cross_val_score", "line_number": 210, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 228, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 228, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 229, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 229, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 230, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 230, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 231, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 231, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 232, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 232, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 233, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 233, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 236, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 236, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 240, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 240, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 241, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 241, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 242, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 242, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 243, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 243, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 246, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 246, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 248, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 248, "usage_type": "name"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 252, "usage_type": "call"}, {"api_name": "seaborn.heatmap", "line_number": 254, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 255, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 255, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 256, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 256, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 257, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 257, "usage_type": "name"}]} +{"seq_id": "28338040", "text": "'''\r\nface++ 眼镜定位, 镜框贴图 , 多图\r\n'''\r\nfrom __future__ import division\r\n\r\nimport requests\r\nfrom json import JSONDecoder\r\nimport cv2\r\nimport numpy as np\r\nfrom matplotlib import pyplot as plt\r\nfrom PIL import Image\r\nfrom scipy import misc\r\nimport os\r\nfrom Dir_make import mk_dir\r\nimport time\r\nimport math\r\n\r\n# 初始化\r\n# 原始镜框信息\r\nglass_filepath = 'D:/Anaconda3/Lib/site-packages/facenet/glass.png'\r\n# 原始镜框定位点\r\n# 定位点\r\nglass_left = {'x': 80, 'y': 43}\r\nglass_right = {'x': 220, 'y': 43}\r\n# 眼镜对应间距\r\nglass_point_distance = glass_right['x'] - glass_left['x']\r\n\r\n\r\ndef faceplusplus_face_detect_api_s(filepath, class_name):\r\n num_images = len(filepath)\r\n glass_point_store_list = []\r\n face_glass_path_list = []\r\n for i in range(num_images):\r\n # 图片路径\r\n # filepath = \"D:/Anaconda3/Lib/site-packages/facenet/data/test_image/1.png\"\r\n\r\n # 原图显示\r\n clean_img = cv2.imread(filepath[i])\r\n # cv2.imshow('clean_image', clean_img)\r\n # cv2.waitKey(1)\r\n\r\n # 人眼检测\r\n left_eye_center, right_eye_center, angle = eye_detect(filepath[i])\r\n if str(left_eye_center) == 'False':\r\n continue\r\n else:\r\n print(str(left_eye_center))\r\n print(str(right_eye_center))\r\n '''\r\n # 标注瞳孔\r\n point_img_left_eye = cv2.circle(clean_img, (left_eye_center['x'], left_eye_center['y']), 2, (0, 0, 255),\r\n -1) # circle(图像,圆心,半径,颜色,填充)\r\n point_img = cv2.circle(point_img_left_eye, (right_eye_center['x'], right_eye_center['y']), 2, (0, 0, 255), -1)\r\n\r\n cv2.imshow('point_image', point_img)\r\n cv2.imwrite('point_image.png', point_img)\r\n cv2.waitKey(1)\r\n '''\r\n # 缩放比例计算,新镜框定位点\r\n new_glass_left, new_glass_right = scale(left_eye_center, right_eye_center)\r\n\r\n # 贴图, 并获取镜框坐标\r\n store_point, face = wear_glass_1(filepath[i], new_glass_left, left_eye_center, angle)\r\n\r\n glass_point_store_list.append(store_point)\r\n # 数组转化为Image类\r\n face_glass = Image.fromarray(face)\r\n\r\n # 保存路径生成\r\n\r\n pre_dir = 'D:/Anaconda3/Lib/site-packages/facenet/data/face_glass_image/'\r\n mk_dir(pre_dir + class_name[i])\r\n\r\n start_index = filepath[i].find(class_name[i] + '_')\r\n\r\n face_glass_path = 'D:/Anaconda3/Lib/site-packages/facenet/data/face_glass_image/' + class_name[i] + '/' + class_name[i] + filepath[i][start_index: -4] + '_glass.png'\r\n print(face_glass_path)\r\n face_glass_path_list.append(face_glass_path)\r\n face_glass.save(face_glass_path, 'png')\r\n\r\n return glass_point_store_list, face_glass_path, pre_dir\r\n\r\n\r\ndef eye_detect(filepath):\r\n '''\r\n 调用Face++ 人眼定位\r\n '''\r\n\r\n http_url = \"https://api-cn.faceplusplus.com/facepp/v3/detect\"\r\n key = \"dHTpFppve9bV3ZqkvJadC74fQql3paRp\"\r\n secret = \"19R5whMXkjXyFCPEivqHPCrelkxfxByi\"\r\n\r\n data = {\"api_key\": key, \"api_secret\": secret, \"return_landmark\": \"1\"}\r\n files = {\"image_file\": open(filepath, \"rb\")}\r\n # print(files)\r\n response = requests.post(http_url, data=data, files=files)\r\n\r\n req_con = response.content.decode('utf-8')\r\n req_dict = JSONDecoder().decode(req_con)\r\n\r\n #print(req_dict)\r\n # print(req_dict['faces'])\r\n #print(req_dict['faces'][0])\r\n #print(req_dict['faces'][0]['landmark'])\r\n\r\n '''\r\n # API调用失败策略\r\n if 'faces' not in req_dict:\r\n landmark_inf = False\r\n left_eye_center = False\r\n right_eye_center = False\r\n angle = False\r\n return left_eye_center, right_eye_center, angle\r\n else:\r\n landmark_inf = req_dict['faces'][0]['landmark']\r\n left_eye_center = landmark_inf['left_eye_center']\r\n right_eye_center = landmark_inf['right_eye_center']\r\n #print('%s: %s' % ('left_eye_center', str(left_eye_center)))\r\n #print('%s: %s' % ('right_eye_center', str(right_eye_center)))\r\n '''\r\n # 远程调用防中断\r\n while 'faces' not in req_dict:\r\n response = requests.post(http_url, data=data, files=files)\r\n\r\n req_con = response.content.decode('utf-8')\r\n req_dict = JSONDecoder().decode(req_con)\r\n\r\n landmark_inf = req_dict['faces'][0]['landmark']\r\n left_eye_center = landmark_inf['left_eye_center']\r\n right_eye_center = landmark_inf['right_eye_center']\r\n\r\n # 求倾斜角度\r\n delta_x = left_eye_center['x'] - right_eye_center['x']\r\n delta_y = left_eye_center['y'] - right_eye_center['y']\r\n tan_ = delta_y / delta_x\r\n print('%s: %f' % ('tan', tan_))\r\n angle = math.atan(tan_)\r\n\r\n return left_eye_center, right_eye_center, angle\r\n\r\n\r\ndef scale(left_eye_center, right_eye_center):\r\n '''\r\n 获取缩放后的镜框定位点\r\n\r\n :param left_eye_center:\r\n :param right_eye_center:\r\n :param glass_point_distance:\r\n :return:\r\n '''\r\n # 瞳孔间距\r\n real_eye_distance = right_eye_center['x'] - left_eye_center['x']\r\n # 放缩比例\r\n k = real_eye_distance / glass_point_distance\r\n print('%s: %f' % ('k', k))\r\n\r\n # 镜框缩放\r\n glass = Image.open(glass_filepath)\r\n w, h = glass.size\r\n # glass_new = misc.imresize(glass, k)\r\n glass.thumbnail((int(w * h), int(h * k)))\r\n glass.save('glass_new.png', 'png')\r\n\r\n # 新镜框定位点\r\n new_glass_left = {}\r\n new_glass_right = {}\r\n\r\n # 定位点缩放\r\n new_glass_left['x'] = int(glass_left['x'] * k)\r\n new_glass_left['y'] = int(glass_left['y'] * k)\r\n new_glass_right['x'] = int(glass_right['x'] * k)\r\n new_glass_right['y'] = int(glass_right['y'] * k)\r\n\r\n return new_glass_left, new_glass_right\r\n\r\n\r\ndef wear_glass_1(filepath, new_glass_left, left_eye_center, angle):\r\n '''\r\n 眼镜贴图方法一:\r\n 根据左眼和镜框左定位点重合,进行像素替换\r\n\r\n :param new_glass_left:\r\n :return:\r\n '''\r\n # 存放镜框位置坐标\r\n store_point = [] # [i, j]\r\n pkg_point = []\r\n '''\r\n begin_point = {}\r\n begin_point['x'] = left_eye_center['x'] - new_glass_left['x']\r\n begin_point['y'] = left_eye_center['y'] - new_glass_left['y']\r\n '''\r\n face = np.array(Image.open(filepath))\r\n\r\n new_glass = np.array(Image.open('glass_new.png'))\r\n h, w, dim = new_glass.shape\r\n H, W, Dim = face.shape\r\n\r\n cos_ = math.cos(angle)\r\n sin_ = math.sin(angle)\r\n\r\n start_time = time.time()\r\n\r\n for i in range(w):\r\n for j in range(h):\r\n if (sum(new_glass[j][i]) <= 510): # 若为镜框点\r\n pkg_point.append([j, i])\r\n\r\n for i in range(W):\r\n for j in range(H):\r\n vec_x_0 = i - left_eye_center['x']\r\n vec_y_0 = j - left_eye_center['y']\r\n vec_x = int(vec_x_0 * cos_ + vec_y_0 * sin_)\r\n vec_y = int(vec_x_0 * (-sin_) + vec_y_0 * cos_)\r\n if [(new_glass_left['y'] + vec_y), (new_glass_left['x'] + vec_x)] in pkg_point:\r\n for a in range(-1, 2):\r\n for b in range(-1, 2):\r\n if 0<=j+b<=159 and 0<=i+a<=159 and (a*b == 0):\r\n face[j+b][i+a] = [0, 0, 0]\r\n if [j+b, i+a] not in store_point:\r\n store_point.append([j+b, i+a])\r\n end_time = time.time()\r\n print(end_time - start_time)\r\n print('%s: %f %s' % ('about', len(store_point) / (16 * 16), '%'))\r\n return store_point, face\r\n\r\n\r\n'''\r\nif __name__ == '__faceplusplus_face_detect_api__':\r\n faceplusplus_face_detect_api()\r\n'''", "sub_path": "Faceplusplus_face_detect_API_s.py", "file_name": "Faceplusplus_face_detect_API_s.py", "file_ext": "py", "file_size_in_byte": 7658, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "cv2.imread", "line_number": 38, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 67, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 67, "usage_type": "name"}, {"api_name": "Dir_make.mk_dir", "line_number": 72, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 96, "usage_type": "call"}, {"api_name": "json.JSONDecoder", "line_number": 99, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 123, "usage_type": "call"}, {"api_name": "json.JSONDecoder", "line_number": 126, "usage_type": "call"}, {"api_name": "math.atan", "line_number": 137, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 158, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 158, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 193, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 193, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 193, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 195, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 195, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 195, "usage_type": "name"}, {"api_name": "math.cos", "line_number": 199, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 200, "usage_type": "call"}, {"api_name": "time.time", "line_number": 202, "usage_type": "call"}, {"api_name": "time.time", "line_number": 222, "usage_type": "call"}]} +{"seq_id": "625198205", "text": "# Yhlin\nimport os\nimport tensorflow as tf\nimport argparse\nimport time\nimport json\nimport glob\nimport shutil\nfrom dataset_factory import get_custom_dataset, CUSTOM_DATASET_INFO\nfrom resnet_model import ResNet, get_block_sizes\n\nos.environ['TF_CPP_MIN_LOG_LEVEL'] = '3'\n\n\ndef get_optimizer(optim_name, lr):\n if optim_name == 'adam':\n return tf.train.AdamOptimizer(lr)\n elif optim_name == 'rmsprop':\n return tf.train.RMSPropOptimizer(lr)\n elif optim_name == 'momentum':\n return tf.train.MomentumOptimizer(lr, momentum=0.9)\n else:\n return tf.train.GradientDescentOptimizer(lr)\n\n\ndef get_learning_rate(global_step, batch_size, num_images,\n boundary_epochs, decay_rates,\n base_lr=0.1, warmup=False, decay=False):\n batches_per_epoch = num_images / batch_size\n if decay:\n boundaries = [int(batches_per_epoch * epoch) for epoch in\n boundary_epochs]\n vals = [base_lr * rate for rate in decay_rates]\n lr = tf.train.piecewise_constant(global_step, boundaries, vals)\n else:\n lr = tf.convert_to_tensor(base_lr, tf.float32)\n if warmup:\n warmup_steps = int(batches_per_epoch * 5)\n warmup_lr = (base_lr * tf.cast(global_step, tf.float32) /\n tf.cast(warmup_steps, tf.float32))\n return tf.cond(pred=global_step < warmup_steps,\n true_fn=lambda: warmup_lr,\n false_fn=lambda: lr)\n return lr\n\n\ndef reset_last_ckpt_dir(ckpt_dir):\n count = 1\n ckpt_dir += '_0'\n while os.path.exists(ckpt_dir):\n index = ckpt_dir.rfind('_')\n ckpt_dir = ckpt_dir[:index] + '_{}'.format(count)\n count += 1\n count -= 1\n index = ckpt_dir.rfind('_')\n ckpt_dir = ckpt_dir[:index] + '_{}'.format(count)\n shutil.rmtree(ckpt_dir)\n return ckpt_dir\n\n\ndef prepare_ckpt_dir(ckpt_dir):\n count = 1\n ckpt_dir += '_0'\n while os.path.exists(ckpt_dir):\n index = ckpt_dir.rfind('_')\n ckpt_dir = ckpt_dir[:index] + '_{}'.format(count)\n count += 1\n return ckpt_dir\n\n\ndef main(args):\n ckpt_dir = args.ckpt_dir\n num_epochs = args.num_epochs\n batch_size = args.batch_size\n data_dir = args.data_dir\n boundaries = [int(x) for x in args.boundaries.split(',')]\n decay_rates = [float(x) for x in args.decay_rates.split(',')]\n\n training = tf.placeholder(tf.bool, shape=[], name='training')\n\n # Dataset\n train_dataset = get_custom_dataset(True,\n data_dir=data_dir,\n batch_size=batch_size)\n val_dataset = get_custom_dataset(False, data_dir=data_dir,\n batch_size=batch_size)\n iterator = tf.data.Iterator.from_structure(train_dataset.output_types,\n train_dataset.output_shapes)\n features, labels = iterator.get_next()\n train_ds_init_op = iterator.make_initializer(train_dataset)\n val_ds_init_op = iterator.make_initializer(val_dataset)\n\n # Model\n resnet_size = 50\n model = ResNet(resnet_size=resnet_size,\n num_classes=CUSTOM_DATASET_INFO.num_classes,\n num_filters=64,\n kernel_size=7,\n conv_stride=2,\n first_pool_size=3,\n first_pool_stride=2,\n block_sizes=get_block_sizes(resnet_size),\n block_strides=[1, 2, 2, 2],\n resnet_version=2,\n data_format='channels_last',\n dtype=tf.float32)\n logits = model(features, training)\n loss = tf.losses.sparse_softmax_cross_entropy(\n logits=logits, labels=labels)\n\n pred = tf.argmax(input=logits, axis=1, name='classes')\n acc, acc_update_op = tf.metrics.accuracy(labels=labels,\n predictions=pred)\n metric_vars = tf.get_collection(tf.GraphKeys.METRIC_VARIABLES)\n metric_vars_init_op = tf.variables_initializer(metric_vars,\n name='metric_vars_init')\n\n global_step = tf.train.get_or_create_global_step()\n lr = get_learning_rate(global_step, batch_size,\n CUSTOM_DATASET_INFO.train,\n boundary_epochs=boundaries,\n decay_rates=decay_rates,\n base_lr=args.lr,\n warmup=args.warmup,\n decay=args.decay)\n\n tf.identity(lr, name='learning_rate')\n\n optimizer = get_optimizer(args.optim_name, lr)\n minimize_op = optimizer.minimize(loss, global_step=global_step)\n update_ops = tf.get_collection(tf.GraphKeys.UPDATE_OPS)\n train_op = tf.group(minimize_op, update_ops)\n\n config = tf.ConfigProto()\n config.gpu_options.allow_growth = True\n\n saver = tf.train.Saver()\n best_acc_history = {}\n best_acc = 0\n with tf.Session(config=config) as sess:\n sess.run(tf.global_variables_initializer())\n last_ckpt = tf.train.latest_checkpoint(ckpt_dir)\n if last_ckpt is not None:\n saver.restore(sess, last_ckpt)\n print('Restore from checkpoint: ', str(last_ckpt.split('-')[-1]))\n\n global_step_value = None\n for epoch in range(num_epochs):\n # Train\n sess.run(metric_vars_init_op)\n sess.run(train_ds_init_op)\n step = 0\n total_loss = 0\n start_time = time.time()\n lr_value = 0\n while True:\n try:\n _, _, loss_value, global_step_value, lr_value = sess.run(\n [train_op, acc_update_op, loss, global_step, lr],\n feed_dict={training: True}\n )\n total_loss += loss_value\n step += 1\n except tf.errors.OutOfRangeError:\n break\n\n acc_value = sess.run(acc)\n msg = 'Epoch: {}\\n'.format(epoch)\n msg += ' Train: Loss {:.4f} Acc {:.4f} global step {}'.format(\n total_loss / step, acc_value, global_step_value)\n msg += ' lr {:.6f}'.format(lr_value)\n msg += '({:.2f}s)'.format(time.time() - start_time)\n print(msg)\n saver.save(sess, os.path.join(ckpt_dir, 'model'),\n global_step=global_step_value)\n\n # Eval\n sess.run(metric_vars_init_op)\n sess.run(val_ds_init_op)\n step = 0\n total_loss = 0\n while True:\n try:\n loss_value, _ = sess.run(\n [loss, acc_update_op],\n feed_dict={training: False})\n total_loss += loss_value\n step += 1\n except tf.errors.OutOfRangeError:\n break\n acc_value = sess.run(acc)\n print(' Valid: Loss {:.4f} Acc {:.4f} '.format(\n total_loss / step, acc_value))\n last_ckpt = tf.train.latest_checkpoint(ckpt_dir)\n if acc_value > best_acc:\n best_acc = acc_value\n best_step = last_ckpt.split('-')[-1]\n best_acc_history[best_step] = '{:.4f}'.format(best_acc)\n print(' Gets a better result. Save it.')\n for filename in glob.glob(last_ckpt + '.*'):\n basename = os.path.basename(filename)\n if not os.path.exists(\n os.path.join(args.ckpt_dir, 'save')):\n os.makedirs(\n os.path.join(args.ckpt_dir, 'save'))\n shutil.copyfile(\n filename,\n os.path.join(args.ckpt_dir, 'save', basename))\n\n with open(os.path.join(args.ckpt_dir, 'save', 'acc.json'),\n 'w', encoding='utf-8') as fp:\n json.dump(best_acc_history, fp)\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument('--num_epochs', type=int, default=80)\n parser.add_argument('--batch_size', type=int, default=32)\n parser.add_argument('--gpus', default='1')\n parser.add_argument('--lr', type=float, default=0.00625)\n parser.add_argument('--ckpt_dir', default='...')\n parser.add_argument('--data_dir',\n default='...')\n parser.add_argument('--warmup', action='store_true', default=False)\n parser.add_argument('--decay', action='store_true', default=False)\n parser.add_argument('--boundaries', default='10,20,30,40')\n parser.add_argument('--decay_rates', default='1,0.1,0.01,0.001,0.0001')\n parser.add_argument('--optim_name', type=str, default='momentum')\n parser.add_argument('--go_on', action='store_true', default=False)\n parser.add_argument('--reset', action='store_true', default=False)\n args = parser.parse_args()\n os.environ['CUDA_VISIBLE_DEVICES'] = args.gpus\n\n if args.reset:\n assert not args.go_on\n args.ckpt_dir = reset_last_ckpt_dir(args.ckpt_dir)\n\n\n\n if not args.go_on:\n args.ckpt_dir = prepare_ckpt_dir(args.ckpt_dir)\n\n args_dict = {}\n for k, v in sorted(vars(args).items()):\n args_dict[k] = str(v)\n if not os.path.exists(args.ckpt_dir):\n os.makedirs(args.ckpt_dir)\n with open(os.path.join(args.ckpt_dir, 'cfg.json'),\n 'w', encoding='utf-8') as fp:\n json.dump(args_dict, fp, indent=4)\n\n main(args)\n", "sub_path": "train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 9628, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "os.environ", "line_number": 12, "usage_type": "attribute"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 17, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 17, "usage_type": "attribute"}, {"api_name": "tensorflow.train.RMSPropOptimizer", "line_number": 19, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 19, "usage_type": "attribute"}, {"api_name": "tensorflow.train.MomentumOptimizer", "line_number": 21, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 21, "usage_type": "attribute"}, {"api_name": "tensorflow.train.GradientDescentOptimizer", "line_number": 23, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 23, "usage_type": "attribute"}, {"api_name": "tensorflow.train.piecewise_constant", "line_number": 34, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 34, "usage_type": "attribute"}, {"api_name": "tensorflow.convert_to_tensor", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 36, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 39, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 39, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 40, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tensorflow.cond", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 79, "usage_type": "call"}, {"api_name": "tensorflow.bool", "line_number": 79, "usage_type": "attribute"}, {"api_name": "dataset_factory.get_custom_dataset", "line_number": 82, "usage_type": "call"}, {"api_name": "dataset_factory.get_custom_dataset", "line_number": 85, "usage_type": "call"}, {"api_name": "tensorflow.data.Iterator.from_structure", "line_number": 87, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 87, "usage_type": "attribute"}, {"api_name": "resnet_model.ResNet", "line_number": 95, "usage_type": "call"}, {"api_name": "dataset_factory.CUSTOM_DATASET_INFO.num_classes", "line_number": 96, "usage_type": "attribute"}, {"api_name": "dataset_factory.CUSTOM_DATASET_INFO", "line_number": 96, "usage_type": "name"}, {"api_name": "resnet_model.get_block_sizes", "line_number": 102, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 106, "usage_type": "attribute"}, {"api_name": "tensorflow.losses.sparse_softmax_cross_entropy", "line_number": 108, "usage_type": "call"}, {"api_name": "tensorflow.losses", "line_number": 108, "usage_type": "attribute"}, {"api_name": "tensorflow.argmax", "line_number": 111, "usage_type": "call"}, {"api_name": "tensorflow.metrics.accuracy", "line_number": 112, "usage_type": "call"}, {"api_name": "tensorflow.metrics", "line_number": 112, "usage_type": "attribute"}, {"api_name": "tensorflow.get_collection", "line_number": 114, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 114, "usage_type": "attribute"}, {"api_name": "tensorflow.variables_initializer", "line_number": 115, "usage_type": "call"}, {"api_name": "tensorflow.train.get_or_create_global_step", "line_number": 118, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 118, "usage_type": "attribute"}, {"api_name": "dataset_factory.CUSTOM_DATASET_INFO.train", "line_number": 120, "usage_type": "attribute"}, {"api_name": "dataset_factory.CUSTOM_DATASET_INFO", "line_number": 120, "usage_type": "name"}, {"api_name": "tensorflow.identity", "line_number": 127, "usage_type": "call"}, {"api_name": "tensorflow.get_collection", "line_number": 131, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 131, "usage_type": "attribute"}, {"api_name": "tensorflow.group", "line_number": 132, "usage_type": "call"}, {"api_name": "tensorflow.ConfigProto", "line_number": 134, "usage_type": "call"}, {"api_name": "tensorflow.train.Saver", "line_number": 137, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 137, "usage_type": "attribute"}, {"api_name": "tensorflow.Session", "line_number": 140, "usage_type": "call"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 141, "usage_type": "call"}, {"api_name": "tensorflow.train.latest_checkpoint", "line_number": 142, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 142, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 154, "usage_type": "call"}, {"api_name": "tensorflow.errors", "line_number": 164, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 172, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 174, "usage_type": "call"}, {"api_name": "os.path", "line_number": 174, "usage_type": "attribute"}, {"api_name": "tensorflow.errors", "line_number": 189, "usage_type": "attribute"}, {"api_name": "tensorflow.train.latest_checkpoint", "line_number": 194, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 194, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 200, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 201, "usage_type": "call"}, {"api_name": "os.path", "line_number": 201, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 202, "usage_type": "call"}, {"api_name": "os.path", "line_number": 202, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 203, "usage_type": "call"}, {"api_name": "os.path", "line_number": 203, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 204, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 205, "usage_type": "call"}, {"api_name": "os.path", "line_number": 205, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 206, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 208, "usage_type": "call"}, {"api_name": "os.path", "line_number": 208, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 210, "usage_type": "call"}, {"api_name": "os.path", "line_number": 210, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 212, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 216, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 232, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 246, "usage_type": "call"}, {"api_name": "os.path", "line_number": 246, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 247, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 248, "usage_type": "call"}, {"api_name": "os.path", "line_number": 248, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 250, "usage_type": "call"}]} +{"seq_id": "210792947", "text": "import numpy as np\nimport problem1\nimport statsmodels.api as sm\nimport matplotlib.pyplot as plt\n\ndef calcLambda(E, v):\n return v*E / ((1.+v)*(1.-2.*v))\n\ndef calcMiu(E, v):\n return E / (2.*(1.+v))\n\ndef calcBeta(H, equivalent_stress, Miu):\n return 9.*Miu**2 / (equivalent_stress**2 * (H + 3.*Miu))\n\ndef calcStressArr(nodeX, nodeY, nodes, originalGrids, dsArr, esArr):\n grids = problem1.generateGrids(nodeX, nodeY, nodes)\n stressArr = []\n for grid, originalGrid, ds, es in zip(grids, originalGrids, dsArr, esArr):\n a = problem1.calcArea(grid)\n B = problem1.calcBmatrix(a, grid)\n du = grid - originalGrid\n strain_t = B * du.reshape((du.size, 1))\n stress_t = D * strain_t\n if es > yield_stress:\n beta = calcBeta(H, es, Miu)\n else:\n beta = 0.\n sigma33 = (Lambda - beta*ds[3]*ds[0]) * strain_t[0] +\\\n (Lambda - beta*ds[3]*ds[1]) * strain_t[1] +\\\n (-beta*ds[3]*ds[2]) * strain_t[2]\n stress_t = np.array(stress_t).reshape(stress_t.size//3, 3).flatten()\n sigma33 = np.array(sigma33).reshape(sigma33.size, 1).flatten()\n stressArr.append(list(stress_t) + list(sigma33))\n return np.array(stressArr).reshape(len(stressArr), 4, 1)\n\ndef calcDpMatrix(deviatoric_stress, equivalent_stress):\n beta = calcBeta(H, equivalent_stress, Miu)\n d11 = Lambda + 2.*Miu - beta*deviatoric_stress[0]**2\n d12 = Lambda - beta*deviatoric_stress[0]*deviatoric_stress[1]\n d13 = -beta * deviatoric_stress[0] * deviatoric_stress[2]\n d22 = Lambda + 2.*Miu - beta*deviatoric_stress[1]**2\n d23 = -beta * deviatoric_stress[1] * deviatoric_stress[2]\n d33 = Miu - beta * deviatoric_stress[2]**2\n Dp = np.matrix([\n [d11[0], d12[0], d13[0]],\n [d12[0], d22[0], d23[0]],\n [d13[0], d23[0], d33[0]]])\n return Dp\n\ndef calcKmatrix(stressArr, D, nodeX, nodeY, gridSize, grids):\n def calcDeviatoricStress(stress):\n mean_stress = (stress[0] + stress[1] + stress[3]) / 3\n deviatoric_stress = [\n stress[0] - mean_stress,\n stress[1] - mean_stress,\n stress[2],\n stress[3] - mean_stress\n ]\n return np.array(deviatoric_stress)\n \n def calcEquivalentStress(deviatoric_stress):\n equivalent_stress = np.sqrt(1.5 * (\n deviatoric_stress[0]**2 +\n deviatoric_stress[1]**2 +\n 2. * deviatoric_stress[2]**2 +\n deviatoric_stress[3]**2)\n )\n return equivalent_stress\n\n nodeNum = nodeX * nodeY\n K = np.matrix(np.zeros((nodeNum*2, nodeNum*2), dtype=np.float64))\n dsArr = []\n esArr = []\n for grid, stress in zip(grids, stressArr):\n _nodes = [problem1.calcNodeNo(node, nodeY, gridSize) for node in grid]\n a = problem1.calcArea(grid)\n b = problem1.calcBmatrix(a, grid)\n\n deviatoric_stress = calcDeviatoricStress(stress)\n equivalent_stress = calcEquivalentStress(deviatoric_stress)\n dsArr.append(deviatoric_stress)\n esArr.append(equivalent_stress)\n\n if equivalent_stress > yield_stress:\n Dp = calcDpMatrix(deviatoric_stress, equivalent_stress)\n k = a * b.transpose() * Dp * b\n else:\n k = a * b.transpose() * D * b\n for i, nodeNoX in enumerate(_nodes):\n x = i * 2\n nodeNoX *= 2\n for j, nodeNoY in enumerate(_nodes):\n y = j * 2\n nodeNoY *= 2\n K[nodeNoX:nodeNoX+2, nodeNoY:nodeNoY+2] += k[x:x+2, y:y+2]\n return K, np.array(dsArr), np.array(esArr)\n\ndef tensile_test(u, stressArr, D, gridSize, nodeX, nodeY, nodes):\n grids = problem1.generateGrids(nodeX, nodeY, nodes)\n K, dsArr, esArr = calcKmatrix(stressArr, D, nodeX, nodeY, gridSize, grids)\n\n f = np.zeros_like(u)\n f -= K * u\n f_ = f[nodeX*2:]\n f_ = np.delete(f_, range(nodeX*(nodeY-2)*2+1, nodeX*(nodeY-1)*2, 2), axis=0)\n\n K_ = K[nodeX*2:, nodeY*2:]\n K_ = np.delete(K_, range(nodeX*(nodeY-2)*2+1, nodeX*(nodeY-1)*2, 2), axis=0)\n K_ = np.delete(K_, range(nodeX*(nodeY-2)*2+1, nodeX*(nodeY-1)*2, 2), axis=1)\n\n u_ = np.linalg.solve(K_, f_)\n u[nodeX*2:nodeX*(nodeY-1)*2] += u_[:-nodeX]\n u[-nodeX*2::2] += u_[-nodeX:]\n\n nodes += u.reshape(u.shape[0]//2, 2)\n\n f = K * u\n\n stressArr = calcStressArr(nodeX, nodeY, nodes, originalGrids, dsArr, esArr)\n return np.sum(f[-nodeX*2+1::2]), stressArr, nodes\n\nif __name__ == '__main__':\n E = 1.06e5 # Young's modulus\n v = .25 # Poisson's ratio\n yield_stress = 300.\n H = E / 20 # hardening modulus\n\n D = problem1.calcDmatrix(E, v)\n Lambda = calcLambda(E, v)\n Miu = calcMiu(E, v)\n\n #boundary\n x0 = 0.; xn = 1.\n y0 = 0.; yn = 1.\n length = yn - y0\n gridSize = .1\n\n nodeX, nodeY, nodes = problem1.generateNodes(x0, xn, y0, yn, gridSize)\n nodeNum = nodeX * nodeY\n originalGrids = problem1.generateGrids(nodeX, nodeY, nodes)\n stressArr = np.zeros((len(originalGrids), 4, 1))\n \n stress_axis = [0.]\n strain_axis = [0.]\n equivalent_stress = 0.\n strain_increment = .0001\n\n u = np.matrix(np.zeros((nodeNum*2, 1), dtype=np.float64))\n u[-nodeX*2+1::2] = strain_increment\n\n for counter in range(1, 45):\n stress_increment, stressArr, nodes = tensile_test(\n u, stressArr, D, gridSize, nodeX, nodeY, nodes)\n stress_axis.append(stress_axis[-1] + stress_increment)\n strain_axis.append((strain_increment*counter) / length)\n\n plt.xlabel('strain')\n plt.ylabel('stress (MPa)')\n plt.plot(strain_axis, stress_axis, 'o')\n plt.show()\n", "sub_path": "report/aoyagi/problem2.py", "file_name": "problem2.py", "file_ext": "py", "file_size_in_byte": 5651, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "problem1.generateGrids", "line_number": 16, "usage_type": "call"}, {"api_name": "problem1.calcArea", "line_number": 19, "usage_type": "call"}, {"api_name": "problem1.calcBmatrix", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.matrix", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.matrix", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 71, "usage_type": "attribute"}, {"api_name": "problem1.calcNodeNo", "line_number": 75, "usage_type": "call"}, {"api_name": "problem1.calcArea", "line_number": 76, "usage_type": "call"}, {"api_name": "problem1.calcBmatrix", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 96, "usage_type": "call"}, {"api_name": "problem1.generateGrids", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.linalg.solve", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 111, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 120, "usage_type": "call"}, {"api_name": "problem1.calcDmatrix", "line_number": 128, "usage_type": "call"}, {"api_name": "problem1.generateNodes", "line_number": 138, "usage_type": "call"}, {"api_name": "problem1.generateGrids", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.matrix", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 148, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 157, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 159, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 159, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 160, "usage_type": "name"}]} +{"seq_id": "532416875", "text": "from functools import wraps\nfrom itertools import chain\n\nfrom lunavl.sdk.estimators.face_estimators.fisheye import Fisheye\nfrom lunavl.sdk.faceengine.setting_provider import DetectorType\nfrom lunavl.sdk.image_utils.image import VLImage\nfrom tests.base import BaseTestClass\nfrom tests.resources import FISHEYE, FROWNING\n\n\ndef warpedSubTests(test):\n @wraps(test)\n def wrappedFunc(*func_args, **func_kwargs):\n for warped in [True, False]:\n with func_args[0].subTest(warped=warped):\n test(*func_args, **func_kwargs, warped=warped)\n\n return wrappedFunc\n\n\nclass TestFisheyeEffect(BaseTestClass):\n \"\"\"\n Test fisheye estimation.\n \"\"\"\n\n @classmethod\n def setup_class(cls):\n super().setup_class()\n cls.detector = cls.faceEngine.createFaceDetector(DetectorType.FACE_DET_V3)\n cls.warper = cls.faceEngine.createFaceWarper()\n cls.fisheyeEstimator = cls.faceEngine.createFisheyeEstimator()\n\n def test_estimate_fisheye_correctness(self):\n \"\"\"\n Test fisheye estimator correctness\n \"\"\"\n estimation = self.estimate(FISHEYE)\n assert estimation.status\n assert 0 <= estimation.score <= 1\n\n def estimate(self, image: str = FROWNING, warped: bool = True) -> Fisheye:\n \"\"\"Estimate fisheye on image\"\"\"\n faceDetection = self.detector.detectOne(VLImage.load(filename=image))\n if warped:\n warp = self.warper.warp(faceDetection)\n estimation = self.fisheyeEstimator.estimate(warp.warpedImage)\n else:\n estimation = self.fisheyeEstimator.estimate(faceDetection)\n assert isinstance(estimation, Fisheye)\n return estimation\n\n def estimateBatch(self, images, warped) -> list[Fisheye]:\n \"\"\"Estimate fisheye on image\"\"\"\n imageDetections = self.detector.detect([VLImage.load(filename=name) for name in images])\n if warped:\n warps = [self.warper.warp(res[0]) for res in imageDetections]\n estimations = self.fisheyeEstimator.estimateBatch(warps)\n else:\n estimations = self.fisheyeEstimator.estimateBatch(list(chain(*imageDetections)))\n assert all(isinstance(estimation, Fisheye) for estimation in estimations)\n return estimations\n\n @warpedSubTests\n def test_estimate_fisheye(self, warped):\n \"\"\"\n Simple fisheye estimation\n \"\"\"\n estimation = self.estimate(FROWNING, warped=warped)\n assert not estimation.status\n assert 0 <= estimation.score <= 1\n\n @warpedSubTests\n def test_fisheye_as_dict(self, warped):\n \"\"\"\n Test method Fisheye.asDict\n \"\"\"\n estimation = self.estimate(FROWNING, warped=warped)\n assert {\n \"status\": estimation.status,\n \"score\": estimation.score,\n } == estimation.asDict()\n\n @warpedSubTests\n def test_estimate_fisheye_batch(self, warped):\n \"\"\"\n Batch fisheye estimation test\n \"\"\"\n estimations = self.estimateBatch([FROWNING, FISHEYE], warped=warped)\n assert not estimations[0].status\n assert estimations[1].status\n", "sub_path": "tests/test_fisheye.py", "file_name": "test_fisheye.py", "file_ext": "py", "file_size_in_byte": 3140, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "functools.wraps", "line_number": 12, "usage_type": "call"}, {"api_name": "tests.base.BaseTestClass", "line_number": 21, "usage_type": "name"}, {"api_name": "lunavl.sdk.faceengine.setting_provider.DetectorType.FACE_DET_V3", "line_number": 29, "usage_type": "attribute"}, {"api_name": "lunavl.sdk.faceengine.setting_provider.DetectorType", "line_number": 29, "usage_type": "name"}, {"api_name": "tests.resources.FISHEYE", "line_number": 37, "usage_type": "argument"}, {"api_name": "tests.resources.FROWNING", "line_number": 41, "usage_type": "name"}, {"api_name": "lunavl.sdk.image_utils.image.VLImage.load", "line_number": 43, "usage_type": "call"}, {"api_name": "lunavl.sdk.image_utils.image.VLImage", "line_number": 43, "usage_type": "name"}, {"api_name": "lunavl.sdk.estimators.face_estimators.fisheye.Fisheye", "line_number": 49, "usage_type": "argument"}, {"api_name": "lunavl.sdk.estimators.face_estimators.fisheye.Fisheye", "line_number": 41, "usage_type": "name"}, {"api_name": "lunavl.sdk.image_utils.image.VLImage.load", "line_number": 54, "usage_type": "call"}, {"api_name": "lunavl.sdk.image_utils.image.VLImage", "line_number": 54, "usage_type": "name"}, {"api_name": "itertools.chain", "line_number": 59, "usage_type": "call"}, {"api_name": "lunavl.sdk.estimators.face_estimators.fisheye.Fisheye", "line_number": 60, "usage_type": "argument"}, {"api_name": "lunavl.sdk.estimators.face_estimators.fisheye.Fisheye", "line_number": 52, "usage_type": "name"}, {"api_name": "tests.resources.FROWNING", "line_number": 68, "usage_type": "argument"}, {"api_name": "tests.resources.FROWNING", "line_number": 77, "usage_type": "argument"}, {"api_name": "tests.resources.FROWNING", "line_number": 88, "usage_type": "name"}, {"api_name": "tests.resources.FISHEYE", "line_number": 88, "usage_type": "name"}]} +{"seq_id": "496799772", "text": "from datetime import datetime\n\nimport colander\nimport pytz\n\nfrom ichnaea.api.exceptions import ParseError\nfrom ichnaea.api.submit.schema_v0 import SUBMIT_V0_SCHEMA\nfrom ichnaea.api.submit.tests.base import BaseSubmitTest\nfrom ichnaea.models import Radio\nfrom ichnaea.tests.base import (\n CeleryAppTestCase,\n TestCase,\n)\nfrom ichnaea.tests.factories import (\n CellShardFactory,\n WifiShardFactory,\n)\nfrom ichnaea import util\n\n\nclass TestSubmitSchema(TestCase):\n\n schema = SUBMIT_V0_SCHEMA\n\n def test_empty(self):\n with self.assertRaises(colander.Invalid):\n self.schema.deserialize({})\n\n def test_empty_wifi_entry(self):\n wifi = WifiShardFactory.build()\n data = self.schema.deserialize({'items': [\n {'lat': wifi.lat, 'lon': wifi.lon, 'wifi': [{}]},\n ]})\n self.assertEqual(data, {'items': []})\n\n def test_minimal(self):\n wifi = WifiShardFactory.build()\n data = self.schema.deserialize(\n {'items': [{'lat': wifi.lat, 'lon': wifi.lon,\n 'wifi': [{'key': 'ab'}]}]})\n self.assertTrue('items' in data)\n self.assertEqual(len(data['items']), 1)\n\n\nclass TestView(BaseSubmitTest, CeleryAppTestCase):\n\n url = '/v1/submit'\n metric_path = 'path:v1.submit'\n status = 204\n radio_id = 'radio'\n cells_id = 'cell'\n\n def _one_cell_query(self, radio=True):\n cell = CellShardFactory.build()\n query = {'lat': cell.lat, 'lon': cell.lon,\n 'cell': [{'mcc': cell.mcc, 'mnc': cell.mnc,\n 'lac': cell.lac, 'cid': cell.cid}]}\n if radio:\n query['cell'][0]['radio'] = cell.radio.name\n return (cell, query)\n\n def test_cell(self):\n now = util.utcnow()\n today = now.replace(hour=0, minute=0, second=0)\n cell = CellShardFactory.build(radio=Radio.umts)\n res = self._post([{\n 'lat': cell.lat,\n 'lon': cell.lon,\n 'time': today.strftime('%Y-%m-%d'),\n 'accuracy': 10.6,\n 'altitude': 123.1,\n 'altitude_accuracy': 7.0,\n 'radio': cell.radio.name,\n 'cell': [{\n 'radio': 'umts', 'mcc': cell.mcc,\n 'mnc': cell.mnc, 'lac': cell.lac, 'cid': cell.cid}],\n }], api_key='test')\n self.assertEqual(res.body, b'')\n\n self._assert_queue_size(1)\n item = self.queue.dequeue(self.queue.queue_key())[0]\n self.assertEqual(item['api_key'], 'test')\n report = item['report']\n timestamp = datetime.utcfromtimestamp(report['timestamp'] / 1000.0)\n timestamp = timestamp.replace(microsecond=0, tzinfo=pytz.UTC)\n self.assertEqual(timestamp, today)\n position = report['position']\n self.assertEqual(position['latitude'], cell.lat)\n self.assertEqual(position['longitude'], cell.lon)\n self.assertEqual(position['accuracy'], 10.6)\n self.assertEqual(position['altitude'], 123.1)\n self.assertEqual(position['altitudeAccuracy'], 7.0)\n cells = report['cellTowers']\n self.assertEqual(cells[0]['radioType'], 'wcdma')\n self.assertEqual(cells[0]['mobileCountryCode'], cell.mcc)\n self.assertEqual(cells[0]['mobileNetworkCode'], cell.mnc)\n self.assertEqual(cells[0]['locationAreaCode'], cell.lac)\n self.assertEqual(cells[0]['cellId'], cell.cid)\n\n def test_wifi(self):\n wifi = WifiShardFactory.build()\n self._post([{\n 'lat': wifi.lat,\n 'lon': wifi.lon,\n 'accuracy': 17.1,\n 'wifi': [{'key': wifi.mac.upper(),\n 'frequency': 2437,\n 'signal': -70,\n 'signalToNoiseRatio': 5,\n 'ssid': 'my-wifi',\n }]\n }])\n\n self._assert_queue_size(1)\n item = self.queue.dequeue(self.queue.queue_key())[0]\n self.assertEqual(item['api_key'], None)\n report = item['report']\n position = report['position']\n self.assertEqual(position['latitude'], wifi.lat)\n self.assertEqual(position['longitude'], wifi.lon)\n self.assertEqual(position['accuracy'], 17.1)\n self.assertFalse('altitude' in position)\n self.assertFalse('altitudeAccuracy' in position)\n wifis = report['wifiAccessPoints']\n self.assertEqual(wifis[0]['macAddress'], wifi.mac.upper())\n self.assertFalse('channel' in wifis[0])\n self.assertEqual(wifis[0]['frequency'], 2437)\n self.assertEqual(wifis[0]['signalStrength'], -70)\n self.assertEqual(wifis[0]['signalToNoiseRatio'], 5)\n self.assertEqual(wifis[0]['ssid'], 'my-wifi')\n\n def test_batches(self):\n batch = 110\n wifis = WifiShardFactory.build_batch(batch)\n items = [{'lat': wifi.lat,\n 'lon': wifi.lon,\n 'wifi': [{'key': wifi.mac}]}\n for wifi in wifis]\n\n # add a bad one, this will just be skipped\n items.append({'lat': 10.0, 'lon': 10.0, 'whatever': 'xx'})\n self._post(items)\n self._assert_queue_size(batch)\n\n def test_error(self):\n wifi = WifiShardFactory.build()\n res = self.app.post_json(\n '/v1/submit',\n [{'lat': wifi.lat, 'lon': wifi.lon, 'cell': []}],\n status=400)\n self.assertEqual(res.json, ParseError.json_body())\n self.check_raven(['ParseError'])\n\n def test_error_missing_latlon(self):\n wifi = WifiShardFactory.build()\n self._post([\n {'lat': wifi.lat,\n 'lon': wifi.lon,\n 'accuracy': 17.0,\n 'wifi': [{'key': wifi.mac}],\n },\n {'wifi': [{'key': wifi.mac}],\n 'accuracy': 16.0},\n ])\n self._assert_queue_size(2)\n", "sub_path": "ichnaea/api/submit/tests/test_submit_v0.py", "file_name": "test_submit_v0.py", "file_ext": "py", "file_size_in_byte": 5815, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "ichnaea.tests.base.TestCase", "line_number": 21, "usage_type": "name"}, {"api_name": "ichnaea.api.submit.schema_v0.SUBMIT_V0_SCHEMA", "line_number": 23, "usage_type": "name"}, {"api_name": "colander.Invalid", "line_number": 26, "usage_type": "attribute"}, {"api_name": "ichnaea.tests.factories.WifiShardFactory.build", "line_number": 30, "usage_type": "call"}, {"api_name": "ichnaea.tests.factories.WifiShardFactory", "line_number": 30, "usage_type": "name"}, {"api_name": "ichnaea.tests.factories.WifiShardFactory.build", "line_number": 37, "usage_type": "call"}, {"api_name": "ichnaea.tests.factories.WifiShardFactory", "line_number": 37, "usage_type": "name"}, {"api_name": "ichnaea.api.submit.tests.base.BaseSubmitTest", "line_number": 45, "usage_type": "name"}, {"api_name": "ichnaea.tests.base.CeleryAppTestCase", "line_number": 45, "usage_type": "name"}, {"api_name": "ichnaea.tests.factories.CellShardFactory.build", "line_number": 54, "usage_type": "call"}, {"api_name": "ichnaea.tests.factories.CellShardFactory", "line_number": 54, "usage_type": "name"}, {"api_name": "ichnaea.util.utcnow", "line_number": 63, "usage_type": "call"}, {"api_name": "ichnaea.util", "line_number": 63, "usage_type": "name"}, {"api_name": "ichnaea.tests.factories.CellShardFactory.build", "line_number": 65, "usage_type": "call"}, {"api_name": "ichnaea.tests.factories.CellShardFactory", "line_number": 65, "usage_type": "name"}, {"api_name": "ichnaea.models.Radio.umts", "line_number": 65, "usage_type": "attribute"}, {"api_name": "ichnaea.models.Radio", "line_number": 65, "usage_type": "name"}, {"api_name": "datetime.datetime.utcfromtimestamp", "line_number": 84, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 84, "usage_type": "name"}, {"api_name": "pytz.UTC", "line_number": 85, "usage_type": "attribute"}, {"api_name": "ichnaea.tests.factories.WifiShardFactory.build", "line_number": 101, "usage_type": "call"}, {"api_name": "ichnaea.tests.factories.WifiShardFactory", "line_number": 101, "usage_type": "name"}, {"api_name": "ichnaea.tests.factories.WifiShardFactory.build_batch", "line_number": 134, "usage_type": "call"}, {"api_name": "ichnaea.tests.factories.WifiShardFactory", "line_number": 134, "usage_type": "name"}, {"api_name": "ichnaea.tests.factories.WifiShardFactory.build", "line_number": 146, "usage_type": "call"}, {"api_name": "ichnaea.tests.factories.WifiShardFactory", "line_number": 146, "usage_type": "name"}, {"api_name": "ichnaea.api.exceptions.ParseError.json_body", "line_number": 151, "usage_type": "call"}, {"api_name": "ichnaea.api.exceptions.ParseError", "line_number": 151, "usage_type": "name"}, {"api_name": "ichnaea.tests.factories.WifiShardFactory.build", "line_number": 155, "usage_type": "call"}, {"api_name": "ichnaea.tests.factories.WifiShardFactory", "line_number": 155, "usage_type": "name"}]} +{"seq_id": "104333630", "text": "import numpy as np\nimport heapq\nimport itertools\nfrom scipy.sparse import csr_matrix\nfrom .dataiterator import DataIterator\nfrom backend import batch_random_choice\nfrom utils import typeassert\nfrom utils import timer\n\n\ndef random_choice(a, size=None, replace=True, p=None, exclusion=None):\n if exclusion is not None:\n if p is None:\n p = np.ones_like(a)\n else:\n p = np.array(p, copy=True)\n p = np.ndarray.flatten(p)\n p[exclusion] = 0\n p = p / np.sum(p)\n sample = np.random.choice(a, size=size, replace=replace, p=p)\n return sample\n\n\n@typeassert(matrix=csr_matrix)\ndef csr_to_user_dict(matrix):\n \"\"\"convert a scipy.sparse.csr_matrix to a dict,\n where the key is row number, and value is the\n non-empty index in each row.\n \"\"\"\n idx_value_dict = {}\n for idx, value in enumerate(matrix):\n if any(value.indices):\n idx_value_dict[idx] = value.indices.copy()\n return idx_value_dict\n\n\n# # TODO rewrite with cython or cpp\n# # @timer\n# @typeassert(matrix=csr_matrix, neg_num=int, fold_neg=bool)\n# def csr_to_pairwise(matrix, neg_num=1, fold_neg=False):\n# user_num, item_num = matrix.shape\n# all_items = np.arange(item_num)\n# user_list, pos_list, neg_list = [], [], []\n# for idx, value in enumerate(matrix):\n# if any(value.indices):\n# pos_items = value.indices.copy()\n# pos_len = len(pos_items)\n# if neg_num > 0: # sample negative items\n# neg_len = neg_num*pos_len\n# neg_item = random_choice(all_items, size=neg_len, exclusion=pos_items)\n# if not fold_neg: # unfold negative items\n# user_list.extend([idx]*neg_len)\n# pos_items = np.reshape(np.tile(pos_items, neg_num), newshape=[-1])\n# pos_list.extend(pos_items)\n# neg_list.extend(np.reshape(neg_item, newshape=[-1]))\n# else: # fold negative items\n# user_list.extend([idx]*pos_len)\n# pos_list.extend(pos_items)\n# neg_item = np.reshape(neg_item, newshape=[pos_len, neg_num]).tolist()\n# neg_list.extend(neg_item)\n# else: # do not sample negative items\n# user_list.extend([idx] * pos_len)\n# pos_list.extend(pos_items)\n#\n# return user_list, pos_list, neg_list\n\n\n@typeassert(matrix=csr_matrix, neg_num=int, fold_neg=bool)\ndef csr_to_pairwise(matrix, neg_num=1, fold_neg=False):\n all_user_list, all_pos_list, all_neg_list = [], [], []\n user_num, item_num = matrix.shape\n all_users = list(range(user_num))\n all_items = np.arange(item_num)\n all_users = DataIterator(all_users, batch_size=1024, shuffle=False, drop_last=False)\n if neg_num > 0: # sample negative items\n for bat_users in all_users:\n pos_items_list = [matrix[user].indices for user in bat_users]\n samples_size = [len(pos)*neg_num for pos in pos_items_list]\n neg_items_list = batch_random_choice(all_items, samples_size, replace=True, exclusion=pos_items_list)\n for idx, user in enumerate(bat_users):\n if not fold_neg: # unfold negative items\n all_user_list.extend([user]*samples_size[idx])\n pos_items = np.tile(pos_items_list[idx], neg_num).reshape([-1])\n all_pos_list.extend(pos_items)\n all_neg_list.extend(neg_items_list[idx].reshape([-1]))\n else: # fold negative items\n pos_len = len(pos_items_list[idx])\n all_user_list.extend([user]*pos_len)\n all_pos_list.extend(pos_items_list[idx].reshape([-1]))\n all_neg_list.extend(neg_items_list[idx].reshape([-1, neg_num]))\n else: # do not sample negative items\n for idx, value in enumerate(matrix):\n pos_items = value.indices\n pos_len = len(pos_items)\n all_user_list.extend([idx] * pos_len)\n all_pos_list.extend(pos_items)\n\n return all_user_list, all_pos_list, all_neg_list\n\n\n# TODO check\n@typeassert(matrix=csr_matrix)\ndef csr_to_pointwise(matrix, neg_num=1):\n user_num, item_num = matrix.shape\n all_items = np.arange(item_num)\n user_list, item_list, label_list = [], [], []\n for idx, value in enumerate(matrix):\n if any(value.indices):\n pos_items = value.indices.copy()\n pos_len = len(pos_items)\n if neg_num == 0:\n user_list.extend([idx]*pos_len)\n item_list.extend(pos_items.tolist())\n label_list.extend(len(pos_items)*[1.0])\n elif neg_num > 0:\n neg_len = neg_num*pos_len\n neg_items = random_choice(all_items, size=neg_len, exclusion=pos_items).tolist()\n\n user_list.extend([idx]*(neg_len+pos_len))\n item_list.extend(pos_items.tolist())\n label_list.extend(pos_len*[1.0])\n item_list.extend(neg_items)\n label_list.extend(neg_len*[0.0])\n else:\n raise ValueError(\"The parameter 'neg_num' is invalid!\")\n\n return user_list, item_list, label_list\n\n\ndef padding(input, value=0, max_len=None):\n if max_len is None:\n max_len = max([len(x) for x in input])\n output = np.full([len(input), max_len], value, dtype=int)\n for idx, x in enumerate(input):\n copy_len = max_len if len(x) > max_len else len(x)\n output[idx][:copy_len] = x[:copy_len]\n return output.tolist()\n\n\ndef argmax_top_k(a, top_k=50):\n ele_idx = heapq.nlargest(top_k, zip(a, itertools.count()))\n return np.array([idx for ele, idx in ele_idx], dtype=np.intc)\n\n", "sub_path": "utils/tools.py", "file_name": "tools.py", "file_ext": "py", "file_size_in_byte": 5782, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "numpy.ones_like", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.ndarray.flatten", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 17, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 20, "usage_type": "attribute"}, {"api_name": "utils.typeassert", "line_number": 24, "usage_type": "call"}, {"api_name": "scipy.sparse.csr_matrix", "line_number": 24, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 73, "usage_type": "call"}, {"api_name": "dataiterator.DataIterator", "line_number": 74, "usage_type": "call"}, {"api_name": "backend.batch_random_choice", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 83, "usage_type": "call"}, {"api_name": "utils.typeassert", "line_number": 68, "usage_type": "call"}, {"api_name": "scipy.sparse.csr_matrix", "line_number": 68, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 105, "usage_type": "call"}, {"api_name": "utils.typeassert", "line_number": 102, "usage_type": "call"}, {"api_name": "scipy.sparse.csr_matrix", "line_number": 102, "usage_type": "name"}, {"api_name": "numpy.full", "line_number": 133, "usage_type": "call"}, {"api_name": "heapq.nlargest", "line_number": 141, "usage_type": "call"}, {"api_name": "itertools.count", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.intc", "line_number": 142, "usage_type": "attribute"}]} +{"seq_id": "164246450", "text": "import json\nimport yaml\n\nfrom flask import request\nfrom flask_restplus import Namespace, Resource\n\nimport helpers.transformations as enhanced\n\nfrom db import dsl\nfrom providers import *\nfrom app.models import Payload\n\n\napi = Namespace('actions', description='Actions Endpoints')\n\n\nclass Actions(Resource):\n\n def post(self, action, uuid):\n \"\"\"Post Actions\"\"\"\n\n payload = Payload.where('uuid', uuid).first()\n\n if payload:\n content = json.loads(payload.payload)\n\n params = enhanced.map(\n dsl.file.integrations.get('actions').get(action),\n lambda t: enhanced.translate(\n t,\n payload=content.get('payload', {}),\n changes=request.get_json()\n )\n )\n\n provider = params.get('provider')\n response = globals()[provider.capitalize()](**params).run()\n\n return response\n\n\n return {}\n\n\napi.add_resource(\n Actions,\n '//',\n methods=['POST'])\n", "sub_path": "app/controllers/actions.py", "file_name": "actions.py", "file_ext": "py", "file_size_in_byte": 1061, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "flask_restplus.Namespace", "line_number": 14, "usage_type": "call"}, {"api_name": "flask_restplus.Resource", "line_number": 17, "usage_type": "name"}, {"api_name": "app.models.Payload.where", "line_number": 22, "usage_type": "call"}, {"api_name": "app.models.Payload", "line_number": 22, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 25, "usage_type": "call"}, {"api_name": "helpers.transformations.map", "line_number": 27, "usage_type": "call"}, {"api_name": "helpers.transformations", "line_number": 27, "usage_type": "name"}, {"api_name": "db.dsl.file.integrations.get", "line_number": 28, "usage_type": "call"}, {"api_name": "db.dsl.file", "line_number": 28, "usage_type": "attribute"}, {"api_name": "db.dsl", "line_number": 28, "usage_type": "name"}, {"api_name": "helpers.transformations.translate", "line_number": 29, "usage_type": "call"}, {"api_name": "helpers.transformations", "line_number": 29, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 32, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 32, "usage_type": "name"}]} +{"seq_id": "518835647", "text": "import sys\nimport json\nimport copy\nimport argparse\nfrom datetime import datetime\nfrom helper import time_string\nfrom helper import daterange\nfrom helper import date_string\nfrom helper import eprint\nfrom test_completion import get_test_completion as get_test_scores\nfrom start_end import commit_data\n\n\ndef jsonify(test_data):\n \"\"\"Formats data for /classProgress endpoint\n\n Converts information in **data** into an appropriately formatted json \n for the /classProgress endpoint \n\n **Args**:\n **data** (dict): A dictionary of the following format: ::\n\n {\n \"name1\": {\n \"Test1\": (\"P\" or \"F\"),\n ...\n },\n ...\n }\n\n **Return**:\n dict: A dictionary of histogram data split into 20% bins: ::\n\n {\n \"0-20%\": int,\n \"20-40%\": int,\n \"40-60%\": int,\n \"60-80%\": int,\n \"80-100%\": int\n }\n\n where each percentage bin contains a value 0 <= int <= 100\n\n \"\"\"\n histogram_data = {\"0-20%\": 0, \"20-40%\": 0, \"40-60%\": 0, \"60-80%\": 0, \"80-100%\": 0}\n for student in test_data:\n info = test_data[student]\n if info[\"total\"] <= 20:\n histogram_data[\"0-20%\"] += 1\n elif info[\"total\"] <= 40:\n histogram_data[\"20-40%\"] += 1\n elif info[\"total\"] <= 60:\n histogram_data[\"40-60%\"] += 1\n elif info[\"total\"] <= 80:\n histogram_data[\"60-80%\"] += 1\n elif info[\"total\"] <= 100:\n histogram_data[\"80-100%\"] += 1\n return json.dumps(histogram_data)\n\n\ndef merge_data(visible, hidden):\n \"\"\"Sums the values in **visible** and **hidden** for each bin\"\"\"\n visible = json.loads(visible)\n hidden = json.loads(hidden)\n for key in visible:\n visible[key] += hidden[key]\n return json.dumps(visible)\n\n\nif __name__ == \"__main__\":\n parser = argparse.ArgumentParser()\n parser.add_argument(\"visible\", help=\"path to visible test score file\")\n parser.add_argument(\"hidden\", help=\"path to hidden test score file\")\n\n args = parser.parse_args()\n\n visible_test_score_file = open(args.visible, \"r\")\n hidden_test_score_file = open(args.hidden, \"r\")\n\n visible_data = get_test_scores(visible_test_score_file)\n hidden_data = get_test_scores(hidden_test_score_file)\n # print(visible_data)\n\n formatted_visible = jsonify(visible_data)\n formatted_hidden = jsonify(hidden_data)\n api_json = merge_data(formatted_visible, formatted_hidden)\n print(api_json)\n", "sub_path": "python/get_class_progress.py", "file_name": "get_class_progress.py", "file_ext": "py", "file_size_in_byte": 2579, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "json.dumps", "line_number": 58, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 63, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 64, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 67, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 71, "usage_type": "call"}, {"api_name": "test_completion.get_test_completion", "line_number": 80, "usage_type": "call"}, {"api_name": "test_completion.get_test_completion", "line_number": 81, "usage_type": "call"}]} +{"seq_id": "544539502", "text": "import logging\nfrom datetime import datetime\n\nfrom telegram.ext import Updater, CommandHandler, CallbackQueryHandler\nfrom telegram import InlineKeyboardButton, InlineKeyboardMarkup\nfrom app.config import API_KEY, PROXY\nfrom app import rzd\nfrom app.model import Subscribes\n\nlogging.basicConfig(format='%(asctime)s - %(levelname)s - %(message)s',\n level=logging.INFO,\n filename='bot.log'\n )\nCHATS_TRAINS = {}\n\n\ndef get_train_button(train):\n return '(%s) %s %s' % (\n train.number, train.departure_time.strftime(\"%H:%M %d.%m\"), train.arrival_time.strftime(\"%H:%M %d.%m\"))\n\n\ndef get_train_info(train):\n result = '(%s) %s \\nОтправление: %s\\nПрибытие: %s\\n\\n' % (\n train.number, train.title, train.departure_time.strftime(\"%H:%M %d.%m.%Y\"),\n train.arrival_time.strftime(\"%H:%M %d.%m.%Y\"))\n return result + '\\n'.join([str(seat) for seat in dict(train.seats).values()])\n\n\ndef get_train(route_from, route_to, route_date):\n with rzd.RzdFetcher() as fetcher:\n train_list = fetcher.trains(\n route_from.upper(),\n route_to.upper(),\n rzd.TimeRange(\n datetime(route_date.year, route_date.month,\n route_date.day,\n 0, 0),\n datetime(route_date.year, route_date.month,\n route_date.day, 23, 59),\n )\n )\n return train_list\n\n\ndef build_menu(buttons,\n n_cols,\n header_buttons=None,\n footer_buttons=None):\n menu = [buttons[i:i + n_cols] for i in range(0, len(buttons), n_cols)]\n if header_buttons:\n menu.insert(0, header_buttons)\n if footer_buttons:\n menu.append(footer_buttons)\n return menu\n\n\ndef get_route(bot, update):\n text = update.message.text.split()\n logging.info(text)\n if len(text) == 1:\n update.message.reply_text('Введите значение после команды')\n rout_from = text[1]\n route_to = text[2]\n route_date = datetime.strptime(text[3], \"%d-%m-%y\")\n trains = get_train(rout_from, route_to, route_date)\n CHATS_TRAINS[update.message.chat.id] = {train.number: train for train in trains}\n button_list = [InlineKeyboardButton(get_train_button(train), callback_data=train.number)\n for train in trains\n ]\n reply_markup = InlineKeyboardMarkup(build_menu(button_list, n_cols=1))\n bot.send_message(update.message.chat.id, \"Список поездов\", reply_markup=reply_markup)\n # except Exception as e:\n # logging.info(e)\n # update.message.reply_text('не правильно введена команда')\n\n\ndef get_route2(bot, update):\n text = update.message.text.split()\n logging.info(text)\n try:\n if len(text) == 1:\n update.message.reply_text('Введите значение после команды')\n rout_from = text[1]\n route_to = text[2]\n route_date = datetime.strptime(text[3], \"%d-%m-%y\")\n trains = get_train(rout_from, route_to, route_date)\n CHATS_TRAINS[update.message.chat.id] = {train.number: train for train in trains}\n message = '\\n'.join([str(train) for train in trains])\n button_list = [InlineKeyboardButton(train.number, callback_data=train.number)\n for train in trains\n ]\n reply_markup = InlineKeyboardMarkup(build_menu(button_list, n_cols=1))\n bot.send_message(update.message.chat.id, \"Список поездов: \\n{}\".format(message), reply_markup=reply_markup)\n except Exception as e:\n logging.info(e)\n update.message.reply_text(\n 'не правильно введена команда\\nвведите команду в формате\\nотправление прибытие дата\\nдата в формате дд-мм-гг')\n\n\ndef callbackHandler(bot, call):\n logging.info(call)\n logging.info(CHATS_TRAINS)\n if call.callback_query:\n if call.callback_query.data:\n if CHATS_TRAINS[call.callback_query.message.chat.id] and call.callback_query.data in CHATS_TRAINS[\n call.callback_query.message.chat.id].keys():\n bot.send_message(chat_id=call.callback_query.message.chat.id, text=get_train_info(\n CHATS_TRAINS[call.callback_query.message.chat.id][call.callback_query.data]))\n\n # bot.edit_message_text(chat_id=call.callback_query.message.chat.id,\n # message_id=call.callback_query.message.message_id, text=call.callback_query.data)\n\n\ndef subscribe(bot, update):\n text = update.message.text.split()\n logging.info(text)\n if len(text) == 1:\n update.message.reply_text('Введите значение после команды')\n route_from = text[1]\n route_to = text[2]\n route_date = datetime.strptime(text[3], \"%d-%m-%y\")\n _subscribe = Subscribes(chat_id=update.message.chat.id, route_from=route_from, route_to=route_to, route_date=route_date)\n\n\ndef main():\n mybot = Updater(API_KEY, request_kwargs=PROXY)\n\n logging.info('Бот запускается')\n dp = mybot.dispatcher\n dp.add_handler(CommandHandler('route', get_route))\n dp.add_handler(CommandHandler('subscribe', subscribe))\n dp.add_handler(CallbackQueryHandler(callbackHandler))\n\n mybot.start_polling()\n mybot.idle()\n\n\nif __name__ == '__main__':\n main()\n", "sub_path": "app/bot.py", "file_name": "bot.py", "file_ext": "py", "file_size_in_byte": 5510, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "logging.basicConfig", "line_number": 10, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 11, "usage_type": "attribute"}, {"api_name": "app.rzd.RzdFetcher", "line_number": 30, "usage_type": "call"}, {"api_name": "app.rzd", "line_number": 30, "usage_type": "name"}, {"api_name": "app.rzd.TimeRange", "line_number": 34, "usage_type": "call"}, {"api_name": "app.rzd", "line_number": 34, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 35, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 38, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 59, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 64, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 64, "usage_type": "name"}, {"api_name": "telegram.InlineKeyboardButton", "line_number": 67, "usage_type": "call"}, {"api_name": "telegram.InlineKeyboardMarkup", "line_number": 70, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 79, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 85, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 85, "usage_type": "name"}, {"api_name": "telegram.InlineKeyboardButton", "line_number": 89, "usage_type": "call"}, {"api_name": "telegram.InlineKeyboardMarkup", "line_number": 92, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 95, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 101, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 102, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 116, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 121, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 121, "usage_type": "name"}, {"api_name": "app.model.Subscribes", "line_number": 122, "usage_type": "call"}, {"api_name": "telegram.ext.Updater", "line_number": 126, "usage_type": "call"}, {"api_name": "app.config.API_KEY", "line_number": 126, "usage_type": "argument"}, {"api_name": "app.config.PROXY", "line_number": 126, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 128, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 130, "usage_type": "call"}, {"api_name": "telegram.ext.CommandHandler", "line_number": 131, "usage_type": "call"}, {"api_name": "telegram.ext.CallbackQueryHandler", "line_number": 132, "usage_type": "call"}]} +{"seq_id": "99193920", "text": "from PySide2 import QtCore\nfrom PySide2 import QtGui\nfrom PySide2 import QtWidgets\nfrom shiboken2 import wrapInstance\n\nimport maya.cmds as cmds\nimport maya.mel as mel\nimport maya.OpenMayaUI as omui\n\n\ndef maya_main_window():\n \"\"\"\n Return the Maya main window widget as a Python object\n \"\"\"\n main_window_ptr = omui.MQtUtil.mainWindow()\n return wrapInstance(long(main_window_ptr), QtWidgets.QWidget)\n\n\nclass TimelineOverlay(QtWidgets.QWidget):\n\n KEYFRAME_COLOR = QtGui.QColor(QtCore.Qt.green)\n\n def __init__(self):\n self.time_control = mel.eval(\"$tempVar = $gPlayBackSlider\")\n time_control_ptr = omui.MQtUtil.findControl(self.time_control)\n time_control_widget = wrapInstance(long(time_control_ptr), QtWidgets.QWidget)\n\n super(TimelineOverlay, self).__init__(time_control_widget)\n\n self.frame_times = [1, 6, 8, 10, 19, 30, 39, 50, 51, 52, 53, 54, 120]\n\n self.set_context_menu_enabled(False)\n\n def add_frame(self):\n current_time = cmds.currentTime(q=True)\n if current_time not in self.frame_times:\n self.frame_times.append(current_time)\n self.update()\n\n def add_frames(self):\n print(\"TODO: Add Frames\")\n\n def remove_frame(self):\n current_time = cmds.currentTime(q=True)\n if current_time in self.frame_times:\n self.frame_times.remove(current_time)\n self.update()\n\n def remove_frames(self):\n print(\"TODO: Remove Frames\")\n\n def set_context_menu_enabled(self, enabled):\n self.context_menu_enabled = enabled\n\n if enabled:\n print(\"TODO: Add Context Menu\")\n\n def paintEvent(self, paint_event):\n parent = self.parentWidget()\n if parent:\n self.setGeometry(parent.geometry())\n\n range_start = cmds.playbackOptions(q=True, minTime=True)\n range_end = cmds.playbackOptions(q=True, maxTime=True)\n displayed_frame_count = range_end - range_start + 1\n\n padding = self.width() * 0.005\n frame_width = (self.width() * 0.99) / displayed_frame_count\n\n frame_height = 0.333 * self.height()\n frame_y = self.height() - frame_height\n\n painter = QtGui.QPainter(self)\n\n pen = painter.pen()\n pen.setWidth(1)\n pen.setColor(TimelineOverlay.KEYFRAME_COLOR)\n painter.setPen(pen)\n\n fill_color = QtGui.QColor(TimelineOverlay.KEYFRAME_COLOR)\n fill_color.setAlpha(63)\n\n for frame_time in self.frame_times:\n frame_x = padding + ((frame_time - range_start) * frame_width) + 0.5\n\n painter.fillRect(frame_x, frame_y, frame_width, frame_height, fill_color)\n painter.drawRect(frame_x, frame_y, frame_width, frame_height)\n\n\n\nif __name__ == \"__main__\":\n try:\n TimelineOverlayDialog.delete_overlays() # pylint: disable=E0601\n except:\n pass\n\n\nclass TimelineOverlayDialog(QtWidgets.QDialog):\n\n WINDOW_TITLE = \"Timeline Overlay\"\n\n timeline_overlay = None\n\n @classmethod\n def delete_overlays(cls):\n if TimelineOverlayDialog.timeline_overlay:\n TimelineOverlayDialog.timeline_overlay.setParent(None)\n TimelineOverlayDialog.timeline_overlay.deleteLater()\n TimelineOverlayDialog.timeline_overlay = None\n\n def __init__(self, parent=maya_main_window()):\n super(TimelineOverlayDialog, self).__init__(parent)\n\n self.setWindowTitle(self.WINDOW_TITLE)\n if cmds.about(ntOS=True):\n self.setWindowFlags(self.windowFlags() ^ QtCore.Qt.WindowContextHelpButtonHint)\n elif cmds.about(macOS=True):\n self.setWindowFlags(QtCore.Qt.Tool)\n\n self.setMinimumSize(280, 160)\n\n self.create_widgets()\n self.create_layout()\n self.create_connections()\n\n self.set_overlay_visible(True)\n\n def create_widgets(self):\n self.overlay_visible_cb = QtWidgets.QCheckBox(\"Show Overlay\")\n\n self.context_menu_cb = QtWidgets.QCheckBox(\"Context Menu Enabled\")\n self.context_menu_cb.setChecked(True)\n\n self.add_frame_btn = QtWidgets.QPushButton(\"Add Frame\")\n self.remove_frame_btn = QtWidgets.QPushButton(\"Remove Frame\")\n\n self.close_btn = QtWidgets.QPushButton(\"Close\")\n\n def create_layout(self):\n frame_layout = QtWidgets.QHBoxLayout()\n frame_layout.setSpacing(4)\n frame_layout.addWidget(self.add_frame_btn)\n frame_layout.addWidget(self.remove_frame_btn)\n frame_layout.addStretch()\n\n overlay_layout = QtWidgets.QVBoxLayout()\n overlay_layout.addWidget(self.overlay_visible_cb)\n overlay_layout.addWidget(self.context_menu_cb)\n overlay_layout.addLayout(frame_layout)\n\n options_grp = QtWidgets.QGroupBox(\"Overlay Options\")\n options_grp.setLayout(overlay_layout)\n\n btn_layout = QtWidgets.QHBoxLayout()\n btn_layout.addStretch()\n btn_layout.addWidget(self.close_btn)\n\n main_layout = QtWidgets.QVBoxLayout(self)\n main_layout.setContentsMargins(2, 2, 2, 2)\n main_layout.addWidget(options_grp)\n main_layout.addStretch()\n main_layout.addLayout(btn_layout)\n\n def create_connections(self):\n self.overlay_visible_cb.toggled.connect(self.set_overlay_visible)\n\n self.close_btn.clicked.connect(self.close)\n\n def set_overlay_visible(self, visible):\n if visible:\n if not TimelineOverlayDialog.timeline_overlay:\n TimelineOverlayDialog.timeline_overlay = TimelineOverlay()\n TimelineOverlayDialog.timeline_overlay.set_context_menu_enabled(self.context_menu_cb.isChecked())\n\n self.context_menu_cb.toggled.connect(TimelineOverlayDialog.timeline_overlay.set_context_menu_enabled)\n self.add_frame_btn.clicked.connect(TimelineOverlayDialog.timeline_overlay.add_frame)\n self.remove_frame_btn.clicked.connect(TimelineOverlayDialog.timeline_overlay.remove_frame)\n\n\n if TimelineOverlayDialog.timeline_overlay:\n TimelineOverlayDialog.timeline_overlay.setVisible(visible)\n\n self.overlay_visible_cb.setChecked(visible)\n\n\n\nif __name__ == \"__main__\":\n\n try:\n overlay_dialog.close() # pylint: disable=E0601\n overlay_dialog.deleteLater()\n except:\n pass\n\n overlay_dialog = TimelineOverlayDialog()\n overlay_dialog.show()\n", "sub_path": "CZ_Tutorials/010_PySide2 for Maya (Vol. 3)/22-pyside2_for_maya_vol_3-custom_maya_overlays_part_4/timeline_overlay_dialog_start.py", "file_name": "timeline_overlay_dialog_start.py", "file_ext": "py", "file_size_in_byte": 6422, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "maya.OpenMayaUI.MQtUtil.mainWindow", "line_number": 15, "usage_type": "call"}, {"api_name": "maya.OpenMayaUI.MQtUtil", "line_number": 15, "usage_type": "attribute"}, {"api_name": "maya.OpenMayaUI", "line_number": 15, "usage_type": "name"}, {"api_name": "shiboken2.wrapInstance", "line_number": 16, "usage_type": "call"}, {"api_name": "PySide2.QtWidgets.QWidget", "line_number": 16, "usage_type": "attribute"}, {"api_name": "PySide2.QtWidgets", "line_number": 16, "usage_type": "name"}, {"api_name": "PySide2.QtWidgets.QWidget", "line_number": 19, "usage_type": "attribute"}, {"api_name": "PySide2.QtWidgets", "line_number": 19, "usage_type": "name"}, {"api_name": "PySide2.QtGui.QColor", "line_number": 21, "usage_type": "call"}, {"api_name": "PySide2.QtGui", "line_number": 21, "usage_type": "name"}, {"api_name": "PySide2.QtCore.Qt", "line_number": 21, "usage_type": "attribute"}, {"api_name": "PySide2.QtCore", "line_number": 21, "usage_type": "name"}, {"api_name": "maya.mel.eval", "line_number": 24, "usage_type": "call"}, {"api_name": "maya.mel", "line_number": 24, "usage_type": "name"}, {"api_name": "maya.OpenMayaUI.MQtUtil.findControl", "line_number": 25, "usage_type": "call"}, {"api_name": "maya.OpenMayaUI.MQtUtil", "line_number": 25, "usage_type": "attribute"}, {"api_name": "maya.OpenMayaUI", "line_number": 25, "usage_type": "name"}, {"api_name": "shiboken2.wrapInstance", "line_number": 26, "usage_type": "call"}, {"api_name": "PySide2.QtWidgets.QWidget", "line_number": 26, "usage_type": "attribute"}, {"api_name": "PySide2.QtWidgets", "line_number": 26, "usage_type": "name"}, {"api_name": "maya.cmds.currentTime", "line_number": 35, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 35, "usage_type": "name"}, {"api_name": "maya.cmds.currentTime", "line_number": 44, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 44, "usage_type": "name"}, {"api_name": "maya.cmds.playbackOptions", "line_number": 63, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 63, "usage_type": "name"}, {"api_name": "maya.cmds.playbackOptions", "line_number": 64, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 64, "usage_type": "name"}, {"api_name": "PySide2.QtGui.QPainter", "line_number": 73, "usage_type": "call"}, {"api_name": "PySide2.QtGui", "line_number": 73, "usage_type": "name"}, {"api_name": "PySide2.QtGui.QColor", "line_number": 80, "usage_type": "call"}, {"api_name": "PySide2.QtGui", "line_number": 80, "usage_type": "name"}, {"api_name": "PySide2.QtWidgets.QDialog", "line_number": 98, "usage_type": "attribute"}, {"api_name": "PySide2.QtWidgets", "line_number": 98, "usage_type": "name"}, {"api_name": "maya.cmds.about", "line_number": 115, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 115, "usage_type": "name"}, {"api_name": "PySide2.QtCore.Qt", "line_number": 116, "usage_type": "attribute"}, {"api_name": "PySide2.QtCore", "line_number": 116, "usage_type": "name"}, {"api_name": "maya.cmds.about", "line_number": 117, "usage_type": "call"}, {"api_name": "maya.cmds", "line_number": 117, "usage_type": "name"}, {"api_name": "PySide2.QtCore.Qt", "line_number": 118, "usage_type": "attribute"}, {"api_name": "PySide2.QtCore", "line_number": 118, "usage_type": "name"}, {"api_name": "PySide2.QtWidgets.QCheckBox", "line_number": 129, "usage_type": "call"}, {"api_name": "PySide2.QtWidgets", "line_number": 129, "usage_type": "name"}, {"api_name": "PySide2.QtWidgets.QCheckBox", "line_number": 131, "usage_type": "call"}, {"api_name": "PySide2.QtWidgets", "line_number": 131, "usage_type": "name"}, {"api_name": "PySide2.QtWidgets.QPushButton", "line_number": 134, "usage_type": "call"}, {"api_name": "PySide2.QtWidgets", "line_number": 134, "usage_type": "name"}, {"api_name": "PySide2.QtWidgets.QPushButton", "line_number": 135, "usage_type": "call"}, {"api_name": "PySide2.QtWidgets", "line_number": 135, "usage_type": "name"}, {"api_name": "PySide2.QtWidgets.QPushButton", "line_number": 137, "usage_type": "call"}, {"api_name": "PySide2.QtWidgets", "line_number": 137, "usage_type": "name"}, {"api_name": "PySide2.QtWidgets.QHBoxLayout", "line_number": 140, "usage_type": "call"}, {"api_name": "PySide2.QtWidgets", "line_number": 140, "usage_type": "name"}, {"api_name": "PySide2.QtWidgets.QVBoxLayout", "line_number": 146, "usage_type": "call"}, {"api_name": "PySide2.QtWidgets", "line_number": 146, "usage_type": "name"}, {"api_name": "PySide2.QtWidgets.QGroupBox", "line_number": 151, "usage_type": "call"}, {"api_name": "PySide2.QtWidgets", "line_number": 151, "usage_type": "name"}, {"api_name": "PySide2.QtWidgets.QHBoxLayout", "line_number": 154, "usage_type": "call"}, {"api_name": "PySide2.QtWidgets", "line_number": 154, "usage_type": "name"}, {"api_name": "PySide2.QtWidgets.QVBoxLayout", "line_number": 158, "usage_type": "call"}, {"api_name": "PySide2.QtWidgets", "line_number": 158, "usage_type": "name"}]} +{"seq_id": "555712479", "text": "# -*- coding: utf-8 -*-\n#\n# Copyright 2017 Xilosopher\n#\n# Author: Moro JoJo\n\n\nimport os\nimport json\nfrom utils.logger import framework_log\n\n\ndef get_config_path():\n os_path = os.getcwd()\n return os_path + '/config/config.json'\n\n\ndef load_config(config_path):\n if config_path is None:\n return {}\n if not os.path.exists(config_path):\n framework_log.error(\"config.json not found in {config_path}\").format(config_path)\n return False\n with open(config_path) as json_file:\n config = json.load(json_file)\n return config\n\n", "sub_path": "utils/config.py", "file_name": "config.py", "file_ext": "py", "file_size_in_byte": 564, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "os.getcwd", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "utils.logger.framework_log.error", "line_number": 22, "usage_type": "call"}, {"api_name": "utils.logger.framework_log", "line_number": 22, "usage_type": "name"}, {"api_name": "json.load", "line_number": 25, "usage_type": "call"}]} +{"seq_id": "545900791", "text": "#!/usr/bin/env python\n# coding=utf-8\nimport numpy as np\n\nfrom ..utils import NAG, GD\nfrom .centralized_optimizer import CentralizedOptimizer\n\nclass ADMM(CentralizedOptimizer):\n '''ADMM for consensus optimization described in http://www.princeton.edu/~yc5/ele522_optimization/lectures/ADMM.pdf'''\n\n def __init__(self, p, n_iters=100, rho=0.1, x_0=None, W=None, local_n_iters=100, delta=None, local_optimizer='NAG', verbose=False):\n super().__init__(p, n_iters, x_0, W, verbose)\n self.rho = rho\n self.local_optimizer = local_optimizer\n self.local_n_iters = local_n_iters\n self.Lambda = np.random.rand(self.dim, self.n_agent)\n self.delta = delta\n\n def update(self):\n self.n_comm[self.t] += 2*self.n_agent\n\n x = np.random.rand(self.dim, self.n_agent)\n z = self.x # Using notations from the tutorial\n\n for i in range(self.n_agent):\n\n def _grad(tmp):\n return self.grad(tmp, i) + self.rho / 2 * (tmp - z) + self.Lambda[:, i] / 2\n\n if self.local_optimizer == \"NAG\":\n x[:, i], _ = NAG(_grad, self.x.copy(), self.L + self.rho, self.sigma + self.rho, self.local_n_iters)\n else:\n if self.delta is not None:\n x[:, i], _ = GD(_grad, self.x.copy(), self.delta, self.local_n_iters)\n else:\n x[:, i], _ = GD(_grad, self.x.copy(), 2/(self.L + self.rho + self.sigma + self.rho), self.local_n_iters)\n\n z = (x + self.Lambda).mean(axis=1)\n for i in range(self.n_agent):\n self.Lambda[:, i] += self.rho * (x[:, i] - self.x)\n self.x = z # Update\n", "sub_path": "optimizers/centralized/ADMM.py", "file_name": "ADMM.py", "file_ext": "py", "file_size_in_byte": 1662, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "centralized_optimizer.CentralizedOptimizer", "line_number": 8, "usage_type": "name"}, {"api_name": "numpy.random.rand", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 16, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 22, "usage_type": "attribute"}, {"api_name": "utils.NAG", "line_number": 31, "usage_type": "call"}, {"api_name": "utils.GD", "line_number": 34, "usage_type": "call"}, {"api_name": "utils.GD", "line_number": 36, "usage_type": "call"}]} +{"seq_id": "491341691", "text": "import importlib\nimport sys\nimport time\nimport numpy as np\nimportlib.reload(sys)\n\nfrom sklearn.grid_search import GridSearchCV \n\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.externals import joblib\n\n\nx_test = []\ny_test = []\n\nff=open('/home/rxx/quant/stock/temp1/20170221','r')\nline1=ff.readline()\nwhile line1:\n\ttoken1=line1.strip().split(' ')\n\ttemp1 = []\n\tfor i in range(1,210):\n\t\ttemp1.append(token1[i])\n\tx_test.append(temp1)\n\tif token1[210] >= '0':\n\t\ty_test.append(1)\n\telse:\n\t\ty_test.append(0)\n\tline1=ff.readline()\nprint (\"data done~\")\n\nclf=joblib.load(\"train_model.m\")\nprediction = clf.predict(x_test)\nprint (\"Accuracy:\\t\", (y_test == clf.predict(x_test)).mean())\n\nprint(prediction == y_test)\nprint(clf.feature_importances_)\nprint('all done')", "sub_path": "数据处理/randomforest_5.py", "file_name": "randomforest_5.py", "file_ext": "py", "file_size_in_byte": 766, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "importlib.reload", "line_number": 5, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib.load", "line_number": 31, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib", "line_number": 31, "usage_type": "name"}]} +{"seq_id": "44016589", "text": "import freenect\nimport cv2\nimport numpy as np\nimport imutils\nfrom cameradata.utils.perspTransform import four_point_transform\nfrom cameradata.utils.get_pts_gui import get_points\nfrom time import sleep\nfrom operator import itemgetter\n\n\ndef get_video(pts):\n array, _ = freenect.sync_get_video()\n array = cv2.cvtColor(array, cv2.COLOR_RGB2BGR)\n return four_point_transform(array, pts)\n\n\ndef get_depth(pts):\n array, _ = freenect.sync_get_depth()\n array = array.astype(np.uint8)\n return four_point_transform(array, pts)\n\n\ndef get_ball_reference(pts_depth, pts_rgb):\n \"\"\"Frame with no balls and no cue\"\"\"\n img_depth = get_depth(pts_depth)\n img = get_video(pts_rgb)\n return img_depth, img\n\n\ndef get_ball_diff(pts, Ra):\n return cv2.absdiff(get_depth(pts), Ra)\n\n\ndef get_ball_contours(pts, Ra):\n ballFrame = get_ball_diff(pts, Ra)\n # cv2.imshow(\"2\", imutils.resize(ballFrame, height=320))\n\n Bw = 29\n Tb = (13 / 16) * Bw\n ballBin = np.zeros(ballFrame.shape, np.uint8)\n ballBin[ballFrame > Tb] = 255\n # cv2.imshow(\"3\", imutils.resize(ballBin, height=320))\n\n # ballSmooth = cv2.medianBlur(ballBin, 5)\n ballErode = cv2.erode(ballBin, np.ones((7, 7), np.uint8))\n # cv2.imshow(\"4\", imutils.resize(ballErode, height=320))\n\n _, contours, hierarchy = cv2.findContours(ballErode, cv2.RETR_TREE, cv2.CHAIN_APPROX_SIMPLE)\n return contours\n\n\ndef draw_contour_circles(img, contours):\n n = 0\n for i in range(len(contours)):\n if cv2.contourArea(contours[i]) > 110:\n (x, y), radius = cv2.minEnclosingCircle(contours[i])\n center = (int(x), int(y))\n radius = int(radius)\n img = cv2.drawContours(img, contours, i, 255, 1)\n img = cv2.circle(img, center, radius + 3, 255, 1)\n img = cv2.rectangle(img, (int(x - 1), int(y - 1)), (int(x + 1), int(y + 1)), 255)\n n += 1\n return img\n\n\ndef get_white_ball(img_rgb, img_depth, contours):\n img_rgb = cv2.resize(img_rgb, (img_depth.shape[1], img_depth.shape[0]))\n # Initialize empty list\n # global max_cnt_index, max_cnt\n lst_intensities = []\n\n # For each list of contour points...\n\n for i in range(len(contours)):\n if cv2.contourArea(contours[i]) > 110:\n # Create a mask image that contains the contour filled in\n cimg = np.zeros_like(img_rgb)\n cimg = cv2.resize(cimg, (img_depth.shape[1], img_depth.shape[0]))\n cv2.drawContours(cimg, contours, i, color=255, thickness=-1)\n\n # Access the image pixels and create a 1D numpy array then add to list\n pts = np.where(cimg == 255)\n lst_intensities.append((i, img_rgb[pts[0], pts[1]]))\n\n # print(lst_intensities)\n\n sum_val = []\n Sum = 0\n for i, cnt in lst_intensities:\n for val in cnt:\n # print(val)\n Sum += sum(val)\n sum_val.append((i, Sum))\n Sum = 0\n #print(sum_val)\n\n max_cnt_index = None\n max_cnt = None\n if len(sum_val) > 0:\n max_cnt_index = max(sum_val, key=itemgetter(1))[0]\n cv2.drawContours(img_rgb, contours, max_cnt_index, color=(255, 120, 0), thickness=-1)\n max_cnt = contours[max_cnt_index]\n\n return max_cnt_index, max_cnt, img_rgb\n\n\npts_depth = get_points(1)\npts_rgb = get_points(0)\nRa, Ra_rgb = get_ball_reference(pts_depth, pts_rgb)\n# cv2.imshow(\"1\", imutils.resize(Ra, height=320))\nsleep(0)\n\nwhile 1:\n contours = get_ball_contours(pts_depth, Ra)\n\n img_depth = get_depth(pts_depth)\n img_rgb = get_video(pts_rgb)\n\n img_rgb = cv2.resize(img_rgb, (img_depth.shape[1], img_depth.shape[0]))\n # image1 = cv2.drawContours(image1, contours, -1, 255, 2)\n\n img_depth = draw_contour_circles(img_depth, contours)\n img_rgb = draw_contour_circles(img_rgb, contours)\n\n cv2.imshow(\"5\", imutils.resize(img_depth, height=320))\n\n max_cnt_idx, max_cnt, img_rgb = get_white_ball(img_rgb, img_depth, contours)\n\n cv2.imshow(\"6\", imutils.resize(img_rgb, height=320))\n\n k = cv2.waitKey(5) & 0xFF\n if k == 27:\n break\ncv2.destroyAllWindows()\n", "sub_path": "cameradata/ball_detect/ball_detect.py", "file_name": "ball_detect.py", "file_ext": "py", "file_size_in_byte": 4082, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "freenect.sync_get_video", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2BGR", "line_number": 13, "usage_type": "attribute"}, {"api_name": "cameradata.utils.perspTransform.four_point_transform", "line_number": 14, "usage_type": "call"}, {"api_name": "freenect.sync_get_depth", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 19, "usage_type": "attribute"}, {"api_name": "cameradata.utils.perspTransform.four_point_transform", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.absdiff", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 40, "usage_type": "attribute"}, {"api_name": "cv2.erode", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 45, "usage_type": "attribute"}, {"api_name": "cv2.findContours", "line_number": 48, "usage_type": "call"}, {"api_name": "cv2.RETR_TREE", "line_number": 48, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_SIMPLE", "line_number": 48, "usage_type": "attribute"}, {"api_name": "cv2.contourArea", "line_number": 55, "usage_type": "call"}, {"api_name": "cv2.minEnclosingCircle", "line_number": 56, "usage_type": "call"}, {"api_name": "cv2.drawContours", "line_number": 59, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 60, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 61, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 67, "usage_type": "call"}, {"api_name": "cv2.contourArea", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 77, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 78, "usage_type": "call"}, {"api_name": "cv2.drawContours", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 82, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 100, "usage_type": "call"}, {"api_name": "cv2.drawContours", "line_number": 101, "usage_type": "call"}, {"api_name": "cameradata.utils.get_pts_gui.get_points", "line_number": 107, "usage_type": "call"}, {"api_name": "cameradata.utils.get_pts_gui.get_points", "line_number": 108, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 111, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 119, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 125, "usage_type": "call"}, {"api_name": "imutils.resize", "line_number": 125, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 129, "usage_type": "call"}, {"api_name": "imutils.resize", "line_number": 129, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 131, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 134, "usage_type": "call"}]} +{"seq_id": "245615139", "text": "import csv\nimport shutil\nfrom splinter import Browser\nimport time\nimport os\nimport sys\nsys.path.append('../')\nimport gsheet\n\n# https://docs.google.com/spreadsheets/d/19_IaGTfPGtGxLoO-Rlc-Ear6A0R0vL0WtZqe5EIskr0/edit#gid=1338425418\n# \"auto realestate report\"\nspreadsheetId = \"19_IaGTfPGtGxLoO-Rlc-Ear6A0R0vL0WtZqe5EIskr0\"\n\ncategories = [\n 'Rent Income',\n 'Mortgage',\n 'Association Fees',\n 'Repairs',\n 'Management Fees',\n 'Commissions',\n 'Taxes',\n 'Utilities',\n 'Insurance',\n 'Cleaning and Maintenance',\n 'Other Income']\n\ncategory_redfines = {'Repairs Income': 'Repairs', 'Utility Income': 'Utilities',\n 'Insurance': 'Insurance',\n 'Cleaning and Maint Income': 'Maintenance'}\n\nignore_categories = ['Owner Draw', 'Owner Contribution']\n\nseries = {\"5650 E Sahara Ave 2063\": \"B\",\n \"8073 Palace Monaco Ave\": \"B\",\n \"10550 W Alexander Rd 1130\": \"A\",\n \"50 E Serene Ave 212\": \"D\"}\nnickname = {\"5650 E Sahara Ave 2063\": \"Terrasanta\",\n \"8073 Palace Monaco Ave\": \"Palace Monaco\",\n \"10550 W Alexander Rd 1130\": \"Chateau Versailles\",\n \"50 E Serene Ave 212\": \"Serene\"}\n\ndownload_dir = \"/Users/eleanorwen/Downloads/\"\n\noutput_dir = './output/'\nraw_dir = './output/raw/'\n\nentity_decoder = {'Cardinal_A': ['ellie.wen@gmail.com', '1'],\n 'Cardinal_B': ['ellie.wen@gmail.com', '2'],\n 'Cardinal_C': ['gegkat@gmail.com', '1'],\n 'Cardinal_D': ['gegkat@gmail.com', '2'],\n 'Ellie_Wen': ['ellie.wen@gmail.com', '3'],\n 'Greg_Katz': ['gegkat@gmail.com', '3'],\n 'You_Me': ['gegkat@gmail.com', '4']}\n\nirow_dict = {'ROS':[0, 1, 3, 9, 10, 12, 14, 16], # Rental Owner Statement\n 'RR': [0, 2, 9, 7] # Rent Roll\n }\n\ndownload_report_name_dict = {'ROS': 'Rental_Owner_Statement.csv',\n 'RR': 'Rent_Roll.csv'}\n\ndef rows2cols(table_data):\n headers = table_data[0]\n data = table_data[1:]\n cols = {}\n for col in headers:\n cols[col] = []\n for row in data:\n for col in range(len(headers)):\n cols[headers[col]].append(row[col])\n return cols\n\ndef str2float(l):\n return [float(i) for i in l]\n\ndef breakout_mortgage(cols):\n N = len(cols['glAccountName'])\n for i in range(N):\n cat = cols['glAccountName'][i]\n payee = cols['payeeName'][i]\n if cat == 'Owner Draw' and 'Cardinal Investments LLC' not in payee:\n cols['glAccountName'][i] = 'Mortgage'\n return cols\n\ndef group(cols):\n props = {}\n N = len(cols['glAccountName'])\n for i in range(N):\n #print(props)\n cat = cols['glAccountName'][i]\n if cat in category_redfines:\n cat = category_redfines[cat]\n amount = cols['amount'][i]\n address = cols['buildingName'][i]\n if address in props:\n if cat in props[address]:\n props[address][cat] += amount\n else:\n if cat not in ignore_categories:\n print(\"category {} not recognized\".format(cat))\n else:\n props[address] = {}\n for cat in categories:\n #print(cat)\n props[address][cat] = 0\n #print(props)\n return props\n\ndef period_time_struct(date):\n curr_time_sec = time.time()\n curr_time_struct = time.localtime(curr_time_sec)\n if date: # date is not empty\n month = date[0]\n year = date[1]\n else:\n month = curr_time_struct.tm_mon\n year = curr_time_struct.tm_year\n\n # Go to previous month if not past the 15th\n if curr_time_struct.tm_mday < 10:\n month -= 1\n # Loop to december if necessary\n if month == 0:\n month = 12\n year -= 1\n day = 15 # Always use the 15th because that is when period always ends\n return time.strptime(\"{}-{}-{}\".format(year, month, day), \"%Y-%m-%d\")\n\ndef report_date_str(time_struct):\n return time.strftime(\"Period ending %B %d, %Y\", time_struct)\n\ndef gsheet_date_str(time_struct):\n return time.strftime(\"%b %Y\", time_struct)\n\n\ndef force_move_file(src, dst):\n src_base = os.path.basename(src)\n if not src_base:\n print(\"force_move_file is gauranteed to work only when the src is a file not a directory\")\n print(src)\n print(src_base)\n\n dst_dir = os.path.dirname(dst)\n dst_base = os.path.basename(dst)\n\n if not dst_base: # empty base, so dst is a directory\n dst_base = os.path.basename(src)\n dst = os.path.join(dst_dir, dst_base)\n\n # remove existing file if neccessary\n if os.path.isfile(dst):\n os.remove(dst)\n\n # create directory if necessary\n if dst_dir and not os.path.isdir(dst_dir):\n os.mkdir(dst_dir)\n\n shutil.move(src, dst)\n\n\n\n\nclass Report:\n def __init__(self, type, in_file = '', out_file='', entity = 'Cardinal_A', date = [], download_new = True):\n\n self.entity = entity\n self.type = type\n\n # get time for file names\n #timestr = time.strftime(\"%Y%m%d-%H%M%S\")\n #timestr = time.strftime(\"%m_%Y\")\n\n self.date_time_struct = period_time_struct(date)\n\n\n\n if in_file:\n self.in_file = in_file\n else:\n self.get_default_name()\n\n print(\"Preparing report for {}\".format(self.in_file))\n\n\n if out_file: # output file name given\n self.out_file = out_file\n else: # create output file name\n self.out_file = output_dir + self.in_file[:-4] + '_Formatted.csv'\n\n # Check if in_file already exists, if not than download from website\n if download_new:\n self.download()\n else:\n if not os.path.isfile(self.in_file):\n self.in_file = raw_dir + self.in_file # try looking in raw dir\n if not os.path.isfile(self.in_file):\n print('{} not found so downloading'.format(self.in_file))\n self.download()\n\n # Convert raw csv file to a more readable format\n self.convert()\n\n # Update google drive sheet with information\n self.update_gsheet()\n print(\"\\n\")\n\n def get_default_name(self):\n if self.type is 'ROS':\n report = 'Rental_Owner_Statement'\n elif self.type is 'RR':\n report = 'Rent_Roll'\n else:\n print('unrecognized type')\n\n datestr = time.strftime(\"%m_%Y\", self.date_time_struct)\n\n self.in_file = self.entity + '_' + report + '_' + datestr + '.csv'\n self.report = report\n self.datestr = datestr\n\n # Use splinter to open a browser and download information from Foster website\n def download(self):\n if self.entity in entity_decoder:\n self.login_id, self.login_number = entity_decoder[self.entity]\n else:\n print(\"Did not recognize entity: {}\".format(self.entity))\n with Browser('chrome') as browser:\n # Log in to Foster site\n url = \"https://foster.managebuilding.com/Manager/PublicPages/Login.aspx?ReturnUrl=%2fmanager%2f\"\n browser.visit(url)\n username = self.login_id\n password = 'wenfamily8'\n browser.find_by_id('txtUserName').first.fill(username)\n browser.find_by_id('txtPassword').first.fill(password)\n browser.find_by_id('btnLogIn').first.click()\n xpath = '//*[@id=\"chooseAccountContainer\"]/li[' + self.login_number + ']/a'\n browser.find_by_xpath(xpath).first.click()\n\n # Download Rent Roll file\n if self.type is \"RR\":\n rent_roll_site = 'https://foster.managebuilding.com/Manager/Reports/ReportsSearch.aspx?category=Residents&query=RentRoll'\n browser.visit(rent_roll_site)\n browser.find_by_id('txtexportTypes').select_by_text('Comma-separated text (.csv)')\n browser.find_by_id('btnSubmit').first.click()\n\n # Download Rental Owner Statement\n if self.type is \"ROS\":\n rental_owner_statement_site = 'https://foster.managebuilding.com/Manager/Reports/ReportsSearch.aspx?category=Financials&query=OwnerStatementNewCalculation'\n browser.visit(rental_owner_statement_site)\n# browser.find_by_id('txtexportTypes').select_by_text('Comma-separated text (.csv)')\n browser.find_by_id('cbincludeDetailTransactions').first.click()\n browser.find_by_id('txtdaterange').select_by_text(report_date_str(self.date_time_struct))\n# browser.find_by_id('btnSubmit').first.click()\n browser.find_by_id('lnkExportAs').click()\n browser.find_by_id('lnkExportCsv').click()\n\n # Sleep to make sure there is time for file(s) to download\n time.sleep(5)\n force_move_file(download_dir + download_report_name_dict[self.type], self.in_file)\n\n # function to read and clean up raw csv report files\n def convert(self):\n in_data = []\n out_data = []\n irow = irow_dict[self.type]\n with open(self.in_file, 'r') as f:\n reader = csv.reader(f)\n with open(self.out_file, 'wt') as fido:\n for row in reader:\n in_data.append(row)\n if row[0].find('CANCELLED') == -1 and row[0].find('SOLD') == -1 and row[0].find(\n 'TRANSFERRED') == -1:\n s = ''\n for i in irow:\n s += row[i] + '|, '\n s = s[:-2]\n s += '\\n'\n fido.write(s)\n next_out_data = s.split('|,')\n next_out_data = [s.strip() for s in next_out_data]\n out_data.append(next_out_data)\n self.in_data = in_data\n self.out_data = out_data\n\n # function to update google sheet\n def update_gsheet(self):\n if len(self.out_data) > 1:\n if self.type is \"ROS\":\n self.update_gsheetROS()\n elif self.type is \"RR\":\n self.update_gsheetRR()\n else:\n print('did not recognize report type')\n else:\n print('out data is empty, nothing to update on gsheet')\n\n def update_gsheetROS(self):\n sheet_name = 'monthly tracking'\n date_str = gsheet_date_str(self.date_time_struct)\n G = gsheet.Gsheet(spreadsheetId)\n # data = self.out_data[1:]\n data = self.out_data\n cols = rows2cols(data)\n cols['amount'] = str2float(cols['amount'])\n cols['amount'] = [-1 * x for x in cols['amount']] # reverse sign of amount\n cols = breakout_mortgage(cols)\n props = group(cols)\n currData = G.query(sheet_name)\n print_data = []\n for prop, propdict in props.items():\n row = []\n row.append(prop)\n row.append(nickname[prop])\n row.append(series[prop])\n row.append(date_str)\n for cat in categories:\n row.append(propdict[cat])\n print_data.append(row)\n row_index = G.firstBlankRow(sheet_name) # start assuming append to end\n count = 1\n for existing_row in currData: # check if address and month already exist\n match = True\n if len(existing_row) > 4:\n for i in range(4):\n if existing_row[i] != row[i]:\n match = False\n break\n else:\n match = False\n\n if match:\n print(\"found match at row {}\".format(count))\n row_index = count\n break\n count += 1\n G.update(print_data, sheet_name + '!A' + str(row_index))\n\n self.G = G\n self.cols = cols\n self.groups = props\n self.print_data = print_data\n\n def update_gsheetRR(self):\n G = gsheet.Gsheet(spreadsheetId)\n # data = self.out_data[1:]\n data = self.out_data\n for r in data:\n r.append(self.in_file)\n row = G.firstBlankRow('rent_roll')\n G.update(data, 'rent_roll!A' + str(row))\n self.G = G\n # G.addSheet(self.rent_roll_file)\n # G.update(self.out_data, \"'\" + self.rent_roll_file + \"'!A1\")\n # G.update(self.out_data, \"A1\")\n\ndef tryReport(type, entity, date, download_new):\n RR = []\n try:\n RR = Report(type=type, entity=entity, date=date)\n except:\n print('Failed to make report: {}, {}, {}'.format(type, entity, date, download_new))\n return RR\n\ndef CurrentCardinalROS():\n type = 'ROS'\n date = []\n download_new = True\n reports = []\n entities = ['Cardinal_A', 'Cardinal_B', 'Cardinal_D']\n for entity in entities:\n reports.append(tryReport(type=type, entity = entity, date=date, download_new = download_new))\n return reports\n\nif __name__ == \"__main__\":\n for i in range(1,13):\n date = [i, 2015]\n timestr = \"{}_{}\".format(date[0], date[1])\n type = 'ROS'\n download_new = True\n\n #entities = ['Carindal_A', 'Cardinal_B', 'Cardinal_C', 'Cardinal_D', 'Ellie_Wen', 'Greg_Katz', 'You_Me']\n #entities = ['Carindal_A', 'Cardinal_B', 'Cardinal_D']\n entities = ['Ellie_Wen', 'Greg_Katz']\n #entities = ['You_Me']\n\n for entity in entities:\n tryReport(type=type, entity=entity, date=date, download_new=download_new)\n\n", "sub_path": "python/realestate/realestate.py", "file_name": "realestate.py", "file_ext": "py", "file_size_in_byte": 13706, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "sys.path.append", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 110, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 111, "usage_type": "call"}, {"api_name": "time.strptime", "line_number": 127, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 130, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 133, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 137, "usage_type": "call"}, {"api_name": "os.path", "line_number": 137, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 143, "usage_type": "call"}, {"api_name": "os.path", "line_number": 143, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 144, "usage_type": "call"}, {"api_name": "os.path", "line_number": 144, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 147, "usage_type": "call"}, {"api_name": "os.path", "line_number": 147, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 148, "usage_type": "call"}, {"api_name": "os.path", "line_number": 148, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 151, "usage_type": "call"}, {"api_name": "os.path", "line_number": 151, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 152, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 155, "usage_type": "call"}, {"api_name": "os.path", "line_number": 155, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 156, "usage_type": "call"}, {"api_name": "shutil.move", "line_number": 158, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 194, "usage_type": "call"}, {"api_name": "os.path", "line_number": 194, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 196, "usage_type": "call"}, {"api_name": "os.path", "line_number": 196, "usage_type": "attribute"}, {"api_name": "time.strftime", "line_number": 215, "usage_type": "call"}, {"api_name": "splinter.Browser", "line_number": 227, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 258, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 267, "usage_type": "call"}, {"api_name": "gsheet.Gsheet", "line_number": 300, "usage_type": "call"}, {"api_name": "gsheet.Gsheet", "line_number": 344, "usage_type": "call"}]} +{"seq_id": "514780202", "text": "from matplotlib.pylab import gca, figure, plot, subplot, title, xlabel, ylabel, xlim,show\nfrom matplotlib.lines import Line2D\nimport segment\nimport fit\nimport sys\n\nfrom numpy import array\ndef draw_plot(data,plot_title):\n plot(data[:,0],data[:,1],alpha=0.8,color='red')\n title(plot_title)\n xlabel(\"Samples\")\n ylabel(\"Signal\")\n xlim((data[:,0][0],data[:,0][-1]))\n\ndef draw_segments(segments,ticks):\n segments=[(x1,y1,x21,y21) for x,(x1,y1),x2,(x21,y21),err in segments]\n ax = gca()\n ax.set_xticklabels(ticks, rotation=20)\n for segment in segments:\n line = Line2D((segment[0],segment[2]),(segment[1],segment[3]))\n ax.add_line(line)\n\ndef wrapOrchestration(title, labels, data, segment_algo, create_segment, compute_error, max_error):\n figure()\n segments = segment_algo(data, create_segment, compute_error, max_error)\n draw_plot(data,title)\n draw_segments(segments,labels)\n print(len(segments))\n print(len(data)/len(segments))\n name=title.split()[0][:3]+title.split()[-1][:3]\n with open(f\"./output/{name}_{labels[0]}_{labels[-1]}_{max_error}.txt\",\"w\") as file:\n for xbeg, (x0,y0), xend, (x1,y1), err in segments:\n file.write(f\"{labels[xbeg]}\\t{x0}\\t{y0}\\n\")\n file.write(f\"{labels[xend]}\\t{x1}\\t{y1}\\t{err}\\n\")\n\n\nMIN=int(sys.argv[1]) if len(sys.argv)>1 else 0\nMAX=int(sys.argv[2]) if len(sys.argv)>2 else MIN+30\nERROR=float(sys.argv[3]) if len(sys.argv)>3 else 1\nwith open(\"example_data/bitcoin_2010-8-16_2021-9-8.txt\") as f:\n file_lines = f.readlines()\n\ndata = [tuple(x.split(\"\\t\")[1:3]) for x in file_lines[MIN:MAX]]\nlabels = [x.split(\"\\t\")[0] for x in file_lines[MIN:MAX]]\ndata = array([(float(x),float(y.strip())) for x,y in data])\nmax_error = ERROR\n\n#sliding window with regression \nwrapOrchestration(\"Sliding window with regression\",labels,data,segment.slidingwindowsegment, fit.regression, fit.sumsquared_error, max_error)\n#bottom-up with regression\nwrapOrchestration(\"Bottom-up with regression\",labels,data,segment.bottomupsegment,fit.regression, fit.sumsquared_error, max_error)\n#top-down with regression\n#wrapOrchestration(\"Top-down with regression\",labels,data,segment.topdownsegment,fit.regression, fit.sumsquared_error, max_error)\n\n#sliding window with simple interpolation\nwrapOrchestration(\"Sliding window with simple interpolation\",labels,data,segment.slidingwindowsegment, fit.interpolate, fit.sumsquared_error, max_error)\n#bottom-up with simple interpolation\nwrapOrchestration(\"Bottom-up with simple interpolation\",labels,data,segment.bottomupsegment,fit.interpolate, fit.sumsquared_error, max_error)\n#top-down with simple interpolation\n#wrapOrchestration(\"Top-down with simple interpolation\",labels,data,segment.topdownsegment,fit.interpolate, fit.sumsquared_error, max_error)\n\nshow()\n\n", "sub_path": "example.py", "file_name": "example.py", "file_ext": "py", "file_size_in_byte": 2798, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "matplotlib.pylab.plot", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.pylab.title", "line_number": 10, "usage_type": "call"}, {"api_name": "matplotlib.pylab.xlabel", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.pylab.ylabel", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pylab.xlim", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pylab.gca", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.lines.Line2D", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pylab.figure", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pylab.title", "line_number": 26, "usage_type": "argument"}, {"api_name": "matplotlib.pylab.title.split", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pylab.title", "line_number": 30, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 37, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 38, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 39, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 45, "usage_type": "call"}, {"api_name": "segment.slidingwindowsegment", "line_number": 49, "usage_type": "attribute"}, {"api_name": "fit.regression", "line_number": 49, "usage_type": "attribute"}, {"api_name": "fit.sumsquared_error", "line_number": 49, "usage_type": "attribute"}, {"api_name": "segment.bottomupsegment", "line_number": 51, "usage_type": "attribute"}, {"api_name": "fit.regression", "line_number": 51, "usage_type": "attribute"}, {"api_name": "fit.sumsquared_error", "line_number": 51, "usage_type": "attribute"}, {"api_name": "segment.slidingwindowsegment", "line_number": 56, "usage_type": "attribute"}, {"api_name": "fit.interpolate", "line_number": 56, "usage_type": "attribute"}, {"api_name": "fit.sumsquared_error", "line_number": 56, "usage_type": "attribute"}, {"api_name": "segment.bottomupsegment", "line_number": 58, "usage_type": "attribute"}, {"api_name": "fit.interpolate", "line_number": 58, "usage_type": "attribute"}, {"api_name": "fit.sumsquared_error", "line_number": 58, "usage_type": "attribute"}, {"api_name": "matplotlib.pylab.show", "line_number": 62, "usage_type": "call"}]} +{"seq_id": "300520338", "text": "from lxml import etree\n\n\n\ndef add_entry(dict, key, value = None):\n if key in dict: # если ключ существует\n if value == None:\n pass\n #'ключ есть, новое значение None - пропускаем'\n return dict\n else:\n if dict.get(key) == None:\n dict.update({key: [value]})\n #'ключ есть, старое значение None - добавляем не None'\n return dict\n else:\n old_value = dict.get(key)\n old_value.append(value)\n dict.update({key: old_value})\n #'ключ есть, старое значение не None - добавляем в список'\n return dict\n else:\n if value == None:\n dict.update({key: value})\n #'ключa нет - добавляем None'\n return dict\n else:\n dict.update({key: [ value]})\n #'ключa нет - добавляем'\n return dict\n\n\n\ndef HeaderCheck(filename):\n result = {}\n tree = etree.parse(filename)\n root = tree.getroot()\n for dashboard in root.iter('dashboard'): #проходим по каждой ноде \n header_check = list()\n\n for zones in dashboard.iter('zones'): #ищем первую ноду с именем \n\n for item in zones.getchildren():\n if item.get('type') == 'layout-basic': # проеряем тип зоны, чтобы исключить floating контейнеры\n\n child = item.find('zone') #vert container\n header_check.append(child.attrib)\n\n\n if child.find('zone') != None: # проверка структуры заголовка\n sub_child = child.find('zone') #horz container\n header_check.append(sub_child.attrib)\n\n # проверка структуры заголовка\n if header_check[0].get('param') == 'vert' and header_check[1].get('param') == 'horz':\n result = add_entry(key = 'Заголовок', dict = result)\n else:\n result = add_entry(key = 'Заголовок', value = ' в дашборде \"{}\" структура не соответствует'.format(dashboard.get('name')) , dict = result)\n\n # проверка высоты заголовка\n if int(header_check[1].get('fixed-size')) == 44:\n result = add_entry(key = 'Заголовок', dict = result)\n else:\n result = add_entry(key = 'Заголовок', value = ' в дашборде \"{}\" высота не равна 44px'.format(dashboard.get('name')) , dict = result)\n\n # проверка ширины логотипа\n logo = sub_child.find('zone')\n\n try:\n if int(logo.get('fixed-size')) == 105:\n result = add_entry(key = 'Заголовок', dict = result)\n else:\n result = add_entry(key = 'Заголовок', value = ' в дашборде \"{}\" ширина логотипа не равна 105px'.format(dashboard.get('name')), dict = result )\n except:\n result = add_entry(key = 'Заголовок', value = ' в дашборде \"{}\" проверка ширины логотипа не удалась'.format(dashboard.get('name')), dict = result )\n\n # проверка отступов логотипа\n try:\n if int(logo.find('..//zone-style/format[@attr=\"margin\"]').get('value')) == 12:\n result = add_entry(key = 'Заголовок')\n else:\n result = add_entry(key = 'Заголовок', value = ' в дашборде \"{}\" внешние отступы логотипа не равны 12px'.format(dashboard.get('name')), dict = result )\n except:\n result = add_entry(key = 'Заголовок', value = ' в дашборде \"{}\" проверка отступов логотипа не удалась'.format(dashboard.get('name')), dict = result )\n\n # проверка центровки логотипа\n try:\n if int(logo.get('is-centered')) == 0:\n result = add_entry(key = 'Заголовок', value = ' в дашборде \"{}\" у логотипа не не стоит флаг \"Center Image\"'.format(dashboard.get('name')), dict = result )\n else:\n pass\n except:\n result = add_entry(key = 'Заголовок', dict = result)\n\n # проверка Fit Image логотипа\n try:\n if int(logo.get('is-scaled')) == 1:\n result = add_entry(key = 'Заголовок', dict = result)\n else:\n result = add_entry(key = 'Заголовок', value = ' в дашборде \"{}\" у логотипа не не стоит флаг \"Fit Image\"'.format(dashboard.get('name')), dict = result )\n except:\n result = add_entry(key = 'Заголовок', value = ' в дашборде \"{}\" проверка параметра \"Fit Image\" не удалась'.format(dashboard.get('name')), dict = result )\n\n # проверка фона #565C61\n if sub_child.find('..//zone-style/format[@attr=\"background-color\"]').get('value').upper() == '#565C61':\n result = add_entry(key = 'Заголовок', dict = result)\n else:\n result = add_entry(key = 'Заголовок', value = ' в дашборде \"{}\" у заголовка фон не \"#565C61\" '.format(dashboard.get('name')), dict = result )\n\n else:\n result = add_entry(key = 'Заголовок', value = ' в дашборде \"{}\" структура не соответствует'.format(dashboard.get('name')) , dict = result)\n\n\n else:\n pass\n return result\n", "sub_path": "uploader/logic.py", "file_name": "logic.py", "file_ext": "py", "file_size_in_byte": 6790, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "lxml.etree.parse", "line_number": 36, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 36, "usage_type": "name"}]} +{"seq_id": "592770237", "text": "from typing import List, Any\n\nfrom hearthstone.entities import Entity\nfrom hearthstone.enums import GameTag\n\nfrom .base_entity import BaseEntity\nfrom .spell_entity import SpellEntity\n\n\nclass HeroEntity(BaseEntity):\n\n def __init__(self, entity: Entity):\n super().__init__(entity)\n self.card_id = 0\n self.atk = 0\n self.max_health = 0\n # 受伤\n self.damage = 0\n # INVALID = 0 施法者CASTER = 1 斗士FIGHTER = 2 TANK = 3 NEUTRAL = 4\n self.lettuce_role = 2 # 斗士\n self.cardrace = None\n self.pos = [0, 0] # 坐标[x, y]\n # 场上位置 从左往右1开始\n\n self.zone_position = 0\n # 意义不明\n self.cost = 0\n self.divine_shield = 0\n # INVALID = 0 部落HORDE = 1 联盟ALLIANCE = 2 中立NEUTRAL = 3\n self.faction = 0\n self.windfury = 0\n self.spell_cnt = 1\n # 被动 一技能 二技能 三技能 ...\n self.spell: List[SpellEntity] = []\n self.spellpower = 0\n self.deathrattle = 0\n # 是否选择了技能\n self.lettuce_has_manually_selected_ability = 0\n # 选了什么技能\n self.lettuce_ability_tile_visual_self_only = 0\n # 技能选择的目标\n self.lettuce_selected_target = 0\n # 经验 55000满级\n self.lettuce_mercenary_experience = 0\n self.skill_seq = None\n self.parse_entity()\n\n def parse_entity(self):\n if self.entity is None:\n return\n super(HeroEntity, self).parse_entity()\n self.card_id = self.entity.card_id\n self.atk = self.get_tag(GameTag.ATK)\n self.max_health = self.get_tag(GameTag.HEALTH)\n self.damage = self.get_tag(GameTag.DAMAGE)\n self.lettuce_role = self.get_tag(GameTag.LETTUCE_ROLE)\n self.cardrace = self.get_tag(GameTag.CARDRACE)\n self.zone_position = self.get_tag(GameTag.ZONE_POSITION)\n self.cost = self.get_tag(GameTag.COST)\n self.divine_shield = self.get_tag(GameTag.DIVINE_SHIELD)\n self.faction = self.get_tag(GameTag.FACTION)\n self.windfury = self.get_tag(GameTag.WINDFURY)\n self.spellpower = self.get_tag(GameTag.SPELLPOWER)\n self.deathrattle = self.get_tag(GameTag.DEATHRATTLE)\n self.lettuce_has_manually_selected_ability = self.get_tag(GameTag.LETTUCE_HAS_MANUALLY_SELECTED_ABILITY)\n self.lettuce_ability_tile_visual_self_only = self.get_tag(GameTag.LETTUCE_ABILITY_TILE_VISUAL_SELF_ONLY)\n self.lettuce_selected_target = self.get_tag(GameTag.LETTUCE_SELECTED_TARGET)\n self.lettuce_mercenary_experience = self.get_tag(GameTag.LETTUCE_MERCENARY_EXPERIENCE)\n\n def set_pos(self, x, y):\n self.pos = [x, y]\n\n def set_skill_seq(self, skills):\n self.skill_seq = skills\n\n def own(self):\n \"\"\"\n 谁的随从\n \"\"\"\n return self.controller == 3\n\n def add_spell(self, spell: SpellEntity):\n self.spell.append(spell)\n\n def get_health(self):\n return self.max_health - self.damage\n\n def basic_attack(self, target, dmg):\n total_dmg = dmg * BaseEntity.damage_advantage[self.lettuce_role][target.lettuce_role]\n target.damage += total_dmg\n return total_dmg\n\n def __str__(self):\n return {'card_id': self.card_id, 'atk': self.atk, 'health': self.get_health(),\n 'zone_pos': self.zone_position}.__str__()\n", "sub_path": "entity/hero_entity.py", "file_name": "hero_entity.py", "file_ext": "py", "file_size_in_byte": 3422, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "base_entity.BaseEntity", "line_number": 10, "usage_type": "name"}, {"api_name": "hearthstone.entities.Entity", "line_number": 12, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 34, "usage_type": "name"}, {"api_name": "spell_entity.SpellEntity", "line_number": 34, "usage_type": "name"}, {"api_name": "hearthstone.enums.GameTag.ATK", "line_number": 53, "usage_type": "attribute"}, {"api_name": "hearthstone.enums.GameTag", "line_number": 53, "usage_type": "name"}, {"api_name": "hearthstone.enums.GameTag.HEALTH", "line_number": 54, "usage_type": "attribute"}, {"api_name": "hearthstone.enums.GameTag", "line_number": 54, "usage_type": "name"}, {"api_name": "hearthstone.enums.GameTag.DAMAGE", "line_number": 55, "usage_type": "attribute"}, {"api_name": "hearthstone.enums.GameTag", "line_number": 55, "usage_type": "name"}, {"api_name": "hearthstone.enums.GameTag.LETTUCE_ROLE", "line_number": 56, "usage_type": "attribute"}, {"api_name": "hearthstone.enums.GameTag", "line_number": 56, "usage_type": "name"}, {"api_name": "hearthstone.enums.GameTag.CARDRACE", "line_number": 57, "usage_type": "attribute"}, {"api_name": "hearthstone.enums.GameTag", "line_number": 57, "usage_type": "name"}, {"api_name": "hearthstone.enums.GameTag.ZONE_POSITION", "line_number": 58, "usage_type": "attribute"}, {"api_name": "hearthstone.enums.GameTag", "line_number": 58, "usage_type": "name"}, {"api_name": "hearthstone.enums.GameTag.COST", "line_number": 59, "usage_type": "attribute"}, {"api_name": "hearthstone.enums.GameTag", "line_number": 59, "usage_type": "name"}, {"api_name": "hearthstone.enums.GameTag.DIVINE_SHIELD", "line_number": 60, "usage_type": "attribute"}, {"api_name": "hearthstone.enums.GameTag", "line_number": 60, "usage_type": "name"}, {"api_name": "hearthstone.enums.GameTag.FACTION", "line_number": 61, "usage_type": "attribute"}, {"api_name": "hearthstone.enums.GameTag", "line_number": 61, "usage_type": "name"}, {"api_name": "hearthstone.enums.GameTag.WINDFURY", "line_number": 62, "usage_type": "attribute"}, {"api_name": "hearthstone.enums.GameTag", "line_number": 62, "usage_type": "name"}, {"api_name": "hearthstone.enums.GameTag.SPELLPOWER", "line_number": 63, "usage_type": "attribute"}, {"api_name": "hearthstone.enums.GameTag", "line_number": 63, "usage_type": "name"}, {"api_name": "hearthstone.enums.GameTag.DEATHRATTLE", "line_number": 64, "usage_type": "attribute"}, {"api_name": "hearthstone.enums.GameTag", "line_number": 64, "usage_type": "name"}, {"api_name": "hearthstone.enums.GameTag.LETTUCE_HAS_MANUALLY_SELECTED_ABILITY", "line_number": 65, "usage_type": "attribute"}, {"api_name": "hearthstone.enums.GameTag", "line_number": 65, "usage_type": "name"}, {"api_name": "hearthstone.enums.GameTag.LETTUCE_ABILITY_TILE_VISUAL_SELF_ONLY", "line_number": 66, "usage_type": "attribute"}, {"api_name": "hearthstone.enums.GameTag", "line_number": 66, "usage_type": "name"}, {"api_name": "hearthstone.enums.GameTag.LETTUCE_SELECTED_TARGET", "line_number": 67, "usage_type": "attribute"}, {"api_name": "hearthstone.enums.GameTag", "line_number": 67, "usage_type": "name"}, {"api_name": "hearthstone.enums.GameTag.LETTUCE_MERCENARY_EXPERIENCE", "line_number": 68, "usage_type": "attribute"}, {"api_name": "hearthstone.enums.GameTag", "line_number": 68, "usage_type": "name"}, {"api_name": "spell_entity.SpellEntity", "line_number": 82, "usage_type": "name"}, {"api_name": "base_entity.BaseEntity.damage_advantage", "line_number": 89, "usage_type": "attribute"}, {"api_name": "base_entity.BaseEntity", "line_number": 89, "usage_type": "name"}]} +{"seq_id": "227136654", "text": "# Data-store record for voter\n\n\n# Import external modules\nimport logging\nimport re\nimport urlparse\n# Import app modules\nfrom configOpenVoterId import const as conf\nfrom secretsOpenVoterId import const as secrets\nimport security\nimport voter\n\n\n# If not unit testing... include gCloud code\nif __name__ != '__main__':\n\n from google.appengine.ext import ndb\n\n # Parent key: none\n # Key: identityHash: long alpha-numeric string\n class IdRecord( ndb.Model ):\n saltPhone = ndb.StringProperty()\n saltSocialSec = ndb.StringProperty()\n saltBirthdate = ndb.StringProperty()\n saltMailedCode = ndb.StringProperty()\n city = ndb.StringProperty()\n verificationHashes = ndb.TextProperty( repeated=True )\n\n\n # Parent key: none\n # Key: string: client/voter type + identity\n class RateRecord( ndb.Model ):\n loginFailuresSinceSuccess = ndb.IntegerProperty( default=0 )\n nextAttemptTime = ndb.IntegerProperty( default=0 ) # Needed to know when wait ends\n resetFailuresOnNextAttempt = ndb.BooleanProperty( default=False )\n \n def allowed( self, now ):\n\n logging.debug( 'RateRecord.allowed() now=' + str(now) + ' nextAttemptTime=' + str(self.nextAttemptTime) \n + ' nextAttemptTime-now = ' + str(self.nextAttemptTime - now)\n + ' loginFailuresSinceSuccess=' + str(self.loginFailuresSinceSuccess) )\n\n return (self.nextAttemptTime <= now)\n\n\n####################################################################################\n# Rate-limit functions\n\nconf.rateRecordTypeClient = 'client'\nconf.rateRecordTypeVoter = 'voter'\n\nconf.rateUseDatastore = False # Turn off persistent storage, and expire records via memcache expiration\nconf.rateUseMemcache = True\nconf.rateMemcacheTimeout = conf.oneDaySec\n\n# Continuing to login and failing will not keep memcache record alive longer than memcache-expiration, \n# because rate-check only reads memcache, and only memcache-writes update expiration-time\n\n\ndef retrieveClientRateLimit( clientIp ):\n recordId = toRateRecordId( conf.rateRecordTypeClient, clientIp )\n if conf.isDev: logging.debug( 'retrieveClientRateLimit() recordId=' + str(recordId) )\n return RateRecord.get_by_id( recordId, use_datastore=conf.rateUseDatastore, use_memcache=conf.rateUseMemcache, memcache_timeout=conf.rateMemcacheTimeout )\n\ndef retrieveVoterRateLimit( voterId ):\n recordId = toRateRecordId( conf.rateRecordTypeVoter, voterId )\n if conf.isDev: logging.debug( 'retrieveVoterRateLimit() recordId=' + str(recordId) )\n return RateRecord.get_by_id( recordId, use_datastore=conf.rateUseDatastore, use_memcache=conf.rateUseMemcache, memcache_timeout=conf.rateMemcacheTimeout )\n\n\n# Modifies/creates and returns rateRecord\ndef updateClientLoginRate( now, success, clientIp, rateRecord ):\n return updateLoginRate( now, success, clientIp, conf.rateRecordTypeClient, rateRecord )\n\ndef updateVoterLoginRate( now, success, voterId, rateRecord ):\n return updateLoginRate( now, success, voterId, conf.rateRecordTypeVoter, rateRecord )\n\ndef updateLoginRate( now, success, recordId, recordType, rateRecord ):\n\n if not rateRecord and ( recordId and recordType ):\n rateRecord = RateRecord( id=toRateRecordId(recordType, recordId) )\n\n if not rateRecord: return None\n \n if success or rateRecord.resetFailuresOnNextAttempt:\n rateRecord.loginFailuresSinceSuccess = 0\n rateRecord.nextAttemptTime = now\n rateRecord.resetFailuresOnNextAttempt = False\n else:\n # Wait-time proportional to number of failures since last login, until wait > 1day ... then reset fail-count\n failuresSinceSuccess = rateRecord.loginFailuresSinceSuccess\n if (not failuresSinceSuccess) or (failuresSinceSuccess < 0): failuresSinceSuccess = 0\n rateRecord.loginFailuresSinceSuccess = failuresSinceSuccess + 1\n\n waitSec = (2 << failuresSinceSuccess)\n if ( waitSec > conf.oneDaySec ):\n rateRecord.nextAttemptTime = now + conf.oneDaySec\n rateRecord.resetFailuresOnNextAttempt = True\n else:\n rateRecord.nextAttemptTime = now + waitSec\n rateRecord.resetFailuresOnNextAttempt = False\n \n # Store record synchronously, because async fails\n if rateRecord: rateRecord.put( use_datastore=conf.rateUseDatastore, use_memcache=conf.rateUseMemcache, memcache_timeout=conf.rateMemcacheTimeout )\n return rateRecord\n\n\ndef toRateRecordId( recordType, recordId ):\n return '{}_{}'.format( recordType, recordId )\n\n\n\n#################################################################################\n# Unit test\n\nimport unittest\n\nclass TestText(unittest.TestCase):\n\n def testRateLimit( self ):\n\n class FakeRateRecord:\n def __init__( self, failures=0 ):\n self.loginFailuresSinceSuccess = failures\n self.nextAttemptTime = 0\n self.resetFailuresOnNextAttempt = False\n\n def put( self, use_datastore=None, use_memcache=None, memcache_timeout=None, use_cache=None ): pass\n\n now = 100\n rateRecord = FakeRateRecord( failures=0 )\n success = True\n recordId = 'recordId'\n recordType = 'recordType'\n rateRecord = updateLoginRate( now, success, recordId, recordType, rateRecord )\n self.assertEqual( rateRecord.loginFailuresSinceSuccess, 0 )\n self.assertEqual( rateRecord.nextAttemptTime, now )\n self.assertFalse( rateRecord.resetFailuresOnNextAttempt )\n\n success = False\n rateRecord = updateLoginRate( now, success, recordId, recordType, rateRecord )\n self.assertEqual( rateRecord.loginFailuresSinceSuccess, 1 )\n self.assertEqual( rateRecord.nextAttemptTime, now+2 )\n self.assertFalse( rateRecord.resetFailuresOnNextAttempt )\n\n rateRecord = FakeRateRecord( failures=50 )\n success = False\n rateRecord = updateLoginRate( now, success, recordId, recordType, rateRecord )\n self.assertEqual( rateRecord.loginFailuresSinceSuccess, 51 )\n self.assertEqual( rateRecord.nextAttemptTime, now + conf.oneDaySec )\n self.assertTrue( rateRecord.resetFailuresOnNextAttempt )\n\n success = False\n rateRecord = updateLoginRate( now, success, recordId, recordType, rateRecord )\n self.assertEqual( rateRecord.loginFailuresSinceSuccess, 0 )\n self.assertEqual( rateRecord.nextAttemptTime, now )\n self.assertFalse( rateRecord.resetFailuresOnNextAttempt )\n\n\nif __name__ == '__main__':\n unittest.main()\n\n\n", "sub_path": "voterRecord.py", "file_name": "voterRecord.py", "file_ext": "py", "file_size_in_byte": 6566, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "google.appengine.ext.ndb.Model", "line_number": 22, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.ndb", "line_number": 22, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.StringProperty", "line_number": 23, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 23, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.StringProperty", "line_number": 24, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 24, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.StringProperty", "line_number": 25, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 25, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.StringProperty", "line_number": 26, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 26, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.StringProperty", "line_number": 27, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 27, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.TextProperty", "line_number": 28, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 28, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.Model", "line_number": 33, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.ndb", "line_number": 33, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.IntegerProperty", "line_number": 34, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 34, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.IntegerProperty", "line_number": 35, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 35, "usage_type": "name"}, {"api_name": "google.appengine.ext.ndb.BooleanProperty", "line_number": 36, "usage_type": "call"}, {"api_name": "google.appengine.ext.ndb", "line_number": 36, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 40, "usage_type": "call"}, {"api_name": "configOpenVoterId.const.rateRecordTypeClient", "line_number": 50, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const", "line_number": 50, "usage_type": "name"}, {"api_name": "configOpenVoterId.const.rateRecordTypeVoter", "line_number": 51, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const", "line_number": 51, "usage_type": "name"}, {"api_name": "configOpenVoterId.const.rateUseDatastore", "line_number": 53, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const", "line_number": 53, "usage_type": "name"}, {"api_name": "configOpenVoterId.const.rateUseMemcache", "line_number": 54, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const", "line_number": 54, "usage_type": "name"}, {"api_name": "configOpenVoterId.const.rateMemcacheTimeout", "line_number": 55, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const", "line_number": 55, "usage_type": "name"}, {"api_name": "configOpenVoterId.const.oneDaySec", "line_number": 55, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const.rateRecordTypeClient", "line_number": 62, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const", "line_number": 62, "usage_type": "name"}, {"api_name": "configOpenVoterId.const.isDev", "line_number": 63, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const", "line_number": 63, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 63, "usage_type": "call"}, {"api_name": "configOpenVoterId.const.rateUseDatastore", "line_number": 64, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const", "line_number": 64, "usage_type": "name"}, {"api_name": "configOpenVoterId.const.rateUseMemcache", "line_number": 64, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const.rateMemcacheTimeout", "line_number": 64, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const.rateRecordTypeVoter", "line_number": 67, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const", "line_number": 67, "usage_type": "name"}, {"api_name": "configOpenVoterId.const.isDev", "line_number": 68, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const", "line_number": 68, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 68, "usage_type": "call"}, {"api_name": "configOpenVoterId.const.rateUseDatastore", "line_number": 69, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const", "line_number": 69, "usage_type": "name"}, {"api_name": "configOpenVoterId.const.rateUseMemcache", "line_number": 69, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const.rateMemcacheTimeout", "line_number": 69, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const.rateRecordTypeClient", "line_number": 74, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const", "line_number": 74, "usage_type": "name"}, {"api_name": "configOpenVoterId.const.rateRecordTypeVoter", "line_number": 77, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const", "line_number": 77, "usage_type": "name"}, {"api_name": "configOpenVoterId.const.oneDaySec", "line_number": 97, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const", "line_number": 97, "usage_type": "name"}, {"api_name": "configOpenVoterId.const.oneDaySec", "line_number": 98, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const", "line_number": 98, "usage_type": "name"}, {"api_name": "configOpenVoterId.const.rateUseDatastore", "line_number": 105, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const", "line_number": 105, "usage_type": "name"}, {"api_name": "configOpenVoterId.const.rateUseMemcache", "line_number": 105, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const.rateMemcacheTimeout", "line_number": 105, "usage_type": "attribute"}, {"api_name": "unittest.TestCase", "line_number": 119, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const.oneDaySec", "line_number": 151, "usage_type": "attribute"}, {"api_name": "configOpenVoterId.const", "line_number": 151, "usage_type": "name"}, {"api_name": "unittest.main", "line_number": 162, "usage_type": "call"}]} +{"seq_id": "562045568", "text": "from prodnet.view import ProdnetView\nfrom prodnet.models import DBSession, Bid, Project, BidRecord\nfrom pyramid.view import view_config, view_defaults\nfrom pyramid.httpexceptions import HTTPBadRequest, HTTPOk, HTTPNotFound\nfrom prodnet.util import _, dictify\n\n\n@view_defaults(renderer='json', permission='user', route_name='project_bid')\nclass BidAPI(ProdnetView):\n def __init__(self, request):\n super().__init__(request)\n project = Project.get_by_id(self.request.matchdict['id'], True)\n if project.user_id == self.user.id or not project.auction:\n raise HTTPBadRequest()\n\n self.bid = project.auction.get_bid(self.user)\n self.bid_active = self.bid.is_active() if self.bid else True\n self.project = project\n\n @view_config(request_method='POST')\n def create(self):\n project = self.project\n\n if project.state != 'running':\n raise HTTPBadRequest()\n\n bid = self.bid\n price = float(self.request.json_body['price'])\n insured = self.request.json_body['insured']\n\n if bid:\n bid.records.insert(0, BidRecord(price=price, insured=insured))\n cause_text = \"Sie haben ihr Gebot für ${pname} auf ${price} ${curr} aktualisiert\"\n text = \"${name} hat sein Gebot für ${pname} auf ${price} ${curr} aktualisiert\"\n else:\n if not project.auction.can_bid(self.user):\n raise HTTPBadRequest()\n try:\n self.user.watched_projects.remove(project)\n except ValueError:\n pass\n bid = Bid(project.auction, self.user, price, insured)\n cause_text = \"Sie haben ein Gebot für ${pname} abgegeben: ${price} ${curr}\"\n text = \"${name} hat ein Gebot für ${pname} abgegeben: ${price} ${curr}\"\n self.bid = bid\n\n mapping = dict(pname=project.name, price=price, curr=_(\"€\"))\n self.event(self.user, _(cause_text, **mapping), project.user, seen=True)\n self.event(project.user, _(text, name=self.user.name, **mapping), self.user, project)\n DBSession.flush()\n return {\n 'bid': dictify(bid, [\n 'id',\n ('records', ['price', 'insured', 'created_at'])\n ]),\n 'project': {\n 'auction': self.project.auction.dictify(all_data=True)\n }\n }\n\n @view_config(request_method='DELETE')\n def delete(self):\n if not self.bid:\n raise HTTPNotFound()\n project = self.project\n if project.state not in ['running', 'frozen']:\n raise HTTPBadRequest()\n\n DBSession.delete(self.bid)\n self.user.watched_projects.append(project)\n\n cause_text = \"Sie haben ihr Gebot für ${pname} zurückgezogen\"\n text = \"${name} hat sein Gebot für ${pname} zurückgezogen\"\n self.event(self.user, _(cause_text, pname=project.name), project.user, project, seen=True)\n self.event(project.user, _(text, name=self.user.name, pname=project.name), self.user, project)\n return {\n 'bid': None,\n 'project': {\n 'auction': self.project.auction.dictify(all_data=True)\n }\n }\n\n @view_config(request_method='POST', route_name='project_bid_accept_mods')\n def accept_mods(self):\n if self.project.state != 'running':\n raise HTTPBadRequest()\n if not self.bid:\n raise HTTPNotFound()\n self.bid.accept_mods()\n return HTTPOk()", "sub_path": "prodnet/views/api/project/bid.py", "file_name": "bid.py", "file_ext": "py", "file_size_in_byte": 3520, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "prodnet.view.ProdnetView", "line_number": 9, "usage_type": "name"}, {"api_name": "prodnet.models.Project.get_by_id", "line_number": 12, "usage_type": "call"}, {"api_name": "prodnet.models.Project", "line_number": 12, "usage_type": "name"}, {"api_name": "pyramid.httpexceptions.HTTPBadRequest", "line_number": 14, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions.HTTPBadRequest", "line_number": 25, "usage_type": "call"}, {"api_name": "prodnet.models.BidRecord", "line_number": 32, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions.HTTPBadRequest", "line_number": 37, "usage_type": "call"}, {"api_name": "prodnet.models.Bid", "line_number": 42, "usage_type": "call"}, {"api_name": "prodnet.util._", "line_number": 47, "usage_type": "call"}, {"api_name": "prodnet.util._", "line_number": 48, "usage_type": "call"}, {"api_name": "prodnet.util._", "line_number": 49, "usage_type": "call"}, {"api_name": "prodnet.models.DBSession.flush", "line_number": 50, "usage_type": "call"}, {"api_name": "prodnet.models.DBSession", "line_number": 50, "usage_type": "name"}, {"api_name": "prodnet.util.dictify", "line_number": 52, "usage_type": "call"}, {"api_name": "pyramid.view.view_config", "line_number": 20, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions.HTTPNotFound", "line_number": 64, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions.HTTPBadRequest", "line_number": 67, "usage_type": "call"}, {"api_name": "prodnet.models.DBSession.delete", "line_number": 69, "usage_type": "call"}, {"api_name": "prodnet.models.DBSession", "line_number": 69, "usage_type": "name"}, {"api_name": "prodnet.util._", "line_number": 74, "usage_type": "call"}, {"api_name": "prodnet.util._", "line_number": 75, "usage_type": "call"}, {"api_name": "pyramid.view.view_config", "line_number": 61, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions.HTTPBadRequest", "line_number": 86, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions.HTTPNotFound", "line_number": 88, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions.HTTPOk", "line_number": 90, "usage_type": "call"}, {"api_name": "pyramid.view.view_config", "line_number": 83, "usage_type": "call"}, {"api_name": "pyramid.view.view_defaults", "line_number": 8, "usage_type": "call"}]} +{"seq_id": "258219933", "text": "# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n# Import\n# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n\n# Import Standard libraries\nimport tensorflow as tf\nimport numpy as np \nimport matplotlib.cm as cm\nfrom IPython.display import Image, display\n\n\n# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n# Functions\n# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\ndef save_and_display_gradcam(img, heatmap,idx, cam_path=\"cam.jpg\", alpha=0.4):\n # Load the original image\n #img = tf.keras.preprocessing.image.load_img(img_path)\n img = tf.keras.preprocessing.image.img_to_array(img)\n\n # Rescale heatmap to a range 0-255\n heatmap = np.uint8(255 * heatmap)\n\n # Use jet colormap to colorize heatmap\n jet = cm.get_cmap(\"jet\")\n\n # Use RGB values of the colormap\n jet_colors = jet(np.arange(256))[:, :3]\n jet_heatmap = jet_colors[heatmap]\n\n # Create an image with RGB colorized heatmap\n jet_heatmap = tf.keras.preprocessing.image.array_to_img(jet_heatmap)\n jet_heatmap = jet_heatmap.resize((img.shape[1], img.shape[0]))\n jet_heatmap = tf.keras.preprocessing.image.img_to_array(jet_heatmap)\n\n # Superimpose the heatmap on original image\n superimposed_img = jet_heatmap * alpha + img\n superimposed_img = tf.keras.preprocessing.image.array_to_img(superimposed_img)\n\n # Save the superimposed image\n superimposed_img.save(\"cam_\" + str(idx) + \".jpg\")\n\n # Display Grad CAM\n #display(Image(cam_path))\n\n\ndef make_gradcam_heatmap(\n model\n , image\n , pred_index = None\n):\n # Check for the latest convolution layer with a 4D output\n for layer in reversed(model.layers):\n if len(layer.output_shape) == 4:\n last_conv_layer_name = layer.name\n break\n\n # First, we create a model that maps the input image to the activations of the last conv layer as well as the output\n # predictions\n grad_model = tf.keras.models.Model(\n [ model.inputs ]\n , [ model.get_layer(last_conv_layer_name).output, model.output ]\n )\n\n # Then, we compute the gradient of the top predicted class for our input image with respect to the activations of\n # the last conv layer\n with tf.GradientTape() as tape:\n last_conv_layer_output, preds = grad_model(image)\n if pred_index is None:\n pred_index = tf.argmax(preds[0])\n class_channel = preds[:, pred_index]\n\n # This is the gradient of the output neuron (top predicted or chosen) with regard to the output feature map of the\n # last conv layer\n grads = tape.gradient(class_channel, last_conv_layer_output)\n\n # This is a vector where each entry is the mean intensity of the gradient over a specific feature map channel\n pooled_grads = tf.reduce_mean(grads, axis=(0, 1, 2))\n\n # We multiply each channel in the feature map array by \"how important this channel is\" with regard to the top\n # predicted class then sum all the channels to obtain the heatmap class activation\n last_conv_layer_output = last_conv_layer_output[0]\n heatmap = last_conv_layer_output @ pooled_grads[..., tf.newaxis]\n heatmap = tf.squeeze(heatmap)\n\n # For visualization purpose, we will also normalize the heatmap between 0 & 1\n heatmap = tf.maximum(heatmap, 0) / tf.math.reduce_max(heatmap)\n return heatmap.numpy()\n", "sub_path": "streamlit/model/interpretability.py", "file_name": "interpretability.py", "file_ext": "py", "file_size_in_byte": 3262, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "tensorflow.keras.preprocessing.image.img_to_array", "line_number": 18, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 18, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.cm.get_cmap", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.cm", "line_number": 24, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 27, "usage_type": "call"}, {"api_name": "tensorflow.keras.preprocessing.image.array_to_img", "line_number": 31, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 31, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.preprocessing.image.img_to_array", "line_number": 33, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 33, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.preprocessing.image.array_to_img", "line_number": 37, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 37, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.models.Model", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 59, "usage_type": "attribute"}, {"api_name": "tensorflow.GradientTape", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.argmax", "line_number": 69, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 77, "usage_type": "call"}, {"api_name": "tensorflow.newaxis", "line_number": 82, "usage_type": "attribute"}, {"api_name": "tensorflow.squeeze", "line_number": 83, "usage_type": "call"}, {"api_name": "tensorflow.maximum", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorflow.math.reduce_max", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 86, "usage_type": "attribute"}]} +{"seq_id": "458874653", "text": "import tokenize\nfrom collections import defaultdict, OrderedDict\nimport six\n\n\nSTATE_LABEL = 256\nops = set(['&=', '<<', '<=', '==', '[', '//=', '%=', '^', ']',\n '~', '//', '|=', '-=', '}', '^=', '!=', '>=', '>>',\n '<', '**=', '>>=', '*=', '+=', ';', ':', '/=', '<<=',\n '>', '=', '|', '{', '**', '@', '&', '%', '+', '*',\n ')', '(', '.', '-', ',', '<>', '...', '->'])\nnormal_tks = set(['NAME', 'STRING', 'NUMBER', 'INDENT',\n 'DEDENT', 'NEWLINE', 'ENDMARKER'])\n\nclass Label(object):\n __slots__ = 'type', 'val'\n\n def __init__(self, type, val=None):\n self.type = type\n self.val = val\n\n @classmethod\n def get_label(cls, label):\n if label[0] == \"'\":\n if label[-1] != \"'\" or len(label) < 3:\n raise Exception('invalid label: %r' % label)\n label = label[1:-1]\n if label in ops:\n return cls(tokenize.OP, label)\n return cls(tokenize.NAME, label)\n elif label in normal_tks:\n return cls(getattr(tokenize, label))\n\n\nclass State(object):\n __slots__ = 'is_final', 'name', 'symbol', 'arcs', 'bootstrap'\n\n def __init__(self, is_final, name=None, symbol=None):\n self.is_final = is_final\n self.name = name\n self.symbol = symbol\n self.arcs = {}\n\n def __repr__(self):\n if self.name is not None:\n return '' % self.name\n return super(State, self).__repr__()\n\n def arc(self, label, state):\n if label in self.arcs:\n raise Exception('duplicated arc')\n self.arcs[label] = state\n\n def build_bootstrap(self):\n if not hasattr(self, 'bootstrap'):\n self.bootstrap = defaultdict(lambda: {})\n for label, state in self.arcs.items():\n if label.type == STATE_LABEL:\n if label.val == self:\n continue\n for t, vals in label.val.build_bootstrap().items():\n for val in vals.keys():\n if val in self.bootstrap[t]:\n raise Exception('duplicated bootstrap')\n self.bootstrap[t][val] = (label.val, state)\n else:\n self.bootstrap[label.type][label.val] = (None, state)\n return self.bootstrap\n\n def generate(self, name, ids):\n myid = ids[self]\n varname = '%s%d' % (name, myid)\n if myid == 0:\n yield '%s = %s\\n' % (varname, self.name)\n else:\n yield '%s = State(%r)\\n' % (varname, self.is_final)\n\n for label, state in self.arcs.items():\n if state not in ids:\n yield from state.generate(name, ids)\n if label.type == STATE_LABEL:\n yield '%s.arc(Label(%d, %s), %s%d)\\n' % (varname,\n label.type, label.val.name, name, ids[state])\n else:\n yield '%s.arc(Label(%d, %r), %s%d)\\n' % (varname,\n label.type, label.val, name, ids[state])\n yield '%s.bootstrap = {' % varname\n for t, vals in self.bootstrap.items():\n yield '\\n %d: {' % t\n for v, (st1, st2) in vals.items():\n yield '\\n %r: (' % v\n if st1 is None:\n yield 'None'\n else:\n yield '%s' % st1.name\n yield ', %s%d),' % (name, ids[st2])\n yield '},'\n yield '}\\n'\n\n\nclass IDs(object):\n def __init__(self):\n self.inc = 0\n self.m = {}\n\n def __getitem__(self, k):\n i = self.m.get(k, None)\n if i is None:\n i = self.inc\n self.inc += 1\n self.m[k] = i\n return i\n\n def __contains__(self, k):\n return k in self.m\n\n\nclass States(object):\n def __init__(self, **states):\n self.all_states = []\n self.states = OrderedDict()\n self.symbols = {} # name -> id\n self.inc = STATE_LABEL\n self.reverse_symbol = {} # id -> name\n for name, state in states.items():\n self[name] = state\n\n def __setitem__(self, name, state):\n self.inc += 1\n self.symbols[name] = self.inc\n self.reverse_symbol[self.inc] = name\n self.states[name] = state\n stack = [state]\n states = set([state])\n while stack:\n st = stack[0]\n stack = stack[1:]\n for s in st.arcs.values():\n if s not in states:\n self.all_states.append(s)\n stack.append(s)\n states.add(s)\n self.all_states.append(state)\n\n def __getitem__(self, k):\n if isinstance(k, int):\n k = self.reverse_symbol[k]\n return self.states[k]\n\n def __iter__(self):\n yield from self.states\n\n def keys(self):\n yield from self.states.keys()\n\n def items(self):\n yield from self.states.items()\n\n def from_dfas(self, dfas):\n states = {}\n def dfa2state(dfa):\n state = states.get(dfa, None)\n if state is not None:\n return state\n state = State(dfa.is_final)\n states[dfa] = state\n for name, d in dfa.arcs.items():\n label = Label.get_label(name)\n if label is None:\n label = Label(STATE_LABEL,\n dfa2state(dfas[name][0]))\n state.arc(label, dfa2state(d))\n return state\n for name, dfa in dfas.items():\n dfa2state(dfa[0])\n for name, dfa in dfas.items():\n state = states[dfa[0]]\n state.name = name\n self[name] = state\n\n def build_bootstrap(self):\n for state in self.all_states:\n state.build_bootstrap()\n\n def generate(self):\n buf = six.StringIO()\n buf.write('# generated by cpy.parser.state.States\\n')\n buf.write('from .grammar.state import State, Label\\n')\n buf.write('from .grammar.symbols import Symbols\\n\\n\\n')\n\n buf.write('class _Symbols(Symbols):\\n')\n for i, name in enumerate(self.states.keys()):\n buf.write(' %s = %d\\n' % (name, i + STATE_LABEL + 1))\n buf.write('symbols = _Symbols()\\n\\n\\n')\n\n for i, (name, state) in enumerate(self.states.items()):\n buf.write('%s = State(%r, %r, %d)\\n' % (\n name, state.is_final, name, STATE_LABEL + i + 1))\n for name, state in self.states.items():\n buf.write('\\n\\n')\n buf.write(''.join(state.generate(name, IDs())))\n return buf.getvalue()\n", "sub_path": "cpy/parser/grammar/state.py", "file_name": "state.py", "file_ext": "py", "file_size_in_byte": 6645, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "tokenize.OP", "line_number": 29, "usage_type": "attribute"}, {"api_name": "tokenize.NAME", "line_number": 30, "usage_type": "attribute"}, {"api_name": "collections.defaultdict", "line_number": 56, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 121, "usage_type": "call"}, {"api_name": "six.StringIO", "line_number": 186, "usage_type": "call"}]} +{"seq_id": "38512234", "text": "import numpy as np\nimport scipy.optimize as opt\nfrom lrCostFunction import lrCostFunction\n\n# ONEVSALL trains multiple logistic regression classifiers and returns all\n# the classifiers in a matrix all_theta, where the i-th row of all_theta \n# corresponds to the classifier for label i\n# [all_theta] = ONEVSALL(X, y, num_labels, lambda) trains num_labels\n# logistic regression classifiers and returns each of these classifiers\n# in a matrix all_theta, where the i-th row of all_theta corresponds \n# to the classifier for label i\n\ndef oneVsAll(X, y, num_labels, l):\n# Some useful variables\n m, n = X.shape\n all_theta = np.zeros((num_labels, n + 1))\n X = np.hstack((np.ones((m, 1)), X))\n initial_theta = np.zeros(n + 1)\n\n # ====================== YOUR CODE HERE ======================\n # Instructions: You should complete the following code to train num_labels\n # logistic regression classifiers with regularization\n # parameter lambda.\n #\n # Hint: theta(:) will return a column vector.\n #\n # Hint: You can use y == c to obtain a vector of 1's and 0's that tell you\n # whether the ground truth is true/false for this class.\n #\n # Note: For this assignment, we recommend using fmincg to optimize the cost\n # function. It is okay to use a for-loop (for c = 1:num_labels) to\n # loop over the different classes.\n #\n # fmincg works similarly to fminunc, but is more efficient when we\n # are dealing with large number of parameters.\n\n for i in range(0, num_labels):\n label = 10 if i == 0 else i\n result = opt.minimize(fun=lrCostFunction, x0=initial_theta, args=(X, (y == label).astype(int), l), method='TNC', jac=True)\n all_theta[i, :] = result.x\n\n\n return all_theta\n", "sub_path": "3. Neural Networks Representation/oneVsAll.py", "file_name": "oneVsAll.py", "file_ext": "py", "file_size_in_byte": 1827, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "numpy.zeros", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 18, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 39, "usage_type": "call"}, {"api_name": "scipy.optimize", "line_number": 39, "usage_type": "name"}, {"api_name": "lrCostFunction.lrCostFunction", "line_number": 39, "usage_type": "name"}]} +{"seq_id": "163709940", "text": "\"\"\"\nCopyright 2021 The Magma Authors.\n\nThis source code is licensed under the BSD-style license found in the\nLICENSE file in the root directory of this source tree.\n\nUnless required by applicable law or agreed to in writing, software\ndistributed under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for the specific language governing permissions and\nlimitations under the License.\n\"\"\"\nimport os\nimport sys\nimport json\n\nimport click\nfrom boto3 import Session\n\nfrom .common import (\n run_command,\n run_playbook,\n print_error_msg,\n print_warning_msg,\n print_success_msg)\n\ndef setup_aws_creds():\n session = Session()\n creds = session.get_credentials()\n if not creds:\n print_error_msg('''\nAWS credentials not configured.\nconfigure through awscli or through orcl\norcl configure set -k aws_access_key_id -v \norcl configure set -k aws_secret_access_key -v \norcl configure set -k region -v \n''')\n sys.exit(1)\n\n frozen_creds = creds.get_frozen_credentials()\n os.environ[\"AWS_ACCESS_KEY_ID\"] = frozen_creds.access_key\n os.environ[\"AWS_SECRET_ACCESS_KEY\"] = frozen_creds.secret_key\n\n\n@click.group(invoke_without_command=True)\n@click.pass_context\ndef cleanup(ctx):\n \"\"\"\n Removes resources deployed for orc8r\n \"\"\"\n tf_destroy = [ \"terraform\", \"destroy\", \"-auto-approve\"]\n\n if ctx.invoked_subcommand is None:\n cmd = \" \".join(tf_destroy)\n click.echo(f\"Following commands will be run during cleanup\\n{cmd}\")\n click.confirm('Do you want to continue with cleanup?', abort=True)\n click.echo(f\"Running {cmd}\")\n rc = run_command(tf_destroy)\n if rc != 0:\n print_error_msg(\"Destroy Failed!!! Attempt cleaning up individual resources using 'orcl cleanup raw' subcommand\")\n return\n\n@cleanup.command()\n@click.pass_context\n@click.option('--dryrun', default=False, is_flag=True, help='Show resources to be cleaned up during raw cleanup')\n@click.option('--state', help='Provide state file containing resource information e.g. terraform.tfstate or terraform.tfstate.backup')\n@click.option('--values', multiple=True, help='Key value pairs. for e.g. cluster_name,orc8r. Can be used multiple times')\ndef raw(ctx, dryrun, state, values):\n \"\"\"\n Individually cleans up resources deployed for orc8r\n \"\"\"\n click.confirm(click.style('This is irreversable!! Do you want to continue with cleanup?', fg='red'), abort=True)\n if state:\n ctx.obj['cleanup_state'] = state\n\n # add additional items\n for config_items in values:\n k, v = config_items.split(\",\")\n ctx.obj[k] = v\n\n extra_vars = json.dumps(ctx.obj)\n cleanup_playbook = \"%s/cleanup.yml\" % ctx.obj[\"playbooks\"]\n playbook_args = [ \"ansible-playbook\", \"-v\", \"-e\", extra_vars]\n\n # Few boto dependent modules in ansible require these values to be\n # setup as environment variables. Hence setting these up.\n setup_aws_creds()\n\n if dryrun:\n tag_args = [\"-t\", \"cleanup_dryrun\"]\n else:\n tag_args = [\"-t\", \"cleanup\"]\n\n rc = run_playbook(playbook_args + tag_args + [cleanup_playbook])\n if rc != 0:\n print_error_msg(\"Failed cleaning up resources!!!\")\n sys.exit(1)\n print_success_msg(\"Successfully cleaned up underlying resources\")", "sub_path": "orc8r/cloud/deploy/orc8r_deployer/docker/root/scripts/cli/cleanup.py", "file_name": "cleanup.py", "file_ext": "py", "file_size_in_byte": 3396, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "boto3.Session", "line_number": 28, "usage_type": "call"}, {"api_name": "common.print_error_msg", "line_number": 31, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 38, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 42, "usage_type": "attribute"}, {"api_name": "click.echo", "line_number": 55, "usage_type": "call"}, {"api_name": "click.confirm", "line_number": 56, "usage_type": "call"}, {"api_name": "click.echo", "line_number": 57, "usage_type": "call"}, {"api_name": "common.run_command", "line_number": 58, "usage_type": "call"}, {"api_name": "common.print_error_msg", "line_number": 60, "usage_type": "call"}, {"api_name": "click.group", "line_number": 45, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 46, "usage_type": "attribute"}, {"api_name": "click.confirm", "line_number": 72, "usage_type": "call"}, {"api_name": "click.style", "line_number": 72, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 81, "usage_type": "call"}, {"api_name": "common.run_playbook", "line_number": 94, "usage_type": "call"}, {"api_name": "common.print_error_msg", "line_number": 96, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 97, "usage_type": "call"}, {"api_name": "common.print_success_msg", "line_number": 98, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 64, "usage_type": "attribute"}, {"api_name": "click.option", "line_number": 65, "usage_type": "call"}, {"api_name": "click.option", "line_number": 66, "usage_type": "call"}, {"api_name": "click.option", "line_number": 67, "usage_type": "call"}]} +{"seq_id": "124128885", "text": "import time\nimport random\nimport string\nimport serial.tools.list_ports\n\nfor port in serial.tools.list_ports.comports():\n if port.vid == 6790 and port.pid == 29987:\n try:\n print(\"Found \" + port.device)\n ser = serial.Serial(port.device, 115200) # open serial port\n app_eui = ''.join(random.SystemRandom().choice(string.hexdigits) for _ in range(16))\n app_key = ''.join(random.SystemRandom().choice(string.hexdigits) for _ in range(32))\n print(\"App EUI will be \" + app_eui)\n print(\"App Key will be \" + app_key)\n print(ser.name) # check which port was really used\n cmd = 'at+set_config=app_eui:' + app_eui + '&app_key:' + app_key + '\\r\\n'\n ser.write(cmd.encode('ASCII'))\n time.sleep(5)\n ser.close()\n print(\"Initialized \" + port.device)\n except:\n print(\"Cannot initialize \" + port.device)\n\n", "sub_path": "install/scripts/reset_loranode.py", "file_name": "reset_loranode.py", "file_ext": "py", "file_size_in_byte": 954, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "serial.tools.list_ports.tools.list_ports.comports", "line_number": 6, "usage_type": "call"}, {"api_name": "serial.tools.list_ports.tools", "line_number": 6, "usage_type": "attribute"}, {"api_name": "serial.tools.list_ports", "line_number": 6, "usage_type": "name"}, {"api_name": "serial.tools.list_ports.Serial", "line_number": 10, "usage_type": "call"}, {"api_name": "serial.tools.list_ports", "line_number": 10, "usage_type": "name"}, {"api_name": "random.SystemRandom", "line_number": 11, "usage_type": "call"}, {"api_name": "string.hexdigits", "line_number": 11, "usage_type": "attribute"}, {"api_name": "random.SystemRandom", "line_number": 12, "usage_type": "call"}, {"api_name": "string.hexdigits", "line_number": 12, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 18, "usage_type": "call"}]} +{"seq_id": "182008044", "text": "#!/usr/bin/python3\n\nimport sqlite3\nimport re\nimport unicodedata\n\nfrom bs4 import BeautifulSoup\nfrom pprint import pprint\nfrom urllib.request import Request, urlopen\nfrom urllib.error import URLError, HTTPError\n\n\nclass amd_fx_cpus_mining:\n def parser(self, url, pn_list):\n\n conn = sqlite3.connect('components.db')\n c = conn.cursor()\n\n user_agent = 'Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:43.0'\n\n try:\n q = Request(url)\n q.add_header('User-Agent', user_agent)\n html = urlopen(q).read()\n except HTTPError as e:\n print('The server couldn\\'t fulfill the request.')\n print('Error code: ', e.code)\n print('Error message: ', e.msg)\n except URLError as e:\n print('We failed to reach a server.')\n print('Reason: ', e.reason)\n else:\n soup = BeautifulSoup(html, \"html.parser\")\n for pn in pn_list:\n sliced_pn = pn[0:9]\n pn_element = soup.find(text=sliced_pn) # in BeatifulSoup 4.4 text became string\n if pn_element is None:\n for column in soup.find_all('td'): # BeautifulSoup won't find text inside
tag\n if sliced_pn in column.get_text():\n pn_element = column # if column with the given pn was found, take the whole element for\n # research\n # try-except block in case cpu was not found in wikipedia\n try:\n cpu_row = pn_element.find_parent('tr')\n except AttributeError:\n c.execute(\"UPDATE cpu SET core_clock = ?, no_of_cores = ?, l2_l3_cache = ?, \"\n \"tdp = ? WHERE pn = ?\", (-1, -1, -1, -1, pn))\n continue\n\n # Declaring variables for the new cpu and the db\n no_of_cores = -1\n no_of_cores_found = False\n core_clock = -1\n core_clock_found = False\n l_cache = -1\n l2_cache_found = False\n tdp = -1\n tdp_found = False\n unicoded_space = unicodedata.normalize(\"NFKD\", '\\xa0')\n i = 0\n\n while not (no_of_cores_found is True and core_clock_found is True and l2_cache_found is True\n and tdp_found is True):\n # If cell was merged, try to find the previous cell that has the desired info\n if i != 0:\n cpu_row = cpu_row.find_previous('tr')\n\n for i, column in enumerate(cpu_row.find_all('td')):\n if pn == 'FD837EWMHKBOX':\n core_clock = 3.3\n core_clock_found = True\n elif pn == 'FD9590FHHKWOF':\n core_clock = 4.7\n core_clock_found = True\n\n # cpu_row is the row of our cpu and main_cpu_row is the row of the info that is shared\n if i == 1 and re.search(r'[0-9][0-9]?/[0-9][0-9]?',\n column.get_text()) and no_of_cores_found is False:\n no_of_cores = column.get_text()\n if no_of_cores.find('/') != -1:\n index = no_of_cores.find('/')\n no_of_cores = no_of_cores[0:index]\n no_of_cores = int(no_of_cores)\n no_of_cores_found = True\n elif 'Hz' in column.get_text() and (\n 'V)' not in column.get_text() or i == 7) and core_clock_found is False:\n core_clock = column.get_text()\n index = core_clock.find('GHz')\n core_clock = core_clock[0:index]\n # Fix the code of spaces (latin-1) for the sqlite3 db\n core_clock = core_clock.replace('\\xa0', unicoded_space)\n core_clock = core_clock.strip()\n core_clock = float(core_clock)\n core_clock_found = True\n elif ('MB' in column.get_text() or 'kB' in column.get_text()) and l2_cache_found is False:\n l2_cache = column.get_text()\n # Fix the code of spaces (latin-1) for the sqlite3 db\n l2_cache = l2_cache.replace('\\xa0', unicoded_space)\n if 'MB' in l2_cache:\n l2_cache = l2_cache.replace('MB', '').strip()\n if '×' in l2_cache:\n index = l2_cache.find('×')\n element1 = int(l2_cache[0:index])\n element2 = int(l2_cache[index + 1:])\n l2_cache = element1 * element2 * 1024\n else:\n l2_cache = int(l2_cache) * 1024\n elif 'kB' in l2_cache:\n l2_cache = l2_cache.replace('kB', '')\n if '×' in l2_cache:\n index = l2_cache.find('×')\n element1 = int(l2_cache[0:index])\n element2 = int(l2_cache[index + 1:])\n l2_cache = element1 * element2\n else:\n l2_cache = int(l2_cache)\n l2_cache_found = True\n # Try to find l3 cache in the next td\n l3_cache = 0\n if 'MB' in cpu_row.find_all('td')[i + 1].get_text() or 'KB' in cpu_row.find_all('td')[\n i + 1].get_text():\n l3_cache = cpu_row.find_all('td')[i + 1].get_text()\n # Fix the code of spaces (latin-1) for the sqlite3 db\n l3_cache = l3_cache.replace('\\xa0', unicoded_space)\n if 'MB' in l3_cache:\n l3_cache = l3_cache.replace('MB', '').strip()\n if '×' in l3_cache:\n index = l3_cache.find('×')\n element1 = int(l3_cache[0:index])\n element2 = int(l3_cache[index + 1:])\n l3_cache = element1 * element2 * 1024\n else:\n l3_cache = int(l3_cache) * 1024\n elif 'KB' in l3_cache:\n l3_cache = l3_cache.replace('KB', '')\n if '×' in l3_cache:\n index = l3_cache.find('×')\n element1 = int(l3_cache[0:index])\n element2 = int(l3_cache[index + 1:])\n l3_cache = element1 * element2\n else:\n l3_cache = int(l2_cache)\n l_cache = l2_cache + l3_cache\n elif re.search(r'^[0-9]{1,4}\\s*W$', column.get_text()) and tdp_found is False:\n tdp = column.get_text()\n tdp = tdp.replace('W', '').strip()\n # Fix the code of spaces (latin-1) for the sqlite3 db\n tdp = tdp.replace('\\xa0', unicoded_space)\n tdp = int(tdp)\n tdp_found = True\n\n c.execute(\"UPDATE cpu SET core_clock = ?, no_of_cores = ?, l2_l3_cache = ?, \"\n \"tdp = ? WHERE pn = ?\", (core_clock, no_of_cores, l_cache, tdp, pn))\n\n conn.commit()\n conn.close()\n\n return\n", "sub_path": "amd_fx_cpus_mining.py", "file_name": "amd_fx_cpus_mining.py", "file_ext": "py", "file_size_in_byte": 8315, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "sqlite3.connect", "line_number": 16, "usage_type": "call"}, {"api_name": "urllib.request.Request", "line_number": 22, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 24, "usage_type": "call"}, {"api_name": "urllib.error.HTTPError", "line_number": 25, "usage_type": "name"}, {"api_name": "urllib.error.URLError", "line_number": 29, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 33, "usage_type": "call"}, {"api_name": "unicodedata.normalize", "line_number": 59, "usage_type": "call"}, {"api_name": "re.search", "line_number": 77, "usage_type": "call"}, {"api_name": "re.search", "line_number": 144, "usage_type": "call"}]} +{"seq_id": "8626824", "text": "from __future__ import print_function\n\nimport argparse\nimport sys\n\nimport onthego.download\nimport onthego.spotify.auth\n\n\ndef download_playlist():\n parser = argparse.ArgumentParser(description=\"Download the tracks of a Spotify playlist from YouTube\")\n parser.add_argument(\"-S\", \"--no-skip\", action='store_true',\n help=\"Don't skip files that were already downloaded.\")\n parser.add_argument(\"-C\", \"--no-convert\", action='store_true',\n help=\"Don't convert audio files to mp3 format.\")\n parser.add_argument(\"playlist\", help=\"Name of playlist. E.g: 'Road music'\")\n parser.add_argument(\"dst\", help=\"Destination directory\")\n args = parser.parse_args()\n\n spotify_client = onthego.spotify.auth.Client()\n try:\n for track in spotify_client.iter_tracks(args.playlist):\n onthego.download.audio(track, args.dst,skip_existing=(not args.no_skip), convert_to_mp3=(not args.no_convert))\n except onthego.spotify.auth.PlaylistNotFound as e:\n print(\"Playlist '%s' was not found. Did you type its name correctly?\" % e.playlist_name)\n sys.exit(1)\n", "sub_path": "onthego/scripts/cli.py", "file_name": "cli.py", "file_ext": "py", "file_size_in_byte": 1104, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 11, "usage_type": "call"}, {"api_name": "onthego.download.spotify.auth.Client", "line_number": 20, "usage_type": "call"}, {"api_name": "onthego.download.spotify", "line_number": 20, "usage_type": "attribute"}, {"api_name": "onthego.download", "line_number": 20, "usage_type": "name"}, {"api_name": "onthego.download.download.audio", "line_number": 23, "usage_type": "call"}, {"api_name": "onthego.download.download", "line_number": 23, "usage_type": "attribute"}, {"api_name": "onthego.download", "line_number": 23, "usage_type": "name"}, {"api_name": "onthego.download.spotify", "line_number": 24, "usage_type": "attribute"}, {"api_name": "onthego.download", "line_number": 24, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 26, "usage_type": "call"}]} +{"seq_id": "644332668", "text": "from __future__ import print_function\nimport numpy as np\nimport preprocessing as proc\nimport argparse\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torch.optim as optim\nfrom torch.autograd import Variable\n\n# Training settings\nparser = argparse.ArgumentParser(description='BASE Model')\nparser.add_argument('--batch-size', type=int, default=64, metavar='N',\n help='input batch size for training (default: 64)')\nparser.add_argument('--test-batch-size', type=int, default=1000, metavar='N',\n help='input batch size for testing (default: 1000)')\nparser.add_argument('--epochs', type=int, default=20, metavar='N',\n help='number of epochs to train (default: 20)')\nparser.add_argument('--lr', type=float, default=0.01, metavar='LR',\n help='learning rate (default: 0.01)')\nparser.add_argument('--momentum', type=float, default=0.5, metavar='M',\n help='SGD momentum (default: 0.5)')\nparser.add_argument('--no-cuda', action='store_true', default=True,\n help='disables CUDA training')\nparser.add_argument('--seed', type=int, default=0, metavar='S',\n help='random seed (default: 0)')\nparser.add_argument('--log-interval', type=int, default=10, metavar='N',\n help='how many batches to wait before logging training status')\nargs = parser.parse_args()\nargs.cuda = not args.no_cuda and torch.cuda.is_available()\n\ntorch.manual_seed(args.seed)\nif args.cuda:\n torch.cuda.manual_seed(args.seed)\n\n\nclass Net(nn.Module):\n def __init__(self):\n super(Net, self).__init__()\n self.conv1 = nn.Conv2d(1, 20, kernel_size=5, stride=1, padding=0)\n self.conv2 = nn.Conv2d(20, 50, kernel_size=5, stride=1, padding=0)\n self.fc1 = nn.Linear(4*4*50, 500)\n self.fc2 = nn.Linear(500, 10)\n self.bn1 = nn.BatchNorm2d(20)\n self.bn2 = nn.BatchNorm2d(50)\n self.bn3 = nn.BatchNorm1d(500)\n self.drop = nn.Dropout(p=0.5)\n\n def forward(self, x):\n x = self.conv1(x)\n x = F.max_pool2d(self.bn1(x), 2)\n x = self.conv2(x)\n x = F.max_pool2d(self.bn2(x), 2)\n x = x.view(-1, 4*4*50)\n x = self.fc1(x)\n x = F.relu(self.bn3(x))\n x = self.fc2(x)\n x = self.drop(x)\n return F.log_softmax(x)\n\nmodel = Net()\n# print(model)\nif args.cuda:\n model.cuda()\n\noptimizer = optim.SGD(model.parameters(), lr=args.lr, momentum=args.momentum)\n\n\ndef generate_data(data, label, batchSize, data_type='train', shuffle=True):\n assert batchSize > 0\n data_len = data.shape[0]\n total_batch = data_len / batchSize + (1 if data_len % batchSize != 0 else 0)\n if shuffle:\n indices = np.random.permutation(data_len)\n data = data[indices]\n label = label[indices]\n for idx in range(total_batch):\n start = idx * batchSize\n end = min((idx + 1) * batchSize, data_len)\n if data_type == 'train':\n yield proc.Normalize(data[start:end], (proc.TRAIN_MEAN,)*(end-start),\n (proc.TRAIN_STD,)*(end-start)), label[start:end]\n else:\n yield proc.Normalize(data[start:end], (proc.TRAIN_MEAN,)*(end-start),\n (proc.TRAIN_STD,)*(end-start)), label[start:end]\n\n\ndef train(epoch, train_data, train_labels, use_data_len=10000):\n model.train() # set to training mode\n batch_idx = 1\n for (_data, _target) in generate_data(train_data[:use_data_len], train_labels[:use_data_len], batchSize=args.batch_size, shuffle=True):\n data = torch.from_numpy(_data)\n target = torch.from_numpy(_target).long()\n if args.cuda:\n data, target = data.cuda(), target.cuda()\n data, target = Variable(data), Variable(target)\n # print(data.size(), target.size())\n optimizer.zero_grad()\n output = model.forward(data)\n # print(output.size())\n loss = F.nll_loss(output, target)\n loss.backward()\n optimizer.step()\n if batch_idx % args.log_interval == 0:\n print('Train Epoch: {} [{:5d}/{} ({:2d}%)]\\tLoss: {:.6f}'.format(\n epoch, batch_idx * len(data), use_data_len,\n int(100. * batch_idx * len(data) / use_data_len), loss.data[0]))\n batch_idx += 1\n\n\ndef test(test_data, test_labels):\n model.eval() # set to evaluation mode\n test_loss = 0\n correct = 0\n for (data, target) in generate_data(test_data, test_labels,\n batchSize=args.batch_size, shuffle=True):\n data = torch.from_numpy(data)\n target = torch.from_numpy(target).long()\n if args.cuda:\n data, target = data.cuda(), target.cuda()\n data, target = Variable(data, volatile=True), Variable(target)\n output = model.forward(data)\n test_loss += F.nll_loss(output, target).data[0]\n pred = output.data.max(1)[1] # get the index of the max log-probability\n correct += pred.eq(target.data).cpu().sum()\n\n test_loss = test_loss\n test_loss /= test_data.shape[0] # loss function already averages over batch size\n print('\\nTest set: Average loss: {:.4f}, Accuracy: {}/{} ({:.1f}%)\\n'.format(\n test_loss, correct, test_data.shape[0],\n 100. * correct / test_data.shape[0]))\n\n\ndef go():\n train_images, train_labels = proc.get_data(\"train\")\n test_images, test_labels = proc.get_data(\"test\")\n for epoch in range(1, args.epochs + 1):\n train(epoch, train_images, train_labels, 10000)\n test(test_images, test_labels)\n\ngo()\n", "sub_path": "CNN_for_Mnist/batchnormalization.py", "file_name": "batchnormalization.py", "file_ext": "py", "file_size_in_byte": 5629, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 12, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 30, "usage_type": "attribute"}, {"api_name": "torch.manual_seed", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.cuda.manual_seed", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 34, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 37, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 37, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 40, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 41, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 42, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 43, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 44, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 45, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 46, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 47, "usage_type": "name"}, {"api_name": "torch.nn.functional.max_pool2d", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 51, "usage_type": "name"}, {"api_name": "torch.nn.functional.max_pool2d", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 53, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 56, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 56, "usage_type": "name"}, {"api_name": "torch.nn.functional.log_softmax", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 59, "usage_type": "name"}, {"api_name": "torch.optim.SGD", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 66, "usage_type": "name"}, {"api_name": "numpy.random.permutation", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 74, "usage_type": "attribute"}, {"api_name": "preprocessing.Normalize", "line_number": 81, "usage_type": "call"}, {"api_name": "preprocessing.TRAIN_MEAN", "line_number": 81, "usage_type": "attribute"}, {"api_name": "preprocessing.TRAIN_STD", "line_number": 82, "usage_type": "attribute"}, {"api_name": "preprocessing.Normalize", "line_number": 84, "usage_type": "call"}, {"api_name": "preprocessing.TRAIN_MEAN", "line_number": 84, "usage_type": "attribute"}, {"api_name": "preprocessing.TRAIN_STD", "line_number": 85, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 93, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.nn.functional.nll_loss", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 101, "usage_type": "name"}, {"api_name": "torch.from_numpy", "line_number": 117, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 118, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 121, "usage_type": "call"}, {"api_name": "torch.nn.functional.nll_loss", "line_number": 123, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 123, "usage_type": "name"}, {"api_name": "preprocessing.get_data", "line_number": 135, "usage_type": "call"}, {"api_name": "preprocessing.get_data", "line_number": 136, "usage_type": "call"}]} +{"seq_id": "157169457", "text": "# Convnet with Data augmentation, Regularization\nimport numpy\nimport keras\nfrom keras.preprocessing.image import ImageDataGenerator\nimport matplotlib.pyplot as plt\nfrom keras.datasets import cifar10\nfrom keras.models import Sequential\nfrom keras.layers import Dense\nfrom keras.layers import Dropout\nfrom keras.layers import Flatten\nfrom keras.constraints import maxnorm\nfrom keras.optimizers import SGD\nfrom keras.layers.convolutional import Conv2D\nfrom keras.layers.convolutional import MaxPooling2D\nfrom keras.utils import np_utils\nfrom keras import backend as K\nK.set_image_dim_ordering('th')\n\n\nclass AccHistory_train(keras.callbacks.Callback):\n def on_train_begin(self, logs={}):\n self.accuracy = []\n\n def on_epoch_end(self, batch, logs={}):\n self.accuracy.append(logs.get('acc'))\n\n\nclass AccHistory_valid(keras.callbacks.Callback):\n def on_train_begin(self, logs={}):\n self.accuracy = []\n\n def on_epoch_end(self, batch, logs={}):\n self.accuracy.append(logs.get('val_acc'))\n\nseed = 7\nnumpy.random.seed(seed)\n\n(X_train, y_train), (X_test, y_test) = cifar10.load_data()\n\nX_train = X_train.astype('float32')\nX_test = X_test.astype('float32')\nX_train = X_train / 255.0\nX_test = X_test / 255.0\n\ny_train = np_utils.to_categorical(y_train)\ny_test = np_utils.to_categorical(y_test)\nnum_classes = y_test.shape[1]\n\n\nmodel = Sequential()\nmodel.add(Conv2D(32, (3, 3), input_shape=(3, 32, 32), activation='relu', padding='same'))\nmodel.add(Dropout(0.2))\nmodel.add(Conv2D(32, (3, 3), activation='relu', padding='same'))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\nmodel.add(Conv2D(64, (3, 3), activation='relu', padding='same'))\nmodel.add(Dropout(0.2))\nmodel.add(Conv2D(64, (3, 3), activation='relu', padding='same'))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\nmodel.add(Conv2D(128, (3, 3), activation='relu', padding='same'))\nmodel.add(Dropout(0.2))\nmodel.add(Conv2D(128, (3, 3), activation='relu', padding='same'))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\nmodel.add(Flatten())\nmodel.add(Dropout(0.2))\nmodel.add(Dense(1024, activation='relu', kernel_constraint=maxnorm(3)))\nmodel.add(Dropout(0.2))\nmodel.add(Dense(512, activation='relu', kernel_constraint=maxnorm(3)))\nmodel.add(Dropout(0.2))\nmodel.add(Dense(num_classes, activation='softmax'))\n# Compile model\nepochs = 25\n\n# lrate = 0.01\n# decay = lrate/epochs\n# sgd = SGD(lr=lrate, momentum=0.9, decay=decay, nesterov=False)\nmodel.compile(loss='categorical_crossentropy', optimizer='adam', metrics=['accuracy'])\nprint(model.summary())\n\n\nt_accuracy_history = AccHistory_train()\nv_accuracy_history = AccHistory_valid()\n\n# using keras ImageDataGenerator class for data augmentation\ndatagen = ImageDataGenerator(\n rotation_range=40,\n width_shift_range=0.2,\n shear_range=0.2,\n horizontal_flip=True,\n fill_mode='nearest')\n\n\n# compute quantities required for featurewise normalization\ndatagen.fit(X_train)\n\n# fits the model on batches with real-time data augmentation:\nmodel.fit_generator(datagen.flow(X_train, y_train, batch_size=100),\n samples_per_epoch=len(X_train), nb_epoch=epochs, validation_data=(X_test, y_test),\n callbacks=[t_accuracy_history, v_accuracy_history])\n\n# model.fit(X_train, y_train, validation_data=(X_test, y_test), epochs=epochs, batch_size=128,\n# callbacks=[t_accuracy_history, v_accuracy_history])\n# Final evaluation of the model\nscores = model.evaluate(X_test, y_test, verbose=0)\nprint(\"Accuracy: %.2f%%\" % (scores[1]*100))\n\n# model.save_weights('first_try_1.h5')\n\nta = t_accuracy_history.accuracy\nva = v_accuracy_history.accuracy\n\nfig, ax = plt.subplots()\nline_1, = ax.plot(ta, label='Inline label')\nline_2, = ax.plot(va, 'r-', label='Inline label')\n# Overwrite the label by calling the method.\nline_1.set_label('training set')\nline_2.set_label('validation set')\nax.legend()\nax.set_ylabel('Accuracy')\nax.set_title('convnet_reg_aug')\nax.set_xlabel('epoch')\nplt.show()", "sub_path": "convnet_reg_aug.py", "file_name": "convnet_reg_aug.py", "file_ext": "py", "file_size_in_byte": 3951, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "keras.backend.set_image_dim_ordering", "line_number": 17, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 17, "usage_type": "name"}, {"api_name": "keras.callbacks", "line_number": 20, "usage_type": "attribute"}, {"api_name": "keras.callbacks", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.random.seed", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 36, "usage_type": "attribute"}, {"api_name": "keras.datasets.cifar10.load_data", "line_number": 38, "usage_type": "call"}, {"api_name": "keras.datasets.cifar10", "line_number": 38, "usage_type": "name"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 45, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 45, "usage_type": "name"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 46, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 46, "usage_type": "name"}, {"api_name": "keras.models.Sequential", "line_number": 50, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 51, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 52, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 53, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.MaxPooling2D", "line_number": 54, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 55, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 56, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 57, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.MaxPooling2D", "line_number": 58, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 59, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 60, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 61, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.MaxPooling2D", "line_number": 62, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 63, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 64, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 65, "usage_type": "call"}, {"api_name": "keras.constraints.maxnorm", "line_number": 65, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 66, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 67, "usage_type": "call"}, {"api_name": "keras.constraints.maxnorm", "line_number": 67, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 68, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 69, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 121, "usage_type": "name"}]} +{"seq_id": "589495372", "text": "from nmt.utils.argument import get_main_argument\r\nfrom nmt.utils.log import get_logger\r\nfrom os.path import join, exists\r\nfrom os import makedirs\r\nimport os\r\nimport torch\r\nimport json\r\n\r\n\r\ndef create_dir(dir_path):\r\n if not exists(dir_path):\r\n makedirs(dir_path)\r\n\r\n\r\nclass Context:\r\n def __init__(self, desc=\"Transformer\", config=None, logger=None):\r\n self.description = desc\r\n\r\n # A dictionary of Config Parameters\r\n self.config = get_main_argument(desc=self.description)\r\n if config is not None:\r\n self.config.update(config)\r\n\r\n self.project_name = self.config[\"project_name\"]\r\n self.phase = self.config[\"phase\"]\r\n\r\n self.project_raw_dir = str(self.config[\"project_raw_dir\"])\r\n create_dir(self.project_raw_dir)\r\n\r\n self.project_log = self.config[\"project_log\"]\r\n if not exists(self.project_log):\r\n self.project_log = join(os.path.dirname(self.project_raw_dir), 'logs', f'{self.phase}.log')\r\n create_dir(os.path.dirname(self.project_log))\r\n\r\n # logger interface\r\n self.isDebug = self.config[\"debug\"] == 'True'\r\n self.logger = get_logger(self.description, self.project_log, self.isDebug) if logger is None else logger\r\n self.logger.debug(\"The logger interface is initited ...\")\r\n\r\n self.project_config = self.config[\"project_config\"]\r\n if exists(self.project_config):\r\n with open(self.project_config) as f:\r\n self.config.update(json.load(f))\r\n else:\r\n create_dir(join(os.path.dirname(self.project_raw_dir), 'configs'))\r\n\r\n self.project_save_config = self.config[\"project_save_config\"]\r\n\r\n if self.project_save_config is True:\r\n config_filepath = join(os.path.dirname(self.project_raw_dir),\r\n 'configs',\r\n f'${self.project_name}_save_config.json')\r\n self.logger.debug(f\"Dump project configration to the file {config_filepath} ...\")\r\n with open(config_filepath, 'w') as config_file:\r\n json.dump(self.config, config_file)\r\n\r\n self.logger.debug(\"The Input Parameters:\")\r\n for key, val in self.config.items():\r\n self.logger.debug(f\"{key} => {val}\")\r\n\r\n self.project_processed_dir = self.config[\"project_processed_dir\"]\r\n create_dir(self.project_processed_dir)\r\n\r\n self.project_checkpoint = str(self.config[\"project_checkpoint\"])\r\n if not exists(self.project_checkpoint):\r\n create_dir(os.path.dirname(self.project_checkpoint))\r\n\r\n # Model Paramters\r\n self.d_model = self.config[\"d_model\"]\r\n self.layers_count = self.config[\"layers_count\"]\r\n self.heads_count = self.config[\"heads_count\"]\r\n self.d_ff = self.config[\"d_ff\"]\r\n self.dropout = self.config[\"dropout\"]\r\n self.label_smoothing = self.config[\"label_smoothing\"]\r\n self.optimizer = self.config[\"optimizer\"]\r\n self.lr = self.config[\"lr\"]\r\n self.clip_grads = self.config[\"clip_grads\"]\r\n self.batch_size = self.config[\"batch_size\"]\r\n self.epochs = self.config[\"epochs\"]\r\n self.vocabulary_size = self.config[\"vocabulary_size\"]\r\n self.padding_index = 0\r\n\r\n self.dataset_limit = self.config[\"dataset_limit\"]\r\n self.print_every = self.config[\"print_every\"]\r\n self.print_every = self.config[\"print_every\"]\r\n\r\n self.source = self.config[\"source\"]\r\n self.num_candidates = self.config[\"num_candidates\"]\r\n self.save_result = self.config[\"save_result\"]\r\n self.share_dictionary = self.config[\"share_dictionary\"]\r\n self.save_every = self.config[\"save_every\"]\r\n\r\n # Trainning Device Set up\r\n self.device = torch.device(self.config[\"device\"])\r\n self.device_id = list(self.config[\"device_id\"])\r\n self.is_cuda = self.config[\"device\"] == 'cuda'\r\n self.is_cpu = self.config[\"device\"] == 'cpu'\r\n self.is_gpu_parallel = self.is_cuda and (len(self.device_id) > 1)\r\n", "sub_path": "nmt/utils/context.py", "file_name": "context.py", "file_ext": "py", "file_size_in_byte": 4082, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "os.path.exists", "line_number": 11, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 12, "usage_type": "call"}, {"api_name": "nmt.utils.argument.get_main_argument", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "nmt.utils.log.get_logger", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 41, "usage_type": "call"}, {"api_name": "json.load", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "torch.device", "line_number": 94, "usage_type": "call"}]} +{"seq_id": "491549771", "text": "import matplotlib.pyplot as plt\nfrom tqdm import tqdm\n\n# fungsi untuk mendapatkan data node dari file\ndef getCoorNodesFromFile(cityOption):\n coorNodes = []\n fNodes = open(f'../data/{cityOption}/NodeData.txt', 'r').readlines()\n for line in fNodes:\n parseLine = [x for x in line.split()]\n x = float(parseLine[1])\n y = float(parseLine[2])\n coorNodes.append((x,y)) \n return coorNodes\n\n# fungsi untuk draw map nya\ndef drawMap(adjListAll, mapNodeToIdx, listPath, jumlahKurir, resTSPkurir, cntAllNode, cityOption):\n coorNodes = getCoorNodesFromFile(cityOption)\n\n if (cityOption==\"Oldenburg\"):\n #print all map\n for i in tqdm(range(cntAllNode)):\n if (len(adjListAll[i])!=0):\n for v in adjListAll[i]:\n x1 = coorNodes[i][0]\n y1 = coorNodes[i][1]\n x2 = coorNodes[v][0]\n y2 = coorNodes[v][1]\n plt.plot([x1,x2],[y1,y2],color=\"k\")\n\n listColor=['c','r','y','m','g','b']\n\n for i in tqdm(range(jumlahKurir)):\n for pairNode in resTSPkurir[i]:\n path = listPath[mapNodeToIdx[pairNode[0]]][mapNodeToIdx[pairNode[1]]]\n predNode = path[0]\n for nodes in path:\n x1 = coorNodes[predNode][0]\n y1 = coorNodes[predNode][1]\n x2 = coorNodes[nodes][0]\n y2 = coorNodes[nodes][1]\n plt.plot([x1,x2],[y1,y2],color=listColor[i%6])\n\n predNode = nodes\n \n \n plt.show()", "sub_path": "src/mapDrawer.py", "file_name": "mapDrawer.py", "file_ext": "py", "file_size_in_byte": 1567, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "tqdm.tqdm", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "tqdm.tqdm", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}]} +{"seq_id": "477688167", "text": "from typing import Dict\n\nfrom bobocep.rules.events.bobo_event import BoboEvent\n\n\nclass PrimitiveEvent(BoboEvent):\n \"\"\"A primitive event.\n\n :param timestamp: The event timestamp indicating when it was first\n generated.\n :type timestamp: int\n\n :param data: The event data, defaults to an empty dict.\n :type data: Dict[str, str], optional\n\n :param event_id: The event ID, defaults to a randomly generated ID.\n :type event_id: str, optional\n \"\"\"\n\n def __init__(self,\n timestamp: int,\n data: Dict[str, str] = None,\n event_id: str = None) -> None:\n super().__init__(timestamp=timestamp,\n data=data,\n event_id=event_id)\n\n def to_dict(self) -> dict:\n \"\"\"\n :return: A dict representation of the object.\n \"\"\"\n\n return {\n self.TIMESTAMP: self.timestamp,\n self.DATA: self.data,\n self.EVENT_ID: self.event_id\n }\n", "sub_path": "bobocep/rules/events/primitive_event.py", "file_name": "primitive_event.py", "file_ext": "py", "file_size_in_byte": 1021, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "bobocep.rules.events.bobo_event.BoboEvent", "line_number": 6, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 22, "usage_type": "name"}]} +{"seq_id": "489698056", "text": "import os\nimport sys\nimport json\nfrom datetime import date\nimport pandas as pd\nimport streamlit as st\n\n\nfrom litmscript.litmanalysis.remittance_downloader import download_cheque_files as download_cheque_files\nfrom litmscript.litmanalysis.generate_row_level_data_from_json_updated_labelling import generate_row_level_data_util\nfrom litmscript.litmanalysis.heading_model import predict_on_test_data as header_prediction\nfrom litmscript.litmanalysis.total_model import predict_on_test_data as total_prediction\nfrom litmscript.litmanalysis.is_remittance import predict_on_test_data as remittance_prediction\nfrom litmscript.litmanalysis.amount_and_reference_no_capture import monitoring_litm_data\n\n\n\ndef flow_code(root_dir,s3_path,account_no):\n \n \n super_directory = root_dir\n model_dir = \"/root/caascript/litmscript/litmanalysis\"\n header_pkl_path = os.path.join(model_dir,\"LITM_heading.pickle\")\n total_pkl_path = os.path.join(model_dir,\"LITM_total.pickle\")\n remittance_pkl_path = os.path.join(model_dir,\"LITM_remittance.pickle\")\n\n\n account_id=str(account_no)\n directory = os.path.join(super_directory,str(account_id))\n\n if not os.path.exists(directory):\n os.makedirs(directory)\n\n \n data = pd.read_csv(s3_path)\n \n with st.spinner(\"Downloading Image files\"):\n download_cheque_files(data, directory)\n print(\"directory path\"+directory)\n \n with st.spinner(\"Generating Prediction files: \"):\n json_csv_write_path = os.path.join(directory,'row_level_data.csv')\n print(\"json path:\"+json_csv_write_path)\n generate_row_level_data_util(directory,json_csv_write_path)\n header_prediction(header_pkl_path,json_csv_write_path,directory)\n print(\"header prediction\")\n header_prediction_path = os.path.join(directory,'is_heading_prediction.csv')\n print(\"header prediction path\")\n total_prediction(total_pkl_path,header_prediction_path,directory)\n print(\"total prediction\")\n total_prediction_path = os.path.join(directory,'is_total_pred.csv')\n print(\"total prediction path\")\n remittance_prediction(remittance_pkl_path,total_prediction_path,directory)\n print(\"remittance prediciton\")\n remittace_prediction_path = os.path.join(directory,'is_remittance_pred.csv')\n print(\"remittance prediction path\")\n monitoring_litm_data(remittace_prediction_path,directory)\n print(\"monitoring Prediction path\")\n\n correctly_closed_path = os.path.join(directory, 'correctly_closed_checks.csv')\n os.chdir(root_dir)\n return correctly_closed_path", "sub_path": "flow_new.py", "file_name": "flow_new.py", "file_ext": "py", "file_size_in_byte": 2591, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 32, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 35, "usage_type": "call"}, {"api_name": "streamlit.spinner", "line_number": 37, "usage_type": "call"}, {"api_name": "litmscript.litmanalysis.remittance_downloader.download_cheque_files", "line_number": 38, "usage_type": "call"}, {"api_name": "streamlit.spinner", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "litmscript.litmanalysis.generate_row_level_data_from_json_updated_labelling.generate_row_level_data_util", "line_number": 44, "usage_type": "call"}, {"api_name": "litmscript.litmanalysis.heading_model.predict_on_test_data", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "litmscript.litmanalysis.total_model.predict_on_test_data", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "litmscript.litmanalysis.is_remittance.predict_on_test_data", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "litmscript.litmanalysis.amount_and_reference_no_capture.monitoring_litm_data", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 61, "usage_type": "call"}]} +{"seq_id": "597830333", "text": "\nfrom django.db import models\n\n\nclass TimeStamp(models.Model):\n created_at = models.DateTimeField(auto_now_add=True)\n\n class Meta:\n abstract = True\n\n def age(self):\n import datetime\n from django.utils.timezone import utc\n now = datetime.datetime.utcnow().replace(tzinfo=utc)\n age_in_minutes = int((now - self.created_at).total_seconds()) / 60\n\n if age_in_minutes < 60:\n value = age_in_minutes\n precision = 'minute'\n elif age_in_minutes < 60 * 24:\n value = age_in_minutes // 60\n precision = 'hour'\n else:\n value = age_in_minutes // (60*24)\n precision = 'day'\n\n age_string ='%d %s%s ago' % (value, precision,\n ('s' if value > 1 else ''))\n return age_string\n", "sub_path": "posts/behaviours.py", "file_name": "behaviours.py", "file_ext": "py", "file_size_in_byte": 836, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "django.db.models.Model", "line_number": 5, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 5, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 6, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 6, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 14, "usage_type": "attribute"}, {"api_name": "django.utils.timezone.utc", "line_number": 14, "usage_type": "name"}]} +{"seq_id": "375281849", "text": "import numpy as np\nimport keras\nimport cv2\nimport os\n# from vidaug import augmentors as va\nfrom matplotlib.pyplot import imread, imshow, show\nfrom tensorflow.keras.preprocessing.image import ImageDataGenerator\nfrom data_helper import calculateRGBdiff\n# from imageai.Detection import ObjectDetection\n\nclass DataGeneratorBKB(keras.utils.Sequence):\n 'Generates data for Keras'\n def __init__(self, list_IDs, labels, batch_size=32, dim=(32,32), n_channels=1,\n n_sequence=4, shuffle=True, path_dataset=None,\n select_joint=[], type_gen='train', type_model='rgb', option=None):\n 'Initialization'\n self.dim = dim\n self.batch_size = batch_size\n self.labels = labels\n self.list_IDs = list_IDs\n self.n_channels = n_channels\n self.n_sequence = n_sequence\n self.shuffle = shuffle\n self.path_dataset = path_dataset\n self.select_joint = select_joint\n self.n_joint = len(select_joint)\n self.option = option\n self.type_gen = type_gen\n self.type_model = type_model\n print(\"all:\", len(self.list_IDs), \" batch per epoch\", int(np.floor(len(self.list_IDs) / self.batch_size)) )\n \n # execution_path = os.getcwd()\n # self.detector = ObjectDetection()\n # self.detector.setModelTypeAsYOLOv3()\n # self.detector.setModelPath( os.path.join(execution_path , \"pretrain/yolo.h5\"))\n # self.detector.loadModel(detection_speed=\"fast\")#detection_speed=\"fast\"\n # self.execution_path = execution_path\n # self.detector = detector\n\n # sometimes = lambda aug: va.Sometimes(0.5, aug) # Used to apply augmentor with 50% probability\n # self.seq = va.SomeOf([ #va.Sequential([\n # # va.RandomCrop(size=(300, 300)), # randomly crop video with a size of (240 x 180)\n # va.RandomRotate(degrees=10), # randomly rotates the video with a degree randomly choosen from [-10, 10] \n # va.RandomTranslate(x=60,y=30), \n # # sometimes(va.Add(value=-100)),\n # # sometimes(va.Pepper(ratio=40)),\n # sometimes(va.Add(value=-60)),\n # sometimes(va.HorizontalFlip()) # horizontally flip the video with 50% probability\n # ], 2)\n\n self.aug_gen = ImageDataGenerator() \n \n self.on_epoch_end()\n\n def __len__(self):\n 'Denotes the number of batches per epoch'\n return int(np.floor(len(self.list_IDs) / self.batch_size))\n\n def on_epoch_end(self):\n 'Updates indexes after each epoch' \n self.indexes = np.arange(len(self.list_IDs))\n if self.shuffle == True:\n np.random.shuffle(self.indexes)\n\n def __getitem__(self, index):\n 'Generate one batch of data'\n # Generate indexes of the batch\n indexes = self.indexes[index*self.batch_size:(index+1)*self.batch_size]\n # Find list of IDs\n list_IDs_temp = [self.list_IDs[k] for k in indexes]\n # Generate data\n X, y = self.__data_generation(list_IDs_temp)\n if self.type_gen == 'predict':\n return X\n else:\n return X, y\n\n def get_sampling_frame(self, len_frames, path_video): \n '''\n Sampling n_sequence frame from video file\n Input: \n len_frames -- number of frames that this video have\n Output: \n index_sampling -- n_sequence frame indexs from sampling algorithm \n '''\n\n # Define maximum sampling rate\n # sample_interval = len_frames//self.n_sequence\n # start_i = 0 #np.random.randint(0, len_frames - sample_interval * self.n_sequence + 1)\n\n if True:#self.type_gen =='train':\n random_sample_range = 10\n if random_sample_range*self.n_sequence > len_frames:\n random_sample_range = len_frames//self.n_sequence\n\n if random_sample_range <= 0:\n print('test:',random_sample_range, len_frames, path_video)\n # Randomly choose sample interval and start frame\n if random_sample_range < 3:\n sample_interval = np.random.randint(1, random_sample_range + 1)\n else:\n sample_interval = np.random.randint(3, random_sample_range + 1)\n\n # sample_interval = np.random.randint(1, random_sample_range + 1)\n\n # temp = len_frames - sample_interval * self.n_sequence + 1\n # if temp <= 0:\n # print(temp, len_frames)\n start_i = np.random.randint(0, len_frames - sample_interval * self.n_sequence + 1)\n \n # Get n_sequence index of frames\n index_sampling = []\n end_i = sample_interval * self.n_sequence + start_i\n for i in range(start_i, end_i, sample_interval):\n if len(index_sampling) < self.n_sequence:\n index_sampling.append(i)\n \n return index_sampling\n\n def get_crop_img(self, frame):\n # detect_image, detections, extract_picture = self.detector.detectObjectsFromImage(input_type=\"array\", input_image=frame, output_type='array', \n # minimum_percentage_probability=10, extract_detected_objects=True )\n print('#################',self.execution_path)\n detections = self.detector.detectObjectsFromImage(input_image=os.path.join(self.execution_path , \"room.jpg\"),\n output_image_path=os.path.join(self.execution_path , \"image2new.jpg\"), minimum_percentage_probability=30)\n max_prob = 0\n max_idx = 0\n for i,eachObject in enumerate(detections):\n if eachObject[\"name\"] == 'person' and eachObject[\"percentage_probability\"] > max_prob:\n max_prob = eachObject[\"percentage_probability\"]\n max_idx = i\n if max_idx > len(detections):\n # if no detection, use black array\n crop_img = np.zeros((*self.dim, self.n_channels))\n else:\n crop_img = extract_picture[max_idx]\n return crop_img\n\n # def calculateRGBdiff(self, sequence_img):\n # 'keep first frame as rgb data, other is use RGBdiff for temporal data'\n # length = len(sequence_img)\n # new_sequence = np.zeros((length,self.dim[0],self.dim[1],self.n_channels))\n\n # # find RGBdiff frame 1 to last frame\n # for i in range(length-1,3,-1): # count down\n # new_sequence[i] = cv2.subtract(sequence_img[i],sequence_img[i-1])\n \n # new_sequence[:4] = sequence_img[:4] # first frame as rgb data\n\n # return new_sequence\n\n def sequence_augment(self, sequence):\n name_list = ['rotate','width_shift','height_shift',\n 'brightness','flip_horizontal','width_zoom',\n 'height_zoom']\n dictkey_list = ['theta','ty','tx',\n 'brightness','flip_horizontal','zy',\n 'zx']\n # dictkey_list = ['ty','tx','zy','zx']\n random_aug = np.random.randint(2, 5) # random 0-4 augmentation method\n pick_idx = np.random.choice(len(dictkey_list), random_aug, replace=False) #\n\n dict_input = {}\n for i in pick_idx:\n if dictkey_list[i] == 'theta':\n # dict_input['theta'] = np.random.randint(-10, 10)\n dict_input['theta'] = np.random.randint(-5,5)\n\n elif dictkey_list[i] == 'ty': # width_shift\n # dict_input['ty'] = np.random.randint(-60, 60)\n dict_input['ty'] = np.random.randint(-20,20)\n\n elif dictkey_list[i] == 'tx': # height_shift\n # dict_input['tx'] = np.random.randint(-30, 30)\n dict_input['tx'] = np.random.randint(-10,10)\n\n elif dictkey_list[i] == 'brightness': \n dict_input['brightness'] = np.random.uniform(0.15,1)\n\n elif dictkey_list[i] == 'flip_horizontal': \n dict_input['flip_horizontal'] = True\n\n elif dictkey_list[i] == 'zy': # width_zoom\n # dict_input['zy'] = np.random.uniform(0.5,1.5)\n dict_input['zy'] = np.random.uniform(0.9,1.3) \n\n elif dictkey_list[i] == 'zx': # height_zoom\n # dict_input['zx'] = np.random.uniform(0.5,1.5)\n dict_input['zx'] = np.random.uniform(0.9,1.3) \n \n sh = sequence.shape\n new_sequence = np.zeros((sh[0],sh[1],sh[2],sh[3]))\n for i in range(sh[0]):\n new_sequence[i] = self.aug_gen.apply_transform(sequence[i],dict_input)\n \n return new_sequence\n \n\n def __data_generation(self, list_IDs_temp):\n 'Generates data containing batch_size samples'\n # Initialization\n X1 = np.empty((self.batch_size, self.n_sequence, *self.dim, self.n_channels)) # X : (n_samples, timestep, *dim, n_channels)\n X2 = np.empty((self.batch_size, self.n_sequence, self.n_joint*3))\n Y = np.empty((self.batch_size), dtype=int)\n\n for i, ID in enumerate(list_IDs_temp): # ID is name of file (2 batch)\n path_video = self.path_dataset + ID + '.mp4'\n # print(path_video)\n path_skeleton = self.path_dataset + ID + '.npy'\n \n # print(path_video)\n \n if self.type_model == '2stream' or self.type_model == 'rgb':\n cap = cv2.VideoCapture(path_video) \n length_file = int(cap.get(cv2.CAP_PROP_FRAME_COUNT)) # get how many frames this video have ,-1 because some bug \n \n if self.type_model == '2stream' or self.type_model == 'skeleton':\n skeleton_data = np.load(path_skeleton)\n length_file = skeleton_data.shape[0]\n\n index_sampling = self.get_sampling_frame(length_file, path_video) # get sampling index \n # print(index_sampling)\n if self.type_model == '2stream' or self.type_model == 'rgb': \n # Get RGB sequence\n for j, n_pic in enumerate(index_sampling):\n cap.set(cv2.CAP_PROP_POS_FRAMES, n_pic) # jump to that index\n ret, frame = cap.read()\n # frame = self.get_crop_img(frame)\n new_image = cv2.resize(frame, self.dim)\n # new_image = frame\n # new_image = new_image/255.0 \n X1[i,j,:,:,:] = new_image\n \n\n if self.type_gen =='train':\n X1[i,] = self.sequence_augment(X1[i,])/255.0*2-1\n else:\n X1[i,] = X1[i,]/255.0*2-1\n\n # cv2.imshow('imgae',X1[i,0])\n # cv2.waitKey(2000)\n\n\n if self.option == 'RGBdiff':\n # print(\"dddddddddddd\")\n X1[i,] = calculateRGBdiff(X1[i,], 0)\n \n\n if self.type_model == '2stream' or self.type_model == 'skeleton':\n # Get skeleton sequence \n skeleton_data = skeleton_data[index_sampling]\n skeleton_data = skeleton_data[:,:,self.select_joint]\n skeleton_data = skeleton_data.reshape(self.n_sequence,self.n_joint*3)\n X2[i] = skeleton_data\n\n # Get label\n Y[i] = self.labels[ID]\n if self.type_model == '2stream' or self.type_model == 'rgb':\n cap.release() \n\n # for i_frame in range(self.n_sequence):\n # cv2.imshow('Frame', X1[i,i_frame])\n # cv2.waitKey(1000)\n\n if self.type_model == 'rgb':\n X = X1\n elif self.type_model == 'skeleton':\n X = X2\n elif self.type_model == '2stream':\n X = [X1, X2]\n\n return X,Y\n\n\n# class DataGenerator2Stream(DataGeneratorBKB):\n# def __init__(self, list_IDs, labels, batch_size=32, dim=(32,32),\n# n_channels=1, n_sequence=4, shuffle=True, path_dataset=None, \n# select_joint=[], type_gen='train'):\n\n# super().__init__(list_IDs, labels, batch_size, dim, n_channels,\n# n_sequence, shuffle, path_dataset, type_gen)\n# self.select_joint = select_joint\n", "sub_path": "data_gen_bkb.py", "file_name": "data_gen_bkb.py", "file_ext": "py", "file_size_in_byte": 12219, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "keras.utils", "line_number": 11, "usage_type": "attribute"}, {"api_name": "numpy.floor", "line_number": 30, "usage_type": "call"}, {"api_name": "tensorflow.keras.preprocessing.image.ImageDataGenerator", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 63, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 100, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 102, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 109, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 124, "usage_type": "call"}, {"api_name": "os.path", "line_number": 124, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 125, "usage_type": "call"}, {"api_name": "os.path", "line_number": 125, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 160, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 161, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 167, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 171, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 175, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 178, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 185, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 185, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 189, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 189, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 192, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 202, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 204, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 214, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_FRAME_COUNT", "line_number": 215, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 218, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_POS_FRAMES", "line_number": 226, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 229, "usage_type": "call"}, {"api_name": "data_helper.calculateRGBdiff", "line_number": 246, "usage_type": "call"}]} +{"seq_id": "199248646", "text": "# Copyright 2021 MosaicML. All Rights Reserved.\n\nfrom copy import deepcopy\n\nimport pytest\nimport torch\n\nfrom composer.algorithms import LayerFreezing, LayerFreezingHparams\nfrom composer.core.state import State\nfrom composer.core.types import Event, Model, Precision\nfrom composer.loggers import Logger\nfrom composer.trainer.trainer_hparams import TrainerHparams\nfrom composer.utils import ensure_tuple\nfrom tests.utils.trainer_fit import train_model\n\n\ndef _generate_state(epoch: int, max_epochs: int, model: Model):\n state = State(\n epoch=epoch,\n step=epoch,\n train_batch_size=64,\n eval_batch_size=64,\n grad_accum=1,\n max_epochs=max_epochs,\n model=model,\n optimizers=(torch.optim.SGD(model.parameters(), lr=0.01, momentum=0.99),),\n precision=Precision.FP32,\n )\n return state\n\n\ndef _check_param_groups(expected_groups, actual_groups):\n assert len(expected_groups) == len(actual_groups), 'Incorrect number of param groups'\n\n for i, expected_group in enumerate(expected_groups):\n assert len(expected_group) == len(actual_groups[i]), \\\n f'Group {i} has the wrong number of parameters'\n\n for j, expected_params in enumerate(expected_group['params']):\n torch.testing.assert_equal(actual_groups[i]['params'][j], expected_params)\n\n\ndef test_freeze_layers_no_freeze(simple_conv_model: Model, noop_dummy_logger: Logger):\n state = _generate_state(epoch=10, max_epochs=100, model=simple_conv_model)\n\n first_optimizer = ensure_tuple(state.optimizers)[0]\n assert first_optimizer is not None\n\n expected_param_groups = deepcopy(first_optimizer.param_groups)\n freezing = LayerFreezing(freeze_start=0.5, freeze_level=1.0)\n freezing.apply(event=Event.EPOCH_END, state=state, logger=noop_dummy_logger)\n updated_param_groups = first_optimizer.param_groups\n\n _check_param_groups(expected_param_groups, updated_param_groups)\n\n\ndef test_freeze_layers_with_freeze(simple_conv_model: Model, noop_dummy_logger: Logger):\n state = _generate_state(epoch=80, max_epochs=100, model=simple_conv_model)\n\n first_optimizer = ensure_tuple(state.optimizers)[0]\n assert first_optimizer is not None\n\n expected_param_groups = deepcopy(first_optimizer.param_groups)\n # The first group should be removed due to freezing\n expected_param_groups[0]['params'] = expected_param_groups[0]['params'][1:]\n freezing = LayerFreezing(freeze_start=0.05, freeze_level=1.0)\n freezing.apply(event=Event.EPOCH_END, state=state, logger=noop_dummy_logger)\n updated_param_groups = first_optimizer.param_groups\n\n _check_param_groups(expected_param_groups, updated_param_groups)\n\n\n@pytest.mark.run_long\n@pytest.mark.timeout(90)\ndef test_layer_freezing_trains(mosaic_trainer_hparams: TrainerHparams):\n mosaic_trainer_hparams.algorithms = [LayerFreezingHparams(freeze_start=.25, freeze_level=1)]\n train_model(mosaic_trainer_hparams, max_epochs=4)\n", "sub_path": "tests/algorithms/test_layer_freezing.py", "file_name": "test_layer_freezing.py", "file_ext": "py", "file_size_in_byte": 2955, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "composer.core.types.Model", "line_number": 17, "usage_type": "name"}, {"api_name": "composer.core.state.State", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.optim.SGD", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 26, "usage_type": "attribute"}, {"api_name": "composer.core.types.Precision.FP32", "line_number": 27, "usage_type": "attribute"}, {"api_name": "composer.core.types.Precision", "line_number": 27, "usage_type": "name"}, {"api_name": "torch.testing.assert_equal", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.testing", "line_number": 40, "usage_type": "attribute"}, {"api_name": "composer.core.types.Model", "line_number": 43, "usage_type": "name"}, {"api_name": "composer.loggers.Logger", "line_number": 43, "usage_type": "name"}, {"api_name": "composer.utils.ensure_tuple", "line_number": 46, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 49, "usage_type": "call"}, {"api_name": "composer.algorithms.LayerFreezing", "line_number": 50, "usage_type": "call"}, {"api_name": "composer.core.types.Event.EPOCH_END", "line_number": 51, "usage_type": "attribute"}, {"api_name": "composer.core.types.Event", "line_number": 51, "usage_type": "name"}, {"api_name": "composer.core.types.Model", "line_number": 57, "usage_type": "name"}, {"api_name": "composer.loggers.Logger", "line_number": 57, "usage_type": "name"}, {"api_name": "composer.utils.ensure_tuple", "line_number": 60, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 63, "usage_type": "call"}, {"api_name": "composer.algorithms.LayerFreezing", "line_number": 66, "usage_type": "call"}, {"api_name": "composer.core.types.Event.EPOCH_END", "line_number": 67, "usage_type": "attribute"}, {"api_name": "composer.core.types.Event", "line_number": 67, "usage_type": "name"}, {"api_name": "composer.trainer.trainer_hparams.TrainerHparams", "line_number": 75, "usage_type": "name"}, {"api_name": "composer.algorithms.LayerFreezingHparams", "line_number": 76, "usage_type": "call"}, {"api_name": "tests.utils.trainer_fit.train_model", "line_number": 77, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 73, "usage_type": "attribute"}, {"api_name": "pytest.mark.timeout", "line_number": 74, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 74, "usage_type": "attribute"}]} +{"seq_id": "616909160", "text": "#\n# @lc app=leetcode id=1557 lang=python3\n#\n# [1557] Minimum Number of Vertices to Reach All Nodes\n#\n\n# @lc code=start\nimport collections\nclass Solution:\n def findSmallestSetOfVertices(self, n: int, edges: List[List[int]]) -> List[int]:\n indgree = collections.Counter()\n for edge in edges:\n indgree[edge[1]]+=1\n result = []\n for i in range(n):\n if indgree[i] == 0:\n result.append(i)\n return result\n \n \n\n \n# @lc code=end\n\n", "sub_path": "py-src/1557.minimum-number-of-vertices-to-reach-all-nodes.py", "file_name": "1557.minimum-number-of-vertices-to-reach-all-nodes.py", "file_ext": "py", "file_size_in_byte": 525, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "collections.Counter", "line_number": 11, "usage_type": "call"}]} +{"seq_id": "414266124", "text": "from flask import Flask\nfrom flask_jwt_extended import JWTManager\nfrom flask_cors import CORS\n\nfrom api_v1.api_user import main as user_api_routes\nfrom api_v1.api_auth import main as auth_api_routes\nfrom api_v1.api_topic import main as topic_api_routes\nfrom api_v1.api_board import main as board_api_routes\nfrom api_v1.api_reply import main as reply_api_routes\nfrom api_v1.api_mail import main as mail_api_routes\nfrom config import secret_key\n\n\ndef configured_app():\n app = Flask(__name__)\n # 设置 secret_key 来使用 flask 自带的 session\n # 这个字符串只要hard to guess就可以了\n app.secret_key = secret_key\n\n # app.register_blueprint(index_routes)\n # app.register_blueprint(topic_routes, url_prefix='/topic')\n # app.register_blueprint(reply_routes, url_prefix='/reply')\n # app.register_blueprint(board_routes, url_prefix='/board')\n # app.register_blueprint(mail_routes, url_prefix='/mail')\n app.register_blueprint(auth_api_routes, url_prefix='/api/v1')\n app.register_blueprint(user_api_routes, url_prefix='/api/v1')\n app.register_blueprint(topic_api_routes, url_prefix='/api/v1')\n app.register_blueprint(board_api_routes, url_prefix='/api/v1')\n app.register_blueprint(reply_api_routes, url_prefix='/api/v1')\n app.register_blueprint(mail_api_routes, url_prefix='/api/v1')\n # Setup the Flask-JWT-Extended extension\n app.config['JWT_SECRET_KEY'] = 'super-secret' # Change this!\n jwt = JWTManager(app)\n CORS(app)\n return app\n\n\nif __name__ == '__main__':\n app = configured_app()\n # 自动 reload jinja\n app.config['TEMPLATES_AUTO_RELOAD'] = True\n app.jinja_env.auto_reload = True\n app.config['SEND_FILE_MAX_AGE_DEFAULT'] = 0\n # debug 模式可以自动加载对代码的变动, 所以不用重启程序\n # host 参数指定为 '0.0.0.0' 可以让别的机器访问代码\n config = dict(\n debug=True,\n host='0.0.0.0',\n port=8000,\n threaded=True,\n )\n app.run(**config)\n", "sub_path": "bbsapi/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1999, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "flask.Flask", "line_number": 15, "usage_type": "call"}, {"api_name": "config.secret_key", "line_number": 18, "usage_type": "name"}, {"api_name": "api_v1.api_auth.main", "line_number": 25, "usage_type": "argument"}, {"api_name": "api_v1.api_user.main", "line_number": 26, "usage_type": "argument"}, {"api_name": "api_v1.api_topic.main", "line_number": 27, "usage_type": "argument"}, {"api_name": "api_v1.api_board.main", "line_number": 28, "usage_type": "argument"}, {"api_name": "api_v1.api_reply.main", "line_number": 29, "usage_type": "argument"}, {"api_name": "api_v1.api_mail.main", "line_number": 30, "usage_type": "argument"}, {"api_name": "flask_jwt_extended.JWTManager", "line_number": 33, "usage_type": "call"}, {"api_name": "flask_cors.CORS", "line_number": 34, "usage_type": "call"}]} +{"seq_id": "162080484", "text": "from django.shortcuts import render\nfrom django.http import JsonResponse\nfrom default.models import Todolist\n\n# Create your views here.\n\ndef index(request):\n return render(request,\"index.html\")\n\n\ndef todogetlist(request):\n dataset = Todolist.objects.all()\n todolist = []\n for item in dataset:\n temp ={\"id\":item.id,\"content\":item.content}\n todolist.append(temp)\n res ={\"todolist\":todolist}\n return JsonResponse(res)\n\n\ndef todoadd(request):\n todo = request.POST['todo']\n Todolist.objects.create(content=todo)\n res ={\"success\":\"true\"}\n return JsonResponse(res)\n\n\ndef todoedit(request):\n id = request.POST['id']\n content = request.POST['todo']\n todo = Todolist.objects.get(id=id)\n todo.content =content\n todo.save()\n res ={\"success\":\"true\"}\n return JsonResponse(res)\n\ndef tododel(request):\n id = request.GET['id']\n Todolist.objects.get(id=id).delete()\n res ={\"success\":\"true\"}\n return JsonResponse(res)", "sub_path": "django框架练习/todolist3/default/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 975, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "django.shortcuts.render", "line_number": 8, "usage_type": "call"}, {"api_name": "default.models.Todolist.objects.all", "line_number": 12, "usage_type": "call"}, {"api_name": "default.models.Todolist.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "default.models.Todolist", "line_number": 12, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 18, "usage_type": "call"}, {"api_name": "default.models.Todolist.objects.create", "line_number": 23, "usage_type": "call"}, {"api_name": "default.models.Todolist.objects", "line_number": 23, "usage_type": "attribute"}, {"api_name": "default.models.Todolist", "line_number": 23, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 25, "usage_type": "call"}, {"api_name": "default.models.Todolist.objects.get", "line_number": 31, "usage_type": "call"}, {"api_name": "default.models.Todolist.objects", "line_number": 31, "usage_type": "attribute"}, {"api_name": "default.models.Todolist", "line_number": 31, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 35, "usage_type": "call"}, {"api_name": "default.models.Todolist.objects.get", "line_number": 39, "usage_type": "call"}, {"api_name": "default.models.Todolist.objects", "line_number": 39, "usage_type": "attribute"}, {"api_name": "default.models.Todolist", "line_number": 39, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 41, "usage_type": "call"}]} +{"seq_id": "3811282", "text": "import requests,time\nfrom lxml import etree\n\n# url = \"https://weixin.sogou.com/weixin?query=%E5%BE%AE%E7%A4%BE%E5%8C%BAe%E5%AE%B6%E9%80%9A%E9%87%91%E8%8A%B1\"\nagent = 'Mozilla/5.0 (Linux; Android 6.0; Nexus 5 Build/MRA58N) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/56.0.2924.87 Mobile Safari/537.36'\nagent1 = \"Mozilla/5.0 (Linux; Android 6.0; Nexus 5 Build/MRA58N) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/71.0.3578.98 Mobile Safari/537.36\"\nheaders = {\n'Host': \"weixin.sogou.com\",\n'Accept': \"text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,image/apng,*/*;q=0.8\",\n'Accept-Encoding': \"gzip, deflate, br\",\n'Accept-Language': \"zh-CN,zh;q=0.9,en;q=0.8\",\n'User-Agent': agent,\n'Connection': \"keep-alive\",\n}\n#读取本地ip.txt中的IP代理\ndef ip_proxy():#\n # encoding=utf-8\n # 随机数,随机读取每一行的数据\n import linecache\n import random\n\n for i in range(1, 2): # for循环几次3\n a = random.randrange(1, 27) # 1-9中生成随机数\n print(a)\n # 从文件poem.txt中对读取第a行的数据\n theline = linecache.getline(r'ip.txt', a)\n theline = theline.strip('\\n')\n print(\"the a row data is:\",theline)\n return theline\n\n# proxy = ip_proxy()\nproxy = '12.191.182.53:9999'\nurl = 'http://www.baidu.com' # 验证ip有效性的指定url\n# url = \"https://weixin.sogou.com/weixin?query=\"+'公众号'\nproxies = {\n 'http': 'http://' + proxy\n}\nprint(proxy)\nresponse = requests.get(url, allow_redirects=False, headers=headers, proxies=proxies)\nprint(response.status_code)\n\n# time.sleep(3)\nif response.status_code == 200:\n with open('test.html', 'w', encoding='utf-8') as f:\n f.write(response.text)\n session = requests.Session()\n print(session)\n seletor = etree.HTML(response.text)\n print(\"current page:\", response.url)\nelse:\n print(\"网页出错了\",response.status_code)\n", "sub_path": "WeChat/Ejiatong/proxyTest.py", "file_name": "proxyTest.py", "file_ext": "py", "file_size_in_byte": 1890, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "random.randrange", "line_number": 23, "usage_type": "call"}, {"api_name": "linecache.getline", "line_number": 26, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 39, "usage_type": "call"}, {"api_name": "requests.Session", "line_number": 46, "usage_type": "call"}, {"api_name": "lxml.etree.HTML", "line_number": 48, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 48, "usage_type": "name"}]} +{"seq_id": "135227536", "text": "from django.shortcuts import render, redirect\r\nfrom User.models import Branch, User\r\n\r\ndef home(request):\r\n return render(request, 'index.html')\r\n\r\ndef register(request):\r\n branches = Branch.objects.all()\r\n print(branches)\r\n return render(request, 'register.html',{\r\n 'branches':branches\r\n })\r\n\r\ndef saveuser(request):\r\n ob= User()\r\n ob.user_name = request.POST.get('uname')\r\n ob.user_email = request.POST.get('email')\r\n ob.user_password = request.POST.get('pwd')\r\n ob.user_type = request.POST.get('type')\r\n branch_slug = request.POST.get('branch')\r\n branch = Branch.objects.filter(branch_slug=branch_slug)\r\n ob.user_branch = branch[0]\r\n ob.save()\r\n return redirect(\"/register\")\r\n \r\n\r\ndef login(request):\r\n if request.method == \"GET\":\r\n return render(request, 'login.html') \r\n else:\r\n email = request.POST.get(\"email\")\r\n pwd = request.POST.get(\"pwd\")\r\n\r\n records = User.objects.filter(user_email=email,user_password=pwd)\r\n count = len(records)\r\n if count == 0:\r\n return redirect(\"/login\") \r\n else:\r\n userob = records[0]\r\n request.session['user'] = {\r\n 'userid': userob.id,\r\n 'name' : userob.user_name,\r\n 'email' : userob.user_email,\r\n 'type' : userob.user_type,\r\n 'branch' : userob.user_branch.branch_slug\r\n }\r\n return redirect(\"/user/home\") ", "sub_path": "CollegeQuera/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1499, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "django.shortcuts.render", "line_number": 5, "usage_type": "call"}, {"api_name": "User.models.Branch.objects.all", "line_number": 8, "usage_type": "call"}, {"api_name": "User.models.Branch.objects", "line_number": 8, "usage_type": "attribute"}, {"api_name": "User.models.Branch", "line_number": 8, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 10, "usage_type": "call"}, {"api_name": "User.models.User", "line_number": 15, "usage_type": "call"}, {"api_name": "User.models.Branch.objects.filter", "line_number": 21, "usage_type": "call"}, {"api_name": "User.models.Branch.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "User.models.Branch", "line_number": 21, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 24, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 29, "usage_type": "call"}, {"api_name": "User.models.User.objects.filter", "line_number": 34, "usage_type": "call"}, {"api_name": "User.models.User.objects", "line_number": 34, "usage_type": "attribute"}, {"api_name": "User.models.User", "line_number": 34, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 37, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 47, "usage_type": "call"}]} +{"seq_id": "233719166", "text": "#!/usr/bin/env python\n\n\"\"\"\n########################################################################\n# Author: Carlos A. Ruiz Perez\n# Email: cruizperez3@gatech.edu\n# Intitution: Georgia Institute of Technology\n# Version: 1.0.0\n# Date: Nov 13, 2020\n\n# Description: Parses compressed dat files and extracts protein\ninformation relevant for annotation purposes.\n########################################################################\n\"\"\"\n\n################################################################################\n\"\"\"---0.0 Import Modules---\"\"\"\nimport gzip\nfrom pathlib import Path\n\n################################################################################\n\"\"\"---1.0 Define Functions---\"\"\"\n\ndef parse_uniprot_dat(dat_file, output_file_table):\n output_folder = Path(output_file_table).parent\n uniprot_to_refseq = output_folder / \"03.uniprot_to_refseq.txt\"\n with gzip.open(dat_file, 'rt') as uniprot, open(output_file_table, 'w') as output_file, open(uniprot_to_refseq, 'a') as uni_to_ref:\n gene_id = \"\"\n accession = \"\"\n gene_name = \"\"\n ko_number = \"\"\n organism = \"\"\n taxonomy = \"\"\n function = \"\"\n compartment = \"\"\n process = \"\"\n interpro = \"\"\n pfam = \"\"\n ec_number = \"\"\n refseq_code = \"\"\n for line in uniprot:\n if line.startswith(\"ID\", 0):\n gene_id = line.split()[1]\n elif \"AC \" in line:\n accession = line.split()[1].replace(\";\",\"\")\n elif line.startswith(\"DE\") and \"Full=\" in line:\n gene_name = line.split(\"Full=\")[1]\n gene_name = gene_name.split(\"{\")[0].strip().replace(\";\",\"\")\n gene_name = gene_name.lower()\n elif \"OS \" in line:\n organism = ' '.join([organism, line.split(\"OS\")[1].strip()]).replace(\".\",\"\")\n elif \"OC \" in line:\n taxonomy = ' '.join([taxonomy, line.split(\"OC\")[1].strip()]).replace(\".\",\"\")\n elif \"DR KO;\" in line:\n ko_number = line.split()[2].replace(\";\", \"\")\n elif \"DR GO;\" in line:\n if \"; F:\" in line:\n code = line.strip().split(\";\")[1]\n code = code.strip()\n if function == \"\":\n function = code\n else:\n function = ''.join([function, \" \", code])\n elif \"; C:\" in line:\n code = line.strip().split(\";\")[1]\n code = code.strip()\n if compartment == \"\":\n compartment = code\n else:\n compartment = ''.join([compartment, \" \", code])\n elif \"; P:\" in line:\n code = line.strip().split(\";\")[1]\n code = code.strip()\n if process == \"\":\n process = code\n else:\n process = ''.join([process, \" \", code])\n elif \"DR InterPro\" in line:\n code = line.strip().split()[2]\n code = code.replace(\";\", \"\")\n if interpro == \"\":\n interpro = code\n else:\n interpro = ' '.join([interpro, code])\n elif \"DR Pfam\" in line:\n code = line.strip().split()[2]\n code = code.replace(\";\", \"\")\n if pfam == \"\":\n pfam = code\n else:\n pfam = ' '.join([pfam, code])\n elif line.startswith(\"DE\") and \"EC=\" in line:\n ec_code = line.strip().split()[1]\n ec_code = ec_code.replace(\";\", \"\")\n ec_code = ec_code.replace(\"EC=\", \"\")\n if ec_number == \"\":\n ec_number = ec_code\n else:\n ec_number = ' '.join([ec_number, ec_code])\n elif line.startswith(\"DR\") and \"RefSeq\" in line:\n refseq_code = line.strip().split()[2]\n refseq_code = refseq_code.replace(\";\", \"\")\n uni_to_ref.write(\"{}\\t{}\\n\".format(gene_id, refseq_code))\n elif \"//\\n\" in line:\n if ko_number == \"\":\n ko_number = \"NA\"\n if organism == \"\":\n organism = \"NA\"\n if function == \"\":\n function = \"NA\"\n if compartment == \"\":\n compartment = \"NA\"\n if process == \"\":\n process = \"NA\"\n if interpro == \"\":\n interpro = \"NA\"\n if pfam == \"\":\n pfam = \"NA\"\n if ec_number == \"NA\":\n ec_number = \"NA\"\n output_file.write(\"{}\\t{}\\t{}\\t{}\\t{}\\t{}\\t{}\\t{}\\t{}\\t{}\\t{}\\t{}\\n\".format(gene_id, \n accession, gene_name, ko_number, organism, taxonomy, function, compartment, process, interpro, pfam, ec_number))\n gene_id = \"\"\n accession = \"\"\n gene_name = \"\"\n ko_number = \"\"\n organism = \"\"\n taxonomy = \"\"\n function = \"\"\n compartment = \"\"\n process = \"\"\n interpro = \"\"\n pfam =\"\"\n ec_number = \"\"\n refseq_code = \"\"\n\n\n################################################################################\n\"\"\"---3.0 Main Function---\"\"\"\n\ndef main():\n import argparse, sys\n # Setup parser for arguments.\n parser = argparse.ArgumentParser(description='''This script parses a Uniprot.dat file and output_files a table with\\n'''\n '''the ID, Accession, Gene Name, Organism, Taxonomy, KEGG ID, Function,\\n\n Compartment, Process, InterPro, and Pfam\\n\n For faster usage in alrge files use gnu parallel (read script file to see how)\\n'''\n '''\\nGlobal mandatory parameters: [Input Uniprot.dat File]\\n'''\n '''\\nOptional Database Parameters: See ''' + sys.argv[0] + ' -h')\n parser.add_argument('-i', '--input', dest='input_dat', action='store', required=True, help='Uniprot.dat file to parse')\n parser.add_argument('-o', '--output_file', dest='output_file_table', action='store', required=False, help='output_file table')\n args = parser.parse_args()\n\n input_dat = args.input_dat\n output_file_table = args.output_file_table\n\n parse_uniprot_dat(input_dat, output_file_table)\n\nif __name__ == \"__main__\":\n main()\n", "sub_path": "independent_scripts/uniprot_dat_parser.py", "file_name": "uniprot_dat_parser.py", "file_ext": "py", "file_size_in_byte": 6786, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pathlib.Path", "line_number": 25, "usage_type": "call"}, {"api_name": "gzip.open", "line_number": 27, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 144, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 149, "usage_type": "attribute"}]} +{"seq_id": "158367951", "text": "from __future__ import print_function\n\nimport os\nimport sys\nfrom time import sleep\nfrom os import environ\nfrom socket import socket\nfrom traceback import format_tb\nfrom socket import AF_INET, SOCK_STREAM\n\n\nfrom circuits.web.errors import httperror\nfrom circuits.web import Controller, Server\nfrom circuits.web.exceptions import NotFound\nfrom circuits import handler, Event, Component\n\nfrom jinja2 import Environment, FileSystemLoader, TemplateNotFound\n\nfrom redisco import connection_setup, get_client\n\n\nfrom models import TodoItem, TodoList\n\n\nDEFAULTS = {\n \"appname\": \"todoapp\",\n \"version\": \"dev\",\n}\n\n\nclass render(Event):\n \"\"\"render Event\"\"\"\n\n\nclass JinjaTemplate(object):\n\n def __init__(self, _name, **context):\n self._name = _name\n self.context = context\n\n @property\n def name(self):\n return self._name\n\n\nclass JinjaRenderer(Component):\n\n channel = \"web\"\n\n def init(self, path, defaults=None):\n self.path = path\n self.defaults = defaults or {}\n\n self.env = Environment(loader=FileSystemLoader(path))\n\n @handler(\"response\", priority=1.0)\n def serialize_response_body(self, event, response):\n template = response.body\n if not isinstance(template, JinjaTemplate):\n return\n\n try:\n request = response.request\n\n try:\n tmpl = self.env.get_template(\"{0}.html\".format(template.name))\n except TemplateNotFound:\n raise NotFound()\n\n ctx = self.defaults.copy()\n ctx.update({\"request\": request, \"response\": response, \"uri\": request.uri})\n\n ctx.update(template.context)\n\n response.body = tmpl.render(**ctx)\n except:\n event.stop()\n evalue, etype, etraceback = sys.exc_info()\n error = (evalue, etype, format_tb(etraceback))\n self.fire(httperror(request, response, 500, error=error))\n\n\nclass Root(Controller):\n\n def GET(self, *args, **kwargs):\n name = (args and args[0]) or \"TODO\"\n todo = TodoList.objects.get_or_create(name=name)\n entries = [entry for entry in todo.entries if not entry.done]\n return JinjaTemplate(\"views/index\", name=name, entries=entries)\n\n\nclass Add(Controller):\n\n channel = \"/add\"\n\n def GET(self, *args, **kwargs):\n return JinjaTemplate(\"views/add\")\n\n def POST(self, *args, **kwargs):\n name = (args and args[0]) or \"TODO\"\n todo = TodoList.objects.get_or_create(name=name)\n todo.add_entry(kwargs[\"title\"])\n return self.redirect(self.uri(\"/\"))\n\n\nclass Update(Controller):\n\n channel = \"/update\"\n\n def done(self, *args, **kwargs):\n id = int(kwargs[\"id\"])\n item = TodoItem.objects.get_by_id(id)\n item.mark_done()\n return self.redirect(self.uri(\"/\"))\n\n\ndef waitfor(host, port, timeout=10):\n sock = socket(AF_INET, SOCK_STREAM)\n\n while sock.connect_ex((host, port)) != 0 and timeout:\n timeout -= 1\n sleep(1)\n\n if timeout <= 0:\n print(\"Timed out waiting for {0:s}:{1:d}\".format(host, port))\n raise SystemExit(1)\n\n\ndef setup_database():\n dbhost = environ.get(\"REDIS_PORT_6379_TCP_ADDR\", \"localhost\")\n dbport = int(environ.get(\"REDIS_PORT_6379_TCP_PORT\", \"6379\"))\n\n print(\"Waiting for Redis Service on {0:s}:{1:d} ...\".format(dbhost, dbport))\n\n waitfor(dbhost, dbport)\n\n print(\"Connecting to Redis on {0:s}:{1:d} ...\".format(dbhost, dbport))\n\n connection_setup(host=dbhost, port=dbport)\n\n print(\"Success!\")\n\n db = get_client()\n\n return db\n\n\nclass TodoApp(Component):\n\n def init(self, db):\n self.db = db\n\n Server((\"0.0.0.0\", 8000)).register(self)\n JinjaRenderer(\"templates\", defaults=DEFAULTS).register(self)\n\n Root().register(self)\n Add().register(self)\n Update().register(self)\n\n def stopped(self, *args):\n print(\"Shutting down...\")\n self.db.save()\n\n\ndef main():\n sys.stdout = os.fdopen(sys.stdout.fileno(), \"w\", 0)\n\n db = setup_database()\n\n TodoApp(db).run()\n\n\nif __name__ == \"__main__\":\n main()\n", "sub_path": "todoapp/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 4105, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "circuits.Event", "line_number": 31, "usage_type": "name"}, {"api_name": "circuits.Component", "line_number": 46, "usage_type": "name"}, {"api_name": "jinja2.Environment", "line_number": 54, "usage_type": "call"}, {"api_name": "jinja2.FileSystemLoader", "line_number": 54, "usage_type": "call"}, {"api_name": "jinja2.TemplateNotFound", "line_number": 67, "usage_type": "name"}, {"api_name": "circuits.web.exceptions.NotFound", "line_number": 68, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 78, "usage_type": "call"}, {"api_name": "traceback.format_tb", "line_number": 79, "usage_type": "call"}, {"api_name": "circuits.web.errors.httperror", "line_number": 80, "usage_type": "call"}, {"api_name": "circuits.handler", "line_number": 56, "usage_type": "call"}, {"api_name": "circuits.web.Controller", "line_number": 83, "usage_type": "name"}, {"api_name": "models.TodoList.objects.get_or_create", "line_number": 87, "usage_type": "call"}, {"api_name": "models.TodoList.objects", "line_number": 87, "usage_type": "attribute"}, {"api_name": "models.TodoList", "line_number": 87, "usage_type": "name"}, {"api_name": "circuits.web.Controller", "line_number": 92, "usage_type": "name"}, {"api_name": "models.TodoList.objects.get_or_create", "line_number": 101, "usage_type": "call"}, {"api_name": "models.TodoList.objects", "line_number": 101, "usage_type": "attribute"}, {"api_name": "models.TodoList", "line_number": 101, "usage_type": "name"}, {"api_name": "circuits.web.Controller", "line_number": 106, "usage_type": "name"}, {"api_name": "models.TodoItem.objects.get_by_id", "line_number": 112, "usage_type": "call"}, {"api_name": "models.TodoItem.objects", "line_number": 112, "usage_type": "attribute"}, {"api_name": "models.TodoItem", "line_number": 112, "usage_type": "name"}, {"api_name": "socket.socket", "line_number": 118, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 118, "usage_type": "argument"}, {"api_name": "socket.SOCK_STREAM", "line_number": 118, "usage_type": "argument"}, {"api_name": "time.sleep", "line_number": 122, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 130, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 130, "usage_type": "name"}, {"api_name": "os.environ.get", "line_number": 131, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 131, "usage_type": "name"}, {"api_name": "redisco.connection_setup", "line_number": 139, "usage_type": "call"}, {"api_name": "redisco.get_client", "line_number": 143, "usage_type": "call"}, {"api_name": "circuits.Component", "line_number": 148, "usage_type": "name"}, {"api_name": "circuits.web.Server", "line_number": 153, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 166, "usage_type": "attribute"}, {"api_name": "os.fdopen", "line_number": 166, "usage_type": "call"}, {"api_name": "sys.stdout.fileno", "line_number": 166, "usage_type": "call"}]} +{"seq_id": "169219636", "text": "from impala.dbapi import connect\nimport glob\nfrom datetime import date\n\nconn = connect(host='node02', port=21050, auth_mechanism='GSSAPI')\ncursor = conn.cursor()\n\ndatabase = \"data_telco\"\n\ncursor.execute(\"use \"+database)\npath = \"data/transformed_data/*.csv\"\nfiles = glob.glob(path)\n\ntoday = str(date.today().strftime(\"%Y%m%d\"))\n\nfor file in files:\n folder_name = str(file).split(\"data/transformed_data/\")[1].split(\".\")[0]\n query = 'CREATE EXTERNAL TABLE IF NOT EXISTS ' + folder_name + '(`ten_kh` string, `dia_chi` string, `sdt` bigint, `thanh_pho` string)' + \" ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' \" + \"LOCATION 'hdfs:///user/hive/warehouse/data_telco_namnn2/data_telco_\" + today + \"/\" + folder_name + \"'\"\n cursor.execute(query)", "sub_path": "vega_data_telco/create_table_impala.py", "file_name": "create_table_impala.py", "file_ext": "py", "file_size_in_byte": 746, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "impala.dbapi.connect", "line_number": 5, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 12, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 14, "usage_type": "name"}]} +{"seq_id": "564878038", "text": "from pylab import *\r\nimport matplotlib as mpl\r\nimport numpy as np\r\n\r\nclass GridPlotter:\r\n \"\"\"\r\n Controls the production of a plot image from\r\n a geodata class. The plot image contains\r\n a series of sub plots.\r\n\r\n Author: M. Gill\r\n Copyright: M. Gill\r\n \"\"\"\r\n \r\n def __init__ (self):\r\n \"\"\"\r\n Constructor\r\n \"\"\"\r\n self.sdev = -9999\r\n self.mean = -9999\r\n self.min = -9999\r\n self.max = -9999\r\n self.shownulls = True\r\n\r\n\r\n def plotcsv(self, filename, delim, title, outfile):\r\n \"\"\"\r\n Entry point to create a plot from a csv file.\r\n The CSV file should be a formatted file, not\r\n a raw data file. Null values should be represented\r\n by nan (which indicates 'not a number').\r\n \r\n filename - a string of the full path and filename\r\n of the source data file\r\n delim - the string delimiter in the file\r\n title - the title for the plot\r\n outfile - a string of the full path and filename\r\n of the out file\r\n \"\"\"\r\n arr = np.loadtxt(filename,\r\n delimiter=delim,\r\n unpack=True)\r\n self.plotgrid(arr, title, outfile)\r\n \r\n\r\n def plotgeodata(self, geodata, title, outfile):\r\n \"\"\"\r\n Entry point to create a plot from a geodata object.\r\n \r\n geodata - the geodata object to be plotted\r\n title - the title for the plot\r\n outfile - a string of the full path and filename\r\n of the out file\r\n \"\"\"\r\n \r\n # Array columns and rows need to be swapped in order\r\n # to be shown properly on the plot (np.transpose())\r\n self.plotgrid(np.transpose(np.array(geodata.data)),\r\n title,\r\n outfile)\r\n\r\n\r\n def plotgrid(self, arr, title, outfile):\r\n \"\"\"\r\n The main controlling method for the production\r\n of a plot file.\r\n \r\n arr - the geodata object to be plotted\r\n title - the title for the plot\r\n outfile - a string of the full path and filename\r\n of the out file\r\n \"\"\"\r\n # Calculate relevant stats\r\n self.calcstats(arr) \r\n\r\n # Find positions of nulls\r\n arrnulls = self.getnullindexes(arr)\r\n\r\n # Clear figure and set global figure settings \r\n plt.clf()\r\n mpl.rc('xtick', labelsize=8)\r\n mpl.rc('ytick', labelsize=8)\r\n mpl.rc('savefig', dpi=300)\r\n\r\n fig = plt.figure(figsize=(10,10))\r\n plt.title(title)\r\n\r\n # Calculate lower and upper limit values for\r\n # the 2 standard deviations plot\r\n sdev_lower = self.mean - 2*self.sdev\r\n if sdev_lower < self.min:\r\n sdev_lower = self.min\r\n sdev_upper = self.mean + 2*self.sdev\r\n if sdev_upper > self.max:\r\n sdev_upper = self.max\r\n\r\n # Create the subplots\r\n self.makeplot(arr, 331, 'nearest', cm.gray, [],\r\n 'Grey nearest', 10,\r\n True, True, arrnulls)\r\n self.makeplot(arr, 332, 'bilinear', cm.gray, [],\r\n 'Grey bilinear', 10,\r\n True, True, arrnulls)\r\n self.makeplot(arr, 333, 'nearest', cm.gray,\r\n [sdev_lower, sdev_upper],\r\n 'Two STDEV nearest ', 10,\r\n True, True, arrnulls)\r\n self.makeplot(arr, 334, 'nearest', cm.jet, [],\r\n 'Colour nearest', 10,\r\n True, True, arrnulls)\r\n self.makeplot(arr, 335, 'bilinear', cm.jet, [],\r\n 'Colour bilinear', 10,\r\n True, True, arrnulls)\r\n self.makeplot(arr, 336, 'bilinear', cm.gray,\r\n [sdev_lower, sdev_upper],\r\n 'Two STDEV bilinear', 10,\r\n True, True, arrnulls)\r\n self.makeplot(arr, 337, 'nearest', cm.jet,\r\n [sdev_lower, sdev_upper],\r\n 'Two STDEV nearest', 10,\r\n True, True, arrnulls)\r\n self.makeplot(arr, 338, 'bilinear', cm.jet,\r\n [sdev_lower, sdev_upper],\r\n 'Two STDEV bilinear', 10,\r\n True, True, arrnulls)\r\n\r\n # Save the figure to file \r\n plt.savefig(outfile)\r\n\r\n \r\n def makeplot(self,\r\n arr,\r\n plotref,\r\n interp,\r\n colmap,\r\n col_lim,\r\n title,\r\n titlesize,\r\n showlegend,\r\n showgrid,\r\n arrnulls):\r\n \"\"\"\r\n Controls the creation of a subplot.\r\n \r\n arr - the 2D array to be plotted\r\n plotref - the subplot position index\r\n interp - the interpolation method (eg bilinear)\r\n colmap - the colormap (eg cm.jet)\r\n col_lim - lower and upper data limits as an array\r\n of format [lower_val, upper_val]\r\n title - the title to be shown for the subplot\r\n titlesize - the font size\r\n showlegend - boolean controlling legend display\r\n showgrid - boolean controlling grid display\r\n arrnulls - 2D array of null value indexes in form\r\n [[r,c],[r,c]]\r\n \"\"\"\r\n ax = plt.subplot(plotref)\r\n ax.axis('off')\r\n imgplot = plt.imshow(arr, origin='lower',\r\n interpolation=interp, cmap=colmap)\r\n if len(col_lim) != 0:\r\n imgplot.set_clim(col_lim[0], col_lim[1])\r\n plt.grid(showgrid)\r\n if showlegend:\r\n plt.colorbar()\r\n plt.title(title, size=titlesize)\r\n if len(arrnulls) > 0:\r\n self.plotnulls(ax, arrnulls)\r\n \r\n\r\n def getnullindexes(self, arr):\r\n \"\"\"\r\n Calculates the row and column indexes for null\r\n values.\r\n\r\n arr - the 2D array\r\n returns - 2D array of null value indexes in form\r\n [[r,c],[r,c]]\r\n \"\"\"\r\n arrnulls = []\r\n for r in range(0, len(arr)):\r\n for c in range(0, len(arr[r])):\r\n if math.isnan(arr[r][c]):\r\n arrnulls.append([r, c])\r\n return arrnulls\r\n\r\n\r\n def plotnulls(self, ax, arrnulls):\r\n \"\"\"\r\n Displays a small red cross at the position\r\n of a null value on the plot\r\n\r\n ax - the plot axes\r\n arrnulls - 2D array of null value indexes in form\r\n [[r,c],[r,c]] \r\n \"\"\"\r\n if self.shownulls:\r\n ax.set_autoscale_on(False)\r\n if len(arrnulls) == 0:\r\n return\r\n\r\n for i in range(0, len(arrnulls)):\r\n ax.plot([arrnulls[i][1]], [arrnulls[i][0]],\r\n 'rx', markersize=3, mew=1)\r\n \r\n\r\n def filternans(self, arr):\r\n \"\"\"\r\n Creates a list of values from a 2D array, filtering\r\n out nan (not a number) values.\r\n\r\n arr - the array of values to be filtered\r\n returns - a filtered list of values \r\n \"\"\"\r\n list = []\r\n for r in range(0, len(arr)):\r\n for c in range(0, len(arr[r])):\r\n if not math.isnan(arr[r][c]):\r\n list.append(arr[r][c])\r\n return np.array(list)\r\n \r\n\r\n def calcstats(self, arr):\r\n \"\"\"\r\n Calculates mean and standard deviation\r\n statistics for an array of values.\r\n\r\n arr - the array for which stats are to be\r\n calculated \r\n \"\"\"\r\n arrfilt = self.filternans(arr)\r\n self.sdev = np.std(arrfilt)\r\n self.mean = np.mean(arrfilt)\r\n self.min = np.min(arrfilt)\r\n self.max = np.max(arrfilt)\r\n\r\n", "sub_path": "geoconverter/src/geoconverter/GridPlotter.py", "file_name": "GridPlotter.py", "file_ext": "py", "file_size_in_byte": 7739, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "numpy.loadtxt", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 233, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 234, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 235, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 236, "usage_type": "call"}]} +{"seq_id": "412212094", "text": "\nfrom mrjob.job import MRJob\n\n# Avoid broken pipe error\nfrom signal import signal, SIGPIPE, SIG_DFL\nsignal(SIGPIPE,SIG_DFL) \n\nclass LJ(MRJob):\n def mapper_init(self):\n self.urls = {}\n with open(\"urls.txt\") as urls:\n for line in urls:\n url, key = line.strip().replace('\"',\"\").split(\",\")\n self.urls[key] = url\n \n def mapper(self, _, lines):\n try:\n yield (lines, self.urls[lines[2:6]])\n except ValueError:\n yield (lines, \"\")\n \nif __name__ == \"__main__\":\n LJ.run()", "sub_path": "week5/lj.py", "file_name": "lj.py", "file_ext": "py", "file_size_in_byte": 574, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "signal.signal", "line_number": 6, "usage_type": "call"}, {"api_name": "signal.SIGPIPE", "line_number": 6, "usage_type": "argument"}, {"api_name": "signal.SIG_DFL", "line_number": 6, "usage_type": "argument"}, {"api_name": "mrjob.job.MRJob", "line_number": 8, "usage_type": "name"}]} +{"seq_id": "69947016", "text": "#Selenium imports here\nfrom selenium import webdriver\nfrom selenium.webdriver.common.keys import Keys\nfrom selenium.webdriver.support import expected_conditions as EC\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.support.wait import WebDriverWait\nfrom bs4 import BeautifulSoup\nimport time\nimport re\nfrom urllib.request import urlopen\n# from urllib.request import urlopen \nimport json\nfrom pandas.io.json import json_normalize\nimport pandas as pd, numpy as np\nimport bs4, requests\n\n\n#Other imports here\nimport os\nimport wget\n\ndriver=webdriver.Chrome('C:/Users/Professional/Downloads/chromedriver_win32/chromedriver.exe')\n\n# Fisiere txt\ninstagram = open('./instagram.txt', 'r') \n\n#Variables\nURL='https://www.instagram.com/'\nURL_BEST = 'https://www.instagram.com/best_chisinau/'\nPOST_URL_PATTERN='https://www.instagram.com/best_chisinau/p/'\npost_xpath_str = \"//a[contains(@href, '/p/')]\"\npost_links = driver.find_elements_by_xpath(post_xpath_str)\npost_link_el = None\n\ndriver.get(URL)\n\n#target username\nusername = WebDriverWait(driver, 10).until(EC.element_to_be_clickable((By.CSS_SELECTOR, \"input[name='username']\")))\npassword = WebDriverWait(driver, 10).until(EC.element_to_be_clickable((By.CSS_SELECTOR, \"input[name='password']\")))\n\n#enter username and password\nusername.clear()\nusername.send_keys(\"massveritas\")\npassword.clear()\npassword.send_keys(\"#123456\")\n\n#target the login button and click it\nbutton = WebDriverWait(driver, 2).until(EC.element_to_be_clickable((By.CSS_SELECTOR, \"button[type='submit']\"))).click()\n\n\n#nadle NOT NOW\nnot_now = WebDriverWait(driver, 10).until(EC.element_to_be_clickable((By.XPATH, '//button[contains(text(), \"Not Now\")]'))).click()\n# not_now2 = WebDriverWait(driver, 10).until(EC.element_to_be_clickable((By.XPATH, '//button[contains(text(), \"Not Now\")]'))).click()\n\ndriver.implicitly_wait(10)\n\ndriver.get(URL_BEST)\n\n# scroll to the bottom of the page\nlenOfPage = driver.execute_script(\"window.scrollTo(0, document.body.scrollHeight);var lenOfPage=document.body.scrollHeight;return lenOfPage;\")\nmatch=False\nwhile(match==False):\n lastCount = lenOfPage\n time.sleep(3)\n lenOfPage = driver.execute_script(\"window.scrollTo(0, document.body.scrollHeight);var lenOfPage=document.body.scrollHeight;return lenOfPage;\")\n if lastCount==lenOfPage:\n match=True\n\n# find all links on the page and if they match '/p' append to list named posts\nposts = []\nlinks = driver.find_elements_by_tag_name('a')\nfor link in links:\n post = link.get_attribute('href')\n if '/p/' in post:\n posts.append( post )\n\nfor post in posts:\n driver.get( post )\n page = requests.get(post)\n soup = BeautifulSoup(page.text, \"lxml\")\n print(soup.find_all('span'))\n\ninstagram = open('./instagram.txt', 'r') \n\n\n# Citire line by line\n# \nLines = instagram.readlines()\n\n#Variabile\ninstagram_array=[]\nroman_numbers=[]\nkeys=[]\n\n#Parsing txts\nfor line in Lines: \n new_line = line.strip().split(':')\n instagram_array.append(new_line)\n\nclass Solution(object):\n def romanToInt(self, s):\n \"\"\"\n :type s: str\n :rtype: int\n \"\"\"\n roman = {'I':1,'V':5,'X':10,'L':50,'C':100,'D':500,'M':1000,'IV':4,'IX':9,'XL':40,'XC':90,'CD':400,'CM':900}\n i = 0\n num = 0\n while i < len(s):\n if i+130*60):\n print('Downloading',csv_fn)\n _download_file(csv_remote_fp, csv_fp)\n\ndef _load_National_data(csv_fp):\n df_arg = pd.read_csv(csv_fp)\n df_arg['LOCATION'] = 'ARGENTINA/' + df_arg['PROVINCIA']\n df_arg = df_arg.drop(columns=['PROVINCIA'])\n df_arg = df_arg.set_index(['TYPE','LOCATION'])\n df_arg = df_arg.rename(columns=lambda colname: pd.to_datetime(colname,format='%d/%m').replace(year=2020))\n\n total_arg = df_arg.groupby(level=[0]).sum()\n total_arg['LOCATION']='ARGENTINA'\n total_arg = total_arg.reset_index().set_index(['TYPE','LOCATION'])\n\n df_arg = pd.concat([df_arg,total_arg]).sort_index()\n df_arg = df_arg[df_arg.columns[:-1]]\n return df_arg\n\ndef _set_location_safe(row):\n location_prefix = 'ARGENTINA/SANTA FE'\n if row['DEPARTMENT']=='##TOTAL':\n return location_prefix\n location_prefix += '/'+row['DEPARTMENT'][3:]\n if row['PLACE'].startswith('#'):\n return location_prefix\n return location_prefix +'/'+ row['PLACE']\n\ndef _load_SantaFe_data(csv_fp):\n df_safe = pd.read_csv(csv_fp)\n df_safe['LOCATION'] = df_safe.apply(_set_location_safe, axis=1)\n df_safe = df_safe[ (df_safe['TYPE']=='CONFIRMADOS') & (df_safe['DEPARTMENT']!='##TOTAL') ]\n df_safe['LOCATION'] = df_safe['LOCATION'].replace({\n 'ARGENTINA/SANTA FE/IRIONDO/CLASSON':'ARGENTINA/SANTA FE/IRIONDO/CLASON',\n 'ARGENTINA/SANTA FE/ROSARIO/VILLA GOB. GALVEZ':'ARGENTINA/SANTA FE/ROSARIO/VILLA GOBERNADOR GALVEZ',\n 'ARGENTINA/SANTA FE/SAN LORENZO/PUERTO GRAL. SAN MARTIN': 'ARGENTINA/SANTA FE/SAN LORENZO/PUERTO GENERAL SAN MARTIN',\n })\n df_safe = df_safe.drop(columns=['DEPARTMENT', 'PLACE'])\n df_safe = df_safe.set_index(['TYPE','LOCATION'])\n df_safe = df_safe.rename(columns=lambda colname: pd.to_datetime(colname,format='%d/%m/%Y'))\n return df_safe\n\ndef _load_data_time_series(df_geoinfo):\n df_arg = _load_National_data(os.path.join(DATA_DIR, 'Argentina_Provinces.csv'))\n df_safe = _load_SantaFe_data(os.path.join(DATA_DIR, 'SantaFe_AllData.csv'))\n df = pd.concat([df_arg,df_safe])\n # Non described dates are 0's\n df = df.fillna(0).sort_index()\n # Set day 0 (prior any date) with all 0's\n day_zero = df.columns[0]-pd.Timedelta(days=1)\n df[day_zero]=0\n df = df[df.columns.sort_values()]\n\n # Add per capita fields\n df_per_capita = pd.merge((df*10000).reset_index(),df_geoinfo[['LOCATION','POPULATION']],on='LOCATION',how='left')\n df_per_capita = df_per_capita.fillna(math.inf).set_index(['TYPE','LOCATION'])\n df_per_capita = df_per_capita.div(df_per_capita['POPULATION'], axis=0)\n df_per_capita = df_per_capita.drop(columns=['POPULATION'])\n df_per_capita.index = df_per_capita.index.map(lambda x : (x[0]+'_PER100K',x[1]) )\n df = pd.concat([df,df_per_capita]).sort_index()\n\n # Calculate number afected subregions\n are_confirmados = df.loc['CONFIRMADOS']>0\n are_confirmados['PARENT_LOCATION'] = are_confirmados.index.map(lambda l : os.path.dirname(l))\n affected_subregions = are_confirmados.groupby('PARENT_LOCATION').sum()\n affected_subregions = affected_subregions.reset_index().rename(columns={'PARENT_LOCATION':'LOCATION'})\n affected_subregions = affected_subregions[ affected_subregions['LOCATION']!='' ]\n affected_subregions['TYPE']='AFFECTED_SUBREGIONS'\n affected_subregions = affected_subregions.set_index(['TYPE','LOCATION'])\n df = pd.concat([df,affected_subregions]).sort_index()\n\n # Calculate difference and differnce ratio with last day\n df_shift = df.shift(axis=1).fillna(0)\n df_diff = df-df_shift\n df_diff.index = df_diff.index.map(lambda x : (x[0]+'_DIFF',x[1]) )\n df_diff_ration = ((df-df_shift)/df_shift).fillna(0)\n df_diff_ration.index = df_diff_ration.index.map(lambda x : (x[0]+'_DIFF_RATIO',x[1]) )\n\n df = pd.concat([df,df_diff,df_diff_ration,affected_subregions])\n\n # Erase non sense columns\n nonsense_columns = [ 'ACTIVOS_PER100K_DIFF_RATIO',\n 'AFFECTED_SUBREGIONS_DIFF_RATIO',\n 'CONFIRMADOS_PER100K_DIFF_RATIO',\n 'MUERTOS_PER100K_DIFF_RATIO',\n 'RECUPERADOS_PER100K_DIFF_RATIO' ]\n df = df[df.index.map(lambda i : i[0] not in nonsense_columns)]\n return df\n\ndef _time_series_melt(df_time_series, df_geoinfo):\n df = pd.melt(df_time_series, id_vars=['TYPE','LOCATION'], value_vars=df_time_series.columns[2:], var_name='date')\n df = df.pivot_table(index=['LOCATION','date'], columns='TYPE', values='value').reset_index()\n df = pd.merge(df,df_geoinfo,on='LOCATION',how='left')\n return df\n\ndef _only_povs(df):\n df = df[ df['LOCATION'].apply(lambda l : l.count('/')==1) ].copy()\n df['LOCATION'] = df['LOCATION'].apply(lambda l : l[10:])\n return df\n\ndef _soon_deprecated_data(df_time_series, df_info):\n df_time_series=_only_povs(df_time_series)\n df_info=_only_povs(df_info)\n df_time_series['2020-03-02 00:00:00']=0.0\n\n df = pd.melt(df_time_series, id_vars=['TYPE','LOCATION'], value_vars=df_time_series.columns[2:], var_name='date')\n df = df[ df['TYPE'].apply(lambda t: t in ['ACTIVOS','CONFIRMADOS','MUERTOS','RECUPERADOS']) ]\n df['TYPE'] = df['TYPE'].replace({\n 'ACTIVOS': 'active',\n 'CONFIRMADOS': 'confirmed',\n 'MUERTOS': 'deceased',\n 'RECUPERADOS': 'recovered',\n })\n df = pd.merge(df,df_info,on='LOCATION')\n df['Province/State']=df['LOCATION']\n df = df.rename(columns={\n 'TYPE':'var',\n 'LAT':'Lat',\n 'LONG':'Long',\n 'LOCATION':'Country/Region',\n 'POPULATION': 'population',\n })\n df = df[ [ 'date', 'Country/Region', 'Province/State', 'var', 'value', 'Lat', 'Long', 'population' ] ]\n df = df.sort_values(by=['Country/Region','date','var'])\n df['value_new'] = df['value'].diff(4)\n df = df.sort_values(by=['date', 'Country/Region','var'])\n df = df[df['date']!='2020-03-02']\n return df\n\ndef _calculate_global_status():\n df_geoinfo = pd.read_csv(os.path.join(DATA_DIR, 'info_general.csv'))\n df_time_series =_load_data_time_series(df_geoinfo).reset_index()\n df_time_series_melt = _time_series_melt(df_time_series,df_geoinfo)\n return {\n 'timestamp': datetime.datetime.today().strftime('%Y-%m-%d-%H:%M:%S'),\n 'geoinfo': df_geoinfo,\n 'time_series': df_time_series,\n 'time_series_melt': df_time_series_melt,\n 'soon_deprecated': _soon_deprecated_data(df_time_series, df_geoinfo)\n }\n\n_global_status = None\ndef backend_update_data():\n global _global_status\n print(\"Updating backend...\")\n _download_expired_data()\n _global_status = _calculate_global_status()\n\ndef backend_global_status_getter(field):\n global _global_status\n return _global_status[field]\n\ndef backend_data_at_date(date):\n global _global_status\n return _global_status['time_series'][date].swaplevel(0,1).unstack()\n\ndef backend_filter_location_by_level(df, level, extract_name=True):\n if level=='LEAF':\n have_childs = set(df['LOCATION'].apply(lambda l : os.path.dirname(l)))\n df = df[ df['LOCATION'].apply(lambda l : l not in have_childs) ]\n else:\n to_level_map = { 'COUNTRY': 0,\n 'PROVINCE': 1,\n 'DEPARTMENT': 2,\n 'CITY': 3 }\n if type(level)==str:\n level = to_level_map[level]\n df = df[ df['LOCATION'].apply(lambda l : l.count('/')) == level ]\n if extract_name:\n df['LOCATION'] = df['LOCATION'].apply(lambda l : os.path.basename(l))\n return df\n", "sub_path": "backend.py", "file_name": "backend.py", "file_ext": "py", "file_size_in_byte": 8628, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "requests.get", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 30, "usage_type": "call"}, {"api_name": "os.stat", "line_number": 30, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 35, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 39, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 45, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 59, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path", "line_number": 74, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 75, "usage_type": "call"}, {"api_name": "pandas.Timedelta", "line_number": 79, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 84, "usage_type": "call"}, {"api_name": "math.inf", "line_number": 85, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 93, "usage_type": "call"}, {"api_name": "os.path", "line_number": 93, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 99, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 108, "usage_type": "call"}, {"api_name": "pandas.melt", "line_number": 120, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 122, "usage_type": "call"}, {"api_name": "pandas.melt", "line_number": 135, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 143, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 160, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 160, "usage_type": "call"}, {"api_name": "os.path", "line_number": 160, "usage_type": "attribute"}, {"api_name": "datetime.datetime.today", "line_number": 164, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 164, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 188, "usage_type": "call"}, {"api_name": "os.path", "line_number": 188, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 199, "usage_type": "call"}, {"api_name": "os.path", "line_number": 199, "usage_type": "attribute"}]} +{"seq_id": "308891261", "text": "from jieba import lcut\nimport re\nimport sys\nimport numpy as np\nfrom memory_profiler import profile\n\n\n# coding=utf-8\n# 读取目标文档\n@profile\ndef readText(path):\n f = open(path, 'r', encoding='utf-8')\n text = f.read()\n # text = []\n # for line in f.readlines(): # 依次读取每行\n # line = line.strip() # 去掉每行头尾空白\n # if not len(line) or line.startswith('#'): # 判断是否是空行或注释行\n # continue\n # text.append(line)\n # text.sort()\n return text # 排序后将文档返回\n\n\n# 利用结巴算法对目标文档进行“分词”处理\n@profile\ndef cut(text):\n words = []\n seg_list = lcut(text, cut_all=False) # 使用jieba下的lcut()方法,精确分割,返回一个列表\n pat = re.compile(u'[a-zA-Z0-9\\u4e00-\\u9fa5]').sub(\" \", \"\") # 将正则表达式转换为内部格式,提高执行效率\n for word in seg_list:\n if re.match(pat, word):\n words.append(word) # 筛选出不含标点符号的结果\n else:\n pass\n return words\n\n\n# 对于两个文档中相同词语进行追加合并进列表\n\ndef mergeWords(t1, t2):\n MergeWords = []\n for i in t1:\n MergeWords.append(i)\n for i in t2:\n if i not in MergeWords:\n MergeWords.append(i)\n return MergeWords\n\n\n# 分别统计两个文档关键词和词频并合并结果转化为向量(vector)\n\ndef countWords(MergeWords, t1, t2):\n list1 = [0 for i in range(len(MergeWords))] # 设定长度并赋值为0\n count1 = dict(zip(MergeWords, list1)) # 设置一个具有���并后词列表的键,但值为零的字典\n for x in t1:\n if x in MergeWords:\n count1[x] += 1 # 遍历合并列表,计算出词频\n else:\n pass\n\n list2 = [0 for i in range(len(MergeWords))]\n count2 = dict(zip(MergeWords, list2))\n for y in t2:\n if y in MergeWords:\n count2[y] += 1\n else:\n pass\n vec1 = list(count1.values()) # 将字典转化为列表类型\n vec2 = list(count2.values())\n return vec1, vec2\n\n\n# 通过向量计算余弦相似度(cosine_similarity)\n\ndef cosine_similarity(v1, v2):\n a = np.array(v1) # 将向量列表转化为数组形式\n b = np.array(v2)\n ma = np.linalg.norm(a) # np.linalg.norm()对数组求整体元素的平方和开根号\n mb = np.linalg.norm(b)\n sim = (np.matmul(a, b)) / (ma * mb) # np.matmul()方法计算内积,结果为余弦相似度\n return sim\n\n\n# 主函数,其包含调用其他函数\n\ndef Main(p1, p2, f) -> object:\n try:\n t1 = cut(readText(p1))\n t2 = cut(readText(p2))\n\n mw = mergeWords(t1, t2)\n v1, v2 = countWords(mw, t1, t2)\n result = cosine_similarity(v1, v2)\n result = np.float(result)*100\n result = round(result, 2) # 保留小数点后两位\n print(\"文本相似度为:\"+str(result)+\"%\")\n fh = open(f, \"a\", encoding='utf-8')\n fh.write(str(p1) + \"与\" + str(p2) + \"的相似度:\" + str(result)+\"%\")\n fh.close()\n except FileNotFoundError:\n print(\"文件不存在!\")\n\n\n# 函数入口\nif __name__ == '__main__':\n path1 = \"\"\n path2 = \"\"\n file_save = \"\"\n try:\n path1 = sys.argv[1] # 实现与命令行交互\n path2 = sys.argv[2]\n file_save = sys.argv[3]\n except IndexError:\n path1 = input(\"请输入正版文件路径:\")\n path2 = input(\"请输入抄袭文件路径:\")\n file_save = input(\"请输入你要保存结果的路径:\")\n Main(path1, path2, file_save)\n", "sub_path": "3119005430/main/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3601, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "memory_profiler.profile", "line_number": 10, "usage_type": "name"}, {"api_name": "jieba.lcut", "line_number": 28, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 29, "usage_type": "call"}, {"api_name": "re.match", "line_number": 31, "usage_type": "call"}, {"api_name": "memory_profiler.profile", "line_number": 25, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 78, "usage_type": "attribute"}, {"api_name": "numpy.linalg.norm", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 79, "usage_type": "attribute"}, {"api_name": "numpy.matmul", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.float", "line_number": 94, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 110, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 111, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 112, "usage_type": "attribute"}]} +{"seq_id": "299678464", "text": "import argparse\nfrom cocos.director import director\n\n\nparser = argparse.ArgumentParser()\nparser.add_argument('--sprite', nargs=1)\nparser.add_argument('--bg', nargs=1)\nparser.add_argument('--level', nargs=1)\nargs = parser.parse_args()\n\ndirector.init()\nscene = None\n\nif args.sprite is not None:\n # To run this test:\n # $ python . --sprite game.sprite.test.StickSprite.run --bg 255,255,255,255\n from importlib import import_module\n from game.scene.tests import SpriteTestScene\n\n module, sheet, sprite = args.sprite[0].rsplit('.', 2)\n module = import_module(module)\n sheet = getattr(module, sheet)\n sprite = getattr(sheet, sprite)\n bg = args.bg[0].split(',') if args.bg else None\n\n scene = SpriteTestScene(sprite, bg)\n\nelif args.level is not None:\n # To run this test:\n # $ python . --level 1\n from game.scene.level import LevelScene\n scene = LevelScene(args.level[0])\n\nelse:\n from game.scene import first_scene\n scene = first_scene()\n\ndirector.run(scene)\n", "sub_path": "__main__.py", "file_name": "__main__.py", "file_ext": "py", "file_size_in_byte": 1000, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 5, "usage_type": "call"}, {"api_name": "cocos.director.director.init", "line_number": 11, "usage_type": "call"}, {"api_name": "cocos.director.director", "line_number": 11, "usage_type": "name"}, {"api_name": "importlib.import_module", "line_number": 21, "usage_type": "call"}, {"api_name": "game.scene.tests.SpriteTestScene", "line_number": 26, "usage_type": "call"}, {"api_name": "game.scene.level.LevelScene", "line_number": 32, "usage_type": "call"}, {"api_name": "game.scene.first_scene", "line_number": 36, "usage_type": "call"}, {"api_name": "cocos.director.director.run", "line_number": 38, "usage_type": "call"}, {"api_name": "cocos.director.director", "line_number": 38, "usage_type": "name"}]} +{"seq_id": "67447102", "text": "import json\nimport streamlit as st\nfrom hydralit import HydraHeadApp\nimport time\nimport requests\nimport logging\n\nURL = \"http://0.0.0.0:8080\"\ncurrent_time = time.strftime(\"%H%M%S-%d%M%y\")\nfile_name = \"transcript_\" + str(current_time)\n\n\nclass SpeechToTextApp(HydraHeadApp):\n\n def __init__(self, title=\"\", **kwargs):\n self.__dict__.update(kwargs)\n self.title = title\n self.logger = logging.getLogger(__name__)\n\n def run(self):\n\n try:\n # ----------------------------------------------------------------\n # Show display of Speech To Text app\n st.title(\"Speech To Text\")\n st.subheader(\"App to transcribe available audio voice to text\")\n st.markdown('

', unsafe_allow_html=True)\n\n _, col2, _ = st.columns((1, 8, 1))\n self.display_app_header(self.title, True)\n # ----------------------------------------------------------------\n\n # ----------------------------------------------------------------\n # selection of acoustic model and language model\n arg_return = self.generate_sidebar()\n\n # upload file audio\n upload_complete, audio_file = self.upload_file(col2)\n\n # transcribe audio\n trans_btn = col2.button(\"Transcribe\")\n\n if upload_complete:\n predict_str = None\n if trans_btn:\n predict_str = self.predict(audio_file)\n\n if predict_str is not None:\n transcript_result = col2.text_area(\"Text Transcript\", value=predict_str[\"word\"], height=300)\n\n if transcript_result is not None:\n self.save_transcript(col2, transcript_result)\n\n except Exception as e:\n st.image(\"./resources/failure.png\", width=100, )\n st.error(\n 'An error has occurred, someone will be punished for your inconvenience, we humbly request you try again.')\n st.error('Error details: {}'.format(e))\n\n def save_transcript(self, col, result):\n if result is not None:\n _result = result.strip()\n\n try:\n col.download_button(label=\"Save transcript\",\n data=_result,\n file_name=f'{file_name}.txt',\n mime='text/csv')\n\n col.success(\"Save transcript successfully\")\n self.logger.info(\"Saved transcript successfully\")\n\n except Exception as e:\n col.error(\"Error saving transcript: {}\".format(e))\n self.logger.error(\"Error saving transcript: {}\".format(e))\n\n def predict(self, audio_file):\n values = {\"file\": (audio_file.name, audio_file, \"audio/wav\")}\n\n st.session_state.text_result = None\n\n if isinstance(values, dict):\n try:\n response = requests.post(f\"{URL}/predict\", files=values)\n st.session_state.text_result = response.json()\n\n if st.session_state.text_result is not None:\n self.logger.info(f\"Predict: {st.session_state.text_result}\")\n else:\n self.logger.warning(\"Predict failed\")\n\n except Exception as e:\n st.error('Error details: {}'.format(e))\n self.logger.error(f\"Error: {e}\")\n\n else:\n st.warning('Predict failed')\n self.logger.info('Predict failed')\n\n return st.session_state.text_result\n\n def generate_sidebar(self):\n if \"predict_arg\" not in st.session_state:\n st.session_state.predict_arg = {\"model\": \"model1\",\n \"lm\": \"CTC + 4-gram\"}\n if \"acoustic_model\" not in st.session_state:\n st.session_state.acoustic_model = \"model1\"\n if \"lm\" not in st.session_state:\n st.session_state.lm_option = \"CTC + 4-gram\"\n\n with st.sidebar:\n acoustic_model_option = st.selectbox(\"Which acoustic model?\",\n ('Model 50k', 'Model 130k'))\n st.info(f\"You selected:{acoustic_model_option}\")\n if acoustic_model_option == \"Model 50k\":\n st.session_state.acoustic_model = \"model1\"\n elif acoustic_model_option == \"Model 130k\":\n st.session_state.acoustic_model = \"model2\"\n st.session_state.lm_option = st.selectbox(\"Which language model?\",\n (\"CTC\", \"CTC + 4-gram\"))\n st.info(f\"You selected:{st.session_state.lm_option}\")\n\n if st.session_state.acoustic_model == \"\" or st.session_state.lm_option == \"\":\n st.sidebar.warning(\"Setting must not be empty\")\n\n if st.session_state.predict_arg != {\"model\": st.session_state.acoustic_model,\n \"lm\": st.session_state.lm_option}:\n st.session_state.predict_arg = {\"model\": st.session_state.acoustic_model,\n \"lm\": st.session_state.lm_option}\n arg_request = json.dumps(st.session_state.predict_arg)\n response = requests.post(f\"{URL}/pattern\", data=arg_request)\n arg_return = response.json()\n\n return arg_return\n\n def upload_file(self, col):\n st.session_state.audio_file = col.file_uploader(\"Upload audio\", type=['wav', 'mp3'])\n if st.session_state.audio_file is not None:\n col.success(\"File uploaded successfully\")\n col.audio(st.session_state.audio_file)\n st.session_state.upload_complete = True\n else:\n col.info(\"Please upload a audio file\")\n st.session_state.upload_complete = False\n\n return st.session_state.upload_complete, st.session_state.audio_file\n\n def display_app_header(self, main_txt, is_sidebar=False):\n \"\"\"\n function to display major headers at user interface\n ----------\n main_txt: str -> the major text to be displayed\n sub_txt: str -> the minor text to be displayed \n is_sidebar: bool -> check if its side panel or major panel\n \"\"\"\n\n html_temp = f\"\"\"\n

{main_txt}

\n
\n \"\"\"\n if is_sidebar:\n st.sidebar.markdown(html_temp, unsafe_allow_html=True)\n else:\n st.markdown(html_temp, unsafe_allow_html=True)\n", "sub_path": "apps/stt_app.py", "file_name": "stt_app.py", "file_ext": "py", "file_size_in_byte": 6551, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "time.strftime", "line_number": 9, "usage_type": "call"}, {"api_name": "hydralit.HydraHeadApp", "line_number": 13, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 18, "usage_type": "call"}, {"api_name": "streamlit.title", "line_number": 25, "usage_type": "call"}, {"api_name": "streamlit.subheader", "line_number": 26, "usage_type": "call"}, {"api_name": "streamlit.markdown", "line_number": 27, "usage_type": "call"}, {"api_name": "streamlit.columns", "line_number": 29, "usage_type": "call"}, {"api_name": "streamlit.image", "line_number": 55, "usage_type": "call"}, {"api_name": "streamlit.error", "line_number": 56, "usage_type": "call"}, {"api_name": "streamlit.error", "line_number": 58, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 80, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 84, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 85, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 87, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 88, "usage_type": "attribute"}, {"api_name": "streamlit.error", "line_number": 93, "usage_type": "call"}, {"api_name": "streamlit.warning", "line_number": 97, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 100, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 103, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 104, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 106, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 107, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 108, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 109, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar", "line_number": 111, "usage_type": "attribute"}, {"api_name": "streamlit.selectbox", "line_number": 112, "usage_type": "call"}, {"api_name": "streamlit.info", "line_number": 114, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 116, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 118, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 119, "usage_type": "attribute"}, {"api_name": "streamlit.selectbox", "line_number": 119, "usage_type": "call"}, {"api_name": "streamlit.info", "line_number": 121, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 121, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 123, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.warning", "line_number": 124, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 124, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 126, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 127, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 128, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 129, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 130, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 130, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 131, "usage_type": "call"}, {"api_name": "streamlit.session_state", "line_number": 137, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 138, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 140, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 141, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 144, "usage_type": "attribute"}, {"api_name": "streamlit.session_state", "line_number": 146, "usage_type": "attribute"}, {"api_name": "streamlit.sidebar.markdown", "line_number": 162, "usage_type": "call"}, {"api_name": "streamlit.sidebar", "line_number": 162, "usage_type": "attribute"}, {"api_name": "streamlit.markdown", "line_number": 164, "usage_type": "call"}]} +{"seq_id": "130151323", "text": "import datetime\nimport glob\nimport os\nfrom jinja2 import Template\n\ndef main():\n pages = []\n# Auto-discovery of content files\n all_content_files = glob.glob(\"content/*.md\")\n\n for page in all_content_files:\n file_name = os.path.basename(page)\n name_only, extension = os.path.splitext(file_name)\n pages.append({\n \"filename\": \"content/\" + file_name,\n \"title\": name_only,\n \"output\": file_name\n })\n# Create final pages using content files with jinja2 templating\n year = datetime.datetime.now().strftime('%Y')\n for page in all_content_files:\n file_name = os.path.basename(page)\n name_only, extension = os.path.splitext(file_name)\n content_page = open(\"content/\" + file_name).read()\n template_html = open(\"templates/base.md\").read()\n template = Template(template_html)\n results = template.render(\n title=name_only,\n content=content_page,\n year=str(year),\n pages=pages,\n )\n open(\"docs/\"+name_only+\".html\",\"w+\").write(results)\n\n# Auto-generated blog pages\n# NOT WORKING!!\n # blog_pages = []\n # all_blog_posts = glob.glob(\"blog/*.md\")\n \n # for blog in all_blog_posts:\n # blog_pages.append(open(blog).read())\n\n # for blog in blog_pages:\n # file_name = os.path.basename(blog)\n # template_html = open(\"templates/blog_base.md\").read()\n # template2 = Template(template_html)\n # results = template.render(\n # blog=blog,\n # year=str(year),\n # )\n # open(\"docs/\"+file_name,\"w+\").write(results)", "sub_path": "utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 1635, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "glob.glob", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 20, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "jinja2.Template", "line_number": 26, "usage_type": "call"}]} +{"seq_id": "358528820", "text": "from __future__ import print_function, unicode_literals\nimport os\nimport sys\nimport tempfile\nimport boto3\nimport json\nfrom glob import glob\nfrom shutil import copyfile\nfrom aws_tools.s3_handler import S3Handler\nfrom general_tools.file_utils import write_file\nfrom door43_tools import templaters\nfrom datetime import datetime, timedelta\n\n\nclass ProjectDeployer(object):\n \"\"\"\n Deploys a project's revision to the door43.org bucket\n\n Read from the project's user dir in the cdn.door43.org bucket\n by applying the door43.org template to the raw html files\n \"\"\"\n\n def __init__(self, cdn_bucket, door43_bucket):\n \"\"\"\n :param string cdn_bucket: \n :param string door43_bucket: \n \"\"\"\n self.cdn_bucket = cdn_bucket\n self.door43_bucket = door43_bucket\n self.cdn_handler = None\n self.door43_handler = None\n self.lambda_client = None\n self.setup_resources()\n\n def setup_resources(self):\n self.cdn_handler = S3Handler(self.cdn_bucket)\n self.door43_handler = S3Handler(self.door43_bucket)\n self.lambda_client = boto3.client('lambda', region_name='us-west-2')\n\n def deploy_revision_to_door43(self, build_log_key):\n \"\"\"\n Deploys a single revision of a project to door43.org\n :param string build_log_key:\n :return bool:\n \"\"\"\n build_log = None\n try:\n build_log = self.cdn_handler.get_json(build_log_key)\n except:\n pass\n\n if not build_log or 'commit_id' not in build_log or 'repo_owner' not in build_log or 'repo_name' not in build_log:\n return False\n\n user = build_log['repo_owner']\n repo_name = build_log['repo_name']\n commit_id = build_log['commit_id'][:10]\n\n s3_commit_key = 'u/{0}/{1}/{2}'.format(user, repo_name, commit_id)\n s3_repo_key = 'u/{0}/{1}'.format(user, repo_name)\n\n source_dir = tempfile.mkdtemp(prefix='source_')\n output_dir = tempfile.mkdtemp(prefix='output_')\n template_dir = tempfile.mkdtemp(prefix='template_')\n\n self.cdn_handler.download_dir(s3_commit_key, source_dir)\n source_dir = os.path.join(source_dir, s3_commit_key)\n\n resource_type = build_log['resource_type']\n if resource_type == 'ulb' or resource_type == 'udb':\n resource_type = 'bible'\n\n # determining the template and templater from the resource_type, use general if not found\n try:\n templater_class = self.str_to_class('templaters.{0}Templater'.format(resource_type.capitalize()))\n template_key = 'templates/{0}.html'.format(resource_type)\n except AttributeError:\n templater_class = templaters.Templater\n template_key = 'templates/obs.html' # Use a generic template here\n\n template_file = os.path.join(template_dir, 'template.html')\n print(\"Downloading {0} to {1}...\".format(template_key, template_file))\n self.door43_handler.download_file(template_key, template_file)\n\n html_files = sorted(glob(os.path.join(source_dir, '*.html')))\n if len(html_files) < 1:\n content = ''\n if len(build_log['errors']) > 0:\n content += \"\"\"\n
\n \n
\n

Critical!

\n

Here is what went wrong with this build:

\n
\n \"\"\"\n content += '
  • ' + '
  • '.join(build_log['errors']) + '
'\n elif len(build_log['warnings']) > 0:\n content += \"\"\"\n
\n \n
\n

Warning!

\n

Here are some problems with this build:

\n
\n \"\"\"\n content += '
  • ' + '
  • '.join(build_log['warnings']) + '
'\n else:\n content += '

{0}

'.format(build_log['message'])\n content += '

No content is available to show for {0} yet.

'.format(repo_name)\n\n html = \"\"\"\n \n \n {0}\n \n \n
{1}
\n \n \"\"\".format(repo_name, content)\n repo_index_file = os.path.join(source_dir, 'index.html')\n write_file(repo_index_file, html)\n\n # merge the source files with the template\n templater = templater_class(source_dir, output_dir, template_file)\n templater.run()\n\n # Copy first HTML file to index.html if index.html doesn't exist\n html_files = sorted(glob(os.path.join(output_dir, '*.html')))\n if len(html_files) > 0:\n index_file = os.path.join(output_dir, 'index.html')\n if not os.path.isfile(index_file):\n copyfile(os.path.join(output_dir, html_files[0]), index_file)\n\n # Copy all other files over that don't already exist in output_dir, like css files\n for filename in sorted(glob(os.path.join(source_dir, '*'))):\n output_file = os.path.join(output_dir, os.path.basename(filename))\n if not os.path.exists(output_file) and not os.path.isdir(filename):\n copyfile(filename, output_file)\n\n # Upload all files to the door43.org bucket\n for root, dirs, files in os.walk(output_dir):\n for f in sorted(files):\n path = os.path.join(root, f)\n if os.path.isdir(path):\n continue\n key = s3_commit_key + path.replace(output_dir, '')\n print(\"Uploading {0} to {1}\".format(path, key))\n self.door43_handler.upload_file(path, key, 0)\n\n # Now we place json files and make an index.html file for the whole repo\n try:\n self.door43_handler.copy(from_key='{0}/project.json'.format(s3_repo_key), from_bucket=self.cdn_bucket)\n self.door43_handler.copy(from_key='{0}/manifest.json'.format(s3_commit_key), to_key='{0}/manifest.json'.format(s3_repo_key))\n self.door43_handler.redirect(s3_repo_key, '/' + s3_commit_key)\n self.door43_handler.redirect(s3_repo_key + '/index.html', '/' + s3_commit_key)\n except Exception:\n pass\n\n return True\n\n def redeploy_all_projects(self, deploy_function):\n i = 0\n one_day_ago = datetime.utcnow() - timedelta(hours=24)\n print(one_day_ago)\n for obj in self.cdn_handler.get_objects(prefix='u/', suffix='build_log.json'):\n i += 1\n last_modified = obj.last_modified.replace(tzinfo=None)\n if one_day_ago <= last_modified:\n continue\n print(\"{0}: {1} {2}\".format(i, obj.key, last_modified))\n self.lambda_client.invoke(\n FunctionName=deploy_function,\n InvocationType='Event',\n LogType='Tail',\n Payload=json.dumps({\n 'cdn_bucket': self.cdn_bucket,\n 'build_log_key': obj.key\n })\n )\n return True\n\n def str_to_class(self, str):\n \"\"\"\n Gets a class from a string.\n\n :param str|unicode str: The string of the class name\n \"\"\"\n return reduce(getattr, str.split(\".\"), sys.modules[__name__])\n", "sub_path": "door43_tools/project_deployer.py", "file_name": "project_deployer.py", "file_ext": "py", "file_size_in_byte": 7849, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "aws_tools.s3_handler.S3Handler", "line_number": 36, "usage_type": "call"}, {"api_name": "aws_tools.s3_handler.S3Handler", "line_number": 37, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 38, "usage_type": "call"}, {"api_name": "tempfile.mkdtemp", "line_number": 62, "usage_type": "call"}, {"api_name": "tempfile.mkdtemp", "line_number": 63, "usage_type": "call"}, {"api_name": "tempfile.mkdtemp", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "door43_tools.templaters.Templater", "line_number": 78, "usage_type": "attribute"}, {"api_name": "door43_tools.templaters", "line_number": 78, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path", "line_number": 121, "usage_type": "attribute"}, {"api_name": "general_tools.file_utils.write_file", "line_number": 122, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 129, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 129, "usage_type": "call"}, {"api_name": "os.path", "line_number": 129, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 131, "usage_type": "call"}, {"api_name": "os.path", "line_number": 131, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 132, "usage_type": "call"}, {"api_name": "os.path", "line_number": 132, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 133, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 133, "usage_type": "call"}, {"api_name": "os.path", "line_number": 133, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 136, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 136, "usage_type": "call"}, {"api_name": "os.path", "line_number": 136, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 137, "usage_type": "call"}, {"api_name": "os.path", "line_number": 137, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 137, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path", "line_number": 138, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 138, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 139, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 142, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 144, "usage_type": "call"}, {"api_name": "os.path", "line_number": 144, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path", "line_number": 145, "usage_type": "attribute"}, {"api_name": "datetime.datetime.utcnow", "line_number": 164, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 164, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 164, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 176, "usage_type": "call"}, {"api_name": "sys.modules", "line_number": 189, "usage_type": "attribute"}]} +{"seq_id": "233163770", "text": "# .......................... Connect to each Sensor of the Node ...........................\nimport csv\nimport board\nimport busio\nimport serial\nimport adafruit_bme280\nimport struct\n#import os\nimport json\nimport datetime\nimport time\nimport smtplib\nfrom subprocess import check_output\n#import pandas as pd\nimport smtplib\nimport json\nimport datetime\n\ndef mail_alert1(): \n fromaddr = email\n toaddrs = email\n msg = 'Subject: {}\\n\\n{}'.format('Indoor Unit ' + sensorParameters['ID'] + ' ' +'PMS5003 error', 'Sensor' + '_' + sensorParameters['ID'] + ' PMS5003 error at ' + currentTime.strftime('%Y%m%d_%H%M%S'))\n\n# Credentials (if needed)\n username = email\n password = pw\n\n# The actual mail send\n server = smtplib.SMTP('smtp.gmail.com:587')\n server.starttls()\n server.login(username,password)\n server.sendmail(fromaddr, toaddrs, msg)\n server.quit()\n\ndef mail_alert2(): \n fromaddr = email\n toaddrs = email_2\n msg = 'Subject: {}\\n\\n{}'.format('Indoor Unit ' + sensorParameters['ID'] + ' ' +'PMS5003 error', 'Sensor' + '_' + sensorParameters['ID'] + ' PMS5003 error at ' + currentTime.strftime('%Y%m%d_%H%M%S'))\n\n# Credentials (if needed)\n username = email\n password = pw\n\n# The actual mail send\n server = smtplib.SMTP('smtp.gmail.com:587')\n server.starttls()\n server.login(username,password)\n server.sendmail(fromaddr, toaddrs, msg)\n server.quit()\n\nwith open('sensorParameters.json') as json_file:\n sensorParameters=json.load(json_file)\n\nwith open('/home/pi/SpokaneSchools/software/Name_1.txt','r') as file:\n email=file.read()\n \nwith open('/home/pi/SpokaneSchools/software/Name_3.txt','r') as file:\n email_2=file.read()\n\nwith open('/home/pi/SpokaneSchools/software/Name_2.txt','r') as file:\n pw=file.read()\n\n\n#Once JSON file is created, open the file to read in sensorParameters\nwith open('/home/pi/SpokaneSchools/Cloud/sensorParameters.json') as json_file:\n sensorParameters=json.load(json_file)\n\n# Create a unique filename for the current date.\ncurrentHour = datetime.datetime.now().hour\ncurrentTime = datetime.datetime.now()\ncurrentDate = currentTime.date()\nfilename = sensorParameters['name'] + '_' + sensorParameters['ID'] + '_' +currentTime.strftime('%Y%m%d_%H%M%S') + '.csv'\n\n### Initialize variables to store in CSV file.\ndata_file = []\n\nwith open('/home/pi/SpokaneSchools/Data/Default_Frequency/' + filename, 'w') as f:\n writer = csv.DictWriter(f, fieldnames = [\"Datetime\", \"PM_0_3\", \"PM_0_5\", 'PM_1', 'PM_2_5', 'PM_5', 'PM_10', 'PM1_standard', 'PM2_5_standard', 'PM10_standard', 'PM1_env', 'PM2_5_env', 'PM10_env'])\n writer.writeheader()\n #init_headers.to_csv(f, header=True)\n f.close()\n\n\n#### Initialize Sensors\nuart = serial.Serial(\"/dev/ttyS0\", baudrate=9600, timeout=3000)\nbuffer = []\n\n# .......................... Acquire and Store Sensor Data ...........................\nwhile True:\n try:\n # If new day, then close current file and open a new file.\n if (datetime.datetime.now().date() != currentDate):\n with open('/home/pi/SpokaneSchools/Data/Default_Frequency/' + filename, 'a') as f:\n wr = csv.writer(f, delimiter = ',')\n wr.writerows(data_file)\n f.close()\n currentHour = datetime.datetime.now().hour\n currentTime = datetime.datetime.now()\n currentDate = currentTime.date()\n data_file = []\n filename = sensorParameters['name'] + '_' + sensorParameters['ID'] + '_' +currentTime.strftime('%Y%m%d_%H%M%S') + '.csv'\n with open('/home/pi/SpokaneSchools/Data/Default_Frequency/' + filename, 'w') as f:\n writer = csv.DictWriter(f, fieldnames = ['Datetime', 'PM_0_3', 'PM_0_5', 'PM_1', 'PM_2_5', 'PM_5', 'PM_10', 'PM1_standard', 'PM2_5_standard', 'PM10_standard', 'PM1_env', 'PM2_5_env', 'PM10_env'])\n writer.writeheader()\n f.close()\n \n if datetime.datetime.now().hour != currentHour:\n with open('/home/pi/SpokaneSchools/Data/Default_Frequency/' + filename, 'a') as f:\n #data_save.to_csv(f, header=False)\n wr = csv.writer(f, delimiter = ',')\n wr.writerows(data_file)\n #np.savetxt(f, data, delimiter = ',')\n f.close()\n currentHour = datetime.datetime.now().hour\n data_file = []\n # Attempts to acquire and decode the data from the PMS5003 particulate matter sensor\n data = uart.read(32) # read up to 32 bytes\n data = list(data)\n\n buffer += data\n while buffer and buffer[0] != 0x42:\n buffer.pop(0)\n\n if len(buffer) > 200:\n buffer = [] # avoid an overrun if all bad data\n if len(buffer) < 32:\n continue\n\n if buffer[1] != 0x4d:\n buffer.pop(0)\n continue\n\n frame_len = struct.unpack(\">H\", bytes(buffer[2:4]))[0]\n if frame_len != 28:\n buffer = []\n continue\n \n frame = struct.unpack(\">HHHHHHHHHHHHHH\", bytes(buffer[4:]))\n\n pm10_standard, pm25_standard, pm100_standard, pm10_env, \\\n pm25_env, pm100_env, particles_03um, particles_05um, particles_10um, \\\n particles_25um, particles_50um, particles_100um, skip, checksum = frame\n \n check = sum(buffer[0:30])\n \n if check != checksum:\n buffer = []\n continue\n\n # create error to see if email is sent (comment out after confirming function)\n #error = unknown_variable\n \n data_line = [datetime.datetime.now().isoformat(), particles_03um, particles_05um, particles_10um, particles_25um, particles_50um, particles_100um, pm10_standard, pm25_standard, pm100_standard, pm10_env, pm25_env, pm100_env]\n data_file.append(data_line)\n # print(type(data_line))\n # print(type(data))\n print(data_line)\n # print(data_file)\n print(\"Current time: \", datetime.datetime.now())\n\n buffer = buffer[32:]\n \n # Close JSON file\n \n except:\n \n currentTime = datetime.datetime.now()\n\n mail_alert1()\n #mail_alert2()\n time.sleep(3600)\n", "sub_path": "python/monitoring/PMS5003_only.py", "file_name": "PMS5003_only.py", "file_ext": "py", "file_size_in_byte": 6294, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "smtplib.SMTP", "line_number": 29, "usage_type": "call"}, {"api_name": "smtplib.SMTP", "line_number": 45, "usage_type": "call"}, {"api_name": "json.load", "line_number": 52, "usage_type": "call"}, {"api_name": "json.load", "line_number": 66, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 69, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 69, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 70, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 70, "usage_type": "attribute"}, {"api_name": "csv.DictWriter", "line_number": 78, "usage_type": "call"}, {"api_name": "serial.Serial", "line_number": 85, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 92, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 92, "usage_type": "attribute"}, {"api_name": "csv.writer", "line_number": 94, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 97, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 97, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 98, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 98, "usage_type": "attribute"}, {"api_name": "csv.DictWriter", "line_number": 103, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 107, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 107, "usage_type": "attribute"}, {"api_name": "csv.writer", "line_number": 110, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 114, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 114, "usage_type": "attribute"}, {"api_name": "struct.unpack", "line_number": 133, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 138, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 153, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 153, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 159, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 159, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 167, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 167, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 171, "usage_type": "call"}]} +{"seq_id": "107681595", "text": "# Program: nearestNeighbor.py\n# Author: OSU Fa16 CS325-400 Project Group 12\n# Description: Greedy TSP algorithm that chooses the next destination\n# from the closest destination among those that haven't\n# yet been visited.\n# Usage: From the command line type the following...\n# /usr/bin/python nearestNeighbor.py inputFileName.txt\n\nimport datetime\nimport math\nimport os\nimport sys\n\n# -----------------------------------------------------------------------------\t\ndef dist(cityOne, cityTwo):\n\tdx = cityOne['x'] - cityTwo['x']\n\tdy = cityOne['y'] - cityTwo['y']\n\tdxSq = math.pow(dx, 2)\n\tdySq = math.pow(dy, 2)\n\treturn int(round(math.sqrt(dxSq + dySq)))\n\n# -----------------------------------------------------------------------------\t\ndef main():\n\t# get input file name from command line\n\tif (len(sys.argv) != 2):\n\t\tprint ('ERROR: Exactly one argument expected. See usage instructions.\\n')\n\t\tquit()\n\telse:\n\t\tinFil = sys.argv[1]\n\t\tif not(os.path.isfile(inFil)):\n\t\t\tprint ('ERROR: File \\'' + str(inFil) + '\\' not found.\\n')\n\t\t\tquit()\n\n\t# open input/output files\n\tbase = os.path.basename(inFil)\n\tinFil = open(base, 'r')\n\t# outFil = open(base + '.tour', 'w')\n\n\t# get cities from input file into a list\n\tcities = []\n\tfor eachLine in inFil:\n\t\teachCity = eachLine.split()\n\t\tthisCity = {'id':int(eachCity[0]), 'x':int(eachCity[1]), 'y':int(eachCity[2])}\n\t\tcities.append(thisCity)\n#\t\tprint(thisCity)\n\tinFil.close()\n\n\t# init adjacency matrix for graph (every city connected to every other city)\n\tadjMatrix = [[-1 for x in range(len(cities))] for y in range(len(cities))]\n\t\n\t# look for tour starting at each possible vertex\n\tminTourDist = sys.maxsize\n\tminTourOrder = []\n\tfor i in range(0, len(cities)):\n#\t\tprint(\"starting at: \" + str(i))\n\t\ttourCities = [x for x in cities]\n\t\ttourOrder = []\n\t\ttourOrder.append(tourCities[i]['id'])\n#\t\tprint(\"tourOrder: \" + str(tourOrder))\n\t\ttourCities.remove(tourCities[i])\n\t\ttourDist = 0\n\t\twhile len(tourCities) > 0:\n#\t\t\tif len(tourCities) % 1000 == 0:\n#\t\t\t\tprint (str(len(tourCities)))\n\t\t\t# find closest city\n\t\t\tthisCity = cities[tourOrder[len(tourOrder) - 1]]\n#\t\t\tprint(\"thisCity: \" + str(thisCity))\n\t\t\tminDist = sys.maxsize\n\t\t\tminCity = -1\n#\t\t\tprint(\"tourCities: \" + str(tourCities))\n\t\t\tfor j in range(0, len(tourCities)):\n\t\t\t\tthisDist = adjMatrix[thisCity['id']][tourCities[j]['id']]\n#\t\t\t\tprint(\"thisDist: \" + str(thisDist))\n\t\t\t\tif thisDist == -1:\n\t\t\t\t\tthisDist = dist(thisCity, tourCities[j])\n#\t\t\t\t\tprint(\"thisDist: \" + str(thisDist))\n\t\t\t\t\tadjMatrix[thisCity['id']][tourCities[j]['id']] = thisDist\n\t\t\t\t\tadjMatrix[tourCities[j]['id']][thisCity['id']] = thisDist\n\t\t\t\tif thisDist < minDist:\n\t\t\t\t\tminDist = thisDist\n\t\t\t\t\tminCity = tourCities[j]\n\t\t\ttourOrder.append(minCity['id'])\n#\t\t\tprint(\"minDist: \" + str(minDist))\n#\t\t\tprint(\"tourOrder: \" + str(tourOrder))\n\t\t\ttourCities.remove(minCity)\n\t\t\ttourDist = tourDist + minDist\n\n\t\tu = cities[tourOrder[0]]['id']\n\t\tv = cities[tourOrder[len(tourOrder) - 1]]['id']\n\t\tthisDist = adjMatrix[u][v]\n\t\tif thisDist == -1:\n\t\t\tthisDist = dist(cities[tourOrder[0]], cities[tourOrder[len(tourOrder) - 1]])\n\t\t\tadjMatrix[u][v] = thisDist\n\t\t\tadjMatrix[v][u] = thisDist\n\t\t\t\n\t\ttourDist = tourDist + thisDist\n#\t\tprint(\"tourOrder: \" + str(tourOrder))\n#\t\tprint(\"starting at: \" + str(i) + \" thisDist:\" + str(tourDist))\n\t\tif tourDist < minTourDist:\n\t\t\tprint(str(datetime.datetime.now()) + \" starting at: \" + str(i) + \" thisDist:\" + str(tourDist))\n\t\t\tminTourDist = tourDist\n\t\t\tminTourOrder = [x for x in tourOrder]\n\n\t\t\t# write output to file\n\t\t\toutFil = open(base + '.tour', 'w')\n\t\t\toutFil.write(str(minTourDist) + '\\n')\n\t\t\titerCities = iter(minTourOrder)\n\t\t\tfor eachCity in iterCities:\n\t\t\t\toutFil.write(str(eachCity) + '\\n')\n\t\t\toutFil.close()\n#\t\telse:\n#\t\t\tprint(str(datetime.datetime.now()) + \" starting at: \" + str(i))\n\n# -----------------------------------------------------------------------------\t\nif __name__ == \"__main__\":\n\tmain()\n", "sub_path": "Alg04_Nearest_Neighbor/nearestNeighbor.py", "file_name": "nearestNeighbor.py", "file_ext": "py", "file_size_in_byte": 3995, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "math.pow", "line_number": 18, "usage_type": "call"}, {"api_name": "math.pow", "line_number": 19, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 20, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 25, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "sys.maxsize", "line_number": 52, "usage_type": "attribute"}, {"api_name": "sys.maxsize", "line_number": 68, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 100, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 100, "usage_type": "attribute"}]} +{"seq_id": "601241735", "text": "#!/usr/bin/env python\nimport MySQLdb\nimport subprocess\nimport time\nfrom flask import Flask\nfrom flask import request\nimport requests\nimport re\nimport sys\nimport json\nimport urllib\ndef send_gm(message):\n\tat=\"68ea42109ec80134adf205b7f1deccdf\"\n\ttest_ID='27306241'\n\turl_send='https://api.groupme.com/v3/groups/'+test_ID+'/messages?token='+at\n\t#message='Fuck+off+jason'\n\turl='https://api.groupme.com/v3/bots/post?bot_id=481baace72a55ebbb9488e296e&text='+message\n\tprint(url)\n\tr=requests.post(url)\n\tprint(r)\n \napp=Flask(__name__)\n@app.route('/',methods=['GET','POST'])\ndef index():\n\tif request.method=='POST':\n\t\ttry:\n\t\t\tdata=request.form['data']\n\t\t\tdata=data.decode('hex')\n\t\t\tmomsn=request.form['momsn']\n\t\t\ttransmit_time=request.form['transmit_time']\n\t\t\tiridium_lat=request.form['iridium_latitude']\n\t\t\tiridium_lon=request.form['iridium_longitude']\n\t\t\tir_cep=request.form['iridium_cep']\n\t\t\tprint(\"Message Number: \"+str(momsn))\n\t\t\tprint(\"Lat: \"+str(iridium_lat))\n\t\t\tprint(\"Lon: \"+str(iridium_lon))\n\t\t\tprint(\"Accuracy: \"+str(ir_cep))\n\t\t\tprint(\"Time: \"+str(transmit_time))\n\t\t\tprint(\"Message: \"+str(data))\n\t\t\tif str(data).startswith(\"$\"):\n\t\t\t\tdata=data[1:]#strip off the leading $\n\t\t\t\tdset=data.split(\",\")\n\t\t\t\tLat=dset[0]\n\t\t\t\tLon=dset[1]\n\t\t\t\tVoltage=dset[2]\n\t\t\t\tStatus=dset[3]\n\t\t\t\tsend_gm(\"Update:\\rLat: \"+str(Lat)+\"\\rLon: \"+str(Lon)+\"\\rVoltage: \"+str(Voltage)+\"\\rStatus: \"+str(Status))\n\t\t\t\tsend_gm(\"http://maps.google.com/maps?q=\"+str(Lat)+\",\"+str(Lon))\n\t\t\telif str(data)==\"\":\n\t\t\t\t#I think a blank message is sent when tring to retreve a message\n\t\t\t\t#either way, no need to send a blank message to groupme\n\t\t\t\treturn \"done\",200\t\n\t\t\telse:\n\t\t\t\tsend_gm(\"Message Received from SatCom: \"+str(data))\t\t\n\t\t\t\tsend_gm(\"Estimated Location: Lat: \"+str(iridium_lat)+\" Lon: \"+str(iridium_lon))\n\t\texcept:\n\t\t\ttry:\n\t\t\t\to=open('/home/pi/SatComBox/WS_Out','w')\n\t\t\t\tdata=request.json\n\t\t\t\tname=data['name']\n\t\t\t\tmessage=data['text']\n\t\t\t\tname=str(name)\n\t\t\t\tmessage=str(message)\n\t\t\t\tif message[0]==\"$\":\n\t\t\t\t\to.write(\"got a $\\n\")\n\t\t\t\t\tif \"Chad\" in name:\n\t\t\t\t\t\to.write(\"and its from me\\n\")\n\t\t\t\t\t\t#send message to sat\n\t\t\t\t\t\tIMEI='300234064380130'\n\t\t\t\t\t\tNAME=\"gibeautc@oregonstate.edu\"\n\t\t\t\t\t\tPASSWORD='myvice12'\n\t\t\t\t\t\tDATA=message[1:]\n\t\t\t\t\t\tparams=urllib.urlencode({'imei':IMEI,'username':NAME,'password':PASSWORD,'data':DATA.encode(\"hex\")})\n\t\t\t\t\t\tf=urllib.urlopen(\"https://core.rock7.com/rockblock/MT\",params)\n\t\t\t\t\t\tprint(f.read())\n\t\t\t\t\t\to.write(str(f.read()))\n\t\t\t\t\t\tsend_gm('Thanks Chad, your message of: \"'+DATA+'\" has been sent to the que')\n\t\t\t\t\telse:\n\t\t\t\t\t\t#not authorized\n\t\t\t\t\t\ttry:\n\t\t\t\t\t\t\tname=name.split(\" \")\n\t\t\t\t\t\t\tname=name[0]\n\t\t\t\t\t\texcept:\n\t\t\t\t\t\t\tname=name\n\t\t\t\t\t\tsend_gm(\"Sorry \"+name+\" , you are not authorized to send messages at this time.....please fuck off\")\n\t\t\t\t#print(\"GM Message from: \"+str(name)+\" ----:\"+str(message))\n\t\t\t\to.close()\n\t\t\texcept:\n\t\t\t\to.write(\"Failed 1\")\n\t\t\t\to.close()\n\t\t\n\t\treturn \"done\",200\nif __name__ == '__main__':\n\tapp.run(debug=True, host='0.0.0.0')\n", "sub_path": "web_service.py", "file_name": "web_service.py", "file_ext": "py", "file_size_in_byte": 2952, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "requests.post", "line_number": 19, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 22, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 25, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 25, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 27, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 27, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 29, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 29, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 30, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 30, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 31, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 31, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 32, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 32, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 33, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 33, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 59, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 59, "usage_type": "name"}, {"api_name": "urllib.urlencode", "line_number": 73, "usage_type": "call"}, {"api_name": "urllib.urlopen", "line_number": 74, "usage_type": "call"}]} +{"seq_id": "271772592", "text": "import numpy as np\nimport netCDF4\nimport matplotlib.pyplot as plt# ; plt.close('all')\nimport bisect\n\nfrom phdpy import settings\nsettings.set_rc()\n\n\n\ndsid = 'x1'\n\ndatafile = settings.datafiles[dsid]\n\nwith netCDF4.Dataset(datafile) as ds:\n \n ang = ds.variables['ANGLE'][:]\n lon = ds.variables['ULONG'][:]\n lat = ds.variables['ULAT'][:]\n \n meanlat = np.mean(lat,axis=1)\n \n fig = plt.figure()\n ax = fig.gca()\n\n for lat0 in np.arange(40,61,5):\n \n j = np.min((bisect.bisect_left(meanlat,lat0),len(meanlat)-1))\n\n ax.plot(np.mod(lon[j,:],360),np.degrees(ang[j,:]),label='{}'.format(lat0))\n \n ax.legend(loc=0)\n\n plt.show()\n", "sub_path": "other/grid_angle.py", "file_name": "grid_angle.py", "file_ext": "py", "file_size_in_byte": 682, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "phdpy.settings.set_rc", "line_number": 7, "usage_type": "call"}, {"api_name": "phdpy.settings", "line_number": 7, "usage_type": "name"}, {"api_name": "phdpy.settings.datafiles", "line_number": 13, "usage_type": "attribute"}, {"api_name": "phdpy.settings", "line_number": 13, "usage_type": "name"}, {"api_name": "netCDF4.Dataset", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 28, "usage_type": "call"}, {"api_name": "bisect.bisect_left", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.mod", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.degrees", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}]} +{"seq_id": "156702334", "text": "# Copyright 2013 Red Hat, Inc.\n# All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\nimport collections\n\nfrom oslo_concurrency import lockutils\nfrom oslo_log import log\nfrom stevedore import dispatch\n\nfrom ironic.common import exception\nfrom ironic.common.i18n import _, _LI, _LW\nfrom ironic.conf import CONF\nfrom ironic.drivers import base as driver_base\nfrom ironic.drivers import fake_hardware\nfrom ironic.drivers import hardware_type\n\n\nLOG = log.getLogger(__name__)\n\nEM_SEMAPHORE = 'extension_manager'\n\n\ndef build_driver_for_task(task, driver_name=None):\n \"\"\"Builds a composable driver for a given task.\n\n Starts with a `BareDriver` object, and attaches implementations of the\n various driver interfaces to it. For classic drivers these all come from\n the monolithic driver singleton, for hardware types - from separate\n driver factories and are configurable via the database.\n\n :param task: The task containing the node to build a driver for.\n :param driver_name: The name of the classic driver or hardware type to use\n as a base, if different than task.node.driver.\n :returns: A driver object for the task.\n :raises: DriverNotFound if node.driver could not be found in either\n \"ironic.drivers\" or \"ironic.hardware.types\" namespaces.\n :raises: InterfaceNotFoundInEntrypoint if some node interfaces are set\n to invalid or unsupported values.\n :raises: IncompatibleInterface if driver is a hardware type and\n the requested implementation is not compatible with it.\n \"\"\"\n node = task.node\n driver_name = driver_name or node.driver\n\n driver_or_hw_type = get_driver_or_hardware_type(driver_name)\n try:\n check_and_update_node_interfaces(\n node, driver_or_hw_type=driver_or_hw_type)\n except exception.MustBeNone as e:\n # NOTE(rloo). This was raised because nodes with classic drivers\n # cannot have any interfaces (except for network and\n # storage) set. However, there was a small window\n # where this was possible so instead of breaking those\n # users totally, we'll spam them with warnings instead.\n LOG.warning(_LW('%s They will be ignored. To avoid this warning, '\n 'please set them to None.'), e)\n\n bare_driver = driver_base.BareDriver()\n _attach_interfaces_to_driver(bare_driver, node, driver_or_hw_type)\n\n return bare_driver\n\n\ndef _attach_interfaces_to_driver(bare_driver, node, driver_or_hw_type):\n \"\"\"Attach interface implementations to a bare driver object.\n\n For classic drivers, copies implementations from the singleton driver\n object, then attaches the dynamic interfaces (network and storage\n interfaces for classic drivers, all interfaces for dynamic drivers\n made of hardware types).\n\n For hardware types, load all interface implementations dynamically.\n\n :param bare_driver: BareDriver instance to attach interfaces to\n :param node: Node object\n :param driver_or_hw_type: classic driver or hardware type instance\n :raises: InterfaceNotFoundInEntrypoint if the entry point was not found.\n :raises: IncompatibleInterface if driver is a hardware type and\n the requested implementation is not compatible with it.\n \"\"\"\n if isinstance(driver_or_hw_type, hardware_type.AbstractHardwareType):\n # For hardware types all interfaces are dynamic\n dynamic_interfaces = _INTERFACE_LOADERS\n else:\n # Copy implementations from the classic driver singleton\n for iface in driver_or_hw_type.all_interfaces:\n impl = getattr(driver_or_hw_type, iface, None)\n setattr(bare_driver, iface, impl)\n\n # NOTE(TheJulia): This list of interfaces to be applied\n # to classic drivers, thus requiring separate treatment.\n dynamic_interfaces = ['network', 'storage']\n\n for iface in dynamic_interfaces:\n impl_name = getattr(node, '%s_interface' % iface)\n impl = get_interface(driver_or_hw_type, iface, impl_name)\n setattr(bare_driver, iface, impl)\n\n\ndef get_interface(driver_or_hw_type, interface_type, interface_name):\n \"\"\"Get interface implementation instance.\n\n For hardware types also validates compatibility.\n\n :param driver_or_hw_type: a hardware type or classic driver instance.\n :param interface_type: name of the interface type (e.g. 'boot').\n :param interface_name: name of the interface implementation from an\n appropriate entry point\n (ironic.hardware.interfaces.).\n :returns: instance of the requested interface implementation.\n :raises: InterfaceNotFoundInEntrypoint if the entry point was not found.\n :raises: IncompatibleInterface if driver_or_hw_type is a hardware type and\n the requested implementation is not compatible with it.\n \"\"\"\n factory = _INTERFACE_LOADERS[interface_type]()\n try:\n impl_instance = factory.get_driver(interface_name)\n except KeyError:\n raise exception.InterfaceNotFoundInEntrypoint(\n iface=interface_name,\n entrypoint=factory._entrypoint_name,\n valid=factory.names)\n\n if not isinstance(driver_or_hw_type, hardware_type.AbstractHardwareType):\n # NOTE(dtantsur): classic drivers do not have notion of compatibility\n return impl_instance\n\n if isinstance(driver_or_hw_type, fake_hardware.FakeHardware):\n # NOTE(dtantsur): special-case fake hardware type to allow testing with\n # any combinations of interface implementations.\n return impl_instance\n\n supported_impls = getattr(driver_or_hw_type,\n 'supported_%s_interfaces' % interface_type)\n if type(impl_instance) not in supported_impls:\n raise exception.IncompatibleInterface(\n interface_type=interface_type, interface_impl=impl_instance,\n hardware_type=driver_or_hw_type.__class__.__name__)\n\n return impl_instance\n\n\ndef default_interface(driver_or_hw_type, interface_type,\n driver_name=None, node=None):\n \"\"\"Calculate and return the default interface implementation.\n\n Finds the first implementation that is supported by the hardware type\n and is enabled in the configuration.\n\n :param driver_or_hw_type: classic driver or hardware type instance object.\n :param interface_type: type of the interface (e.g. 'boot').\n :param driver_name: entrypoint name of the driver_or_hw_type object. Is\n used for exception message.\n :param node: the identifier of a node. If specified, is used for exception\n message.\n :returns: an entrypoint name of the calculated default implementation.\n :raises: InterfaceNotFoundInEntrypoint if the entry point was not found.\n :raises: NoValidDefaultForInterface if no default interface can be found.\n \"\"\"\n factory = _INTERFACE_LOADERS[interface_type]\n is_hardware_type = isinstance(driver_or_hw_type,\n hardware_type.AbstractHardwareType)\n # Explicit interface defaults\n additional_defaults = {\n 'network': 'flat' if CONF.dhcp.dhcp_provider == 'neutron' else 'noop',\n 'storage': 'noop'\n }\n\n # The fallback default from the configuration\n impl_name = getattr(CONF, 'default_%s_interface' % interface_type)\n if impl_name is None:\n impl_name = additional_defaults.get(interface_type)\n\n if impl_name is not None:\n # Check that the default is correct for this type\n get_interface(driver_or_hw_type, interface_type, impl_name)\n elif is_hardware_type:\n supported = getattr(driver_or_hw_type,\n 'supported_%s_interfaces' % interface_type)\n # Mapping of classes to entry points\n enabled = {obj.__class__: name for (name, obj) in factory().items()}\n\n # Order of the supported list matters\n for impl_class in supported:\n try:\n impl_name = enabled[impl_class]\n break\n except KeyError:\n continue\n\n if impl_name is None:\n # NOTE(rloo). No i18n on driver_type_str because translating substrings\n # on their own may cause the final string to look odd.\n if is_hardware_type:\n driver_type_str = 'hardware type'\n else:\n driver_type_str = 'driver'\n driver_name = driver_name or driver_or_hw_type.__class__.__name__\n node_info = \"\"\n if node is not None:\n node_info = _(' node %s with') % node\n raise exception.NoValidDefaultForInterface(\n interface_type=interface_type, driver_type=driver_type_str,\n driver=driver_name, node_info=node_info)\n\n return impl_name\n\n\ndef check_and_update_node_interfaces(node, driver_or_hw_type=None):\n \"\"\"Ensure that node interfaces (e.g. for creation or updating) are valid.\n\n Updates (but doesn't save to the database) hardware interfaces with\n calculated defaults, if they are not provided.\n\n This function is run on node updating and creation, as well as each time\n a driver instance is built for a node.\n\n :param node: node object to check and potentially update\n :param driver_or_hw_type: classic driver or hardware type instance object;\n will be detected from node.driver if missing\n :returns: True if any changes were made to the node, otherwise False\n :raises: InterfaceNotFoundInEntrypoint on validation failure\n :raises: NoValidDefaultForInterface if the default value cannot be\n calculated and is not provided in the configuration\n :raises: DriverNotFound if the node's driver or hardware type is not found\n :raises: MustBeNone if one or more of the node's interface\n fields were specified when they should not be.\n \"\"\"\n if driver_or_hw_type is None:\n driver_or_hw_type = get_driver_or_hardware_type(node.driver)\n is_hardware_type = isinstance(driver_or_hw_type,\n hardware_type.AbstractHardwareType)\n\n if is_hardware_type:\n factories = _INTERFACE_LOADERS.keys()\n else:\n # Only network and storage interfaces are dynamic for classic drivers\n factories = ['network', 'storage']\n\n # These are interfaces that cannot be specified via the node. E.g.,\n # for classic drivers, none are allowed except for network & storage.\n not_allowed_ifaces = driver_base.ALL_INTERFACES - set(factories)\n\n updates = node.obj_what_changed()\n # Result - whether the node object was modified\n result = False\n\n bad_interface_fields = []\n for iface in not_allowed_ifaces:\n field_name = '%s_interface' % iface\n # NOTE(vsaienko): reset *_interface fields that shouldn't exist for\n # classic driver, only when driver was changed and field not set\n # explicitly\n if 'driver' in updates and field_name not in updates:\n setattr(node, field_name, None)\n result = True\n # NOTE(dtantsur): objects raise NotImplementedError on accessing fields\n # that are known, but missing from an object. Thus, we cannot just use\n # getattr(node, field_name, None) here.\n elif field_name in node:\n impl_name = getattr(node, field_name)\n if impl_name is not None:\n bad_interface_fields.append(field_name)\n\n if bad_interface_fields:\n raise exception.MustBeNone(node=node.uuid, driver=node.driver,\n node_fields=','.join(bad_interface_fields))\n\n # Walk through all dynamic interfaces and check/update them\n for iface in factories:\n field_name = '%s_interface' % iface\n # NOTE(dtantsur): objects raise NotImplementedError on accessing fields\n # that are known, but missing from an object. Thus, we cannot just use\n # getattr(node, field_name, None) here.\n if field_name in node:\n impl_name = getattr(node, field_name)\n if impl_name is not None:\n # Check that the provided value is correct for this type\n get_interface(driver_or_hw_type, iface, impl_name)\n # Not changing the result, proceeding with the next interface\n continue\n\n impl_name = default_interface(driver_or_hw_type, iface,\n driver_name=node.driver, node=node.uuid)\n\n # Set the calculated default and set result to True\n setattr(node, field_name, impl_name)\n result = True\n\n return result\n\n\ndef get_driver_or_hardware_type(name):\n \"\"\"Get driver or hardware type by its entry point name.\n\n First, checks the hardware types namespace, then checks the classic\n drivers namespace. The first object found is returned.\n\n :param name: entry point name.\n :returns: An instance of a hardware type or a classic driver.\n :raises: DriverNotFound if neither hardware type nor classic driver found.\n \"\"\"\n try:\n return get_hardware_type(name)\n except exception.DriverNotFound:\n return get_driver(name)\n\n\ndef get_hardware_type(hardware_type):\n \"\"\"Get a hardware type instance by name.\n\n :param hardware_type: the name of the hardware type to find\n :returns: An instance of ironic.drivers.hardware_type.AbstractHardwareType\n :raises: DriverNotFound if requested hardware type cannot be found\n \"\"\"\n try:\n return HardwareTypesFactory().get_driver(hardware_type)\n except KeyError:\n raise exception.DriverNotFound(driver_name=hardware_type)\n\n\n# TODO(dtantsur): rename to get_classic_driver\ndef get_driver(driver_name):\n \"\"\"Simple method to get a ref to an instance of a driver.\n\n Driver loading is handled by the DriverFactory class. This method\n conveniently wraps that class and returns the actual driver object.\n\n :param driver_name: the name of the driver class to load\n :returns: An instance of a class which implements\n ironic.drivers.base.BaseDriver\n :raises: DriverNotFound if the requested driver_name could not be\n found in the \"ironic.drivers\" namespace.\n\n \"\"\"\n\n try:\n factory = DriverFactory()\n return factory.get_driver(driver_name)\n except KeyError:\n raise exception.DriverNotFound(driver_name=driver_name)\n\n\ndef _get_all_drivers(factory):\n \"\"\"Get all drivers for `factory` as a dict name -> driver object.\"\"\"\n # NOTE(jroll) I don't think this needs to be ordered, but\n # ConductorManager.init_host seems to depend on this behavior (or at\n # least the unit tests for it do), and it can't hurt much to keep it\n # that way.\n return collections.OrderedDict((name, factory[name].obj)\n for name in factory.names)\n\n\ndef drivers():\n \"\"\"Get all drivers.\n\n :returns: Dictionary mapping driver name to driver object.\n \"\"\"\n return _get_all_drivers(DriverFactory())\n\n\ndef hardware_types():\n \"\"\"Get all hardware types.\n\n :returns: Dictionary mapping hardware type name to hardware type object.\n \"\"\"\n return _get_all_drivers(HardwareTypesFactory())\n\n\ndef interfaces(interface_type):\n \"\"\"Get all interfaces for a given interface type.\n\n :param interface_type: String, type of interface to fetch for.\n :returns: Dictionary mapping interface name to interface object.\n \"\"\"\n return _get_all_drivers(_INTERFACE_LOADERS[interface_type]())\n\n\ndef enabled_supported_interfaces(hardware_type):\n \"\"\"Get usable interfaces for a given hardware type.\n\n For a given hardware type, find the intersection of enabled and supported\n interfaces for each interface type. This is the set of interfaces that are\n usable for this hardware type.\n\n :param hardware_type: The hardware type object to search.\n :returns: a dict mapping interface types to a list of enabled and supported\n interface names.\n \"\"\"\n mapping = dict()\n for interface_type in driver_base.ALL_INTERFACES:\n supported = set()\n supported_ifaces = getattr(hardware_type,\n 'supported_%s_interfaces' % interface_type)\n for name, iface in interfaces(interface_type).items():\n if iface.__class__ in supported_ifaces:\n supported.add(name)\n mapping[interface_type] = supported\n return mapping\n\n\nclass BaseDriverFactory(object):\n \"\"\"Discover, load and manage the drivers available.\n\n This is subclassed to load both main drivers and extra interfaces.\n \"\"\"\n\n # NOTE(deva): loading the _extension_manager as a class member will break\n # stevedore when it loads a driver, because the driver will\n # import this file (and thus instantiate another factory).\n # Instead, we instantiate a NameDispatchExtensionManager only\n # once, the first time DriverFactory.__init__ is called.\n _extension_manager = None\n\n # Entrypoint name containing the list of all available drivers/interfaces\n _entrypoint_name = None\n # Name of the [DEFAULT] section config option containing a list of enabled\n # drivers/interfaces\n _enabled_driver_list_config_option = ''\n # This field will contain the list of the enabled drivers/interfaces names\n # without duplicates\n _enabled_driver_list = None\n # Template for logging loaded drivers\n _logging_template = _LI(\"Loaded the following drivers: %s\")\n\n def __init__(self):\n if not self.__class__._extension_manager:\n self.__class__._init_extension_manager()\n\n def __getitem__(self, name):\n return self._extension_manager[name]\n\n def get_driver(self, name):\n return self[name].obj\n\n # NOTE(deva): Use lockutils to avoid a potential race in eventlet\n # that might try to create two driver factories.\n @classmethod\n @lockutils.synchronized(EM_SEMAPHORE)\n def _init_extension_manager(cls):\n # NOTE(deva): In case multiple greenthreads queue up on this lock\n # before _extension_manager is initialized, prevent\n # creation of multiple NameDispatchExtensionManagers.\n if cls._extension_manager:\n return\n enabled_drivers = getattr(CONF, cls._enabled_driver_list_config_option,\n [])\n\n # Check for duplicated driver entries and warn the operator\n # about them\n counter = collections.Counter(enabled_drivers).items()\n duplicated_drivers = []\n cls._enabled_driver_list = []\n for item, cnt in counter:\n if not item:\n LOG.warning(\n _LW('An empty driver was specified in the \"%s\" '\n 'configuration option and will be ignored. Please '\n 'fix your ironic.conf file to avoid this warning '\n 'message.'), cls._enabled_driver_list_config_option)\n continue\n if cnt > 1:\n duplicated_drivers.append(item)\n cls._enabled_driver_list.append(item)\n if duplicated_drivers:\n LOG.warning(_LW('The driver(s) \"%s\" is/are duplicated in the '\n 'list of enabled_drivers. Please check your '\n 'configuration file.'),\n ', '.join(duplicated_drivers))\n\n # NOTE(deva): Drivers raise \"DriverLoadError\" if they are unable to be\n # loaded, eg. due to missing external dependencies.\n # We capture that exception, and, only if it is for an\n # enabled driver, raise it from here. If enabled driver\n # raises other exception type, it is wrapped in\n # \"DriverLoadError\", providing the name of the driver that\n # caused it, and raised. If the exception is for a\n # non-enabled driver, we suppress it.\n def _catch_driver_not_found(mgr, ep, exc):\n # NOTE(deva): stevedore loads plugins *before* evaluating\n # _check_func, so we need to check here, too.\n if ep.name in cls._enabled_driver_list:\n if not isinstance(exc, exception.DriverLoadError):\n raise exception.DriverLoadError(driver=ep.name, reason=exc)\n raise exc\n\n def _check_func(ext):\n return ext.name in cls._enabled_driver_list\n\n cls._extension_manager = (\n dispatch.NameDispatchExtensionManager(\n cls._entrypoint_name,\n _check_func,\n invoke_on_load=True,\n on_load_failure_callback=_catch_driver_not_found,\n propagate_map_exceptions=True))\n\n # NOTE(deva): if we were unable to load any configured driver, perhaps\n # because it is not present on the system, raise an error.\n if (sorted(cls._enabled_driver_list) !=\n sorted(cls._extension_manager.names())):\n found = cls._extension_manager.names()\n names = [n for n in cls._enabled_driver_list if n not in found]\n # just in case more than one could not be found ...\n names = ', '.join(names)\n raise exception.DriverNotFoundInEntrypoint(\n names=names, entrypoint=cls._entrypoint_name)\n\n # warn for any untested/unsupported/deprecated drivers or interfaces\n cls._extension_manager.map(cls._extension_manager.names(),\n _warn_if_unsupported)\n\n LOG.info(cls._logging_template, cls._extension_manager.names())\n\n @property\n def names(self):\n \"\"\"The list of driver names available.\"\"\"\n return self._extension_manager.names()\n\n def items(self):\n \"\"\"Iterator over pairs (name, instance).\"\"\"\n return ((ext.name, ext.obj) for ext in self._extension_manager)\n\n\ndef _warn_if_unsupported(ext):\n if not ext.obj.supported:\n LOG.warning(_LW('Driver \"%s\" is UNSUPPORTED. It has been deprecated '\n 'and may be removed in a future release.'), ext.name)\n\n\nclass DriverFactory(BaseDriverFactory):\n _entrypoint_name = 'ironic.drivers'\n _enabled_driver_list_config_option = 'enabled_drivers'\n\n\nclass HardwareTypesFactory(BaseDriverFactory):\n _entrypoint_name = 'ironic.hardware.types'\n _enabled_driver_list_config_option = 'enabled_hardware_types'\n _logging_template = _LI(\"Loaded the following hardware types: %s\")\n\n\n_INTERFACE_LOADERS = {\n name: type('%sInterfaceFactory' % name.capitalize(),\n (BaseDriverFactory,),\n {'_entrypoint_name': 'ironic.hardware.interfaces.%s' % name,\n '_enabled_driver_list_config_option':\n 'enabled_%s_interfaces' % name,\n '_logging_template':\n _LI(\"Loaded the following %s interfaces: %%s\") % name})\n for name in driver_base.ALL_INTERFACES\n}\n\n\n# TODO(dtantsur): This factory is still used explicitly in many places,\n# refactor them later to use _INTERFACE_LOADERS.\nNetworkInterfaceFactory = _INTERFACE_LOADERS['network']\nStorageInterfaceFactory = _INTERFACE_LOADERS['storage']\n", "sub_path": "ironic/common/driver_factory.py", "file_name": "driver_factory.py", "file_ext": "py", "file_size_in_byte": 23853, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "oslo_log.log.getLogger", "line_number": 30, "usage_type": "call"}, {"api_name": "oslo_log.log", "line_number": 30, "usage_type": "name"}, {"api_name": "ironic.common.exception.MustBeNone", "line_number": 61, "usage_type": "attribute"}, {"api_name": "ironic.common.exception", "line_number": 61, "usage_type": "name"}, {"api_name": "ironic.common.i18n._LW", "line_number": 67, "usage_type": "call"}, {"api_name": "ironic.drivers.base.BareDriver", "line_number": 70, "usage_type": "call"}, {"api_name": "ironic.drivers.base", "line_number": 70, "usage_type": "name"}, {"api_name": "ironic.drivers.hardware_type.AbstractHardwareType", "line_number": 93, "usage_type": "attribute"}, {"api_name": "ironic.drivers.hardware_type", "line_number": 93, "usage_type": "name"}, {"api_name": "ironic.common.exception.InterfaceNotFoundInEntrypoint", "line_number": 131, "usage_type": "call"}, {"api_name": "ironic.common.exception", "line_number": 131, "usage_type": "name"}, {"api_name": "ironic.drivers.hardware_type.AbstractHardwareType", "line_number": 136, "usage_type": "attribute"}, {"api_name": "ironic.drivers.hardware_type", "line_number": 136, "usage_type": "name"}, {"api_name": "ironic.drivers.fake_hardware.FakeHardware", "line_number": 140, "usage_type": "attribute"}, {"api_name": "ironic.drivers.fake_hardware", "line_number": 140, "usage_type": "name"}, {"api_name": "ironic.common.exception.IncompatibleInterface", "line_number": 148, "usage_type": "call"}, {"api_name": "ironic.common.exception", "line_number": 148, "usage_type": "name"}, {"api_name": "ironic.drivers.hardware_type.AbstractHardwareType", "line_number": 174, "usage_type": "attribute"}, {"api_name": "ironic.drivers.hardware_type", "line_number": 174, "usage_type": "name"}, {"api_name": "ironic.conf.CONF.dhcp", "line_number": 177, "usage_type": "attribute"}, {"api_name": "ironic.conf.CONF", "line_number": 177, "usage_type": "name"}, {"api_name": "ironic.conf.CONF", "line_number": 182, "usage_type": "argument"}, {"api_name": "ironic.common.i18n._", "line_number": 213, "usage_type": "call"}, {"api_name": "ironic.common.exception.NoValidDefaultForInterface", "line_number": 214, "usage_type": "call"}, {"api_name": "ironic.common.exception", "line_number": 214, "usage_type": "name"}, {"api_name": "ironic.drivers.hardware_type.AbstractHardwareType", "line_number": 244, "usage_type": "attribute"}, {"api_name": "ironic.drivers.hardware_type", "line_number": 244, "usage_type": "name"}, {"api_name": "ironic.drivers.base.ALL_INTERFACES", "line_number": 254, "usage_type": "attribute"}, {"api_name": "ironic.drivers.base", "line_number": 254, "usage_type": "name"}, {"api_name": "ironic.common.exception.MustBeNone", "line_number": 278, "usage_type": "call"}, {"api_name": "ironic.common.exception", "line_number": 278, "usage_type": "name"}, {"api_name": "ironic.common.exception.DriverNotFound", "line_number": 317, "usage_type": "attribute"}, {"api_name": "ironic.common.exception", "line_number": 317, "usage_type": "name"}, {"api_name": "ironic.drivers.hardware_type", "line_number": 329, "usage_type": "argument"}, {"api_name": "ironic.common.exception.DriverNotFound", "line_number": 331, "usage_type": "call"}, {"api_name": "ironic.common.exception", "line_number": 331, "usage_type": "name"}, {"api_name": "ironic.drivers.hardware_type", "line_number": 331, "usage_type": "name"}, {"api_name": "ironic.common.exception.DriverNotFound", "line_number": 353, "usage_type": "call"}, {"api_name": "ironic.common.exception", "line_number": 353, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 362, "usage_type": "call"}, {"api_name": "ironic.drivers.base.ALL_INTERFACES", "line_number": 403, "usage_type": "attribute"}, {"api_name": "ironic.drivers.base", "line_number": 403, "usage_type": "name"}, {"api_name": "ironic.drivers.hardware_type", "line_number": 405, "usage_type": "argument"}, {"api_name": "ironic.common.i18n._LI", "line_number": 436, "usage_type": "call"}, {"api_name": "ironic.conf.CONF", "line_number": 458, "usage_type": "argument"}, {"api_name": "collections.Counter", "line_number": 463, "usage_type": "call"}, {"api_name": "ironic.common.i18n._LW", "line_number": 469, "usage_type": "call"}, {"api_name": "ironic.common.i18n._LW", "line_number": 478, "usage_type": "call"}, {"api_name": "ironic.common.exception.DriverLoadError", "line_number": 495, "usage_type": "attribute"}, {"api_name": "ironic.common.exception", "line_number": 495, "usage_type": "name"}, {"api_name": "ironic.common.exception.DriverLoadError", "line_number": 496, "usage_type": "call"}, {"api_name": "ironic.common.exception", "line_number": 496, "usage_type": "name"}, {"api_name": "stevedore.dispatch.NameDispatchExtensionManager", "line_number": 503, "usage_type": "call"}, {"api_name": "stevedore.dispatch", "line_number": 503, "usage_type": "name"}, {"api_name": "ironic.common.exception.DriverNotFoundInEntrypoint", "line_number": 518, "usage_type": "call"}, {"api_name": "ironic.common.exception", "line_number": 518, "usage_type": "name"}, {"api_name": "oslo_concurrency.lockutils.synchronized", "line_number": 451, "usage_type": "call"}, {"api_name": "oslo_concurrency.lockutils", "line_number": 451, "usage_type": "name"}, {"api_name": "ironic.common.i18n._LW", "line_number": 539, "usage_type": "call"}, {"api_name": "ironic.common.i18n._LI", "line_number": 551, "usage_type": "call"}, {"api_name": "ironic.common.i18n._LI", "line_number": 561, "usage_type": "call"}, {"api_name": "ironic.drivers.base.ALL_INTERFACES", "line_number": 562, "usage_type": "attribute"}, {"api_name": "ironic.drivers.base", "line_number": 562, "usage_type": "name"}]} +{"seq_id": "13589017", "text": "import _pickle\n\nimport numpy as np\nimport torch\n\nfrom fastNLP.action.action import Action\nfrom fastNLP.action.action import RandomSampler, Batchifier\nfrom fastNLP.action.tester import POSTester\nfrom fastNLP.modules.utils import seq_mask\n\n\nclass BaseTrainer(Action):\n \"\"\"Base trainer for all trainers.\n Trainer receives a model and data, and then performs training.\n\n Subclasses must implement the following abstract methods:\n - prepare_input\n - mode\n - define_optimizer\n - data_forward\n - grad_backward\n - get_loss\n \"\"\"\n\n def __init__(self, train_args):\n \"\"\"\n :param train_args: dict of (key, value)\n\n The base trainer requires the following keys:\n - epochs: int, the number of epochs in training\n - validate: bool, whether or not to validate on dev set\n - batch_size: int\n - pickle_path: str, the path to pickle files for pre-processing\n \"\"\"\n super(BaseTrainer, self).__init__()\n self.n_epochs = train_args[\"epochs\"]\n self.validate = train_args[\"validate\"]\n self.batch_size = train_args[\"batch_size\"]\n self.pickle_path = train_args[\"pickle_path\"]\n self.model = None\n self.iterator = None\n self.loss_func = None\n self.optimizer = None\n\n def train(self, network):\n \"\"\"General Training Steps\n :param network: a model\n\n The method is framework independent.\n Work by calling the following methods:\n - prepare_input\n - mode\n - define_optimizer\n - data_forward\n - get_loss\n - grad_backward\n - update\n Subclasses must implement these methods with a specific framework.\n \"\"\"\n # prepare model and data\n self.model = network\n data_train, data_dev, data_test, embedding = self.prepare_input(self.pickle_path)\n\n # define tester over dev data\n valid_args = {\"save_output\": True, \"validate_in_training\": True, \"save_dev_input\": True,\n \"save_loss\": True, \"batch_size\": self.batch_size, \"pickle_path\": self.pickle_path}\n validator = POSTester(valid_args)\n\n # main training epochs\n iterations = len(data_train) // self.batch_size\n for epoch in range(self.n_epochs):\n\n # turn on network training mode; define optimizer; prepare batch iterator\n self.mode(test=False)\n self.define_optimizer()\n self.iterator = iter(Batchifier(RandomSampler(data_train), self.batch_size, drop_last=True))\n\n # training iterations in one epoch\n for step in range(iterations):\n batch_x, batch_y = self.batchify(data_train)\n\n prediction = self.data_forward(network, batch_x)\n\n loss = self.get_loss(prediction, batch_y)\n self.grad_backward(loss)\n self.update()\n\n if self.validate:\n if data_dev is None:\n raise RuntimeError(\"No validation data provided.\")\n validator.test(network)\n\n # finish training\n\n def prepare_input(self, data_path):\n \"\"\"\n To do: Load pkl files of train/dev/test and embedding\n \"\"\"\n data_train = _pickle.load(open(data_path + \"data_train.pkl\", \"rb\"))\n data_dev = _pickle.load(open(data_path + \"data_dev.pkl\", \"rb\"))\n data_test = _pickle.load(open(data_path + \"data_test.pkl\", \"rb\"))\n embedding = _pickle.load(open(data_path + \"embedding.pkl\", \"rb\"))\n return data_train, data_dev, data_test, embedding\n\n def mode(self, test=False):\n \"\"\"\n Tell the network to be trained or not.\n :param test: bool\n \"\"\"\n raise NotImplementedError\n\n def define_optimizer(self):\n \"\"\"\n Define framework-specific optimizer specified by the models.\n \"\"\"\n raise NotImplementedError\n\n def update(self):\n \"\"\"\n Perform weight update on a model.\n\n For PyTorch, just call optimizer to update.\n \"\"\"\n raise NotImplementedError\n\n def data_forward(self, network, x):\n \"\"\"\n Forward pass of the data.\n :param network: a model\n :param x: input feature matrix and label vector\n :return: output by the models\n\n For PyTorch, just do \"network(*x)\"\n \"\"\"\n raise NotImplementedError\n\n def grad_backward(self, loss):\n \"\"\"\n Compute gradient with link rules.\n :param loss: a scalar where back-prop starts\n\n For PyTorch, just do \"loss.backward()\"\n \"\"\"\n raise NotImplementedError\n\n def get_loss(self, predict, truth):\n \"\"\"\n Compute loss given prediction and ground truth.\n :param predict: prediction label vector\n :param truth: ground truth label vector\n :return: a scalar\n \"\"\"\n if self.loss_func is None:\n if hasattr(self.model, \"loss\"):\n self.loss_func = self.model.loss\n else:\n self.define_loss()\n return self.loss_func(predict, truth)\n\n def define_loss(self):\n \"\"\"\n Assign an instance of loss function to self.loss_func\n E.g. self.loss_func = nn.CrossEntropyLoss()\n \"\"\"\n raise NotImplementedError\n\n def batchify(self, data):\n \"\"\"\n 1. Perform batching from data and produce a batch of training data.\n 2. Add padding.\n :param data: list. Each entry is a sample, which is also a list of features and label(s).\n E.g.\n [\n [[word_11, word_12, word_13], [label_11. label_12]], # sample 1\n [[word_21, word_22, word_23], [label_21. label_22]], # sample 2\n ...\n ]\n :return batch_x: list. Each entry is a list of features of a sample. [batch_size, max_len]\n batch_y: list. Each entry is a list of labels of a sample. [batch_size, num_labels]\n \"\"\"\n indices = next(self.iterator)\n batch = [data[idx] for idx in indices]\n batch_x = [sample[0] for sample in batch]\n batch_y = [sample[1] for sample in batch]\n batch_x = self.pad(batch_x)\n return batch_x, batch_y\n\n @staticmethod\n def pad(batch, fill=0):\n \"\"\"\n Pad a batch of samples to maximum length.\n :param batch: list of list\n :param fill: word index to pad, default 0.\n :return: a padded batch\n \"\"\"\n max_length = max([len(x) for x in batch])\n for idx, sample in enumerate(batch):\n if len(sample) < max_length:\n batch[idx] = sample + [fill * (max_length - len(sample))]\n return batch\n\n\nclass ToyTrainer(BaseTrainer):\n \"\"\"\n deprecated\n \"\"\"\n\n def __init__(self, train_args):\n super(ToyTrainer, self).__init__(train_args)\n self.test_mode = False\n self.weight = np.random.rand(5, 1)\n self.bias = np.random.rand()\n self._loss = 0\n self._optimizer = None\n\n def prepare_input(self, data):\n return data[:, :-1], data[:, -1]\n\n def mode(self, test=False):\n self.model.mode(test)\n\n def data_forward(self, network, x):\n return np.matmul(x, self.weight) + self.bias\n\n def grad_backward(self, loss):\n loss.backward()\n\n def get_loss(self, pred, truth):\n self._loss = np.mean(np.square(pred - truth))\n return self._loss\n\n def define_optimizer(self):\n self._optimizer = torch.optim.SGD(self.model.parameters(), lr=0.01)\n\n def update(self):\n self._optimizer.step()\n\n\nclass WordSegTrainer(BaseTrainer):\n \"\"\"\n deprecated\n \"\"\"\n\n def __init__(self, train_args):\n super(WordSegTrainer, self).__init__(train_args)\n self.id2word = None\n self.word2id = None\n self.id2tag = None\n self.tag2id = None\n\n self.lstm_batch_size = 8\n self.lstm_seq_len = 32 # Trainer batch_size == lstm_batch_size * lstm_seq_len\n self.hidden_dim = 100\n self.lstm_num_layers = 2\n self.vocab_size = 100\n self.word_emb_dim = 100\n\n self.hidden = (self.to_var(torch.zeros(2, self.lstm_batch_size, self.word_emb_dim)),\n self.to_var(torch.zeros(2, self.lstm_batch_size, self.word_emb_dim)))\n\n self.optimizer = None\n self._loss = None\n\n self.USE_GPU = False\n\n def to_var(self, x):\n if torch.cuda.is_available() and self.USE_GPU:\n x = x.cuda()\n return torch.autograd.Variable(x)\n\n def prepare_input(self, data):\n \"\"\"\n perform word indices lookup to convert strings into indices\n :param data: list of string, each string contains word + space + [B, M, E, S]\n :return\n \"\"\"\n word_list = []\n tag_list = []\n for line in data:\n if len(line) > 2:\n tokens = line.split(\"#\")\n word_list.append(tokens[0])\n tag_list.append(tokens[2][0])\n self.id2word = list(set(word_list))\n self.word2id = {word: idx for idx, word in enumerate(self.id2word)}\n self.id2tag = list(set(tag_list))\n self.tag2id = {tag: idx for idx, tag in enumerate(self.id2tag)}\n words = np.array([self.word2id[w] for w in word_list]).reshape(-1, 1)\n tags = np.array([self.tag2id[t] for t in tag_list]).reshape(-1, 1)\n return words, tags\n\n def mode(self, test=False):\n if test:\n self.model.eval()\n else:\n self.model.train()\n\n def data_forward(self, network, x):\n \"\"\"\n :param network: a PyTorch model\n :param x: sequence of length [batch_size], word indices\n :return:\n \"\"\"\n x = x.reshape(self.lstm_batch_size, self.lstm_seq_len)\n output, self.hidden = network(x, self.hidden)\n return output\n\n def define_optimizer(self):\n self.optimizer = torch.optim.SGD(self.model.parameters(), lr=0.01, momentum=0.85)\n\n def get_loss(self, predict, truth):\n truth = torch.Tensor(truth)\n self._loss = torch.nn.CrossEntropyLoss(predict, truth)\n return self._loss\n\n def grad_backward(self, network):\n self.model.zero_grad()\n self._loss.backward()\n torch.nn.utils.clip_grad_norm(self.model.parameters(), 5, norm_type=2)\n\n def update(self):\n self.optimizer.step()\n\n\nclass POSTrainer(BaseTrainer):\n \"\"\"\n Trainer for Sequence Modeling\n\n \"\"\"\n def __init__(self, train_args):\n super(POSTrainer, self).__init__(train_args)\n self.vocab_size = train_args[\"vocab_size\"]\n self.num_classes = train_args[\"num_classes\"]\n self.max_len = None\n self.mask = None\n\n def prepare_input(self, data_path):\n \"\"\"\n To do: Load pkl files of train/dev/test and embedding\n \"\"\"\n data_train = _pickle.load(open(data_path + \"/data_train.pkl\", \"rb\"))\n data_dev = _pickle.load(open(data_path + \"/data_train.pkl\", \"rb\"))\n return data_train, data_dev, 0, 1\n\n def data_forward(self, network, x):\n \"\"\"\n :param network: the PyTorch model\n :param x: list of list, [batch_size, max_len]\n :return y: [batch_size, num_classes]\n \"\"\"\n seq_len = [len(seq) for seq in x]\n x = torch.Tensor(x).long()\n self.batch_size = x.size(0)\n self.max_len = x.size(1)\n self.mask = seq_mask(seq_len, self.max_len)\n y = network(x)\n return y\n\n def mode(self, test=False):\n if test:\n self.model.eval()\n else:\n self.model.train()\n\n def define_optimizer(self):\n self.optimizer = torch.optim.SGD(self.model.parameters(), lr=0.01, momentum=0.9)\n\n def grad_backward(self, loss):\n self.model.zero_grad()\n loss.backward()\n\n def update(self):\n self.optimizer.step()\n\n def get_loss(self, predict, truth):\n \"\"\"\n Compute loss given prediction and ground truth.\n :param predict: prediction label vector, [batch_size, num_classes]\n :param truth: ground truth label vector, [batch_size, max_len]\n :return: a scalar\n \"\"\"\n truth = torch.Tensor(truth)\n if self.loss_func is None:\n if hasattr(self.model, \"loss\"):\n self.loss_func = self.model.loss\n else:\n self.define_loss()\n loss, prediction = self.loss_func(predict, truth, self.mask, self.batch_size, self.max_len)\n return loss\n\n\nif __name__ == \"__name__\":\n train_args = {\"epochs\": 1, \"validate\": False, \"batch_size\": 3, \"pickle_path\": \"./\"}\n trainer = BaseTrainer(train_args)\n data_train = [[[1, 2, 3, 4], [0]] * 10] + [[[1, 3, 5, 2], [1]] * 10]\n trainer.batchify(data=data_train)\n", "sub_path": "fastNLP/action/trainer.py", "file_name": "trainer.py", "file_ext": "py", "file_size_in_byte": 12869, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "fastNLP.action.action.Action", "line_number": 12, "usage_type": "name"}, {"api_name": "fastNLP.action.tester.POSTester", "line_number": 67, "usage_type": "call"}, {"api_name": "fastNLP.action.action.Batchifier", "line_number": 76, "usage_type": "call"}, {"api_name": "fastNLP.action.action.RandomSampler", "line_number": 76, "usage_type": "call"}, {"api_name": "_pickle.load", "line_number": 99, "usage_type": "call"}, {"api_name": "_pickle.load", "line_number": 100, "usage_type": "call"}, {"api_name": "_pickle.load", "line_number": 101, "usage_type": "call"}, {"api_name": "_pickle.load", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 211, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 211, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 212, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 212, "usage_type": "attribute"}, {"api_name": "numpy.matmul", "line_number": 223, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 229, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 229, "usage_type": "call"}, {"api_name": "torch.optim.SGD", "line_number": 233, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 233, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 258, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 259, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 267, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 267, "usage_type": "attribute"}, {"api_name": "torch.autograd.Variable", "line_number": 269, "usage_type": "call"}, {"api_name": "torch.autograd", "line_number": 269, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 288, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 289, "usage_type": "call"}, {"api_name": "torch.optim.SGD", "line_number": 309, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 309, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 312, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 313, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 313, "usage_type": "attribute"}, {"api_name": "torch.nn.utils.clip_grad_norm", "line_number": 319, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 319, "usage_type": "attribute"}, {"api_name": "_pickle.load", "line_number": 341, "usage_type": "call"}, {"api_name": "_pickle.load", "line_number": 342, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 352, "usage_type": "call"}, {"api_name": "fastNLP.modules.utils.seq_mask", "line_number": 355, "usage_type": "call"}, {"api_name": "torch.optim.SGD", "line_number": 366, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 366, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 382, "usage_type": "call"}]} +{"seq_id": "579439161", "text": "from django.db.models import Q\nfrom rest_framework.pagination import PageNumberPagination\nfrom rest_framework.response import Response\nfrom rest_framework.views import APIView\n\nimport npo_law\nfrom npo_law.models import NPOLaw, NPOLawFavorite\nfrom npo_law.serializers import NPOLawSerializer\n\n\nclass NPOLawAPIView(APIView, PageNumberPagination):\n allow_methods = ['GET', 'POST']\n serializer_class = NPOLawSerializer\n\n def get(self, request, *args, **kwargs):\n query = request.query_params.get('query', '')\n npolaw = NPOLaw.objects.filter(Q(title__icontains=query) |\n Q(description__icontains=query))\n\n results = self.paginate_queryset(npolaw,\n request,\n view=self)\n return self.get_paginated_response(self.serializer_class(results,\n many=True,\n context={'request': request}).data)\n\n def post(self, request, *args, **kwargs):\n title = request.data.get('title')\n description = request.data.get('description')\n created_date = request.data.get('created_date')\n file = request.data.get('file')\n npolaw = NPOLaw.objects.create(title=title,\n description=description,\n created_date=created_date,\n file=file)\n npolaw.save()\n return Response(data=self.serializer_class(npo_law).data,\n status=status.HTTP_201_CREATED)\n\n\nclass NPOLawDetailAPIView(APIView):\n allow_methods = ['GET', 'PUT', 'DELETE']\n serializer_class = NPOLawSerializer\n\n def get(self, request, id):\n law = NPOLaw.objects.get(id=id)\n return Response(data=self.serializer_class(law).data)\n\n def put(self, request, id):\n law = NPOLaw.objects.get(id=id)\n title = request.data.get('title')\n description = request.data.get('description')\n file = request.data.get('file')\n law.title = title\n law.description = description\n law.file = file\n law.save()\n return Response(data=self.serializer_class(law).data,\n status=status.HTTP_202_ACCEPTED)\n\n def delete(self, request, id):\n law = NPOLaw.objects.get(id=id)\n law.delete()\n return Response(status=status.HTTP_204_NO_CONTENT)\n\n\nclass NPOLawFavoriteAPIView(APIView):\n allow_methods = ['GET', 'PUT', 'DELETE']\n serializer_class = NPOLawSerializer\n\n def get(self, request, id):\n npolaw = NPOLaw.objects.get(id=id)\n return Response(data=self.serializer_class(npolaw).data)\n\n def put(self, request, id):\n npolaw = NPOLaw.objects.get(id=id)\n text = request.data.get('text')\n created_date = request.data.get('created_date')\n npolaw.text = text\n npolaw.created_date = created_date\n npolaw.save()\n return Response(data=self.serializer_class(npolaw).data,\n status=status.HTTP_202_ACCEPTED)\n\n def delete(self, request, id):\n npolaw = NPOLaw.objects.get(id=id)\n npolaw.delete()\n return Response(status=status.HTTP_204_NO_CONTENT)\n\n", "sub_path": "npo_law/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3342, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "rest_framework.views.APIView", "line_number": 11, "usage_type": "name"}, {"api_name": "rest_framework.pagination.PageNumberPagination", "line_number": 11, "usage_type": "name"}, {"api_name": "npo_law.serializers.NPOLawSerializer", "line_number": 13, "usage_type": "name"}, {"api_name": "npo_law.models.NPOLaw.objects.filter", "line_number": 17, "usage_type": "call"}, {"api_name": "npo_law.models.NPOLaw.objects", "line_number": 17, "usage_type": "attribute"}, {"api_name": "npo_law.models.NPOLaw", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 18, "usage_type": "call"}, {"api_name": "npo_law.models.NPOLaw.objects.create", "line_number": 32, "usage_type": "call"}, {"api_name": "npo_law.models.NPOLaw.objects", "line_number": 32, "usage_type": "attribute"}, {"api_name": "npo_law.models.NPOLaw", "line_number": 32, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 37, "usage_type": "call"}, {"api_name": "rest_framework.views.APIView", "line_number": 41, "usage_type": "name"}, {"api_name": "npo_law.serializers.NPOLawSerializer", "line_number": 43, "usage_type": "name"}, {"api_name": "npo_law.models.NPOLaw.objects.get", "line_number": 46, "usage_type": "call"}, {"api_name": "npo_law.models.NPOLaw.objects", "line_number": 46, "usage_type": "attribute"}, {"api_name": "npo_law.models.NPOLaw", "line_number": 46, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 47, "usage_type": "call"}, {"api_name": "npo_law.models.NPOLaw.objects.get", "line_number": 50, "usage_type": "call"}, {"api_name": "npo_law.models.NPOLaw.objects", "line_number": 50, "usage_type": "attribute"}, {"api_name": "npo_law.models.NPOLaw", "line_number": 50, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 58, "usage_type": "call"}, {"api_name": "npo_law.models.NPOLaw.objects.get", "line_number": 62, "usage_type": "call"}, {"api_name": "npo_law.models.NPOLaw.objects", "line_number": 62, "usage_type": "attribute"}, {"api_name": "npo_law.models.NPOLaw", "line_number": 62, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 64, "usage_type": "call"}, {"api_name": "rest_framework.views.APIView", "line_number": 67, "usage_type": "name"}, {"api_name": "npo_law.serializers.NPOLawSerializer", "line_number": 69, "usage_type": "name"}, {"api_name": "npo_law.models.NPOLaw.objects.get", "line_number": 72, "usage_type": "call"}, {"api_name": "npo_law.models.NPOLaw.objects", "line_number": 72, "usage_type": "attribute"}, {"api_name": "npo_law.models.NPOLaw", "line_number": 72, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 73, "usage_type": "call"}, {"api_name": "npo_law.models.NPOLaw.objects.get", "line_number": 76, "usage_type": "call"}, {"api_name": "npo_law.models.NPOLaw.objects", "line_number": 76, "usage_type": "attribute"}, {"api_name": "npo_law.models.NPOLaw", "line_number": 76, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 82, "usage_type": "call"}, {"api_name": "npo_law.models.NPOLaw.objects.get", "line_number": 86, "usage_type": "call"}, {"api_name": "npo_law.models.NPOLaw.objects", "line_number": 86, "usage_type": "attribute"}, {"api_name": "npo_law.models.NPOLaw", "line_number": 86, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 88, "usage_type": "call"}]} +{"seq_id": "465123219", "text": "import requests\nimport os\nimport time\nimport sys\nimport subprocess\nfrom PIL import Image\n\nwhile True: # making a loop\n t = requests.get('https://min-api.cryptocompare.com/data/pricemultifull?fsyms=BTC&tsyms=USD,EUR')\n os.system('clear')\n s = t.json()['DISPLAY']['BTC']['USD']['FROMSYMBOL']\n p = t.json()['DISPLAY']['BTC']['USD']['PRICE']\n m = t.json()['DISPLAY']['BTC']['USD']['MKTCAP']\n h = t.json()['DISPLAY']['BTC']['USD']['HIGH24HOUR']\n l = t.json()['DISPLAY']['BTC']['USD']['LOW24HOUR']\n c = t.json()['DISPLAY']['BTC']['USD']['CHANGE24HOUR']\n pct = t.json()['DISPLAY']['BTC']['USD']['CHANGEPCT24HOUR']\n o = t.json()['DISPLAY']['BTC']['USD']['OPEN24HOUR']\n \n print(\"\\n\"+\"\\n\"+\"\\n\"+\"\\n\"+\"\\n\"+\"\\n\"+\"\\n\")\n print(\"\\n\"+\"\\n\"+\"\\n\")\n \n print(\"Bitcoin (BTC) \"+s+\"\\n\")\n print(\"BTC/USD: \"+p+\"\\n\")\n print(\"Market Cap: \"+m+\"\\n\")\n print(\"24 Hour High: \"+h+\"\\n\")\n print(\"24 Hour Low: \"+l+\"\\n\")\n print(\"24 Hour Change: \"+c+\"\\n\")\n if(o>p):\n print(\"24 Hour Percent Change: \"+pct+\"%\"+\"\\n\")\n pink = subprocess.Popen([\"display\", \"pink.jpg\"])\n else:\n print(\"24 Hour Percent Change: +\"+pct+\"%\"+\"\\n\")\n green = subprocess.Popen([\"display\", \"green.jpg\"])\n \n for i in range(30):\n timeTill = (\"Update in: \"+str(30-i)+\" seconds\")\n print(timeTill)\n time.sleep(1)\n sys.stdout.write(\"\\033[F\") #back to previous line\n sys.stdout.write(\"\\033[K\") #erase line\n if(o>p):\n pink.kill()\n else:\n green.kill()", "sub_path": "btc_usd.py", "file_name": "btc_usd.py", "file_ext": "py", "file_size_in_byte": 1526, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "requests.get", "line_number": 9, "usage_type": "call"}, {"api_name": "os.system", "line_number": 10, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 31, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 34, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 39, "usage_type": "call"}, {"api_name": "sys.stdout.write", "line_number": 40, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 40, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 41, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 41, "usage_type": "attribute"}]} +{"seq_id": "386171184", "text": "#!/usr/bin/python\r\n# -*- coding: utf-8 -*-\r\nimport xdrlib, sys\r\nimport os\r\nimport xlrd\r\nimport queue\r\nimport urllib\r\nimport urllib.request as urllib2\r\nimport re\r\nimport threading\r\nimport time\r\n\r\n\r\ndef download_from_hmbd(name):\r\n # name = 'Nicotinic acid'\r\n name_quote = urllib.parse.quote_plus(name)\r\n # print(name)\r\n url = 'http://www.hmdb.ca/unearth/q?utf8=%E2%9C%93&query=' + name_quote + '&searcher=metabolites&button='\r\n # print(url)\r\n content = urllib2.urlopen(url).read().decode()\r\n\r\n m = re.search(\"hit-name.{40}\", content)\r\n m = re.split('\"', m.group())\r\n ID = re.split('/', m[2])[-1]\r\n\r\n url = 'http://www.hmdb.ca/metabolites/' + ID\r\n content = urllib2.urlopen(url).read().decode()\r\n m = re.search(\".{40}1H NMR Spectrum\", content)\r\n nmrID = re.search(\"/spectra/nmr_one_d/[0-9]*\", m.group()).group()\r\n\r\n url = 'http://www.hmdb.ca' + nmrID\r\n content = urllib2.urlopen(url).read().decode()\r\n m = re.search(\"List of chemical shift values for the spectrum.{200}\", content)\r\n m = re.search(\"a href=.*\", m.group())\r\n file_url = re.split('\"', m.group())[1]\r\n\r\n data = urllib2.urlopen(file_url).read().decode()\r\n\r\n f = open('database/' + name + '.txt', 'w')\r\n f.write(data)\r\n f.close()\r\n # print(content)\r\n\r\n\r\ndef open_excel(file='metabolites list.xlsx'):\r\n try:\r\n data = xlrd.open_workbook(file)\r\n return data\r\n except Exception as e:\r\n print(str(e))\r\n\r\n\r\n# 根据索引获取Excel表格中的数据 参数:file:Excel文件路径 colnameindex:表头列名所在行的所以 ,by_index:表的索引\r\ndef excel_table_byindex(file='metabolites list.xlsx', colnameindex=0, by_index=0):\r\n data = open_excel(file)\r\n table = data.sheets()[by_index]\r\n nrows = table.nrows # 行数\r\n ncols = table.ncols # 列数\r\n colnames = table.row_values(colnameindex) # 某一行数据\r\n list = []\r\n for rownum in range(1, nrows):\r\n\r\n row = table.row_values(rownum)\r\n if row:\r\n app = {}\r\n for i in range(len(colnames)):\r\n app[colnames[i]] = row[i]\r\n list.append(app)\r\n return list\r\n\r\nclass myThread (threading.Thread): #继承父类threading.Thread\r\n def __init__(self, threadID, name, counter):\r\n threading.Thread.__init__(self)\r\n self.threadID = threadID\r\n self.name = name\r\n self.counter = counter\r\n def run(self): #把要执行的代码写到run函数里面 线程在创建后会直接运行run函数 \r\n print( \"\\nStarting \" + self.name)\r\n download_from_hmbd(self.name)\r\n print( \"\\nExiting \" + self.name)\r\n\r\ndef main():\r\n tables = excel_table_byindex()\r\n q = queue.Queue()\r\n for row in tables:\r\n name_check = row['name']\r\n name_path = os.getcwd()+ '\\\\database\\\\'+ name_check+'.txt'\r\n print('Check ' +name_check + '...')\r\n if(os.path.exists(name_path)):\r\n continue\r\n else:\r\n q.put(name_check )\r\n\r\n #print(row)\r\n #print(row['name'] )\r\n \r\n while not q.empty():\r\n name_uodate = q.get()\r\n thread1 = myThread(1, name_uodate, 1).start()\r\n #download_from_hmbd(name_check)\r\n\r\n print( \"Exiting Main Thread\")\r\n\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n print(\"Update finish!\")\r\n", "sub_path": "Metabolite Qulification/database_update.py", "file_name": "database_update.py", "file_ext": "py", "file_size_in_byte": 3346, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "urllib.parse.quote_plus", "line_number": 16, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 16, "usage_type": "attribute"}, {"api_name": "urllib.request.urlopen", "line_number": 20, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 20, "usage_type": "name"}, {"api_name": "re.search", "line_number": 22, "usage_type": "call"}, {"api_name": "re.split", "line_number": 23, "usage_type": "call"}, {"api_name": "re.split", "line_number": 24, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 27, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 27, "usage_type": "name"}, {"api_name": "re.search", "line_number": 28, "usage_type": "call"}, {"api_name": "re.search", "line_number": 29, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 32, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 32, "usage_type": "name"}, {"api_name": "re.search", "line_number": 33, "usage_type": "call"}, {"api_name": "re.search", "line_number": 34, "usage_type": "call"}, {"api_name": "re.split", "line_number": 35, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 37, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 37, "usage_type": "name"}, {"api_name": "xlrd.open_workbook", "line_number": 47, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 71, "usage_type": "attribute"}, {"api_name": "threading.Thread.__init__", "line_number": 73, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 73, "usage_type": "attribute"}, {"api_name": "queue.Queue", "line_number": 84, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 89, "usage_type": "call"}, {"api_name": "os.path", "line_number": 89, "usage_type": "attribute"}]} +{"seq_id": "299474241", "text": "import argparse\nimport time\nfrom PIL import Image\nimport string\nimport random\nimport os\n\nparser = argparse.ArgumentParser(description=\"Create matrix styled effect with an use of a image mask\")\nparser.add_argument('-mask', metavar='m', help=\"Path of an image mask\")\nparser.add_argument('--x', metavar='x', type=int, help=\"x size of the final output\", default=-1)\nparser.add_argument('--y', metavar='y', type=int, help=\"y size of the finak output\", default=-1)\nparser.add_argument('--iterations', metavar='i', type=int, help=\"how many iterations should the animation go trough, 0 for infinite\", default=0)\nparser.add_argument('--delay', metavar='d', type=float, help=\"how many seconds to wait after each iteration\", default=1)\n\nargs = parser.parse_args()\n\nif args.iterations < 0:\n raise Exception(\"Please enter a positive number for iterations\")\n\nsizeResult = os.popen(\"stty size\", 'r').read().split()\n\nwindowX = 0\nwindowY = 0\n\nif sizeResult.__len__() != 0:\n windowY, windowX = sizeResult\n windowX = int(windowX)\n windowY = int(windowY)\n\ninfinite = args.iterations == 0\nimagePath = args.mask\nresultWidth = args.x\nresultHeight = args.y\ndelay = args.delay\n\nim = Image.open(imagePath)\npixels = im.load()\nwidth, height = im.size\n\nif resultWidth == -1:\n resultWidth = width\n\nif resultHeight == -1:\n resultHeight = height\n\nwidthStep = width / resultWidth\nheightStep = height / resultHeight\n\noffsetX = int((windowX - resultWidth) / 2)\noffsetY = int((windowY - resultHeight) / 2)\n\nprint(offsetX, offsetY)\n\nsymbolPool = string.ascii_letters + string.digits + \"`-=[]\\'\\\\/.,<;~!@#$%^&*()_+}{|\\\":?>\"\n\nwhile True:\n\n render = \"\"\n\n for y in range(resultHeight):\n\n for i in range(int(offsetX)):\n render += \" \"\n\n for x in range(resultWidth):\n imgX = int(x * widthStep)\n imgY = int(y * heightStep)\n pix = pixels[imgX, imgY]\n\n if pix == 1:\n render += random.choice(symbolPool)\n else:\n render += \" \"\n\n render += \"\\n\"\n\n for i in range(int(offsetY)):\n render += \"\\n\"\n\n print(render, end=\"\")\n\n time.sleep(delay)\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2144, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 8, "usage_type": "call"}, {"api_name": "os.popen", "line_number": 20, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 36, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 36, "usage_type": "name"}, {"api_name": "string.ascii_letters", "line_number": 54, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 54, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 71, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 82, "usage_type": "call"}]} +{"seq_id": "452302143", "text": "import logging\nimport re\nimport math\nfrom typing import Callable, List, Tuple, Type, Iterable, Dict\nimport itertools\nfrom overrides import overrides\nimport argparse\nimport sys\nimport os\nimport tarfile\nimport torch\nimport torch.nn.init\nimport numpy as np\nimport pprint\nimport json\n\nfrom allennlp.common import Registrable\nfrom allennlp.common.params import Params\nfrom allennlp.common.checks import ConfigurationError\nlogger = logging.getLogger(__name__) # pylint: disable=invalid-name\n\n\ndef parse_args():\n \"\"\"Parse command line arguments.\n\n Args:\n None.\n\n Returns:\n A argparse.Namespace object which contains all parsed argument values.\n \"\"\"\n parser = argparse.ArgumentParser(description=__doc__)\n parser.add_argument(\n '--print-tensors', help='whether print tensors', default=False)\n parser.add_argument(\n '--export-unknown-token',\n help='whether export unkown token',\n default=False)\n parser.add_argument('--serialization-dir', help='serialization dir path')\n parser.add_argument(\n '--weights-file-name', help='weigth file name', default='best.th')\n parser.add_argument(\n '--embedder-name',\n help='embedder name',\n default='bert.embeddings.word_embeddings.weight')\n parser.add_argument(\n '--vocab-file', help='vocab file', default='vocabulary')\n parser.add_argument(\n '--output-embedding-file',\n help='output embedding file name',\n default='exported_embedding.tar.gz')\n args = parser.parse_args()\n return args\n\n\ndef load_weights(serialization_dir, weights_file_name, embedder_name):\n weights_file_path = os.path.join(serialization_dir, weights_file_name)\n weights: Dict[str, torch.Tensor] = torch.load(weights_file_path)\n for name, weight in weights.items():\n print(name)\n embedder_weight = weights[embedder_name]\n print(embedder_weight.size())\n return embedder_weight.data.cpu().numpy()[:, :]\n\n\ndef load_vocab(serialization_dir, vocab_file):\n vocab_file_path = os.path.join(serialization_dir, vocab_file)\n tokens = []\n with open(vocab_file_path, 'r') as f:\n for line in f:\n tokens.append(line.strip())\n tokens[0] = \"@@PADDING@@\"\n return tokens\n\n\ndef save_embedding_file(weights, tokens, serialization_dir,\n output_embedding_file):\n embedding_fn = os.path.join(serialization_dir, 'exported_embedding.txt')\n with open(embedding_fn, 'w') as f:\n size, dim = weights.shape\n f.write('{} {}\\n'.format(size, dim))\n for i in range(size):\n f.write('{} {}\\n'.format(\n tokens[i], ' '.join([str(s) for s in weights[i].tolist()])))\n with tarfile.open(\n os.path.join(serialization_dir, output_embedding_file),\n 'w:gz') as tar_file:\n tar_file.add(embedding_fn, 'exported_embedding.txt')\n\n\ndef print_tensors(serialization_dir, weights_file_name):\n weights_file_path = os.path.join(serialization_dir, weights_file_name)\n weights: Dict[str, torch.Tensor] = torch.load(weights_file_path)\n arr = []\n map = {}\n prefix = '_inner_model'\n for name, weight in weights.items():\n if name.startswith('_transformer'):\n arr.append(name)\n key = '{}.{}'.format(prefix, name)\n map[key] = name\n pp = pprint.PrettyPrinter(indent=4)\n print(json.dumps(map))\n print('|'.join(arr))\n\n\ndef main():\n args = parse_args()\n if args.print_tensors:\n print_tensors(args.serialization_dir, args.weights_file_name)\n weights = load_weights(args.serialization_dir, args.weights_file_name,\n args.embedder_name)\n tokens = load_vocab(args.serialization_dir, args.vocab_file)\n print(len(tokens))\n print(weights.shape[0])\n assert len(tokens) == weights.shape[0]\n save_embedding_file(weights, tokens, args.serialization_dir,\n args.output_embedding_file)\n\n\nif __name__ == \"__main__\":\n sys.exit(main())\n", "sub_path": "scripts/export_embedding_file.py", "file_name": "export_embedding_file.py", "file_ext": "py", "file_size_in_byte": 4002, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "logging.getLogger", "line_number": 20, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 58, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 58, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path", "line_number": 78, "usage_type": "attribute"}, {"api_name": "tarfile.open", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path", "line_number": 92, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 93, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 93, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 93, "usage_type": "call"}, {"api_name": "pprint.PrettyPrinter", "line_number": 102, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 103, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 122, "usage_type": "call"}]} +{"seq_id": "348941006", "text": "import os\nimport warnings\nfrom pickle import PicklingError\n\nimport numpy as np\nimport pandas as pd\nimport cloudpickle\n\nfrom .data_processing import read_data_with_metadata\nfrom .data_processing import preprocess_time\n\n\ndef make_model_env_class(system_env_object):\n \"\"\"Make the ModelEnv class dynamically inherit from the system env class.\n\n Parameters\n ----------\n system_env_object : object\n AI Gym environment of the system.\n\n Returns\n -------\n ModelEnv : object\n See docstring of the ModelEnv class for more details.\n \"\"\"\n\n class ModelEnv(system_env_object):\n \"\"\"Open AI gym env.\n\n Environment associated to the learnt model of a model based RL\n strategy. See the docstring of the workflow step method for the\n description of the history.\n\n Parameters\n ----------\n submission_path : string\n Path of the submission.\n problem_module : object\n Problem module.\n reward_func : function\n Function taking as input observations and returning as output the\n associated reward. Can also depend on the taken actions.\n metadata : dictionary\n Providing the names of observations, actions and restart variables.\n Associated keys respectively are 'observation', 'action' and\n 'restart_name'.\n output_dir : string\n Path of the output directory, where to find the data to train the\n model.\n seed : int\n Seed of the RNG used for this environment.\n \"\"\"\n\n def __init__(self, submission_path, problem_module, reward_func,\n metadata, output_dir, seed=None):\n self.submission_path = submission_path\n self.reward_func = reward_func\n self.metadata = metadata\n self.output_dir = output_dir\n super(ModelEnv, self).seed(seed)\n\n # only storing needed problem_module attributes as problem_module\n # can be problematic to pickle\n self.n_burn_in = problem_module._n_burn_in\n self.workflow_step = problem_module.workflow.step\n self.get_train_data = problem_module.get_train_data\n self.train_submission = problem_module.workflow.train_submission\n self._get_column_names(metadata)\n\n # set history to None, this attribute is used to check if an env\n # has an history that needs to be set\n self.history = None\n\n def _get_column_names(self, metadata):\n self.action_names = metadata['action']\n self.observation_names = metadata['observation']\n self.restart_name = metadata['restart_name']\n\n def set_history(self, observation, restart):\n \"\"\"Reset the history with the given observations\n\n Parameters\n ----------\n observation : array, shape (n_features)\n Observations.\n restart : int\n Whether the observation is the first of an episode. This is\n used to know the history the model can use.\n \"\"\"\n # reset history\n history_col_names = (['time'] + self.observation_names +\n self.action_names + [self.restart_name])\n n_action_features = len(self.action_names)\n # XXX we use 0 as an arbitrary start time\n # each sample of the history is made of one observation and one\n # action, the action being the one selected after the given\n # observation.\n # the unknown next action is set to NaN for now and will be\n # replaced by the action of the next call to step.\n history = np.r_[\n 0, observation.ravel(), [np.nan] * n_action_features, restart]\n history = history.reshape(1, -1)\n history = (pd.DataFrame(data=history, columns=history_col_names)\n .set_index('time'))\n self.history = history\n\n def add_observation_to_history(self, observation, restart):\n \"\"\"Update history with the new given observation\n\n Parameters\n ----------\n observation : array, shape (n_features)\n Observations.\n restart : int\n Whether the observation is the first of an episode. This is\n used to know the history the model can use.\n \"\"\"\n if restart:\n self.set_history(observation, restart)\n else:\n history_df = self.history\n # the action is set to NaN for now and will be updated when\n # needed\n new_sample_col_names = (\n ['time'] + self.observation_names +\n self.action_names + [self.restart_name])\n new_time = history_df.index[-1] + 1\n\n # the unknown next action is set to NaN for now and will be\n # replaced by the action of the next call to step.\n n_action_features = len(self.action_names)\n new_sample = np.r_[\n new_time, observation,\n [np.nan] * n_action_features,\n restart]\n new_sample = new_sample.reshape(1, -1)\n new_sample = pd.DataFrame(\n data=new_sample,\n columns=new_sample_col_names).set_index('time')\n history_df = pd.concat([history_df, new_sample], axis=0)\n # keep history size less than n_burn_in + 1 samples\n self.history = history_df.iloc[-(self.n_burn_in + 1):]\n\n def add_action_to_history(self, action):\n \"\"\"Update history with the given action.\n\n Add this action to the last observation of the history.\n\n Parameters\n ----------\n action : array\n Action.\n \"\"\"\n action_col_num = self.history.columns.get_indexer(\n self.action_names)\n self.history.iloc[-1, action_col_num] = action\n\n def reset(self):\n \"\"\"Reset method of the environment.\n\n The history of the model is also reset.\n\n Returns\n -------\n observation : numpy array, shape (n_observations,)\n The passed observation if not None or a new observation.\n \"\"\"\n observation = super(ModelEnv, self).reset()\n self.set_history(observation, 1)\n\n return observation\n\n def _workflow_step(self, history, seed=None):\n \"\"\"Compute step from history history.\n\n Parameters\n ----------\n history : pandas DataFrame\n History. Contains past data and the new action.\n\n seed : int\n Seed of the RNG used to sample the new observation.\n\n Returns\n -------\n observation : pandas DataFrame\n The sampled observation.\n \"\"\"\n observations = self.workflow_step(\n self.model, history, random_state=seed)\n return observations\n\n def step(self, action):\n \"\"\"Step function of the model environment.\n\n The history of the environment is used by the model for the\n dynamics prediction and updated at each step with the given action\n and returned observations.\n Note that done is returned for compatibility but is always set to\n 0 as we do not consider early terminations when using the model.\n\n Parameters\n ----------\n action : numpy array, shape (n_action_features,)\n The action to be taken.\n\n Returns\n -------\n observation : numpy array, shape (n_observations,)\n The sampled observations.\n\n reward : float\n Reward computed from the taken action and the obtained\n observations.\n\n done : int\n 0 is always returned.\n\n info : dict\n Empty dict, used for compatibility with AI Gym API.\n \"\"\"\n self.add_action_to_history(action)\n\n observation = self._workflow_step(self.history, seed=None)\n observation = observation.to_numpy().ravel()\n reward = self.reward_func(np.r_[observation, action])\n\n self.add_observation_to_history(observation, 0)\n\n # we do not terminate early when using the models and thus always\n # return 0 for the done variable\n done = 0\n\n return observation, reward, done, {}\n\n def train_model(self, epoch):\n \"\"\"Update model with collected data.\n\n Parameters\n ----------\n epoch : int\n Epoch of the main loop. Used to know how many traces should be\n used to update the model.\n \"\"\"\n output_dir = self.output_dir\n metadata = self.metadata\n\n # get all previous traces, concatenate them and update model\n trace_paths = [os.path.join(output_dir, f'epoch_{i}', 'trace.csv')\n for i in range(epoch + 1)]\n trace_dfs = []\n for trace_path in trace_paths:\n trace_df = read_data_with_metadata(trace_path, metadata)\n trace_dfs.append(trace_df)\n all_traces = pd.concat(trace_dfs, axis=0).reset_index(drop=True)\n all_traces = preprocess_time(all_traces, metadata)\n\n epoch_output_dir = os.path.join(output_dir, f'epoch_{epoch}')\n training_data_dir = os.path.join(epoch_output_dir, 'data')\n if not os.path.exists(training_data_dir):\n os.makedirs(training_data_dir)\n\n all_traces.to_csv(os.path.join(training_data_dir, 'X_train.csv'))\n X_train, y_train = self.get_train_data(epoch_output_dir)\n trained_model = self.train_submission(\n self.submission_path, X_train, y_train)\n\n # saving trained model, will raise an error if a model cannot be\n # pickled\n model_filename = os.path.join(\n epoch_output_dir, 'trained_submission.pkl')\n with open(model_filename, 'wb') as f:\n try:\n cloudpickle.dump(trained_model, f)\n except PicklingError:\n msg = ('Using dill instead of cloudpickle to pickle '\n 'trained submission.')\n warnings.warn(msg)\n import dill\n dill.dump(trained_model, f)\n\n self.model = trained_model\n\n def __getstate__(self):\n \"\"\"Needed to override this method of the parent class.\n\n Sometimes the parent class, the system environment object,\n implements its own __getstate__ method and makes the copy of\n the ModelEnv object fail.\"\"\"\n return self.__dict__.copy()\n\n def __setstate__(self, state):\n \"\"\"Needed to override this method of the parent class.\n\n Sometimes the parent class, the system environment object,\n implements its own __setstate__ method and makes the copy of\n the ModelEnv object fail.\"\"\"\n self.__dict__.update(state)\n\n return ModelEnv\n", "sub_path": "mbrl-tools/mbrltools/model_env.py", "file_name": "model_env.py", "file_ext": "py", "file_size_in_byte": 11494, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "numpy.r_", "line_number": 100, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 101, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.r_", "line_number": 132, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 134, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 137, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.r_", "line_number": 226, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 249, "usage_type": "call"}, {"api_name": "os.path", "line_number": 249, "usage_type": "attribute"}, {"api_name": "data_processing.read_data_with_metadata", "line_number": 253, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 255, "usage_type": "call"}, {"api_name": "data_processing.preprocess_time", "line_number": 256, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 258, "usage_type": "call"}, {"api_name": "os.path", "line_number": 258, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 259, "usage_type": "call"}, {"api_name": "os.path", "line_number": 259, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 260, "usage_type": "call"}, {"api_name": "os.path", "line_number": 260, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 261, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 263, "usage_type": "call"}, {"api_name": "os.path", "line_number": 263, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 270, "usage_type": "call"}, {"api_name": "os.path", "line_number": 270, "usage_type": "attribute"}, {"api_name": "cloudpickle.dump", "line_number": 274, "usage_type": "call"}, {"api_name": "pickle.PicklingError", "line_number": 275, "usage_type": "name"}, {"api_name": "warnings.warn", "line_number": 278, "usage_type": "call"}, {"api_name": "dill.dump", "line_number": 280, "usage_type": "call"}]} +{"seq_id": "129875519", "text": "from codecs import open\nfrom os import path\nimport sys\nfrom setuptools import setup, find_packages\nfrom setuptools.command.test import test as TestCommand\n\nBASE_DIR = path.abspath(path.dirname(__file__))\n\n\nclass PyTest(TestCommand):\n\n def finalize_options(self):\n TestCommand.finalize_options(self)\n\n self.test_args = []\n self.test_suite = True\n\n def run_tests(self):\n import pytest\n\n sys.exit(pytest.main(self.test_args))\n\n\ntests_require = ['pytest', 'mock', 'pretend']\n\nsetup(\n name='ppm',\n version='0.0.1',\n description='A package manager for Python',\n long_description='A package manager for Python',\n url='https://github.com/patrickporto/ppm',\n author='Patrick Porto',\n author_email='ppm-dev@groups.google.com',\n license='MIT',\n classifiers=[\n 'Development Status :: 3 - Alpha',\n 'Intended Audience :: Developers',\n 'Topic :: Software Development :: Build Tools',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.4',\n 'Programming Language :: Python :: 3.5',\n ],\n keywords='pip easy_install distutils setuptools wheel virtualenv',\n packages=find_packages(exclude=['contrib', 'docs', 'tests']),\n extras_require={\n 'dev': ['check-manifest'],\n 'test': ['coverage'],\n 'testing': tests_require,\n },\n entry_points={\n 'console_scripts': [\n 'sample=sample:main',\n ],\n },\n\n tests_require=tests_require,\n cmdclass={'test': PyTest},\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1687, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "os.path.abspath", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 7, "usage_type": "call"}, {"api_name": "setuptools.command.test.test", "line_number": 10, "usage_type": "name"}, {"api_name": "setuptools.command.test.test.finalize_options", "line_number": 13, "usage_type": "call"}, {"api_name": "setuptools.command.test.test", "line_number": 13, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 21, "usage_type": "call"}, {"api_name": "pytest.main", "line_number": 21, "usage_type": "call"}, {"api_name": "setuptools.setup", "line_number": 26, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 47, "usage_type": "call"}]} +{"seq_id": "330811557", "text": "\"\"\"\nGraph comparing different call times.\n\"\"\"\nimport matplotlib.pyplot as plt\nimport numpy as np\n\nfrom convertible_bond import dS_total as dS, payoff, C, T\nfrom model import FDEModel\nfrom plot import plotmain\n\ndef main():\n S = np.linspace(0, 200, 200 * 8 + 1)\n Sl = 0\n Su = 250\n N = 128 * T\n K = 8 * (Su - Sl)\n Sk = (K * (S - Sl) / (Su - Sl)).astype(int)\n legend = []\n label = \"$\\\\Omega^c = \\\\{%i\\\\}$\"\n\n model = FDEModel(N, dS, payoff)\n fig = plt.figure()\n\n ax = fig.add_subplot(111)\n for i in range(1, 5):\n C.times = [i]\n ax.plot(S, model.price(Sl, Su, K).V[0][Sk])\n legend.append(label % i)\n plt.xlabel(\"Stock Price\")\n plt.ylabel(\"Convertible Bond Price\")\n plt.legend(legend)\n\n\nif __name__ == \"__main__\":\n plotmain(main)\n", "sub_path": "src/fig_varTc.py", "file_name": "fig_varTc.py", "file_ext": "py", "file_size_in_byte": 792, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "numpy.linspace", "line_number": 12, "usage_type": "call"}, {"api_name": "convertible_bond.T", "line_number": 15, "usage_type": "name"}, {"api_name": "model.FDEModel", "line_number": 21, "usage_type": "call"}, {"api_name": "convertible_bond.dS_total", "line_number": 21, "usage_type": "argument"}, {"api_name": "convertible_bond.payoff", "line_number": 21, "usage_type": "argument"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "convertible_bond.C.times", "line_number": 26, "usage_type": "attribute"}, {"api_name": "convertible_bond.C", "line_number": 26, "usage_type": "name"}, {"api_name": "model.price", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "plot.plotmain", "line_number": 35, "usage_type": "call"}]} +{"seq_id": "39067422", "text": "from collections import namedtuple # For using namedtuple\n\n# Cluster's protocol uses fifteen different message types, each defined as a Python namedtuple.\nAccepted = namedtuple('Accepted', ['slot', 'ballot_num'])\nAccept = namedtuple('Accept', ['slot', 'ballot_num', 'proposal'])\nDecision = namedtuple('Decision', ['slot', 'proposal'])\nInvoked = namedtuple('Invoded', ['client_id', 'output'])\nInvoke = namedtuple('Invode', ['caller', 'client_id', 'input_value'])\nJoin = namedtuple('Join', [])\nActive = namedtuple('Active', [])\nPrepare = namedtuple('Prepare', ['ballot_num'])\nPromise = namedtuple('Promise', ['ballot_num', 'accepted_proposals'])\nPropose = namedtuple('Propose', ['slot', 'proposal'])\nWelcome = namedtuple('Welcome', ['state', 'slot', 'decisions'])\nDecided = namedtuple('Decided', ['slot'])\nPreempted = namedtuple('Preempted', ['slot', 'preempted_by'])\nAdopted = namedtuple('Adopted', ['ballot_num', 'accepted_proposals'])\nAccepting = namedtuple('Accepting', ['leader'])\n\n# Cluster uses two data types named to correspond to the protocol description\nProposal = namedtuple('Proposal', ['caller', 'client_id', 'input'])\nBallot = namedtuple('Ballot', ['n', 'leader'])\n\n# timeouts for various messages\nJOIN_RETRNSMIT = 0.7\nCATCHUP_INTERVAL = 0.6\nACCEPT_RETRANSMIT = 1.0\nPREPARE_RETRANSMIT = 1.0\nINVOKE_RETRANSMIT = 0.5\nLEADER_TIMEOUT = 1.0\nNULL_BALLOT = Ballot(-1, -1)\nNOOP_PROPOSAL = Proposal(None, None, None)\n", "sub_path": "paxos.py", "file_name": "paxos.py", "file_ext": "py", "file_size_in_byte": 1421, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "collections.namedtuple", "line_number": 4, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 5, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 6, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 7, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 8, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 9, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 10, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 11, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 12, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 13, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 14, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 15, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 16, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 17, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 18, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 21, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 22, "usage_type": "call"}]} +{"seq_id": "580062429", "text": "from django.conf.urls import url\nfrom deals import views\n\nurlpatterns = [\n url(r'^contacts/(?P[0-9]+)/$', views.deal_contacts),\n url(r'^total/$', views.get_sum_of_budgets),\n url(r'^dashboard/$', views.get_report),\n url(r'^dashboard/deals/$', views.get_deal_report),\n url(r'^dashboard/contacts/$', views.get_contact_report),\n url(r'^stage/(?P[0-9]+)/$', views.stage_detail),\n url(r'^funnels/$', views.funnel_list_l),\n url(r'^funnel/(?P[0-9]+)/$', views.funnel_detail),\n url(r'^stages/(?P[0-9]+)/$', views.funnel_list_tel),\n url(r'^stages/$', views.funnel_list_front),\n url(r'^stage/$', views.stage_list),\n url(r'^funnel/$', views.funnel_list),\n url(r'^delete/$', views.delete_deals),\n url(r'^change_stage/$', views.change_statuses),\n url(r'^change_responsible/$', views.change_responsible),\n url(r'^get_related/$', views.get_deals_co),\n url(r'^(?P[0-9]+)/$', views.deal_detail),\n url(r'^average/$', views.calculate_average_in_Stage),\n url(r'^average2/$', views.calculate_average),\n url(r'^complete/$', views.complete),\n url(r'^d_related/$', views.delete_relations),\n url(r'^hola/$', views.hola),\n url(r'^sql_query/$', views.sql_query),\n url(r'^count/$', views.sql_query_count),\n url(r'^choose_main/$', views.choose_main),\n url(r'^$', views.deal_list),\n]\n", "sub_path": "deals/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1355, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "django.conf.urls.url", "line_number": 5, "usage_type": "call"}, {"api_name": "deals.views.deal_contacts", "line_number": 5, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 5, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 6, "usage_type": "call"}, {"api_name": "deals.views.get_sum_of_budgets", "line_number": 6, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 6, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "deals.views.get_report", "line_number": 7, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 7, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "deals.views.get_deal_report", "line_number": 8, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 8, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "deals.views.get_contact_report", "line_number": 9, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 9, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "deals.views.stage_detail", "line_number": 10, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 10, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "deals.views.funnel_list_l", "line_number": 11, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 11, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "deals.views.funnel_detail", "line_number": 12, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 12, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "deals.views.funnel_list_tel", "line_number": 13, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 13, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 14, "usage_type": "call"}, {"api_name": "deals.views.funnel_list_front", "line_number": 14, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 14, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 15, "usage_type": "call"}, {"api_name": "deals.views.stage_list", "line_number": 15, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 15, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 16, "usage_type": "call"}, {"api_name": "deals.views.funnel_list", "line_number": 16, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 16, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 17, "usage_type": "call"}, {"api_name": "deals.views.delete_deals", "line_number": 17, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 17, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 18, "usage_type": "call"}, {"api_name": "deals.views.change_statuses", "line_number": 18, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 18, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 19, "usage_type": "call"}, {"api_name": "deals.views.change_responsible", "line_number": 19, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 19, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 20, "usage_type": "call"}, {"api_name": "deals.views.get_deals_co", "line_number": 20, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 20, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 21, "usage_type": "call"}, {"api_name": "deals.views.deal_detail", "line_number": 21, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 21, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 22, "usage_type": "call"}, {"api_name": "deals.views.calculate_average_in_Stage", "line_number": 22, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 22, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 23, "usage_type": "call"}, {"api_name": "deals.views.calculate_average", "line_number": 23, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 23, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 24, "usage_type": "call"}, {"api_name": "deals.views.complete", "line_number": 24, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 24, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 25, "usage_type": "call"}, {"api_name": "deals.views.delete_relations", "line_number": 25, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 25, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 26, "usage_type": "call"}, {"api_name": "deals.views.hola", "line_number": 26, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 26, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 27, "usage_type": "call"}, {"api_name": "deals.views.sql_query", "line_number": 27, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 27, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 28, "usage_type": "call"}, {"api_name": "deals.views.sql_query_count", "line_number": 28, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 28, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 29, "usage_type": "call"}, {"api_name": "deals.views.choose_main", "line_number": 29, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 29, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 30, "usage_type": "call"}, {"api_name": "deals.views.deal_list", "line_number": 30, "usage_type": "attribute"}, {"api_name": "deals.views", "line_number": 30, "usage_type": "name"}]} +{"seq_id": "390143183", "text": "from stripe import Charge\nfrom stripe.resource import convert_to_stripe_object, populate_headers\n\nimport stripe\n\nimport sys\ncurrent_module = sys.modules[__name__]\n\napi_key = None\napi_base = 'https://api.pandapay.io'\n\nclass AuthenticationError(Exception):\n pass\n\nclass APIError(stripe.error.APIError):\n pass\n\nclass APIRequestor(stripe.api_requestor.APIRequestor):\n def request(self, *args, **kwargs):\n try:\n return super(APIRequestor, self).request(*args, **kwargs)\n except stripe.error.APIError as e:\n raise APIError(e.message, e.http_body, e.http_status, e.json_body)\n\nclass StripeCharge(Charge):\n @classmethod\n def create(cls, api_key=None, idempotency_key=None,\n stripe_account=None, **params):\n api_key_to_use = api_key or current_module.api_key\n\n if not api_key_to_use:\n raise AuthenticationError('No API key provided. (HINT: set your API key using '\n '\"pandaecs.api_key = \"). You can find your API Keys '\n 'on the the PandaPay web interface. See '\n 'https://www.pandapay.io/docs/api-reference for details, or email '\n 'support@pandapay.io if you have any questions.')\n\n\n requestor = APIRequestor(\n api_key_to_use,\n account=stripe_account,\n api_base=api_base\n )\n\n url = cls.class_url()\n headers = populate_headers(idempotency_key)\n\n params_to_use = dict()\n stripe_params = dict(params)\n\n params_to_use['donation_amount'] = stripe_params.pop('donation_amount', None)\n params_to_use['destination'] = stripe_params.pop('destination_ein', None)\n params_to_use['receipt_email'] = stripe_params.pop('receipt_email', None)\n params_to_use['currency'] = stripe_params.get('currency', None)\n params_to_use['stripe_params'] = stripe_params\n\n response, returned_api_key = requestor.request('post', url, params_to_use, headers)\n\n return StripeCharge.construct_from(response, api_key_to_use, stripe_account)\n\n @classmethod\n def class_url(cls):\n return '/v1/donations'\n\n @classmethod\n def construct_from(cls, resp, api_key, stripe_account):\n panda_charge = cls(resp['id'], api_key, stripe_account)\n\n for k, v in resp.iteritems():\n if k == 'stripe_response':\n next\n\n panda_charge.__setitem__(k, v)\n\n panda_charge.original_charge = Charge.construct_from(\n resp.get('stripe_response', {}),\n api_key,\n stripe_account\n )\n\n return panda_charge\n\n", "sub_path": "pandaecs/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 2734, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "sys.modules", "line_number": 7, "usage_type": "attribute"}, {"api_name": "stripe.error", "line_number": 15, "usage_type": "attribute"}, {"api_name": "stripe.api_requestor", "line_number": 18, "usage_type": "attribute"}, {"api_name": "stripe.error", "line_number": 22, "usage_type": "attribute"}, {"api_name": "stripe.Charge", "line_number": 25, "usage_type": "name"}, {"api_name": "stripe.resource.populate_headers", "line_number": 46, "usage_type": "call"}, {"api_name": "stripe.Charge.construct_from", "line_number": 75, "usage_type": "call"}, {"api_name": "stripe.Charge", "line_number": 75, "usage_type": "name"}]} +{"seq_id": "248653920", "text": "# MIT License\n#\n# Copyright (C) The Adversarial Robustness Toolbox (ART) Authors 2021\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated\n# documentation files (the \"Software\"), to deal in the Software without restriction, including without limitation the\n# rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit\n# persons to whom the Software is furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in all copies or substantial portions of the\n# Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE\n# WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,\n# TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n# SOFTWARE.\n\"\"\"\nThis module implements EoT of changes in brightness by addition of uniformly sampled delta.\n\"\"\"\nimport logging\nfrom typing import Tuple, Union, TYPE_CHECKING, Optional\n\nimport numpy as np\n\nfrom art.preprocessing.expectation_over_transformation.tensorflow import EoTTensorFlowV2\n\nif TYPE_CHECKING:\n import tensorflow as tf\n\nlogger = logging.getLogger(__name__)\n\n\nclass EoTBrightnessTensorFlow(EoTTensorFlowV2):\n \"\"\"\n This module implements EoT of changes in brightness by addition of uniformly sampled delta.\n \"\"\"\n\n def __init__(\n self,\n nb_samples: int,\n clip_values: Tuple[float, float],\n delta: Union[float, Tuple[float, float]],\n apply_fit: bool = False,\n apply_predict: bool = True,\n ) -> None:\n \"\"\"\n Create an instance of EoTBrightnessTensorFlow.\n\n :param nb_samples: Number of random samples per input sample.\n :param clip_values: Tuple of float representing minimum and maximum values of input `(min, max)`.\n :param delta: Range to sample the delta (addition) to the pixel values to adjust the brightness. A single float\n is translated to range [-delta, delta] or a tuple of floats is used to create sampling range\n [delta[0], delta[1]]. The applied delta is sampled uniformly from this range for each image.\n :param apply_fit: True if applied during fitting/training.\n :param apply_predict: True if applied during predicting.\n \"\"\"\n super().__init__(\n apply_fit=apply_fit, apply_predict=apply_predict, nb_samples=nb_samples, clip_values=clip_values\n )\n\n self.delta = delta\n self.delta_range = (-delta, delta) if isinstance(delta, (int, float)) else delta\n self._check_params()\n\n def _transform(\n self, x: \"tf.Tensor\", y: Optional[\"tf.Tensor\"], **kwargs\n ) -> Tuple[\"tf.Tensor\", Optional[\"tf.Tensor\"]]:\n \"\"\"\n Transformation of an image with randomly sampled brightness.\n\n :param x: Input samples.\n :param y: Label of the samples `x`.\n :return: Transformed samples and labels.\n \"\"\"\n import tensorflow as tf # lgtm [py/repeated-import]\n\n delta_i = np.random.uniform(low=self.delta_range[0], high=self.delta_range[1])\n return tf.clip_by_value(x + delta_i, clip_value_min=self.clip_values[0], clip_value_max=self.clip_values[1]), y\n\n def _check_params(self) -> None:\n\n # pylint: disable=R0916\n if not isinstance(self.delta, (int, float, tuple)) or (\n isinstance(self.delta, tuple)\n and (\n len(self.delta) != 2\n or not isinstance(self.delta[0], (int, float))\n or not isinstance(self.delta[1], (int, float))\n or self.delta[0] > self.delta[1]\n )\n ):\n raise ValueError(\"The argument `delta` has to be a float or tuple of two float values as (min, max).\")\n", "sub_path": "art/preprocessing/expectation_over_transformation/natural_corruptions/brightness/tensorflow.py", "file_name": "tensorflow.py", "file_ext": "py", "file_size_in_byte": 4081, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "typing.TYPE_CHECKING", "line_number": 28, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 31, "usage_type": "call"}, {"api_name": "art.preprocessing.expectation_over_transformation.tensorflow.EoTTensorFlowV2", "line_number": 34, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 42, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 43, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 43, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 67, "usage_type": "name"}, {"api_name": "numpy.random.uniform", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 78, "usage_type": "attribute"}, {"api_name": "tensorflow.clip_by_value", "line_number": 79, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 68, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 68, "usage_type": "name"}]} +{"seq_id": "494923273", "text": "import sys\nimport os\n\nimport math\nimport time\nimport glob\nimport json\nimport numpy as np\n\nimport torch\nimport argparse\n\nimport sys\n\nsys.path.append(\"../dataset/\")\n\nfrom numpy2midi_mix import numpy2midi\nfrom dictionary_mix import genre\nfrom model import CMT\n\nnum_songs = 10\n\n\ndef cal_control_error(err_note_number_list, err_beat_number_list):\n print(\"err_note_number_list\", err_note_number_list)\n print(\"err_beat_number_list\", err_beat_number_list)\n print(\"strength control error\", np.mean(err_note_number_list) / 1.83)\n print(\"density control error\", np.mean(err_beat_number_list) / 10.90)\n\n\ndef generate():\n # path\n parser = argparse.ArgumentParser(description=\"Demo of argparse\")\n parser.add_argument('-c', '--ckpt', default=\"../exp/loss_8_params.pt\")\n parser.add_argument('-f', '--files', required=True)\n parser.add_argument('-g', '--gpus')\n args = parser.parse_args()\n \n if args.gpus is not None:\n if not args.gpus.isnumeric():\n raise RuntimeError('Only 1 GPU is needed for inference')\n os.environ['CUDA_VISIBLE_DEVICES'] = args.gpus\n else:\n os.environ['CUDA_VISIBLE_DEVICES'] = '0'\n\n path_saved_ckpt = args.ckpt\n filelist = glob.glob(args.files)\n # outdir\n\n decoder_n_class = [18, 3, 18, 129, 18, 6, 20, 102, 5025]\n init_n_token = [7, 1, 6]\n\n # log\n\n # init model\n net = torch.nn.DataParallel(CMT(decoder_n_class, init_n_token))\n\n # load model\n print('[*] load model from:', path_saved_ckpt)\n if torch.cuda.is_available():\n net.cuda()\n net.eval()\n net.load_state_dict(torch.load(path_saved_ckpt))\n else:\n net.eval()\n net.load_state_dict(torch.load(path_saved_ckpt, map_location=torch.device('cpu')))\n\n if len(filelist) == 0:\n raise RuntimeError('no npz file in ' + str(filelist))\n\n for file_name in filelist:\n # gen\n start_time = time.time()\n song_time_list = []\n words_len_list = []\n\n cnt_tokens_all = 0\n\n\n sidx = 0\n\n while sidx < num_songs:\n try:\n print(\"new song\")\n start_time = time.time()\n vlog_npz = np.load(file_name)['input']\n \n vlog_npz = vlog_npz[vlog_npz[:, 2] != 1]\n print(vlog_npz)\n\n res, err_note_number_list, err_beat_number_list = net(is_train=False, vlog=vlog_npz, C=0.7)\n \n cal_control_error(err_note_number_list, err_beat_number_list)\n\n numpy2midi(f\"{file_name}_{sidx}\", res[:, [1, 0, 2, 3, 4, 5, 6]].astype(np.int32))\n song_time = time.time() - start_time\n word_len = len(res)\n print('song time:', song_time)\n print('word_len:', word_len)\n words_len_list.append(word_len)\n song_time_list.append(song_time)\n\n sidx += 1\n except KeyboardInterrupt:\n raise ValueError(' [x] terminated.')\n\n\nif __name__ == '__main__':\n print(\"inference\")\n generate()\n \n", "sub_path": "src/gen_midi_conditional.py", "file_name": "gen_midi_conditional.py", "file_ext": "py", "file_size_in_byte": 3076, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "sys.path.append", "line_number": 15, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 28, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 33, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 42, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 44, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.nn.DataParallel", "line_number": 56, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 56, "usage_type": "attribute"}, {"api_name": "model.CMT", "line_number": 56, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 60, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 66, "usage_type": "call"}, {"api_name": "time.time", "line_number": 73, "usage_type": "call"}, {"api_name": "time.time", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy2midi_mix.numpy2midi", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 95, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 96, "usage_type": "call"}]} +{"seq_id": "647376621", "text": "# www.coursera.org/learn/intro-to-deep-learning/ keras-task\n# by help https://machinelearningmastery.com/handwritten-digit-recognition-using-convolutional-neural-networks-python-keras/\n\nfrom keras.datasets import mnist\nfrom matplotlib import pyplot as plt\nfrom keras.models import Sequential\nfrom keras.layers import Dropout, Dense\nfrom keras.utils import np_utils\nimport numpy as np\nfrom tensorflow.python.client import device_lib\n\ndef main():\n (X_train, y_train), (X_test, y_test) = mnist.load_data()\n # plt.imshow(X_train[0], plt.get_cmap('gray'))\n # plt.show()\n\n # print(X_train.shape)\n\n seed = 7\n np.random.seed(seed)\n flat_size = X_train[0].shape[0] * X_train[0].shape[1]\n X_train = X_train.reshape(X_train.shape[0], flat_size).astype('float32')\n X_test = X_test.reshape(X_test.shape[0], flat_size).astype('float32')\n # print(X_train.shape,type(X_train), X_train.dtype)\n # print(X_train[0])\n\n X_train = X_train / 255.0\n X_test = X_test / 255.0\n\n # print(X_train[0])\n\n # print(y_train[0])\n y_train = np_utils.to_categorical(y_train)\n y_test = np_utils.to_categorical(y_test)\n # print(y_train[0])\n\n model = make_baseline_mnist_model()\n\n model.fit(X_train, y_train, validation_data=(X_test, y_test), epochs=10, batch_size=200, verbose=2)\n score = model.evaluate(X_test, y_test)\n print(score)\n\ndef make_baseline_mnist_model():\n n_input = 784\n n_classes = 10\n\n model = Sequential()\n model.add(Dense(n_input, input_dim=n_input, kernel_initializer='normal', activation='relu'))\n model.add(Dense(units=n_classes, kernel_initializer='normal', activation='softmax'))\n\n model.compile(loss='categorical_crossentropy', optimizer='adam', metrics=['accuracy'])\n return model\n\ndef check_evices():\n print(device_lib.list_local_devices())\n\nif __name__ == '__main__':\n main()\n # check_evices()", "sub_path": "KerasTests/Coursera/Keras-task.py", "file_name": "Keras-task.py", "file_ext": "py", "file_size_in_byte": 1877, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "keras.datasets.mnist.load_data", "line_number": 13, "usage_type": "call"}, {"api_name": "keras.datasets.mnist", "line_number": 13, "usage_type": "name"}, {"api_name": "numpy.random.seed", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 20, "usage_type": "attribute"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 33, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 33, "usage_type": "name"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 34, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 34, "usage_type": "name"}, {"api_name": "keras.models.Sequential", "line_number": 47, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 48, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 49, "usage_type": "call"}, {"api_name": "tensorflow.python.client.device_lib.list_local_devices", "line_number": 55, "usage_type": "call"}, {"api_name": "tensorflow.python.client.device_lib", "line_number": 55, "usage_type": "name"}]} +{"seq_id": "580350504", "text": "# coding: utf-8\n\n\"\"\"\n Carbon DLS API\n\n Welcome to the Carbon DLS API docs! You can find all relevant documentation here: https://github.com/carbon3d/carbon3d-api # noqa: E501\n\n The version of the OpenAPI document: 0.0.8\n Contact: api-list@carbon3d.com\n Generated by: https://openapi-generator.tech\n\"\"\"\n\n\nimport pprint\nimport re # noqa: F401\n\nimport six\n\nfrom carbon3d.configuration import Configuration\n\n\nclass PartGenealogyPrintInfo(object):\n \"\"\"NOTE: This class is auto generated by OpenAPI Generator.\n Ref: https://openapi-generator.tech\n\n Do not edit the class manually.\n \"\"\"\n\n \"\"\"\n Attributes:\n openapi_types (dict): The key is attribute name\n and the value is attribute type.\n attribute_map (dict): The key is attribute name\n and the value is json key in definition.\n \"\"\"\n openapi_types = {\n 'queued_at': 'datetime',\n 'print_uuid': 'str',\n 'printer_serial': 'str',\n 'platform_serial': 'str',\n 'cassette_serial': 'str',\n 'resin_lot_number': 'str',\n 'started_at': 'datetime',\n 'finished_at': 'datetime',\n 'error': 'str'\n }\n\n attribute_map = {\n 'queued_at': 'queued_at',\n 'print_uuid': 'print_uuid',\n 'printer_serial': 'printer_serial',\n 'platform_serial': 'platform_serial',\n 'cassette_serial': 'cassette_serial',\n 'resin_lot_number': 'resin_lot_number',\n 'started_at': 'started_at',\n 'finished_at': 'finished_at',\n 'error': 'error'\n }\n\n def __init__(self, queued_at=None, print_uuid=None, printer_serial=None, platform_serial=None, cassette_serial=None, resin_lot_number=None, started_at=None, finished_at=None, error=None, local_vars_configuration=None): # noqa: E501\n \"\"\"PartGenealogyPrintInfo - a model defined in OpenAPI\"\"\" # noqa: E501\n if local_vars_configuration is None:\n local_vars_configuration = Configuration()\n self.local_vars_configuration = local_vars_configuration\n\n self._queued_at = None\n self._print_uuid = None\n self._printer_serial = None\n self._platform_serial = None\n self._cassette_serial = None\n self._resin_lot_number = None\n self._started_at = None\n self._finished_at = None\n self._error = None\n self.discriminator = None\n\n if queued_at is not None:\n self.queued_at = queued_at\n if print_uuid is not None:\n self.print_uuid = print_uuid\n if printer_serial is not None:\n self.printer_serial = printer_serial\n if platform_serial is not None:\n self.platform_serial = platform_serial\n if cassette_serial is not None:\n self.cassette_serial = cassette_serial\n if resin_lot_number is not None:\n self.resin_lot_number = resin_lot_number\n if started_at is not None:\n self.started_at = started_at\n if finished_at is not None:\n self.finished_at = finished_at\n if error is not None:\n self.error = error\n\n @property\n def queued_at(self):\n \"\"\"Gets the queued_at of this PartGenealogyPrintInfo. # noqa: E501\n\n\n :return: The queued_at of this PartGenealogyPrintInfo. # noqa: E501\n :rtype: datetime\n \"\"\"\n return self._queued_at\n\n @queued_at.setter\n def queued_at(self, queued_at):\n \"\"\"Sets the queued_at of this PartGenealogyPrintInfo.\n\n\n :param queued_at: The queued_at of this PartGenealogyPrintInfo. # noqa: E501\n :type: datetime\n \"\"\"\n\n self._queued_at = queued_at\n\n @property\n def print_uuid(self):\n \"\"\"Gets the print_uuid of this PartGenealogyPrintInfo. # noqa: E501\n\n\n :return: The print_uuid of this PartGenealogyPrintInfo. # noqa: E501\n :rtype: str\n \"\"\"\n return self._print_uuid\n\n @print_uuid.setter\n def print_uuid(self, print_uuid):\n \"\"\"Sets the print_uuid of this PartGenealogyPrintInfo.\n\n\n :param print_uuid: The print_uuid of this PartGenealogyPrintInfo. # noqa: E501\n :type: str\n \"\"\"\n\n self._print_uuid = print_uuid\n\n @property\n def printer_serial(self):\n \"\"\"Gets the printer_serial of this PartGenealogyPrintInfo. # noqa: E501\n\n\n :return: The printer_serial of this PartGenealogyPrintInfo. # noqa: E501\n :rtype: str\n \"\"\"\n return self._printer_serial\n\n @printer_serial.setter\n def printer_serial(self, printer_serial):\n \"\"\"Sets the printer_serial of this PartGenealogyPrintInfo.\n\n\n :param printer_serial: The printer_serial of this PartGenealogyPrintInfo. # noqa: E501\n :type: str\n \"\"\"\n\n self._printer_serial = printer_serial\n\n @property\n def platform_serial(self):\n \"\"\"Gets the platform_serial of this PartGenealogyPrintInfo. # noqa: E501\n\n\n :return: The platform_serial of this PartGenealogyPrintInfo. # noqa: E501\n :rtype: str\n \"\"\"\n return self._platform_serial\n\n @platform_serial.setter\n def platform_serial(self, platform_serial):\n \"\"\"Sets the platform_serial of this PartGenealogyPrintInfo.\n\n\n :param platform_serial: The platform_serial of this PartGenealogyPrintInfo. # noqa: E501\n :type: str\n \"\"\"\n\n self._platform_serial = platform_serial\n\n @property\n def cassette_serial(self):\n \"\"\"Gets the cassette_serial of this PartGenealogyPrintInfo. # noqa: E501\n\n\n :return: The cassette_serial of this PartGenealogyPrintInfo. # noqa: E501\n :rtype: str\n \"\"\"\n return self._cassette_serial\n\n @cassette_serial.setter\n def cassette_serial(self, cassette_serial):\n \"\"\"Sets the cassette_serial of this PartGenealogyPrintInfo.\n\n\n :param cassette_serial: The cassette_serial of this PartGenealogyPrintInfo. # noqa: E501\n :type: str\n \"\"\"\n\n self._cassette_serial = cassette_serial\n\n @property\n def resin_lot_number(self):\n \"\"\"Gets the resin_lot_number of this PartGenealogyPrintInfo. # noqa: E501\n\n\n :return: The resin_lot_number of this PartGenealogyPrintInfo. # noqa: E501\n :rtype: str\n \"\"\"\n return self._resin_lot_number\n\n @resin_lot_number.setter\n def resin_lot_number(self, resin_lot_number):\n \"\"\"Sets the resin_lot_number of this PartGenealogyPrintInfo.\n\n\n :param resin_lot_number: The resin_lot_number of this PartGenealogyPrintInfo. # noqa: E501\n :type: str\n \"\"\"\n\n self._resin_lot_number = resin_lot_number\n\n @property\n def started_at(self):\n \"\"\"Gets the started_at of this PartGenealogyPrintInfo. # noqa: E501\n\n\n :return: The started_at of this PartGenealogyPrintInfo. # noqa: E501\n :rtype: datetime\n \"\"\"\n return self._started_at\n\n @started_at.setter\n def started_at(self, started_at):\n \"\"\"Sets the started_at of this PartGenealogyPrintInfo.\n\n\n :param started_at: The started_at of this PartGenealogyPrintInfo. # noqa: E501\n :type: datetime\n \"\"\"\n\n self._started_at = started_at\n\n @property\n def finished_at(self):\n \"\"\"Gets the finished_at of this PartGenealogyPrintInfo. # noqa: E501\n\n\n :return: The finished_at of this PartGenealogyPrintInfo. # noqa: E501\n :rtype: datetime\n \"\"\"\n return self._finished_at\n\n @finished_at.setter\n def finished_at(self, finished_at):\n \"\"\"Sets the finished_at of this PartGenealogyPrintInfo.\n\n\n :param finished_at: The finished_at of this PartGenealogyPrintInfo. # noqa: E501\n :type: datetime\n \"\"\"\n\n self._finished_at = finished_at\n\n @property\n def error(self):\n \"\"\"Gets the error of this PartGenealogyPrintInfo. # noqa: E501\n\n\n :return: The error of this PartGenealogyPrintInfo. # noqa: E501\n :rtype: str\n \"\"\"\n return self._error\n\n @error.setter\n def error(self, error):\n \"\"\"Sets the error of this PartGenealogyPrintInfo.\n\n\n :param error: The error of this PartGenealogyPrintInfo. # noqa: E501\n :type: str\n \"\"\"\n\n self._error = error\n\n def to_dict(self):\n \"\"\"Returns the model properties as a dict\"\"\"\n result = {}\n\n for attr, _ in six.iteritems(self.openapi_types):\n value = getattr(self, attr)\n if isinstance(value, list):\n result[attr] = list(map(\n lambda x: x.to_dict() if hasattr(x, \"to_dict\") else x,\n value\n ))\n elif hasattr(value, \"to_dict\"):\n result[attr] = value.to_dict()\n elif isinstance(value, dict):\n result[attr] = dict(map(\n lambda item: (item[0], item[1].to_dict())\n if hasattr(item[1], \"to_dict\") else item,\n value.items()\n ))\n else:\n result[attr] = value\n\n return result\n\n def to_str(self):\n \"\"\"Returns the string representation of the model\"\"\"\n return pprint.pformat(self.to_dict())\n\n def __repr__(self):\n \"\"\"For `print` and `pprint`\"\"\"\n return self.to_str()\n\n def __eq__(self, other):\n \"\"\"Returns true if both objects are equal\"\"\"\n if not isinstance(other, PartGenealogyPrintInfo):\n return False\n\n return self.to_dict() == other.to_dict()\n\n def __ne__(self, other):\n \"\"\"Returns true if both objects are not equal\"\"\"\n if not isinstance(other, PartGenealogyPrintInfo):\n return True\n\n return self.to_dict() != other.to_dict()\n", "sub_path": "v1/python/carbon3d/models/part_genealogy_print_info.py", "file_name": "part_genealogy_print_info.py", "file_ext": "py", "file_size_in_byte": 9750, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "carbon3d.configuration.Configuration", "line_number": 63, "usage_type": "call"}, {"api_name": "six.iteritems", "line_number": 289, "usage_type": "call"}, {"api_name": "pprint.pformat", "line_number": 311, "usage_type": "call"}]} +{"seq_id": "128042459", "text": "'''\nCopyright 2014 Hewlett-Packard\n\nLicensed under the Apache License, Version 2.0 (the \"License\");\nyou may not use this file except in compliance with the License.\nYou may obtain a copy of the License at\n\n http://www.apache.org/licenses/LICENSE-2.0\n\nUnless required by applicable law or agreed to in writing, software\ndistributed under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for the specific language governing permissions and\nlimitations under the License.\n\nThis product includes cryptographic software written by Eric Young\n(eay@cryptsoft.com). This product includes software written by Tim\nHudson (tjh@cryptsoft.com).\n========================================================================\n\nFreezer Backup modes related functions\n'''\n\nfrom freezer.lvm import lvm_snap, lvm_snap_remove\nfrom freezer.tar import tar_backup, gen_tar_command\nfrom freezer.swift import add_object, manifest_upload\nfrom freezer.utils import gen_manifest_meta, add_host_name_ts_level\n\nfrom multiprocessing import Process, Queue\nimport logging\nimport os\n\n\ndef backup_mode_mysql(backup_opt_dict, time_stamp, manifest_meta_dict):\n '''\n Execute a MySQL DB backup. currently only backup with lvm snapshots\n are supported. This mean, just before the lvm snap vol is created,\n the db tables will be flushed and locked for read, then the lvm create\n command will be executed and after that, the table will be unlocked and\n the backup will be executed. It is important to have the available in\n backup_args.mysql_conf_file the file where the database host, name, user,\n and password are set.\n '''\n try:\n import MySQLdb\n except ImportError as error:\n logging.critical('[*] Error: please install MySQLdb module')\n raise ImportError('[*] Error: please install MySQLdb module')\n\n if not backup_opt_dict.mysql_conf_file:\n logging.critical(\n '[*] MySQL Error: please provide a valid config file')\n raise ValueError\n # Open the file provided in backup_args.mysql_conf_file and extract the\n # db host, name, user and password.\n db_user = db_host = db_pass = False\n with open(backup_opt_dict.mysql_conf_file, 'r') as mysql_file_fd:\n for line in mysql_file_fd:\n if 'host' in line:\n db_host = line.split('=')[1].strip()\n continue\n elif 'user' in line:\n db_user = line.split('=')[1].strip()\n continue\n elif 'password' in line:\n db_pass = line.split('=')[1].strip()\n continue\n\n # Initialize the DB object and connect to the db according to\n # the db mysql backup file config\n try:\n backup_opt_dict.mysql_db_inst = MySQLdb.connect(\n host=db_host, user=db_user, passwd=db_pass)\n except Exception as error:\n logging.critical('[*] MySQL Error: {0}'.format(error))\n raise Exception\n\n # Execute LVM backup\n backup_mode_fs(backup_opt_dict, time_stamp, manifest_meta_dict)\n\n\ndef backup_mode_mongo(backup_opt_dict, time_stamp, manifest_meta_dict):\n '''\n Execute the necessary tasks for file system backup mode\n '''\n try:\n from pymongo import MongoClient\n except ImportError:\n logging.critical('[*] Error: please install pymongo module')\n raise ImportError('[*] Error: please install pymongo module')\n\n logging.info('[*] MongoDB backup is being executed...')\n logging.info('[*] Checking is the localhost is Master/Primary...')\n mongodb_port = '27017'\n local_hostname = backup_opt_dict.hostname\n db_host_port = '{0}:{1}'.format(local_hostname, mongodb_port)\n mongo_client = MongoClient(db_host_port)\n master_dict = dict(mongo_client.admin.command(\"isMaster\"))\n mongo_me = master_dict['me']\n mongo_primary = master_dict['primary']\n\n if mongo_me == mongo_primary:\n backup_mode_fs(backup_opt_dict, time_stamp, manifest_meta_dict)\n else:\n logging.warning('[*] localhost {0} is not Master/Primary,\\\n exiting...'.format(local_hostname))\n return True\n\n\ndef backup_mode_fs(backup_opt_dict, time_stamp, manifest_meta_dict):\n '''\n Execute the necessary tasks for file system backup mode\n '''\n\n logging.info('[*] File System backup is being executed...')\n lvm_snap(backup_opt_dict)\n # Extract some values from arguments that will be used later on\n # Initialize swift client object, generate container segments name\n # and extract backup name\n sw_connector = backup_opt_dict.sw_connector\n\n # Execute a tar gzip of the specified directory and return\n # small chunks (default 128MB), timestamp, backup, filename,\n # file chunk index and the tar meta-data file\n\n # Generate a string hostname, backup name, timestamp and backup level\n file_name = add_host_name_ts_level(backup_opt_dict, time_stamp)\n meta_data_backup_file = u'tar_metadata_{0}'.format(file_name)\n\n (backup_opt_dict, tar_command, manifest_meta_dict) = gen_tar_command(\n opt_dict=backup_opt_dict, time_stamp=time_stamp,\n remote_manifest_meta=manifest_meta_dict)\n # Initialize a Queue for a maximum of 2 items\n tar_backup_queue = Queue(maxsize=2)\n tar_backup_stream = Process(\n target=tar_backup, args=(\n backup_opt_dict, tar_command, tar_backup_queue,))\n tar_backup_stream.daemon = True\n tar_backup_stream.start()\n\n add_object_stream = Process(\n target=add_object, args=(\n backup_opt_dict, tar_backup_queue, file_name, time_stamp))\n add_object_stream.daemon = True\n add_object_stream.start()\n\n tar_backup_stream.join()\n tar_backup_queue.put(\n ({False : False}))\n tar_backup_queue.close()\n add_object_stream.join()\n\n (backup_opt_dict, manifest_meta_dict, tar_meta_to_upload,\n tar_meta_prev) = gen_manifest_meta(\n backup_opt_dict, manifest_meta_dict, meta_data_backup_file)\n\n manifest_file = u''\n meta_data_abs_path = '{0}/{1}'.format(\n backup_opt_dict.workdir, tar_meta_prev)\n # Upload swift manifest for segments\n if backup_opt_dict.upload:\n if not backup_opt_dict.no_incremental:\n # Upload tar incremental meta data file and remove it\n logging.info('[*] Uploading tar meta data file: {0}'.format(\n tar_meta_to_upload))\n with open(meta_data_abs_path, 'r') as meta_fd:\n sw_connector.put_object(\n backup_opt_dict.container, tar_meta_to_upload, meta_fd)\n # Removing tar meta data file, so we have only one authoritative\n # version on swift\n logging.info('[*] Removing tar meta data file: {0}'.format(\n meta_data_abs_path))\n os.remove(meta_data_abs_path)\n # Upload manifest to swift\n manifest_upload(\n manifest_file, backup_opt_dict, file_name, manifest_meta_dict)\n\n # Unmount and remove lvm snapshot volume\n lvm_snap_remove(backup_opt_dict)\n", "sub_path": "freezer/backup.py", "file_name": "backup.py", "file_ext": "py", "file_size_in_byte": 7065, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "logging.critical", "line_number": 47, "usage_type": "call"}, {"api_name": "logging.critical", "line_number": 51, "usage_type": "call"}, {"api_name": "MySQLdb.connect", "line_number": 72, "usage_type": "call"}, {"api_name": "logging.critical", "line_number": 75, "usage_type": "call"}, {"api_name": "logging.critical", "line_number": 89, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 92, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 93, "usage_type": "call"}, {"api_name": "pymongo.MongoClient", "line_number": 97, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 105, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 115, "usage_type": "call"}, {"api_name": "freezer.lvm.lvm_snap", "line_number": 116, "usage_type": "call"}, {"api_name": "freezer.utils.add_host_name_ts_level", "line_number": 127, "usage_type": "call"}, {"api_name": "freezer.tar.gen_tar_command", "line_number": 130, "usage_type": "call"}, {"api_name": "multiprocessing.Queue", "line_number": 134, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 135, "usage_type": "call"}, {"api_name": "freezer.tar.tar_backup", "line_number": 136, "usage_type": "name"}, {"api_name": "multiprocessing.Process", "line_number": 141, "usage_type": "call"}, {"api_name": "freezer.swift.add_object", "line_number": 142, "usage_type": "name"}, {"api_name": "freezer.utils.gen_manifest_meta", "line_number": 154, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 164, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 171, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 173, "usage_type": "call"}, {"api_name": "freezer.swift.manifest_upload", "line_number": 175, "usage_type": "call"}, {"api_name": "freezer.lvm.lvm_snap_remove", "line_number": 179, "usage_type": "call"}]} +{"seq_id": "151759114", "text": "#!/usr/bin/env python\n\nfrom __future__ import print_function\nfrom builtins import str\nfrom builtins import range\nimport os\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom astropy.coordinates import SkyCoord\nimport astropy.units as u\nfrom surveys_db import SurveysDB\nfontsize=16 # adjust to taste\nfrom matplotlib import rc\nrc('font',**{'family':'serif','serif':['Times'],'size':fontsize})\nrc('text', usetex=True)\nimport sys\nimport datetime\n\norg=180\n\ndef cc(ra,dec):\n ra=np.array(ra)\n\n x = np.remainder(ra+360-org,360) # shift RA values\n ind = x>180\n x[ind] -=360 # scale conversion to [-180, 180]\n x=-x # reverse the scale: East to the left\n\n return np.radians(x),np.radians(dec)\n\ndef plot_select(r,sf,label,**kwargs):\n\n ra=[]\n dec=[]\n r_in=[]\n r_out=[]\n for f in r:\n if sf(f):\n ra.append(f['ra'])\n dec.append(f['decl'])\n r_in.append(f)\n else:\n r_out.append(f)\n\n ra_r,dec_r=cc(ra,dec)\n plt.scatter(ra_r,dec_r,label=label,**kwargs)\n print(\"%-20s : %i\" % (label,len(r_in)))\n return r_in,r_out\n \nwith SurveysDB(readonly=True) as sdb:\n sdb.cur.execute('select * from fields where dr2')\n #sdb.cur.execute('select * from fields where status!=\"Not started\"')\n results=sdb.cur.fetchall()\n \nprint(len(results),'fields are in dr2')\n \nfig = plt.figure(figsize=(16, 8))\nfig.add_subplot(111, projection='aitoff')\n\n# GP\n\nfor b in [-10,0,10]:\n\n lon=np.linspace(-180,180,1000)\n lat=b*np.ones_like(lon)\n\n sc=SkyCoord(l=lon,b=lat,unit=(u.deg,u.deg),frame='galactic')\n\n ra=np.array(sc.icrs.ra)\n dec=np.array(sc.icrs.dec)\n\n ra_r,dec_r=cc(ra,dec)\n \n plt.scatter(ra_r,dec_r,color='blue',s=5,label='MW' if b==0 else None)\n\n#DR2 area\nravals = []\ndecvals = []\nfor r in results:\n ravals.append(r['ra'])\n decvals.append(r['decl'])\n\nra_r,dec_r=cc(ravals,decvals)\nplt.scatter(ra_r,dec_r,marker='o',color='blue',alpha=0.2,zorder=-5,edgecolors='none',s=50,label='DR2')\n\n_,r=plot_select(results,lambda r:r['dr1']==1,label='Complete (DR1)',color='magenta')\n_,r=plot_select(r,lambda r:r['lgz']==1,label='Complete (internal)',color='orange')\n_,r=plot_select(r,lambda r:r['gz_status']=='Paused',label='Paused',color='red')\n_,r=plot_select(r,lambda r:r['gz_status']=='Complete',label='Complete',color='green')\n_,r=plot_select(r,lambda r:r['gz_status'] in ['In progress'],label='In progress',color='cyan')\n_,r=plot_select(r,lambda r:r['gz_status'] in ['Created','Downloading'],label='Ready to upload',color='yellow')\n#_,r=plot_select(r,lambda r:r['gz_status'] in ['Failed'],label='Failed',color='red')\n\n\nax=plt.gca()\n\ntick_labels = np.array([150, 120, 90, 60, 30, 0, 330, 300, 270, 240, 210])\ntick_labels = np.remainder(tick_labels+360+org,360)\ntick_labels = list(tick_labels)\nfor i in range(0,len(tick_labels)):\n tick_labels[i] = ''+str(tick_labels[i])+'\\ \\ \\ \\ \\ \\ \\ \\ \\ \\ \\ \\ \\ \\ \\ \\ \\ \\ \\ \\,'\n\nax.set_xticklabels(tick_labels,verticalalignment='top',rotation='vertical')\n\nplt.xlabel('R.A.')\nplt.ylabel('Decl.')\nplt.grid(True)\nplt.legend(loc=4)\nplt.tight_layout()\nplt.title('RGZ status at '+datetime.datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\"),loc='right')\nif len(sys.argv)>1:\n plt.savefig(sys.argv[1],dpi=250)\nelse:\n plt.show()\n", "sub_path": "plot_db_rgz.py", "file_name": "plot_db_rgz.py", "file_ext": "py", "file_size_in_byte": 3271, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "matplotlib.rc", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.remainder", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.radians", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "surveys_db.SurveysDB", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.ones_like", "line_number": 65, "usage_type": "call"}, {"api_name": "astropy.coordinates.SkyCoord", "line_number": 67, "usage_type": "call"}, {"api_name": "astropy.units.deg", "line_number": 67, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 67, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.remainder", "line_number": 98, "usage_type": "call"}, {"api_name": "builtins.range", "line_number": 100, "usage_type": "call"}, {"api_name": "builtins.str", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 106, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 110, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 110, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 110, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 111, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 112, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}]} +{"seq_id": "180182086", "text": "from mock import Mock, patch\nfrom nose.tools import assert_raises\nfrom unittest import TestCase\n\nfrom stretch.models import LoadBalancer\n\n\nclass TestLoadBalancer(TestCase):\n def setUp(self):\n self.client = Mock()\n patcher = patch('stretch.agent.supervisors.endpoint_supervisor_client',\n return_value=self.client)\n patcher.start()\n self.addCleanup(patcher.stop)\n\n for attr in ('group', 'save'):\n self.patch_lb(attr, mock=True)\n\n self.group = Mock()\n\n self.lb = self.create_lb()\n\n def patch_lb(self, attr, mock=False):\n patcher = patch('stretch.models.LoadBalancer.%s' % attr)\n if mock:\n mock_obj = patcher.start()\n self.addCleanup(patcher.stop)\n return mock_obj\n else:\n return patcher\n\n def create_lb(self):\n with self.patch_lb('backend') as backend:\n backend.create_lb.return_value = ('2.2.2.2', 443)\n return LoadBalancer.create(self.group, 'p', 'http', {'k': 'v'})\n\n def test_create(self):\n lb = self.create_lb()\n\n lb.group.environment.system.config_manager.set.assert_called_with(\n lb.config_key, '{\"host\": \"2.2.2.2\", \"port\": 443}')\n self.client.add_group.assert_called_with(lb.group.pk,\n lb.group.config_key)\n lb.save.assert_called_with()\n\n def test_add_endpoint(self):\n backend = self.patch_lb('backend', mock=True)\n with self.patch_lb('_apply_endpoint'):\n endpoint = {'host': '1.1.1.1', 'ports': {'http': 80}}\n self.lb.add_endpoint(endpoint)\n self.lb._apply_endpoint.assert_called_with(backend.lb_add_endpoint,\n endpoint)\n\n def test_remove_endpoint(self):\n backend = self.patch_lb('backend', mock=True)\n with self.patch_lb('_apply_endpoint'):\n endpoint = {'host': '1.1.1.1', 'ports': {'http': 80}}\n self.lb.remove_endpoint(endpoint)\n self.lb._apply_endpoint.assert_called_with(\n backend.lb_remove_endpoint, endpoint)\n\n def test__apply_endpoint(self):\n endpoint = {'host': '1.1.1.1', 'ports': {'http': 80}}\n func = Mock()\n\n self.lb.port_name = 'foo'\n self.lb._apply_endpoint(func, endpoint)\n\n assert not func.called\n\n self.lb.port_name = 'http'\n self.lb._apply_endpoint(func, endpoint)\n\n func.assert_called_with(self.lb, '1.1.1.1', 80)\n\n def test_backend(self):\n self.lb.group.environment.backend = None\n with assert_raises(Exception):\n self.lb.backend\n self.lb.group.environment.backend = 'backend'\n self.assertEquals(self.lb.backend, 'backend')\n\n def test_config_key(self):\n config = self.lb.group.environment.system.config_manager\n config.get_lb_key.return_value = 'key'\n self.assertEquals(self.lb.config_key, 'key')\n config.get_lb_key.assert_called_with(self.lb)\n\n def test_pre_delete(self):\n LoadBalancer.pre_delete(Mock(), self.lb)\n self.client.remove_group.assert_called_with(self.lb.group.pk)\n config = self.lb.group.environment.system.config_manager\n config.delete.assert_called_with(self.lb.config_key)\n self.lb.backend.delete_lb.assert_called_with(self.lb)\n", "sub_path": "tests/stretch/models/test_loadbalancer.py", "file_name": "test_loadbalancer.py", "file_ext": "py", "file_size_in_byte": 3338, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "unittest.TestCase", "line_number": 8, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 10, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 11, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 19, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 24, "usage_type": "call"}, {"api_name": "stretch.models.LoadBalancer.create", "line_number": 35, "usage_type": "call"}, {"api_name": "stretch.models.LoadBalancer", "line_number": 35, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 64, "usage_type": "call"}, {"api_name": "nose.tools.assert_raises", "line_number": 78, "usage_type": "call"}, {"api_name": "stretch.models.LoadBalancer.pre_delete", "line_number": 90, "usage_type": "call"}, {"api_name": "stretch.models.LoadBalancer", "line_number": 90, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 90, "usage_type": "call"}]} +{"seq_id": "462433667", "text": "import numpy as np\r\nimport cv2\r\n\r\ncap = cv2.VideoCapture(0)\r\nimg_counter = 0\r\n\r\ndef change_res(width, height):\r\n cap.set(3, width)\r\n cap.set(4, height)\r\n\r\ndef rescale_frame(frame, percent=75):\r\n width = int(frame.shape[1] * percent/ 100)\r\n height = int(frame.shape[0] * percent/ 100)\r\n dim = (width, height)\r\n return cv2.resize(frame, dim, interpolation =cv2.INTER_AREA)\r\n\r\n# 720p\r\nchange_res(1280, 720)\r\n\r\n\r\nwhile(True):\r\n ret, frame = cap.read()\r\n k = cv2.waitKey(20)\r\n\r\n gray = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)\r\n frame75 = rescale_frame(frame, percent=10)\r\n\r\n cv2.imshow('frame', gray)\r\n\r\n if k & 0xFF == ord('q'):\r\n # Q to quit\r\n break\r\n\r\n elif k%256 == 32:\r\n # SPACE to take picture\r\n img_name = \"opencv_frame_{}.png\".format(img_counter)\r\n cv2.imwrite(img_name, frame)\r\n print(\"{} written!\".format(img_name))\r\n img_counter += 1\r\n \r\ncap.release()\r\ncv2.destroyAllWindows()", "sub_path": "OpenCV_test.py", "file_name": "OpenCV_test.py", "file_ext": "py", "file_size_in_byte": 976, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "cv2.VideoCapture", "line_number": 4, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.INTER_AREA", "line_number": 15, "usage_type": "attribute"}, {"api_name": "cv2.waitKey", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 25, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 37, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 42, "usage_type": "call"}]} +{"seq_id": "423418169", "text": "# -*- coding: utf-8 -*-\n#\n# Copyright (c) 2016, ParaTools, Inc.\n# All rights reserved.\n#\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted provided that the following conditions are met:\n# (1) Redistributions of source code must retain the above copyright notice,\n# this list of conditions and the following disclaimer.\n# (2) Redistributions in binary form must reproduce the above copyright notice,\n# this list of conditions and the following disclaimer in the documentation\n# and/or other materials provided with the distribution.\n# (3) Neither the name of ParaTools, Inc. nor the names of its contributors may\n# be used to endorse or promote products derived from this software without\n# specific prior written permission.\n#\n# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\"\n# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE\n# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE\n# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE\n# FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL\n# DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR\n# SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER\n# CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,\n# OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE\n# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.\n#\n\"\"\"Test functions.\n\nFunctions used for unit tests of error.py.\n\"\"\"\n\nimport os\nimport sys\nfrom pylint import epylint\nfrom tau import TAU_HOME\nfrom tau import tests\n\n\nclass PylintTest(tests.TestCase):\n \"\"\"Runs Pylint to make sure the code scores at least 9.0\"\"\"\n \n def run_pylint(self, *args):\n pylint_args = '--rcfile=' + os.path.join(TAU_HOME, \"pylintrc\") + ' ' + ' '.join(args) \n fstdout, fstderr = epylint.py_run(pylint_args, return_std=True, script='pylint')\n stdout = ''.join(line for line in fstdout)\n stderr = ''.join(line for line in fstderr)\n return stdout, stderr\n \n def test_pylint_version(self):\n stdout, stderr = self.run_pylint('--version')\n sys.stdout.write(stdout)\n sys.stderr.write(stderr)\n self.assertFalse(stderr)\n version_parts = stdout.split(',')[0].split('pylint ')[1].split('.')\n version = tuple(int(x) for x in version_parts)\n self.assertGreaterEqual(version, (1, 5, 2), \"Pylint version %s is too old!\" % str(version))\n \n def test_pylint(self):\n stdout, stderr = self.run_pylint(os.path.join(TAU_HOME, \"packages\", \"tau\"))\n sys.stdout.write(stdout)\n sys.stderr.write(stderr)\n self.assertFalse(stderr)\n self.assertIn('Your code has been rated at', stdout)\n score = float(stdout.split('Your code has been rated at')[1].split('/10')[0])\n self.assertGreaterEqual(score, 9.0, \"Pylint score %s/10 is too low!\" % score)\n\n", "sub_path": "packages/tau/tests/test_pylint.py", "file_name": "test_pylint.py", "file_ext": "py", "file_size_in_byte": 3068, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "tau.tests.TestCase", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tau.tests", "line_number": 40, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "tau.TAU_HOME", "line_number": 44, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "pylint.epylint.py_run", "line_number": 45, "usage_type": "call"}, {"api_name": "pylint.epylint", "line_number": 45, "usage_type": "name"}, {"api_name": "sys.stdout.write", "line_number": 52, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 52, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 53, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 60, "usage_type": "call"}, {"api_name": "tau.TAU_HOME", "line_number": 60, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 61, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 61, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 62, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 62, "usage_type": "attribute"}]} +{"seq_id": "464626994", "text": "\"\"\"\n.. module:: extractor\n :synopsis: Embedding extractor module\n\n.. moduleauthor:: Justin Shenk \n\n\n\"\"\"\n\nimport logging\nimport os\nfrom typing import Union, Optional, Tuple\nimport warnings\n\nimport closely\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom PIL import Image\nfrom sklearn.decomposition import PCA\nimport torch\nimport torch.nn as nn\nimport torchvision\nimport torchvision.transforms as transforms\nimport torchvision.transforms.functional as TF\nimport torch.utils.data as utils\n\nfrom .dataset import PILDataset, ImageFolder\nfrom .models import BasicAutoencoder, UnNormalize\n\nwarnings.filterwarnings(\"ignore\", message=\"Palette images with Transparency\")\nlog = logging.getLogger(__name__)\n\n\nclass EmbeddingExtractor:\n \"\"\"Extract embeddings from data with models and allow visualization.\n\n Attributes:\n trainloader (torch loader)\n evalloader (torch loader)\n model (torch.nn.Module)\n embeddings (np.ndarray)\n\n \"\"\"\n\n def __init__(\n self,\n input: Union[str, np.ndarray],\n num_channels: int = 3,\n num_epochs: int = 2,\n batch_size: int = 32,\n show: bool = False,\n plot_embeddings=False,\n show_path: bool = False,\n show_train: bool = False,\n z_dim: int = 8,\n metric: str = \"cosine\",\n model: Optional[torch.nn.Module] = None,\n db: Optional = None,\n embeddings_path: Optional[str] = False,\n save_embeddings: Optional[bool] = False\n ):\n \"\"\"Inits EmbeddingExtractor with input, either `str` or `np.ndarray`, performs training and validation.\n\n Args:\n input (np.ndarray or str): data\n num_channels (int): grayscale = 1, color = 3\n num_epochs (int): more is better (generally)\n batch_size (int): number of images per batch\n show (bool): show closest pairs\n show_path (bool): show path of duplicates\n show_train (bool): show intermediate training results\n z_dim (int): compression size\n metric (str): distance metric for :meth:`scipy.spatial.distance.cdist` (eg, euclidean, cosine, hamming, etc.)\n model (torch.nn.Module, optional): class implementing same methods as :class:`~simages.BasicAutoencoder`\n db_conn_string (str): Mongodb connection string\n embeddings_path (str): path to load embeddings\n save_embeddings (str): saves embeddings in current directory\n\n \"\"\"\n self.num_epochs = num_epochs\n self._batch_size = batch_size\n self._show = show\n self._show_path = show_path\n self._show_train = show_train\n self._db = db\n\n self._device = torch.device(\"cuda:0\" if torch.cuda.is_available() else \"cpu\")\n self._num_channels = num_channels\n\n self._metric = metric\n self._z_dim = z_dim\n self._hw = 48\n\n self._mean = [0.5] * self._num_channels\n self._std = [0.25] * self._num_channels\n\n train_transforms = transforms.Compose(\n [\n transforms.RandomResizedCrop(self._hw),\n transforms.RandomHorizontalFlip(),\n transforms.ToTensor(),\n transforms.Normalize(self._mean, self._std),\n ]\n )\n basic_transforms = transforms.Compose(\n [\n transforms.Resize(self._hw),\n transforms.CenterCrop(self._hw),\n transforms.ToTensor(),\n transforms.Normalize(self._mean, self._std),\n ]\n )\n\n def is_valid(path):\n img_extensions = [\n \".jpg\",\n \".jpeg\",\n \".png\",\n \".ppm\",\n \".bmp\",\n \".pgm\",\n \".tif\",\n \".gif\",\n \".octet-stream\",\n ]\n\n _, file_extension = os.path.splitext(path)\n valid_ext = file_extension.lower() in img_extensions\n if not valid_ext:\n return False\n try:\n Image.open(path).verify()\n except Exception as e:\n log.info(f\"Skipping {os.path.basename(path)}: {e}\")\n return False\n return True\n\n if isinstance(input, str):\n data_dir = os.path.abspath(input)\n self.root = data_dir\n self.train_dataset = ImageFolder(\n data_dir, transform=train_transforms, is_valid_file=is_valid\n )\n self.eval_dataset = ImageFolder(\n data_dir, transform=basic_transforms, is_valid_file=is_valid\n )\n self.trainloader = torch.utils.data.DataLoader(\n self.train_dataset, batch_size=batch_size, shuffle=True, num_workers=4\n )\n self.evalloader = torch.utils.data.DataLoader(\n self.eval_dataset, batch_size=batch_size, shuffle=False, num_workers=4\n )\n elif isinstance(input, np.ndarray):\n self.trainloader = self._tensor_dataloader(\n input, train_transforms, shuffle=True\n )\n self.evalloader = self._tensor_dataloader(\n input, basic_transforms, shuffle=False\n )\n\n if not torch.cuda.is_available():\n log.info(\n \"Note: No GPU found, using CPU. Performance is improved on a CUDA-device.\"\n )\n\n if model is not None:\n self.model = model\n else:\n self.model = BasicAutoencoder(num_channels=num_channels, z_dim=z_dim)\n\n if torch.cuda.device_count() > 1:\n log.info(\"Let's use\", torch.cuda.device_count(), \"GPUs!\")\n model = nn.DataParallel(self.model)\n\n self.model.to(self._device)\n self._distance = nn.MSELoss()\n self._optimizer = torch.optim.Adam(self.model.parameters(), weight_decay=1e-5)\n\n if embeddings_path:\n log.info(\"skipping training, loading embeddings from file\")\n self.embeddings = np.load(embeddings_path)\n log.info(f\"Embeddings shape {self.embeddings.shape}\")\n else:\n embeddings_path = f'embeddings_epochs-{num_epochs}.npy'\n self.train()\n self.eval()\n if save_embeddings:\n log.info(f\"saving embeddings to {embeddings_path}\")\n np.save(embeddings_path, self.embeddings)\n\n\n\n def _truncate_middle(self, string: str, n: int) -> str:\n if len(string) <= n:\n # string is already short-enough\n return string\n # half of the size, minus the 3 .'s\n n_2 = int(int(n) / 2 - 3)\n # whatever's left\n n_1 = int(n - n_2 - 3)\n return f\"{string[:n_1]}...{string[-n_2:]}\"\n\n def get_image(self, index: int) -> torch.Tensor:\n result = self.evalloader.dataset[index]\n if isinstance(result, tuple):\n return result[0].cpu()\n else:\n return result.cpu()\n\n def _tensor_dataloader(\n self,\n array: np.ndarray,\n transforms: torchvision.transforms.Compose,\n shuffle: bool = True,\n ) -> utils.DataLoader:\n log.debug(f\"INFO: data shape: {array.shape} (Target: N x C x H x W)\")\n if array.ndim == 3:\n log.debug(\n f\"Converting to grayscale dataset of dims {array.shape[0]} x 1 x {array.shape[1]} x {array.shape[2]}\"\n )\n array = array[:, np.newaxis, ...]\n log.debug(f\"New shape: {array.shape}\")\n\n tensor = torch.Tensor(array)\n pil_list = [TF.to_pil_image(array.squeeze()) for array in tensor]\n dataset = PILDataset(pil_list, transform=transforms)\n dataloader = utils.DataLoader(\n dataset, batch_size=self._batch_size, shuffle=shuffle\n )\n return dataloader\n\n def train(self):\n \"\"\"Train autoencoder to build embeddings of dataset. Final embeddings are created in\n :meth:`~simages.extractor.EmbeddingExtractor.eval`.\n\n \"\"\"\n log.info(\n f\"Building embeddings for {len(self.evalloader.dataset)} images. This may take some time...\"\n )\n\n for epoch in range(self.num_epochs):\n for data in self.trainloader:\n if isinstance(data, list):\n data = data[0]\n img = data.to(self._device)\n # ===================forward=====================\n output, embedding = self.model(img)\n\n loss = self._distance(output, img)\n # ===================backward====================\n self._optimizer.zero_grad()\n loss.backward()\n self._optimizer.step()\n\n if self._show_train:\n try:\n img_array = img.cpu()[0]\n output_array = output.detach().cpu()[0]\n\n grid_img = torchvision.utils.make_grid(\n [img_array, output_array], nrow=1\n )\n self.show(\n grid_img,\n title=f\"Building embeddings: epoch [{epoch+1}/{self.num_epochs}]\",\n block=False,\n y_labels=[(2, \"Original\"), (5, \"Reconstruction\")],\n )\n except Exception as e:\n log.error(f\"{e}\")\n\n # ===================log========================\n log.info(\n \"epoch [{}/{}], loss:{:.4f}\".format(epoch + 1, self.num_epochs, loss)\n )\n\n def eval(self):\n \"\"\"Evaluate reconstruction of embeddings built in `train`.\"\"\"\n embeddings = []\n imgs = []\n\n # Change model to `eval` mode so weights are frozen\n self.model.eval()\n\n for data in self.evalloader:\n if isinstance(data, list):\n data = data[0]\n img = data.to(self._device)\n imgs.append(img)\n # ===================forward=====================\n output, embedding = self.model(img)\n embeddings.append(embedding)\n\n loss = self._distance(output, img)\n # ===================backward====================\n self._optimizer.zero_grad()\n loss.backward()\n self._optimizer.step()\n\n if self._show_train:\n try:\n img_array = img.cpu()[0]\n output_array = output.detach().cpu()[0]\n\n grid_img = torchvision.utils.make_grid(\n [img_array, output_array], nrow=1\n )\n self.show(\n grid_img,\n title=f\"Reconstruction\",\n y_labels=[(2, \"Original\"), (5, \"Reconstruction\")],\n )\n except Exception as e:\n log.error(f\"{e}\")\n\n # ===================log========================\n log.info(\"eval, loss:{:.4f}\".format(loss))\n\n self.embeddings = torch.cat(embeddings).detach().cpu().numpy()\n\n def duplicates(\n self, n: int = 10, quantile: float = None\n ) -> Tuple[np.ndarray, np.ndarray]:\n \"\"\"Identify `n` closest pairs of images, or quantile (for example, closest 0.05).\n\n Args:\n n (int): number of pairs\n quantile (float): quantile of total combination (suggested range: 0.001 - 0.01)\n \"\"\"\n nr_embeddings = len(self.embeddings)\n min_pairs = (nr_embeddings * (nr_embeddings - 1)) // 2\n\n n = min(n, min_pairs)\n\n if quantile is not None:\n pairs, distances = closely.solve(\n self.embeddings, quantile=quantile, metric=self._metric\n )\n else:\n pairs, distances = closely.solve(self.embeddings, n=n, metric=self._metric)\n\n return pairs, distances\n\n @staticmethod\n def channels_last(img: np.ndarray) -> np.ndarray:\n \"\"\"Move channels from first to last by swapping axes.\"\"\"\n img_t = np.transpose(img, (1, 2, 0))\n return img_t\n\n def show(\n self,\n img: Union[torch.Tensor, np.ndarray],\n title: str = \"\",\n block: bool = True,\n y_labels=None,\n unnormalize=True,\n ):\n \"\"\"Plot `img` with `title`.\n\n Args:\n img (torch.Tensor or np.ndarray): Image to plot\n title (str): plot title\n block (bool): block matplotlib plot until window closed\n \"\"\"\n if unnormalize:\n img = self.unnormalize(img)\n if isinstance(img, torch.Tensor):\n npimg = img.detach().numpy()\n elif isinstance(img, np.ndarray):\n pass\n else:\n raise NotImplementedError(f\"{type(img)}\")\n\n if img.shape[0] in [1, 2, 3]:\n npimg = self.channels_last(npimg).squeeze()\n\n fig, ax = plt.subplots(1, 1)\n plt.title(f\"{title}\")\n ax.imshow(npimg, interpolation=\"nearest\")\n if y_labels is not None:\n labels = [item.get_text() for item in ax.get_xticklabels()]\n for idx, label in y_labels:\n labels[idx] = label\n ax.set_yticklabels(labels)\n\n plt.show(block=block)\n\n def show_images(self, indices: Union[list, int], title=\"\"):\n \"\"\"Plot images (from validation data) at `indices` with `title`\"\"\"\n if isinstance(indices, int):\n indices = [indices]\n tensors = [self.get_image(idx) for idx in indices]\n self.show(torchvision.utils.make_grid(tensors), title=title)\n\n\n def color_embeddings(self, image_paths, path_colors='train-blue_val-green'):\n \"\"\"Color embeddings by path, eg, 'train' or 'val'\"\"\"\n colors = []\n path_color_mapping = {}\n for path_color in path_colors.split('_'):\n path, color = path_color.split('-')\n path_color_mapping[path] = color\n\n # Iterate over each image path and assign a color based on the presence of 'train' or 'val'\n for path in image_paths:\n color = 'gray'\n for directory, directory_color in path_color_mapping.items():\n if directory in path:\n color = directory_color\n break\n\n colors.append(color)\n return colors\n\n def plot_embeddings(self, title=\"\", path_colors=None):\n \"\"\"Plot embeddings (from validation data), hover to see image using bokeh\"\"\"\n from bokeh.models import ColumnDataSource, HoverTool\n from bokeh.plotting import figure, show, output_file\n\n if os.name == 'nt': # workaround for permissions issue\n output_file('simages-show-script.html', title=title)\n\n embeddings = self.embeddings\n validation_image_paths = self.evalloader.dataset.samples\n\n colors = ['gray'] * len(validation_image_paths)\n if path_colors:\n colors = self.color_embeddings(validation_image_paths, path_colors=path_colors)\n\n # Get PCA of embeddings\n pca = PCA(n_components=2)\n pca.fit(embeddings)\n embeddings = pca.transform(embeddings)\n\n # Create scatter plot of embeddings\n source = ColumnDataSource(\n data=dict(\n x=embeddings[:, 0],\n y=embeddings[:, 1],\n image_path=validation_image_paths,\n image_path_relative=[os.path.relpath(p, self.root) for p in validation_image_paths],\n colors=colors\n )\n )\n hover = HoverTool(\n tooltips=\"\"\"\n
\n
\n \"@image_path\"\n @image_path_relative\n
\n
\n \"\"\"\n )\n p = figure(\n width=800,\n height=800,\n title=title,\n tools=[hover, \"pan\", \"wheel_zoom\", \"box_zoom\", \"reset\", \"save\"],\n )\n p.circle(\"x\", \"y\", size=5, source=source, fill_alpha=0.8, line_color='colors', fill_color='colors')\n show(p)\n\n\n def get_embedding(self, index: int) -> torch.Tensor:\n \"\"\"Get embedding at `index` of eval/embedding\"\"\"\n return torch.Tensor(self.embeddings[index])\n\n def image_paths(self, indices, short=True):\n \"\"\"Get path to image at `index` of eval/embedding\n\n Args:\n indices Union[int,list]: indices of embeddings in dataset\n short (bool): truncate filepath to 30 charachters\n\n Returns:\n paths (str or list of str): paths to images in image folder\n\n \"\"\"\n if isinstance(indices, (int, np.int_, np.int64)): # <-- add np.int64 here\n indices = [indices]\n\n paths = []\n for index in indices:\n path = self.evalloader.dataset.samples[index]\n if short:\n path = self._truncate_middle(os.path.basename(path), 30)\n paths.append(path)\n\n if len(paths) == 1:\n return paths[0] # backward compatibility\n return paths\n\n @torch.no_grad()\n def show_duplicates(self, n=5, path=False) -> (np.ndarray, np.ndarray):\n \"\"\"Show duplicates from comparison of embeddings. Uses `closely` package to get pairs.\n\n Args:\n n (int): how many closest pairs to identify\n path (bool): Plot pairs of images with abbreviated paths\n\n Returns:\n pairs (np.ndarray): pairs as indices\n distances (np.ndarray): distances of pairs\n\n \"\"\"\n show_path = path or self._show_path\n pairs, distances = self.duplicates(n=n)\n\n # Plot pairs\n for idx, pair in enumerate(pairs):\n img0 = self.get_image(pair[0])\n img1 = self.get_image(pair[1])\n img0_reconst = self.decode(index=pair[0], astensor=True)[0]\n img1_reconst = self.decode(index=pair[1], astensor=True)[0]\n pair_details = (\n f\"{self.image_paths(pair[0])}\\n{self.image_paths(pair[1])}\"\n if show_path\n else pair\n )\n title = f\"{pair_details}, dist={distances[idx]:.2f}\"\n self.show(\n torchvision.utils.make_grid(\n [img0, img1, img0_reconst, img1_reconst], nrow=2\n ),\n title=title,\n y_labels=[(2, \"Original\"), (5, \"Reconstruction\")],\n )\n\n return pairs, distances\n\n\n def unnormalize(self, image: torch.Tensor) -> torch.Tensor:\n \"\"\"Unnormalize an image.\n\n Args:\n image (:class:`torch.Tensor`)\n\n Returns:\n image (:class:`torch.Tensor`)\n\n \"\"\"\n unorm = UnNormalize(mean=self._mean, std=self._std)\n return unorm(image)\n\n @torch.no_grad()\n def decode(\n self,\n embedding: Optional[np.ndarray] = None,\n index: Optional[int] = None,\n show: bool = False,\n astensor: bool = False,\n ) -> np.ndarray:\n \"\"\"Decode embeddings at `index` or pass `embedding` directly\n\n Args:\n embedding (np.ndarray, optional): embedding of image\n index (int): index (of validation set / embeddings) to decode\n show (bool): plot the results\n astensor (bool): keep as torch.Tensor\n\n Returns:\n image (np.ndarray or torch.Tensor): reconstructed image from embedding\n\n \"\"\"\n self.model.eval()\n\n if embedding is None:\n embedding = self.embeddings[index]\n\n emb = np.expand_dims(embedding, 0) # add batch axis\n\n # Check if has direct access to `decode` method\n if not hasattr(self.model, \"decode\"):\n image, _ = self.model.module.decode(torch.Tensor(emb).to(self._device))\n else:\n image, _ = self.model.decode(torch.Tensor(emb).to(self._device))\n\n image = self.unnormalize(image)\n\n if show:\n grid_img = torchvision.utils.make_grid(image)\n self.show(grid_img, title=index)\n\n if astensor:\n return image.detach().cpu()\n return image.detach().cpu().numpy()\n", "sub_path": "src/simages/extractor.py", "file_name": "extractor.py", "file_ext": "py", "file_size_in_byte": 20172, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "warnings.filterwarnings", "line_number": 30, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 31, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 47, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 47, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 57, "usage_type": "name"}, {"api_name": "torch.nn", "line_number": 57, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 58, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 59, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 60, "usage_type": "name"}, {"api_name": "torch.device", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 87, "usage_type": "attribute"}, {"api_name": "torchvision.transforms.Compose", "line_number": 97, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 97, "usage_type": "name"}, {"api_name": "torchvision.transforms.RandomResizedCrop", "line_number": 99, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 99, "usage_type": "name"}, {"api_name": "torchvision.transforms.RandomHorizontalFlip", "line_number": 100, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 100, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 101, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 101, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 102, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 102, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 105, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 105, "usage_type": "name"}, {"api_name": "torchvision.transforms.Resize", "line_number": 107, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 107, "usage_type": "name"}, {"api_name": "torchvision.transforms.CenterCrop", "line_number": 108, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 108, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 109, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 109, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 110, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 110, "usage_type": "name"}, {"api_name": "os.path.splitext", "line_number": 127, "usage_type": "call"}, {"api_name": "os.path", "line_number": 127, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 132, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 132, "usage_type": "name"}, {"api_name": "os.path.basename", "line_number": 134, "usage_type": "call"}, {"api_name": "os.path", "line_number": 134, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 139, "usage_type": "call"}, {"api_name": "os.path", "line_number": 139, "usage_type": "attribute"}, {"api_name": "dataset.ImageFolder", "line_number": 141, "usage_type": "call"}, {"api_name": "dataset.ImageFolder", "line_number": 144, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 147, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 147, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 150, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 150, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 153, "usage_type": "attribute"}, {"api_name": "torch.cuda.is_available", "line_number": 161, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 161, "usage_type": "attribute"}, {"api_name": "models.BasicAutoencoder", "line_number": 169, "usage_type": "call"}, {"api_name": "torch.cuda.device_count", "line_number": 171, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 171, "usage_type": "attribute"}, {"api_name": "torch.cuda.device_count", "line_number": 172, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 172, "usage_type": "attribute"}, {"api_name": "torch.nn.DataParallel", "line_number": 173, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 173, "usage_type": "name"}, {"api_name": "torch.nn.MSELoss", "line_number": 176, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 176, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 177, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 177, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 189, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 203, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 212, "usage_type": "attribute"}, {"api_name": "torchvision.transforms", "line_number": 213, "usage_type": "attribute"}, {"api_name": "numpy.newaxis", "line_number": 221, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 224, "usage_type": "call"}, {"api_name": "torchvision.transforms.functional.to_pil_image", "line_number": 225, "usage_type": "call"}, {"api_name": "torchvision.transforms.functional", "line_number": 225, "usage_type": "name"}, {"api_name": "dataset.PILDataset", "line_number": 226, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 226, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 227, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 227, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 215, "usage_type": "attribute"}, {"api_name": "torch.utils.data", "line_number": 215, "usage_type": "name"}, {"api_name": "torchvision.utils.make_grid", "line_number": 260, "usage_type": "call"}, {"api_name": "torchvision.utils", "line_number": 260, "usage_type": "attribute"}, {"api_name": "torchvision.utils.make_grid", "line_number": 305, "usage_type": "call"}, {"api_name": "torchvision.utils", "line_number": 305, "usage_type": "attribute"}, {"api_name": "torch.cat", "line_number": 319, "usage_type": "call"}, {"api_name": "closely.solve", "line_number": 336, "usage_type": "call"}, {"api_name": "closely.solve", "line_number": 340, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 323, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 323, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 345, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 347, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 352, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 352, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 352, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 367, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 369, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 377, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 377, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 378, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 378, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 386, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 386, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 388, "usage_type": "name"}, {"api_name": "torchvision.utils.make_grid", "line_number": 393, "usage_type": "call"}, {"api_name": "torchvision.utils", "line_number": 393, "usage_type": "attribute"}, {"api_name": "os.name", "line_number": 420, "usage_type": "attribute"}, {"api_name": "bokeh.plotting.output_file", "line_number": 421, "usage_type": "call"}, {"api_name": "sklearn.decomposition.PCA", "line_number": 431, "usage_type": "call"}, {"api_name": "bokeh.models.ColumnDataSource", "line_number": 436, "usage_type": "call"}, {"api_name": "os.path.relpath", "line_number": 441, "usage_type": "call"}, {"api_name": "os.path", "line_number": 441, "usage_type": "attribute"}, {"api_name": "bokeh.models.HoverTool", "line_number": 445, "usage_type": "call"}, {"api_name": "bokeh.plotting.figure", "line_number": 455, "usage_type": "call"}, {"api_name": "bokeh.plotting.show", "line_number": 462, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 467, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 465, "usage_type": "attribute"}, {"api_name": "numpy.int_", "line_number": 480, "usage_type": "attribute"}, {"api_name": "numpy.int64", "line_number": 480, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 487, "usage_type": "call"}, {"api_name": "os.path", "line_number": 487, "usage_type": "attribute"}, {"api_name": "torchvision.utils.make_grid", "line_number": 523, "usage_type": "call"}, {"api_name": "torchvision.utils", "line_number": 523, "usage_type": "attribute"}, {"api_name": "torch.no_grad", "line_number": 494, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 495, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 533, "usage_type": "attribute"}, {"api_name": "models.UnNormalize", "line_number": 543, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 549, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 549, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 550, "usage_type": "name"}, {"api_name": "numpy.expand_dims", "line_number": 571, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 575, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 577, "usage_type": "call"}, {"api_name": "bokeh.plotting.show", "line_number": 581, "usage_type": "name"}, {"api_name": "torchvision.utils.make_grid", "line_number": 582, "usage_type": "call"}, {"api_name": "torchvision.utils", "line_number": 582, "usage_type": "attribute"}, {"api_name": "torch.no_grad", "line_number": 546, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 553, "usage_type": "attribute"}]} +{"seq_id": "430805123", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sun May 2 16:44:54 2021\n\n@author: lijia\n\"\"\"\n\n\"\"\"\n准备测试环境\n\"\"\"\ntry:\n import os\n import json\n import logging\n #设置日志格式\n LOG_FORMAT = \"%(asctime)s - %(levelname)s - %(message)s\"\n DATE_FORMAT = \"%m/%d/%Y %H:%M:%S %p\"\n logging.basicConfig(level=logging.DEBUG, format=LOG_FORMAT, datefmt=DATE_FORMAT)\nexcept:\n print(\"请配置好logging环境,以便于正常输出log日志\")\n\ntry:\n import unittest\n logging.info(\"unitest库成功导入\")\nexcept:\n logging.warning(\"请配置好unittest测试环境!\")\n\ntry:\n from tkinter import *\n from tkinter import ttk\n from tkinter.filedialog import*\n import tkinter\n from tkinter.messagebox import askokcancel,showinfo,WARNING\n import ast\n logging.info(\"tkinter、ast库成功导入\")\nexcept:\n logging.warning(\"请配置好tkinter、ast环境!\")\n \n# 文件功能测试类\nclass fileTest(unittest.TestCase):\n def setUp(self):\n #测试前需执行的操作\n self.jsonFile = \"test.json\"\n self.txtFile = \"test.txt\"\n with open(self.jsonFile,\"w\") as f:\n json.dump({'comment_text': '中国平安(SH601318)太惨了上海机场(SH600009)'}, f)\n logging.info(\"成功创建json文件\")\n with open(self.txtFile, \"w\") as f:\n logging.info(\"成功创建txt文件\")\n \n def tearDown(self):\n #测试用例执行完后所需执行的操作\n if os.path.exists(self.jsonFile):\n os.remove(self.jsonFile)\n logging.info(\"成功删除json文件\")\n else:\n logging.warning(\"没有找到json文件!\")\n \n if os.path.exists(self.txtFile):\n os.remove(self.txtFile)\n logging.info(\"成功删除txt文件\")\n else:\n logging.warning(\"没有找到txt文件!\")\n\n def testImportFile(self): # 导入json文件功能测试\n #具体的测试脚本\n name = importfile(Listbox()).split('/')[-1]\n self.assertEqual(self.jsonFile[-4:], name[-4:], \"导入文件失败\")\n\n def testNewFile(self): # 新建文件功能测试\n try:\n new_file()\n logging.info(\"新建文件功能正常!\")\n except:\n logging.warning(\"新建文件功能出现错误!\")\n \n def testSaveFile(self): # 导出文件功能测试\n try:\n filesave() \n logging.info(\"导出文件功能正常!\")\n except:\n logging.warning(\"导出文件功能出现错误!\")\n \n def testSaveAsFile(self): # 文件另存功能测试\n try:\n filesaveas() \n logging.info(\"文件另存功能正常!\")\n except:\n logging.warning(\"文件另存功能出现错误!\") \n \n \n \nif __name__ == \"__main__\":\n try:\n from datagui import importfile, new_file, filesave, filesaveas\n logging.info(\"测试代码成功导入\")\n except:\n logging.warning(\"测试代码未成功导入,请检查!\")\n unittest.main()", "sub_path": "test/fileTest/fileTest.py", "file_name": "fileTest.py", "file_ext": "py", "file_size_in_byte": 3102, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "logging.basicConfig", "line_number": 18, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 18, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 24, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 26, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 35, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 37, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 40, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 46, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 47, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 54, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 55, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 60, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 61, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 63, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 73, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 75, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 80, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 82, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 87, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 89, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 96, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 98, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 99, "usage_type": "call"}]} +{"seq_id": "617742170", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\nimport os\nimport re\nimport sys\nimport codecs\nimport json\nimport time\nimport filecmp\nimport urllib.request\nfrom datetime import datetime\nfrom bs4 import BeautifulSoup, Tag\nfrom elasticsearch import Elasticsearch\nfrom apscheduler.schedulers.blocking import BlockingScheduler\n\n\n# URL定義\nurls = [#sns情報\n 'http://vip.blomaga.jp/articles/51112.html', #2ch\n 'http://ameblo.jp/takapon-jp/entry-10277141534.html', #blog\n 'https://twitter.com/takapon_jp?lang=ja', #twitter\n 'https://ja-jp.facebook.com/horieofficial', #facebook\n 'https://www.instagram.com/takapon_jp', #instagram\n #財務情報\n 'https://profile.yahoo.co.jp/consolidate/1766', #yahoo finance,\n 'http://www.nikkei.com/nkd/company/kessan/?scode=1766&ba=1', #日系企業情報\n 'http://www.msn.com/ja-jp/money/stockdetails/analysis/fi-133.1.1766.TKS', #msnマネー\n 'http://gyokai-search.com/3-kensetu.htm', #業界動向サーチ\n 'http://www.token.co.jp/ir/pdf/quarterly17tanshin_3r.pdf' #pdf\n ]\n\ndef job_function():\n # webからHTMLファイルを取得\n for url in urls:\n req = urllib.request.Request(url)\n response = urllib.request.urlopen(req)\n data = response.read()\n soup = BeautifulSoup(data, \"html5lib\")\n html = soup.prettify(\"utf-8\")\n\n #ヘッダ情報取得\n #for Header in str (response.info ()).rstrip ().splitlines ():\n #lastmodified = response.headers['last-modified']\n # print (Header)\n\n # urlから不要な文字列を除去\n url=re.sub(r'https://', \"\", url)\n url=re.sub(r'http://', \"\", url)\n url=re.sub(r'/', \"_\", url)\n url=re.sub(r'.html', \"\", url)\n url=re.sub('\\n', \" \", url)\n\n # datetimeを取得し、ファイル名に連結\n # tdatetime = datetime.now()\n # tstr = tdatetime.strftime('%Y-%m-%d %H:%M:%S')\n # url = \"{} {}\".format(url, tstr)\n\n # 出力ファイルオープン\n if url[-4:] == '.pdf':\n url = re.sub(r'.pdf', \"\", url)\n f = open('./url/{}.pdf'.format(url), 'wb')\n f.write(data)\n f.close()\n # ファイルの内容を比較\n if os.path.exists('./url/{}_latest.pdf'.format(url)):\n tf = filecmp.cmp('./url/{}.pdf'.format(url), './url/{}_latest.pdf'.format(url))\n if tf == True:\n print (url + ' 更新無し')\n os.remove('./url/{}.pdf'.format(url))\n else:\n print (url + ' 更新有り')\n os.remove('./url/{}_latest.pdf'.format(url))\n os.rename('./url/{}.pdf'.format(url), './url/{}_latest.pdf'.format(url))\n else:\n os.rename('./url/{}.pdf'.format(url), './url/{}_latest.pdf'.format(url))\n\n else:\n f = open('./url/{}.html'.format(url), 'wb')\n f.write(html)\n f.close()\n # ファイルの内容を比較\n if os.path.exists('./url/{}_latest.html'.format(url)):\n tf = filecmp.cmp('./url/{}.html'.format(url), './url/{}_latest.html'.format(url))\n if tf == True:\n print (url + ' 更新無し')\n os.remove('./url/{}.html'.format(url))\n else:\n print (url + ' 更新有り')\n os.remove('./url/{}_latest.html'.format(url))\n os.rename('./url/{}.html'.format(url), './url/{}_latest.html'.format(url))\n else:\n os.rename('./url/{}.html'.format(url), './url/{}_latest.html'.format(url))\n\n # 出力ファイルオープン(日経Web)\n if url.startswith('www.nikkei.com'):\n f = open('./txt/{}.txt'.format(url), 'wt')\n\n # 正式社名\n n1 = soup.find(text='正式社名')\n n2 = n1.next('td').contents[0]\n f.write(n2 + '\\n')\n\n'''\n f.close()\n\n # 出力ファイルオープン(yahoo finance)\n if url.startswith('profile.yahoo'):\n f = open('./txt/{}.txt'.format(url), 'wt')\n f.write(url[-4:] + '\\n')\n \n # 自己資本比率\n y1 = soup.find(text='自己資本比率').findNext('td').contents[0]\n # nikimae = soup.find(text='自己資本比率').findNext('td').findNext('td').contents[0]\n f.write(y1 + '\\n')\n # ROE(自己資本利益率)\n y2 = soup.find(text='ROE(自己資本利益率)').findNext('td').contents[0]\n f.write(y2 + '\\n')\n # ROA(総資産利益率)\n y3 = soup.find(text='ROA(総資産利益率)').findNext('td').contents[0]\n f.write(y3 + '\\n')\n # 当期利益\n y4 = soup.find(text='当期利益').findNext('td').contents[0]\n f.write(y4 + '\\n')\n \n f.close()\n'''\n\n #es = Elasticsearch(\"localhost:9200\")\n #es.index(index=\"sns\", doc_type=\"2ch\", body=html_data)\n\nif __name__ == \"__main__\":\n #一定間隔でurl取得job実行\n sc = BlockingScheduler(standalone=True, coalesce=True)\n sc.add_job(job_function, \"interval\" ,seconds=30)\n #定時にurl取得job実行\n# sc = BlockingScheduler(timezone=\"UTC\")\n# sc.add_job(job_function, \"cron\", hour=0, minute=0)\n sc.start()\n", "sub_path": "batch/batch_main.py", "file_name": "batch_main.py", "file_ext": "py", "file_size_in_byte": 5182, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "urllib.request.request.Request", "line_number": 35, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 35, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 35, "usage_type": "name"}, {"api_name": "urllib.request.request.urlopen", "line_number": 36, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 36, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 36, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 38, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 47, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 48, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 49, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 50, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 51, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path", "line_number": 65, "usage_type": "attribute"}, {"api_name": "filecmp.cmp", "line_number": 66, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 69, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 72, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 73, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "filecmp.cmp", "line_number": 83, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 86, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 89, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 90, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 92, "usage_type": "call"}, {"api_name": "apscheduler.schedulers.blocking.BlockingScheduler", "line_number": 133, "usage_type": "call"}]} +{"seq_id": "290943971", "text": "\"\"\"\nSQuAD with Bidirectional Encoder Representations from Transformers\n\n=========================================================================================\n\nBERT base model exporter\n\n@article{devlin2018bert,\n title={BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding},\n author={Devlin, Jacob and Chang, Ming- \\\n Wei and Lee, Kenton and Toutanova, Kristina},\n journal={arXiv preprint arXiv:1810.04805},\n year={2018}\n}\n\"\"\"\n\n# coding=utf-8\n\n# Licensed to the Apache Software Foundation (ASF) under one\n# or more contributor license agreements. See the NOTICE file\n# distributed with this work for additional information\n# regarding copyright ownership. The ASF licenses this file\n# to you under the Apache License, Version 2.0 (the\n# \"License\"); you may not use this file except in compliance\n# with the License. You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing,\n# software distributed under the License is distributed on an\n# \"AS IS\" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY\n# KIND, either express or implied. See the License for the\n# specific language governing permissions and limitations\n# under the License.\n# pylint:disable=redefined-outer-name,logging-format-interpolation\n\nimport argparse\nimport collections\nimport json\nimport logging\nimport os\nimport io\nimport copy\nimport random\nimport time\nimport warnings\nimport ctypes\n\nos.environ[\"MXNET_GPU_WORKER_NTHREADS\"] = \"1\"\nos.environ[\"MXNET_COPY_WORKER_NTHREADS\"] = \"1\"\n\nimport numpy as np\nimport mxnet as mx\n\nimport gluonnlp as nlp\nfrom gluonnlp.data import SQuAD\nfrom data.qa import SQuADTransform, preprocess_dataset\nfrom bert_qa_evaluate import get_F1_EM, predict, PredResult\nfrom export import hybrid_bert, hybrid_bert_old\n\nnp.random.seed(6)\nrandom.seed(6)\nmx.random.seed(6)\n\nsend_lib = ctypes.CDLL('./libsend_data.so')\n\nlog = logging.getLogger('gluonnlp')\nlog.setLevel(logging.DEBUG)\nformatter = logging.Formatter(\n fmt='%(levelname)s:%(name)s:%(asctime)s %(message)s', datefmt='%H:%M:%S')\n\n\nparser = argparse.ArgumentParser(description='BERT QA model exporter for inference.')\n\nparser.add_argument('--dtype',\n type=str,\n help='Type to use. Currently only float16 is supported.')\n\nparser.add_argument('--model_parameters',\n type=str,\n default=None,\n help='Model parameter file')\n\nparser.add_argument('--bert_model',\n type=str,\n default='bert_12_768_12',\n help='BERT model name. Currently only bert_12_768_12 is supported.')\n\nparser.add_argument('--uncased',\n action='store_false',\n help='if not set, inputs are converted to lower case.')\n\nparser.add_argument('--max_seq_length',\n type=int,\n default=384,\n help='The maximum total input sequence length after WordPiece tokenization.'\n 'Sequences longer than this will be truncated, and sequences shorter '\n 'than this will be padded. default is 384')\n\nparser.add_argument('--doc_stride',\n type=int,\n default=128,\n help='When splitting up a long document into chunks, how much stride to '\n 'take between chunks. default is 128')\n\nparser.add_argument('--max_query_length',\n type=int,\n default=64,\n help='The maximum number of tokens for the question. Questions longer than '\n 'this will be truncated to this length. default is 64')\n\nparser.add_argument('--gpu',\n type=int,\n default=0,\n help='which gpu to use. GPU(0) is used if not set.')\n\nparser.add_argument('--sentencepiece',\n type=str,\n default=None,\n help='Path to the sentencepiece .model file for both tokenization and vocab.')\n\nparser.add_argument('--gemms_compute_type',\n type=str,\n default='float16',\n help='Precision or compute type to use in cublas forward GEMM. Default float16')\n\nargs = parser.parse_args()\n\nconsole = logging.StreamHandler()\nconsole.setLevel(logging.INFO)\nconsole.setFormatter(formatter)\nlog.addHandler(console)\n\nlog.info(args)\n\nmodel_name = args.bert_model\ndataset_name = 'book_corpus_wiki_en_uncased'\nmodel_parameters = args.model_parameters\nlower = args.uncased\n\ntest_batch_size = 1\nctx = mx.gpu(args.gpu)\n\nmax_seq_length = args.max_seq_length\ndoc_stride = args.doc_stride\nmax_query_length = args.max_query_length\n\nif max_seq_length <= max_query_length + 3:\n raise ValueError('The max_seq_length (%d) must be greater than max_query_length '\n '(%d) + 3' % (max_seq_length, max_query_length))\n\n# vocabulary and tokenizer\nif args.sentencepiece:\n logging.info('loading vocab file from sentence piece model: %s', args.sentencepiece)\n if dataset_name:\n warnings.warn('Both --dataset_name and --sentencepiece are provided. '\n 'The vocabulary will be loaded based on --sentencepiece.')\n vocab = nlp.vocab.BERTVocab.from_sentencepiece(args.sentencepiece)\n dataset_name = None\nelse:\n vocab = None\n\nuse_fp16 = (args.dtype == 'float16')\nif not use_fp16:\n raise ValueError(\"Currently only float16 is supported.\")\n\nhybrid_bert_old.init_fast_softmax()\norig_bert, _ = hybrid_bert_old.get_hybrid_model(\n name=model_name,\n dataset_name=dataset_name,\n vocab=vocab,\n pretrained=False,\n ctx=ctx,\n use_pooler=False,\n use_decoder=False,\n use_classifier=False,\n seq_length=args.max_seq_length,\n use_FP16=use_fp16)\norig_net = hybrid_bert_old.HybridBERTForQA(bert=orig_bert, use_FP16=use_fp16)\nif model_parameters:\n nlp.utils.load_parameters(orig_net, model_parameters, ctx=ctx, cast_dtype=None)\n\nhybrid_bert.init_fast_multiheadattn_and_softmax(args.gemms_compute_type)\nhybrid_bert.detach_addbias_and_set_gemms_compute_type(args.gemms_compute_type)\nbert, vocab = hybrid_bert.get_hybrid_model(\n name=model_name,\n dataset_name=dataset_name,\n vocab=vocab,\n pretrained=False,\n ctx=ctx,\n use_pooler=False,\n use_decoder=False,\n use_classifier=False,\n seq_length=args.max_seq_length,\n use_FP16=use_fp16)\n\n\ndef convert_arg_params(net_arg_params, loaded_arg_params, num_heads):\n for k, v in net_arg_params.items():\n k = k.replace('hybridbertencoder1', 'hybridbertencoder0')\n k = k.replace('hybridbertmodel1', 'hybridbertmodel0')\n k = k.replace('hybridbertforqa1', 'hybridbertforqa0')\n\n if k.endswith('ffn_1_bias_alone'):\n oldname = k[:-6]\n ffn_1_bias = loaded_arg_params[oldname].data().reshape(shape=(1, 1, -1))\n v.set_data(ffn_1_bias.astype(v.dtype))\n elif k.endswith('ffn_2_bias_alone'):\n oldname = k[:-6]\n ffn_2_bias = loaded_arg_params[oldname].data().reshape(shape=(1, 1, -1))\n v.set_data(ffn_2_bias.astype(v.dtype))\n elif k.endswith('proj_bias_alone'):\n oldname = k[:-6]\n proj_bias = loaded_arg_params[oldname].data().reshape(shape=(1, 1, -1))\n v.set_data(proj_bias.astype(v.dtype))\n elif k.endswith('proj_inweight'):\n assert k[:-13] + 'query_weight' in loaded_arg_params\n assert k[:-13] + 'key_weight' in loaded_arg_params\n assert k[:-13] + 'value_weight' in loaded_arg_params\n q_weight = loaded_arg_params[k[:-13] + 'query_weight'].data().reshape(shape=(num_heads, -1, 0),\n reverse=True)\n k_weight = loaded_arg_params[k[:-13] + 'key_weight'].data().reshape(shape=(num_heads, -1, 0),\n reverse=True)\n v_weight = loaded_arg_params[k[:-13] + 'value_weight'].data().reshape(shape=(num_heads, -1, 0),\n reverse=True)\n all_weight = mx.nd.concat(q_weight, k_weight, v_weight, dim=-2)\n all_weight = mx.nd.reshape(all_weight, shape=(-1, 0), reverse=True)\n v.set_data(all_weight.astype(v.dtype))\n elif k.endswith('proj_inbias'):\n assert k[:-11] + 'query_bias' in loaded_arg_params\n assert k[:-11] + 'key_bias' in loaded_arg_params\n assert k[:-11] + 'value_bias' in loaded_arg_params\n q_bias = loaded_arg_params[k[:-11] + 'query_bias'].data().reshape(shape=(num_heads, -1),\n reverse=True)\n k_bias = loaded_arg_params[k[:-11] + 'key_bias'].data().reshape(shape=(num_heads, -1),\n reverse=True)\n v_bias = loaded_arg_params[k[:-11] + 'value_bias'].data().reshape(shape=(num_heads, -1),\n reverse=True)\n all_bias = mx.nd.stack(q_bias, k_bias, v_bias, axis=1)\n all_bias = mx.nd.reshape(all_bias, shape=(-1))\n v.set_data(all_bias.astype(v.dtype))\n else:\n v.set_data(loaded_arg_params[k].data().astype(v.dtype))\n\nnet = hybrid_bert.HybridBERTForQA(bert=bert, use_FP16=use_fp16)\nnet.initialize(mx.init.Xavier(), ctx=ctx)\nnet.hybridize(static_alloc=True, static_shape=True)\nif use_fp16:\n net.cast('float16')\nif model_parameters:\n convert_arg_params(net.collect_params(), orig_net.collect_params(), 12)\nif args.sentencepiece:\n tokenizer = nlp.data.BERTSPTokenizer(args.sentencepiece, vocab, lower=lower)\nelse:\n tokenizer = nlp.data.BERTTokenizer(vocab=vocab, lower=lower)\n\ndef _transposed_pad_arrs_to_max_length(arrs, pad_axis, pad_val, use_shared_mem, dtype):\n if isinstance(arrs[0], mx.nd.NDArray):\n dtype = arrs[0].dtype if dtype is None else dtype\n arrs = [arr.asnumpy() for arr in arrs]\n elif not isinstance(arrs[0], np.ndarray):\n arrs = [np.asarray(ele) for ele in arrs]\n else:\n dtype = arrs[0].dtype if dtype is None else dtype\n\n original_length = [ele.shape[pad_axis] for ele in arrs]\n max_size = max(original_length)\n\n ret_shape = list(arrs[0].shape)\n ret_shape[pad_axis] = max_size\n ret_shape = (ret_shape[0], len(arrs)) + tuple(ret_shape[1:])\n ret = np.full(shape=ret_shape, fill_value=pad_val, dtype=dtype)\n for i, arr in enumerate(arrs):\n if arr.shape[pad_axis] == max_size:\n ret[:,i] = arr\n else:\n slices = [slice(None) for _ in range(arr.ndim)]\n slices[pad_axis] = slice(0, arr.shape[pad_axis])\n if slices[pad_axis].start != slices[pad_axis].stop:\n slices = [slice(i, i + 1)] + slices\n ret[:, tuple(slices)] = arr\n ctx = mx.Context('cpu_shared', 0) if use_shared_mem else mx.cpu()\n ret = mx.nd.array(ret, ctx=ctx, dtype=dtype)\n original_length = mx.nd.array(original_length, ctx=ctx, dtype=np.int32)\n\n return ret, original_length\n\nclass TransposedPad(object):\n def __init__(self, axis=0, pad_val=None, ret_length=False, dtype=None):\n self._axis = axis\n assert isinstance(axis, int), 'axis must be an integer! ' \\\n 'Received axis=%s, type=%s.' % (str(axis),\n str(type(axis)))\n self._pad_val = 0 if pad_val is None else pad_val\n self._ret_length = ret_length\n self._dtype = dtype\n self._warned = False\n if pad_val is None:\n warnings.warn(\"padding value is not given and will be set automatically to 0\")\n\n def __call__(self, data):\n if isinstance(data[0], mx.nd.NDArray) and not self._warned:\n self._warned = True\n warnings.warn(\"Using Pad with NDArrays is discouraged for speed reasons...\")\n\n if isinstance(data[0], (mx.nd.NDArray, np.ndarray, list)):\n padded_arr, original_length = _transposed_pad_arrs_to_max_length(data, self._axis,\n self._pad_val, True,\n self._dtype)\n if self._ret_length:\n return padded_arr, original_length\n else:\n return padded_arr\n else:\n raise NotImplementedError\n\nbatchify_fn = nlp.data.batchify.Tuple(\n nlp.data.batchify.Stack(),\n TransposedPad(axis=0, pad_val=vocab[vocab.padding_token]),\n TransposedPad(axis=0, pad_val=vocab[vocab.padding_token]),\n nlp.data.batchify.Stack('float32'),\n nlp.data.batchify.Stack('float32'),\n nlp.data.batchify.Stack('float32'))\n\ndef send_data(arr_cpu, arr_gpu):\n cpu_handle = arr_cpu.handle\n gpu_handle = arr_gpu.handle\n if len(arr_cpu.shape) == 1:\n nbytes = 4 * arr_cpu.shape[0]\n elif len(arr_cpu.shape) == 2:\n nbytes = 4 * arr_cpu.shape[0] * arr_cpu.shape[1]\n ptr_cpu = ctypes.c_void_p()\n ptr_gpu = ctypes.c_void_p()\n mx.base._LIB.MXNDArrayGetData(arr_cpu.handle, ctypes.byref(ptr_cpu))\n mx.base._LIB.MXNDArrayGetData(arr_gpu.handle, ctypes.byref(ptr_gpu))\n send_lib.send_data(ptr_cpu, ptr_gpu, ctypes.c_size_t(nbytes))\n\ndef send_data_to_GPU(input_cpu, token_types_cpu, valid_length_cpu, input_gpu, token_types_gpu, valid_length_gpu):\n # Send data to the GPU using stream 0 in order to avoid\n # cudaStreamSynchronize call overhead that would happen\n # when naively copying the data to GPU via MXNet\n send_data(input_cpu, input_gpu)\n send_data(token_types_cpu, token_types_gpu)\n send_data(valid_length_cpu, valid_length_gpu)\n\ndef recv_data(arr_gpu, arr_cpu, nbytes, sync=False):\n gpu_handle = arr_gpu.handle\n ptr_gpu = ctypes.c_void_p()\n mx.base._LIB.MXNDArrayGetData(arr_gpu.handle, ctypes.byref(ptr_gpu))\n if sync:\n send_lib.recv_data_sync(arr_cpu, ptr_gpu, ctypes.c_size_t(nbytes))\n else:\n send_lib.recv_data_async(arr_cpu, ptr_gpu, ctypes.c_size_t(nbytes))\n\ndef recv_data_from_GPU(start_gpu, end_gpu, start_cpu, end_cpu):\n if use_fp16:\n nbytes = 2 * start_gpu.shape[0] * start_gpu.shape[1]\n else:\n nbytes = 4 * start_gpu.shape[0] * start_gpu.shape[1]\n recv_data(start_gpu, start_cpu, nbytes)\n recv_data(end_gpu, end_cpu, nbytes, sync=True)\n\ndef export():\n \"\"\"Evaluate the model on validation dataset.\n \"\"\"\n log.info('Loading dev data...')\n dev_data = SQuAD('dev', version='1.1')\n log.info('Number of records in dev data:{}'.format(len(dev_data)))\n\n dev_dataset = dev_data.transform(\n SQuADTransform(\n copy.copy(tokenizer),\n max_seq_length=max_seq_length,\n doc_stride=doc_stride,\n max_query_length=max_query_length,\n is_pad=False,\n is_training=False)._transform, lazy=False)\n\n dev_data_transform, _ = preprocess_dataset(\n dev_data, SQuADTransform(\n copy.copy(tokenizer),\n max_seq_length=max_seq_length,\n doc_stride=doc_stride,\n max_query_length=max_query_length,\n is_pad=True,\n is_training=False))\n log.info('The number of examples after preprocessing:{}'.format(\n len(dev_data_transform)))\n\n dev_dataloader = mx.gluon.data.DataLoader(\n dev_data_transform,\n batchify_fn=batchify_fn,\n num_workers=4, batch_size=test_batch_size,\n shuffle=False, last_batch='discard')\n\n total_iters=0\n total_time=0.0\n total_len = 0\n\n inputs_GPU = mx.nd.zeros((max_seq_length, test_batch_size), ctx=ctx)\n token_types_GPU = mx.nd.zeros((max_seq_length, test_batch_size), ctx=ctx)\n valid_length_GPU = mx.nd.zeros((test_batch_size,), ctx=ctx)\n\n if use_fp16:\n pred_start = np.zeros((test_batch_size, max_seq_length), dtype=np.float16)\n pred_end = np.zeros((test_batch_size, max_seq_length), dtype=np.float16)\n else:\n raise ValueError\n pred_start = np.zeros((test_batch_size, max_seq_length), dtype=np.float32)\n pred_end = np.zeros((test_batch_size, max_seq_length), dtype=np.float32)\n\n start_ptr = pred_start.ctypes.data_as(ctypes.c_void_p)\n end_ptr = pred_end.ctypes.data_as(ctypes.c_void_p)\n\n for data in dev_dataloader:\n example_ids, inputs, token_types, valid_length, _, _ = data\n mx.nd.waitall()\n tic = time.time()\n\n send_data_to_GPU(inputs, token_types, valid_length, inputs_GPU, token_types_GPU, valid_length_GPU)\n out1, out2 = net(inputs_GPU,\n token_types_GPU,\n valid_length_GPU)\n mx.nd.waitall()\n recv_data_from_GPU(out1, out2, start_ptr, end_ptr)\n\n toc = time.time()\n\n net.export('fp16_bert_' + str(max_seq_length), 0)\n break\n\nif __name__ == '__main__':\n if model_parameters:\n export()\n else:\n raise RuntimeError(\"Need to provide model_parameters option\")\n", "sub_path": "nvidia-examples/gluon/bert_inference/export_symbolic_model.py", "file_name": "export_symbolic_model.py", "file_ext": "py", "file_size_in_byte": 17243, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "os.environ", "line_number": 49, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 50, "usage_type": "attribute"}, {"api_name": "numpy.random.seed", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 61, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 62, "usage_type": "call"}, {"api_name": "mxnet.random.seed", "line_number": 63, "usage_type": "call"}, {"api_name": "mxnet.random", "line_number": 63, "usage_type": "attribute"}, {"api_name": "ctypes.CDLL", "line_number": 65, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 67, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 68, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 69, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 73, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 129, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 130, "usage_type": "attribute"}, {"api_name": "mxnet.gpu", "line_number": 142, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 154, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 156, "usage_type": "call"}, {"api_name": "gluonnlp.vocab.BERTVocab.from_sentencepiece", "line_number": 158, "usage_type": "call"}, {"api_name": "gluonnlp.vocab", "line_number": 158, "usage_type": "attribute"}, {"api_name": "export.hybrid_bert_old.init_fast_softmax", "line_number": 167, "usage_type": "call"}, {"api_name": "export.hybrid_bert_old", "line_number": 167, "usage_type": "name"}, {"api_name": "export.hybrid_bert_old.get_hybrid_model", "line_number": 168, "usage_type": "call"}, {"api_name": "export.hybrid_bert_old", "line_number": 168, "usage_type": "name"}, {"api_name": "export.hybrid_bert_old.HybridBERTForQA", "line_number": 179, "usage_type": "call"}, {"api_name": "export.hybrid_bert_old", "line_number": 179, "usage_type": "name"}, {"api_name": "gluonnlp.utils.load_parameters", "line_number": 181, "usage_type": "call"}, {"api_name": "gluonnlp.utils", "line_number": 181, "usage_type": "attribute"}, {"api_name": "export.hybrid_bert.init_fast_multiheadattn_and_softmax", "line_number": 183, "usage_type": "call"}, {"api_name": "export.hybrid_bert", "line_number": 183, "usage_type": "name"}, {"api_name": "export.hybrid_bert.detach_addbias_and_set_gemms_compute_type", "line_number": 184, "usage_type": "call"}, {"api_name": "export.hybrid_bert", "line_number": 184, "usage_type": "name"}, {"api_name": "export.hybrid_bert.get_hybrid_model", "line_number": 185, "usage_type": "call"}, {"api_name": "export.hybrid_bert", "line_number": 185, "usage_type": "name"}, {"api_name": "mxnet.nd.concat", "line_number": 226, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 226, "usage_type": "attribute"}, {"api_name": "mxnet.nd.reshape", "line_number": 227, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 227, "usage_type": "attribute"}, {"api_name": "mxnet.nd.stack", "line_number": 239, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 239, "usage_type": "attribute"}, {"api_name": "mxnet.nd.reshape", "line_number": 240, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 240, "usage_type": "attribute"}, {"api_name": "export.hybrid_bert.HybridBERTForQA", "line_number": 245, "usage_type": "call"}, {"api_name": "export.hybrid_bert", "line_number": 245, "usage_type": "name"}, {"api_name": "mxnet.init.Xavier", "line_number": 246, "usage_type": "call"}, {"api_name": "mxnet.init", "line_number": 246, "usage_type": "attribute"}, {"api_name": "gluonnlp.data.BERTSPTokenizer", "line_number": 253, "usage_type": "call"}, {"api_name": "gluonnlp.data", "line_number": 253, "usage_type": "attribute"}, {"api_name": "gluonnlp.data.BERTTokenizer", "line_number": 255, "usage_type": "call"}, {"api_name": "gluonnlp.data", "line_number": 255, "usage_type": "attribute"}, {"api_name": "mxnet.nd", "line_number": 258, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 261, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 262, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 272, "usage_type": "call"}, {"api_name": "mxnet.Context", "line_number": 282, "usage_type": "call"}, {"api_name": "mxnet.cpu", "line_number": 282, "usage_type": "call"}, {"api_name": "mxnet.nd.array", "line_number": 283, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 283, "usage_type": "attribute"}, {"api_name": "mxnet.nd.array", "line_number": 284, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 284, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 284, "usage_type": "attribute"}, {"api_name": "warnings.warn", "line_number": 299, "usage_type": "call"}, {"api_name": "data.qa", "line_number": 302, "usage_type": "name"}, {"api_name": "mxnet.nd", "line_number": 302, "usage_type": "attribute"}, {"api_name": "warnings.warn", "line_number": 304, "usage_type": "call"}, {"api_name": "data.qa", "line_number": 306, "usage_type": "name"}, {"api_name": "mxnet.nd", "line_number": 306, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 306, "usage_type": "attribute"}, {"api_name": "data.qa", "line_number": 307, "usage_type": "argument"}, {"api_name": "gluonnlp.data.batchify.Tuple", "line_number": 317, "usage_type": "call"}, {"api_name": "gluonnlp.data", "line_number": 317, "usage_type": "attribute"}, {"api_name": "gluonnlp.data.batchify.Stack", "line_number": 318, "usage_type": "call"}, {"api_name": "gluonnlp.data", "line_number": 318, "usage_type": "attribute"}, {"api_name": "gluonnlp.data.batchify.Stack", "line_number": 321, "usage_type": "call"}, {"api_name": "gluonnlp.data", "line_number": 321, "usage_type": "attribute"}, {"api_name": "gluonnlp.data.batchify.Stack", "line_number": 322, "usage_type": "call"}, {"api_name": "gluonnlp.data", "line_number": 322, "usage_type": "attribute"}, {"api_name": "gluonnlp.data.batchify.Stack", "line_number": 323, "usage_type": "call"}, {"api_name": "gluonnlp.data", "line_number": 323, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 332, "usage_type": "call"}, {"api_name": "ctypes.c_void_p", "line_number": 333, "usage_type": "call"}, {"api_name": "mxnet.base._LIB.MXNDArrayGetData", "line_number": 334, "usage_type": "call"}, {"api_name": "mxnet.base", "line_number": 334, "usage_type": "attribute"}, {"api_name": "ctypes.byref", "line_number": 334, "usage_type": "call"}, {"api_name": "mxnet.base._LIB.MXNDArrayGetData", "line_number": 335, "usage_type": "call"}, {"api_name": "mxnet.base", "line_number": 335, "usage_type": "attribute"}, {"api_name": "ctypes.byref", "line_number": 335, "usage_type": "call"}, {"api_name": "ctypes.c_size_t", "line_number": 336, "usage_type": "call"}, {"api_name": "ctypes.c_void_p", "line_number": 348, "usage_type": "call"}, {"api_name": "mxnet.base._LIB.MXNDArrayGetData", "line_number": 349, "usage_type": "call"}, {"api_name": "mxnet.base", "line_number": 349, "usage_type": "attribute"}, {"api_name": "ctypes.byref", "line_number": 349, "usage_type": "call"}, {"api_name": "ctypes.c_size_t", "line_number": 351, "usage_type": "call"}, {"api_name": "ctypes.c_size_t", "line_number": 353, "usage_type": "call"}, {"api_name": "gluonnlp.data.SQuAD", "line_number": 367, "usage_type": "call"}, {"api_name": "data.qa.SQuADTransform", "line_number": 371, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 372, "usage_type": "call"}, {"api_name": "data.qa.preprocess_dataset", "line_number": 379, "usage_type": "call"}, {"api_name": "data.qa.SQuADTransform", "line_number": 380, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 381, "usage_type": "call"}, {"api_name": "mxnet.gluon.data.DataLoader", "line_number": 390, "usage_type": "call"}, {"api_name": "mxnet.gluon", "line_number": 390, "usage_type": "attribute"}, {"api_name": "mxnet.nd.zeros", "line_number": 400, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 400, "usage_type": "attribute"}, {"api_name": "mxnet.nd.zeros", "line_number": 401, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 401, "usage_type": "attribute"}, {"api_name": "mxnet.nd.zeros", "line_number": 402, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 402, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 405, "usage_type": "call"}, {"api_name": "numpy.float16", "line_number": 405, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 406, "usage_type": "call"}, {"api_name": "numpy.float16", "line_number": 406, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 409, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 409, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 410, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 410, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 412, "usage_type": "attribute"}, {"api_name": "ctypes.c_void_p", "line_number": 413, "usage_type": "attribute"}, {"api_name": "data.qa", "line_number": 415, "usage_type": "name"}, {"api_name": "data.qa", "line_number": 416, "usage_type": "name"}, {"api_name": "mxnet.nd.waitall", "line_number": 417, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 417, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 418, "usage_type": "call"}, {"api_name": "mxnet.nd.waitall", "line_number": 424, "usage_type": "call"}, {"api_name": "mxnet.nd", "line_number": 424, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 427, "usage_type": "call"}]} +{"seq_id": "274218601", "text": "\nfrom openstack import connection\nimport sys\nimport os\n#import openstack\nfrom time import sleep\n\nauth_args = { 'auth_url': 'https://jpe2.jiocloud.com:5000/v3',\n 'user_domain_name': 'default',\n 'project_domain_name':'default',\n 'project_name': 'JioPhone-Push-Prod',\n 'username': '',\n 'password': '' }\n\nif __name__ == \"__main__\":\n conn = connection.Connection(verify=False,**auth_args)\n\n############## Server Details ######################\n\nname = 'prod-jmn-push-rep'\n\nflavor_id = 'm1.medium'\nflavor = conn.compute.find_flavor(flavor_id)\nprint ('Falvor : ' + flavor.name)\n\nimage_id = 'Centos-7.3'\nimage = conn.compute.find_image(image_id)\nprint ('image : '+ image.name)\n\nkey_name = 'PushProd'\nkeypair = conn.compute.find_keypair(key_name)\nprint ('Keypair : '+ keypair.name)\n\nnetname = 'PushProdFinal'\nnetwork = conn.network.find_network(netname)\nprint ('network name : ' + network.name)\n\n\nprint ('======================================')\n\nazone = 'JPHONE-ROW3-POD'\n\n#root volume size\nrvolsize= 10\n\n#sec volume size\nsvolsize = 50\n\n############################ iterator ####################################\n\n#name of server to start from\nsernum = 52\n\n#number of server you want to create\nsercount = 15\n\n\n############################# volume and server creation ###################################\n \nfor i in range (sernum, sernum+sercount):\n nzone=(i%3)+1\n zone=azone+str(nzone) \n sname= name + '-' + str(i)\n volname= name + '-' + str(i)\n \n vol = conn.block_store.create_volume(name=volname, size=rvolsize, image_id=image.id, is_bootable='boot', availabilty_zone='nova')\n print ('volume : ' + vol.name)\n\n for i in range(6):\n sleep(6)\n vol1 = conn.block_store.get_volume(vol.id)\n if vol1.status=='available': break\n\n\n block_device_mapping = [{'boot_index':'0', 'source_type':'volume', \n 'destination_type':'volume','uuid':vol.id,\n 'delete_on_termination':False}]\n\n\n server = conn.compute.create_server(name=sname, flavor_id=flavor.id, key_name=key_name, networks=[{\"uuid\" : network.id}], availability_zone=zone, block_device_mapping=block_device_mapping )\n\n server = conn.compute.wait_for_server(server)\n\n############################### SG, FIP and vdb2 ###########################################\n security_group = 'Restricted'\n sg = conn.network.find_security_group(security_group)\n print ('Security Group : ' + sg.name)\n conn.compute.add_security_group_to_server(server.id,sg)\n \n### below code will assign the existing free floating IP\n### if uncommented pleae comment create floating IP lines\n# fip=conn.network.ips(fixed_ip_address='None')\n# for ip in fip:\n# if ip.fixed_ip_address==None:\n# float_ip=ip.floating_ip_address \n# print ('floating ip : ' + float_ip)\n# break\n# conn.compute.add_floating_ip_to_server(server.id,float_ip)\n\n\n### below code will create new floating IP and assign it to server\n### if used please comment the above floating IP code.\n extnet=conn.network.find_network('ext-net')\n float_ip=conn.network.create_ip(floating_network_id=extnet.id)\n print ('floating_ip : ' + float_ip.floating_ip_address)\n conn.compute.add_floating_ip_to_server(server.id,float_ip.floating_ip_address)\n\n \n vol2_name = 'data' + volname\n print ('secondary vol : ' + vol2_name)\n vol2 = conn.block_store.create_volume(name=vol2_name, size=svolsize, availabilty_zone='nova')\n for i in range(5):\n sleep(5)\n vol2 = conn.block_store.get_volume(vol2.id)\n if vol2.status=='available': break\n \n conn.compute.create_volume_attachment(server.id,volumeId=vol2.id)\n\n print ('server ready, happy deploying!')\n \n print ('==============================================')\n\n #apend the floating ip to a file for ansible use.\n file = open(\"newcephost\",\"a+\") \n fip = float_ip.floating_ip_address\n file.write(fip)\n file.close()\n\n\n\n#######################################################################################################\n#######################################################################################################\n#######################################################################################################\n#######################################################################################################\n\n\n", "sub_path": "vm_creation_pushprod.py", "file_name": "vm_creation_pushprod.py", "file_ext": "py", "file_size_in_byte": 4256, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "openstack.connection.Connection", "line_number": 16, "usage_type": "call"}, {"api_name": "openstack.connection", "line_number": 16, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 70, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 113, "usage_type": "call"}]} +{"seq_id": "57953406", "text": "#!/usr/bin/env python\n\n\"\"\"\n.. module:: convert\n :synopsis: used to create info.txt and the .txt files.\n\n\"\"\"\nimport sys\nimport os\nimport argparse\n\nargparser = argparse.ArgumentParser(description = \n'create info.txt, txname.txt, twiki.txt and sms.py')\nargparser.add_argument ('-utilsPath', '--utilsPath', \nhelp = 'path to the package smodels_utils',\\\ntype = str )\nargparser.add_argument ('-smodelsPath', '--smodelsPath', \nhelp = 'path to the package smodels_utils',\\\ntype = str )\nargs = argparser.parse_args()\n\nif args.utilsPath:\n utilsPath = args.utilsPath\nelse:\n databaseRoot = '../../../'\n sys.path.append(os.path.abspath(databaseRoot))\n from utilsPath import utilsPath\n utilsPath = databaseRoot + utilsPath\nif args.smodelsPath:\n sys.path.append(os.path.abspath(args.smodelsPath))\n\nsys.path.append(os.path.abspath(utilsPath))\nfrom smodels_utils.dataPreparation.inputObjects import MetaInfoInput,DataSetInput\nfrom smodels_utils.dataPreparation.databaseCreation import databaseCreator\nfrom smodels_utils.dataPreparation.massPlaneObjects import x, y, z\n\n#+++++++ global info block ++++++++++++++\ninfo \t\t\t = MetaInfoInput('ATLAS-SUSY-2016-16')\ninfo.url \t\t = 'https://atlas.web.cern.ch/Atlas/GROUPS/PHYSICS/PAPERS/SUSY-2016-16/'\ninfo.sqrts \t\t = 13\ninfo.lumi \t\t = 36.1\ninfo.prettyName = '1L stop'\ninfo.private \t = False\ninfo.arxiv \t\t = 'https://arxiv.org/abs/1711.11520'\ninfo.contact \t = 'atlas-phys-susy-conveners@cern.ch'\ninfo.publication = 'JHEP 06 (2018) 108'\n\nT2tt = {\n'name' \t\t : ['T2tt','T2ttoff','T2bbffff'],\n'info' \t\t :{'figure' \t\t: ['Fig.20', 'Fig.21'],\n\t\t\t 'figureUrl' \t\t: ['https://atlas.web.cern.ch/Atlas/GROUPS/PHYSICS/PAPERS/SUSY-2016-16/fig_20.png', 'https://atlas.web.cern.ch/Atlas/GROUPS/PHYSICS/PAPERS/SUSY-2016-16/fig_21.png'],\n\t\t\t 'dataUrl' \t\t: ['https://www.hepdata.net/record/ins1639856?version=4&table=Table60','https://www.hepdata.net/record/ins1639856?version=4&table=Table61']},\n'sources'\t :{'expExcl'\t\t: ['orig/HEPData-ins1639856-v4-Table_16.csv','orig/HEPData-ins1639856-v4-Table_19.csv'],\n\t\t\t 'obsExcl'\t\t: ['orig/HEPData-ins1639856-v4-Table_17.csv','orig/HEPData-ins1639856-v4-Table_20.csv'],\n\t\t\t 'upLimit'\t\t: ['orig/HEPData-ins1639856-v4-Table_60.csv','orig/HEPData-ins1639856-v4-Table_61.csv']},\n'constraint' : ['[[[t]],[[t]]]','[[[b, W]],[[b, W]]]','27./8.*[[[b, l, nu]],[[b, jet, jet]]]'],\n'massPlane' : [2*[[x, y]],2*[[x, y]]]}\n\nT6bbWW = {\n'name' \t\t : 'T6bbWW',\n'info' \t\t :{'figure' \t\t: 'Fig.23', \n\t\t\t 'figureUrl' \t\t: 'https://atlas.web.cern.ch/Atlas/GROUPS/PHYSICS/PAPERS/SUSY-2016-16/fig_23.png', \n\t\t\t 'dataUrl' \t\t: 'https://www.hepdata.net/record/ins1639856?version=4&table=Table70'},\n'sources'\t :{'expExcl'\t\t: 'orig/HEPData-ins1639856-v4-Table_28.csv',\n\t\t\t 'obsExcl'\t\t: 'orig/HEPData-ins1639856-v4-Table_29.csv',\n\t\t\t 'upLimit'\t\t: 'orig/HEPData-ins1639856-v4-Table_70.csv'},\n'constraint' : '[[[b],[W]],[[b],[W]]]',\n'massPlane' : 2*[[x, x - 10, y]]}\n\n\nDATA = [T2tt]\n\n#+++++++ dataset block ++++++++++++++\ndataset = DataSetInput('data')\ndataset.setInfo(dataType = 'upperLimit', dataId = None)\n\n\nfor TX in DATA:\n\t#+++++++ next txName block ++++++++++++++\n\tnewTx \t\t\t\t\t\t\t= dataset.addTxName(TX['name'][0])\n\tnewTx.checked \t\t\t\t\t= 'False'\n\tnewTx.constraint \t\t\t\t= TX['constraint'][0]\n\tnewTx.conditionDescription \t\t= None\n\tnewTx.condition \t\t\t\t= None\n\tnewTx.source \t\t\t\t\t= 'ATLAS'\n\t#+++++++ next txName block ++++++++++++++\n\tnewTxOff1 \t\t\t\t\t\t= dataset.addTxName(TX['name'][1])\n\tnewTxOff1.checked \t\t\t\t= 'False'\n\tnewTxOff1.constraint \t\t\t= TX['constraint'][1]\n\tnewTxOff1.conditionDescription \t= None\n\tnewTxOff1.condition \t\t\t= None\n\tnewTxOff1.massConstraint \t\t= [['80 <= dm < 169.0'], ['80 <= dm < 169.0']]\n\tnewTxOff1.source \t\t\t\t= 'ATLAS'\n\t#+++++++ next mass plane block ++++++++++++++\n\tnewTxOff2\t\t\t\t\t\t= dataset.addTxName(TX['name'][2])\n\tnewTxOff2.checked\t\t\t\t= 'False'\n\tnewTxOff2.constraint\t\t\t= TX['constraint'][2]\n\tnewTxOff2.conditionDescription \t= None\n\tnewTxOff2.condition\t\t\t\t= None\n\tnewTxOff2.source\t\t\t\t= 'ATLAS'\n\tnewTxOff2.massConstraint\t\t= [['dm < 80'], ['dm < 80']]\n\n\t#for i in range(len(TX['info']['figure'])):\n\ti = 0\n\n\t#+++++++ next mass plane block ++++++++++++++\n\tnewPlane \t\t\t\t\t\t= newTx.addMassPlane(TX['massPlane'][i])\n\tnewPlane.figure \t\t\t\t= TX['info']['figure'][i]\n\tnewPlane.figureUrl \t\t\t\t= TX['info']['figureUrl'][i]\n\tnewPlane.dataUrl \t\t\t\t= TX['info']['dataUrl'][i]\n\tnewPlane.setSources(dataLabels \t= ['expExclusion', 'obsExclusion', 'upperLimits'],\n\t\t\t\t\tdataFiles \t\t= [TX['sources']['expExcl'][i], TX['sources']['obsExcl'][i], TX['sources']['upLimit'][i]],\n\t\t\t\t\tunits\t\t\t= [ None, None, 'pb' ],\n\t\t\t\t \tcoordinates \t= [ {x: 0, y: 1, 'value': None}, {x: 0, y: 1, 'value': None}, {x : 1, y: 0, 'value' :2} ],\n\t \tdataFormats \t= ['csv', 'csv', 'csv'])\n\n\tnewTxOff1.addMassPlane(newPlane)\n\tnewTxOff2.addMassPlane(newPlane)\n\nDATA = [T6bbWW]\n\nfor TX in DATA:\n\t#+++++++ next txName block ++++++++++++++\n\tnewTx \t\t\t\t\t\t\t= dataset.addTxName(TX['name'])\n\tnewTx.checked \t\t\t\t\t= 'False'\n\tnewTx.constraint \t\t\t\t= TX['constraint']\n\tnewTx.conditionDescription \t\t= None\n\tnewTx.condition \t\t\t\t= None\n\tnewTx.source \t\t\t\t\t= 'ATLAS'\n\t#+++++++ next mass plane block ++++++++++++++\n\tnewPlane \t\t\t\t\t\t= newTx.addMassPlane(TX['massPlane'])\n\tnewPlane.figure \t\t\t\t= TX['info']['figure']\n\tnewPlane.figureUrl \t\t\t\t= TX['info']['figureUrl']\n\tnewPlane.dataUrl \t\t\t\t= TX['info']['dataUrl']\n\tnewPlane.setSources(dataLabels \t= ['expExclusion', 'obsExclusion', 'upperLimits'],\n\t\t\t\t\tdataFiles \t\t= [TX['sources']['expExcl'], TX['sources']['obsExcl'], TX['sources']['upLimit']],\n\t\t\t\t\tunits\t\t\t= [ None, None, 'pb' ],\n\t\t\t\t \tcoordinates \t= [ {x: 0, y: 1, 'value': None}, {x: 0, y: 1, 'value': None}, {x : 1, y: 0, 'value' :2} ],\n \tdataFormats \t= ['csv', 'csv', 'csv'])\n\ndatabaseCreator.create()\n", "sub_path": "smodels-database/13TeV/ATLAS/ATLAS-SUSY-2016-16/convert.py", "file_name": "convert.py", "file_ext": "py", "file_size_in_byte": 5748, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 12, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 26, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "utilsPath.utilsPath", "line_number": 28, "usage_type": "name"}, {"api_name": "sys.path.append", "line_number": 30, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 32, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 32, "usage_type": "call"}, {"api_name": "utilsPath.utilsPath", "line_number": 32, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "smodels_utils.dataPreparation.inputObjects.MetaInfoInput", "line_number": 38, "usage_type": "call"}, {"api_name": "smodels_utils.dataPreparation.massPlaneObjects.x", "line_number": 57, "usage_type": "name"}, {"api_name": "smodels_utils.dataPreparation.massPlaneObjects.y", "line_number": 57, "usage_type": "name"}, {"api_name": "smodels_utils.dataPreparation.massPlaneObjects.x", "line_number": 68, "usage_type": "name"}, {"api_name": "smodels_utils.dataPreparation.massPlaneObjects.y", "line_number": 68, "usage_type": "name"}, {"api_name": "smodels_utils.dataPreparation.inputObjects.DataSetInput", "line_number": 74, "usage_type": "call"}, {"api_name": "smodels_utils.dataPreparation.massPlaneObjects.x", "line_number": 114, "usage_type": "name"}, {"api_name": "smodels_utils.dataPreparation.massPlaneObjects.y", "line_number": 114, "usage_type": "name"}, {"api_name": "smodels_utils.dataPreparation.massPlaneObjects.x", "line_number": 138, "usage_type": "name"}, {"api_name": "smodels_utils.dataPreparation.massPlaneObjects.y", "line_number": 138, "usage_type": "name"}, {"api_name": "smodels_utils.dataPreparation.databaseCreation.databaseCreator.create", "line_number": 141, "usage_type": "call"}, {"api_name": "smodels_utils.dataPreparation.databaseCreation.databaseCreator", "line_number": 141, "usage_type": "name"}]} +{"seq_id": "164020121", "text": "# -*- coding: utf-8 -*-\n'''\n提交命令:\n/opt/spark-2.0.2/bin/spark-submit \\\n--master yarn \\\n--deploy-mode client \\\n--driver-memory 15g \\\n--queue project.hongjing \\\nhdfs_to_hbase.py {version}\n'''\n\nimport sys\nimport os\nimport json\nfrom pyspark.sql import SparkSession\nfrom pyspark.conf import SparkConf\n\n\ndef get_spark_session():\n conf = SparkConf()\n conf.setMaster('yarn-client')\n conf.set(\"spark.yarn.am.cores\", 7)\n conf.set(\"spark.executor.memory\", \"40g\")\n conf.set(\"spark.executor.instances\", 30)\n conf.set(\"spark.executor.cores\", 8)\n conf.set(\"spark.python.worker.memory\", \"2g\")\n conf.set(\"spark.default.parallelism\", 2000)\n conf.set(\"spark.sql.shuffle.partitions\", 2000)\n conf.set(\"spark.broadcast.blockSize\", 1024) \n conf.set(\"spark.shuffle.file.buffer\", '512k')\n conf.set(\"spark.speculation\", True)\n conf.set(\"spark.speculation.quantile\", 0.98)\n\n spark = SparkSession \\\n .builder \\\n .appName(\"hgongjing2_hdfs_to_hbase\") \\\n .config(conf = conf) \\\n .enableHiveSupport() \\\n .getOrCreate() \n \n return spark\n \ndef spark_data_flow(): \n raw_rdd = spark.sparkContext.textFile(\n (\"{path}/\"\n \"all_company_feature/\"\n \"{version}\").format(path=IN_PATH,\n version=RELATION_VERSION)\n ).map(\n json.loads\n ).cache()\n\n data = raw_rdd.take(1)\n columns = sorted(data[0].keys())\n \n def get_feature(row):\n row_key = row['bbd_qyxx_id']\n row_data = [json.dumps(row[k], ensure_ascii=False) \n if k != 'bbd_qyxx_id' \n and k != 'company_name' \n else row[k]\n for k in columns]\n row_data.insert(0, row_key)\n \n return '\\t'.join(row_data)\n \n prd_rdd = raw_rdd.map(\n get_feature\n ).coalesce(\n 500\n )\n \n return prd_rdd\n\ndef run():\n prd_rdd = spark_data_flow()\n \n os.system(\n (\"hadoop fs -rmr \" \n \"{path}/\"\n \"hdfs_to_hbase\"\n \"/{version}\").format(path=OUT_PATH, \n version=RELATION_VERSION))\n \n prd_rdd.saveAsTextFile(\n (\"{path}/\"\n \"hdfs_to_hbase\"\n \"/{version}\").format(path=OUT_PATH,\n version=RELATION_VERSION),\n compressionCodecClass='org.apache.hadoop.io.compress.GzipCodec')\n\nif __name__ == '__main__':\n #中间结果版本\n RELATION_VERSION = sys.argv[1]\n #输入路径\n IN_PATH = \"/user/antifraud/hongjing2/dataflow/step_four/tid/prd/\"\n #输出路径\n OUT_PATH = \"/user/antifraud/hongjing2/dataflow/step_four/tid/prd/\"\n \n spark = get_spark_session()\n \n run()", "sub_path": "src/step_four/tid/prd/hdfs_to_hbase.py", "file_name": "hdfs_to_hbase.py", "file_ext": "py", "file_size_in_byte": 2710, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "pyspark.conf.SparkConf", "line_number": 20, "usage_type": "call"}, {"api_name": "pyspark.sql.SparkSession.builder.appName", "line_number": 34, "usage_type": "call"}, {"api_name": "pyspark.sql.SparkSession.builder", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pyspark.sql.SparkSession", "line_number": 34, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 50, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 58, "usage_type": "call"}, {"api_name": "os.system", "line_number": 78, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 94, "usage_type": "attribute"}]} +{"seq_id": "92049337", "text": "#!/usr/bin/env python\n\"\"\"\nTest the creation of redundant features (see sklearn source)\n\nThe structure of X is columns of [informative, redundant, nuisance] features\n\"\"\"\n\nimport numpy as np\nfrom scipy import stats\nfrom sklearn.preprocessing import MinMaxScaler\nfrom sympy import symbols\n\nfrom synthetic_data.synthetic_data import (generate_redundant_features,\n make_tabular_data,\n transform_to_distribution)\n\nnp.random.seed(111)\nnp.set_printoptions(precision=11)\nseed = 1234\n\n\ndef test_redundant():\n # define symbols\n x1, x2 = symbols(\"x1 x2\")\n\n # define expression\n expr = x1\n\n # define mapping from symbols to column of X\n col_map = {x1: 0, x2: 1}\n\n # baseline 2D data, no noise\n cov = np.array([[1.0, 0.0], [0.0, 1.0]])\n\n # generate synthetic data with 2 redundant columns\n seed = 1234\n generator = np.random.RandomState(seed)\n\n n_samples = 3\n n_informative = 2\n n_redundant = 2\n X, _, _, _ = make_tabular_data(\n n_samples=n_samples,\n n_informative=n_informative,\n n_redundant=n_redundant,\n n_nuisance=0,\n cov=cov,\n col_map=col_map,\n expr=expr,\n p_thresh=0.5,\n # random_state=generator,\n seed=seed,\n )\n print(\"in test results for X - \")\n print(X)\n\n # replicate the redundant features\n # replicate the random state - initialize, run multivariate...\n generator = np.random.RandomState(seed)\n means = np.zeros(n_informative)\n mvnorm = stats.multivariate_normal(mean=means, cov=cov)\n x = mvnorm.rvs(n_samples, random_state=seed)\n norm = stats.norm()\n x_cont = norm.cdf(x)\n\n # this duplicates the generate_redundant_features function\n B = 2 * generator.rand(n_informative, n_redundant) - 1\n print(\"in test - B\")\n print(B)\n # x_cont = X[:, :n_informative]\n print(\"in test - x\")\n print(x_cont)\n\n x_redundant = np.dot(x_cont, B)\n\n scaler = MinMaxScaler(feature_range=[-1, 1])\n x_redundant_scaled = scaler.fit_transform(x_redundant)\n print(\" - scaled - \")\n print(x_redundant_scaled)\n\n x_slice_redundant = X[:, -n_redundant:]\n\n # print(\"in test script - x_redundant\")\n # print(x_redundant)\n\n # check that they match\n assert np.allclose(x_redundant_scaled, x_slice_redundant, rtol=1e-05, atol=1e-08)\n", "sub_path": "tests/test_redundant.py", "file_name": "test_redundant.py", "file_ext": "py", "file_size_in_byte": 2374, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "numpy.random.seed", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 17, "usage_type": "attribute"}, {"api_name": "numpy.set_printoptions", "line_number": 18, "usage_type": "call"}, {"api_name": "sympy.symbols", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 37, "usage_type": "attribute"}, {"api_name": "synthetic_data.synthetic_data.make_tabular_data", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 59, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 60, "usage_type": "call"}, {"api_name": "scipy.stats.multivariate_normal", "line_number": 61, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 61, "usage_type": "name"}, {"api_name": "scipy.stats.norm", "line_number": 63, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 63, "usage_type": "name"}, {"api_name": "numpy.dot", "line_number": 74, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.allclose", "line_number": 87, "usage_type": "call"}]} +{"seq_id": "258254065", "text": "import MySQLdb\nimport os.path\n\ndef guardaRelatorio():\n\tdb = MySQLdb.connect(host=\"localhost\",user=\"root\",passwd=\"1234\",db=\"pedidos\")\n\tcur = db.cursor()\n\ttry:\n\t\tnumPedido = input(\"Insira o número do pedido: \")\n\t\tcur.execute(\"select relatorio from pedido where numPedido=%s\", [numPedido])\n\t\trelatorio = cur.fetchall()\n\t\tif relatorio:\n\t\t\tfor row in relatorio:\n\t\t\t\tif row[0] :\n\t\t\t\t\tpath = \"./relatorios/\"\n\t\t\t\t\tfile_name = path + numPedido + \".txt\"\n\t\t\t\t\tfile = open(file_name, \"w\")\n\t\t\t\t\tpedido = \"Pedido:\" + numPedido + \"\\n\"\n\t\t\t\t\tfile.write(pedido)\n\t\t\t\t\tfile.write(row[0])\n\t\t\t\t\tfile.close()\n\t\t\t\telse :\n\t\t\t\t\tprint(\"Não existe relatório\")\n\t\telse:\n\t\t\tprint(\"Não existe pedido com esse número\")\n\texcept(MySQLdb.Error, MySQLdb.Warning) as e:\n\t\tprint(e)\n\tdb.close()\n\n\ndef inserePedido():\n\tdb = MySQLdb.connect(host=\"localhost\",user=\"root\",passwd=\"1234\",db=\"pedidos\")\n\tcur = db.cursor()\n\ttry:\n\t\tdata = input(\"Insira a data com o seguinte formato DDMMYYYY :\")\n\t\thora = input(\"Insira a hora com o seguinte formato HHMM :\")\n\t\tnumDoente = input(\"Insira o número do doente:\")\n\t\tnumProcesso = input(\"Insira o número do processo:\")\n\t\tmorada = input(\"Insira a morada:\")\n\t\ttelefone = input(\"Insira o número de telefone:\")\n\t\tnumEpisodio = input(\"Insira o número do episódio:\")\n\t\ttipoEpisodio = input(\"Insira o tipo de episódio:\")\n\t\tinfoClinica = input(\"Insira, caso desejável, qualquer informação clínica:\")\n\t\tif infoClinica == \"\":\n\t\t\tinfoClinica = None\n\t\tquery = \"insert into pedido values (null,%s,%s,%s,%s,%s,%s,%s,%s, %s, null, 1)\"\n\t\tcur.execute(query,(data, hora, numDoente, numProcesso, morada, telefone, numEpisodio, tipoEpisodio, infoClinica))\n\t\tdb.commit()\n\t\tprint(\"O pedido foi inserido com sucesso.\")\n\texcept(MySQLdb.Error, MySQLdb.Warning) as e:\n\t\tprint(e)\n\tdb.close()\n\ndef updatePedido():\n\tdb = MySQLdb.connect(host=\"localhost\",user=\"root\",passwd=\"1234\",db=\"pedidos\")\n\tvirgula = False\n\tcontador = 0\n\tpar = list()\n\tcur = db.cursor()\n\ttry:\n\t\tquery = \"update pedido set \"\n\t\tnumPedido = input(\"Insira o número do pedido a modificar: \")\n\t\tcur.execute(\"select * from pedido where numPedido=%s\", [numPedido])\n\t\tlista = cur.fetchall()\n\t\tif lista:\n\t\t\tprint(\"De seguida, insira os parâmetros que deseja modificar, ou clique Enter se não quiser alterar esse parâmetro.\")\n\t\t\tdata = input(\"Insira a data com o seguinte formato DD-MM-YYYY :\")\n\t\t\thora = input(\"Insira a hora com o seguinte formato HH:MM :\")\n\t\t\tmorada = input(\"Insira a morada:\")\n\t\t\ttelefone = input(\"Insira o número de telefone:\")\n\t\t\ttipoEpisodio = input(\"Insira o tipo de episódio:\")\n\t\t\tinfoClinica = input(\"Insira qualquer informação clínica:\")\n\t\t\tif data:\n\t\t\t\tvirgula = True\n\t\t\t\tcontador = contador + 1\n\t\t\t\tquery = query + \" data=%s \"\n\t\t\t\tpar.insert(contador, data)\n\t\t\tif hora:\n\t\t\t\tcontador = contador + 1\n\t\t\t\tif virgula == True:\n\t\t\t\t\tquery = query + \",\"\n\t\t\t\tquery = query + \" hora=%s \"\n\t\t\t\tvirgula = True\n\t\t\t\tpar.insert(contador, hora)\n\t\t\tif morada:\n\t\t\t\tcontador = contador + 1\n\t\t\t\tif virgula == True:\n\t\t\t\t\tquery = query + \",\"\n\t\t\t\tquery = query + \" morada=%s \"\n\t\t\t\tvirgula = True\n\t\t\t\tpar.insert(contador, morada)\n\t\t\tif telefone:\n\t\t\t\tcontador = contador + 1\n\t\t\t\tif virgula == True:\n\t\t\t\t\tquery = query + \",\"\n\t\t\t\tquery = query + \" telefone=%s\"\n\t\t\t\tvirgula = True\n\t\t\t\tpar.insert(contador, telefone)\n\t\t\tif tipoEpisodio:\n\t\t\t\tcontador = contador + 1\n\t\t\t\tif virgula == True:\n\t\t\t\t\tquery = query + \",\"\n\t\t\t\tquery = query + \" tipoEpisodio=%s\"\n\t\t\t\tvirgula = True\n\t\t\t\tpar.insert(contador, tipoEpisodio)\n\t\t\tif infoClinica:\n\t\t\t\tcontador = contador + 1\n\t\t\t\tif virgula == True:\n\t\t\t\t\tquery = query + \",\"\n\t\t\t\tquery = query + \" infoClinica=%s\"\n\t\t\t\tvirgula = True\n\t\t\t\tpar.insert(contador, infoClinica)\n\t\t\t\t\n\t\t\tquery = query + \" , estado = 2 where numPedido=%s\"\n\n\t\t\tif contador == 1:\n\t\t\t\tcur.execute(query, (par[0], numPedido))\n\t\t\telif contador == 2:\n\t\t\t\tcur.execute(query, (par[0], par[1], numPedido))\n\t\t\telif contador == 3:\n\t\t\t\tcur.execute(query, (par[0], par[1], par[2], numPedido))\n\t\t\telif contador == 4:\n\t\t\t\tcur.execute(query, (par[0], par[1], par[2], par[3], numPedido))\n\t\t\telif contador == 5:\n\t\t\t\tcur.execute(query, (par[0], par[1], par[2], par[3], par[4], numPedido))\n\t\t\telif contador == 6:\n\t\t\t\tcur.execute(query, (par[0], par[1], par[2], par[3], par[4], par[5], numPedido))\n\t\t\telif contador == 0:\n\t\t\t\tprint(\"Não foram efetuadas quaisquer alterações\")\n\t\t\tif contador != 0 :\n\t\t\t\tprint(\"Os parâmetros foram alterados com sucesso\")\n\t\t\tdb.commit()\n\n\t\t\tupd = \"insert into listaenviar values(null, %s, %s) \"\n\t\t\tcur.execute(upd, (numPedido, 2))\n\t\t\tdb.commit()\n\t\telse:\n\t\t\tprint(\"O número de pedido não existe.\")\n\t\t\n\texcept(MySQLdb.Error, MySQLdb.Warning) as e:\n\t\tprint(e)\n\tdb.close()\n\ndef cancelaPedido():\n\tdb = MySQLdb.connect(host=\"localhost\",user=\"root\",passwd=\"1234\",db=\"pedidos\")\n\tcur = db.cursor()\n\ttry:\n\t\tnumPedido = input(\"Insira o número do pedido a cancelar: \")\n\t\tcur.execute(\"select * from pedido where numPedido=%s\", [numPedido])\n\t\tlista = cur.fetchall()\n\t\tif lista:\n\t\t\tupdate = \"update pedido set estado=3 where numPedido = %s\"\n\t\t\tcur.execute(update,[numPedido])\n\t\t\tdb.commit()\n\t\t\tupd = \"insert into listaenviar values(null, %s, %s) \"\n\t\t\tcur.execute(upd, (numPedido, 3))\n\t\t\tdb.commit()\n\t\t\tprint(\"O pedido foi cancelado com sucesso.\")\n\t\telse:\n\t\t\tprint(\"O número de pedido não existe.\")\n\texcept(MySQLdb.Error, MySQLdb.Warning) as e:\n\t\tprint(e)\n\tdb.close()\n\ndef main():\n\twhile True:\n\t\tprint(\"=========================\")\n\t\tprint(\"Insira a opção desejada:\")\n\t\tprint(\"1:Inserir pedido\")\n\t\tprint(\"2:Modificar pedido\")\n\t\tprint(\"3:Cancelar pedido\")\n\t\tprint(\"4:Guardar relatório\")\n\t\tprint(\"0:Sair\")\n\t\tprint(\"=========================\")\n\t\topcao = input(\"Opção:\")\n\n\t\tif opcao == '0':\n\t\t\tbreak;\n\t\tif opcao == \"1\":\n\t\t\tinserePedido()\n\t\telif opcao == \"2\":\n\t\t\tupdatePedido()\n\t\telif opcao == \"3\":\n\t\t\tcancelaPedido()\n\t\telif opcao == \"4\":\n\t\t\tguardaRelatorio()\n\t\telif True:\n\t\t\tprint(\"Opção inválida\")\n\nif __name__ ==\"__main__\":\n\tmain()", "sub_path": "python/utilizador.py", "file_name": "utilizador.py", "file_ext": "py", "file_size_in_byte": 5892, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "MySQLdb.connect", "line_number": 5, "usage_type": "call"}, {"api_name": "MySQLdb.Error", "line_number": 25, "usage_type": "attribute"}, {"api_name": "MySQLdb.Warning", "line_number": 25, "usage_type": "attribute"}, {"api_name": "MySQLdb.connect", "line_number": 31, "usage_type": "call"}, {"api_name": "MySQLdb.Error", "line_number": 49, "usage_type": "attribute"}, {"api_name": "MySQLdb.Warning", "line_number": 49, "usage_type": "attribute"}, {"api_name": "MySQLdb.connect", "line_number": 54, "usage_type": "call"}, {"api_name": "MySQLdb.Error", "line_number": 139, "usage_type": "attribute"}, {"api_name": "MySQLdb.Warning", "line_number": 139, "usage_type": "attribute"}, {"api_name": "MySQLdb.connect", "line_number": 144, "usage_type": "call"}, {"api_name": "MySQLdb.Error", "line_number": 160, "usage_type": "attribute"}, {"api_name": "MySQLdb.Warning", "line_number": 160, "usage_type": "attribute"}]} +{"seq_id": "149267431", "text": "import requests\nimport json\nimport psutil\nimport os\n\nif __name__ == \"__main__\":\n pass\n # 获取id\n\n headers = {\n 'Origin': 'http://scxk.nmpa.gov.cn:81',\n 'Referer': 'http://scxk.nmpa.gov.cn:81/xk/',\n 'User-Agent': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/89.0.4389.90 Safari/537.36',\n }\n\n id_list = []\n\n # http://scxk.nmpa.gov.cn:81/xk/itownet/portalAction.do?method=getXkzsList\n\n url = \"http://scxk.nmpa.gov.cn:81/xk/itownet/portalAction.do?method=getXkzsList\"\n for i in range(1, 10):\n page_index = i\n\n data = {\n 'on': 'true',\n 'page': page_index,\n 'pageSize': '15',\n 'productName': '',\n 'conditionType': '1',\n 'applyname': '',\n 'applysn': '',\n }\n json_ids = requests.post(url, data=data, headers=headers).json()\n for dic in json_ids['list']:\n id_list.append(dic['ID'])\n\n print(id_list)\n\n # 获取详情\n items_list = []\n url_info = \"http://scxk.nmpa.gov.cn:81/xk/itownet/portalAction.do?method=getXkzsById\"\n for item_id in id_list:\n data_info = {\n 'id': item_id,\n }\n headers_info = {\n 'Origin': 'http://scxk.nmpa.gov.cn:81',\n 'Referer': 'http://scxk.nmpa.gov.cn:81/xk/itownet/portal/dzpz.jsp?id=' + item_id,\n 'User-Agent': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/89.0.4389.90 Safari/537.36',\n }\n response_item = requests.post(\n url=url_info, data=data_info, headers=headers_info)\n # response_item.encoding = 'utf-8'\n item_info = response_item.json()\n # print(item_info)\n # print(json.dumps(item_info))\n print(u'当前占用:%.4f GB'% (psutil.Process(os.getpid()).memory_info().rss/1024/1024/1024))\n items_list.append(item_info) \n print(u'全部读取到列表时占用:%.4f GB'% (psutil.Process(os.getpid()).memory_info().rss/1024/1024/1024))\n\n with open('items.json', 'w',) as fp:\n json.dump(items_list, fp, ensure_ascii=False)\n print(u'当前占用:%.4f GB'% (psutil.Process(os.getpid()).memory_info().rss/1024/1024/1024))\n\n print(u'当前占用:%.4f GB'% (psutil.Process(os.getpid()).memory_info().rss/1024/1024/1024))\n\n", "sub_path": "scxk.py", "file_name": "scxk.py", "file_ext": "py", "file_size_in_byte": 2340, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "requests.post", "line_number": 33, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 51, "usage_type": "call"}, {"api_name": "psutil.Process", "line_number": 57, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 57, "usage_type": "call"}, {"api_name": "psutil.Process", "line_number": 59, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 59, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 62, "usage_type": "call"}, {"api_name": "psutil.Process", "line_number": 63, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 63, "usage_type": "call"}, {"api_name": "psutil.Process", "line_number": 65, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 65, "usage_type": "call"}]} +{"seq_id": "95209037", "text": "import tornado.web\nimport logging\nimport socket\nimport time\nimport os\n\nfrom tornado.web import asynchronous\n\nfrom cmstats.handlers import BaseHandler\nfrom cmstats import __version__, summary\nfrom cmstats.model.schema.devices import Device\n\nPAGE_TEMPLATE = \"\"\"\n\n \n \n \n CMStatsServer (v%(version)s) | %(hostname)s:%(port)s\n
\n %(status)s\n
\n totalCheckins - %(totalCheckins)s\n
\n lastCheckin - %(lastCheckin)s seconds ago\n
\n lastGAReport - %(lastGAReport)s seconds ago\n
\n freeSpace - %(freeSpace)sMB\n
\n databaseExceptions - %(databaseExceptions)s\n
\n databaseQueueSize - %(databaseQueueSize)s\n \n\n\"\"\"\n\nclass PingHandler(BaseHandler):\n @asynchronous\n def get(self):\n status = '** OK **'\n\n lastGAReport = int(time.time()) - self.health['lastGAReport']\n if lastGAReport > 600:\n status = '** ERROR **'\n\n lastCheckin = int(time.time()) - self.health['lastCheckin']\n if lastCheckin > 600:\n status = '** ERROR **'\n\n totalCheckins = summary.submits\n if totalCheckins < 100:\n status = '** ERROR **'\n\n st = os.statvfs(\"/\")\n freeSpace = (st.f_bsize * st.f_bavail) / (1024**2)\n if freeSpace < 512:\n status = '** ERROR **'\n\n databaseExceptions = summary.databaseExceptions\n if databaseExceptions > 1000:\n status = '** ERROR **'\n\n databaseQueueSize = self.queue('database').qsize()\n if databaseQueueSize > 2000:\n status = '** ERROR **'\n\n content = PAGE_TEMPLATE % {\n 'version': __version__,\n 'hostname': socket.gethostname(),\n 'port': self.application.settings['port'],\n 'status': status,\n 'lastGAReport': lastGAReport,\n 'lastCheckin': lastCheckin,\n 'freeSpace': freeSpace,\n 'totalCheckins': totalCheckins,\n 'databaseExceptions': databaseExceptions,\n 'databaseQueueSize': databaseQueueSize\n }\n\n self.write(content)\n self.finish()\n", "sub_path": "cmstats/handlers/ping.py", "file_name": "ping.py", "file_ext": "py", "file_size_in_byte": 2195, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "cmstats.handlers.BaseHandler", "line_number": 37, "usage_type": "name"}, {"api_name": "time.time", "line_number": 42, "usage_type": "call"}, {"api_name": "time.time", "line_number": 46, "usage_type": "call"}, {"api_name": "cmstats.summary.submits", "line_number": 50, "usage_type": "attribute"}, {"api_name": "cmstats.summary", "line_number": 50, "usage_type": "name"}, {"api_name": "os.statvfs", "line_number": 54, "usage_type": "call"}, {"api_name": "cmstats.summary.databaseExceptions", "line_number": 59, "usage_type": "attribute"}, {"api_name": "cmstats.summary", "line_number": 59, "usage_type": "name"}, {"api_name": "cmstats.__version__", "line_number": 68, "usage_type": "name"}, {"api_name": "socket.gethostname", "line_number": 69, "usage_type": "call"}, {"api_name": "tornado.web.asynchronous", "line_number": 38, "usage_type": "name"}]} +{"seq_id": "280861965", "text": "import pandas as pd\nimport datetime as dt\n\nfrom flask import Flask\nfrom flask_restful import Resource, Api, reqparse\nfrom dataclasses import dataclass\nfrom sklearn.preprocessing import MinMaxScaler\n\napp = Flask(__name__)\napi = Api(app)\n\n\n@dataclass\nclass Fields:\n path = r\"/home/maciek/Data-science-webapp/data/cdr_d.csv\"\n models = ['MSE']\n\n\nclass Data(Resource, Fields):\n path = Fields.path\n\n def get(self):\n df = pd.read_csv(self.path)\n df = df.to_dict()\n return {'data': df}, 200\n\n\nclass Describe(Resource, Fields):\n path = Fields.path\n\n def get(self):\n df = pd.read_csv(self.path)\n df = df.dropna(inplace=True)\n return {'data_description': df.describe().to_dict()}, 200\n\n\nclass Sort(Resource, Fields):\n path = Fields.path\n\n def patch(self):\n df = pd.read_csv(self.path)\n\n parser = reqparse.RequestParser()\n\n parser.add_argument('sort_by', required=True)\n\n args = parser.parse_args()\n\n allowed_dic = {'date': 'Data',\n 'close': 'Zamkniecie',\n 'open': 'Otwarcie',\n 'volume': 'Wolumen',\n 'highest': 'Najwyzszy',\n 'lowest': 'Najnizszy'}\n\n if args['sort_by'] in allowed_dic.keys():\n\n df = df.sort_values(allowed_dic[args['sort_by']])\n\n df.to_csv(self.path, index=False)\n\n return {'sorted_data': df.to_dict()}, 200\n else:\n return {'message': \"Niepoprawna wartość\"}, 400\n\n\nclass Types(Resource, Fields):\n path = Fields.path\n\n def get(self):\n df = pd.read_csv(self.path)\n\n parser = reqparse.RequestParser()\n parser.add_argument('int', required=False)\n parser.add_argument('float', required=False)\n parser.add_argument('bool', required=False)\n\n args = parser.parse_args()\n\n if args['int']:\n return {'selected_dtypes': df.select_dtypes('int64').to_dict()}, 200\n elif args['float']:\n return {'selected_dtypes': df.select_dtypes('float64').to_dict()}, 200\n elif args['bool']:\n return {'selected_dtypes': df.select_dtypes('bool').to_dict()}, 200\n else:\n return {'message': \"Błędny format danych\"}, 400\n\n\nclass Head(Resource, Fields):\n path = Fields.path\n\n def get(self):\n df = pd.read_csv(self.path)\n return {'data_head': df.head().to_dict()}, 200\n\n\nclass Tail(Resource, Fields):\n path = Fields.path\n\n def get(self):\n df = pd.read_csv(self.path)\n return {'data_tail': df.tail().to_dict()}, 200\n\n\nclass Models(Resource, Fields):\n models = Fields.models\n\n def get(self):\n return {'models': self.models}, 200\n\n def put(self):\n parser = reqparse.RequestParser()\n parser.add_argument('model', required=True)\n\n args = parser.parse_args()\n\n if args['model'] in self.models:\n return {'message': f\"Model {args['model']} znajduje się już na liście.\"}, 401\n else:\n self.models.append(args['model'])\n\n return {'models': self.models}, 200\n\n def delete(self):\n parser = reqparse.RequestParser()\n parser.add_argument('model', required=True)\n\n args = parser.parse_args()\n\n if args['model'] in self.models:\n\n self.models.remove(args['model'])\n\n return {'models': self.models}, 200\n else:\n return {'message': f\"Nie można usunąć {args['model']} ponieważ nie ma go w liście.\"}, 404\n\n\nclass Preprocess(Resource, Fields):\n\n def put(self):\n\n df = pd.read_csv(self.path)\n\n parser = reqparse.RequestParser()\n parser.add_argument('new_data_path', required=False)\n parser.add_argument('new_column_path', required=False)\n\n args = parser.parse_args()\n\n new_df = pd.read_csv(args['new_data_path'])\n new_col = args[\"new_column_path\"]\n\n if list(new_df['Data'])[0] in list(df['Data'])[-1]:\n return {'message': f\"Data {new_df['Data'][0]} znajduje się już w datasecie.\"}, 401\n else:\n df = df.append(new_df, ignore_index=True)\n\n if new_col.head(1) in df.columns:\n return {'message': f\"Kolumna {new_col.head(1)} już istnieje.\"}, 401\n else:\n df = df.join(new_col)\n\n df.to_csv(self.path, index=False)\n return {'data': df.to_dict()}, 200\n\n def patch(self):\n df = pd.read_csv(self.path)\n\n parser = reqparse.RequestParser()\n parser.add_argument('from_year', required=False)\n parser.add_argument('to_year', required=False)\n parser.add_argument('convert_date', required=False)\n\n args = parser.parse_args()\n\n if args['from_year']:\n for idx, row in df.iterrows():\n if args['from_year'] not in row[\"Data\"]:\n return {'message': f\"Wprowadzona data {args['from_year']} nie występuje w datasecie.\"}, 404\n else:\n df = df[idx + 1:,:]\n \n df.to_csv(self.path, index=False)\n \n return {'data': df.to_dict()}, 200\n \n if args['to_year']:\n for idx, row in df.iterrows():\n if args['to_year'] not in row['Data']:\n return {'message': f\"Wprowadzona data {args['to_year']} nie występuje w datascie.\"}, 404\n else:\n df = df[:idx - 1,:]\n \n df.to_csv(self.path, index=False)\n \n return {'data': df.to_dict()}, 200\n\n if args['convert_date']:\n date = dt.datetime.strptime()\n\n df['Data'] = pd.to_datetime(df['Data'], format=\"%Y-%m-%d\")\n\n df.to_csv(self.path, index=False)\n\n return {'message': \"Prawidłowo skonwertowano typ str na datetime\", 'data': df.to_dict()}, 200\n\n def delete(self):\n\n df = pd.read_csv(self.path)\n\n parser = reqparse.RequestParser()\n parser.add_argument('date', required=False)\n parser.add_argument('column', required=False)\n\n args = parser.parse_args()\n\n if args['date']:\n if args['date'] in list(df['Data']):\n\n df = df[df['Data'] != args['date']]\n\n df.to_csv(self.path, index=False)\n\n return {'data': df.to_dict()}, 200\n else:\n return {'message': f\"Nie znaleziono daty {args['data']}\"}, 404\n\n if args['column']:\n if args['column'] in df.columns:\n\n df = df.drop(args['column'])\n\n df.to_csv(self.path, index=False)\n\n return {'data': df.to_dict()}, 200\n else:\n return {'message': f\"Nie znaleziono kolumny {args['column']}\"}, 404\n\n\nclass Regression(Resource):\n pass\n\n\napi.add_resource(Data, \"/data\")\napi.add_resource(Types, \"/data/types\")\napi.add_resource(Head, \"/data/head\")\napi.add_resource(Tail, \"/data/tail\")\napi.add_resource(Describe, \"/data/describe\")\napi.add_resource(Preprocess, \"/data/preprocess\")\napi.add_resource(Sort, \"/data/preprocess/sort\")\napi.add_resource(Models, \"/models\")\napi.add_resource(Regression, \"/models/regression\")\n\nif __name__ == '__main__':\n app.run()\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 7306, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "flask.Flask", "line_number": 9, "usage_type": "call"}, {"api_name": "flask_restful.Api", "line_number": 10, "usage_type": "call"}, {"api_name": "dataclasses.dataclass", "line_number": 13, "usage_type": "name"}, {"api_name": "flask_restful.Resource", "line_number": 19, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 23, "usage_type": "call"}, {"api_name": "flask_restful.Resource", "line_number": 28, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 32, "usage_type": "call"}, {"api_name": "flask_restful.Resource", "line_number": 37, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 41, "usage_type": "call"}, {"api_name": "flask_restful.reqparse.RequestParser", "line_number": 43, "usage_type": "call"}, {"api_name": "flask_restful.reqparse", "line_number": 43, "usage_type": "name"}, {"api_name": "flask_restful.Resource", "line_number": 67, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 71, "usage_type": "call"}, {"api_name": "flask_restful.reqparse.RequestParser", "line_number": 73, "usage_type": "call"}, {"api_name": "flask_restful.reqparse", "line_number": 73, "usage_type": "name"}, {"api_name": "flask_restful.Resource", "line_number": 90, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 94, "usage_type": "call"}, {"api_name": "flask_restful.Resource", "line_number": 98, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 102, "usage_type": "call"}, {"api_name": "flask_restful.Resource", "line_number": 106, "usage_type": "name"}, {"api_name": "flask_restful.reqparse.RequestParser", "line_number": 113, "usage_type": "call"}, {"api_name": "flask_restful.reqparse", "line_number": 113, "usage_type": "name"}, {"api_name": "flask_restful.reqparse.RequestParser", "line_number": 126, "usage_type": "call"}, {"api_name": "flask_restful.reqparse", "line_number": 126, "usage_type": "name"}, {"api_name": "flask_restful.Resource", "line_number": 140, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 144, "usage_type": "call"}, {"api_name": "flask_restful.reqparse.RequestParser", "line_number": 146, "usage_type": "call"}, {"api_name": "flask_restful.reqparse", "line_number": 146, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 152, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 169, "usage_type": "call"}, {"api_name": "flask_restful.reqparse.RequestParser", "line_number": 171, "usage_type": "call"}, {"api_name": "flask_restful.reqparse", "line_number": 171, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 201, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 201, "usage_type": "attribute"}, {"api_name": "pandas.to_datetime", "line_number": 203, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 211, "usage_type": "call"}, {"api_name": "flask_restful.reqparse.RequestParser", "line_number": 213, "usage_type": "call"}, {"api_name": "flask_restful.reqparse", "line_number": 213, "usage_type": "name"}, {"api_name": "flask_restful.Resource", "line_number": 242, "usage_type": "name"}]} +{"seq_id": "278116870", "text": "import requests, asyncio\n\n\nasync def down(url):\n r = requests.get(url)\n print('下载 %s,长度为 %d' % (url, len(r.text)))\n\n\nif __name__ == '__main__':\n urls = ['http://www.baidu.com', 'http://www.163.com', 'http://www.sina.cm.cn']\n tasks = [down(i) for i in urls]\n loop = asyncio.get_event_loop()\n loop.run_until_complete(asyncio.wait(tasks))\n loop.close()", "sub_path": "MACHINE/ThreadAndProcecss/SynergeticProcess/demo6.py", "file_name": "demo6.py", "file_ext": "py", "file_size_in_byte": 382, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "requests.get", "line_number": 5, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 12, "usage_type": "call"}, {"api_name": "asyncio.wait", "line_number": 13, "usage_type": "call"}]} +{"seq_id": "435847796", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nThis file is a AditrorecruitCoop spider created on top of the ATSSpider\nscrapy crawl aditrorecruit_coop -a mining_job_id=999 -a iteration=1 -a extract=1 -a url=\"https://www.aditrorecruit.com/External/OJCustomer3/AssignmentList.aspx?guidGroup=c0ecf946-9253-4544-80be-2f1a555d34e1\"\n\nsample url:\n https://www.aditrorecruit.com/External/OJCustomer3/AssignmentList.aspx?guidGroup=c0ecf946-9253-4544-80be-2f1a555d34e1\n\"\"\"\n\nfrom re import compile\nfrom urlparse import urljoin\n\nfrom scrapy.http import Request\nfrom scrapy.selector import Selector\n\nfrom brightcorp.base.atsspiders import ATSSpider\nfrom brightcorp.items import BrightcorpItemLoader\nfrom brightcorp.processors import Prefix, HtmlFormatter, ConvertDateString\n\n\nclass AditrorecruitCoop(ATSSpider):\n\n name = 'aditrorecruit_coop'\n ref_re = compile(\"guid=([A-z0-9-]*)\")\n job_link_re = compile(\"Popmiddle\\('(.*?)'\")\n\n def parse(self, response):\n sel = Selector(response)\n\n jobs = sel.xpath(\"//tr[@class='listitem_odd' or @class='listitem_even']\")\n for job in jobs:\n job_link = job.xpath(\"td[1]/a/@href\").extract()\n if job_link:\n res = self.job_link_re.search(job_link[0])\n if res:\n job_url = urljoin(response.url, res.group(1))\n meta = {\n 'title': job.xpath(\"td[1]//text()\").extract(),\n 'company': job.xpath(\"td[2]/text()\").extract(),\n 'location': job.xpath(\"td[3]/text()\").extract(),\n 'date': job.xpath(\"td[4]/text()\").extract(),\n 'exp_date': job.xpath(\"td[5]/text()\").extract(),\n }\n yield Request(\n url=job_url, meta=meta, callback=self.parse_job_callback()\n )\n\n def parse_job(self, response):\n loader = BrightcorpItemLoader(response=response)\n\n if not hasattr(self, 'logo_url'):\n sel = Selector(response)\n self.logo_url = sel.xpath(\"//div[@id='rightcol-inner']/img/@src\").extract()\n\n loader.add_value('url', response.url)\n loader.add_value('logo_url', self.logo_url)\n loader.add_value('title', response.meta['title'])\n loader.add_value('company', response.meta['company'])\n loader.add_value('location', response.meta['location'])\n loader.add_value('date', response.meta['date'], ConvertDateString(\"%Y-%m-%d\"))\n loader.add_value('expiration_date', response.meta['exp_date'], ConvertDateString(\"%Y-%m-%d\"))\n loader.add_xpath('description', \"//div[@id='rightcol-inner']//h1/following-sibling::node()\", HtmlFormatter())\n loader.add_value('referencenumber', response.url, Prefix(\"%s-\" % self.name), re=self.ref_re)\n yield loader.load_item()\n", "sub_path": "brightcorp/brightcorp/spiders/aditrorecruit_coop.py", "file_name": "aditrorecruit_coop.py", "file_ext": "py", "file_size_in_byte": 2855, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "brightcorp.base.atsspiders.ATSSpider", "line_number": 21, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 24, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 25, "usage_type": "call"}, {"api_name": "scrapy.selector.Selector", "line_number": 28, "usage_type": "call"}, {"api_name": "urlparse.urljoin", "line_number": 36, "usage_type": "call"}, {"api_name": "scrapy.http.Request", "line_number": 44, "usage_type": "call"}, {"api_name": "brightcorp.items.BrightcorpItemLoader", "line_number": 49, "usage_type": "call"}, {"api_name": "scrapy.selector.Selector", "line_number": 52, "usage_type": "call"}, {"api_name": "brightcorp.processors.ConvertDateString", "line_number": 60, "usage_type": "call"}, {"api_name": "brightcorp.processors.ConvertDateString", "line_number": 61, "usage_type": "call"}, {"api_name": "brightcorp.processors.HtmlFormatter", "line_number": 62, "usage_type": "call"}, {"api_name": "brightcorp.processors.Prefix", "line_number": 63, "usage_type": "call"}]} +{"seq_id": "277401962", "text": "import csv\n\nfrom django.core.management import BaseCommand\n\nfrom phr.establecimiento.models import Diresa, Red\n\n\nclass Command(BaseCommand):\n def handle(self, *args, **options):\n with open('redes.csv') as f:\n \"\"\"La estructura del .csv debe ser codigo_red, nombre_red, codigo disa.\"\"\"\n rows = csv.reader(f)\n total = 0\n creados = 0\n actualizados = 0\n for row in rows:\n total += 1\n diresa = None\n try:\n diresa = Diresa.objects.get(codigo=row[2])\n except Diresa.DoesNotExist:\n self.stdout.write('No existe la Diresa con código: {}'.format(row[2]))\n except Diresa.MultipleObjectsReturned:\n self.stdout.write('Múltiples coincidencias para la Diresa con código: {}'.format(row[2]))\n if not diresa:\n continue\n try:\n red = Red.objects.get(codigo=row[0], diresa=diresa)\n red.nombre = row[1]\n red.diresa = diresa\n red.save()\n actualizados += 1\n except Red.DoesNotExist:\n Red.objects.create(\n codigo=row[0],\n nombre=row[1],\n diresa=diresa\n )\n creados += 1\n self.stdout.write('Creando red: {} {} {}'.format(row[0], row[1], row[2]))\n self.stdout.write('Total: {}'.format(total))\n self.stdout.write('Creados: {}'.format(creados))\n self.stdout.write('Actualizados: {}'.format(actualizados))\n", "sub_path": "phr/establecimiento/management/commands/actualiza_red.py", "file_name": "actualiza_red.py", "file_ext": "py", "file_size_in_byte": 1723, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "django.core.management.BaseCommand", "line_number": 8, "usage_type": "name"}, {"api_name": "csv.reader", "line_number": 12, "usage_type": "call"}, {"api_name": "phr.establecimiento.models.Diresa.objects.get", "line_number": 20, "usage_type": "call"}, {"api_name": "phr.establecimiento.models.Diresa.objects", "line_number": 20, "usage_type": "attribute"}, {"api_name": "phr.establecimiento.models.Diresa", "line_number": 20, "usage_type": "name"}, {"api_name": "phr.establecimiento.models.Diresa.DoesNotExist", "line_number": 21, "usage_type": "attribute"}, {"api_name": "phr.establecimiento.models.Diresa", "line_number": 21, "usage_type": "name"}, {"api_name": "phr.establecimiento.models.Diresa.MultipleObjectsReturned", "line_number": 23, "usage_type": "attribute"}, {"api_name": "phr.establecimiento.models.Diresa", "line_number": 23, "usage_type": "name"}, {"api_name": "phr.establecimiento.models.Red.objects.get", "line_number": 28, "usage_type": "call"}, {"api_name": "phr.establecimiento.models.Red.objects", "line_number": 28, "usage_type": "attribute"}, {"api_name": "phr.establecimiento.models.Red", "line_number": 28, "usage_type": "name"}, {"api_name": "phr.establecimiento.models.Red.DoesNotExist", "line_number": 33, "usage_type": "attribute"}, {"api_name": "phr.establecimiento.models.Red", "line_number": 33, "usage_type": "name"}, {"api_name": "phr.establecimiento.models.Red.objects.create", "line_number": 34, "usage_type": "call"}, {"api_name": "phr.establecimiento.models.Red.objects", "line_number": 34, "usage_type": "attribute"}, {"api_name": "phr.establecimiento.models.Red", "line_number": 34, "usage_type": "name"}]} +{"seq_id": "362542103", "text": "import json\nimport os\nfrom datetime import date\nfrom json import JSONDecodeError\n\nimport requests\n\nfrom api import sentry\nfrom api.oracle import create_connection\nfrom api.patrimony.business import select_id_patrimony_by_type, exec_proc_insert_patrimony\nfrom api.user_relationship.business import exec_proc_insert_user_relationship\nfrom api.users.business import exec_proc_insert_account_bank, exec_proc_insert_user_custom, \\\n select_reference_date_relationship\nfrom api.util.string import *\nfrom api.util.user import mount_payload_matera\nfrom decorators.trace_error import capture_sentry_result_list\n\n\nclass ResgisterUser:\n def __init__(self):\n self.url = os.environ.get('MATERA_URL') + \\\n \"/sdbanco/api/v1/pessoas\"\n self.url_account = os.environ.get('MATERA_URL') + \\\n \"/sdconta/api/v2/contas\"\n self.headers = {\n 'X-Access-Key': os.environ.get('MATERA_API_KEY')\n }\n\n self.id_person = None\n self.account = {}\n\n @staticmethod\n @capture_sentry_result_list\n def insert_user_relationship(cursor, id_user, level, type, id_related, payload) -> list:\n\n list_errors = []\n try:\n reference_date = select_reference_date_relationship(cursor)\n except Exception:\n reference_date = date.today()\n\n user_rel_params = [\n id_user,\n level,\n payload[\"nome\"],\n payload[\"nomeResumido\"],\n payload[\"inscricao\"],\n type,\n None,\n id_related,\n 0,\n reference_date,\n ' ',\n None,\n None,\n None\n ]\n\n result_proc = exec_proc_insert_user_relationship(\n cursor, user_rel_params)\n if not result_proc:\n list_errors.append(\n 'Error insert user relationship with id {}'.format(id_related))\n\n return list_errors\n\n @capture_sentry_result_list\n def alter_force_complete_register(self, id_user, payload) -> list:\n\n payload[\"indObrigaCadCompleto\"] = \"S\"\n response = requests.post(\n self.url + \"/{}\".format(id_user), json=payload, headers=self.headers)\n\n list_errors = []\n\n desc_error = 'Error updating user Force Complete Register:'\n if response.status_code not in (200, 201):\n try:\n response_json = response.json()\n error = json.dumps(response_json)\n desc_error += error\n list_errors.append(desc_error)\n except ValueError:\n desc_error += response.text\n list_errors.append(desc_error)\n except Exception:\n desc_error += response.reason\n list_errors.append(desc_error)\n\n return list_errors\n\n @capture_sentry_result_list\n def insert_patrimony(self, cursor, id_user, list_patrimony) -> list:\n list_errors = []\n\n if not list_patrimony:\n list_errors.append(\n \"Error insert patrimony: Patrimony not specified.\")\n return list_errors\n\n for item_patrimony in list_patrimony:\n id_type_patrimony = select_id_patrimony_by_type(\n cursor, item_patrimony['type_patrimony'])\n if id_type_patrimony:\n result_proc = exec_proc_insert_patrimony(\n cursor, id_user, id_type_patrimony[0], item_patrimony[\n 'description_patrimony'],\n item_patrimony['value_patrimony'])\n if not result_proc:\n list_errors.append(\"Error insert patrimony: {}\".format(str({\n 'type_patrimony': item_patrimony['type_patrimony'],\n 'description_patrimony': item_patrimony['description_patrimony'],\n 'value_patrimony': item_patrimony['value_patrimony']\n })))\n else:\n list_errors.append(\n \"Error insert patrimony: type {} nonexistent\".format(item_patrimony['type_patrimony']))\n return list_errors\n\n @staticmethod\n @capture_sentry_result_list\n def insert_bank_account(cursor, id_user, list_bank_acount, name, cpf_person) -> list:\n list_errors = []\n\n if not list_bank_acount:\n list_errors.append(\n \"Error insert bank account: Accounts not specified.\")\n return list_errors\n\n for item_account in list_bank_acount:\n if 'poupança' in item_account['type'].lower():\n if not item_account['check_joint_account']:\n type_account = 2\n else:\n type_account = 12\n else:\n if not item_account['check_joint_account']:\n type_account = 1\n else:\n type_account = 11\n\n result_proc_account = exec_proc_insert_account_bank(cursor,\n id_user, item_account['id_bank'],\n item_account['num_agency'],\n item_account['num_current_account'],\n name, cpf_person, type_account)\n\n if not result_proc_account:\n list_errors.append(\"Error insert bank account: {}\".format(str({\n 'id_bank': item_account['id_bank'],\n 'num_current_account': item_account['num_current_account'],\n 'num_agency': item_account['num_agency']\n })))\n return list_errors\n\n @capture_sentry_result_list\n def insert_custom_person(self, id_user, indicator, investment_agent) -> list:\n\n list_custom = []\n\n if indicator and investment_agent:\n list_custom.append(\n {\"field\": \"INDICADOR\", \"value\": indicator.upper()})\n list_custom.append(\n {\"field\": \"AGENTE_DE_INVESTIMENTO\", \"value\": investment_agent.upper()})\n else:\n list_custom.append(\n {\"field\": \"INDICADOR\", \"value\": \"LEONARDO MARQUES HOJAIJ\"})\n list_custom.append(\n {\"field\": \"AGENTE_DE_INVESTIMENTO\", \"value\": \"LEONARDO MARQUES HOJAIJ\"})\n\n list_errors = []\n\n if list_custom:\n payload = {\n \"idPessoa\": id_user\n }\n for item_custom in list_custom:\n payload[\"campo\"] = item_custom['field']\n payload[\"valor\"] = item_custom['value']\n response = requests.post(\n \"{}/custom\".format(self.url), json=payload, headers=self.headers)\n if response.status_code not in (200, 201):\n list_errors.append(\n 'Error insert custom person: {}'.format(str(item_custom)))\n return list_errors\n\n @capture_sentry_result_list\n def register_person(self, payload) -> list:\n list_errors = []\n response = requests.post(self.url, json=payload, headers=self.headers)\n if response.status_code in (200, 201):\n result = response.json()\n\n if 'data' in result:\n if 'idPessoa' in result['data']:\n self.id_person = result['data']['idPessoa']\n return list_errors\n\n else:\n try:\n response_json = response.json()\n except JSONDecodeError:\n list_errors.append(\n 'Error insert person: {}'.format(response.text))\n return list_errors\n if 'error' in response_json:\n result = []\n for item in response.json()['error']:\n result.append(item['message'])\n\n list_errors.append(\n 'Error insert person: {}'.format(', '.join(result)))\n return list_errors\n\n list_errors.append(\n 'Error insert person: {}'.format(response.reason))\n return list_errors\n\n if not self.id_person:\n list_errors.append('Error insert person: {}'.format(\n 'id person not provisioned'))\n return list_errors\n\n @capture_sentry_result_list\n def insert_account(self, id_person, id_coowner, nome, banker, account_type) -> list:\n list_errors = []\n # opening account\n payload_account = {\n \"nomeUsuario\": nome[:30],\n \"dataCadastro\": convert_date(datetime.now()),\n \"contas\": [\n {\n \"agencia\": \"001\",\n \"numConta\": None,\n \"gerente\": banker,\n \"tipoConta\": 2,\n \"primeiroTitular\": id_person,\n \"tipoAssociacao\": account_type\n }\n ]\n }\n if account_type == 'C' and id_coowner:\n payload_account['contas'][0][\"coTitulares\"] = [{\n \"idTitular\": int(id_coowner)\n }]\n\n response_account = requests.post(\n self.url_account, json=payload_account, headers=self.headers)\n\n if response_account.status_code not in (200, 201):\n lst_error_account = []\n if 'error' in response_account.json():\n for item in response_account.json()['error']:\n lst_error_account.append(item['message'])\n\n msg_result_account = 'Error insert account: {}'.format(\n ', '.join(lst_error_account))\n list_errors.append(msg_result_account)\n else:\n account = response_account.json()\n if 'data' in account and 'contas' in account['data'] and account['data']['contas']:\n account = account['data']['contas'][0]\n self.account = {\n 'agency': account['agencia'],\n 'account': account['numConta']\n }\n else:\n list_errors.append('Error insert account: {}'.format(\n 'account data not provisioned'))\n\n return list_errors\n\n # TODO implementar\n def delete_patrimony(self, cursor, id_person):\n pass\n\n def delete_bank_account(self, cursor, id_person):\n pass\n\n def delete_relationship(self, cursor, id_person):\n pass\n\n def delete_custom(self, cursor, id_person):\n pass\n\n def get_account(self, id_person):\n list_errors = []\n response = requests.get(self.url_account, params={\n 'idCliente': id_person}, headers=self.headers)\n if response.status_code in (200, 201):\n response_json = response.json()\n for account in response_json:\n self.account = {\n 'agency': account['agencia'],\n 'account': account['numConta']\n }\n return list_errors\n else:\n list_errors.append(\n 'Error get user account: {}'.format(response.text))\n\n return list_errors\n\n # -------------------\n\n def register_joint_account(self, data):\n\n result = {}\n list_errors = []\n owner = data.get('owner')\n coowner = data.get('coowner')\n\n cod_adviser_owner, cod_adviser_coowner = owner.get(\n 'cod_advisor'), coowner.get('cod_advisor')\n flag_error = owner.get('flag_error')\n flag_error_coowner = coowner.get('flag_error')\n id_person_owner, id_person_coowner = cod_adviser_owner, cod_adviser_coowner\n\n try:\n session_oracle = create_connection()\n cursor = session_oracle.cursor()\n\n payload_owner, banker_owner, indicator_owner, investment_agent_owner = mount_payload_matera(\n owner)\n payload_coowner, banker_coowner, indicator_coowner, investment_agent_coowner = mount_payload_matera(\n coowner)\n\n flag_error = 2 if not all(\n [cod_adviser_owner, cod_adviser_coowner]) and flag_error != 2 else flag_error\n\n if cod_adviser_owner is None:\n error = self.register_person(payload_owner)\n if self.id_person:\n id_person_owner = self.id_person\n self.id_person = None\n if error:\n return {'error': error}, 500\n\n if flag_error == 2:\n list_patrimony_owner = owner.get('list_patrimony')\n list_account_bank_owner = owner.get(\"list_account_bank\")\n\n rollback_stack = [\n {\n 'func': self.insert_patrimony,\n 'args': (cursor, id_person_owner, list_patrimony_owner)\n },\n {\n 'func': self.insert_bank_account,\n 'args': (cursor, id_person_owner, list_account_bank_owner, payload_owner['nome'],\n payload_owner['inscricao'])\n },\n {\n 'func': self.insert_custom_person,\n 'args': (id_person_owner, indicator_owner, investment_agent_owner)\n },\n {\n 'func': self.insert_user_relationship,\n 'args': (cursor, id_person_owner, 10, 'J', 1, payload_owner)\n },\n ]\n\n while rollback_stack:\n invoke = rollback_stack.pop(0)\n ex = invoke['func'](*invoke['args'])\n if ex:\n rollback_stack = []\n list_errors.extend(ex)\n\n if not list_errors:\n session_oracle.commit()\n flag_error = 3\n else:\n session_oracle.rollback()\n\n if cod_adviser_coowner is None:\n error = self.register_person(payload_coowner)\n if self.id_person:\n id_person_coowner = self.id_person\n self.id_person = None\n if error:\n return {'error': error}, 500\n\n if flag_error_coowner == 2:\n list_patrimony_coowner = coowner.get('list_patrimony')\n list_account_bank_coowner = coowner.get(\"list_account_bank\")\n\n rollback_stack = [\n {\n 'func': self.insert_patrimony,\n 'args': (cursor, id_person_coowner, list_patrimony_coowner)\n },\n {\n 'func': self.insert_bank_account,\n 'args': (cursor, id_person_coowner, list_account_bank_coowner, payload_owner['nome'],\n payload_owner['inscricao'])\n },\n {\n 'func': self.insert_custom_person,\n 'args': (id_person_coowner, indicator_coowner, investment_agent_coowner)\n },\n {\n 'func': self.insert_user_relationship,\n 'args': (cursor, id_person_coowner, 10, 'J', 1, payload_coowner)\n },\n {\n 'func': self.insert_user_relationship,\n 'args': (cursor, id_person_owner, 2, 'F', id_person_coowner, payload_owner)\n },\n ]\n\n while rollback_stack:\n invoke = rollback_stack.pop(0)\n ex = invoke['func'](*invoke['args'])\n if ex:\n rollback_stack = []\n list_errors.extend(ex)\n\n if not list_errors:\n session_oracle.commit()\n flag_error_coowner = 3\n else:\n session_oracle.rollback()\n\n both_ids = all([id_person_coowner, id_person_owner])\n\n if both_ids and flag_error == 3 and flag_error_coowner == 3:\n\n # force complete register to be false\n list_errors.extend(self.alter_force_complete_register(\n id_person_owner, payload_owner))\n list_errors.extend(self.alter_force_complete_register(\n id_person_coowner, payload_coowner))\n\n if not list_errors:\n flag_error = 4\n\n if both_ids and flag_error == 4:\n # opening account\n errors_account = self.insert_account(id_person_owner, id_person_coowner,\n payload_owner[\"nome\"],\n banker_owner, 'C')\n\n if not errors_account:\n flag_error = 0\n result['account'] = self.account\n else:\n list_errors.extend(errors_account)\n\n elif both_ids and flag_error == 0:\n # get user account info\n errors_account = self.get_account(id_person_owner)\n if not errors_account:\n result['account'] = self.account\n else:\n list_errors.extend(errors_account)\n\n errors_account = self.get_account(id_person_coowner)\n if not errors_account:\n result['account'] = self.account\n else:\n list_errors.extend(errors_account)\n\n if session_oracle:\n session_oracle.close()\n\n if id_person_owner and id_person_coowner:\n result['data'] = {\n 'idPessoa': id_person_owner,\n 'idPessoaCotitular': id_person_coowner\n }\n\n result['flag_error'] = flag_error\n\n if list_errors or flag_error != 0:\n result['error'] = ', '.join(list_errors)\n return result, 500\n\n return result, 201\n except Exception as e:\n sentry.captureException()\n result = {'error': str(e), 'flag_error': flag_error}\n if id_person_owner and id_person_coowner:\n result['data'] = {'idPessoa': id_person_owner,\n 'idPessoaCotitular': id_person_coowner}\n return result, 500\n\n def register_individual_account(self, data):\n\n result = {}\n list_errors = []\n cod_advisor = data.get('cod_advisor')\n flag_error = data.get('flag_error')\n id_person = cod_advisor\n\n try:\n payload, banker, indicator, investment_agent = mount_payload_matera(\n data)\n\n session_oracle = create_connection()\n cursor = session_oracle.cursor()\n\n flag_error = 2 if not cod_advisor and flag_error != 2 else flag_error\n\n if not cod_advisor:\n error = self.register_person(payload)\n if self.id_person:\n id_person = self.id_person\n self.id_person = None\n if error:\n session_oracle.close()\n return {'error': error}, 500\n\n if id_person and flag_error == 2:\n # exec proc patrimony\n list_patrimony = data.get(\"list_patrimony\")\n list_account_bank = data.get(\"list_account_bank\")\n\n rollback_stack = [\n {\n 'func': self.insert_patrimony,\n 'args': (cursor, id_person, list_patrimony)\n },\n {\n 'func': self.insert_bank_account,\n 'args': (cursor, id_person, list_account_bank, payload['nome'], payload['inscricao'])\n },\n {\n 'func': self.insert_custom_person,\n 'args': (id_person, indicator, investment_agent)\n },\n {\n 'func': self.insert_user_relationship,\n 'args': (cursor, id_person, 10, 'J', 1, payload)\n }\n ]\n\n while rollback_stack:\n invoke = rollback_stack.pop(0)\n ex = invoke['func'](*invoke['args'])\n if ex:\n rollback_stack = []\n list_errors.extend(ex)\n\n if not list_errors:\n session_oracle.commit()\n flag_error = 3\n else:\n session_oracle.rollback()\n\n if id_person and flag_error == 3:\n\n # force complete register to be false\n list_errors.extend(\n self.alter_force_complete_register(id_person, payload))\n\n if not list_errors:\n flag_error = 4\n\n if id_person and flag_error == 4:\n # opening account\n errors_account = self.insert_account(\n id_person, None, payload[\"nome\"], banker, 'N')\n if not errors_account:\n flag_error = 0\n result['account'] = self.account\n else:\n list_errors.extend(errors_account)\n elif id_person and flag_error == 0:\n # get user account info\n errors_account = self.get_account(id_person)\n if not errors_account:\n result['account'] = self.account\n else:\n list_errors.extend(errors_account)\n\n if session_oracle:\n session_oracle.close()\n\n if id_person:\n result['data'] = {\n 'idPessoa': id_person\n }\n\n result['flag_error'] = flag_error\n\n if list_errors or flag_error != 0:\n result['error'] = ', '.join(list_errors)\n return result, 500\n\n return result, 201\n except Exception as e:\n sentry.captureException()\n result = {'error': str(e), 'flag_error': flag_error}\n if id_person:\n result['data'] = {'idPessoa': id_person}\n return result, 500\n\n def register(self, data):\n # Verifica se possui se payload é cadastro conta conjunta\n if data.get('coowner'):\n return self.register_joint_account(data)\n else:\n return self.register_individual_account(data)\n\n def register_individual_account_from_coowner(self, data):\n \n result = {}\n list_errors = []\n cod_advisor = data.get('cod_advisor')\n flag_error = data.get('flag_error')\n id_person = cod_advisor\n\n try:\n payload, banker, indicator, investment_agent = mount_payload_matera(\n data)\n\n session_oracle = create_connection()\n cursor = session_oracle.cursor()\n \n flag_error = None\n\n if id_person:\n\n list_account_bank = data.get(\"list_account_bank\")\n \n flag_error = 3\n if list_account_bank:\n rollback_stack = [\n {\n 'func': self.insert_bank_account,\n 'args': (cursor, id_person, list_account_bank, payload['nome'], payload['inscricao'])\n }\n ]\n\n while rollback_stack:\n invoke = rollback_stack.pop(0)\n ex = invoke['func'](*invoke['args'])\n if ex:\n rollback_stack = []\n list_errors.extend(ex)\n\n if not list_errors:\n session_oracle.commit()\n else:\n session_oracle.rollback()\n flag_error = None\n\n\n if id_person and flag_error == 3:\n # force complete register to be false\n list_errors.extend(\n self.alter_force_complete_register(id_person, payload))\n\n if not list_errors:\n flag_error = 4\n\n if id_person and flag_error == 4:\n # opening account\n errors_account = self.insert_account(\n id_person, None, payload[\"nome\"], banker, 'N')\n if not errors_account:\n flag_error = 0\n result['account'] = self.account\n else:\n list_errors.extend(errors_account)\n\n elif id_person and flag_error == 0:\n # get user account info\n errors_account = self.get_account(id_person)\n if not errors_account:\n result['account'] = self.account\n else:\n list_errors.extend(errors_account)\n\n if session_oracle:\n session_oracle.close()\n\n if id_person:\n result['data'] = {\n 'idPessoa': id_person\n }\n\n result['flag_error'] = flag_error\n\n if list_errors or flag_error != 0:\n result['error'] = ', '.join(list_errors)\n return result, 500\n\n return result, 201\n except Exception as e:\n sentry.captureException()\n result = {'error': str(e), 'flag_error': flag_error}\n if id_person:\n result['data'] = {'idPessoa': id_person}\n return result, 500\n\n def alter(self, data, payload=None):\n id_user = data.get('cod_advisor')\n if id_user:\n if(not payload):\n payload, banker, indicator, investment_agent = mount_payload_matera(\n data)\n payload[\"indObrigaCadCompleto\"] = \"S\"\n\n response = requests.post(\n self.url + \"/{}\".format(id_user), json=payload, headers=self.headers)\n\n if response.status_code in (200, 201):\n return {\n 'message': 'OK'\n }, 201\n else:\n if 'error' in response.json():\n result = []\n for item in response.json()['error']:\n result.append(item['message'])\n\n msg_result = 'Matera: ' + ', '.join(result)\n sentry.captureMessage(msg_result)\n\n return msg_result, 500\n return 'Matera: ' + response.reason, 500\n return 'Cod. Advisor deve ser informado!', 403\n", "sub_path": "Allgoo/andbank-service/api/users/services/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 26917, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "os.environ.get", "line_number": 21, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 23, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 26, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 26, "usage_type": "attribute"}, {"api_name": "api.users.business.select_reference_date_relationship", "line_number": 38, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 40, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 40, "usage_type": "name"}, {"api_name": "api.user_relationship.business.exec_proc_insert_user_relationship", "line_number": 59, "usage_type": "call"}, {"api_name": "decorators.trace_error.capture_sentry_result_list", "line_number": 33, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 71, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 80, "usage_type": "call"}, {"api_name": "decorators.trace_error.capture_sentry_result_list", "line_number": 67, "usage_type": "name"}, {"api_name": "api.patrimony.business.select_id_patrimony_by_type", "line_number": 102, "usage_type": "call"}, {"api_name": "api.patrimony.business.exec_proc_insert_patrimony", "line_number": 105, "usage_type": "call"}, {"api_name": "decorators.trace_error.capture_sentry_result_list", "line_number": 92, "usage_type": "name"}, {"api_name": "api.users.business.exec_proc_insert_account_bank", "line_number": 142, "usage_type": "call"}, {"api_name": "decorators.trace_error.capture_sentry_result_list", "line_number": 121, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 181, "usage_type": "call"}, {"api_name": "decorators.trace_error.capture_sentry_result_list", "line_number": 156, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 191, "usage_type": "call"}, {"api_name": "json.JSONDecodeError", "line_number": 203, "usage_type": "name"}, {"api_name": "decorators.trace_error.capture_sentry_result_list", "line_number": 188, "usage_type": "name"}, {"api_name": "datetime.now", "line_number": 231, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 248, "usage_type": "call"}, {"api_name": "decorators.trace_error.capture_sentry_result_list", "line_number": 225, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 289, "usage_type": "call"}, {"api_name": "api.oracle.create_connection", "line_number": 321, "usage_type": "call"}, {"api_name": "api.util.user.mount_payload_matera", "line_number": 324, "usage_type": "call"}, {"api_name": "api.util.user.mount_payload_matera", "line_number": 326, "usage_type": "call"}, {"api_name": "api.sentry.captureException", "line_number": 482, "usage_type": "call"}, {"api_name": "api.sentry", "line_number": 482, "usage_type": "name"}, {"api_name": "api.util.user.mount_payload_matera", "line_number": 498, "usage_type": "call"}, {"api_name": "api.oracle.create_connection", "line_number": 501, "usage_type": "call"}, {"api_name": "api.sentry.captureException", "line_number": 594, "usage_type": "call"}, {"api_name": "api.sentry", "line_number": 594, "usage_type": "name"}, {"api_name": "api.util.user.mount_payload_matera", "line_number": 616, "usage_type": "call"}, {"api_name": "api.oracle.create_connection", "line_number": 619, "usage_type": "call"}, {"api_name": "api.sentry.captureException", "line_number": 693, "usage_type": "call"}, {"api_name": "api.sentry", "line_number": 693, "usage_type": "name"}, {"api_name": "api.util.user.mount_payload_matera", "line_number": 703, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 707, "usage_type": "call"}, {"api_name": "api.sentry.captureMessage", "line_number": 721, "usage_type": "call"}, {"api_name": "api.sentry", "line_number": 721, "usage_type": "name"}]} +{"seq_id": "309651190", "text": "#\n# Hello World server in Python\n# Binds REP socket to tcp://*:5555\n# Expects b\"Hello\" from client, replies with b\"World\"\n#\n\nimport time\nimport zmq\nimport cv2\nimport sys\nimport os\nimport numpy as np\nimport base64\nimport io\nfrom PIL import Image\nfrom sys import platform\nimport argparse\nimport math\n\ncontext = zmq.Context()\nsocket = context.socket(zmq.REP)\nsocket.bind(\"tcp://*:6555\")\n\n\ndef get_keypoint_string(my_arr):\n my_str = str(int(my_arr[0][0])) + ',' + str(int(my_arr[0][1]))\n for i in range(1,25):\n my_str += ',' + str(int(my_arr[i][0])) + ',' + str(int(my_arr[i][1]))\n return my_str\n\n# Take in base64 string and return PIL image\ndef stringToRGB(base64_string):\n imgdata = base64.b64decode(str(base64_string))\n image = Image.open(io.BytesIO(imgdata))\n #image = image.save(\"geeks.jpg\") \n return cv2.cvtColor(np.array(image),cv2.COLOR_BGR2RGB)\n# convert PIL Image to an RGB image( technically a numpy array ) that's compatible with opencv\n\ntry:\n # Import Openpose (Windows/Ubuntu/OSX)\n dir_path = os.path.dirname(os.path.realpath(__file__))\n try:\n # Windows Import\n if platform == \"win32\":\n # Change these variables to point to the correct folder (Release/x64 etc.)\n sys.path.append(dir_path + '/../../python/openpose/Release');\n os.environ['PATH'] = os.environ['PATH'] + ';' + dir_path + '/../../x64/Release;' + dir_path + '/../../bin;'\n import pyopenpose as op\n else:\n # Change these variables to point to the correct folder (Release/x64 etc.)\n sys.path.append('../../python');\n # If you run `make install` (default path is `/usr/local/python` for Ubuntu), you can also access the OpenPose/python module from there. This will install OpenPose and the python library at your desired installation path. Ensure that this is in your python path in order to use it.\n # sys.path.append('/usr/local/python')\n from openpose import pyopenpose as op\n except ImportError as e:\n print('Error: OpenPose library could not be found. Did you enable `BUILD_PYTHON` in CMake and have this Python script in the right folder?')\n raise e\n\n # Flags\n parser = argparse.ArgumentParser()\n parser.add_argument(\"--image_dir\", default=\"../../../examples/media/\", help=\"Process a directory of images. Read all standard formats (jpg, png, bmp, etc.).\")\n parser.add_argument(\"--no_display\", default=False, help=\"Enable to disable the visual display.\")\n args = parser.parse_known_args()\n\n # Custom Params (refer to include/openpose/flags.hpp for more parameters)\n params = dict()\n params[\"model_folder\"] = \"../../../models/\"\n\n # Add others in path?\n for i in range(0, len(args[1])):\n curr_item = args[1][i]\n if i != len(args[1])-1: next_item = args[1][i+1]\n else: next_item = \"1\"\n if \"--\" in curr_item and \"--\" in next_item:\n key = curr_item.replace('-','')\n if key not in params: params[key] = \"1\"\n elif \"--\" in curr_item and \"--\" not in next_item:\n key = curr_item.replace('-','')\n if key not in params: params[key] = next_item\n\n # Construct it from system arguments\n # op.init_argv(args[1])\n # oppython = op.OpenposePython()\n\n\n ##################### Kalman filter ##################\n '''\n It has 3 input parameters\n dynam_params :the dimension of the state space here is 4\n measure_param : The dimension of the measurement value is 2 here\n control_params:The dimension of the control vector, the default is 0。Since there are no control variables in this model it is also 0\n kalman.processNoiseCov :It is the noise of the model system. The larger the noise, the more unstable the prediction result, and the easier it is to access the predicted value of the model system, and the greater the single-step change. Conversely, if the noise is small, the prediction result is not much different from the previous calculation result.\n kalman.measurementNoiseCov:Is the covariance matrix of the measurement system, the smaller the variance, the closer the prediction result is to the measured value\n '''\n kalman_RAnkle = cv2.KalmanFilter(4,2)\n kalman_RAnkle.measurementMatrix = np.array([[1,0,0,0],[0,1,0,0]],np.float32)\n kalman_RAnkle.transitionMatrix = np.array([[1,0,1,0],[0,1,0,1],[0,0,1,0],[0,0,0,1]], np.float32)\n kalman_RAnkle.processNoiseCov = np.array([[1,0,0,0],[0,1,0,0],[0,0,1,0],[0,0,0,1]], np.float32) * 1e-4\n kalman_RAnkle.measurementNoiseCov = np.array([[1,0],[0,1]], np.float32) * 0.005\n kalman_RAnkle.errorCovPost = np.array([[1,0],[0,1]], np.float32) * 1\n\n kalman_RToe = cv2.KalmanFilter(4,2)\n kalman_RToe.measurementMatrix = np.array([[1,0,0,0],[0,1,0,0]],np.float32)\n kalman_RToe.transitionMatrix = np.array([[1,0,1,0],[0,1,0,1],[0,0,1,0],[0,0,0,1]], np.float32)\n kalman_RToe.processNoiseCov = np.array([[1,0,0,0],[0,1,0,0],[0,0,1,0],[0,0,0,1]], np.float32) * 1e-4\n kalman_RToe.measurementNoiseCov = np.array([[1,0],[0,1]], np.float32) * 0.005\n kalman_RToe.errorCovPost = np.array([[1,0],[0,1]], np.float32) * 1\n\n ori_RAnkle = np.array([[0],[0]],np.float32)\n pre_RAnkle = np.array([[0],[0]],np.float32)\n ori_RToe = np.array([[0],[0]],np.float32)\n pre_RToe = np.array([[0],[0]],np.float32)\n\n # Starting OpenPose\n opWrapper = op.WrapperPython()\n opWrapper.configure(params)\n opWrapper.start()\n\n\n while True:\n # Wait for next request from client\n message = socket.recv()\n\n\n message_string = message.decode(\"utf-8\")\n frame = stringToRGB(message_string)\n #print(image1)\n print(frame.shape)\n\n if(frame.shape[0]!=480 or frame.shape[1]!=640):\n print(\"image is not in right size , width should be 960 and height 540\")\n break\n\n frame_crop = frame.copy()\n datum = op.Datum()\n datum.cvInputData = frame_crop\n opWrapper.emplaceAndPop([datum])\n\n cv2_img = datum.cvOutputData.copy()\n #cv2.putText(cv2_img, \"FPS: %f\" % (1.0/(time.time() - fps_time)), (30, 30), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 255, 0), 2)\n #cv2.imshow('frame', cv2_img)\n #cv2.waitKey(1)\n #fps_time = time.time()\n\n #print(\"Body keypoints: \\n\" + str(datum.poseKeypoints))\n #print(datum.poseKeypoints.dtype) # float32\n #print(datum.poseKeypoints.shape) # (people_num, 25, 3)\n\n if (len(datum.poseKeypoints.shape) == 3): # detect human\n key_points = datum.poseKeypoints[0]\n\n R_deg = '999'\n L_deg = '999'\n\n R_ankel_x = key_points[11][0]\n R_ankel_y = key_points[11][1]\n R_toe_x = key_points[22][0]\n R_toe_y = key_points[22][1]\n\n L_ankel_x = key_points[14][0]\n L_ankel_y = key_points[14][1]\n L_toe_x = key_points[19][0]\n L_toe_y = key_points[19][1]\n\n\n R_hip_x = key_points[9][0]\n R_hip_y = key_points[9][1]\n R_knee_x = key_points[10][0]\n R_knee_y = key_points[10][1]\n\n L_hip_x = key_points[12][0]\n L_hip_y = key_points[12][1]\n L_knee_x = key_points[13][0]\n L_knee_y = key_points[13][1]\n \n R_leg_len = 0\n L_leg_len = 0\n\n ####################### Kalman filter #######################\n # if(key_points[11][0] != 0 or key_points[11][1] != 0):\n # ori_RAnkle = np.array([[key_points[11][0]],[key_points[11][1]]], np.float32)\n # kalman_RAnkle.correct(ori_RAnkle)\n # pre_RAnkle = kalman_RAnkle.predict()\n \n # if(key_points[22][0] != 0 or key_points[22][1] != 0):\n # ori_RToe = np.array([[key_points[22][0]],[key_points[22][1]]], np.float32)\n # kalman_RToe.correct(ori_RToe)\n # pre_RToe = kalman_RToe.predict()\n\n # key_points[11][0] = pre_RAnkle[0,0]\n # key_points[11][1] = pre_RAnkle[1,0]\n # key_points[22][0] = pre_RToe[0,0]\n # key_points[22][1] = pre_RToe[1,0]\n\n # R_ankel_x = pre_RAnkle[0,0]\n # R_ankel_y = pre_RAnkle[1,0]\n # R_toe_x = pre_RToe[0,0]\n # R_toe_y = pre_RToe[1,0]\n\n keypoint_str = get_keypoint_string(datum.poseKeypoints[0])\n #print(keypoint_str)\n\n if ((R_toe_x==0.0 and R_toe_y==0.0) or (R_ankel_x==0.0 and R_ankel_y==0.0)): # No ankle or thumb detected\n #print('deg = 999')\n pass\n else:\n if (abs(R_toe_x - R_ankel_x) < 1): # If x1 is too close to x2, it means vertical\n R_deg = '0'\n #print('zero!!')\n elif (abs(R_toe_y - R_ankel_y) < 1):\n if R_ankel_x > R_toe_x:\n R_deg = '90'\n else:\n R_deg = '-90'\n else:\n c = math.sqrt((R_toe_x - R_ankel_x)**2 + ((R_toe_y - R_ankel_y)**2))\n b = abs((R_toe_y - R_ankel_y))\n #print(str(c) + ', ' + str(b))\n R_deg = int(math.degrees(math.acos((b/c)))) # Adjacent / hypotenuse\n if (R_deg > 0) and (R_deg < 90):\n if R_ankel_x > R_toe_x: # turn right\n R_deg = str(R_deg)\n else: # turn left\n R_deg = str(-R_deg)\n #print(R_deg)\n\n if ((L_toe_x==0.0 and L_toe_y==0.0) or (L_ankel_x==0.0 and L_ankel_y==0.0)): # No ankle or thumb detected\n #print('no L_deg')\n pass\n else:\n if (abs(L_toe_x - L_ankel_x) < 1): # If x1 is too close to x2, it means vertical\n L_deg = '0'\n #print('zero!!')\n elif (abs(L_toe_y - L_ankel_y) < 1):\n if L_ankel_x > L_toe_x:\n L_deg = '90'\n else:\n L_deg = '-90'\n else:\n c = math.sqrt((L_toe_x - L_ankel_x)**2 + ((L_toe_y - L_ankel_y)**2))\n b = abs((L_toe_y - L_ankel_y))\n #print(str(c) + ', ' + str(b))\n L_deg = int(math.degrees(math.acos((b/c)))) # adjacent/hypoteneuse\n if (L_deg > 0) and (L_deg < 90):\n if L_ankel_x > L_toe_x: # turn right\n L_deg = str(L_deg)\n else: # turn left\n L_deg = str(-L_deg)\n #print(L_deg)\n #print(str(R_ankel_x) + ' | ' + str(R_ankel_y) + ' | ' + str(R_toe_x) + ' | ' + str(R_toe_y))\n\n R_leg_len = str(int(math.sqrt((R_hip_x - R_knee_x)**2 + (R_hip_y - R_knee_y)**2) + math.sqrt((R_knee_x - R_ankel_x)**2 + (R_knee_y - R_ankel_y)**2)))\n L_leg_len = str(int(math.sqrt((L_hip_x - L_knee_x)**2 + (L_hip_y - L_knee_y)**2) + math.sqrt((L_knee_x - L_ankel_x)**2 + (L_knee_y - L_ankel_y)**2)))\n\n with open('shoe_index.txt', 'r') as f:\n temp = f.readline()\n #print(temp)\n keypoint_str_byte = bytes(keypoint_str, 'ascii') + b',' + bytes(R_deg, 'ascii') + b',' + bytes(L_deg, 'ascii') + b',' + bytes(R_leg_len, 'ascii') + b',' + bytes(L_leg_len, 'ascii') + b',' + bytes(temp.replace('\\n', ''), 'ascii')\n #keypoint_str_byte = bytes(keypoint_str, 'ascii') + b',' + bytes(R_deg, 'ascii') + b',' + bytes(L_deg, 'ascii') + b',' + bytes(R_leg_len, 'ascii') + b',' + bytes(L_leg_len, 'ascii')\n socket.send(keypoint_str_byte)\n else:\n socket.send(b'0' + b',0'*49 + b',999,999,100,100,-1') # total : 25 (x, y), so a total of 50 coordinate values\n time.sleep(.01)\nexcept Exception as e:\n print(e)\n sys.exit(-1)\n", "sub_path": "server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 12149, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "zmq.Context", "line_number": 20, "usage_type": "call"}, {"api_name": "zmq.REP", "line_number": 21, "usage_type": "attribute"}, {"api_name": "base64.b64decode", "line_number": 33, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 34, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 34, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 36, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 36, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 41, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 44, "usage_type": "name"}, {"api_name": "sys.path.append", "line_number": 46, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 47, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 51, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 60, "usage_type": "call"}, {"api_name": "cv2.KalmanFilter", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 96, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 97, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 98, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 99, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 100, "usage_type": "attribute"}, {"api_name": "cv2.KalmanFilter", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 103, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 104, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 105, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 106, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 107, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 109, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 110, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 111, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 112, "usage_type": "attribute"}, {"api_name": "openpose.pyopenpose.WrapperPython", "line_number": 115, "usage_type": "call"}, {"api_name": "openpose.pyopenpose", "line_number": 115, "usage_type": "name"}, {"api_name": "openpose.pyopenpose.Datum", "line_number": 135, "usage_type": "call"}, {"api_name": "openpose.pyopenpose", "line_number": 135, "usage_type": "name"}, {"api_name": "math.sqrt", "line_number": 216, "usage_type": "call"}, {"api_name": "math.degrees", "line_number": 219, "usage_type": "call"}, {"api_name": "math.acos", "line_number": 219, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 240, "usage_type": "call"}, {"api_name": "math.degrees", "line_number": 243, "usage_type": "call"}, {"api_name": "math.acos", "line_number": 243, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 252, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 253, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 263, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 266, "usage_type": "call"}]} +{"seq_id": "24459948", "text": "import moderngl\nfrom demosys import context, geometry, resources\nfrom demosys.opengl import FBO, DepthTexture, Texture2D, samplers\nfrom pyrr import matrix44\n\n\nclass PointLight:\n \"\"\"A point light and its properties\"\"\"\n def __init__(self, position, radius):\n self.matrix = None\n self._position = position\n self.position = position\n self.radius = radius\n\n @property\n def position(self):\n return self._position\n\n @position.setter\n def position(self, pos):\n self._position = pos\n self.matrix = matrix44.create_from_translation(pos)\n\n\nclass DeferredRenderer:\n\n def __init__(self, width, height, gbuffer=None, lightbuffer=None):\n self.ctx = context.ctx()\n\n self.width = width\n self.height = height\n self.size = (width, height)\n self.depth_sampler = samplers.create(\n texture_compare_mode=False,\n min_filter=moderngl.LINEAR, mag_filter=moderngl.LINEAR\n )\n\n # FBOs\n self.gbuffer = gbuffer\n self.lightbuffer = lightbuffer\n\n # Light Info\n self.point_lights = []\n\n # Create geometry buffer if not supplied\n depth_buffer = DepthTexture(self.size)\n\n if not self.gbuffer:\n self.gbuffer = FBO.create_from_textures(\n [\n Texture2D.create(self.size, 4, dtype='f1'),\n Texture2D.create(self.size, 3, dtype='f2'),\n ],\n depth_buffer=depth_buffer,\n )\n\n if not self.lightbuffer:\n self.lightbuffer = FBO.create_from_textures(\n [Texture2D.create(self.size, 4)],\n # depth_buffer=depth_buffer,\n )\n\n # Unit cube for point lights (cube with radius 1.0)\n self.unit_cube = geometry.cube(width=2, height=2, depth=2)\n self.point_light_shader = resources.shaders.get(\"deferred/light_point.glsl\", create=True)\n\n # Debug draw lights\n self.debug_shader = resources.shaders.get(\"deferred/debug.glsl\", create=True)\n\n # Combine shader\n self.combine_shader = resources.shaders.get(\"deferred/combine.glsl\", create=True)\n self.quad = geometry.quad_fs()\n\n def draw_buffers(self, near, far):\n \"\"\"\n Draw framebuffers for debug purposes.\n We need to supply near and far plane so the depth buffer can be linearized when visualizing.\n\n :param near: Projection near value\n :param far: Projection far value\n \"\"\"\n self.ctx.disable(moderngl.DEPTH_TEST)\n\n self.gbuffer.draw_color_layer(layer=0, pos=(0.0, 0.0), scale=(0.25, 0.25))\n self.gbuffer.draw_color_layer(layer=1, pos=(0.5, 0.0), scale=(0.25, 0.25))\n self.gbuffer.draw_depth(near, far, pos=(1.0, 0.0), scale=(0.25, 0.25))\n self.lightbuffer.draw_color_layer(layer=0, pos=(1.5, 0.0), scale=(0.25, 0.25))\n\n def add_point_light(self, position, radius):\n \"\"\"Add point light\"\"\"\n self.point_lights.append(PointLight(position, radius))\n\n def render_lights(self, camera_matrix, projection):\n \"\"\"Render light volumes\"\"\"\n # Draw light volumes from the inside\n self.ctx.enable(moderngl.CULL_FACE)\n self.ctx.front_face = 'cw'\n\n # No depth testing\n self.ctx.disable(moderngl.DEPTH_TEST)\n\n # Enable additive blending\n self.ctx.enable(moderngl.BLEND)\n self.ctx.blend_func = moderngl.ONE, moderngl.ONE\n\n with self.lightbuffer:\n for light in self.point_lights:\n # Calc light properties\n light_size = light.radius\n m_light = matrix44.multiply(light.matrix, camera_matrix)\n # Draw the light volume\n self.point_light_shader.uniform(\"m_proj\", projection.tobytes())\n self.point_light_shader.uniform(\"m_light\", m_light.astype('f4').tobytes())\n self.gbuffer.color_buffers[1].use(location=0)\n self.point_light_shader.uniform(\"g_normal\", 0)\n self.depth_sampler.use(location=1)\n self.gbuffer.depth_buffer.use(location=1)\n self.point_light_shader.uniform(\"g_depth\", 1)\n self.point_light_shader.uniform(\"screensize\", (self.width, self.height))\n self.point_light_shader.uniform(\"proj_const\", projection.projection_constants)\n self.point_light_shader.uniform(\"radius\", light_size)\n self.unit_cube.draw(self.point_light_shader)\n\n self.depth_sampler.release()\n\n self.ctx.disable(moderngl.BLEND)\n self.ctx.disable(moderngl.CULL_FACE)\n\n def render_lights_debug(self, camera_matrix, projection):\n \"\"\"Render outlines of light volumes\"\"\"\n self.ctx.enable(moderngl.BLEND)\n self.ctx.blend_func = moderngl.SRC_ALPHA, moderngl.ONE_MINUS_SRC_ALPHA\n\n for light in self.point_lights:\n m_mv = matrix44.multiply(light.matrix, camera_matrix)\n light_size = light.radius\n self.debug_shader.uniform(\"m_proj\", projection.tobytes())\n self.debug_shader.uniform(\"m_mv\", m_mv.astype('f4').tobytes())\n self.debug_shader.uniform(\"size\", light_size)\n self.unit_cube.draw(self.debug_shader, mode=moderngl.LINE_STRIP)\n\n self.ctx.disable(moderngl.BLEND)\n\n def render_geometry(self, cam_matrix, projection):\n raise NotImplementedError(\"render_geometry() not implemented\")\n\n def combine(self):\n \"\"\"Combine diffuse and light buffer\"\"\"\n self.gbuffer.color_buffers[0].use(location=0)\n self.combine_shader.uniform(\"diffuse_buffer\", 0)\n self.lightbuffer.color_buffers[0].use(location=1)\n self.combine_shader.uniform(\"light_buffer\", 1)\n self.quad.draw(self.combine_shader)\n\n def clear(self):\n \"\"\"clear all buffers\"\"\"\n self.gbuffer.clear()\n self.lightbuffer.clear()\n", "sub_path": "demosys/deferred/renderer.py", "file_name": "renderer.py", "file_ext": "py", "file_size_in_byte": 5923, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pyrr.matrix44.create_from_translation", "line_number": 22, "usage_type": "call"}, {"api_name": "pyrr.matrix44", "line_number": 22, "usage_type": "name"}, {"api_name": "demosys.context.ctx", "line_number": 28, "usage_type": "call"}, {"api_name": "demosys.context", "line_number": 28, "usage_type": "name"}, {"api_name": "demosys.opengl.samplers.create", "line_number": 33, "usage_type": "call"}, {"api_name": "demosys.opengl.samplers", "line_number": 33, "usage_type": "name"}, {"api_name": "moderngl.LINEAR", "line_number": 35, "usage_type": "attribute"}, {"api_name": "demosys.opengl.DepthTexture", "line_number": 46, "usage_type": "call"}, {"api_name": "demosys.opengl.FBO.create_from_textures", "line_number": 49, "usage_type": "call"}, {"api_name": "demosys.opengl.FBO", "line_number": 49, "usage_type": "name"}, {"api_name": "demosys.opengl.Texture2D.create", "line_number": 51, "usage_type": "call"}, {"api_name": "demosys.opengl.Texture2D", "line_number": 51, "usage_type": "name"}, {"api_name": "demosys.opengl.Texture2D.create", "line_number": 52, "usage_type": "call"}, {"api_name": "demosys.opengl.Texture2D", "line_number": 52, "usage_type": "name"}, {"api_name": "demosys.opengl.FBO.create_from_textures", "line_number": 58, "usage_type": "call"}, {"api_name": "demosys.opengl.FBO", "line_number": 58, "usage_type": "name"}, {"api_name": "demosys.opengl.Texture2D.create", "line_number": 59, "usage_type": "call"}, {"api_name": "demosys.opengl.Texture2D", "line_number": 59, "usage_type": "name"}, {"api_name": "demosys.geometry.cube", "line_number": 64, "usage_type": "call"}, {"api_name": "demosys.geometry", "line_number": 64, "usage_type": "name"}, {"api_name": "demosys.resources.shaders.get", "line_number": 65, "usage_type": "call"}, {"api_name": "demosys.resources.shaders", "line_number": 65, "usage_type": "attribute"}, {"api_name": "demosys.resources", "line_number": 65, "usage_type": "name"}, {"api_name": "demosys.resources.shaders.get", "line_number": 68, "usage_type": "call"}, {"api_name": "demosys.resources.shaders", "line_number": 68, "usage_type": "attribute"}, {"api_name": "demosys.resources", "line_number": 68, "usage_type": "name"}, {"api_name": "demosys.resources.shaders.get", "line_number": 71, "usage_type": "call"}, {"api_name": "demosys.resources.shaders", "line_number": 71, "usage_type": "attribute"}, {"api_name": "demosys.resources", "line_number": 71, "usage_type": "name"}, {"api_name": "demosys.geometry.quad_fs", "line_number": 72, "usage_type": "call"}, {"api_name": "demosys.geometry", "line_number": 72, "usage_type": "name"}, {"api_name": "moderngl.DEPTH_TEST", "line_number": 82, "usage_type": "attribute"}, {"api_name": "moderngl.CULL_FACE", "line_number": 96, "usage_type": "attribute"}, {"api_name": "moderngl.DEPTH_TEST", "line_number": 100, "usage_type": "attribute"}, {"api_name": "moderngl.BLEND", "line_number": 103, "usage_type": "attribute"}, {"api_name": "moderngl.ONE", "line_number": 104, "usage_type": "attribute"}, {"api_name": "pyrr.matrix44.multiply", "line_number": 110, "usage_type": "call"}, {"api_name": "pyrr.matrix44", "line_number": 110, "usage_type": "name"}, {"api_name": "moderngl.BLEND", "line_number": 126, "usage_type": "attribute"}, {"api_name": "moderngl.CULL_FACE", "line_number": 127, "usage_type": "attribute"}, {"api_name": "moderngl.BLEND", "line_number": 131, "usage_type": "attribute"}, {"api_name": "moderngl.SRC_ALPHA", "line_number": 132, "usage_type": "attribute"}, {"api_name": "moderngl.ONE_MINUS_SRC_ALPHA", "line_number": 132, "usage_type": "attribute"}, {"api_name": "pyrr.matrix44.multiply", "line_number": 135, "usage_type": "call"}, {"api_name": "pyrr.matrix44", "line_number": 135, "usage_type": "name"}, {"api_name": "moderngl.LINE_STRIP", "line_number": 140, "usage_type": "attribute"}, {"api_name": "moderngl.BLEND", "line_number": 142, "usage_type": "attribute"}]} +{"seq_id": "561164450", "text": "from django.shortcuts import render\nfrom django.http import HttpResponseRedirect\nfrom django.template.response import TemplateResponse\nfrom site_joao.uploads import handle_uploaded_file\nfrom site_joao.models import *\nfrom site_joao.form import *\nfrom django.shortcuts import render, get_object_or_404\n\n\n\n# Create your views here.\n\ndef main(request):\n if verification(request):\n user = True\n try:\n superAdm = Usuario.objects.get(id= request.session['id'], superAdm= 'S')\n adm = True\n except:\n pass\n\n try:\n moderador = Usuario.objects.get(id= request.session['id'], moderador= 'S')\n moderador = True\n except:\n pass\n \n mainCarrosel = Conteudo.objects.filter(status= 'S').order_by('-id')[0]\n carrosel = Conteudo.objects.filter(status= 'S').order_by('-id')[1:3]\n mainCard = Conteudo.objects.filter(status= 'S').order_by('-id')[3:7]\n card = Conteudo.objects.filter(status= 'S').order_by('-id')[7:11]\n conteudo = Conteudo.objects.filter(status= 'S').order_by('-id')[11:]\n\n buscar = request.GET.get('pesquisa')\n if buscar:\n conteudo = Conteudo.objects.filter(titulo__icontains= buscar, status= 'S')\n return TemplateResponse(request, 'aprovar.html', locals())\n \n return TemplateResponse(request, 'home.html', locals())\n\ndef deletarConteudo(request):\n if verification(request):\n try:\n superAdm = Usuario.objects.get(id= request.session['id'], superAdm= 'S')\n id = request.GET.get('id')\n conteudo = Conteudo.objects.get(id=id)\n conteudo.delete()\n return HttpResponseRedirect('/aprovar')\n except:\n return HttpResponseRedirect('/')\n else:\n return HttpResponseRedirect('/')\n\ndef aprovarConteudo(request):\n if verification(request):\n try:\n superAdm = Usuario.objects.get(id= request.session['id'], superAdm= 'S')\n id = request.GET.get('id')\n conteudo = Conteudo.objects.get(id=id)\n conteudo.status = 'S'\n conteudo.save()\n return HttpResponseRedirect('/aprovar')\n except:\n return HttpResponseRedirect('/aprovar')\n else:\n return HttpResponseRedirect('/')\n\ndef conteudo(request):\n if verification(request):\n user = True\n try:\n superAdm = Usuario.objects.get(id= request.session['id'], superAdm= 'S')\n adm = True\n except:\n pass\n\n try:\n moderador = Usuario.objects.get(id= request.session['id'], moderador= 'S')\n moderador = True\n except:\n pass\n \n buscar = request.GET.get('pesquisa')\n if buscar:\n conteudo = Conteudo.objects.filter(titulo__icontains= buscar, status= 'S')\n return TemplateResponse(request, 'aprovar.html', locals())\n\n nome = request.GET.get('nome')\n conteudo = Conteudo.objects.get(titulo= nome)\n if conteudo.status == 'N':\n status = True\n if conteudo.legenda == 'S':\n legenda = 'Contém legenda'\n\n link = conteudo.linkYoutube.split('watch?v=')\n try:\n link = link[0]+'embed/'+link[1]\n except IndexError:\n link = ''\n\n return TemplateResponse(request, 'conteudo.html', locals())\n\ndef moderadores(request):\n if verification(request):\n user = True\n try:\n superAdm = Usuario.objects.get(id= request.session['id'], superAdm= 'S')\n adm = True\n except:\n pass\n\n try:\n moderador = Usuario.objects.get(id= request.session['id'], moderador= 'S')\n moderador = True\n except:\n return HttpResponseRedirect('/')\n \n usuarios = Usuario.objects.all()\n buscar = request.GET.get('pesquisa')\n if buscar:\n usuarios = Usuario.objects.filter(nome__icontains= buscar)\n if not usuarios:\n usuarios = Usuario.objects.filter(login__icontains= buscar)\n\n return TemplateResponse(request, 'moderadores.html', locals())\n\n else:\n return HttpResponseRedirect('/')\n \n return TemplateResponse(request, 'moderadores.html', locals())\n\ndef tornarAdm(request):\n if verification(request):\n try:\n id = request.GET.get('id')\n usuario = Usuario.objects.get(id=id)\n usuario.superAdm = 'S'\n usuario.save()\n return HttpResponseRedirect('/moderadores')\n except:\n return HttpResponseRedirect('/moderadores')\n else:\n return HttpResponseRedirect('/')\n\ndef retirarAdm(request):\n if verification(request):\n try:\n id = request.GET.get('id')\n usuario = Usuario.objects.get(id=id)\n usuario.superAdm = 'N'\n usuario.save()\n return HttpResponseRedirect('/moderadores')\n except:\n return HttpResponseRedirect('/moderadores')\n else:\n return HttpResponseRedirect('/')\n\ndef verification(request):\n try:\n if request.session['id']:\n return True\n except KeyError:\n return False\n\ndef logout(request):\n try:\n del request.session['id']\n except KeyError:\n pass\n return HttpResponseRedirect('/')\n\ndef login(request):\n if verification(request):\n return HttpResponseRedirect('/')\n else:\n buscar = request.GET.get('pesquisa')\n if buscar:\n conteudo = Conteudo.objects.filter(titulo__icontains= buscar, status= 'S')\n return TemplateResponse(request, 'aprovar.html', locals())\n\n if request.method == 'POST':\n login = request.POST.get('Login')\n senha = request.POST.get('Senha')\n \n user = get_object_or_404(Usuario, login= login, senha= senha) \n # user = Usuario.objects.get(login= login, senha= senha)\n if user:\n request.session['id'] = user.id\n return HttpResponseRedirect('/')\n else:\n msg = 'Usuário não Cadsatrado'\n\n if user:\n request.session['id'] = user.id\n return HttpResponseRedirect('/')\n \n return TemplateResponse(request, 'login.html', locals())\n\ndef register(request):\n buscar = request.GET.get('pesquisa')\n if buscar:\n conteudo = Conteudo.objects.filter(titulo__icontains= buscar, status= 'S')\n return TemplateResponse(request, 'aprovar.html', locals())\n\n form = Form_Register()\n if request.method == 'POST':\n form = Form_Register(request.POST)\n if form.is_valid():\n form.save()\n return HttpResponseRedirect('/entrar')\n else:\n msg = 'Usuário já Cadsatrado'\n return TemplateResponse(request, 'register.html', locals())\n\ndef aprovar(request):\n if verification(request):\n user = True\n\n buscar = request.GET.get('pesquisa')\n if buscar:\n conteudo = Conteudo.objects.filter(titulo__icontains= buscar, status= 'N')\n return TemplateResponse(request, 'aprovar.html', locals())\n\n try:\n superAdm = Usuario.objects.get(id= request.session['id'], superAdm= 'S')\n adm = True\n conteudo = Conteudo.objects.all().filter(status= 'N').order_by('-id')\n except:\n return HttpResponseRedirect('/')\n \n try:\n moderador = Usuario.objects.get(id= request.session['id'], moderador= 'S')\n moderador = True\n except:\n pass\n\n else:\n return HttpResponseRedirect('/')\n return TemplateResponse(request, 'aprovar.html', locals())\n\ndef adicionar(request):\n if verification(request):\n user = True\n\n buscar = request.GET.get('pesquisa')\n if buscar:\n conteudo = Conteudo.objects.filter(titulo__icontains= buscar, status= 'S')\n return TemplateResponse(request, 'aprovar.html', locals())\n\n try:\n superAdm = Usuario.objects.get(id= request.session['id'], superAdm= 'S')\n adm = True\n except:\n pass\n\n try:\n moderador = Usuario.objects.get(id= request.session['id'], moderador= 'S')\n moderador = True\n except:\n pass\n \n editar = False\n try:\n id = request.GET.get('id')\n conteudo_existente = Conteudo.objects.get(id=int(id))\n new = FormConteudo(instance=conteudo_existente)\n editar = True\n except:\n new = FormConteudo()\n if request.method == 'POST':\n if editar:\n new = FormConteudo(request.POST, request.FILES, instance=conteudo_existente)\n else:\n new = FormConteudo(request.POST, request.FILES)\n editarArquivo = False\n if request.FILES.get('myFile'):\n editarArquivo = True\n nomeImg = 'imagem.' + str(request.FILES.get('myFile')).split('.')[-1]\n if 'jpg' in nomeImg:\n nomeImg = 'imagem.png'\n tituloArquivo = request.POST.get('titulo')\n\n if new.is_valid():\n conteudo = new.save()\n if editarArquivo:\n upload_imagem = handle_uploaded_file(request.FILES['myFile'], nomeImg, tituloArquivo)\n conteudo.imagem = upload_imagem\n conteudo.status = 'N'\n conteudo.save()\n return HttpResponseRedirect('/')\n else:\n print('Não salvou!!')\n return TemplateResponse(request, 'add.html', locals())\n else:\n return HttpResponseRedirect('/entrar')\n", "sub_path": "site_joao/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 9701, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "django.template.response.TemplateResponse", "line_number": 37, "usage_type": "call"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 39, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 48, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 50, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 52, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 62, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 64, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 66, "usage_type": "call"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 86, "usage_type": "call"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 101, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 116, "usage_type": "call"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 125, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 128, "usage_type": "call"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 130, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 139, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 141, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 143, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 152, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 154, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 156, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 170, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 174, "usage_type": "call"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 179, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 185, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 189, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 195, "usage_type": "call"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 197, "usage_type": "call"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 203, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 210, "usage_type": "call"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 213, "usage_type": "call"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 222, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 229, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 238, "usage_type": "call"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 239, "usage_type": "call"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 248, "usage_type": "call"}, {"api_name": "site_joao.uploads.handle_uploaded_file", "line_number": 286, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 290, "usage_type": "call"}, {"api_name": "django.template.response.TemplateResponse", "line_number": 293, "usage_type": "call"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 295, "usage_type": "call"}]} +{"seq_id": "229621337", "text": "import pandas as pd\nimport matplotlib.pyplot as plt\nfrom sklearn.preprocessing import MinMaxScaler\nimport numpy as np\nwo_test = pd.read_csv('without_test_thomas.csv', engine='python')\nwo_train = pd.read_csv('without_train_thomas.csv', engine='python')\nw_test = pd.read_csv('with_test_thomas.csv', engine='python')\nw_train = pd.read_csv('with_train_thomas.csv', engine='python')\nbitcoin = pd.read_csv('./data/combined.csv', usecols=[5], engine='python')\n\n# plt.plot([1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19], np.array(wo_test)[:,5], label=\"without Google Trends\")\n# plt.plot([1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19], np.array(w_test)[:,5], label=\"with Google Trends\")\n# plt.legend(loc='best')\n# plt.title('Without Google Trends - Train Set')\n# plt.ylabel('Loss')\n# plt.xlabel('Lookback (Hours)')\n# plt.show()\n\n\n##plot bitcoin, google trends)\ndf = pd.read_csv('./data/combined.csv', usecols=[2,3,4,5,6], engine='python')\nscaler = MinMaxScaler(feature_range=(0, 1))\ndf = scaler.fit_transform(df)\nprint(df)\nplt.plot(df[:,0], linewidth=0.2, label=\"Bitcoin Closing Price\")\nplt.plot(df[:,4], linewidth=0.2, label=\"Google Queries\")\nplt.title('Bitcoin price and Google Search Data')\nplt.ylabel('Normalized Data')\nplt.xlabel('Time (Hours)')\nplt.legend(loc='best')\nplt.show()", "sub_path": "main/plot_exploration.py", "file_name": "plot_exploration.py", "file_ext": "py", "file_size_in_byte": 1281, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pandas.read_csv", "line_number": 5, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 6, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 7, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 8, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 9, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 21, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MinMaxScaler", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}]} +{"seq_id": "383693009", "text": "import tensorflow.contrib.eager as tfe\nimport tensorflow.contrib.summary as tfs\nfrom myCode.helper_functions import extract_words_from_tree, compute_max_arity\nfrom myCode.models import *\nfrom tensorflow_trees.definition import Tree\nimport tensorflow as tf\nfrom nltk.translate.bleu_score import corpus_bleu\nimport myCode.shared_POS_words_lists as shared_list\n\ndef train_model(FLAGS, decoder, encoder, input_train, target_train,input_val, target_val,\n optimizer, beta,lamb,clipping,batch_size,n_exp, name,val_all_captions,\n tree_encoder, tree_decoder, final=False):\n\n best_n_it = 0\n best_loss = 100\n best_matched_word=0\n best_matched_pos=0\n best_struct=0\n best_bleu=0\n if final:\n FLAGS.max_iter = 2000\n FLAGS.check_every = 10\n\n #tensorboard\n summary_writer = tfs.create_file_writer(FLAGS.model_dir+name+\"/\" +str(n_exp), flush_millis=1000)\n summary_writer.set_as_default()\n\n with tfs.always_record_summaries():\n for i in range(FLAGS.max_iter):\n loss_struct=0\n loss_value=0\n loss_POS = 0\n loss_word = 0\n\n #shuffle dataset at beginning of each iteration\n len_input = len(input_train) if type(input_train)==list else input_train.shape[0]\n #input_train,target_train = shuffle_data(input_train,target_train,len_input)\n for j in range(0,len_input,batch_size):\n with tfe.GradientTape() as tape:\n\n current_batch_input=input_train[j:j+batch_size]\n current_batch_target = target_train[j:j+batch_size]\n\n # encode and decode datas\n batch_enc = encoder(current_batch_input)\n root_emb = batch_enc.get_root_embeddings() if tree_encoder else batch_enc\n if tree_decoder:\n batch_dec = decoder(encodings=root_emb, targets=current_batch_target,n_it=i)\n # compute global loss\n loss_struct_miniBatch, loss_values_miniBatch = batch_dec.reconstruction_loss()\n loss_value__miniBatch = loss_values_miniBatch[\"POS_tag\"] + loss_values_miniBatch[\"word\"]\n loss_miniBatch = loss_value__miniBatch+loss_struct_miniBatch\n\n #compute minibatch loss\n loss_struct += loss_struct_miniBatch\n loss_value += loss_value__miniBatch\n loss_POS += loss_values_miniBatch[\"POS_tag\"]\n loss_word += loss_values_miniBatch[\"word\"]\n else:\n loss_single_word=0\n hidden = decoder.reset_state(batch_size=current_batch_target.shape[0])\n dec_input = tf.expand_dims([shared_list.word_idx['']] * current_batch_target.shape[0], 1)\n for h in range(1, current_batch_target.shape[1]):\n predictions, hidden = decoder(dec_input, root_emb, hidden)\n loss_single_word +=loss_function (current_batch_target[:, h],predictions)\n dec_input = tf.expand_dims(tf.argmax(current_batch_target[:, h],axis=-1), 1)\n loss_miniBatch = (loss_single_word/int(current_batch_target.shape[1]))\n loss_word += loss_miniBatch\n\n variables = encoder.variables + decoder.variables\n\n #compute h and w norm for regularization\n h_norm= tf.norm(root_emb)\n w_norm=0\n for w in variables:\n norm = tf.norm(w)\n if norm >= 0.001:\n w_norm += norm\n\n # compute gradient\n grad = tape.gradient(loss_miniBatch+ beta*w_norm +lamb*h_norm, variables)\n gnorm = tf.global_norm(grad)\n grad, _ = tf.clip_by_global_norm(grad, clipping, gnorm)\n tfs.scalar(\"norms/grad\", gnorm)\n tfs.scalar(\"norms/h_norm\", h_norm)\n tfs.scalar(\"norms/w_norm\", w_norm)\n\n # apply optimizer on gradient\n optimizer.apply_gradients(zip(grad, variables), global_step=tf.train.get_or_create_global_step())\n\n\n loss_struct /= (int(int(len_input)/batch_size)+1)\n loss_value /= (int(int(len_input)/batch_size)+1)\n loss_POS /= (int(int(len_input)/batch_size)+1)\n loss_word /= (int(int(len_input)/batch_size)+1)\n loss = loss_struct+loss_value\n print(name,\":iterartion\",i,loss,loss_word,loss_POS)\n\n tfs.scalar(\"loss/loss_struc\", loss_struct)\n tfs.scalar(\"loss/loss_value\", loss_value)\n tfs.scalar(\"loss/loss_value_POS\", loss_POS)\n tfs.scalar(\"loss/loss_value_word\", loss_word)\n\n\n # print stats\n if i % FLAGS.check_every == 0:\n #var_to_save = encoder.variables+encoder.weights + decoder.variables+decoder.weights + optimizer.variables()\n #tfe.Saver(var_to_save).save(checkpoint_prefix,global_step=tf.train.get_or_create_global_step())\n\n if not tree_encoder:\n input_val = tf.Variable(input_val)\n input_val = tf.squeeze(input_val)\n\n batch_val_enc = encoder(input_val)\n if tree_encoder:\n batch_val_enc = batch_val_enc.get_root_embeddings()\n\n if tree_decoder:\n batch_val_dec = decoder(encodings=batch_val_enc,targets=target_val,n_it=i)\n loss_struct_val, loss_values_validation = batch_val_dec.reconstruction_loss()\n loss_validation = loss_struct_val + loss_values_validation[\"POS_tag\"]+loss_values_validation[\"word\"]\n tfs.scalar(\"loss/validation/loss_struc\", loss_struct_val)\n tfs.scalar(\"loss/validation/loss_value\", loss_validation)\n tfs.scalar(\"loss/validation/loss_value_POS\", loss_values_validation[\"POS_tag\"])\n tfs.scalar(\"loss/validation/loss_value_word\", loss_values_validation[\"word\"])\n\n print(\"iteration \", i, \" supervised:\\nloss train value is \", loss_word, \" loss train POS is \", loss_POS , \"\\n\",\n \" loss validation word is \", loss_values_validation[\"word\"], \" loss validation POS is \", loss_values_validation[\"POS_tag\"])\n\n #get unsupervised validation loss\n batch_unsuperv = decoder(encodings=batch_val_enc)\n s_avg, v_avg, tot_pos_uns, matched_pos_uns, total_word_uns ,matched_word_uns= \\\n Tree.compare_trees(target_val, batch_unsuperv.decoded_trees)\n pred_sentences = extract_words_from_tree(batch_unsuperv.decoded_trees)\n tfs.scalar(\"overlaps/unsupervised/struct_avg\", s_avg)\n tfs.scalar(\"overlaps/unsupervised/value_avg\", v_avg)\n tfs.scalar(\"overlaps/unsupervised/total_POS\", tot_pos_uns)\n tfs.scalar(\"overlaps/unsupervised/matched_POS\", matched_pos_uns)\n tfs.scalar(\"overlaps/unsupervised/total_words\", total_word_uns)\n tfs.scalar(\"overlaps/unsupervised/matched_words\", matched_word_uns)\n else:\n pred_sentences= decoder.sampling(batch_val_enc,wi=shared_list.word_idx,\n iw=shared_list.idx_word,max_length=int(target_val.shape[1]))\n\n bleu_1 = corpus_bleu(val_all_captions,pred_sentences,weights=(1.0,))\n bleu_2 = corpus_bleu(val_all_captions,pred_sentences,weights=(0.5,0.5))\n bleu_3 = corpus_bleu(val_all_captions,pred_sentences,weights=(1/3,1/3,1/3))\n bleu_4 = corpus_bleu(val_all_captions,pred_sentences,weights=(0.25,0.25,0.25,0.25))\n tfs.scalar(\"bleu/blue-1\", bleu_1)\n tfs.scalar(\"bleu/blue-2\", bleu_2)\n tfs.scalar(\"bleu/blue-3\", bleu_3)\n tfs.scalar(\"bleu/blue-4\", bleu_4)\n\n if tree_decoder:\n print(\"iteration \", i, \" unsupervised:\\n\", matched_pos_uns,\" out of \", tot_pos_uns, \" POS match\",\n \"that is a perc of\", (matched_pos_uns/tot_pos_uns)*100, \" \" ,matched_word_uns, \" out of \",total_word_uns,\n \"word match that is a percentage of \", (matched_word_uns/total_word_uns)*100, \" struct val \", s_avg,\n \" bleu-1 \", bleu_1,\" bleu-2 \", bleu_2,\" bleu-3 \", bleu_3,\" bleu-4 \", bleu_4)\n else:\n print(name,\":iteration \", i,\" bleu-1 \", bleu_1,\" bleu-2 \", bleu_2,\" bleu-3 \", bleu_3,\" bleu-4 \", bleu_4)\n loss_validation=0\n matched_word_uns=0\n matched_pos_uns=0\n s_avg=0\n\n if best_bleu < bleu_1:\n #update best results\n best_bleu = bleu_1\n best_loss = loss_validation\n best_matched_word=matched_word_uns\n best_matched_pos = matched_pos_uns\n best_n_it = i\n best_struct = s_avg\n #predictions\n elif best_loss > loss_validation:\n best_loss=loss_validation\n #else:\n # break\n\n return best_matched_word,best_matched_pos, best_struct,best_bleu, best_n_it\n\n\ndef loss_function(real, pred):\n mask = tf.math.logical_not(tf.math.equal(tf.argmax(real,axis=-1), 0))\n loss_ = tf.nn.softmax_cross_entropy_with_logits_v2(labels=real, logits=pred)\n #loss_ = K.categorical_crossentropy(real, pred, from_logits=True)\n mask = tf.cast(mask, dtype=loss_.dtype)\n loss_ *= mask\n return tf.reduce_mean(loss_)\n\n\ndef validation(input_train, target_train ,input_val, target_val,parameters, FLAGS,input_tree, target_tree, name: str,val_all_captions) :\n\n #open file\n f= open(name+\".txt\",\"ab\", buffering=0)\n\n #compute max_arity\n image_max_arity, input_train, sen_max_arity = compute_max_arity(input_train, input_tree, target_train, target_tree)\n\n #selected actual parameter to try\n i=0\n emb_tree_size = parameters[0][0]\n max_node_count = parameters[1][0]\n max_depth = parameters[2][0]\n cut_arity = parameters[3][0]\n for lamb in parameters[4]:\n for b in parameters[5]:\n beta = b\n hidden_coeff = parameters[6][0]\n learning_rate = parameters[7][0]\n clipping = parameters[8][0]\n batch_size = parameters[9][0]\n batch_size = pow(2,batch_size)\n emb_word_size = emb_tree_size\n for hid in parameters[11]:\n #hidden_word = int(WordValue.representation_shape*hid)\n hidden_word= int(emb_word_size*hid)\n print(hidden_word)\n\n activation = getattr(tf.nn, FLAGS.activation)\n\n decoder, encoder = get_encoder_decoder(emb_tree_size=emb_tree_size,cut_arity=cut_arity,max_arity=max(image_max_arity,\n sen_max_arity),max_node_count=max_node_count,max_depth=max_depth,hidden_coeff=hidden_coeff,\n activation=activation,image_tree=input_tree,sentence_tree=target_tree,emb_word_size=emb_word_size,\n hidden_word=hidden_word)\n\n #train\n optimizer = tf.train.AdamOptimizer(learning_rate=learning_rate)\n\n matched_word_uns,matched_pos_uns, s_avg, bleu, best_n_it= train_model(FLAGS=FLAGS,decoder=decoder,\n encoder=encoder,\n input_train=input_train,target_train=target_train,\n input_val=input_val, target_val=target_val,optimizer=optimizer,\n beta=beta,lamb=lamb,clipping=clipping,batch_size=batch_size,n_exp=i,name=name,\n tree_encoder =not(input_tree==None), tree_decoder = not(target_tree==None),final=False,\n val_all_captions=val_all_captions)\n\n string = \"\\n\" +str(i) +\")models with parameters emb_tree_size \" + str (emb_tree_size) + \" max node count \" + str(max_node_count) + \\\n \" max_depth \" + str(max_depth) + \" cut arity \" + str(cut_arity) + \\\n \" emb_word_size \" + str(emb_word_size) + \" hidden_word_dim \" + str(hidden_word) +\\\n \" lamdda \" + str(lamb) + \" beta \" + str(beta) + \\\n \" hidden coeff \" + str(hidden_coeff) +\" learn rate \" + str(learning_rate) + \" clipping \"+ str(clipping) + \\\n \" batch size \" + str(batch_size) + \" ,matched word unsupervised \" + str(matched_word_uns) +\\\n \" ,matched POS unsupervised \" + str(matched_pos_uns) + \" and struct accuracy \" + str(s_avg) + \\\n \" bleu-1 \"+str(bleu)+\" in \"+ str(best_n_it) + \" itertions\\n\"\n\n f.write(str.encode(string))\n print (\"experiment \" + str(i) + \" out of 27 finished\\n\")\n i+=1\n", "sub_path": "myCode/validation.py", "file_name": "validation.py", "file_ext": "py", "file_size_in_byte": 13184, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "tensorflow.contrib.summary.create_file_writer", "line_number": 25, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 25, "usage_type": "name"}, {"api_name": "tensorflow.contrib.summary.always_record_summaries", "line_number": 28, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 28, "usage_type": "name"}, {"api_name": "tensorflow.contrib.eager.GradientTape", "line_number": 39, "usage_type": "call"}, {"api_name": "tensorflow.contrib.eager", "line_number": 39, "usage_type": "name"}, {"api_name": "tensorflow.expand_dims", "line_number": 62, "usage_type": "call"}, {"api_name": "myCode.shared_POS_words_lists.word_idx", "line_number": 62, "usage_type": "attribute"}, {"api_name": "myCode.shared_POS_words_lists", "line_number": 62, "usage_type": "name"}, {"api_name": "tensorflow.expand_dims", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.argmax", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.norm", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.norm", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.global_norm", "line_number": 82, "usage_type": "call"}, {"api_name": "tensorflow.clip_by_global_norm", "line_number": 83, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 84, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 84, "usage_type": "name"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 85, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 85, "usage_type": "name"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 86, "usage_type": "name"}, {"api_name": "tensorflow.train.get_or_create_global_step", "line_number": 89, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 89, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 99, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 99, "usage_type": "name"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 100, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 100, "usage_type": "name"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 101, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 101, "usage_type": "name"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 102, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 102, "usage_type": "name"}, {"api_name": "tensorflow.Variable", "line_number": 111, "usage_type": "call"}, {"api_name": "tensorflow.squeeze", "line_number": 112, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 122, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 122, "usage_type": "name"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 123, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 123, "usage_type": "name"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 124, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 124, "usage_type": "name"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 125, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 125, "usage_type": "name"}, {"api_name": "tensorflow_trees.definition.Tree.compare_trees", "line_number": 133, "usage_type": "call"}, {"api_name": "tensorflow_trees.definition.Tree", "line_number": 133, "usage_type": "name"}, {"api_name": "myCode.helper_functions.extract_words_from_tree", "line_number": 134, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 135, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 135, "usage_type": "name"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 136, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 136, "usage_type": "name"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 137, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 137, "usage_type": "name"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 138, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 138, "usage_type": "name"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 139, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 139, "usage_type": "name"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 140, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 140, "usage_type": "name"}, {"api_name": "myCode.shared_POS_words_lists.word_idx", "line_number": 142, "usage_type": "attribute"}, {"api_name": "myCode.shared_POS_words_lists", "line_number": 142, "usage_type": "name"}, {"api_name": "myCode.shared_POS_words_lists.idx_word", "line_number": 143, "usage_type": "attribute"}, {"api_name": "myCode.shared_POS_words_lists", "line_number": 143, "usage_type": "name"}, {"api_name": "nltk.translate.bleu_score.corpus_bleu", "line_number": 145, "usage_type": "call"}, {"api_name": "nltk.translate.bleu_score.corpus_bleu", "line_number": 146, "usage_type": "call"}, {"api_name": "nltk.translate.bleu_score.corpus_bleu", "line_number": 147, "usage_type": "call"}, {"api_name": "nltk.translate.bleu_score.corpus_bleu", "line_number": 148, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 149, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 149, "usage_type": "name"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 150, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 150, "usage_type": "name"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 151, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 151, "usage_type": "name"}, {"api_name": "tensorflow.contrib.summary.scalar", "line_number": 152, "usage_type": "call"}, {"api_name": "tensorflow.contrib.summary", "line_number": 152, "usage_type": "name"}, {"api_name": "tensorflow.math.logical_not", "line_number": 184, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 184, "usage_type": "attribute"}, {"api_name": "tensorflow.math.equal", "line_number": 184, "usage_type": "call"}, {"api_name": "tensorflow.argmax", "line_number": 184, "usage_type": "call"}, {"api_name": "tensorflow.nn.softmax_cross_entropy_with_logits_v2", "line_number": 185, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 185, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 187, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 189, "usage_type": "call"}, {"api_name": "myCode.helper_functions.compute_max_arity", "line_number": 198, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 220, "usage_type": "attribute"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 228, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 228, "usage_type": "attribute"}]} +{"seq_id": "504181465", "text": "# -*- coding: utf-8 -*-\nimport re\nimport json\nimport random\nimport time\nimport scrapy\nfrom urllib.parse import quote, unquote\nfrom ZhihuColumn.items import ZhihucolumnItem\n\nempty_word = 'null'\n\n\nclass ZhColumnSpider(scrapy.Spider):\n name = 'zh_column'\n allowed_domains = ['zhihu.com', 'baidu.com']\n start_urls = ['https://www.zhihu.com/']\n\n def parse(self, response):\n file = open(r'./a.txt', 'r+')\n line = file.readlines()\n len_len = len(line)\n for word in range(len_len):\n kw = line[word].replace('(', '').replace(',1)', '').strip()\n print(kw)\n new_url = 'https://www.zhihu.com/search?q=%s&type=column' % quote(kw)\n item = ZhihucolumnItem()\n item['kw'] = kw\n time.sleep(random.random() * 2)\n yield scrapy.Request(url=new_url, callback=self.parse_init_page, meta={'item': item})\n\n def parse_init_page(self, response):\n kw = response.meta['item']['kw']\n text = response.text\n search_hash_id = re.compile(r'search_hash_id=(.*?)&show_all_topics=').findall(text)[0]\n print(search_hash_id)\n hash_url = 'https://www.zhihu.com/api/v4/search_v3?t=column&q={}&correction=1&' \\\n 'offset=5&limit=10&show_all_topics=0&search_hash_id='.format(quote(kw)) + search_hash_id\n par = re.compile('\"(.*?)\"/'\n '

(.*?)
创建者:'\n '(.*?) 关注'\n '(.*?) 文章')\n info = par.findall(text)\n if not info:\n item = response.meta['item']\n item['title'] = empty_word\n item['description'] = empty_word\n item['id_no'] = empty_word\n item['articles_count'] = empty_word\n item['followers'] = empty_word\n item['avatar_url'] = empty_word\n item['creator_name'] = empty_word\n item['creator_url'] = empty_word\n yield item\n else:\n for i in info:\n item = response.meta['item']\n kw = item['kw']\n item['title'] = i[3] if i[3] else empty_word\n item['description'] = i[5] if i[5] else empty_word\n item['id_no'] = str(info.index(i) + 1)\n item['articles_count'] = i[11] if i[11] else empty_word\n item['followers'] = i[9] if i[9] else empty_word\n item['avatar_url'] = i[1] if i[1] else empty_word\n if '' in i[7]:\n item['creator_name'] = re.compile(r'(.*?)').findall(i[7])[0]\n else:\n item['creator_name'] = i[7] if i[7] else empty_word\n item['creator_url'] = i[6] if i[6] else empty_word\n if item['id_no'] == '5':\n yield item\n yield scrapy.Request(url=hash_url, callback=self.parse_ajax, meta={'kw': kw})\n else:\n yield item\n\n def parse_ajax(self, response):\n kw = unquote(re.compile('q=(.*?)&').findall(response.url)[0])\n hash_url = response.url\n # page = int((int(re.compile('offset=(.*?)&limit=10').findall(hash_url)[0]) - 5) / 10)\n # hash_url = hash_url.replace(str(page), '%s' % str(int(5 + (page + 1) * 10)))\n # hash_url = 'https://www.zhihu.com/api/v4/search_v3?t=column&q=%E5%B0%8F%E7%B1%B3&correction=1&' \\\n # 'offset={}&limit=10&show_all_topics=0&search_hash_id='.format(str(5 + init_i * 10)) + hash_id\n print(hash_url)\n print(json.loads(response.text))\n result = json.loads(response.text)\n try:\n if result['paging']['is_end'] == 'false' or 'False':\n for res in result['data']:\n item = ZhihucolumnItem()\n item['kw'] = kw\n item['title'] = res['highlight']['title'] if res['highlight']['title'] else empty_word\n item['description'] = res['highlight']['description'] if res['highlight']['description'] else empty_word\n item['id_no'] = res['object']['id'] if res['object']['id'] else empty_word\n item['articles_count'] = res['object']['articles_count'] if res['object']['articles_count'] else empty_word\n item['followers'] = res['object']['followers'] if res['object']['followers'] else empty_word\n item['avatar_url'] = res['object']['avatar_url'] if res['object']['avatar_url'] else empty_word\n item['creator_name'] = res['object']['author']['name'] if res['object']['author']['name'] else empty_word\n item['creator_url'] = res['object']['author']['url_token'] if res['object']['author']['url_token'] else empty_word\n yield item\n yield scrapy.Request(url=result['paging']['next'], callback=self.parse_ajax)\n except TypeError:\n print(kw + '----finished!!!')\n else:\n # item = ZhihucolumnItem()\n # item['kw'] = kw\n # item['title'] = empty_word\n # item['description'] = empty_word\n # item['id_no'] = empty_word\n # item['articles_count'] = empty_word\n # item['followers'] = empty_word\n # item['avatar_url'] = empty_word\n # item['creator_name'] = empty_word\n # item['creator_url'] = empty_word\n # yield item\n pass\n", "sub_path": "ZhihuColumn/ZhihuColumn/spiders/zh_column.py", "file_name": "zh_column.py", "file_ext": "py", "file_size_in_byte": 6585, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "scrapy.Spider", "line_number": 13, "usage_type": "attribute"}, {"api_name": "urllib.parse.quote", "line_number": 25, "usage_type": "call"}, {"api_name": "ZhihuColumn.items.ZhihucolumnItem", "line_number": 26, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 28, "usage_type": "call"}, {"api_name": "random.random", "line_number": 28, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 29, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 34, "usage_type": "call"}, {"api_name": "urllib.parse.quote", "line_number": 37, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 38, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 72, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 78, "usage_type": "call"}, {"api_name": "urllib.parse.unquote", "line_number": 83, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 83, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 90, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 91, "usage_type": "call"}, {"api_name": "ZhihuColumn.items.ZhihucolumnItem", "line_number": 95, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 106, "usage_type": "call"}]} +{"seq_id": "458773544", "text": "#!/usr/bin/env python\n\n'''\neasysfs.py from https://raw.githubusercontent.com/isaacovercast/easySFS/master/easySFS.py\nmodified by annabel beichman -- november 2018\nso that sites that are 0/1 across a population get excluded\nand so that there isn't an interactive portion of the script that stops it running remotely \nthis script only retains bi-allelic SNPs.\n~~~easySFS.abModified.3.noInteract.Exclude01Sites.HetFiltering.20181121.py~~~\nmodified by meixi lin -- june 2020 \n1. check input function not allowing vcf file has samples not recorded in pop\n#### note: you must have your projection values be in the same order as the populations are in your popMap file ########\n'''\nfrom __future__ import print_function\nimport matplotlib\nmatplotlib.use('PDF')\nfrom collections import Counter\nfrom collections import OrderedDict\nfrom itertools import combinations\nimport pandas as pd\nimport numpy as np\nimport argparse\nimport shutil\nimport gzip\nimport copy\nimport dadi\nimport sys\nimport os\n\nimport matplotlib.pyplot as plt\nfrom matplotlib.backends.backend_pdf import PdfPages\n\n\ndef dadi_preview_projections(dd, pops, ploidy, fold):\n\tmsg = \"\"\"\n\tRunning preview mode. We will print out the results for # of segregating sites\n\tfor multiple values of projecting down for each population. The dadi\n\tmanual recommends maximizing the # of seg sites for projections, but also\n\ta balance must be struck between # of seg sites and sample size.\n\n\tFor each population you should choose the value of the projection that looks\n\tbest and then rerun easySFS with the `--proj` flag.\n\t\"\"\"\n\tprint(msg)\n\tfor pop in pops:\n\t\tprint(pop)\n\t\tseg_sites = {}\n\t\t## Calculate projections for up to 2 x the number of samples,\n\t\t## so this is assuming populations are diploid.\n\t\t## The +1 makes it possible to see preview including all samples per pop\n\t\tfor x in range(2, ploidy*len(pops[pop])+1):\n\t\t\tfs = dadi.Spectrum.from_data_dict(dd, [pop], [x], polarized=fold)\n\t\t\ts = fs.S()\n\t\t\tseg_sites[x] = round(s)\n\t\t\tprint(\"({}, {})\".format(x, round(s)), end=\"\\t\")\n\t\tprint(\"\")\n\t\t## Old way that's a little uglier\n\t\t#for x in range(2,len(pops[pop])):\n\t\t# print(seg_sites[x], end=\"\\t\")\n\t\tprint(\"\\n\")\n\n\ndef dadi_oneD_sfs_per_pop(dd, pops, proj, unfold, outdir, prefix, dtype):\n\tdadi_dir = os.path.join(outdir, \"dadi\")\n\tfsc_dir = os.path.join(outdir, \"fastsimcoal2\")\n\tM_or_D = \"D\" if unfold else \"M\"\n\tfor i, pop in enumerate(pops):\n\t\tprint(\"Doing 1D sfs - {}\".format(pop))\n\t\tdadi_sfs_file = os.path.join(dadi_dir, pop+\"-\"+str(proj[i])+\".sfs\")\n\n\t\tfs = dadi.Spectrum.from_data_dict(dd, [pop], [proj[i]], mask_corners=True, polarized=unfold)\n\n\t\t## Do int bins rather than float\n\t\tif dtype == \"int\":\n\t\t\tdat = np.rint(np.array(fs.data))\n\t\t\tfs = dadi.Spectrum(dat, data_folded=fs.folded, mask=fs.mask, fill_value=0, dtype=int)\n\n\t\tfs.to_file(dadi_sfs_file)\n\n\t\t## Convert each 1D sfs to fsc format\n\t\tfsc_oneD_filename = os.path.join(fsc_dir, pop+\"_{}AFpop0.obs\".format(M_or_D))\n\t\twith open(fsc_oneD_filename, 'w') as outfile:\n\t\t\toutfile.write(\"1 observation\\n\")\n\t\t\toutfile.write(\"\\t\".join([\"d0_\"+str(x) for x in xrange(proj[i]+1)]) + \"\\n\")\n\t\t\t## Grab the fs data from the dadi sfs\n\t\t\twith open(dadi_sfs_file) as infile:\n\t\t\t\toutfile.write(infile.readlines()[1])\n\t\t\t\toutfile.write(\"\\n\")\n\n\ndef dadi_twoD_sfs_combinations(dd, pops, proj, unfold, outdir, prefix, dtype, verbose):\n\tdadi_dir = os.path.join(outdir, \"dadi\")\n\tfsc_dir = os.path.join(outdir, \"fastsimcoal2\")\n\tM_or_D = \"D\" if unfold else \"M\"\n\t## All combinations of pairs of populations\n\tpopPairs = list(combinations(pops, 2))\n\t## All combinations of corresponding projection values\n\t## This is hackish, it really relies on the fact that `combinations`\n\t## is generating combinations in the same order for pops and projs\n\tprojPairs = list(combinations(proj, 2))\n\t## Make a dict for pop indices. this is a mapping of population labels\n\t## to values (ie. {'pop1':1, 'pop2',2}) for labeling the fsc file names\n\tpopidx = {}\n\tfor i, pop in enumerate(pops):\n\t\tpopidx[pop] = i\n\tif verbose: print(\"Population pairs - {}\".format(popPairs))\n\tif verbose: print(\"Projections for each pop pair - {}\".format(projPairs))\n\tfor i, pair in enumerate(popPairs):\n\t\tprint(\"Doing 2D sfs - {}\".format(pair))\n\t\tdadi_joint_filename = os.path.join(dadi_dir, \"-\".join(pair)+\".sfs\")\n\t\tfs = dadi.Spectrum.from_data_dict(dd, list(pair), list(projPairs[i]), polarized=unfold)\n\n\t\t## Do int bins rather than float\n\t\tif dtype == \"int\":\n\t\t\tdat = np.rint(np.array(fs.data))\n\t\t\tfs = dadi.Spectrum(dat, data_folded=fs.folded, mask=fs.mask, fill_value=0, dtype=int)\n\n\t\tfs.to_file(dadi_joint_filename)\n\n\t\t## Convert each 2D sfs to fsc format\n\t\t## NB: FSC joint format file names look like this: _jointMAFpop1_0.obs\n\t\t## Where the first pop specified is listed in the rows and the second pop\n\t\t## specified is listed in the columns.\n\t\tfsc_twoD_filename = os.path.join(fsc_dir, prefix+\"_joint{}AFpop{}_{}.obs\".format(M_or_D, popidx[pair[0]], popidx[pair[1]]))\n\t\twith open(fsc_twoD_filename, 'w') as outfile:\n\t\t\toutfile.write(\"1 observation\\n\")\n\t\t\t## Format column headers (i.e. d0_0 d0_1 d0_2 .. d0_n for deme 0 up to sample size of n)\n\t\t\toutfile.write(\"\\t\" + \"\\t\".join([\"d{}_\".format(popidx[pair[0]]) + str(x) for x in xrange(projPairs[i][1]+1)]) + \"\\n\") \n\n\t\t\t## Format row headers\n\t\t\trow_headers = [\"d{}_\".format(popidx[pair[1]]) + str(x) for x in xrange(projPairs[i][0]+1)]\n\t\t\t## Read in the joint fs from dadi and format it nice for fsc\n\t\t\twith open(dadi_joint_filename) as infile:\n\t\t\t\t## Get the second line of the dadi-style sfs which contains the data\n\t\t\t\trow_data = infile.readlines()[1].split()\n\t\t\t\t## The length of each row is determined by the number of columns which == the size of the projection for pop2\n\t\t\t\t## Have to add 1 to the value of the projection because xrange stops after 'n' elements\n\t\t\t\t## but we want all n+1 elements from 0,1,2,..,n\n\t\t\t\trow_size = projPairs[i][1] + 1\n\t\t\t\t## Slice the row data into evenly sized chunks based on the number of columns\n\t\t\t\trows = [row_data[i:i + row_size] for i in xrange(0, len(row_data), row_size)]\n\t\t\t\t## Sanity check. Make sure the number of rows you got is the same number you're expecting\n\t\t\t\t## to get (# rows should == size of pop0 projection)\n\t\t\t\tif not len(row_headers) == len(rows):\n\t\t\t\t\tprint(\"FSC Joint SFS failed for - {}\".format(pair))\n\t\t\t\t\tprint(\"Row headers - {}\".format(row_headers))\n\t\t\t\t\tprint(\"Row data - {}\".format(rows))\n\t\t\t\t\tprint(\"Len header / data\\n{}\\t{}\".format(len(row_headers), len(rows)))\n\t\t\t\t\treturn\n\t\t\t\telse:\n\t\t\t\t\tpass\n\t\t\t\t## Write out each row to the file\n\t\t\t\tfor i, row_head in enumerate(row_headers):\n\t\t\t\t\toutfile.write(row_head + \"\\t\" + \" \".join(rows[i]) + \"\\n\")\n\n\ndef dadi_multiSFS(dd, pops, proj, unfold, outdir, prefix, dtype):\n\tprint(\"Doing multiSFS for all pops\")\n\tdadi_dir = os.path.join(outdir, \"dadi\")\n\tfsc_dir = os.path.join(outdir, \"fastsimcoal2\")\n\tdadi_multi_filename = os.path.join(dadi_dir, \"-\".join(pops)+\".sfs\")\n\n\t## Get the multiSFS\n\tfs = dadi.Spectrum.from_data_dict(dd, pops, proj, polarized=unfold)\n\n\t## Do int bins rather than float\n\tif dtype == \"int\":\n\t\tdat = np.rint(np.array(fs.data))\n\t\tfs = dadi.Spectrum(dat, data_folded=fs.folded, mask=fs.mask, fill_value=0, dtype=int)\n\n\t## Write out the dadi file\n\tfs.to_file(dadi_multi_filename)\n\t\n\t## Convert to fsc multiSFS format\n\tfsc_multi_filename = os.path.join(fsc_dir, prefix + \"_MSFS.obs\")\n\twith open(fsc_multi_filename, 'w') as outfile:\n\t\toutfile.write(\"1 observations. No. of demes and sample sizes are on next line.\\n\")\n\t\toutfile.write(str(len(pops)) + \"\\t\" + \" \".join([str(x) for x in proj]) + \"\\n\") \n\t\twith open(dadi_multi_filename) as infile:\n\t\t\toutfile.write(infile.readlines()[1])\n\t\t\toutfile.write(\"\\n\")\n\treturn dadi_multi_filename\n\n\ndef dadi_to_momi(infile, outdir=None, verbose=False):\n\ttry:\n\t\timport momi\n\texcept:\n\t\tif verbose: print(\"Install momi to get momi-style sfs conversion as well.\")\n\t\treturn\n\tif not outdir == None:\n\t\tmomi_dir = os.path.join(outdir, \"momi\")\n\t\tif not os.path.exists(momi_dir):\n\t\t\tos.mkdir(momi_dir)\n\t\toutfile = infile.split(\".sfs\")[0] + \"_momi.sfs\"\n\t\toutfile = os.path.join(outdir, outfile.split(\"/\")[-1])\n\telse:\n\t\toutfile = infile + \"_momi.sfs\"\n\n\tdat = open(infile).readlines()\n\t## Get rid of comment lines\n\tdat = [x.strip() for x in dat if not x.startswith(\"#\")]\n\tif not len(dat) == 3:\n\t\traise Exception(\"Malformed dadi sfs {}.\\n Must have 3 lines, yours has {}\".format(infile, len(dat)))\n\n\t## Parse the info line into nsamps per pop (list of ints), folding flag, and pop names list (if they are given)\n\tinfo = dat[0].split()\n\tnsamps = []\n\t## Keep carving off values as long as they cast successfully as int\n\tfor i in info:\n\t\ttry:\n\t\t\tnsamps.append(int(i))\n\t\texcept:\n\t\t\tpass\n\tnsamps = np.array(nsamps)\n\tpops = [x.replace('\"', '') for x in info[len(nsamps)+1:]]\n\tfolded = info[len(nsamps)]\n\tfolded = False if \"un\" in folded else True\n\tif not len(pops) == len(nsamps):\n\t\tprint(\"Number of populations doesn't agree with number of pop names, using generic names.\")\n\t\tpops = [\"pop\"+x for x in range(len(nsamps))]\n\tif verbose: print(\"Info nsamps={} folded={} pops={}\".format(nsamps, folded, pops))\n\n\t## Get mask\n\tmask = list(map(int, dat[2].split()))\n\tif verbose: print(mask)\n\n\t## Get sfs, and reshape based on sample sizes\n\tsfs = list(map(float, dat[1].split()))\n\tif verbose: print(sfs)\n\tlength = np.ma.array(sfs, mask=mask).sum()\n\tsfs = np.array(sfs).reshape(nsamps)\n\tif verbose: print(\"length {}\".format(length))\n\tif verbose: print(sfs)\n\n\t## Get counts per sfs bin\n\tcounts = Counter()\n\tfor sfsbin in product(*[range(y) for y in [x for x in nsamps]]):\n\t\t## Ignore monomorphic snps\n\t\t## nsamps minus 1 here because of the off by one diff between number\n\t\t## of bins in the sfs and actual number of samples\n\t\tif sfsbin == tuple(nsamps-1) or sfsbin == tuple([0] * len(nsamps)):\n\t\t\tcontinue\n\t\t## ignore zero bin counts\n\t\tif sfs[sfsbin] == 0:\n\t\t\tcontinue\n\t\tif verbose: print(sfsbin, sfs[sfsbin]),\n\t\tcounts.update({sfsbin:sfs[sfsbin]})\n\tif verbose: print(\"nbins {}\".format(len(counts)))\n\n\t## Convert SFS bin style into momi config style\n\tconfigs = pd.DataFrame(index=range(len(counts)), columns=pops)\n \n\tlocus_info = []\n\tfor i, c in enumerate(counts):\n\t\t## (n-1) here because nbins in dadi sfs is n+1\n\t\tcfg = np.array([[(n-1)-x, x] for x,n in zip(c, nsamps)])\n\t\tconfigs.iloc[i] = [list(map(int, list(x))) for x in cfg]\n\t\tlocus_info.append([0, i, counts[c]])\n\tif verbose: print(\"n_snps {}\".format(np.sum([x[2] for x in locus_info])))\n\n\t## Finally build the momi style sfs dictionary and write it out\n\tmomi_sfs = {\"sampled_pops\":pops,\n\t\t\"folded\":folded,\n\t\t\"length\":int(length),\n\t\t\"configs\":configs.values.tolist(),\n\t\t\"(locus,config_id,count)\":locus_info}\n\n\twith open(outfile, 'w') as out:\n\t\tout.write(\"{}\".format(json.dumps(momi_sfs)))\n\t## make it pretty\n\tsfs = momi.Sfs.load(outfile)\n\t## Fold if unfolded\n\tif folded: sfs = sfs.fold()\n\tsfs.dump(outfile)\n\n\ndef oneD_sfs_per_pop(dd, pops, outdir, prefix):\n\tfor pop in pops:\n\t\tallele_counts = [dd[x][\"calls\"][pop] for x in dd.keys()]\n# print(allele_counts) \n\t\tcounts = Counter([x[1] for x in allele_counts])\n\t\tprint(pop, counts)\n\t\tcounts = Counter([x[0] for x in allele_counts])\n\t\tprint(pop, counts)\n\n\ndef make_datadict(genotypes, pops, maxHetFilter,verbose=False,ploidy=1):\n\tdd = {}\n\thetFailSiteCounter=0\n\t## Get genotype counts for each population\n\tfor row in genotypes.iterrows():\n\t\t## iterrows() returns a tuple for some reason\n\t\trow = row[1]\n\n\t\tcalls = {}\n\t\tfor pop in pops.keys():\n\t\t\t## If there is a bunch of info associated w/ each snp then\n\t\t\t## just carve it off for now.\n\t\t\tpop_genotypes = [row[x].split(\":\")[0] for x in pops[pop]]\n\t\t\tref_count = sum([x == \"0\" or x == \"0/0\" or x == \"0|0\" for x in pop_genotypes]) * ploidy\n\t\t\talt_count = sum([x == \"1\" or x == \"1/1\" or x == \"1|1\" for x in pop_genotypes]) * ploidy\n\t\t\t## Haploids shouldn't have hets in the vcf \n\t\t\thet_count = sum([x == \"1/0\" or x == \"0/1\" or x == \"1|0\" or x == \"0|1\" for x in pop_genotypes])\n\n\t\t\tref_count += het_count\n\t\t\talt_count += het_count\n\t\t\t# 20181121: AB adding a condition to account for sites that are 0/1 across all individuals in a population. don't want to include those sites as called, since they indicate some sort of problem. 20181122, AB updated her modification to account for lines with ./., which would mean that 0/1 might be all the calls, but might not equal the length of the genotypes. Instead adding a hom ref and hom alt count and saying if they are both zero (no 0/0 or 0/1 sites, but homAlt isn't zero it means that all available calls for that pop are 0/1)\n\t\t\t# AB adding a hom count \n\t\t\thomRef_count=sum([x == \"0\" or x == \"0/0\" or x == \"0|0\" for x in pop_genotypes])\n\t\t\thomAlt_count=sum([x == \"1\" or x == \"1/1\" or x == \"1|1\" for x in pop_genotypes])\n\t\t\tcalled_gts=sum([x!=\"./.\" for x in pop_genotypes])\n\t\t\t#noCall_count=sum([x == \"./.\" for x in pop_genotypes])\n\t\t\t#if homRef_count==0 and homAlt_count==0 and het_count!=0:\n\t\t\t\t#print(\"found an all 0/1 site for \"+str(pop)+str(pop_genotypes))\n\t\t\t\t#calls[pop] =(0,0) # set it as though it's no-call for that population\n\t\t\tif het_count !=0 and het_count >= called_gts*float(maxHetFilter):\n\t\t\t\tprint(\"found a site with >=\"+str(float(maxHetFilter)*100)+\"% of all calls hets. het count = \"+str(het_count)+\" genotypes: \"+ str(pop_genotypes) +\"\\ndadi call would be: \" +str(ref_count)+\",\"+str(alt_count))\n\t\t\t\t#hetFailSiteCounter += 1\n\t\t\t\tcalls[pop] =(0,0) # set it as though it's no-call for that population\n\n\t\t\telse:\n\t\t\t\tcalls[pop] = (ref_count, alt_count)\n\n\t\tdd[row[\"#CHROM\"]+\"-\"+row[\"POS\"]] =\\\n\t\t\t{\"segregating\":[row[\"REF\"], row[\"ALT\"]],\\\n\t\t\t\"calls\":calls,\\\n\t\t\t\"outgroup_allele\":row[\"REF\"]}\n\t\t#print(str(hetFailSiteCounter))\n\treturn dd\n\n\ndef read_input(vcf_name, all_snps=False, verbose=False):\n\n\t## Counter to track which locus we're evaluating and a list\n\t## to hold all lines for each locus so we can randomly\n\t## select one snp per locus if necessary\n\tcur_loc_number = -1\n\tcur_loc_snps = []\n\n\t## use gzip? \n\tif vcf_name.endswith(\".gz\"):\n\t\tofunc = gzip.open\n\telse: \n\t\tofunc = open\n\tinfile = ofunc(vcf_name, 'r')\n\tlines = infile.readlines()\n\tinfile.close()\n\n\tfor line in lines:\n\t\tif line.startswith(\"#CHROM\"):\n\t\t\theader = line\n\n\t## Just get the data lines, not the comments\n\tlines = [x for x in lines if not x.startswith('#')]\n\tif verbose:\n\t\tprint(\" Number of snps in input file: {}\".format(len(lines)))\n\n\t## Randomly select one snp per locus\n\tif not all_snps:\n\t\tprint(\" Sampling one snp per locus\")\n\t\tloci_nums = set([x.split()[0] for x in lines])\n\t\tloc_dict = {}\n\t\tfor loc in loci_nums:\n\t\t\tloc_dict[loc] = []\n\n\t\t## populate the loci dict\n\t\tfor line in lines:\n\t\t\tloc_dict[line.split()[0]].append(line)\n\n\t\tlines = []\n\t\tfor loc in loc_dict.values():\n\t\t\tline = np.random.choice(loc, 1)[0]\n\t\t\tlines.append(line)\n\n\t\t## Sanity check.\n\t\t## Some snp calling pipelines use the vcf Chrom/pos information to\n\t\t## convey read/snp info per locus (ipyrad), some just assign\n\t\t## all snps to one chrom and use pos/ID (tassel).\n\t\t## If the user chooses to randomly sample one snp per block and the\n\t\t## VCF doesn't use Chrom to indicate RAD loci then it'll just\n\t\t## sample one snp for the whole dataset.\n\t\tif len(loc_dict) == 1:\n\t\t\tmsg = \"\"\"\n\tVCF file uses non-standard Chrom/pos information.\n\tWe assume that Chrom indicates RAD loci and pos indicates snps within each locus \n\tThe VCF file passed does not have rad locus info in the Chrom field.\n\n\tYou can re-run the easySFS conversion with the `-a` flag to use all snps in the conversion.\"\"\"\n\t\t\tsys.exit(msg)\n\n\t\tif verbose:\n\t\t\tprint(\" Using n independent snps: {}\".format(len(lines)))\n\n\n\t## lines now here has a list of either all snps in the input\n\t## or a subset that includes only one snp per locus\n\tgenotypes = pd.DataFrame([x.split() for x in lines], columns=header.split())\n\treturn genotypes\n\n\ndef get_inds_from_input(vcf_name, verbose):\n\t# Read in the vcf file and grab the line with the individual names\n\t# Add the 'U' to handle opening files in universal mode, squashes the\n\t# windows/mac/linux newline issue.\n\t## use gzip? \n\tindnames = []\n\tif vcf_name.endswith(\".gz\"):\n\t\tofunc = gzip.open\n\telse: \n\t\tofunc = open\n\ttry:\n\t\twith ofunc(vcf_name, 'rU') as infile:\n\t\t\tfor line in infile:\n\t\t\t\tif line.startswith('#'):\n\t\t\t\t\tif line.startswith('#CHROM'):\n\t\t\t\t\t\trow = line.strip().split()\n\t\t\t\t\t\t# VCF file format spec says that if the file contains genotype\n\t\t\t\t\t\t# data then \"FORMAT\" will be the last column header before\n\t\t\t\t\t\t# sample IDs start\n\t\t\t\t\t\tstartcol = row.index('FORMAT')\n\t\t\t\t\t\tindnames = [x for x in row[startcol+1:]]\n\t\t\t\t\telse:\n\t\t\t\t\t\tpass\n\t\t\t\telse:\n\t\t\t\t\tbreak\n\texcept Exception as inst:\n\t\tmsg = \"\"\"\n\tProblem reading individuals from input VCF file.\"\"\"\n\t\tprint(\"Error - {}\".format(inst))\n\t\traise\n\n\tif not indnames:\n\t\traise Exception(\"No sample names found in the input vcf. Check vcf file formatting.\")\n\treturn indnames\n\n\t\ndef check_inputs(ind2pop, indnames, pops):\n\t## Make sure all samples are present in both pops file and VCF, give the user the option\n\t## to bail out if something is goofy\n\tpop_set = set(ind2pop.keys())\n\tvcf_set = set(indnames) \n\t## Return error if vcf_set is not a subset of pop_set (ie. samples without pop designation)\n\tif not vcf_set.issubset(pop_set):\n\t\tprint(\"Samples in VCF not present in pops file: {}\\n\".format(\", \".join(vcf_set.difference(pop_set))))\n\t\tsys.exit('Samples in VCF without pop designation')\n\t## remove pop_set values \n\tif not pop_set == vcf_set:\n\t\tprint(\"\\nSamples in pops file not present in VCF: {}\\n\".format(\", \".join(pop_set.difference(vcf_set))))\n\t\t## Remove the offending samples from ind2pop\n\t\t\t# in this case \"pop\" is popping off the samples that are in the diff bet popset and vcfset\n\t\tmap(ind2pop.pop, pop_set.difference(vcf_set))\n\t\t## Remove the offending samples from the pops dict\n\t\t# here k is population name (e.g. ENP; GOC) v is the list of individuals belonging to that population \n\t\tfor k,v in pops.items():\n\t\t\tfor ind in pop_set.difference(vcf_set):\n\t\t\t\t# try a safer change:\n\t\t\t\tif ind in v:\n\t\t\t\t\tv.remove(ind) # doesn't return anything, just removes the entry from v; ab modifie to have this happen in two steps (correct)\n\t\t\t\t\tpops[k] = v\n\t\tfor k, v in pops.items():\n\t\t\tif not v:\n\t\t\t\tprint(\"Empty population, removing - {}\\n\".format(k))\n\t\t\t\tpops.pop(k)\n\t\tfor key,value in pops.items():\n\t\t\tprint(str(len(value)) + ' Surviving individuals for {0}: {1}\\n'.format(key,value))\n\t\t## Return Error if there is offending we \n\t\t# AB: 20181121: removing this part of the script so that there isn't an interactive portion\n\t\t#cont = raw_input(\"\\nContinue, excluding samples not in both pops file and VCF? (yes/no)\\n\")\n\t\t#while not cont in [\"yes\", \"no\"]:\n\t\t\t#cont = raw_input(\"\\nContinue, excluding samples not in both pops file and VCF? (yes/no)\\n\")\n\t\t#if cont == \"no\":\n\t\t\t#sys.exit()\n\treturn ind2pop, indnames, pops\n\n\ndef get_populations(pops_file, verbose=False):\n\t# Here we need to read in the individual population\n\t# assignments file and do this:\n\t# - populate the locs dictionary with each incoming population name\n\t# - populate another dictionary with individuals assigned to populations\n\t# Add the 'U' to handle opening files in universal mode, squashes the\n\t# windows/mac/linux newline issue.\n\n\ttry:\n\t\twith open(pops_file, 'rU') as popsfile:\n\t\t\tind2pop = {}\n\t\t\tpops = OrderedDict()\n\t\t\n\t\t\tlines = popsfile.readlines()\n\t\t\t## Get all the populations\n\t\t\tfor line in lines:\n\t\t\t\tpops.setdefault(line.split()[1], [])\n\t\t\n\t\t\tfor line in lines:\n\t\t\t\tind = line.split()[0]\n\t\t\t\tpop = line.split()[1]\n\t\t\t\tind2pop[ind] = pop\n\t\t\t\tpops[pop].append(ind)\n\n\t\tprint(\"Processing {} populations - {}\".format(len( pops ), pops.keys()))\n\t\tif(verbose):\n\t\t\tfor pop,ls in pops.items():\n\t\t\t\tprint(pop, ls)\n\n\texcept Exception as inst:\n\t\tmsg = \"\"\"\n\tProblem reading populations file. The file should be plain text with one\n\tindividual name and one population name per line, separated by any amount of\n\twhite space. There should be no header line in this file. \n\tAn example looks like this:\n\n\t\tind1 pop1\n\t\tind2 pop1\n\t\tind3 pop2\n\t\tind4 pop2\"\"\"\n\t\tprint(msg)\n\t\tprint(\" File you specified is: \".format(pops_file))\n\t\tprint(\" Error - {}\".format(inst))\n\t\traise\n\n\treturn ind2pop, pops\n\n\ndef parse_command_line():\n\tparser = argparse.ArgumentParser(\n\t\tformatter_class=argparse.RawDescriptionHelpFormatter,\n\t\tepilog=\"\"\"\\n\n\t\"\"\")\n\n\tparser.add_argument(\"-a\", dest=\"all_snps\", action='store_true', \n\t\thelp=\"Keep all snps within each RAD locus (ie. do _not_ randomly sample 1 snp per locus).\")\n\n\tparser.add_argument(\"-i\", dest=\"vcf_name\", required=True, \n\t\thelp=\"name of the VCF input file being converted\")\n\n\tparser.add_argument(\"-p\", dest=\"populations\", required=True, \n\t\thelp=\"Input file containing population assignments per individual\")\n\n\tparser.add_argument(\"--proj\", dest=\"projections\", \n\t\thelp=\"List of values for projecting populations down to different sample sizes\")\n\n\tparser.add_argument(\"--preview\", dest=\"preview\", action='store_true',\n\t\thelp=\"Preview the number of segragating sites per population for different projection values.\")\n\n\tparser.add_argument(\"-o\", dest=\"outdir\", default='output', \n\t\thelp=\"Directory to write output SFS to\")\n\n\tparser.add_argument(\"--ploidy\", dest=\"ploidy\", type=int, default=2,\n\t\thelp=\"Specify ploidy. Default is 2. Only other option is 1 for haploid.\")\n\n\tparser.add_argument(\"--prefix\", dest=\"prefix\", \n\t\thelp=\"Prefix for all output SFS files names.\")\n\n\tparser.add_argument(\"--unfolded\", dest=\"unfolded\", action='store_true', \n\t\thelp=\"Generate unfolded SFS. This assumes that your vcf file is accurately polarized.\")\n\n\tparser.add_argument(\"--dtype\", dest=\"dtype\", default=\"float\",\n\t\thelp=\"Data type for use in output sfs. Options are `int` and `float`. Default is `float`.\")\n\n\tparser.add_argument(\"--GQ\", dest=\"GQual\", \n\t\thelp=\"minimum genotype quality tolerated\", default=20)\n\n\tparser.add_argument(\"-f\", dest=\"force\", action='store_true',\n\t\thelp=\"Force overwriting directories and existing files.\")\n\n\tparser.add_argument(\"-v\", dest=\"verbose\", action='store_true',\n\t\thelp=\"Set verbosity. Dump tons of info to the screen\")\n\t\n\tparser.add_argument(\"-maxHetFilter\", dest=\"maxHetFilter\", default=1.0,\n\t\thelp=\"Fraction of called genotypes per population that are heterozygous (0/1). e.g. -maxHetFilter 0.8 would exclude any site that has >=80 percent of called genotypes 0/1 within a population. Default is 1.0 which only removes sites that are all 0/1 within the population. If your SFS is U-shaped after projection, I recommend lowering the max threshold to .7-.9.\")\n\n\t## if no args then return help message\n\tif len(sys.argv) == 1:\n\t\tparser.print_help()\n\t\tsys.exit(1)\n\n\t## parse args\n\targs = parser.parse_args()\n\treturn args\n\ndef init(args):\n\t## Set up output directory and output prefix\n\toutdir = args.outdir\n\tif os.path.exists(outdir) and args.force == False:\n\t\tprint(\"\\nOutput directory exists. Use -f to override.\\n\")\n\t\tsys.exit()\n\tif os.path.exists(outdir):\n\t\tshutil.rmtree(outdir)\n\tos.mkdir(outdir)\n\tos.mkdir(os.path.join(outdir, \"dadi\"))\n\tos.mkdir(os.path.join(outdir, \"fastsimcoal2\"))\n\n\tif not args.prefix:\n\t\tprefix = args.vcf_name.split('/')[-1].split('.')[0]\n\telse:\n\t\tprefix = args.prefix\n\tif args.verbose:\n\t\tprint(\"Prefix - {}\".format(prefix))\n\n\treturn outdir, prefix\n\n\ndef main():\n\targs = parse_command_line()\n\n\tif args.verbose:\n\t\tprint(\"Input Arguments:\\n\\t{}\".format(args))\n\n\t## Set up output directory and output prefix\n\tif args.preview:\n\t\tif args.verbose: print(\"Doing preview so skipping directory initialization\")\n\telse:\n\t\toutdir, prefix = init(args)\n\n\t## Get populations and populations assignments for individuals\n\t## ind2pop - a dictionary mapping individuals to populations\n\t## pops - a dictionary of populations and all inds w/in them\n\tind2pop, pops = get_populations(args.populations, args.verbose)\n\n\t## Read in the names of individuals present in the vcf file\n\tindnames = get_inds_from_input(args.vcf_name, args.verbose)\n\n\t## Check whether inds exist in the population mapping and input vcf\n\t## files. Give the user an opportunity to bail if there is a mismatch.\n\tif not args.force:\n\t\tind2pop, indnames, pops = check_inputs(ind2pop, indnames, pops)\n\n\t## Reads the vcf and returns a pandas dataframe\n\tgenotypes = read_input(args.vcf_name, all_snps=args.all_snps,\n\t\t\t\t\t\t\tverbose=args.verbose)\n\t## Convert dataframe to dadi-style datadict\n\tdd = make_datadict(genotypes, pops=pops, ploidy=args.ploidy, verbose=args.verbose,maxHetFilter=args.maxHetFilter)\n\t## Don't write the datadict to the file for preview mode\n\tif not args.preview:\n\t\twith open(os.path.join(args.outdir, \"datadict.txt\"), 'w') as outfile:\n\t\t\tfor x,y in dd.items():\n\t\t\t\toutfile.write(x+str(y)+\"\\n\")\n\t\n\t## Do preview of various projections to determine good values\n\tif args.preview:\n\t\tdadi_preview_projections(dd, pops, ploidy=args.ploidy, fold=args.unfolded)\n\t\tsys.exit()\n\n\telif args.projections:\n\t\t## Validate values passed in for projecting\n\t\tproj = [int(x) for x in args.projections.split(\",\")]\n\t\tif not len(pops) == len(proj):\n\n\t\t\tmsg = \"You must pass in the same number of values for projection as you have populations specified\"\n\t\t\tmsg += \"\\n\\nN pops = {}\\nN projections = {}\\nProjections = {}\".format(len(pops), len(proj), proj)\n\t\t\tsys.exit(msg)\n\n\t\t## Create 1D sfs for each population\n\t\tdadi_oneD_sfs_per_pop(dd, pops, proj=proj, unfold=args.unfolded, outdir=outdir, prefix=prefix, dtype=args.dtype)\n\n\t\t## Create pairwise 2D sfs for each population\n\t\tdadi_twoD_sfs_combinations(dd, pops, proj=proj, unfold=args.unfolded,\\\n\t\t\t\t\t\t\t\toutdir=outdir, prefix=prefix, dtype=args.dtype, verbose=args.verbose)\n\n\t\t## Create the full multiSFS for all popuations combined\n\t\tsfs_file = dadi_multiSFS(dd, pops, proj=proj, unfold=args.unfolded, outdir=outdir, prefix=prefix, dtype=args.dtype)\n\n\t\ttry:\n\t\t\timport momi\n\t\t\t## Create momi-style sfs\n\t\t\tdadi_to_momi(infile=sfs_file, outdir=outdir, verbose=args.verbose)\n\t\texcept:\n\t\t\t## Can't create momi file at this point because we're locked to python2 \n\t\t\t## because of dadi. \n\t\t\tpass\n\n\telse:\n\t\tprint(\"Either --preview or --proj must be specified.\")\n\nif __name__ == \"__main__\":\n\tmain()\n\n", "sub_path": "demographic_inference/get_neutral_SFS_pipeline/easySFS_ab.py", "file_name": "easySFS_ab.py", "file_ext": "py", "file_size_in_byte": 25917, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "matplotlib.use", "line_number": 16, "usage_type": "call"}, {"api_name": "dadi.Spectrum.from_data_dict", "line_number": 52, "usage_type": "call"}, {"api_name": "dadi.Spectrum", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path", "line_number": 65, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path", "line_number": 69, "usage_type": "attribute"}, {"api_name": "dadi.Spectrum.from_data_dict", "line_number": 71, "usage_type": "call"}, {"api_name": "dadi.Spectrum", "line_number": 71, "usage_type": "attribute"}, {"api_name": "numpy.rint", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 75, "usage_type": "call"}, {"api_name": "dadi.Spectrum", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path", "line_number": 92, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 93, "usage_type": "call"}, {"api_name": "os.path", "line_number": 93, "usage_type": "attribute"}, {"api_name": "itertools.combinations", "line_number": 96, "usage_type": "call"}, {"api_name": "itertools.combinations", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 110, "usage_type": "call"}, {"api_name": "os.path", "line_number": 110, "usage_type": "attribute"}, {"api_name": "dadi.Spectrum.from_data_dict", "line_number": 111, "usage_type": "call"}, {"api_name": "dadi.Spectrum", "line_number": 111, "usage_type": "attribute"}, {"api_name": "numpy.rint", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 115, "usage_type": "call"}, {"api_name": "dadi.Spectrum", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 124, "usage_type": "call"}, {"api_name": "os.path", "line_number": 124, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 159, "usage_type": "call"}, {"api_name": "os.path", "line_number": 159, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 160, "usage_type": "call"}, {"api_name": "os.path", "line_number": 160, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 161, "usage_type": "call"}, {"api_name": "os.path", "line_number": 161, "usage_type": "attribute"}, {"api_name": "dadi.Spectrum.from_data_dict", "line_number": 164, "usage_type": "call"}, {"api_name": "dadi.Spectrum", "line_number": 164, "usage_type": "attribute"}, {"api_name": "numpy.rint", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 168, "usage_type": "call"}, {"api_name": "dadi.Spectrum", "line_number": 169, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 175, "usage_type": "call"}, {"api_name": "os.path", "line_number": 175, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 192, "usage_type": "call"}, {"api_name": "os.path", "line_number": 192, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 193, "usage_type": "call"}, {"api_name": "os.path", "line_number": 193, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 194, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 196, "usage_type": "call"}, {"api_name": "os.path", "line_number": 196, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 215, "usage_type": "call"}, {"api_name": "numpy.ma.array", "line_number": 231, "usage_type": "call"}, {"api_name": "numpy.ma", "line_number": 231, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 232, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 237, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 252, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 257, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 260, "usage_type": "call"}, {"api_name": "momi.Sfs.load", "line_number": 272, "usage_type": "call"}, {"api_name": "momi.Sfs", "line_number": 272, "usage_type": "attribute"}, {"api_name": "collections.Counter", "line_number": 282, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 284, "usage_type": "call"}, {"api_name": "gzip.open", "line_number": 343, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 373, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 373, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 390, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 398, "usage_type": "call"}, {"api_name": "gzip.open", "line_number": 409, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 446, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 488, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 526, "usage_type": "call"}, {"api_name": "argparse.RawDescriptionHelpFormatter", "line_number": 527, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 574, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 576, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 585, "usage_type": "call"}, {"api_name": "os.path", "line_number": 585, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 587, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 588, "usage_type": "call"}, {"api_name": "os.path", "line_number": 588, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 589, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 590, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 591, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 591, "usage_type": "call"}, {"api_name": "os.path", "line_number": 591, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 592, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 592, "usage_type": "call"}, {"api_name": "os.path", "line_number": 592, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 636, "usage_type": "call"}, {"api_name": "os.path", "line_number": 636, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 643, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 652, "usage_type": "call"}]} +{"seq_id": "624920107", "text": "import sys\nimport pika\n\n# Set up the exchange environment\nconnection = pika.BlockingConnection(pika.ConnectionParameters('localhost'))\nchannel = connection.channel()\n\n# Establish the exchange\nchannel.exchange_declare(exchange='sensor_exchange', exchange_type='topic')\n\n# Establish a queue called\nqueue_name = sys.argv[1]\nresult = channel.queue_declare(queue=queue_name, exclusive=False, durable=True)\nqueue_name = result.method.queue\n\nbinding_keys = sys.argv[1:]\n\nif not binding_keys:\n sys.stderr.write(\"Usage: %s [binding_key]...\\n\" % sys.argv[0])\n sys.exit(1)\n\n\nfor binding_key in binding_keys:\n channel.queue_bind(\n exchange='sensor_exchange', queue=queue_name, routing_key=binding_key)\n\nprint(' [*] PiExchange is now running...')\n", "sub_path": "Pi_Startup/exchange.py", "file_name": "exchange.py", "file_ext": "py", "file_size_in_byte": 750, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pika.BlockingConnection", "line_number": 5, "usage_type": "call"}, {"api_name": "pika.ConnectionParameters", "line_number": 5, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 12, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 16, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 19, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 19, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 19, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 20, "usage_type": "call"}]} +{"seq_id": "135385089", "text": "from datetime import date\nfrom io import StringIO\nimport requests\nimport json\nfrom bs4 import BeautifulSoup\n\n# Retrieve current date to use year as maxYear in range\ncurrent_date = date.today()\n\n#Prompt to take required form number input\nirs_Search_Term = input(\"Please enter which form you would like to search (Only plain text without quotations allowed. Multiple form types allowed with comma as required delimiter.): \")\nsearchList = irs_Search_Term.split(\",\")\n\n# Declare an empty list to eventually store list of lists for multiple form types\nlistOfSingleAndMulitpleFormTypes= []\n\nfor j in searchList: \n # Declare an empty list to eventually store single list of forms one type at a time\n listOfForms = []\n\n # Declare and initialize url formatted to dynamically add irs_Search_Term to make GET request in next lines.\n url = f'https://apps.irs.gov/app/picklist/list/priorFormPublication.html?resultsPerPage=200&sortColumn=sortOrder&indexOfFirstRow=0&criteria=formNumber&value={j.strip()}&isDescending=false'\n\n # Declare and initialize BeautifulSoup variable to parse (html) GET request made in previous lines.\n response = requests.get(url)\n\n soup = BeautifulSoup(response.text, 'html.parser')\n\n # Declare and initialize desired table to iterate through\n table = soup.find_all(class_='picklist-dataTable')[0]\n records = table.find_all('tr')\n\n # Create a class to format/clean required data from web scraping. \n class formatterClass:\n def __init__(self, form_number, form_title, min_year, max_year):\n self.form_number = form_number.strip()\n self.form_title = form_title.strip()\n self.min_year = min_year.strip()\n self.max_year = max_year\n\n # Method to represent objects of this class as string\n def __repr__(self):\n return json.dumps({\"form_number\": self.form_number, \"form_title\": self.form_title,\"min_year\": self.min_year,\"max_year\": self.max_year}, indent=4)\n\n # Iterate through records\n for i in records:\n # Declare and initialize LeftCellSpacer, MiddleCellSpacer, and EndCellSpacer classes in picklistTable div.\n # After if statement and try/except blocks, we can retrieve new values of these variables.\n form_number = 'Form Number Placeholder'\n form_title = 'Form Title Placeholder'\n min_year = 'Min Year Placeholder'\n max_year = 'Max Year Placeholder'\n\n # If condition to retrieve form_number records that match exactly with the irs_Search_Term input.\n # For Example, the input \"Form W-2\" will only return exact \"Form W-2\" results.\n # This is to ensure proper handling of expected final results because the IRS web page\n # will return search terms whose strings end differently like \"Form W-2 P\" or \"Form W-2GU\"\n if (i.find('a').contents[0] == j.strip()):\n form_number = i.find('a').contents[0]\n try:\n form_title = i.find(class_='MiddleCellSpacer').get_text()\n except AttributeError:\n print (\"\")\n try:\n min_year = i.find(class_='EndCellSpacer').get_text()\n except AttributeError:\n print (\"\")\n max_year = current_date.year\n\n if form_title != 'Form Title Placeholder':\n # If not true, append object to list because the variable was updated.\n listOfForms.append(formatterClass(form_number, form_title, min_year, max_year))\n listOfSingleAndMulitpleFormTypes.append(listOfForms)\n\nprint(listOfSingleAndMulitpleFormTypes)\n\n", "sub_path": "searchIrsTaxForms.py", "file_name": "searchIrsTaxForms.py", "file_ext": "py", "file_size_in_byte": 3618, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "datetime.date.today", "line_number": 8, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 8, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 25, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 27, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 43, "usage_type": "call"}]} +{"seq_id": "115194781", "text": "'''\nCreated on Nov 12, 2013\n\n@author: DamienBlack\n'''\nfrom PyQt5 import QtWidgets\n\nclass Model():\n def __init__(self):\n self.orig = \"\"\n \n def findLongestSubStr(self):\n currentSub = ''\n currentlongest = ''\n uniqueChar = {} #Dictionary of char : index\n if ( self.orig == '' ):\n return ''\n if len( self.orig ) == 1:\n return self.orig\n for i,l in enumerate(self.orig):\n if l not in currentSub:\n uniqueChar.update({l:i})\n currentSub += l\n currentlongest = self.longerWord(currentSub, currentlongest)\n else:\n uniqueChar.update({l:i})\n currentSub = l\n return currentlongest\n \n def longerWord(self,str1,str2):\n if ( len(str1) > len(str2)):\n return str1\n else:\n return str2 \n \n\nclass View(QtWidgets.QWidget):\n def __init__(self, parent = None):\n QtWidgets.QWidget.__init__(self, parent)\n \n self.wordLabel = QtWidgets.QLabel(\"What is the word?\")\n self.wordLineEdit = QtWidgets.QLineEdit()\n self.wordButton = QtWidgets.QPushButton(\"&Find_Substring\")\n \n mainLayout = QtWidgets.QGridLayout()\n mainLayout.addWidget(self.wordLabel,0,0)\n mainLayout.addWidget(self.wordLineEdit,2,0)\n mainLayout.addWidget(self.wordButton,1,1)\n \n self.setLayout(mainLayout)\n self.setWindowTitle(\"Dat Substr though\")\n \nclass Controller():\n def __init__(self, model, view):\n self.model = model\n self.view = view\n \n self.view.wordButton.clicked.connect( self.on_wordButton_clicked )\n \n def on_wordButton_clicked(self):\n self.model.orig = self.view.wordLineEdit.text()\n print( self.model.findLongestSubStr() )\n \n def begin(self):\n self.view.show()\n\n\n\n\nif __name__ == '__main__':\n\n import sys\n \n app = QtWidgets.QApplication(sys.argv)\n model = Model()\n view = View()\n controller = Controller(model, view)\n controller.begin()\n\n sys.exit(app.exec_())", "sub_path": "Alg_SubstringNoDup/src/String_SubnoDup/Method1.py", "file_name": "Method1.py", "file_ext": "py", "file_size_in_byte": 2151, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 37, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 37, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QWidget.__init__", "line_number": 39, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 39, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 39, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 41, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 41, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 42, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 42, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 43, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 43, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QGridLayout", "line_number": 45, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 45, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 74, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 74, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 74, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 80, "usage_type": "call"}]} +{"seq_id": "20655991", "text": "#coding=utf8\nimport numpy as np\nimport json\nimport sys\n\nreload(sys)\nsys.setdefaultencoding('utf8')\n\nBasePath = sys.path[0]\nclass Vocab(object):\n def __init__(self, vocab_file):\n self._word_to_id = {}\n self._id_to_word = {}\n self._count = 0 # keeps track of total number of words in the Vocab\n with open(vocab_file, 'r') as vocab_f:\n for line in vocab_f:\n pieces = line.split()\n if len(pieces) != 2:\n print('Warning: incorrectly formatted line in vocabulary file: %s\\n'%line)\n continue\n w = pieces[0]\n count = pieces[1]\n if w in self._word_to_id:\n raise Exception('Duplicated word in vocabulary file: %s' % w)\n if int(w) == 0:\n continue\n self._word_to_id[int(w)] = int(count)\n # self._id_to_word[self._count] = w\n self._count += 1\n\n print(\"Finished constructing vocabulary of %i total words. \"%(self._count))\n def word2count(self, word):\n \"\"\" Returns the id (integer) of a word (string). Returns [UNK] id if word is OOV.\"\"\"\n return self._word_to_id[word]\n\n\ndef one_hot_Vocab(keys):\n one_dict = dict()\n _count = 0\n for key in keys:\n if key in one_dict:\n raise Exception(\"Duplicated word in vocabulary file: %s\" %key)\n if key == 0:\n continue\n one_dict[key] = _count\n _count += 1\n return one_dict\n\n\n\n\n\ndef read_from_json(path_file):\n with open(path_file, 'rb') as jd:\n data = json.loads(jd.read())\n return data\ndef writejson2file(dict_data, save_path):\n json_data = json.dumps(dict_data, ensure_ascii=False)\n with open(save_path, 'wb') as f:\n f.write(json_data)\n\ndef is_utf8(s):\n\ttry:\n\t\ts.decode('utf-8')\n\t\treturn True\n\texcept UnicodeError:\n\t\treturn False\n\ndef is_ascii(s):\n\treturn all(ord(c) < 128 for c in s)\n\ndef is_number(s):\n\ttry:\n\t\tfloat(s)\n\t\treturn True\n\texcept ValueError:\n\t\tpass\n\ttry:\n\t\timport unicodedata\n\t\tunicodedata.numeric(s)\n\t\treturn True\n\texcept (TypeError, ValueError):\n\t\tpass\n\treturn False #\n\ndef save_data2rawdata(data_path, save_path):\n data = read_from_json(data_path)\n save_dict = dict()\n return_x = list()\n return_y = list()\n for line in data:\n # print(line)\n # print(' '.join(line['content']))\n return_x.append(' '.join(line['content']))\n return_y.append(line['result'])\n\n save_dict['content'] = return_x\n save_dict['result'] = return_y\n\n assert len(save_dict['content']) == len(save_dict['result']), \"the len does not match\"\n with open(save_path, 'wb') as rsjd:\n json_data = json.dumps(save_dict)\n rsjd.write(json_data)\n\ndef get_dev_train_data(x_path, y_path):\n x_data = read_from_json(x_path)['x']\n print(\"finished get x_data\")\n y_data = np.loadtxt(y_path)\n print(\"finished get y_data\")\n return np.array(x_data), y_data\n\n\n\n\n\n# def get_dev_train_data(data_path, result_path, vocab_file):\n# vocab = Vocab(vocab_file)\n# print(vocab._word_to_id)\n# x_data = read_from_json(data_path)\n# print(\"~~~~~~~~~~~~~~~~~~~~~~get_x~~~~~~~~~~~~~~~~~~~\")\n# y_data = read_from_json(result_path)\n# print(\"~~~~~~~~~~~~~~~~~~~~~~get_y~~~~~~~~~~~~~~~~~~~\")\n# return_x = list()\n# return_y = list()\n# for line in x_data:\n# blank_line = ' '.join(line['content'])\n# return_x.append(blank_line)\n#\n# for line in y_data['base']:\n# y_test = np.array([0] * 15)\n# one_hot = [vocab.word2id(word) for word in line]\n# y_test[one_hot] = 1\n# # print(y_test)\n# return_y.append(y_test)\n# return np.array(return_x), np.array(return_y)\ndef get_y_label(n):\n vocab_path = BasePath + \"/other_data/vocab.txt\"\n save_path = BasePath + \"/other_data/raw_split_json_data.txt\"\n data = read_from_json(save_path)\n vocab = Vocab(vocab_path)\n print(\"the vocab is : \")\n print(vocab._word_to_id)\n sorted_vocab = sorted(vocab._word_to_id.items(), key=lambda e: e[1], reverse=True)\n print(\"the sorted vocab is : \")\n print(sorted_vocab)\n topn_list = [10, 15, 20, 30, 40, 45, 50, 60, 70]\n topn = topn_list[n]\n # 字典keys集合\n keys = [tuple[0] for tuple in sorted_vocab[:topn]]\n print(\"the keys set len is : {}\".format(len(keys)))\n print(keys)#\n one_hot_vocab = one_hot_Vocab(keys)\n print(\"the one hot dict is : \")\n print(one_hot_vocab)\n with open(BasePath + \"/other_data/one_hot_vocab_\" + str(topn)+ \".txt\", 'wb') as f:\n f.write(json.dumps(one_hot_vocab))\n\n # 对高频结果进行筛选\n data['result'] = [list(set(result) & set(keys)) for result in data['result']]\n return_y = list()\n for line in data['result']:\n one_hot = [one_hot_vocab[word] for word in line]\n print(one_hot)\n # one_hot = np.array([1,1,1,1])\n return_y.append(np.array(one_hot))\n return np.array(return_y)\n\n\n\n\n\n\nif __name__ == \"__main__\":\n\n data_path = BasePath + \"/other_data/json_data.txt\"\n vocab_path = BasePath + \"/other_data/vocab.txt\"\n save_path = BasePath + \"/other_data/raw_split_json_data.txt\"\n\n # 将json_data保存为原始数据,即保存为案例对应的所有法条\n # save_data2rawdata(data_path, save_path)\n\n # 对法条进行筛选,仅保留出现次数高于一定频率的词\n # 加载json_data数据\n\n data = read_from_json(save_path)\n min_frequence = 3000\n # 加载筛选出的法条字典\n vocab = Vocab(vocab_path)\n\n print(\"the vocab is : \")\n print(vocab._word_to_id)\n\n\n sorted_vocab = sorted(vocab._word_to_id.items(), key=lambda e: e[1], reverse=True)\n print(\"the sorted vocab is : \")\n print(sorted_vocab)\n topn_list = [5, 8, 10, 15, 20, 30, 40, 45, 50, 60, 70]\n topn = topn_list[1]\n\n # 字典keys集合\n keys = [tuple[0] for tuple in sorted_vocab[:topn]]\n print(\"the keys set len is : {}\".format(len(keys)))\n print(keys)#\n one_hot_vocab = one_hot_Vocab(keys)\n print(\"the one hot dict is : \")\n print(one_hot_vocab)\n with open(BasePath + \"/other_data/one_hot_vocab_\" + str(topn)+ \".txt\", 'wb') as f:\n f.write(json.dumps(one_hot_vocab))\n\n # # 对高频结果进行筛选\n # data['result'] = [list(set(result) & set(keys)) for result in data['result']]\n # # return_x = [' '.join(sentence) for sentence in data['content']]\n # # return_x = data['content']\n #\n # # return_y = list()\n # # for line in data['result']:\n # # y_test = np.array([0] * len(keys))\n # # one_hot = [one_hot_vocab[word] for word in line]\n # # y_test[one_hot] = 1\n # # return_y.append(y_test)\n # #\n # # # train_dev_x_file_path = BasePath + \"/other_data/train_dev_data_x.txt\"\n # # train_dev_y_file_path = BasePath + \"/other_data/train_dev_data_y_topn\"+ str(topn) +\".txt\"\n # # print(train_dev_y_file_path)\n # # print(data['result'][10])\n # # print(return_y[10])\n # # # # writejson2file({'x':return_x}, train_dev_x_file_path)\n # # np.savetxt(train_dev_y_file_path,np.array(return_y, dtype=np.int32))\n #\n # return_y = list()\n # for line in data['result']:\n # one_hot = [one_hot_vocab[word] for word in line]\n # print(one_hot)\n # # one_hot = np.array([1,1,1,1])\n # return_y.append(np.array(one_hot))\n # print(np.array(return_y))\n #\n # train_dev_y_file_path = BasePath + \"/other_data/train_dev_data_y_topn\"+ str(topn) + \"_rawtest.txt\"\n # print(train_dev_y_file_path)\n # # np.savetxt(train_dev_y_file_path, np.array(return_y))\n # np.array(return_y).tofile(\"filename.bin\")\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "make_train_dev_data.py", "file_name": "make_train_dev_data.py", "file_ext": "py", "file_size_in_byte": 7676, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "sys.setdefaultencoding", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 55, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 58, "usage_type": "call"}, {"api_name": "unicodedata.numeric", "line_number": 80, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 110, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 156, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 166, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 208, "usage_type": "call"}]} +{"seq_id": "623303077", "text": "# -*- coding: utf-8 -*-\nimport cv2\n\n\ndef main():\n # Camera caputure\n cap1 = cv2.VideoCapture(0)\n\n while True:\n im1 = cap1.read()[1] # Get Frame cam1\n im1_gray = cv2.cvtColor(im1, cv2.COLOR_RGB2GRAY) # Tran.to gray cam1 \n\n # thresholding\n thresh1 = 75\n max_pixel1 = 255\n ret, im1_bin = cv2.threshold(im1_gray,\n thresh1,\n max_pixel1,\n cv2.THRESH_BINARY)\n \n # Set window\n cv2.imshow(\"Original\",im1)\n cv2.imshow(\"GrayScale\",im1_gray)\n cv2.imshow(\"Binary\",im1_bin)\n # Loop out\n if cv2.waitKey(10) > 0:\n cap1.release()\n cv2.destroyAllWindows()\n break\n\nif __name__ == '__main__':\n main()\n", "sub_path": "CameraImageProcessing.py", "file_name": "CameraImageProcessing.py", "file_ext": "py", "file_size_in_byte": 818, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "cv2.VideoCapture", "line_number": 7, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2GRAY", "line_number": 11, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 19, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 22, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 28, "usage_type": "call"}]} +{"seq_id": "379830623", "text": "# -*- coding:utf-8-*-\nimport tensorflow as tf\nimport numpy as np\nimport os\nimport sys\nfrom tqdm import tqdm\n\n#tfrecord에 몇개의 example이 있는지 반환하는 함수\ndef get_num_records(tf_record_file):\n return len([x for x in tf.python_io.tf_record_iterator(tf_record_file)])\n\n# 파라미터로 test, train용 데이터를 가진 폴더를 말해줘야한다.\nnum = len(sys.argv)\nif num != 2:\n\tprint('wrong parameters')\n\texit()\n\n# train과 test용 파일들의 파일명 불러오기. + tfrecord 에 들어있는 파일 개수 세기\npath = sys.argv[1]\nif path[-1] != '/':\n\tpath += '/'\n\nprint('train 파일 개수 세는중...')\ntrain_files = os.listdir(path + 'train')\ntrain_files_num = 0\nfor i in tqdm(range(len(train_files))):\n\ttrain_files[i] = path + 'train/'+train_files[i]\n\ttrain_files_num += get_num_records(train_files[i])\nprint('train 음악파일 개수 : {}'.format(train_files_num))\n\nprint('test 파일 개수 세는중...')\ntest_files = os.listdir(path + 'test')\ntest_files_num = 0\nfor i in tqdm(range(len(test_files))):\n\ttest_files[i] = path + 'test/'+test_files[i]\n\ttest_files_num += get_num_records(test_files[i])\nprint('test 음악파일 개수 : {}'.format(test_files_num))\nprint()\n\nbatch_size = 100 # 한번에 뽑아올 파일 개수\n\nwith tf.device('/cpu:0'):\n\t# queue runner를 위한 사전 준비\n\tfilename_queue1 = tf.train.string_input_producer(\n\t train_files, shuffle=False, name='filename_queue1')\n\treader1 = tf.TFRecordReader()\n\tkey1, value1 = reader1.read(filename_queue1)\n\txy1 = tf.parse_single_example(value1, features={'label': tf.FixedLenFeature(\n\t [8], tf.float32), 'mfcc': tf.FixedLenFeature([103360], tf.float32)})\n\ttrain_x_batch, train_y_batch = tf.train.batch(\n\t [xy1['mfcc'], xy1['label']], batch_size=batch_size)\n\n\tfilename_queue2 = tf.train.string_input_producer(\n\t test_files, shuffle=False, name='filename_queue2')\n\treader2 = tf.TFRecordReader()\n\tkey2, value2 = reader2.read(filename_queue2)\n\txy2 = tf.parse_single_example(value2, features={'label': tf.FixedLenFeature(\n\t [8], tf.float32), 'mfcc': tf.FixedLenFeature([103360], tf.float32)})\n\ttest_x_batch, test_y_batch = tf.train.batch(\n\t [xy2['mfcc'], xy2['label']], batch_size=batch_size)\n\n\tgenNum = 8 # 장�� 개수\n\tkeep_prob = tf.placeholder(tf.float32) # dropout 세기 조절용\n\n\tX = tf.placeholder(tf.float32, shape=[None, 103360])\n\t#normalize 과정\n\tX_min = tf.reduce_min(X)\n\tX_max = tf.reduce_max(X)\n\tnorm_X = (X - X_min) / (X_max - X_min)\n\tX_img = tf.reshape(norm_X, [-1, 40, 2584, 1])\n\tY = tf.placeholder(tf.float32, shape=[None, genNum])\n\n\twith tf.name_scope('layer1') as scope:\n\t W1 = tf.Variable(tf.random_normal(\n\t [3, 3, 1, 32], stddev=0.1)) # filter size\n\t layer1 = tf.nn.conv2d(X_img, W1, strides=[1, 1, 1, 1], padding='SAME') # convolution\n\t layer1 = tf.nn.relu(layer1) # relu\n\t layer1 = tf.nn.max_pool(layer1, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME') # max pooling\n\n\t w1_histo = tf.summary.histogram('weights1', W1)\n\t layer1_histo = tf.summary.histogram('layer1', layer1)\n\n\twith tf.name_scope('layer2') as scope:\n\t W2 = tf.Variable(tf.random_normal([3, 3, 32, 32], stddev=0.1)) # filter size\n\t layer2 = tf.nn.conv2d(layer1, W2, strides=[1, 1, 1, 1], padding='SAME') # convolution\n\t layer2 = tf.nn.relu(layer2) # relu\n\t layer2 = tf.nn.max_pool(layer2, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME') # max pooling\n\n\t w2_histo = tf.summary.histogram('weights2', W2)\n\t layer2_histo = tf.summary.histogram('layer2', layer2)\n\n\twith tf.name_scope('FullyConnected') as scope:\n\t layer2 = tf.reshape(layer2, [-1, 10 * 646 * 32])\n\t W = tf.get_variable(\"W\", shape=[10 * 646 * 32, genNum], initializer=tf.contrib.layers.xavier_initializer())\n\t b = tf.Variable(tf.random_normal([genNum]))\n\t H = tf.nn.softmax(tf.matmul(layer2, W) + b)\n\n\t w_histo = tf.summary.histogram('weights', W)\n\t b_histo = tf.summary.histogram('biases', b)\n\t hypothesis_histo = tf.summary.histogram('hypothesis', H)\n\n\tcost = tf.reduce_mean(-tf.reduce_sum(Y * tf.log(H), axis=1))\n\tcost_scalar = tf.summary.scalar('cost', cost)\n\n\toptimizer = tf.train.AdamOptimizer(learning_rate=1e-5) # nan 뜨면 러닝레이트 좀 줄여라\n\ttrain = optimizer.minimize(cost)\n\n\tprediction = tf.argmax(H, 1)\n\tcorrect_prediction = tf.equal(prediction, tf.argmax(Y, 1))\n\taccuracy = tf.reduce_mean(tf.cast(correct_prediction, tf.float32))\n\taccuracy_scalar = tf.summary.scalar('accuracy', accuracy)\n\nwith tf.Session() as sess:\n\tsess.run(tf.global_variables_initializer())\n\t# queue runner 사용할 때 준비해야 하는 것들\n\tcoord = tf.train.Coordinator()\n\tthreads = tf.train.start_queue_runners(coord=coord, sess=sess)\n\n\t# tensorboard 관련\n\tsummary = tf.summary.merge_all()\n\twriter = tf.summary.FileWriter('./logs/CNN_0_0001')\n\twriter.add_graph(sess.graph)\n\n\ttotalTestFileNum = test_files_num\n\ttotalTrainingFileNum = train_files_num\n\ttraining_epochs = 15 # 에폭 횟수\n\n\tprint('훈련 시작, 총 에폭 수 : {}, 배치 크기 : {}'.format(training_epochs, batch_size))\n\tavg_cost = 0\n\tfor epoch in range(training_epochs):\n\t\ttotal_batch = int(totalTrainingFileNum / batch_size)\n\n\t\tfor i in tqdm(range(total_batch)):\n\t\t\tbatch_xs, batch_ys = sess.run([train_x_batch, train_y_batch])\n\t\t\tc, _ = sess.run([cost, train], feed_dict = {X:batch_xs, Y:batch_ys})\n\t\t\tavg_cost += c / total_batch\n\n\t\tprint('epoch : %04d'%(epoch+1), 'cost : %09f'%avg_cost)\n\n\tprint('테스트 시작')\n\tavg_acc = 0\n\ttotal_batch = int(totalTestFileNum / batch_size)\n\tfor i in tqdm(range(total_batch)):\n\t\tbatch_xs, batch_ys = sess.run([test_x_batch, test_y_batch])\n\t\tacc = accuracy.eval(session=sess, feed_dict={X:batch_xs, Y:batch_ys})\n\t\tavg_acc += acc / total_batch\n\tprint('accuracy : {}'.format(avg_acc))\n\tcoord.request_stop()\n\tcoord.join(threads)\n", "sub_path": "learning/CNN.py", "file_name": "CNN.py", "file_ext": "py", "file_size_in_byte": 5815, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "tensorflow.python_io.tf_record_iterator", "line_number": 10, "usage_type": "call"}, {"api_name": "tensorflow.python_io", "line_number": 10, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 13, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 24, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 26, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 32, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 34, "usage_type": "call"}, {"api_name": "tensorflow.device", "line_number": 42, "usage_type": "call"}, {"api_name": "tensorflow.train.string_input_producer", "line_number": 44, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 44, "usage_type": "attribute"}, {"api_name": "tensorflow.TFRecordReader", "line_number": 46, "usage_type": "call"}, {"api_name": "tensorflow.parse_single_example", "line_number": 48, "usage_type": "call"}, {"api_name": "tensorflow.FixedLenFeature", "line_number": 48, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 49, "usage_type": "attribute"}, {"api_name": "tensorflow.FixedLenFeature", "line_number": 49, "usage_type": "call"}, {"api_name": "tensorflow.train.batch", "line_number": 50, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 50, "usage_type": "attribute"}, {"api_name": "tensorflow.train.string_input_producer", "line_number": 53, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 53, "usage_type": "attribute"}, {"api_name": "tensorflow.TFRecordReader", "line_number": 55, "usage_type": "call"}, {"api_name": "tensorflow.parse_single_example", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.FixedLenFeature", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 58, "usage_type": "attribute"}, {"api_name": "tensorflow.FixedLenFeature", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.train.batch", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 59, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 63, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 63, "usage_type": "attribute"}, {"api_name": "tensorflow.placeholder", "line_number": 65, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 65, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_min", "line_number": 67, "usage_type": "call"}, {"api_name": "tensorflow.reduce_max", "line_number": 68, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 70, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 71, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 71, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 74, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 74, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 76, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.relu", "line_number": 77, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 77, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.max_pool", "line_number": 78, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 78, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.histogram", "line_number": 80, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 80, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.histogram", "line_number": 81, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 81, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 83, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 84, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 84, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d", "line_number": 85, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 85, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.relu", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 86, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.max_pool", "line_number": 87, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 87, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.histogram", "line_number": 89, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 89, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.histogram", "line_number": 90, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 90, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 92, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 93, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 94, "usage_type": "call"}, {"api_name": "tensorflow.contrib.layers.xavier_initializer", "line_number": 94, "usage_type": "call"}, {"api_name": "tensorflow.contrib", "line_number": 94, "usage_type": "attribute"}, {"api_name": "tensorflow.Variable", "line_number": 95, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 95, "usage_type": "call"}, {"api_name": "tensorflow.nn.softmax", "line_number": 96, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 96, "usage_type": "attribute"}, {"api_name": "tensorflow.matmul", "line_number": 96, "usage_type": "call"}, {"api_name": "tensorflow.summary.histogram", "line_number": 98, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 98, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.histogram", "line_number": 99, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 99, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.histogram", "line_number": 100, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 100, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_mean", "line_number": 102, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 102, "usage_type": "call"}, {"api_name": "tensorflow.log", "line_number": 102, "usage_type": "call"}, {"api_name": "tensorflow.summary.scalar", "line_number": 103, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 103, "usage_type": "attribute"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 105, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 105, "usage_type": "attribute"}, {"api_name": "tensorflow.argmax", "line_number": 108, "usage_type": "call"}, {"api_name": "tensorflow.equal", "line_number": 109, "usage_type": "call"}, {"api_name": "tensorflow.argmax", "line_number": 109, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 110, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 110, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 110, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.scalar", "line_number": 111, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 111, "usage_type": "attribute"}, {"api_name": "tensorflow.Session", "line_number": 113, "usage_type": "call"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 114, "usage_type": "call"}, {"api_name": "tensorflow.train.Coordinator", "line_number": 116, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 116, "usage_type": "attribute"}, {"api_name": "tensorflow.train.start_queue_runners", "line_number": 117, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 117, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.merge_all", "line_number": 120, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 120, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.FileWriter", "line_number": 121, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 121, "usage_type": "attribute"}, {"api_name": "tqdm.tqdm", "line_number": 133, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 143, "usage_type": "call"}]} +{"seq_id": "451291822", "text": "\"\"\"\n Create an inventory of content on an ArcGIS Portal\n\n\n\"\"\"\nimport os\nfrom collections import defaultdict\nfrom arcgis.gis import GIS\nfrom arcgis.mapping import WebMap\nfrom config import Config\n\nVERSION = '1.0'\npath,exe = os.path.split(__file__)\nmyname = exe + ' ' + VERSION\n\nexclude_esri = '-owner:esri -owner:esri_apps'\n\ndef get_apps(items):\n dtype = {} # A dictionary of all item types\n applications = [\n 'Application',\n 'Code Attachment',\n 'Dashboard',\n 'Desktop Application Template',\n 'Desktop Application',\n 'Web Experience', \n 'Web Mapping Application',\n ]\n for item in items: \n if item.type in applications:\n # if not ('zhunt' in item.owner or \"@CLATSOP\" in item.owner):\n if not item.type in dtype:\n dtype[item.type] = {}\n dtype[item.type][item.id] = item\n return dtype\n\n\ndef generate_html(dtype):\n for itype,items in dtype.items():\n print('

%s

' % itype)\n count = 0\n print(\"\")\n for id in items.keys():\n count += 1\n item = items[id]\n\n url = item.url\n if url:\n if url[0:2] == '//':\n url = \"https:\" + url\n url = 'URL ' % url\n else:\n url = 'no url'\n \n homepage = '%s' % (item.homepage, str(item.title))\n\n print(''.format(count, homepage, item.owner, item.access))\n# keywords = str(', '.join(item.typeKeywords))\n# print('keywords:', keywords)\n print('
{0:3n}{1}{2}{3}
')\n print()\n\n return\n\ndef inventory_maps(gis, query=''):\n q = query + ' ' + exclude_esri\n list_of_maps = gis.content.search(q, item_type='web map', max_items=-1)\n print(\"Maps found %d\" % len(list_of_maps))\n \n # Build a dictionary with each layer as the index\n # and a list of the maps that the layer participates in\n layer_dict = defaultdict(list)\n\n for item in list_of_maps:\n # Look up the layers.\n wm = WebMap(item)\n mapId = wm.item.id\n for l in wm.layers:\n try:\n layer_dict[l.itemId].append(mapId)\n pass\n except Exception as e:\n layer_dict[l.id].append(mapId)\n pass\n\n # Each item is indexed by a layer id and contains a list of the maps containing that id.\n print(layer_dict)\n\n # Now make another dictoinary that is indexed by type.\n dtype = defaultdict(dict)\n for item in list_of_maps: \n dtype[item.type][item.id] = item\n\n print(dtype)\n\n\ndef inventory_services(gis) -> None:\n interesting_services = list()\n interesting_types = ['Map Service', 'Feature Service']\n urls = list()\n\n myservers = gis.admin.servers.list()\n for f in myservers[0].services.folders:\n services = myservers[0].services.list(folder=f)\n print(\"Checking folder=\\\"%s\\\"; %d services.\" % (f, len(services)))\n for s in services:\n properties = s.iteminformation.properties\n try:\n if properties['type'] in interesting_types:\n interesting_services.append(s)\n else:\n print(properties['title'], ':', properties['type'])\n except KeyError:\n if 'GPServer' in s.url:\n continue\n if 'GeometryServer' in s.url:\n continue\n if 'VectorTileServer' in s.url:\n continue\n if 'ImageServer' in s.url:\n continue\n urls.append(s.url)\n\n # These did not have proprties,\n # look like mostly Hosted\n #print(urls)\n\n for s in interesting_services:\n properties = s.iteminformation.properties\n if properties['type'] == 'Map Service':\n print(s.url)\n continue\n else:\n print(properties)\n\nif __name__ == \"__main__\":\n\n # Weird stuff happens if these are not defined.\n assert(Config.PORTAL_URL)\n assert(Config.PORTAL_USER)\n assert(Config.PORTAL_PASSWORD)\n assert(Config.SERVER_URL)\n\n # See arcgis.gis.ContentManager\n # For query definition, refer to http://bitly.com/1fJ8q31\n #q = \"title:Clatsop County Template\"\n #q = \"owner:bwilson@CLATSOP\"\n gis = GIS(Config.PORTAL_URL, Config.PORTAL_USER, Config.PORTAL_PASSWORD)\n #inventory_maps(gis)\n\n types = [\n 'AppBuilder Extension', \n 'Application', \n 'Desktop Application','Desktop Application Template', \n 'Site Application', \n 'Document Link', \n 'Administrative Report',\n 'Form', 'Site Page',\n 'CSV', 'Microsoft Excel', 'Microsoft Word', 'PDF', 'KML',\n 'Map Area', \n 'WMS', 'WMTS',\n 'Code Attachment', 'Code Sample',\n 'Geoprocessing Service',\n 'Dashboard', 'StoryMap', 'StoryMap Theme',\n 'Web Experience', 'Web Mapping Application', \n 'Image', \n 'Geometry Service',\n 'Feature Service',\n 'Shapefile', \n 'Layer Package', 'Tile Package', \n 'SQLite Geodatabase',\n 'Vector Tile Service', 'Vector Tile Package', \n 'Web Scene', 'Service Definition', 'Map Service', \n 'Web Map', ]\n cm = gis.content\n\n q = 'title:EGDB_surveys -owner:esri -owner:esri_apps -owner:esri_nav'\n\n items = cm.search(q, item_type='Feature Service', max_items=-1)\n print(\"Feature Services\", len(items))\n for item in items:\n print(item.title, item.type)\n try:\n for l in item.layers:\n print(l)\n continue\n except Exception as e:\n pass\n\n items = cm.search(q, item_type='Map Service', max_items=-1)\n print(\"Map Services\", len(items))\n for item in items:\n print(item.title)\n for layer in item.layers:\n print(layer, layer.source)\n\n\n inventory_services(gis)\n\n print(\"That's all!\")\n\n# q = \"NOT owner:esri_apps\"\n# items = gis.content.search(q, outside_org=False, max_items=5000)\n# print(\"Items found %d\" % len(items))\n# dtype = get_apps(items)\n# print(dtype)\n# generate_html(dtype)\n\n# That's all!\n", "sub_path": "build_inventory.py", "file_name": "build_inventory.py", "file_ext": "py", "file_size_in_byte": 6334, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "os.path.split", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "collections.defaultdict", "line_number": 72, "usage_type": "call"}, {"api_name": "arcgis.mapping.WebMap", "line_number": 76, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 90, "usage_type": "call"}, {"api_name": "config.Config.PORTAL_URL", "line_number": 139, "usage_type": "attribute"}, {"api_name": "config.Config", "line_number": 139, "usage_type": "name"}, {"api_name": "config.Config.PORTAL_USER", "line_number": 140, "usage_type": "attribute"}, {"api_name": "config.Config", "line_number": 140, "usage_type": "name"}, {"api_name": "config.Config.PORTAL_PASSWORD", "line_number": 141, "usage_type": "attribute"}, {"api_name": "config.Config", "line_number": 141, "usage_type": "name"}, {"api_name": "config.Config.SERVER_URL", "line_number": 142, "usage_type": "attribute"}, {"api_name": "config.Config", "line_number": 142, "usage_type": "name"}, {"api_name": "arcgis.gis.GIS", "line_number": 148, "usage_type": "call"}, {"api_name": "config.Config.PORTAL_URL", "line_number": 148, "usage_type": "attribute"}, {"api_name": "config.Config", "line_number": 148, "usage_type": "name"}, {"api_name": "config.Config.PORTAL_USER", "line_number": 148, "usage_type": "attribute"}, {"api_name": "config.Config.PORTAL_PASSWORD", "line_number": 148, "usage_type": "attribute"}]} +{"seq_id": "185526909", "text": "import random\r\nimport numpy\r\nimport pymysql\r\nfrom deap import base\r\nfrom deap import creator\r\nfrom deap import algorithms\r\nfrom deap import tools\r\nimport matplotlib.pyplot as plt\r\n\r\nclass Produto():\r\n def __init__(self, nome, espaco, peso, importancia, quantidade):\r\n self.nome = nome\r\n self.espaco = espaco\r\n self.peso = peso\r\n self.importancia = importancia\r\n self.quantidade = quantidade\r\n \r\nlista_produtos = []\r\nconexao = pymysql.connect(host='localhost', user='root', passwd='root', db='produtos_negocios')\r\ncursor = conexao.cursor()\r\ncursor.execute('select nome, espaco, peso, importancia, quantidade from produtos')\r\nfor produto in cursor:\r\n #print(produto[3])\r\n for i in range(produto[4]):\r\n lista_produtos.append(Produto(produto[0], produto[1], produto[2], produto[3], produto[4]))\r\n\r\ncursor.close()\r\nconexao.close()\r\n\r\nnomes = []\r\nespacos = []\r\npesos = []\r\nimportancias = []\r\n\r\nfor produto in lista_produtos:\r\n nomes.append(produto.nome)\r\n espacos.append(produto.espaco)\r\n pesos.append(produto.peso)\r\n importancias.append(produto.importancia)\r\n \r\nlimite_espaco = 45000\r\nlimite_peso = 10000\r\n\r\ntoolbox = base.Toolbox()\r\ncreator.create(\"FitnessMax\", base.Fitness, weights=(1.0, ))\r\ncreator.create(\"Individual\", list, fitness=creator.FitnessMax)\r\ntoolbox.register(\"attr_bool\", random.randint, 0, 1)\r\ntoolbox.register(\"individual\", tools.initRepeat, creator.Individual,\r\n toolbox.attr_bool, n=len(espacos))\r\ntoolbox.register(\"population\", tools.initRepeat, list, toolbox.individual)\r\n\r\ndef avaliacao(individual):\r\n nota = 0\r\n soma_espacos = 0\r\n soma_pesos = 0\r\n for i in range(len(individual)):\r\n if individual[i] == 1:\r\n nota += importancias[i]\r\n soma_espacos += espacos[i]\r\n soma_pesos += pesos[i]\r\n if soma_espacos > limite_espaco or soma_pesos > limite_peso:\r\n nota = 1\r\n return nota,\r\n \r\ntoolbox.register(\"evaluate\", avaliacao)\r\ntoolbox.register(\"mate\", tools.cxOnePoint)\r\ntoolbox.register(\"mutate\", tools.mutFlipBit, indpb = 0.05)\r\ntoolbox.register(\"select\", tools.selRoulette)\r\n\r\nif __name__ == \"__main__\":\r\n random.seed()\r\n populacao = toolbox.population(n = 500)\r\n probabilidade_crossover = 2.0\r\n probabilidade_mutacao = 0.05\r\n numero_geracoes = 500\r\n \r\n estatisticas = tools.Statistics(key=lambda individuo: individuo.fitness.values)\r\n estatisticas.register(\"max\", numpy.max)\r\n estatisticas.register(\"min\", numpy.min)\r\n estatisticas.register(\"med\", numpy.mean)\r\n estatisticas.register(\"std\", numpy.std)\r\n \r\n populacao, info = algorithms.eaSimple(populacao, toolbox,\r\n probabilidade_crossover,\r\n probabilidade_mutacao,\r\n numero_geracoes, estatisticas)\r\n melhores = tools.selBest(populacao, 1)\r\n for individuo in melhores:\r\n print(individuo)\r\n print(individuo.fitness)\r\n #print(individuo[1])\r\n soma = 0\r\n soma_pesos = 0\r\n soma_espacos = 0\r\n for i in range(len(lista_produtos)):\r\n if individuo[i] == 1:\r\n soma += importancias[i]\r\n soma_pesos += pesos[i]\r\n soma_espacos += espacos[i]\r\n print(\"Nome: %s\" % (lista_produtos[i].nome))\r\n \r\n print(\"Melhor solução: %s\" % soma)\r\n print(\"Espaço usado: %s\" % soma_espacos)\r\n print(\"Peso total: %s\" % soma_pesos)\r\n \r\n valores_grafico = info.select(\"max\")\r\n plt.plot(valores_grafico)\r\n plt.title(\"Acompanhamento dos valores\")\r\n plt.show()\r\n \r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n", "sub_path": "Trabalho Final.py", "file_name": "Trabalho Final.py", "file_ext": "py", "file_size_in_byte": 3717, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pymysql.connect", "line_number": 19, "usage_type": "call"}, {"api_name": "deap.base.Toolbox", "line_number": 44, "usage_type": "call"}, {"api_name": "deap.base", "line_number": 44, "usage_type": "name"}, {"api_name": "deap.creator.create", "line_number": 45, "usage_type": "call"}, {"api_name": "deap.creator", "line_number": 45, "usage_type": "name"}, {"api_name": "deap.base.Fitness", "line_number": 45, "usage_type": "attribute"}, {"api_name": "deap.base", "line_number": 45, "usage_type": "name"}, {"api_name": "deap.creator.create", "line_number": 46, "usage_type": "call"}, {"api_name": "deap.creator", "line_number": 46, "usage_type": "name"}, {"api_name": "deap.creator.FitnessMax", "line_number": 46, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 47, "usage_type": "attribute"}, {"api_name": "deap.tools.initRepeat", "line_number": 48, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 48, "usage_type": "name"}, {"api_name": "deap.creator.Individual", "line_number": 48, "usage_type": "attribute"}, {"api_name": "deap.creator", "line_number": 48, "usage_type": "name"}, {"api_name": "deap.tools.initRepeat", "line_number": 50, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 50, "usage_type": "name"}, {"api_name": "deap.tools.cxOnePoint", "line_number": 66, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 66, "usage_type": "name"}, {"api_name": "deap.tools.mutFlipBit", "line_number": 67, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 67, "usage_type": "name"}, {"api_name": "deap.tools.selRoulette", "line_number": 68, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 68, "usage_type": "name"}, {"api_name": "random.seed", "line_number": 71, "usage_type": "call"}, {"api_name": "deap.tools.Statistics", "line_number": 77, "usage_type": "call"}, {"api_name": "deap.tools", "line_number": 77, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 78, "usage_type": "attribute"}, {"api_name": "numpy.min", "line_number": 79, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 80, "usage_type": "attribute"}, {"api_name": "numpy.std", "line_number": 81, "usage_type": "attribute"}, {"api_name": "deap.algorithms.eaSimple", "line_number": 83, "usage_type": "call"}, {"api_name": "deap.algorithms", "line_number": 83, "usage_type": "name"}, {"api_name": "deap.tools.selBest", "line_number": 87, "usage_type": "call"}, {"api_name": "deap.tools", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}]} +{"seq_id": "250572561", "text": "\"\"\"\nModule for /vulnerabilities API endpoint\n\"\"\"\n\nimport dateutil.parser\n\nimport connexion\nfrom peewee import Case, fn, JOIN, SQL\n\nfrom common.peewee_model import CveAffectedSystemsCache, CveMetadata, CveImpact\nfrom .base import parse_int_list, GetRequest\nfrom .list_view import ListView\n\n\nclass CvesListView(ListView):\n \"\"\"Database select for CVEs affecting user systems\"\"\"\n\n def __init__(self, list_args, query_args, uri, args={}): # pylint: disable=dangerous-default-value\n join_type = JOIN.INNER\n cve_count = CveAffectedSystemsCache.systems_affected\n if query_args['hide_satellite_managed']:\n cve_count = CveAffectedSystemsCache.direct_systems_affected\n if 'show_all' in args and args['show_all']:\n join_type = JOIN.RIGHT_OUTER\n cve_count = fn.COALESCE(cve_count, 0)\n query = (\n CveAffectedSystemsCache\n .select(cve_count.alias(\"systems_affected\"),\n CveMetadata.cve.alias(\"cve_name\"),\n CveMetadata.cvss3_score,\n CveMetadata.cvss2_score,\n CveMetadata.impact_id,\n CveMetadata.public_date,\n CveMetadata.description.alias(\"cve_description\"))\n .join(CveMetadata, join_type,\n on=((CveAffectedSystemsCache.cve == CveMetadata.cve)\n & (CveAffectedSystemsCache.rh_account == query_args[\"rh_account_number\"])))\n )\n if query_args['hide_satellite_managed'] and not ('show_all' in args and args['show_all']):\n query = query.where(CveAffectedSystemsCache.direct_systems_affected > 0)\n if 'cvss_from' in args and args['cvss_from']:\n query = query.where(CveMetadata.cvss3_score >= args['cvss_from'])\n if 'cvss_to' in args and args['cvss_to']:\n query = query.where(CveMetadata.cvss3_score <= args['cvss_to'])\n if 'public_from' in args and args['public_from']:\n query = query.where(CveMetadata.public_date >= args['public_from'])\n if 'public_to' in args and args['public_to']:\n query = query.where(CveMetadata.public_date <= args['public_to'])\n if 'severity' in args and args['severity']:\n query = query.where(CveMetadata.impact_id << args['severity'])\n query = query.dicts()\n sortable_columns = {\n \"systems_affected\": SQL('systems_affected'),\n \"synopsis\": CveMetadata.cve,\n \"public_date\": CveMetadata.public_date,\n # This assumes we only show one score, and that cvss3 wins over cvss2\n \"cvss_score\": Case(None, ((CveMetadata.cvss3_score.is_null(True), CveMetadata.cvss2_score),), \\\n CveMetadata.cvss3_score),\n \"cvss3_score\": CveMetadata.cvss3_score,\n \"cvss2_score\": CveMetadata.cvss2_score,\n \"impact\": CveMetadata.impact_id\n }\n filterable_columns = {\n \"synopsis\": CveMetadata.cve,\n \"description\": CveMetadata.description\n }\n super(CvesListView, self).__init__(query, sortable_columns, filterable_columns, list_args, args, uri)\n\n\ndef _prepare_impact_id_map():\n impact_id_map = {}\n for row in CveImpact.select().dicts():\n impact_id_map[row['id']] = row['name']\n return impact_id_map\n\n\ndef assign_cves(cves_view):\n \"\"\"Provides CVE details\"\"\"\n result = []\n impact_id_map = _prepare_impact_id_map()\n for row in cves_view:\n entry = dict()\n res = {}\n entry[\"systems_affected\"] = row[\"systems_affected\"]\n entry[\"synopsis\"] = row[\"cve_name\"]\n entry[\"public_date\"] = row[\"public_date\"].isoformat() if row[\"public_date\"] else ''\n entry[\"impact\"] = impact_id_map[row[\"impact_id\"]]\n entry[\"description\"] = row[\"cve_description\"]\n # Store everything we know about CVSS - maybe UI needs to decide what to show\n entry[\"cvss2_score\"] = str(row[\"cvss2_score\"]) if row['cvss2_score'] is not None else ''\n entry[\"cvss3_score\"] = str(row[\"cvss3_score\"]) if row['cvss3_score'] is not None else ''\n res[\"type\"] = \"cve\"\n res[\"id\"] = row[\"cve_name\"]\n res[\"attributes\"] = entry\n result.append(res)\n return result\n\n\nclass GetCves(GetRequest):\n \"\"\"GET to /v1/vulnerabilites/cves\"\"\"\n\n _endpoint_name = r'/v1/vulnerabilities/cves'\n\n @classmethod\n def handle_get(cls, **kwargs):\n \"\"\"Grabs all CVEs user is affected by\"\"\"\n args_desc = [{'arg_name': 'cvss_from', 'convert_func': None},\n {'arg_name': 'cvss_to', 'convert_func': None},\n {'arg_name': 'public_from', 'convert_func': dateutil.parser.parse},\n {'arg_name': 'public_to', 'convert_func': dateutil.parser.parse},\n {'arg_name': 'show_all', 'convert_func': None},\n {'arg_name': 'severity', 'convert_func': parse_int_list}]\n args = cls._parse_arguments(kwargs, args_desc)\n list_arguments = cls._parse_list_arguments(kwargs)\n cves_view = CvesListView(list_arguments, {\"rh_account_number\": connexion.context['user'],\n 'hide_satellite_managed': cls.hide_satellite_managed()},\n connexion.request.path, args)\n res = {}\n res[\"meta\"] = cves_view.get_metadata()\n res[\"links\"] = cves_view.get_pagination_links()\n res[\"data\"] = cls._format_data(list_arguments[\"data_format\"], assign_cves(cves_view))\n return res\n", "sub_path": "manager/vulnerabilities_handler.py", "file_name": "vulnerabilities_handler.py", "file_ext": "py", "file_size_in_byte": 5558, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "list_view.ListView", "line_number": 15, "usage_type": "name"}, {"api_name": "peewee.JOIN.INNER", "line_number": 19, "usage_type": "attribute"}, {"api_name": "peewee.JOIN", "line_number": 19, "usage_type": "name"}, {"api_name": "common.peewee_model.CveAffectedSystemsCache.systems_affected", "line_number": 20, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveAffectedSystemsCache", "line_number": 20, "usage_type": "name"}, {"api_name": "common.peewee_model.CveAffectedSystemsCache.direct_systems_affected", "line_number": 22, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveAffectedSystemsCache", "line_number": 22, "usage_type": "name"}, {"api_name": "peewee.JOIN.RIGHT_OUTER", "line_number": 24, "usage_type": "attribute"}, {"api_name": "peewee.JOIN", "line_number": 24, "usage_type": "name"}, {"api_name": "peewee.fn.COALESCE", "line_number": 25, "usage_type": "call"}, {"api_name": "peewee.fn", "line_number": 25, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 35, "usage_type": "argument"}, {"api_name": "common.peewee_model.CveAffectedSystemsCache.select", "line_number": 27, "usage_type": "call"}, {"api_name": "common.peewee_model.CveAffectedSystemsCache", "line_number": 27, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata.cve.alias", "line_number": 29, "usage_type": "call"}, {"api_name": "common.peewee_model.CveMetadata.cve", "line_number": 29, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 29, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata.cvss3_score", "line_number": 30, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 30, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata.cvss2_score", "line_number": 31, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 31, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata.impact_id", "line_number": 32, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 32, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata.public_date", "line_number": 33, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 33, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata.description.alias", "line_number": 34, "usage_type": "call"}, {"api_name": "common.peewee_model.CveMetadata.description", "line_number": 34, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 34, "usage_type": "name"}, {"api_name": "common.peewee_model.CveAffectedSystemsCache.cve", "line_number": 36, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveAffectedSystemsCache", "line_number": 36, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata.cve", "line_number": 36, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 36, "usage_type": "name"}, {"api_name": "common.peewee_model.CveAffectedSystemsCache.rh_account", "line_number": 37, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveAffectedSystemsCache", "line_number": 37, "usage_type": "name"}, {"api_name": "common.peewee_model.CveAffectedSystemsCache.direct_systems_affected", "line_number": 40, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveAffectedSystemsCache", "line_number": 40, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata.cvss3_score", "line_number": 42, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 42, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata.cvss3_score", "line_number": 44, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 44, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata.public_date", "line_number": 46, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 46, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata.public_date", "line_number": 48, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 48, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata.impact_id", "line_number": 50, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 50, "usage_type": "name"}, {"api_name": "peewee.SQL", "line_number": 53, "usage_type": "call"}, {"api_name": "common.peewee_model.CveMetadata.cve", "line_number": 54, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 54, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata.public_date", "line_number": 55, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 55, "usage_type": "name"}, {"api_name": "peewee.Case", "line_number": 57, "usage_type": "call"}, {"api_name": "common.peewee_model.CveMetadata.cvss3_score.is_null", "line_number": 57, "usage_type": "call"}, {"api_name": "common.peewee_model.CveMetadata.cvss3_score", "line_number": 57, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 57, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata.cvss2_score", "line_number": 57, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata.cvss3_score", "line_number": 58, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 58, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata.cvss3_score", "line_number": 59, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 59, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata.cvss2_score", "line_number": 60, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 60, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata.impact_id", "line_number": 61, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 61, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata.cve", "line_number": 64, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 64, "usage_type": "name"}, {"api_name": "common.peewee_model.CveMetadata.description", "line_number": 65, "usage_type": "attribute"}, {"api_name": "common.peewee_model.CveMetadata", "line_number": 65, "usage_type": "name"}, {"api_name": "common.peewee_model.CveImpact.select", "line_number": 72, "usage_type": "call"}, {"api_name": "common.peewee_model.CveImpact", "line_number": 72, "usage_type": "name"}, {"api_name": "base.GetRequest", "line_number": 99, "usage_type": "name"}, {"api_name": "dateutil.parser.parser", "line_number": 109, "usage_type": "attribute"}, {"api_name": "dateutil.parser", "line_number": 109, "usage_type": "name"}, {"api_name": "dateutil.parser.parser", "line_number": 110, "usage_type": "attribute"}, {"api_name": "dateutil.parser", "line_number": 110, "usage_type": "name"}, {"api_name": "base.parse_int_list", "line_number": 112, "usage_type": "name"}, {"api_name": "connexion.context", "line_number": 115, "usage_type": "attribute"}, {"api_name": "connexion.request", "line_number": 117, "usage_type": "attribute"}]} +{"seq_id": "287149564", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Sat Apr 28 17:57:31 2018\r\n\r\n@author: Kirill\r\n\"\"\"\r\n\r\n\r\nimport csv\r\nimport math\r\nimport os\r\n\r\nimport pandas as pd\r\nimport numpy as np\r\n\r\nfrom sklearn.model_selection import train_test_split\r\nfrom sklearn.model_selection import cross_val_score\r\nfrom sklearn import ensemble\r\nfrom sklearn.ensemble import VotingClassifier\r\nfrom sklearn.linear_model import SGDClassifier\r\nfrom sklearn.linear_model import LogisticRegression\r\nfrom sklearn.naive_bayes import GaussianNB\r\nfrom sklearn.discriminant_analysis import QuadraticDiscriminantAnalysis\r\nfrom sklearn.tree import DecisionTreeClassifier\r\nfrom sklearn.svm import SVC\r\nfrom sklearn.neural_network import MLPClassifier\r\nfrom sklearn.neighbors import KNeighborsClassifier\r\nfrom sklearn import preprocessing\r\n\r\n\r\n\r\n#url_data = 'C:/Users/Кирилл/retnna/Data/csv_data/raw_nr/'\r\n#file_name = 'Search_Andrii2_all_gaze.csv'\r\n\r\nurl_data = 'C:/Users/Кирилл/retnna/Data/csv_data/notreading/'\r\nfile_name = 'pct2_olga.csv' #'faces_anastasia.csv'\r\ndata_path = ['C:/Users/Кирилл/retnna/Work_data/notreading/', 'C:/Users/Кирилл/retnna/Work_data/reading/']\r\n\r\n#data = url_data + file_name\r\n\r\n\r\ndef read_files(data_path):\r\n file_number = 0\r\n female_file_to_work = list()\r\n male_file_to_work = list()\r\n for path in data_path:\r\n for filename in os.listdir(path):\r\n if filename.startswith(\"male_\"):\r\n file_number += 1\r\n male_file_to_work.append(path + filename)\r\n elif filename.startswith(\"female_\"):\r\n file_number += 1\r\n female_file_to_work.append(path + filename)\r\n else:\r\n continue\r\n return (female_file_to_work, male_file_to_work, file_number)\r\n\r\n\r\ndef read_data(url_data, file_name):\r\n f_open = open(url_data + file_name)\r\n return pd.read_csv(f_open, sep='\\t', encoding='cp1251')\r\n\r\n\r\ndef read_data_from_file(url_file_name):\r\n with open(url_file_name, 'r') as file:\r\n reader = csv.reader(file, delimiter='\\t')\r\n next(reader)\r\n for line in reader:\r\n yield line\r\n\r\n\r\ndef temp_trans(temp_tank):\r\n from copy import deepcopy\r\n temp = deepcopy(temp_tank)\r\n return temp\r\n\r\n\r\ndef separator(url_file_name):\r\n temp_result_list =list()\r\n result_list = list()\r\n generator = read_data_from_file(url_file_name)\r\n line = next(generator)\r\n FPOGDS = line[3]\r\n FPOGID = line[5]\r\n FPOGDS_current = FPOGDS\r\n FPOGID_current = FPOGID\r\n while line:\r\n if FPOGDS_current == FPOGDS and FPOGID_current == FPOGID:\r\n temp_result_list.append(line)\r\n FPOGDS = line[3]\r\n FPOGID = line[5]\r\n else:\r\n result_list.append(temp_trans(temp_result_list))\r\n temp_result_list.clear()\r\n FPOGDS = line[3]\r\n FPOGID = line[5]\r\n FPOGDS_current = FPOGDS\r\n FPOGID_current = FPOGID\r\n try:\r\n line = next(generator)\r\n except StopIteration:\r\n line = None\r\n return result_list\r\n\r\n\r\ndef all_data_reading(file_name_list):\r\n result_list = list()\r\n for url_file_name in file_name_list:\r\n result_list += separator(url_file_name)\r\n return result_list\r\n\r\n\r\ndef distance(x, y, x0, y0):\r\n dist = math.sqrt((x-x0)**2 + (y-y0)**2)\r\n return dist\r\n \r\n\r\ndef one_line_dist(session_list):\r\n temp = 0\r\n for line in session_list:\r\n if float(line[9]) != 0.0 and float(line[6]) != 0.0:\r\n temp += distance(\r\n float(line[1]), \r\n float(line[2]), \r\n float(line[7]), \r\n float(line[8]))\r\n else:\r\n break\r\n return temp\r\n\r\n\r\ndef one_line_speed(session_list):\r\n temp = list()\r\n for line in session_list:\r\n if float(line[9]) != 0.0 and float(line[6]) != 0.0:\r\n dist = distance(\r\n float(line[1]), \r\n float(line[2]), \r\n float(line[7]), \r\n float(line[8]))\r\n time = float(line[4])\r\n temp.append(dist/time)\r\n else:\r\n break\r\n try:\r\n sum(temp)/len(temp)\r\n except ZeroDivisionError:\r\n return 0\r\n else:\r\n return sum(temp)/len(temp)\r\n\r\n\r\ndef one_line_std_speed(session_list):\r\n temp = list()\r\n for line in session_list:\r\n if float(line[9]) != 0.0 and float(line[6]) != 0.0:\r\n dist = distance(\r\n float(line[1]), \r\n float(line[2]), \r\n float(line[7]), \r\n float(line[8]))\r\n time = float(line[4])\r\n temp.append(dist/time)\r\n else:\r\n break\r\n arr = np.array(temp)\r\n return np.std(arr, axis=0)\r\n \r\n\r\ndef one_line_average_length(session_list):\r\n temp = list()\r\n for line in session_list:\r\n if float(line[9]) != 0.0 and float(line[6]) != 0.0:\r\n dist = distance(\r\n float(line[1]), \r\n float(line[2]), \r\n float(line[7]), \r\n float(line[8]))\r\n temp.append(dist)\r\n try:\r\n sum(temp)/len(temp)\r\n except ZeroDivisionError:\r\n return 0\r\n else:\r\n return sum(temp)/len(temp)\r\n\r\n\r\ndef one_line_normal_average_length(session_list):\r\n temp = list()\r\n for line in session_list:\r\n if float(line[9]) != 0.0 and float(line[6]) != 0.0:\r\n temp.append(math.fabs(float(line[1]) - float(line[7])))\r\n try:\r\n sum(temp)/len(temp)\r\n except ZeroDivisionError:\r\n return 0\r\n else:\r\n return sum(temp)/len(temp) \r\n\r\n\r\ndef one_line_normal_vertical_average_length(session_list):\r\n temp = list()\r\n for line in session_list:\r\n if float(line[9]) != 0.0 and float(line[6]) != 0.0:\r\n temp.append(math.fabs(float(line[2]) - float(line[8])))\r\n try:\r\n sum(temp)/len(temp)\r\n except ZeroDivisionError:\r\n return 0\r\n else:\r\n return sum(temp)/len(temp)\r\n\r\n\r\ndef one_line_std_sinus(session_list):\r\n temp = list()\r\n for line in session_list:\r\n if float(line[9]) != 0.0 and float(line[6]) != 0.0:\r\n hypotenuse = distance(\r\n float(line[1]), \r\n float(line[2]), \r\n float(line[7]), \r\n float(line[8]))\r\n catheter = float(line[2]) - float(line[8])\r\n temp.append(catheter/hypotenuse)\r\n else:\r\n break\r\n# try:\r\n# sum(temp)/len(temp)\r\n# except ZeroDivisionError:\r\n# return 0\r\n# else:\r\n# return sum(temp)/len(temp)\r\n return np.std(np.array(temp), axis=0)\r\n\r\n\r\ndef inter_line_dist(session_list):\r\n temp = 0\r\n index_max = len(session_list)\r\n for index in range(1, index_max):\r\n if float(session_list[index][9]) != 0 and float(session_list[index][6]) != 0:\r\n temp += distance(\r\n float(session_list[index - 1][7]), \r\n float(session_list[index - 1][8]),\r\n float(session_list[index][1]),\r\n float(session_list[index][2]))\r\n else:\r\n break \r\n return temp\r\n\r\n\r\ndef inter_line_speed(session_list):\r\n temp = list()\r\n index_max = len(session_list)\r\n for index in range(1, index_max):\r\n if float(session_list[index][9]) != 0.0 and float(session_list[index][6]) != 0.0:\r\n dist = distance(\r\n float(session_list[index - 1][7]), \r\n float(session_list[index - 1][8]),\r\n float(session_list[index][1]),\r\n float(session_list[index][2]))\r\n time = float(session_list[index][0]) - float(session_list[index - 1][0])\r\n temp.append(dist/time)\r\n else:\r\n break\r\n try:\r\n sum(temp)/len(temp)\r\n except ZeroDivisionError:\r\n return 0\r\n else:\r\n return sum(temp)/len(temp)\r\n\r\n\r\ndef inter_line_std_speed(session_list):\r\n temp = list()\r\n index_max = len(session_list)\r\n for index in range(1, index_max):\r\n if float(session_list[index][9]) != 0.0 and float(session_list[index][6]) != 0.0:\r\n dist = distance(\r\n float(session_list[index - 1][7]), \r\n float(session_list[index - 1][8]),\r\n float(session_list[index][1]),\r\n float(session_list[index][2]))\r\n time = float(session_list[index][0]) - float(session_list[index - 1][0])\r\n temp.append(dist/time)\r\n else:\r\n break\r\n arr = np.array(temp)\r\n return np.std(arr, axis=0)\r\n\r\n\r\ndef inter_line_average_length(session_list):\r\n temp = list()\r\n index_max = len(session_list)\r\n for index in range(1, index_max):\r\n if float(session_list[index][9]) != 0.0 and float(session_list[index][6]) != 0.0:\r\n dist = distance(\r\n float(session_list[index - 1][7]), \r\n float(session_list[index - 1][8]),\r\n float(session_list[index][1]),\r\n float(session_list[index][2]))\r\n temp.append(dist)\r\n try:\r\n sum(temp)/len(temp)\r\n except ZeroDivisionError:\r\n return 0\r\n else:\r\n return sum(temp)/len(temp)\r\n\r\n\r\ndef inter_line_normal_average_length(session_list):\r\n temp = list()\r\n index_max = len(session_list)\r\n for index in range(1, index_max):\r\n if float(session_list[index][9]) != 0.0 and float(session_list[index][6]) != 0.0:\r\n temp.append(math.fabs(float(session_list[index - 1][7]) - float(session_list[index][1])))\r\n try:\r\n sum(temp)/len(temp)\r\n except ZeroDivisionError:\r\n return 0\r\n else:\r\n return sum(temp)/len(temp)\r\n\r\n\r\ndef inter_line_normal_vertical_average_length(session_list):\r\n temp = list()\r\n index_max = len(session_list)\r\n for index in range(1, index_max):\r\n if float(session_list[index][9]) != 0.0 and float(session_list[index][6]) != 0.0:\r\n temp.append(math.fabs(float(session_list[index - 1][8]) - float(session_list[index][2])))\r\n try:\r\n sum(temp)/len(temp)\r\n except ZeroDivisionError:\r\n return 0\r\n else:\r\n return sum(temp)/len(temp)\r\n\r\n\r\ndef inter_line_std_sinus(session_list):\r\n temp = list()\r\n index_max = len(session_list)\r\n for index in range(1, index_max):\r\n if float(session_list[index][9]) != 0.0 and float(session_list[index][6]) != 0.0:\r\n hypotenuse = distance(\r\n float(session_list[index - 1][7]), \r\n float(session_list[index - 1][8]),\r\n float(session_list[index][1]),\r\n float(session_list[index][2]))\r\n catheter = float(session_list[index - 1][8]) - float(session_list[index][2])\r\n temp.append(catheter/hypotenuse)\r\n# try:\r\n# sum(temp)/len(temp)\r\n# except ZeroDivisionError:\r\n# return 0\r\n# else:\r\n# return sum(temp)/len(temp)\r\n \r\n return np.std(np.array(temp), axis=0)\r\n\r\n\r\ndef total_time_calc(session_list):\r\n temp = 0\r\n index_max = len(session_list)\r\n for index in range(1, index_max):\r\n if float(session_list[index][9]) != 0.0 and float(session_list[index][6]) != 0.0:\r\n temp += (float(session_list[index][0]) - float(session_list[index - 1][0]))\r\n else:\r\n break\r\n return temp\r\n\r\n\r\ndef fix_time_calc(session_list):\r\n temp = 0\r\n index_max = len(session_list)\r\n for index in range(1, index_max):\r\n if float(session_list[index][9]) != 0.0 and float(session_list[index][6]) != 0.0:\r\n temp += math.fabs(float(session_list[index][4]) - float(session_list[index - 1][4]))\r\n else:\r\n break\r\n return temp\r\n\r\n\r\ndef nominal_session_time(session_list):\r\n return math.fabs(float(session_list[-1][0]) - float(session_list[0][0]))\r\n\r\n \r\ndef session_collapsing(session_list):\r\n horizontal_distance = one_line_dist(session_list)\r\n vertical_distance = inter_line_dist(session_list)\r\n average_horizontal_distance = one_line_average_length(session_list)\r\n average_verticall_distance = inter_line_average_length(session_list)\r\n total_time = total_time_calc(session_list)#float(session_list[-1][7])\r\n fix_time_change = fix_time_calc(session_list)\r\n inline_speed = one_line_speed(session_list)\r\n between_line_speed = inter_line_speed(session_list)\r\n inline_normal_horizontal_distance = one_line_normal_average_length(session_list)\r\n inline_normal_vertical_distance = one_line_normal_vertical_average_length(session_list)\r\n interline_normal_horizontal_distance = inter_line_normal_average_length(session_list)\r\n session_time = nominal_session_time(session_list)\r\n inter_line_sinus = inter_line_std_sinus(session_list)\r\n one_line_sinus = one_line_std_sinus(session_list)\r\n inline_std_speed = one_line_std_speed(session_list)\r\n interline_std_speed = inter_line_std_speed(session_list)\r\n# total_blinks = math.fabs(float(session_list[0][12]) - float(session_list[-1][12]))\r\n return [horizontal_distance, \r\n vertical_distance,\r\n average_horizontal_distance,\r\n average_verticall_distance,\r\n total_time, \r\n fix_time_change, \r\n inline_speed,\r\n between_line_speed,\r\n inline_normal_horizontal_distance,\r\n interline_normal_horizontal_distance,\r\n inline_normal_vertical_distance,\r\n session_time,\r\n inter_line_sinus,\r\n one_line_sinus,\r\n inline_std_speed,\r\n interline_std_speed]\r\n# interline_std_speed,\r\n# total_blinks]\r\n\r\n\r\ndef result_data_cooking(res, param):\r\n res_list = list()\r\n for session in res:\r\n res_list.append([param] + session_collapsing(session))\r\n return res_list\r\n \r\n\r\ndef main(data_path):\r\n female_file_to_work, male_file_to_work, file_number = read_files(data_path)\r\n data_list = [female_file_to_work, male_file_to_work]\r\n for data in data_list:\r\n if data_list.index(data) == 0:\r\n param = 0\r\n female_list = result_data_cooking(all_data_reading(data), param)\r\n elif data_list.index(data) == 1:\r\n param = 1\r\n male_list = result_data_cooking(all_data_reading(data), param)\r\n return female_list + male_list\r\n\r\n\r\ndef final_list_clean(final_list):\r\n result = list()\r\n for line in final_list:\r\n if float(line[1]) != 0.0:\r\n if float(line[2]) != 0.0:\r\n if float(line[5]) != 0.0:\r\n if float(line[6]) != 0.0:\r\n result.append(line)\r\n return result\r\n \r\n\r\ndef list_to_frame(final_list):\r\n df = pd.DataFrame(final_list, columns=['Target',\r\n 'X0', \r\n 'X1', \r\n 'X2', \r\n 'X3', \r\n 'X4', \r\n 'X5', \r\n 'X6', \r\n 'X7', \r\n 'X8',\r\n 'X9',\r\n 'X10',\r\n 'X11',\r\n 'X12',\r\n 'X13',\r\n 'X14',\r\n 'X15'])\r\n# 'X15',\r\n# 'X16'])\r\n return df\r\n\r\n\r\ndef soft_voting(df_res, y):\r\n clf2 = ensemble.RandomForestClassifier(n_estimators=5000,\r\n max_depth=9,\r\n max_features=1, \r\n random_state=42, \r\n n_jobs=-1)\r\n clf3 = ensemble.GradientBoostingClassifier(n_estimators=5000, \r\n learning_rate=0.0001, \r\n max_depth=9, \r\n random_state=42)\r\n clf1 = GaussianNB(priors=None)\r\n clf4 = SGDClassifier(max_iter=10000,\r\n alpha=0.001,\r\n tol=1e-4, \r\n shuffle=True, \r\n penalty='l2', \r\n loss='log')\r\n clf5 = DecisionTreeClassifier(max_depth=5)\r\n clf6 = SVC(kernel=\"linear\", C=0.025, probability=True)\r\n clf7 = SVC(gamma=2, C=1, probability=True)\r\n clf8 = LogisticRegression(C=1e5)\r\n clf9 = MLPClassifier(hidden_layer_sizes=(35, 30, 25, 20, 15, 10), alpha=0.0001, max_iter=40000, activation='logistic')\r\n clf10 = ensemble.AdaBoostClassifier()\r\n clf11 = KNeighborsClassifier(3)\r\n clf12 = QuadraticDiscriminantAnalysis() #GaussianProcessClassifier(1.0 * RBF(1.0))\r\n eclf = VotingClassifier(estimators=[('gau', clf1), ('rfc', clf2), \r\n ('gbs', clf3), ('sgdc', clf4),\r\n ('dtc', clf5), ('svm_linear', clf6),\r\n ('svm_gamma', clf7), ('LogReg', clf8),\r\n ('Neuro_mlp', clf9), ('AdaBoost', clf10),\r\n ('KNeighbors', clf11), ('QDA', clf12)], \r\n voting='soft', weights=[1,1.2,1.2,1,1.2,1,1,1.2,1.2,1,1,1])\r\n for clf, label in zip([clf1, clf2, clf3, clf4, clf5, clf6, clf7, clf8, clf9, clf10, clf11, clf12, eclf], \r\n ['GaussianNB', \r\n 'RandomForestClassifier', \r\n 'GradientBoosting', \r\n 'SGDClassifier', \r\n 'DecisionTreeClassifier',\r\n 'SVC_linear',\r\n 'SVC_gamma',\r\n 'LogReg',\r\n 'Neuro_MLP',\r\n 'AdaBoost',\r\n 'KNeighbors',\r\n 'QuadraticDA',\r\n 'Ensemble']):\r\n scores = cross_val_score(clf, df_res, y, cv=10, n_jobs=-1, scoring='roc_auc')\r\n print(\"ROC_AUC scoring: %0.3f (+/- %0.3f) [%s]\" % (scores.mean(), scores.std(), label))\r\n\r\n\r\ndef select_params_for_SGDClassifier(df_res, y):\r\n alpha_list = [0.00001, 0.0001, 0.001, 0.01, 0.1, 1.0, 10, 100]\r\n MAX_iter = [100, 1000, 2000, 5000, 10000]\r\n loss_list = ['hinge', 'log', 'modified_huber', 'squared_hinge', 'perceptron']\r\n for alph in alpha_list:\r\n for it_r in MAX_iter:\r\n for loss_cr in loss_list:\r\n label = 'SGDClassifier'\r\n clf = SGDClassifier(alpha=alph, max_iter=it_r, loss=loss_cr)\r\n scores = cross_val_score(clf, df_res, y, cv=10, n_jobs=-1, scoring='roc_auc')\r\n print(\"ROC_AUC scoring: %0.3f (+/- %0.3f) [%s] [%s] [%s] [%s]\" % (scores.mean(), \r\n scores.std(), \r\n label, \r\n alph, \r\n it_r, \r\n loss_cr))\r\n\r\n\r\ndef select_params_for_LogisticRegression(df_res, y):\r\n CN_list = [0.0001, 0.001, 0.01, 0.1, 1.0, 10, 100, 1000, 10000, 100000]\r\n MAX_iter = [10, 100, 1000, 10000, 100000]\r\n solvers = ['newton-cg', 'lbfgs', 'sag']\r\n for CN in CN_list:\r\n for it_r in MAX_iter:\r\n for solve in solvers:\r\n label = 'LogReg'\r\n clf = LogisticRegression(C=CN, max_iter=it_r, solver=solve)\r\n scores = cross_val_score(clf, df_res, y, cv=10, n_jobs=-1, scoring='roc_auc')\r\n print(\"ROC_AUC scoring: %0.3f (+/- %0.3f) [%s] [%s] [%s] [%s]\" % (scores.mean(), \r\n scores.std(), \r\n label, \r\n CN, \r\n it_r, \r\n solve))\r\n\r\n\r\nif __name__ == '__main__':\r\n# frame = read_data(url_data, file_name)\r\n final_list = main(data_path)\r\n cleaned_final = final_list_clean(final_list)\r\n total_dataframe = list_to_frame(cleaned_final)\r\n\r\n y = total_dataframe['Target']\r\n X = total_dataframe.drop('Target', 1)\r\n\r\n \r\n## X_proc = (X - X.mean()) / X.std()\r\n# \r\n# poly = preprocessing.PolynomialFeatures(interaction_only=True)\r\n# poly.fit(X)\r\n# X_proc0 = poly.transform(X)\r\n#\r\n## min_max_scaler = preprocessing.MinMaxScaler()\r\n## X_proc = min_max_scaler.fit_transform(X)\r\n#\r\n## standart_scaler = preprocessing.StandardScaler() #0.629\r\n## X_proc = standart_scaler.fit_transform(X_proc0)\r\n# \r\n# quantily = preprocessing.QuantileTransformer(output_distribution='uniform') #0.635\r\n# X_proc = quantily.fit_transform(X_proc0)\r\n# \r\n## quantily = preprocessing.QuantileTransformer(output_distribution='normal') #0.634\r\n## X_proc = quantily.fit_transform(X_proc0)\r\n# \r\n# X_proc = np.delete(X_proc, np.s_[0:1], axis=1)\r\n# \r\n## normalizer = preprocessing.Normalizer() #0.628\r\n## X_proc = normalizer.fit_transform(X_proc0)\r\n# \r\n## X_proc = preprocessing.RobustScaler(quantile_range=(25, 75)).fit_transform(X) #0.626\r\n X_proc = X\r\n \r\n X_train, X_test, y_train, y_test = train_test_split(X_proc, y, \r\n test_size=0.3, \r\n random_state=42, \r\n shuffle=True)\r\n soft_voting(X_train, y_train)\r\n \r\n print('*******************')\r\n print('\\n')\r\n \r\n standart_scaler = preprocessing.StandardScaler() #0.629\r\n X_proc = standart_scaler.fit_transform(X)\r\n X_train, X_test, y_train, y_test = train_test_split(X_proc, y, \r\n test_size=0.3, \r\n random_state=42, \r\n shuffle=True) \r\n soft_voting(X_train, y_train) \r\n# select_C_for_LogReg(X_train, y_train)\r\n \r\n\r\n#parameters={\r\n#'learning_rate': [\"constant\", \"invscaling\", \"adaptive\"],\r\n#'hidden_layer_sizes': [(100,1), (100,2), (100,3)],\r\n#'alpha': [10.0 ** -np.arange(1, 7)],\r\n#'activation': [\"logistic\", \"relu\", \"Tanh\"]\r\n#}\r\n#\r\n#clf = gridSearchCV(estimator=MLPClassifier,param_grid=parameters,n_jobs=-1,verbose=2,cv=10)\r\n#https://datascience.stackexchange.com/questions/19768/how-to-implement-pythons-mlpclassifier-with-gridsearchcv\r\n#http://scikit-learn.org/stable/auto_examples/neural_networks/plot_mlp_alpha.html\r\n\r\n\r\n \r\n# female_file_to_work, male_file_to_work, file_number = read_files(data_path)\r\n# data_list = [female_file_to_work, male_file_to_work]\r\n# for data in data_list:\r\n# if data_list.index(data) == 0:\r\n# param = 0\r\n# female_list = result_data_cooking(all_data_reading(data), param)\r\n# elif data_list.index(data) == 1:\r\n# param = 1\r\n# male_list = result_data_cooking(all_data_reading(data), param)\r\n# final_list = female_list + male_list\r\n \r\n# res = all_data_reading(file_names_list)\r\n# final_list = result_data_cooking(res)\r\n# frame = read_data(url_data, file_name)\r\n \r\n# clf4 = SGDClassifier(max_iter=10000, \r\n# tol=1e-4, \r\n# shuffle=True, \r\n# penalty='l2', \r\n# loss='log')\r\n# columns=['X0', 'X1', 'X2', 'X3', 'X4', 'X5', 'X6', 'X7', 'X8']\r\n# X1 = X[columns]", "sub_path": "attempt_0_3.py", "file_name": "attempt_0_3.py", "file_ext": "py", "file_size_in_byte": 24096, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "os.listdir", "line_number": 47, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 61, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 66, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 74, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 166, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 167, "usage_type": "call"}, {"api_name": "math.fabs", "line_number": 192, "usage_type": "call"}, {"api_name": "math.fabs", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 233, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 233, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 287, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 288, "usage_type": "call"}, {"api_name": "math.fabs", "line_number": 315, "usage_type": "call"}, {"api_name": "math.fabs", "line_number": 329, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 357, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 357, "usage_type": "call"}, {"api_name": "math.fabs", "line_number": 376, "usage_type": "call"}, {"api_name": "math.fabs", "line_number": 383, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 456, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 479, "usage_type": "call"}, {"api_name": "sklearn.ensemble", "line_number": 479, "usage_type": "name"}, {"api_name": "sklearn.ensemble.GradientBoostingClassifier", "line_number": 484, "usage_type": "call"}, {"api_name": "sklearn.ensemble", "line_number": 484, "usage_type": "name"}, {"api_name": "sklearn.naive_bayes.GaussianNB", "line_number": 488, "usage_type": "call"}, {"api_name": "sklearn.linear_model.SGDClassifier", "line_number": 489, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 495, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 496, "usage_type": "call"}, {"api_name": "sklearn.svm.SVC", "line_number": 497, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 498, "usage_type": "call"}, {"api_name": "sklearn.neural_network.MLPClassifier", "line_number": 499, "usage_type": "call"}, {"api_name": "sklearn.ensemble.AdaBoostClassifier", "line_number": 500, "usage_type": "call"}, {"api_name": "sklearn.ensemble", "line_number": 500, "usage_type": "name"}, {"api_name": "sklearn.neighbors.KNeighborsClassifier", "line_number": 501, "usage_type": "call"}, {"api_name": "sklearn.discriminant_analysis.QuadraticDiscriminantAnalysis", "line_number": 502, "usage_type": "call"}, {"api_name": "sklearn.ensemble.VotingClassifier", "line_number": 503, "usage_type": "call"}, {"api_name": "sklearn.model_selection.cross_val_score", "line_number": 524, "usage_type": "call"}, {"api_name": "sklearn.linear_model.SGDClassifier", "line_number": 536, "usage_type": "call"}, {"api_name": "sklearn.model_selection.cross_val_score", "line_number": 537, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LogisticRegression", "line_number": 554, "usage_type": "call"}, {"api_name": "sklearn.model_selection.cross_val_score", "line_number": 555, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 600, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 609, "usage_type": "call"}, {"api_name": "sklearn.preprocessing", "line_number": 609, "usage_type": "name"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 611, "usage_type": "call"}]} +{"seq_id": "168598430", "text": "import requests\r\nimport pandas as pd\r\nimport numpy as np\r\nfrom bs4 import BeautifulSoup\r\nimport datetime\r\n\r\ndef money_ticket():\r\n\r\n def get_money_ticket5d(target,date1,date2):\r\n URL_5d=\"https://concords.moneydj.com/z/zc/zcn/zcn.djhtm?a=\"+target+\"&c=\"+date1+\"&d=\"+date2\r\n \r\n response=requests.get(URL_5d)\r\n soup=BeautifulSoup(response.text,\"html.parser\")\r\n #print(soup.prettify())\r\n text=soup.findAll(\"td\",class_=\"t3n1\")\r\n date=soup.findAll(\"td\",class_=\"t3n0\")\r\n day=[]\r\n ticket=[]\r\n \r\n money=[]\r\n money_usage=[]\r\n for i in range(len(text)):\r\n if(i%14==4):\r\n money.append(text[i].text)\r\n if(i%14==6):\r\n money_usage.append(text[i].text)\r\n if(i%14==11):\r\n ticket.append(text[i].text)\r\n for i in range(len(date)):\r\n date[i]=date[i].text\r\n del date[0]\r\n df={'date':date,'target':target,'money':money,'money_usage':money_usage,'ticket':ticket}\r\n #return df\r\n return pd.DataFrame(data=df)\r\n '''\r\n table=pd.read_excel(\"融資融券.xlsx\",usecols=\"A:G\")\r\n \r\n table['代號']=table['代號'].fillna('null')\r\n table=table[~table['代號'].isin(['null'])]\r\n \r\n print(table['代號'])\r\n today=datetime.date.today()\r\n\r\n for i in range(len(table)):\r\n \r\n if(not table.iloc[i]['更新日期']==today):\r\n table.loc[i,'代號']=int(table.loc[i,'代號'])\r\n \r\n print(table.iloc[i]['代號'])\r\n back=get_money_ticket5d(str(table.iloc[i]['代號']))\r\n back1=get_money_ticket20d(str(table.iloc[i]['代號']))\r\n table.loc[i,'5日融資']=back[0]\r\n table.loc[i,'5日融券']=back[1]\r\n table.loc[i,'20日融資']=back1[0]\r\n table.loc[i,'20日融券']=back1[1]\r\n table.loc[i,'更新日期']=today\r\n if(i%100==0): \r\n table.to_excel(\"融資融券.xlsx\",index=False)\r\n table.loc[i,'更新日期']=today\r\n table.to_excel(\"融資融券.xlsx\",index=False)\r\n '''\r\n list=pd.read_excel(\"上市公司列表.xlsx\",usecols=\"A\")\r\n list['代號']=list['代號'].fillna('null')\r\n list=list[~list['代號'].isin(['null'])]\r\n table=get_money_ticket5d(str(list.loc[0,'代號']),\"2021-3-3\",\"2021-6-4\")\r\n for i in range(1,len(list)):\r\n target=str(list.loc[i,'代號'])\r\n print(target)\r\n \r\n table=table.append(get_money_ticket5d(str(list.loc[i,'代號']),\"2021-3-3\",\"2021-6-4\"))\r\n print(table)\r\n table.to_excel(\"融資融券.xlsx\",index=False)\r\n \r\nmoney_ticket()\r\n\r\n\r\n\r\n", "sub_path": "backend/融資融券.py", "file_name": "融資融券.py", "file_ext": "py", "file_size_in_byte": 2672, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "requests.get", "line_number": 12, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 13, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 34, "usage_type": "call"}, {"api_name": "pandas.read_excel", "line_number": 62, "usage_type": "call"}]} +{"seq_id": "374020101", "text": "\"\"\"\nStation text-based search\n\"\"\"\n\n# stdlib\nfrom functools import lru_cache\nfrom typing import Iterable, List, Tuple\n\n# library\nfrom rapidfuzz import fuzz, process # type: ignore\n\n# module\nfrom avwx.load_utils import LazyCalc\nfrom avwx.station.meta import STATIONS\nfrom avwx.station.station import Station, station_filter\n\n\nTYPE_ORDER = [\n \"large_airport\",\n \"medium_airport\",\n \"small_airport\",\n \"seaplane_base\",\n \"heliport\",\n \"balloonport\",\n \"weather_station\",\n]\n\n\ndef _format_search(airport: dict, keys: Iterable[str]) -> str:\n values = [airport.get(k) for k in keys]\n return \" - \".join(k for k in values if k)\n\n\ndef _build_corpus() -> List[str]:\n keys = (\"icao\", \"iata\", \"city\", \"state\", \"name\")\n return [_format_search(s, keys) for s in STATIONS.values()]\n\n\n_CORPUS = LazyCalc(_build_corpus)\n\n\ndef _sort_key(result: Tuple[Station, int]) -> Tuple[int, ...]:\n station, score = result\n try:\n type_order = TYPE_ORDER.index(station.type)\n except ValueError:\n type_order = 10\n return (score, 10 - type_order)\n\n\n@lru_cache(maxsize=128)\ndef search(\n text: str,\n limit: int = 10,\n is_airport: bool = False,\n sends_reports: bool = True,\n) -> List[Station]:\n \"\"\"Text search for stations against codes, name, city, and state\n\n Results may be shorter than limit value\n \"\"\"\n results = process.extract(\n text, _CORPUS.value, limit=limit * 20, scorer=fuzz.token_set_ratio\n )\n results = [(Station.from_icao(k[:4]), s) for k, s, _ in results]\n results.sort(key=_sort_key, reverse=True)\n results = [s for s, _ in results if station_filter(s, is_airport, sends_reports)]\n return results[:limit] if len(results) > limit else results\n", "sub_path": "avwx/station/search.py", "file_name": "search.py", "file_ext": "py", "file_size_in_byte": 1721, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "typing.Iterable", "line_number": 29, "usage_type": "name"}, {"api_name": "avwx.station.meta.STATIONS.values", "line_number": 36, "usage_type": "call"}, {"api_name": "avwx.station.meta.STATIONS", "line_number": 36, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 34, "usage_type": "name"}, {"api_name": "avwx.load_utils.LazyCalc", "line_number": 39, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 42, "usage_type": "name"}, {"api_name": "avwx.station.station.Station", "line_number": 42, "usage_type": "name"}, {"api_name": "rapidfuzz.process.extract", "line_number": 62, "usage_type": "call"}, {"api_name": "rapidfuzz.process", "line_number": 62, "usage_type": "name"}, {"api_name": "rapidfuzz.fuzz.token_set_ratio", "line_number": 63, "usage_type": "attribute"}, {"api_name": "rapidfuzz.fuzz", "line_number": 63, "usage_type": "name"}, {"api_name": "avwx.station.station.Station.from_icao", "line_number": 65, "usage_type": "call"}, {"api_name": "avwx.station.station.Station", "line_number": 65, "usage_type": "name"}, {"api_name": "avwx.station.station.station_filter", "line_number": 67, "usage_type": "call"}, {"api_name": "functools.lru_cache", "line_number": 51, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 57, "usage_type": "name"}, {"api_name": "avwx.station.station.Station", "line_number": 57, "usage_type": "name"}]} +{"seq_id": "474554072", "text": "#!/usr/bin/env python\n\n##############################################################################\n# Copyright (c) 2017 Rajesh Kudaka.\n#\n# All rights reserved. This program and the accompanying materials\n# are made available under the terms of the Apache License, Version 2.0\n# which accompanies this distribution, and is available at\n# http://www.apache.org/licenses/LICENSE-2.0\n##############################################################################\n\n# Unittest for yardstick.benchmark.core.report\n\nfrom __future__ import print_function\n\nfrom __future__ import absolute_import\n\nimport unittest\nimport uuid\n\ntry:\n from unittest import mock\nexcept ImportError:\n import mock\n\nfrom yardstick.benchmark.core import report\nfrom yardstick.cmd.commands import change_osloobj_to_paras\n\nFAKE_YAML_NAME = 'fake_name'\nFAKE_TASK_ID = str(uuid.uuid4())\nFAKE_DB_FIELDKEYS = [{'fieldKey': 'fake_key'}]\nFAKE_TIME = '0000-00-00T00:00:00.000000Z'\nFAKE_DB_TASK = [{'fake_key': 0.000, 'time': FAKE_TIME}]\nFAKE_TIMESTAMP = ['fake_time']\nDUMMY_TASK_ID = 'aaaaaa-aaaaaaaa-aaaaaaaaaa-aaaaaa'\n\n\nclass ReportTestCase(unittest.TestCase):\n\n def setUp(self):\n super(ReportTestCase, self).setUp()\n self.param = change_osloobj_to_paras({})\n self.param.yaml_name = [FAKE_YAML_NAME]\n self.param.task_id = [FAKE_TASK_ID]\n self.rep = report.Report()\n\n @mock.patch('yardstick.benchmark.core.report.Report._get_tasks')\n @mock.patch('yardstick.benchmark.core.report.Report._get_fieldkeys')\n @mock.patch('yardstick.benchmark.core.report.Report._validate')\n def test_generate_success(self, mock_valid, mock_keys, mock_tasks):\n mock_tasks.return_value = FAKE_DB_TASK\n mock_keys.return_value = FAKE_DB_FIELDKEYS\n self.rep.generate(self.param)\n mock_valid.assert_called_once_with(FAKE_YAML_NAME, FAKE_TASK_ID)\n self.assertEqual(1, mock_tasks.call_count)\n self.assertEqual(1, mock_keys.call_count)\n\n def test_invalid_yaml_name(self):\n self.assertRaisesRegexp(ValueError, \"yaml*\", self.rep._validate,\n 'F@KE_NAME', FAKE_TASK_ID)\n\n def test_invalid_task_id(self):\n self.assertRaisesRegexp(ValueError, \"task*\", self.rep._validate,\n FAKE_YAML_NAME, DUMMY_TASK_ID)\n\n @mock.patch('api.utils.influx.query')\n def test_task_not_found(self, mock_query):\n mock_query.return_value = []\n self.rep.yaml_name = FAKE_YAML_NAME\n self.rep.task_id = FAKE_TASK_ID\n self.assertRaisesRegexp(KeyError, \"Task ID\", self.rep._get_fieldkeys)\n self.assertRaisesRegexp(KeyError, \"Task ID\", self.rep._get_tasks)\n", "sub_path": "tests/unit/benchmark/core/test_report.py", "file_name": "test_report.py", "file_ext": "py", "file_size_in_byte": 2670, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "uuid.uuid4", "line_number": 30, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 38, "usage_type": "attribute"}, {"api_name": "yardstick.cmd.commands.change_osloobj_to_paras", "line_number": 42, "usage_type": "call"}, {"api_name": "yardstick.benchmark.core.report.Report", "line_number": 45, "usage_type": "call"}, {"api_name": "yardstick.benchmark.core.report", "line_number": 45, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 47, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 48, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 49, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 66, "usage_type": "call"}]} +{"seq_id": "39746856", "text": "from shutil import which\nfrom colorama import Fore, Back, Style\nfrom os import path\nimport os.path\nimport socket\nimport subprocess\nimport sys\nimport socket\nimport os\nimport argparse\nimport time\n\n\nbanner = \"\"\"\n\n\n\n ██████ ██▓███ ▓██ ██▓ ██░ ██ █ ██ ███▄ █ ▄▄▄█████▓\n▒██ ▒ ▓██░ ██▒▒██ ██▒▓██░ ██▒ ██ ▓██▒ ██ ▀█ █ ▓ ██▒ ▓▒\n░ ▓██▄ ▓██░ ██▓▒ ▒██ ██░▒██▀▀██░▓██ ▒██░▓██ ▀█ ██▒▒ ▓██░ ▒░\n ▒ ██▒▒██▄█▓▒ ▒ ░ ▐██▓░░▓█ ░██ ▓▓█ ░██░▓██▒ ▐▌██▒░ ▓██▓ ░ \n▒██████▒▒▒██▒ ░ ░ ░ ██▒▓░░▓█▒░██▓▒▒█████▓ ▒██░ ▓██░ ▒██▒ ░ \n▒ ▒▓▒ ▒ ░▒▓▒░ ░ ░ ██▒▒▒ ▒ ░░▒░▒░▒▓▒ ▒ ▒ ░ ▒░ ▒ ▒ ▒ ░░ \n░ ░▒ ░ ░░▒ ░ ▓██ ░▒░ ▒ ░▒░ ░░░▒░ ░ ░ ░ ░░ ░ ▒░ ░ \n░ ░ ░ ░░ ▒ ▒ ░░ ░ ░░ ░ ░░░ ░ ░ ░ ░ ░ ░ v 1.0 \n ░ ░ ░ ░ ░ ░ ░ ░ \n ░ ░ by c0deNinja\n\n\n\n\"\"\"\n\nprint(Fore.RED + banner)\nprint(Fore.WHITE)\n\ndef commands(cmd):\n try:\n subprocess.check_call(cmd, shell=True)\n except:\n pass\n\nparser = argparse.ArgumentParser()\ngroup = parser.add_mutually_exclusive_group()\n\ngroup.add_argument('-sv', '--save', action='store',\n help=\"save output to file\")\n\nparser.add_argument('-s',\n type=str, help='scan for subdomains',\n metavar='domain.com')\n\nparser.add_argument('-j',\n type=str, help='find javascript files',\n metavar='domain.com')\n\nparser.add_argument('-d', '--dns',\n type=str, help='scan for dns records',\n metavar='domain.com')\n\nparser.add_argument('-p', '--probe',\n type=str, help='probe domains.',\n metavar='domains.txt')\n\nparser.add_argument('-a', '--aquatone',\n type=str, help='take screenshots of domains.',\n metavar='domains.txt')\n\nparser.add_argument('-r', '--redirects',\n type=str, help='links getting redirected',\n metavar='domains.txt')\n\n\n\nargs = parser.parse_args()\n\nif args.s:\n if args.save:\n print(Fore.CYAN + \"Saving output to {}...\".format(args.save))\n cmd = f\"subfinder -d {args.s}\"\n p = subprocess.Popen(cmd, shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)\n out, err = p.communicate()\n out = out.decode() \n with open(f\"{args.save}\", \"w\") as subfinder:\n subfinder.writelines(out)\n if path.exists(f\"{args.save}\"):\n print(Fore.GREEN + \"DONE!\")\n if not path.exists(f\"{args.save}\"):\n print(Fore.RED + \"ERROR!\")\n sys.exit(1)\n cmd = f\"./scripts/spotter.sh {args.s} | uniq | sort\"\n p = subprocess.Popen(cmd, shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)\n spotterout, err = p.communicate()\n spotterout = spotterout.decode()\n with open(f\"{args.save}\", \"a\") as spotter:\n spotter.writelines(spotterout)\n cmd = f\"./scripts/certsh.sh {args.s} | uniq | sort\"\n p = subprocess.Popen(cmd, shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)\n certshout, err = p.communicate()\n certshout = certshout.decode()\n else:\n commands(f\"subfinder -d {args.s}\")\n commands(f\"./scripts/spotter.sh {args.s} | uniq | sort\")\n commands(f\"./scripts/certsh.sh {args.s} | uniq | sort\") \n\nif args.j:\n if args.save:\n print(Fore.CYAN + \"Saving output to {}\".format(args.save))\n commands(f\"echo {args.j} | waybackurls | grep '\\\\.js$' | uniq | sort >> {args.save}\")\n commands(f\"echo {args.j} | gau | grep -Eo 'https?://\\\\S+?\\\\.js' | anew >> {args.save}\")\n if path.exists(f\"{args.save}\"):\n print(Fore.GREEN + \"DONE!\")\n if not path.exists(f\"{args.save}\"):\n print(Fore.RED + \"ERROR!\")\n else:\n commands(f\"echo {args.j} | waybackurls | grep '\\\\.js$' | anew\")\n commands(f\"echo {args.j} | gau | grep -Eo 'https?://\\\\S+?\\\\.js' | anew\")\n\n\nif args.dns:\n if args.save:\n print(Fore.CYAN + \"Saving output to {}...\".format(args.save))\n commands(f\"cat {args.dns} | dnsx -silent -a -resp >> {args.save}\")\n commands(f\"cat {args.dns} | dnsx -silent -ns -resp >> {args.save}\")\n commands(f\"cat {args.dns} | dnsx -silent -cname -resp >> {args.save}\")\n else:\n print(Fore.CYAN + \"Printing A records...\\n\")\n time.sleep(2)\n commands(f\"cat {args.dns} | dnsx -silent -a -resp\\n\")\n print(Fore.CYAN + \"Printing NS Records...\\n\")\n time.sleep(2)\n commands(f\"cat {args.dns} | dnsx -silent -ns -resp\\n\")\n print(Fore.CYAN + \"Printing CNAME records...\\n\")\n time.sleep(2)\n commands(f\"cat {args.dns} | dnsx -silent -cname -resp\\n\") \n \n\nif args.probe:\n if args.save:\n print(Fore.CYAN + \"Saving output to {}...\".format(args.save))\n commands(f'cat {args.probe} | httprobe | anew >> {args.save}')\n if path.exists(f\"{args.save}\"):\n print(Fore.GREEN + \"DONE!\")\n if not path.exists(f\"{args.save}\"):\n print(Fore.RED + \"ERROR!\")\n else:\n commands(f'sudo cat {args.probe} | httprobe | anew') \n\nif args.aquatone:\n commands(f\"cat {args.aquatone} | aquatone\")\n\nif args.redirects:\n if args.save:\n print(Fore.CYAN + \"Saving output to {}}..\".format(args.save))\n commands(f\"cat {args.redirects} | httpx -silent -location -mc 301,302 | anew >> {args.save}\")\n if path.exists(f\"{args.save}\"):\n print(Fore.GREEN + \"DONE!\")\n if not path.exists(f\"{args.save}\"):\n print(Fore.RED + \"ERROR!\")\n else:\n commands(f\"cat {args.redirects} | httpx -silent -location -mc 301,302\") \n\n\n \n\n\n\n", "sub_path": "spyhunt.py", "file_name": "spyhunt.py", "file_ext": "py", "file_size_in_byte": 6349, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "colorama.Fore.RED", "line_number": 33, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 33, "usage_type": "name"}, {"api_name": "colorama.Fore.WHITE", "line_number": 34, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 34, "usage_type": "name"}, {"api_name": "subprocess.check_call", "line_number": 38, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 42, "usage_type": "call"}, {"api_name": "colorama.Fore.CYAN", "line_number": 78, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 78, "usage_type": "name"}, {"api_name": "subprocess.Popen", "line_number": 80, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 80, "usage_type": "attribute"}, {"api_name": "subprocess.STDOUT", "line_number": 80, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path", "line_number": 85, "usage_type": "name"}, {"api_name": "colorama.Fore.GREEN", "line_number": 86, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 86, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path", "line_number": 87, "usage_type": "name"}, {"api_name": "colorama.Fore.RED", "line_number": 88, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 88, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 89, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 91, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 91, "usage_type": "attribute"}, {"api_name": "subprocess.STDOUT", "line_number": 91, "usage_type": "attribute"}, {"api_name": "subprocess.Popen", "line_number": 97, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 97, "usage_type": "attribute"}, {"api_name": "subprocess.STDOUT", "line_number": 97, "usage_type": "attribute"}, {"api_name": "colorama.Fore.CYAN", "line_number": 107, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 107, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 110, "usage_type": "call"}, {"api_name": "os.path", "line_number": 110, "usage_type": "name"}, {"api_name": "colorama.Fore.GREEN", "line_number": 111, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 111, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path", "line_number": 112, "usage_type": "name"}, {"api_name": "colorama.Fore.RED", "line_number": 113, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 113, "usage_type": "name"}, {"api_name": "colorama.Fore.CYAN", "line_number": 121, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 121, "usage_type": "name"}, {"api_name": "colorama.Fore.CYAN", "line_number": 126, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 126, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 127, "usage_type": "call"}, {"api_name": "colorama.Fore.CYAN", "line_number": 129, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 129, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 130, "usage_type": "call"}, {"api_name": "colorama.Fore.CYAN", "line_number": 132, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 132, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 133, "usage_type": "call"}, {"api_name": "colorama.Fore.CYAN", "line_number": 139, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 139, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 141, "usage_type": "call"}, {"api_name": "os.path", "line_number": 141, "usage_type": "name"}, {"api_name": "colorama.Fore.GREEN", "line_number": 142, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 142, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 143, "usage_type": "call"}, {"api_name": "os.path", "line_number": 143, "usage_type": "name"}, {"api_name": "colorama.Fore.RED", "line_number": 144, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 144, "usage_type": "name"}, {"api_name": "colorama.Fore.CYAN", "line_number": 153, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 153, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 155, "usage_type": "call"}, {"api_name": "os.path", "line_number": 155, "usage_type": "name"}, {"api_name": "colorama.Fore.GREEN", "line_number": 156, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 156, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 157, "usage_type": "call"}, {"api_name": "os.path", "line_number": 157, "usage_type": "name"}, {"api_name": "colorama.Fore.RED", "line_number": 158, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 158, "usage_type": "name"}]} +{"seq_id": "4543404", "text": "\"\"\"doc.\"\"\"\nfrom flask import Blueprint, flash, redirect, render_template, request, url_for\n\nfrom app import db\nfrom customers.forms import CustomerForm, PortForm\nfrom helpers import object_list\nfrom models import Customer, Port, Tag\n\ncustomers = Blueprint('customers', __name__, template_folder='templates')\n\n\n@customers.route('/')\ndef index():\n \"\"\"doc.\"\"\"\n customers = Customer.query.order_by(Customer.c_created_timestamp.desc())\n return customer_list('customers/index.html', customers)\n\n\n@customers.route('//')\ndef detail(slug):\n \"\"\"doc.\"\"\"\n customer = Customer.query.filter(Customer.slug == slug).first_or_404()\n return render_template('customers/detail.html', customer=customer)\n\n\n@customers.route('/tags//')\ndef tag_detail(slug):\n \"\"\"doc.\"\"\"\n tag = Tag.query.filter(Tag.slug == slug).first_or_404()\n customers = tag.customers.order_by(Customer.c_created_timestamp.desc())\n return customer_list('customers/tag_detail.html', customers, tag=tag)\n\n\n@customers.route('/tags/')\ndef tag_index():\n \"\"\"doc.\"\"\"\n tags = Tag.query.order_by(Tag.name)\n return object_list('customers/tag_index.html', tags)\n\n\ndef customer_list(template, query, **context):\n \"\"\"doc.\"\"\"\n valid_statuses = (Customer.STATUS_PUBLIC, Customer.STATUS_DRAFT)\n query = query.filter(Customer.status.in_(valid_statuses))\n if request.args.get('q'):\n search = request.args['q']\n query = query.filter(\n (Customer.body.contains(search)) |\n (Customer.title.contains(search)))\n\n return object_list(template, query, **context)\n\n\ndef get_customer_or_404(slug):\n \"\"\"doc.\"\"\"\n valid_statuses = (Customer.STATUS_PUBLIC, Customer.STATUS_DRAFT)(Customer.query\n .filter(\n (Customer.slug == slug) &\n (Customer.status.in_(valid_statuses)))\n .first_or_404())\n\n\n@customers.route('/create/', methods=['GET', 'POST'])\ndef create():\n \"\"\"doc.\"\"\"\n if request.method == 'POST':\n form = CustomerForm(request.form)\n if form.validate():\n customer = form.save_customer(Customer())\n db.session.add(customer)\n db.session.commit()\n flash('customer \"%s\" created successfully.' %\n customer.title, 'success')\n return redirect(url_for('customers.detail', slug=customer.slug))\n else:\n form = CustomerForm()\n\n return render_template('customers/create.html', form=form)\n\n\n@customers.route('//edit/', methods=['GET', 'POST'])\ndef edit(slug):\n \"\"\"doc.\"\"\"\n customer = Customer.query.filter(Customer.slug == slug).first_or_404()\n if request.method == 'POST':\n form = CustomerForm(request.form, obj=customer)\n if form.validate():\n customer = form.save_customer(customer)\n db.session.add(customer)\n db.session.commit()\n flash('customer \"%s\" created successfully.' %\n customer.title, 'success')\n return redirect(url_for('customers.detail', slug=customer.slug))\n else:\n form = CustomerForm(obj=customer)\n\n return render_template('customers/edit.html', customer=customer, form=form)\n\n\n@customers.route('//delete/', methods=['GET', 'POST'])\ndef delete(slug):\n \"\"\"doc.\"\"\"\n customer = Customer.query.filter(Customer.slug == slug).first_or_404()\n if request.method == 'POST':\n customer.status = customer.STATUS_DELETED\n db.session.add(customer)\n db.session.commit()\n flash('customer \"%s\" created successfully.' %\n customer.title, 'success')\n return redirect(url_for('customers.index'))\n\n return render_template('customers/delete.html', customer=customer)\n", "sub_path": "app/customers/blueprint.py", "file_name": "blueprint.py", "file_ext": "py", "file_size_in_byte": 3938, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "customers.forms", "line_number": 9, "usage_type": "name"}, {"api_name": "flask.Blueprint", "line_number": 9, "usage_type": "call"}, {"api_name": "customers.forms", "line_number": 15, "usage_type": "name"}, {"api_name": "models.Customer.query.order_by", "line_number": 15, "usage_type": "call"}, {"api_name": "models.Customer.query", "line_number": 15, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 15, "usage_type": "name"}, {"api_name": "models.Customer.c_created_timestamp.desc", "line_number": 15, "usage_type": "call"}, {"api_name": "models.Customer.c_created_timestamp", "line_number": 15, "usage_type": "attribute"}, {"api_name": "customers.forms", "line_number": 16, "usage_type": "argument"}, {"api_name": "customers.forms.route", "line_number": 12, "usage_type": "call"}, {"api_name": "customers.forms", "line_number": 12, "usage_type": "name"}, {"api_name": "models.Customer.query.filter", "line_number": 22, "usage_type": "call"}, {"api_name": "models.Customer.query", "line_number": 22, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 22, "usage_type": "name"}, {"api_name": "models.Customer.slug", "line_number": 22, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 23, "usage_type": "call"}, {"api_name": "customers.forms.route", "line_number": 19, "usage_type": "call"}, {"api_name": "customers.forms", "line_number": 19, "usage_type": "name"}, {"api_name": "models.Tag.query.filter", "line_number": 29, "usage_type": "call"}, {"api_name": "models.Tag.query", "line_number": 29, "usage_type": "attribute"}, {"api_name": "models.Tag", "line_number": 29, "usage_type": "name"}, {"api_name": "models.Tag.slug", "line_number": 29, "usage_type": "attribute"}, {"api_name": "customers.forms", "line_number": 30, "usage_type": "name"}, {"api_name": "models.Customer.c_created_timestamp.desc", "line_number": 30, "usage_type": "call"}, {"api_name": "models.Customer.c_created_timestamp", "line_number": 30, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 30, "usage_type": "name"}, {"api_name": "customers.forms", "line_number": 31, "usage_type": "argument"}, {"api_name": "customers.forms.route", "line_number": 26, "usage_type": "call"}, {"api_name": "customers.forms", "line_number": 26, "usage_type": "name"}, {"api_name": "models.Tag.query.order_by", "line_number": 37, "usage_type": "call"}, {"api_name": "models.Tag.query", "line_number": 37, "usage_type": "attribute"}, {"api_name": "models.Tag", "line_number": 37, "usage_type": "name"}, {"api_name": "models.Tag.name", "line_number": 37, "usage_type": "attribute"}, {"api_name": "helpers.object_list", "line_number": 38, "usage_type": "call"}, {"api_name": "customers.forms.route", "line_number": 34, "usage_type": "call"}, {"api_name": "customers.forms", "line_number": 34, "usage_type": "name"}, {"api_name": "models.Customer.STATUS_PUBLIC", "line_number": 43, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 43, "usage_type": "name"}, {"api_name": "models.Customer.STATUS_DRAFT", "line_number": 43, "usage_type": "attribute"}, {"api_name": "models.Customer.status.in_", "line_number": 44, "usage_type": "call"}, {"api_name": "models.Customer.status", "line_number": 44, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 44, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 45, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 45, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 45, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 46, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 46, "usage_type": "name"}, {"api_name": "models.Customer.body.contains", "line_number": 48, "usage_type": "call"}, {"api_name": "models.Customer.body", "line_number": 48, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 48, "usage_type": "name"}, {"api_name": "models.Customer.title.contains", "line_number": 49, "usage_type": "call"}, {"api_name": "models.Customer.title", "line_number": 49, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 49, "usage_type": "name"}, {"api_name": "helpers.object_list", "line_number": 51, "usage_type": "call"}, {"api_name": "models.Customer.STATUS_PUBLIC", "line_number": 56, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 56, "usage_type": "name"}, {"api_name": "models.Customer.STATUS_DRAFT", "line_number": 56, "usage_type": "attribute"}, {"api_name": "models.Customer.query.filter", "line_number": 56, "usage_type": "call"}, {"api_name": "models.Customer.query", "line_number": 56, "usage_type": "attribute"}, {"api_name": "models.Customer.slug", "line_number": 58, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 58, "usage_type": "name"}, {"api_name": "models.Customer.status.in_", "line_number": 59, "usage_type": "call"}, {"api_name": "models.Customer.status", "line_number": 59, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 59, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 66, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 66, "usage_type": "name"}, {"api_name": "customers.forms.CustomerForm", "line_number": 67, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 67, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 67, "usage_type": "name"}, {"api_name": "models.Customer", "line_number": 69, "usage_type": "call"}, {"api_name": "app.db.session.add", "line_number": 70, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 70, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 70, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 71, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 71, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 71, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 72, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 74, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 74, "usage_type": "call"}, {"api_name": "customers.forms.CustomerForm", "line_number": 76, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 78, "usage_type": "call"}, {"api_name": "customers.forms.route", "line_number": 63, "usage_type": "call"}, {"api_name": "customers.forms", "line_number": 63, "usage_type": "name"}, {"api_name": "models.Customer.query.filter", "line_number": 84, "usage_type": "call"}, {"api_name": "models.Customer.query", "line_number": 84, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 84, "usage_type": "name"}, {"api_name": "models.Customer.slug", "line_number": 84, "usage_type": "attribute"}, {"api_name": "flask.request.method", "line_number": 85, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 85, "usage_type": "name"}, {"api_name": "customers.forms.CustomerForm", "line_number": 86, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 86, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 86, "usage_type": "name"}, {"api_name": "app.db.session.add", "line_number": 89, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 89, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 89, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 90, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 90, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 90, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 91, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 93, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 93, "usage_type": "call"}, {"api_name": "customers.forms.CustomerForm", "line_number": 95, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 97, "usage_type": "call"}, {"api_name": "customers.forms.route", "line_number": 81, "usage_type": "call"}, {"api_name": "customers.forms", "line_number": 81, "usage_type": "name"}, {"api_name": "models.Customer.query.filter", "line_number": 103, "usage_type": "call"}, {"api_name": "models.Customer.query", "line_number": 103, "usage_type": "attribute"}, {"api_name": "models.Customer", "line_number": 103, "usage_type": "name"}, {"api_name": "models.Customer.slug", "line_number": 103, "usage_type": "attribute"}, {"api_name": "flask.request.method", "line_number": 104, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 104, "usage_type": "name"}, {"api_name": "app.db.session.add", "line_number": 106, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 106, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 106, "usage_type": "name"}, {"api_name": "app.db.session.commit", "line_number": 107, "usage_type": "call"}, {"api_name": "app.db.session", "line_number": 107, "usage_type": "attribute"}, {"api_name": "app.db", "line_number": 107, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 108, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 110, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 110, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 112, "usage_type": "call"}, {"api_name": "customers.forms.route", "line_number": 100, "usage_type": "call"}, {"api_name": "customers.forms", "line_number": 100, "usage_type": "name"}]} +{"seq_id": "19366683", "text": "import requests \nimport random\nimport re\nfrom requests.exceptions import ReadTimeout,RequestException\n\npro = ['219.141.153.41:80','118.190.95.35:9001','118.190.95.43:9001'] \n\nhead = {\n 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/58.0.3029.110 Safari/537.36'\n}\nurl = 'https://book.douban.com' # 你用于测试自己ip的网站\ntry:\n\trequest = requests.get(url, proxies={'http':random.choice(pro)}, headers=head,timeout=5) # 让问这个网页 随机生成一个ip\n\trequest.encoding = request.apparent_encoding # 设置编码 encoding 返回的是请求头编码 apparent_encoding 是从内容网页中分析出的响应内容编码方式\n\thtml = request.text # 输出返回的内容\n\tprint(request)\n\t# pattern = re.compile('.*?href=\"(.*?)\"\\s+title=\"(.*?)\".*?more-meta.*?author\">(.*?).*?year\">(.*?).*?',re.S)\n\tresult = re.findall('
.*?(.*?)',html,re.S)\n\t# result = re.findall(pattern,html)\n\tprint(result)\n\nexcept ReadTimeout:\n\tprint(\"ReadTimeout\")\nexcept RequestException:\n\tprint(\"RequestException\")", "sub_path": "爬虫/豆瓣图书.py", "file_name": "豆瓣图书.py", "file_ext": "py", "file_size_in_byte": 1124, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "requests.get", "line_number": 13, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 13, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 18, "usage_type": "call"}, {"api_name": "re.S", "line_number": 18, "usage_type": "attribute"}, {"api_name": "requests.exceptions.ReadTimeout", "line_number": 22, "usage_type": "name"}, {"api_name": "requests.exceptions.RequestException", "line_number": 24, "usage_type": "name"}]} +{"seq_id": "545963484", "text": "from importlib import import_module\n\n\nSITES = {\n 'qiubai': 'http://www.qiushibaike.com/',\n 'baisi': 'http://www.budejie.com/'\n}\n\ndef AnyTransfer(key, sites=SITES):\n m = import_module('.'.join(['extractors', key]))\n url = sites[key]\n results = m.transfer(url)\n return results\n\n\n\ndef PrepareData():\n results = []\n for (k,v) in SITES.items():\n results = results + AnyTransfer(k)\n return results\n\ndef main():\n print( PrepareData() )\n\nif __name__ == '__main__':\n main()", "sub_path": "spider/xiaohua/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 504, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "importlib.import_module", "line_number": 10, "usage_type": "call"}]} +{"seq_id": "32631903", "text": "from elasticsearch import Elasticsearch\nfrom elasticsearch.helpers import *\n\ndef load_elastic(index_name, hosts):\n es = Elasticsearch()\n actions = []\n\n for host in hosts:\n for i in host.iterrows():\n action = {\n \"_index\": index_name,\n \"_type\": \"appcompat\",\n \"_source\": i[1].to_json(date_format='iso')\n }\n\n actions.append(action)\n\n if len(actions) > 0:\n bulk(es, actions)", "sub_path": "loader/load_elastic.py", "file_name": "load_elastic.py", "file_ext": "py", "file_size_in_byte": 473, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "elasticsearch.Elasticsearch", "line_number": 5, "usage_type": "call"}]} +{"seq_id": "379028675", "text": "from O365 import Account\nfrom O365 import MSGraphProtocol\nfrom O365.message import Message\nfrom O365.mailbox import MailBox\nimport time\nimport select\n#credentials = ('my_client_id', 'my_client_secret')\n#credentials = ('42a56f23-f453-441d-88e3-3acb123e511c', '1Gi_.VGTn15FQrjMe~pkQ1_B2_zHeXW58W')\ntry:\n credentials = ('3f708f99-56a8-42e3-a606-70dcdc7dfc92',\n 'YHqg8CVt~62w1rm0BJp-jbT3Vm1NIy5~~D')\n#scp = ['https://graph.microsoft.com/.default']\n\n scp = [\"https://graph.microsoft.com/Mail.Send\"]\n# the default protocol will be Microsoft Graph\n# the default authentication method will be \"on behalf of a user\"\n\n#account = Account(credentials)\n#account = Account(credentials, auth_flow_type='credentials', tenant_id='29d0dd74-80aa-401f-8e81-a864b29c30bd')\n account = Account(credentials, auth_flow_type='credentials',\n tenant_id='57192160-2d87-4183-b297-a03d2e69f430')\n\n# if account.authenticate():\n# if account.authenticate(scopes=scp):\n\n#protocol_graph = MSGraphProtocol()\n#scopes_graph = protocol.get_scopes_for('message all')\n# scopes here are: ['https://graph.microsoft.com/Mail.ReadWrite', 'https://graph.microsoft.com/Mail.Send']\n#account = Account(credentials, scopes=scopes_graph)\n\n fromAddress = \"demouser1@smtpclientdemo.onmicrosoft.com\"\n toAddressList = [\"lokesh.padmanabhaiah@toshiba-tsip.com\"]\n\n if account.authenticate():\n print(\"Bingo!! Authenticated!\")\n\n #m = account.new_message(\"demouser1@smtpclientdemo.onmicrosoft.com\")\n m = account.new_message(fromAddress)\n print(\"New Message m is created... m type is \", type(m))\n m.sender.address = fromAddress\n m.to.add(toAddressList)\n m.body = 'Test Mail from python OAUTH2 api. With Attachment'\n m.subject = \"Test mail from script_123\"\n print(\"m.attachment type is \", type(m.attachments))\n # m.attachments.add(\"./attachments/send1.pdf\")\n # m.send()\n\n print(\"Email sent successfully...\")\n mailbox = account.mailbox(fromAddress)\n inbox = mailbox.inbox_folder()\n\n print(\"Messages from Inbox are as follows...\")\n print(\"-----------------------------------\")\n index = 0\n for message in inbox.get_messages():\n index = index + 1\n # print(message)\n #print(\"Message Body \",\tmessage.body)\n print(\"To: \",\tmessage.to.add)\n print(\"Subject:\", message.subject)\n print(\"Attachment \")\n print(\"Attachment type is \", type(message.attachments))\n print(\"Saving attachment as eml file... \")\n fileName = \"Bingo\" + str(index) + \".eml\"\n # message.save_as_eml(fileName)\n print(\"Mime content is \", type(message.get_mime_content()))\n print(\"Message type is \", type(message))\n print(\"Messaget attachment type is \", type(message.attachments.download_attachments(\n )), \"Value is \", message.attachments.download_attachments())\n print(\"************************************\")\n if message.attachments.download_attachments():\n print(\"you somehow need to get the total count... to save all attachments\", type(\n message.attachments))\n for fileAttachment in message.attachments:\n print(\"INside loop..... to save attachments..... \")\n fileAttachment.save()\n print(\"All attachment for this mail is saved.... now deleting this....\")\n # message.delete()\n else:\n print(\"There is no attachment for this email\")\n #print(\"Now will delete this message.... \")\n #print(\"To Delete this email, uncomment next line in script...\")\n # message.delete()\n def polling_wait(x): return select.select([], [], [], x)\n polling_wait(5)\n print(\"-----------------------------------\")\n\n #sent_folder = mailbox.sent_folder()\n # for message in sent_folder.get_messages():\n # print(message)\n else:\n print(\"NOT AUTHENTICATED....\")\nexcept NameError:\n print(\"var not defined/wrong address\")\nexcept RuntimeError:\n print(\"run time error\")\nexcept SyntaxError as err:\n raise SyntaxError(\"Synatx\") from err\nexcept OSError:\n print(\"file not found\")\nexcept ImportError:\n print(\" Import error\")\nexcept KeyboardInterrupt as key:\n raise KeyboardInterrupt(\"Keyboard Interrupt\") from key\nexcept TimeoutError:\n print(\"Time Out error\")\nexcept OSError as exc:\n raise OSError(\"failed to save file\") from exc\nexcept FileNotFoundError:\n print(\"file not found\")\nexcept:\n print(\"error\")\n", "sub_path": "1.py", "file_name": "1.py", "file_ext": "py", "file_size_in_byte": 4619, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "O365.Account", "line_number": 20, "usage_type": "call"}, {"api_name": "select.select", "line_number": 84, "usage_type": "call"}]} +{"seq_id": "469664737", "text": "#!/usr/bin/env python3\n\nimport pigpio\nimport time\n\nfrom phy import phytogpio\nfrom motor import motor\n\nfrom pidcontroller import PIDController\n\ntry:\n shm = open('/dev/shm/MPU', 'r')\nexcept:\n print(\"MPUD doesn't seem to be started???\")\n exit()\n\npi = pigpio.pi()\n\n# When MPUD is running, reads the values from SHM location\ndef getMPUVals():\n global shm\n shm.seek(0, 0)\n output = shm.readline()\n a,b,c = output.split()\n return float(a),float(b),float(c)\n\nm1 = motor(pi, 22, 13, 15)\nm2 = motor(pi, 24, 18, 16)\nm1.rotate(0)\nm2.rotate(0)\n\ndef botMoveForward(power):\n m1.rotate(power)\n m2.rotate(power)\n\ndef botMoveBackward(power):\n # power = power * 1.5\n m1.rotate(power)\n m2.rotate(power)\n\ndef botEquilibrium():\n m1.rotate(0)\n m2.rotate(0)\n\nPID = PIDController(P=50, I=0.01, D=1)\n#import pdb; pdb.set_trace()\ntry:\n while True:\n a,b,c = getMPUVals()\n print(f'{a} {b} {c}')\n PIDx = PID.step(b)\n print(PIDx)\n if PIDx < 0.0:\n botMoveBackward(PIDx)\n elif PIDx > 0.0:\n botMoveForward(PIDx)\n else:\n botEquilibrium()\n\n #m1.rotate(255)\n #m2.rotate(255)\n #time.sleep(0.001)\nexcept:\n pass\n\nfinally:\n m1.rotate(0)\n m2.rotate(0)\n pass\n", "sub_path": "selfbalancing.py", "file_name": "selfbalancing.py", "file_ext": "py", "file_size_in_byte": 1279, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pigpio.pi", "line_number": 17, "usage_type": "call"}, {"api_name": "motor.motor", "line_number": 27, "usage_type": "call"}, {"api_name": "motor.motor", "line_number": 28, "usage_type": "call"}, {"api_name": "pidcontroller.PIDController", "line_number": 45, "usage_type": "call"}]} +{"seq_id": "45631548", "text": "# -*- coding: utf-8 -*-\n\n\"\"\"\nSystem utilities\n\"\"\"\n\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\nfrom __future__ import unicode_literals\n\n\nimport traceback\nimport os\nimport sys\n\n\nif sys.version_info >= (3, 5):\n import subprocess\nelse:\n import subprocess32 as subprocess\n\n\nSTOP_FILE_NAME = \".inlineplzstop\"\n# track commands we've already run so that we don't re-run them\nPREVIOUS_INSTALL_COMMANDS = []\n\n\ndef should_stop():\n return os.path.isfile(os.path.join(os.getcwd(), STOP_FILE_NAME))\n\n\ndef run_command(command, log_on_fail=False, log_all=False, timeout=120):\n print('Running: \"{}\"'.format(\" \".join(command)))\n shell = False\n if os.name == \"nt\":\n shell = True\n popen_kwargs = {\n \"args\": command,\n \"stdin\": subprocess.PIPE,\n \"stdout\": subprocess.PIPE,\n \"stderr\": subprocess.PIPE,\n \"shell\": shell,\n \"env\": os.environ,\n \"universal_newlines\": True,\n \"timeout\": timeout,\n }\n if sys.version_info[0] >= 3 and sys.version_info[1] >= 6:\n popen_kwargs[\"encoding\"] = \"utf-8\"\n try:\n proc = subprocess.run(**popen_kwargs)\n except subprocess.TimeoutExpired:\n print(\"Timeout: {}\".format(command))\n return 0, \"\"\n\n stdout, stderr = proc.stdout, proc.stderr\n output = \"{}\\n{}\".format(stdout, stderr).strip()\n if output and ((log_on_fail and proc.returncode) or log_all):\n print(output)\n sys.stdout.flush()\n return proc.returncode, output\n\n\ndef installed(config):\n try:\n returncode, _ = run_command(config.get(\"help\"))\n return returncode == 0\n\n except (subprocess.CalledProcessError, OSError):\n return False\n\n\ndef install_linter(config):\n install_cmds = config.get(\"install\")\n for install_cmd in install_cmds:\n if install_cmd in PREVIOUS_INSTALL_COMMANDS:\n continue\n\n PREVIOUS_INSTALL_COMMANDS.append(install_cmd)\n if not installed(config):\n try:\n print(\"-\" * 80)\n run_command(install_cmd, log_all=True)\n except OSError:\n print(\n \"Install failed: {0}\\n{1}\".format(\n install_cmd, traceback.format_exc()\n )\n )\n else:\n return\n\n\nHERE = os.path.dirname(__file__)\n\n\nif sys.platform == \"win32\":\n JAVA_SEP = \";\"\nelse:\n JAVA_SEP = \":\"\n\n\ndef vendored_path(path):\n # we use a relpath on windows because the colon in windows drive letter paths messes with java classpaths\n if sys.platform == \"win32\":\n return os.path.normpath(\n os.path.relpath(\n os.path.join(os.path.dirname(HERE), \"bin\", path), os.getcwd()\n )\n )\n\n return os.path.normpath(\n os.path.abspath(os.path.join(os.path.dirname(HERE), \"bin\", path))\n )\n", "sub_path": "inlineplz/util/system.py", "file_name": "system.py", "file_ext": "py", "file_size_in_byte": 2898, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "sys.version_info", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 30, "usage_type": "call"}, {"api_name": "os.name", "line_number": 36, "usage_type": "attribute"}, {"api_name": "subprocess32.PIPE", "line_number": 40, "usage_type": "attribute"}, {"api_name": "subprocess32.PIPE", "line_number": 41, "usage_type": "attribute"}, {"api_name": "subprocess32.PIPE", "line_number": 42, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 44, "usage_type": "attribute"}, {"api_name": "sys.version_info", "line_number": 48, "usage_type": "attribute"}, {"api_name": "subprocess32.run", "line_number": 51, "usage_type": "call"}, {"api_name": "subprocess32.TimeoutExpired", "line_number": 52, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 60, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 60, "usage_type": "attribute"}, {"api_name": "subprocess32.CalledProcessError", "line_number": 69, "usage_type": "attribute"}, {"api_name": "traceback.format_exc", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path", "line_number": 94, "usage_type": "attribute"}, {"api_name": "sys.platform", "line_number": 97, "usage_type": "attribute"}, {"api_name": "sys.platform", "line_number": 105, "usage_type": "attribute"}, {"api_name": "os.path.normpath", "line_number": 106, "usage_type": "call"}, {"api_name": "os.path", "line_number": 106, "usage_type": "attribute"}, {"api_name": "os.path.relpath", "line_number": 107, "usage_type": "call"}, {"api_name": "os.path", "line_number": 107, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path", "line_number": 108, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 108, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path.normpath", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path", "line_number": 112, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path", "line_number": 113, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 113, "usage_type": "call"}]} +{"seq_id": "238066525", "text": "import gevent.monkey\ngevent.monkey.patch_all()\nfrom signalrcore.hub_connection_builder import HubConnectionBuilder\nimport logging\nimport sys\n\nhandler = logging.StreamHandler()\nhandler.setLevel(logging.DEBUG)\n\ndef signup():\n hub_connection.send(\"Signup\", [\"Andy\"])\n\nif __name__ == ('__main__'):\n hub_connection = HubConnectionBuilder()\\\n .with_url(\"http://localhost:5000/riskhub\", options={\"verify_ssl\": False}) \\\n .configure_logging(logging.DEBUG, socket_trace=True, handler=handler) \\\n .with_automatic_reconnect({\n \"type\": \"interval\",\n \"keep_alive_interval\": 10,\n \"intervals\": [1, 3, 5, 6, 7, 87, 3]\n }).build()\n\n\nhub_connection.on_open(lambda: print(\"we are connected\"))\nhub_connection.on_close(lambda: print(\"we are not connected\"))\n\n\nprint('trying to start connection')\nhub_connection.start()\nsignup()\ninput()\n\n\nhub_connection.stop()\n\n", "sub_path": "Risk.Signalr.PythonClient/PyClient.py", "file_name": "PyClient.py", "file_ext": "py", "file_size_in_byte": 894, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "gevent.monkey.monkey.patch_all", "line_number": 2, "usage_type": "call"}, {"api_name": "gevent.monkey.monkey", "line_number": 2, "usage_type": "attribute"}, {"api_name": "gevent.monkey", "line_number": 2, "usage_type": "name"}, {"api_name": "logging.StreamHandler", "line_number": 7, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 8, "usage_type": "attribute"}, {"api_name": "signalrcore.hub_connection_builder.HubConnectionBuilder", "line_number": 14, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 16, "usage_type": "attribute"}]} +{"seq_id": "272920176", "text": "import csv\nfrom math import atan\nimport math\nimport re\nfrom event import CustomizableEvent\nimport rpw\nfrom rpw import revit, db, ui, DB, UI\nfrom rpw.ui.forms import CheckBox, Separator, Button, FlexForm, Label, TextBox\nfrom rpw.exceptions import RevitExceptions\nfrom pyrevit.forms import WPFWindow\nfrom System.Collections.ObjectModel import ObservableCollection\nfrom collections import Iterable, defaultdict\n\n__doc__ = \"Create Civil Model Along an Alignment\"\n__title__ = \"Linear Modelling\"\n__author__ = \"DY Lim\"\n__persistentengine__ = True\n\n\"\"\"\nGlobal Parameter Start\n\"\"\"\n_cond_alignment = False\n_cond_plot = False\n_placement_pts = []\nalingment, updatedAlignment = False, False\n\"\"\"\"\nGlobal Parameter End\n\"\"\"\n\"\"\"\nGeneral Functions Start\n\"\"\"\ndef ft2mm(ft):\n return ft * 304.8\n\ndef mm2ft(mm):\n return mm * 0.00328084\n \ndef Flatten(x):\n if isinstance(x, Iterable):\n return [a for i in x for a in Flatten(i)]\n else:\n return [x]\n\ndef sublist(iterable, interval=2, overlap=1):\n # Create list subset ex)[[0,1],[1,2],[2,3]] from [0,1,2,3]\n if interval == 1:\n return False\n placement_pts_set = []\n i = 0\n while(True):\n placement_pts_set.append(iterable[i : i + interval])\n i = i + (interval - overlap)\n if len(iterable) <= i:\n break\n return placement_pts_set[:-1]\n\"\"\"\nGeneral Functions End\n\"\"\"\n##################################################################################################\n##################################################################################################\nclass FamOpt1(DB.IFamilyLoadOptions):\n def __init__(self):\n pass\n def OnFamilyFound(self,familyInUse, overwriteParameterValues):\n return True\n def OnSharedFamilyFound(self,familyInUse, source, overwriteParameterValues):\n return True\n\ndef selectAlignment():\n global alingment, updatedAlignment\n with db.Transaction(\"Align - selection\"):\n ref = ui.Pick.pick_element(\"Select Alignment Model\", False)\n try:\n alingment = revit.doc.GetElement(ref.ElementId)\n updatedAlignment = True\n except:\n updatedAlignment = False\n\ndef createModels(self, origin, XVec, ZAxis, taskName):\n try:\n allsymbol = []\n for i in self.SegmentContents:\n if not i.IsExcluded:\n allsymbol.append(self.FamSymbolDic[i.Type])\n for i in list(set(allsymbol)):\n AddFamilyParameter(i.Family, \"AlignID\")\n with db.Transaction(\"Align - Create Linear Model\"):\n Plane = [DB.Plane.CreateByNormalAndOrigin(normal, o) for normal, o in zip(ZAxis, origin)]\n sketchPln = [DB.SketchPlane.Create(revit.doc, p).GetPlaneReference() for p in Plane]\n RefPoints = [DB.PointOnPlane.NewPointOnPlane(revit.doc, sp, o, x.Negate()) for sp, o, x in zip(sketchPln, origin, XVec)]\n symbol = []\n createdFamily = []\n selected = \"\"\n insertedID = []\n for i in self.SegmentContents:\n if not i.IsExcluded:\n selected = \"{},{}\".format(selected, i.Index)\n symbol = self.FamSymbolDic[i.Type]\n group = i.Group.Split(\"-\")\n if self.isReversed:\n group = list(reversed(group))\n e = DB.AdaptiveComponentInstanceUtils.CreateAdaptiveComponentInstance(revit.doc, symbol)\n id = e.UniqueId\n insertedID.append(id)\n i.ID = id\n i.IsAdded = True\n createdFamily.append(e)\n for parameter in e.Parameters:\n if parameter.Definition.Name == \"AlignID\":\n parameter.Set(\"{}/{}\".format(i.Group, id))\n AdaptivePointIDs = DB.AdaptiveComponentInstanceUtils.GetInstancePlacementPointElementRefIds(e)\n for g, id in zip(group, AdaptivePointIDs):\n AdaptivePoint = revit.doc.GetElement(id)\n AdaptivePoint.SetPointElementReference(RefPoints[int(g)])\n index = self.AddedObjectContents.Count\n self.AddedObjectContents.Add(AddedObjectFormat(index, taskName, selected[1:], len(createdFamily), insertedID))\n parameters = {}\n familyList = defaultdict(list)\n for i in createdFamily:\n temp = getParameters(i)\n for key, value in temp.items():\n familyList[key].append(value[\"FamilyName\"])\n parameters[key] = value\n self.PTable[taskName] = ObservableCollection[object]()\n for key, value in parameters.items():\n self.PTable[taskName].Add((ParameterTableFormat(index, ','.join(map(str, list(set(familyList[key])))), value[\"Name\"], value[\"Type\"], value[\"Object\"])))\n self.SegmentTable.ItemsSource = None\n self.SegmentTable.ItemsSource = self.SegmentContents\n log(self, \"Add Family Success\")\n except Exception as e:\n UI.TaskDialog.Show(\"Error\", \"{}\".format(e))\n\ndef AddFamilyParameter(family, name):\n try:\n famdoc = revit.doc.EditFamily(family)\n t = DB.Transaction(famdoc)\n for i in famdoc.FamilyManager.Parameters:\n if name == i.Definition.Name:\n return False\n t.Start(\"Add Parameter for Internal Use\")\n famdoc.FamilyManager.AddParameter(name, DB.BuiltInParameterGroup.PG_TEXT, DB.ParameterType.Text, True)\n t.Commit()\n famdoc.LoadFamily(revit.doc, FamOpt1())\n famdoc.Close(True)\n except Exception as e:\n # if t.GetStatus() == DB.TransactionStatus.Started:\n # t.Commit()\n UI.TaskDialog.Show(\"Error\", \"{}\".format(e))\n\ndef SetParameter(self, parameterSet):\n try:\n with db.Transaction(\"Align - Set Parameter\"):\n elementIds = {}\n ElemInActiveView = DB.FilteredElementCollector(revit.doc, revit.doc.ActiveView.Id).ToElements()\n for element in ElemInActiveView:\n parameter = element.GetParameters(\"AlignID\")\n if parameter:\n elementIds[parameter[0].AsString().split(\"/\")[-1]] = element\n for index, id in zip(self.AddedObjectTable.SelectedItem.Items.split(\",\"), self.AddedObjectTable.SelectedItem.ID):\n element = elementIds[id]\n for p in parameterSet:\n parameter = element.GetParameters(p[0])\n if parameter:\n try:\n if p[3]:\n value = self.parameterSet[p[2]][int(index) + 1]\n else:\n value = self.parameterSet[p[2]][int(index)]\n except Exception as e:\n # UI.TaskDialog.Show(\"P[3] Value\", \"{}\".format(e))\n value = p[2]\n value = convertByType(p[1], value, p[4])\n parameter[0].Set(value)\n except Exception as e:\n UI.TaskDialog.Show(\"Error\", \"{}\".format(e))\n\ndef getParameters(family):\n definitions = {i.Id : {\"FamilyName\" : family.Symbol.Family.Name, \"Name\": i.Definition.Name, \"Type\": i.Definition.ParameterType.ToString(), \"Object\" : i} for i in family.Parameters if i.IsReadOnly == False}\n return definitions\n\ndef get_element(of_class, of_category):\n collect = db.Collector(of_class=of_class, of_category=of_category)\n collect_list = collect.get_elements()\n return collect_list\n\ndef convertByType(RevitType, Value, IsNegative):\n string = [\"Text\"]\n number = [\"Length\", \"Angle\"]\n if RevitType in string:\n if IsNegative:\n return \"-{}\".format(Value)\n else:\n return \"{}\".format(Value)\n elif RevitType in number:\n try:\n if IsNegative:\n return -float(Value)\n else:\n return float(Value)\n except:\n return 0\n\ndef log(self, line):\n self.Log.Text = \"{}\\n{}\".format(self.Log.Text, line)\n self.Log.ScrollToEnd()\n\ndef log_p(self, line):\n self.Log_P.Text = \"{}\\n{}\".format(self.Log_P.Text, line)\n self.Log_P.ScrollToEnd()\n\n\n##################################################################################################\n##################################################################################################\n\ncustomizable_event = CustomizableEvent()\n\nclass AlignmentPointTableFormat:\n def __init__(self, index, name, station, elevation, slope, id):\n self.Index = index\n self.Name = name\n self.Station = station\n self.Elevation = elevation\n self.Slope = slope\n self.ID = id\n\nclass SegmentTableFormat:\n def __init__(self, index, group=None, familytype=None, isexcluded=False, isadded=False, id=None):\n self.Index = index\n self.Group = group\n self.Type = familytype\n self.IsExcluded = isexcluded\n self.IsAdded = isadded\n self.ID = id\n\nclass AddedObjectFormat:\n def __init__(self, index, name, items, total, id):\n self.Index = index\n self.Name = name\n self.Items = items\n self.Total = total\n self.ID = id\n\nclass ParameterTableFormat:\n def __init__(self, index, familyname, parameter, valuetype, object, isincluded=False, customvalue=None, IsStaggered=False, IsNagative=False):\n self.Index = index\n self.FamilyName = familyname\n self.Parameter = parameter\n self.ValueType = valuetype\n self.Object = object\n self.IsIncluded = isincluded\n self.CustomValue = customvalue\n self.IsStaggered = IsStaggered\n self.IsNagative = IsNagative\n\n# A simple WPF form used to call the ExternalEvent\nclass form_window(WPFWindow):\n def __init__(self, xaml_file_name):\n WPFWindow.__init__(self, xaml_file_name)\n self.Show()\n FamilySymbols = get_element(\"FamilySymbol\", \"OST_GenericModel\")\n FamilySymbolName = [symbol.name for symbol in FamilySymbols]\n self.FamSymbolDic = {}\n for symbol, name in zip(FamilySymbols, FamilySymbolName):\n self.FamSymbolDic[name] = symbol.unwrap()\n \n self.AlignmentPointContents = ObservableCollection[object]()\n self.AlignmentPointTable.ItemsSource = self.AlignmentPointContents\n\n self.SegmentContents = ObservableCollection[object]()\n self.SegmentTable.ItemsSource = self.SegmentContents\n\n self.AddedObjectContents = ObservableCollection[object]()\n self.AddedObjectTable.ItemsSource = self.AddedObjectContents\n\n self.parameterValues = [\"Custom\", True, False]\n self.PTable = {}\n\n self.FamilyType.ItemsSource = FamilySymbolName\n self.Combo_FamilySymbol.ItemsSource = FamilySymbolName\n self.Combo_CustomValue.ItemsSource = self.parameterValues\n \n log(self, \"Ready\")\n\n def NumberValidationTextBox(self, sender, e):\n #Accecpt only Integer\n if re.search(\"[^0-9.-]+\", e.Text):\n e.Handled = True\n else:\n e.Handled = False\n\n def Clk_SelectAlignment(self, sender, e):\n customizable_event.raise_event(selectAlignment)\n #Log\n log(self, \"Revit Object is selected.\")\n\n def Clk_RefreshAlignmentTable(self, sender, e):\n global alingment, updatedAlignment\n self.AlignmentPointContents.Clear()\n if updatedAlignment:\n try:\n#################Get All Reference Points\n family = alingment.Symbol.Family\n famdoc = revit.doc.EditFamily(family)\n collector = DB.FilteredElementCollector(famdoc)\n refPoints = collector.OfCategory(DB.BuiltInCategory.OST_ReferencePoints).ToElements()\n################Get Name Parameter to display point information to the table\n name = [e.Name.split(\"/\") for e in refPoints if e.Name]\n pointType = list(set([i[1] for i in name]))\n self.PointDic = defaultdict(list)\n for t in pointType:\n for refpt, n in zip(refPoints, name):\n if t in n:\n self.PointDic[t].append([refpt, n])\n self.referenceName = ui.forms.SelectFromList(\"Select Refernce\",pointType)\n selectedPointSet = self.PointDic[self.referenceName]\n self.RefPoint_XVec = []\n self.origin = []\n self.parameterSet = defaultdict(list)\n for i, ptset in enumerate(selectedPointSet):\n origin = ptset[0].GetCoordinateSystem().Origin\n self.origin.append(origin)\n for key, value in self.PointDic.items():\n if key == self.referenceName:\n True\n else:\n self.parameterSet[\"DistanceTo{}\".format(key)].append(self.PointDic[key][i][0].GetCoordinateSystem().Origin.DistanceTo(origin))\n self.RefPoint_XVec.append(ptset[0].GetCoordinateSystem().BasisX.Normalize().Negate())\n index = self.AlignmentPointContents.Count\n offset = round(ft2mm(ptset[0].GetCoordinateSystem().Origin.Z) * 0.001, 5)\n format = AlignmentPointTableFormat(index, ptset[1][1], ptset[1][0], offset, ptset[1][2], ptset[1][-1])\n self.AlignmentPointContents.Add(format)\n #Parameters\n for i in self.AlignmentPointContents:\n self.parameterSet[\"Station\"].append(i.Station)\n self.parameterSet[\"Elevation\"].append(i.Elevation)\n self.parameterSet[\"Slope(%)\"].append(i.Slope)\n self.parameterSet[\"Slope(rad)\"].append(math.atan(float(i.Slope)/100))\n self.parameterSet[\"Slope(deg)\"].append(math.degrees(math.atan(float(i.Slope)/100)))\n for key in self.parameterSet.keys():\n self.parameterValues.append(key)\n log(self, \"Error: {}\".format(self.parameterSet[\"Slope(deg)\"]))\n################Get 3D Alignment and attrubutes for modelling\n collector = DB.FilteredElementCollector(famdoc)\n lines = collector.OfCategory(DB.BuiltInCategory.OST_Lines).ToElements()\n for line in lines:\n if line.GeometryCurve.GetType().Name == \"HermiteSpline\":\n alignment = line.GeometryCurve\n parameters = [p for p in alignment.Parameters]\n transform_Alignment = []\n for p in parameters:\n transform_Alignment.append(alignment.ComputeDerivatives(p, False))\n self.ZAxis = []\n for t, xvec in zip(transform_Alignment, self.RefPoint_XVec):\n tangent_temp = t.BasisX.Normalize()\n self.ZAxis.append(tangent_temp.CrossProduct(xvec.Normalize()))\n################Close Family Document\n famdoc.Close(False)\n log(self, \"Alignment Loaded\")\n updatedAlignment = False\n except Exception as e:\n famdoc.Close(False)\n log(self, \"Error: {}\".format(e))\n return False\n \n def Clk_UpdateSegmentTable(self, sender, e):\n try:\n self.SegmentContents.Clear()\n subSetNum = int(ui.forms.TextInput(\"Divide List\",\"2\", \"Group points every (Only Number) :\"))\n referencePoints = []\n selectedPointSet = self.PointDic[self.referenceName]\n for ptset in selectedPointSet:\n referencePoints.append(ptset[0])\n if subSetNum == 1:\n UI.TaskDialog.Show(\"Error\", \"Divide Number must be more than 1\")\n return False\n referencePointsIndex = sublist(range(len(referencePoints)), subSetNum)\n referencePoints = sublist(referencePoints, subSetNum)\n log(self, \"{}\\n {}\\n {}\\n\".format(referencePoints, referencePointsIndex, selectedPointSet))\n for i in referencePointsIndex:\n index = self.SegmentContents.Count\n gorup = \"-\".join(str(num) for num in i)\n self.SegmentContents.Add(SegmentTableFormat(index, gorup))\n except Exception as e:\n log(self, \"Error : {}\".format(e))\n \n def Clk_UpdateTable(self, sender, e):\n try:\n if self.Start.Text == \"\" or self.End.Text == \"\":\n self.Start.Text = \"0\"\n self.End.Text = \"{}\".format(self.SegmentContents.Count - 1)\n startIndex = int(self.Start.Text)\n endIndex = int(self.End.Text)\n for i in self.SegmentTable.SelectedItems:\n i.Type = self.Combo_FamilySymbol.SelectedValue\n for i in self.SegmentContents:\n if i.IsExcluded == True:\n i.Type = \"\"\n if startIndex <= i.Index <= endIndex:\n i.IsExcluded = False\n else:\n i.IsExcluded = True\n i.Type = \"\"\n #To Refresh Table\n self.SegmentTable.ItemsSource = None \n self.SegmentTable.ItemsSource = self.SegmentContents\n except Exception as e:\n log(self, \"{}\".format(e))\n \n def Clk_AddToRevit(self, sender, e):\n try:\n components = [CheckBox('ReverseProfile', 'Reverse Profile?'),\n Label('Enter Task Name:'), TextBox('TaskName', Text=\"Batch{}\".format(self.AddedObjectContents.Count)),\n Separator(), Button('Select')]\n form = FlexForm('Modelling Configuration', components)\n form.show()\n if form.values[\"ReverseProfile\"]:\n self.RefPoint_XVec = [xvec.Negate() for xvec in self.RefPoint_XVec]\n self.isReversed = True\n else:\n self.isReversed = False\n taskName = form.values[\"TaskName\"]\n self.parameterValues.append(taskName)\n for i in self.AddedObjectContents:\n if i.Items == taskName:\n UI.TaskDialog.Show(\"Error\", \"{} is already in the table.\".format(taskName))\n return False\n customizable_event.raise_event(createModels, self, self.origin, self.RefPoint_XVec, self.ZAxis, taskName)\n except Exception as e:\n log(self, \"{}\".format(e))\n \n def ViewDetailTable(self, sender, e):\n try:\n Name = self.AddedObjectTable.SelectedItem.Name\n self.ParameterTable.ItemsSource = self.PTable[Name]\n self.ParameterTable.ScrollViewer.HorizontalScrollBarVisibility=\"Visible\"\n self.ParameterTable.ScrollViewer.CanContentScroll=\"True\"\n except Exception as e:\n log_p(self, \"{}\".format(e))\n \n def valueUpdated(self, sender, e):\n try:\n Name = self.ParameterTable.SelectedItem.CustomValue\n if Name == \"Custom\":\n newvalue = ui.forms.TextInput(\"SetCustomValue\",\"\",\"Custom Value\")\n self.parameterValues.append(newvalue)\n self.Combo_CustomValue.ItemsSource = self.parameterValues\n\n except Exception as e:\n log_p(self, \"{}\".format(e))\n \n def Clk_SetParameter(self, sender, e):\n try:\n currentParamTable = self.ParameterTable.ItemsSource\n appliedParameter = []\n for i in currentParamTable:\n if i.IsIncluded:\n appliedParameter.append([i.Parameter, i.ValueType, i.CustomValue, i.IsStaggered, i.IsNagative])\n customizable_event.raise_event(SetParameter, self, appliedParameter)\n except Exception as e:\n log_p(self, \"{}\".format(e))\n \n def Clk_ExportParameter(self, sender, e):\n try:\n # currentParamTable = self.ParameterTable.ItemsSource\n # for key, value in self.parameterSet.items():\n # log_p(self, \"{}, {}\".format(key, value))\n with open(\"Export.csv\", \"w\") as f:\n writer = csv.writer(f)\n for key, value in self.parameterSet.items():\n temp = list(value)\n temp.insert(0, key)\n writer.writerow(temp)\n log_p(self, \"Exported\")\n except Exception as e:\n log_p(self, \"{}\".format(e))\n\n\nform = form_window(\"ui.xaml\")\nif False:\n a = select_element()\n b = set_geometry(a)\n UI.TaskDialog.Show(\"Stat\", \"{},{}\".format(a, b))\n c = cal_chainage(b)\n UI.TaskDialog.Show(\"Stat\", \"{}\".format(c))", "sub_path": "script/Python/Align.extension/Align.tab/Create.panel/CreateLinearModel.pushbutton/script.py", "file_name": "script.py", "file_ext": "py", "file_size_in_byte": 20661, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "collections.Iterable", "line_number": 39, "usage_type": "argument"}, {"api_name": "rpw.DB.IFamilyLoadOptions", "line_number": 61, "usage_type": "attribute"}, {"api_name": "rpw.DB", "line_number": 61, "usage_type": "name"}, {"api_name": "rpw.db.Transaction", "line_number": 71, "usage_type": "call"}, {"api_name": "rpw.db", "line_number": 71, "usage_type": "name"}, {"api_name": "rpw.ui.Pick.pick_element", "line_number": 72, "usage_type": "call"}, {"api_name": "rpw.ui.Pick", "line_number": 72, "usage_type": "attribute"}, {"api_name": "rpw.ui", "line_number": 72, "usage_type": "name"}, {"api_name": "rpw.revit.doc.GetElement", "line_number": 74, "usage_type": "call"}, {"api_name": "rpw.revit.doc", "line_number": 74, "usage_type": "attribute"}, {"api_name": "rpw.revit", "line_number": 74, "usage_type": "name"}, {"api_name": "rpw.db.Transaction", "line_number": 87, "usage_type": "call"}, {"api_name": "rpw.db", "line_number": 87, "usage_type": "name"}, {"api_name": "rpw.DB.Plane.CreateByNormalAndOrigin", "line_number": 88, "usage_type": "call"}, {"api_name": "rpw.DB.Plane", "line_number": 88, "usage_type": "attribute"}, {"api_name": "rpw.DB", "line_number": 88, "usage_type": "name"}, {"api_name": "rpw.DB.SketchPlane.Create", "line_number": 89, "usage_type": "call"}, {"api_name": "rpw.DB.SketchPlane", "line_number": 89, "usage_type": "attribute"}, {"api_name": "rpw.DB", "line_number": 89, "usage_type": "name"}, {"api_name": "rpw.revit.doc", "line_number": 89, "usage_type": "attribute"}, {"api_name": "rpw.revit", "line_number": 89, "usage_type": "name"}, {"api_name": "rpw.DB.PointOnPlane.NewPointOnPlane", "line_number": 90, "usage_type": "call"}, {"api_name": "rpw.DB.PointOnPlane", "line_number": 90, "usage_type": "attribute"}, {"api_name": "rpw.DB", "line_number": 90, "usage_type": "name"}, {"api_name": "rpw.revit.doc", "line_number": 90, "usage_type": "attribute"}, {"api_name": "rpw.revit", "line_number": 90, "usage_type": "name"}, {"api_name": "rpw.DB.AdaptiveComponentInstanceUtils.CreateAdaptiveComponentInstance", "line_number": 102, "usage_type": "call"}, {"api_name": "rpw.DB.AdaptiveComponentInstanceUtils", "line_number": 102, "usage_type": "attribute"}, {"api_name": "rpw.DB", "line_number": 102, "usage_type": "name"}, {"api_name": "rpw.revit.doc", "line_number": 102, "usage_type": "attribute"}, {"api_name": "rpw.revit", "line_number": 102, "usage_type": "name"}, {"api_name": "rpw.DB.AdaptiveComponentInstanceUtils.GetInstancePlacementPointElementRefIds", "line_number": 111, "usage_type": "call"}, {"api_name": "rpw.DB.AdaptiveComponentInstanceUtils", "line_number": 111, "usage_type": "attribute"}, {"api_name": "rpw.DB", "line_number": 111, "usage_type": "name"}, {"api_name": "rpw.revit.doc.GetElement", "line_number": 113, "usage_type": "call"}, {"api_name": "rpw.revit.doc", "line_number": 113, "usage_type": "attribute"}, {"api_name": "rpw.revit", "line_number": 113, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 118, "usage_type": "call"}, {"api_name": "System.Collections.ObjectModel.ObservableCollection", "line_number": 124, "usage_type": "name"}, {"api_name": "rpw.UI.TaskDialog.Show", "line_number": 131, "usage_type": "call"}, {"api_name": "rpw.UI.TaskDialog", "line_number": 131, "usage_type": "attribute"}, {"api_name": "rpw.UI", "line_number": 131, "usage_type": "name"}, {"api_name": "rpw.revit.doc.EditFamily", "line_number": 135, "usage_type": "call"}, {"api_name": "rpw.revit.doc", "line_number": 135, "usage_type": "attribute"}, {"api_name": "rpw.revit", "line_number": 135, "usage_type": "name"}, {"api_name": "rpw.DB.Transaction", "line_number": 136, "usage_type": "call"}, {"api_name": "rpw.DB", "line_number": 136, "usage_type": "name"}, {"api_name": "rpw.DB.BuiltInParameterGroup", "line_number": 141, "usage_type": "attribute"}, {"api_name": "rpw.DB", "line_number": 141, "usage_type": "name"}, {"api_name": "rpw.DB.ParameterType", "line_number": 141, "usage_type": "attribute"}, {"api_name": "rpw.revit.doc", "line_number": 143, "usage_type": "attribute"}, {"api_name": "rpw.revit", "line_number": 143, "usage_type": "name"}, {"api_name": "rpw.UI.TaskDialog.Show", "line_number": 148, "usage_type": "call"}, {"api_name": "rpw.UI.TaskDialog", "line_number": 148, "usage_type": "attribute"}, {"api_name": "rpw.UI", "line_number": 148, "usage_type": "name"}, {"api_name": "rpw.db.Transaction", "line_number": 152, "usage_type": "call"}, {"api_name": "rpw.db", "line_number": 152, "usage_type": "name"}, {"api_name": "rpw.DB.FilteredElementCollector", "line_number": 154, "usage_type": "call"}, {"api_name": "rpw.DB", "line_number": 154, "usage_type": "name"}, {"api_name": "rpw.revit.doc", "line_number": 154, "usage_type": "attribute"}, {"api_name": "rpw.revit", "line_number": 154, "usage_type": "name"}, {"api_name": "rpw.UI.TaskDialog.Show", "line_number": 175, "usage_type": "call"}, {"api_name": "rpw.UI.TaskDialog", "line_number": 175, "usage_type": "attribute"}, {"api_name": "rpw.UI", "line_number": 175, "usage_type": "name"}, {"api_name": "rpw.db.Collector", "line_number": 182, "usage_type": "call"}, {"api_name": "rpw.db", "line_number": 182, "usage_type": "name"}, {"api_name": "event.CustomizableEvent", "line_number": 215, "usage_type": "call"}, {"api_name": "pyrevit.forms.WPFWindow", "line_number": 256, "usage_type": "name"}, {"api_name": "pyrevit.forms.WPFWindow.__init__", "line_number": 258, "usage_type": "call"}, {"api_name": "pyrevit.forms.WPFWindow", "line_number": 258, "usage_type": "name"}, {"api_name": "System.Collections.ObjectModel.ObservableCollection", "line_number": 266, "usage_type": "name"}, {"api_name": "System.Collections.ObjectModel.ObservableCollection", "line_number": 269, "usage_type": "name"}, {"api_name": "System.Collections.ObjectModel.ObservableCollection", "line_number": 272, "usage_type": "name"}, {"api_name": "re.search", "line_number": 286, "usage_type": "call"}, {"api_name": "rpw.revit.doc.EditFamily", "line_number": 303, "usage_type": "call"}, {"api_name": "rpw.revit.doc", "line_number": 303, "usage_type": "attribute"}, {"api_name": "rpw.revit", "line_number": 303, "usage_type": "name"}, {"api_name": "rpw.DB.FilteredElementCollector", "line_number": 304, "usage_type": "call"}, {"api_name": "rpw.DB", "line_number": 304, "usage_type": "name"}, {"api_name": "rpw.DB.BuiltInCategory", "line_number": 305, "usage_type": "attribute"}, {"api_name": "rpw.DB", "line_number": 305, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 309, "usage_type": "call"}, {"api_name": "rpw.ui.forms.SelectFromList", "line_number": 314, "usage_type": "call"}, {"api_name": "rpw.ui.forms", "line_number": 314, "usage_type": "attribute"}, {"api_name": "rpw.ui", "line_number": 314, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 318, "usage_type": "call"}, {"api_name": "math.atan", "line_number": 337, "usage_type": "call"}, {"api_name": "math.degrees", "line_number": 338, "usage_type": "call"}, {"api_name": "math.atan", "line_number": 338, "usage_type": "call"}, {"api_name": "rpw.DB.FilteredElementCollector", "line_number": 343, "usage_type": "call"}, {"api_name": "rpw.DB", "line_number": 343, "usage_type": "name"}, {"api_name": "rpw.DB.BuiltInCategory", "line_number": 344, "usage_type": "attribute"}, {"api_name": "rpw.DB", "line_number": 344, "usage_type": "name"}, {"api_name": "rpw.ui.forms.TextInput", "line_number": 368, "usage_type": "call"}, {"api_name": "rpw.ui.forms", "line_number": 368, "usage_type": "attribute"}, {"api_name": "rpw.ui", "line_number": 368, "usage_type": "name"}, {"api_name": "rpw.UI.TaskDialog.Show", "line_number": 374, "usage_type": "call"}, {"api_name": "rpw.UI.TaskDialog", "line_number": 374, "usage_type": "attribute"}, {"api_name": "rpw.UI", "line_number": 374, "usage_type": "name"}, {"api_name": "rpw.ui.forms.CheckBox", "line_number": 411, "usage_type": "call"}, {"api_name": "rpw.ui.forms.Label", "line_number": 412, "usage_type": "call"}, {"api_name": "rpw.ui.forms.TextBox", "line_number": 412, "usage_type": "call"}, {"api_name": "rpw.ui.forms.Separator", "line_number": 413, "usage_type": "call"}, {"api_name": "rpw.ui.forms.Button", "line_number": 413, "usage_type": "call"}, {"api_name": "rpw.ui.forms.FlexForm", "line_number": 414, "usage_type": "call"}, {"api_name": "rpw.UI.TaskDialog.Show", "line_number": 425, "usage_type": "call"}, {"api_name": "rpw.UI.TaskDialog", "line_number": 425, "usage_type": "attribute"}, {"api_name": "rpw.UI", "line_number": 425, "usage_type": "name"}, {"api_name": "rpw.ui.forms.TextInput", "line_number": 444, "usage_type": "call"}, {"api_name": "rpw.ui.forms", "line_number": 444, "usage_type": "attribute"}, {"api_name": "rpw.ui", "line_number": 444, "usage_type": "name"}, {"api_name": "csv.writer", "line_number": 468, "usage_type": "call"}, {"api_name": "rpw.UI.TaskDialog.Show", "line_number": 482, "usage_type": "call"}, {"api_name": "rpw.UI.TaskDialog", "line_number": 482, "usage_type": "attribute"}, {"api_name": "rpw.UI", "line_number": 482, "usage_type": "name"}, {"api_name": "rpw.UI.TaskDialog.Show", "line_number": 484, "usage_type": "call"}, {"api_name": "rpw.UI.TaskDialog", "line_number": 484, "usage_type": "attribute"}, {"api_name": "rpw.UI", "line_number": 484, "usage_type": "name"}]} +{"seq_id": "49263397", "text": "import requests\nimport json\n\n\nclass Poster:\n def __init__(self):\n self.url = \"http://127.0.0.1:5000/\"\n\n def execute(self, ending, payload):\n url = self.url + ending\n r = requests.post(url=url, data=payload)\n print(r.url)\n print(r.text)\n", "sub_path": "basic_api_tester.py", "file_name": "basic_api_tester.py", "file_ext": "py", "file_size_in_byte": 277, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "requests.post", "line_number": 11, "usage_type": "call"}]} +{"seq_id": "261783792", "text": "import re\nfrom datetime import datetime\n\ndef urlify(s):\n\ts = re.sub(r\"[^\\w\\s]\", '', s)\n\ts = re.sub(r\"\\s+\", '-', s)\n\treturn s\n\ntitle = input('title: ')\ntime=str(datetime.now())\n\ncontent='''---\nlayout: post\ntitle: %s\ndate: %s\nupdate_date: \nsummary: \ncategories: blog\npermalink: /blog/%s/\n---\n\n'''%(title, time[:10],urlify(title))\n\nf = open('blog/'+urlify(title)+'.md','w')\nf.write(content)\nf.close\n", "sub_path": "wr.py", "file_name": "wr.py", "file_ext": "py", "file_size_in_byte": 421, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "re.sub", "line_number": 5, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 6, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 10, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 10, "usage_type": "name"}]} +{"seq_id": "105350244", "text": "from django.shortcuts import render, get_object_or_404, redirect, reverse\nfrom django.contrib.auth.models import User\nfrom django.contrib.auth.decorators import login_required\nfrom django.contrib.auth.mixins import LoginRequiredMixin\nfrom django.contrib import messages\nfrom django.db.models import Q\nfrom django.views import View\nfrom django.views.generic import ListView, DetailView, CreateView, UpdateView, DeleteView\nfrom .models import Blog, BlogView, Like, Category\nfrom .forms import BlogCommentForm, BlogForm\n\n\nclass BlogListView(ListView):\n model = Blog\n template_name = 'blog/blog_list.html'\n ordering = ['-publish_date']\n context_object_name = 'all_blogs'\n\n def get_context_data(self, **kwargs):\n feature_blog = Blog.objects.filter(featured=True)\n category_menu = Category.objects.all()\n context = super().get_context_data(**kwargs)\n context['feature_blog'] = feature_blog\n context['category_menu'] = category_menu\n return context\n\n\nclass SearchView(View):\n template_name = 'blog/blog_search.html'\n\n def get(self, request, *args, **kwargs):\n all_blogs = Blog.objects.all()\n category_menu = Category.objects.all()\n query = None\n\n if 's' in request.GET:\n query = request.GET['s']\n if not query:\n messages.error(request, \"Sorry! No Input? Try again Please\")\n return redirect(reverse('blog:list'))\n search = Q(title__icontains=query) | Q(content__icontains=query)\n all_blogs = all_blogs.filter(search)\n\n context = {\n 'search_words': query,\n 'all_blogs': all_blogs,\n 'category_menu': category_menu,\n }\n return render(request, 'blog/blog_search.html', context)\n\n\ndef CategoryView(request, category):\n category_blogs = Blog.objects.filter(category=category)\n category_menu = Category.objects.all()\n all_blogs = Blog.objects.all()\n context = {\n 'category': category,\n 'category_blogs': category_blogs,\n 'category_menu': category_menu,\n 'all_blogs': all_blogs\n }\n return render(request, 'blog/blog_categories.html', context)\n\n\nclass BlogDetailView(DetailView):\n model = Blog\n\n def post(self, *args, **kwargs):\n \"\"\"Adding comments to blogs.\"\"\"\n form = BlogCommentForm(self.request.POST)\n if form.is_valid():\n blog = self.get_object()\n blogcomment = form.instance\n blogcomment.user = self.request.user\n blogcomment.blog = blog\n blogcomment.save()\n return redirect(\"details\", slug=blog.slug)\n return redirect(\"details\", slug=self.get_object().slug)\n\n def get_context_data(self, **kwargs):\n context = super().get_context_data(**kwargs)\n context.update({\n 'form': BlogCommentForm()\n\n })\n return context\n\n def get_object(self, **kwargs):\n \"\"\"Counts the number of authenticated users view the blog.\"\"\"\n object = super().get_object(**kwargs)\n if self.request.user.is_authenticated:\n BlogView.objects.get_or_create(user=self.request.user, blog=object)\n return object\n\n\nclass BlogCreateView(LoginRequiredMixin, CreateView):\n \"\"\"\n Uses the 'blog_form.html' as it needs the inputs,\n The context is changed to 'create', \n \"\"\"\n form_class = BlogForm\n model = Blog\n success_url = '/'\n\n def form_valid(self, form):\n form.instance.author = self.request.user\n return super().form_valid(form)\n\n def get_context_data(self, **kwargs):\n context = super().get_context_data(**kwargs)\n context.update({\n 'view_type': 'create'\n })\n return context\n\n\nclass BlogUpdateView(LoginRequiredMixin, UpdateView):\n \"\"\"\n Uses the 'blog_form.html' as it needs the inputs,\n The context is changed to 'Update', \n (Logic by Mat @ JustDjango) Understood and implemented.\n \"\"\"\n form_class = BlogForm\n model = Blog\n success_url = '/blog/'\n\n def get_context_data(self, **kwargs):\n context = super().get_context_data(**kwargs)\n context.update({\n 'view_type': 'update'\n })\n return context\n\n\nclass BlogDeleteView(LoginRequiredMixin, DeleteView):\n model = Blog\n success_url = '/blog/'\n\n\nclass BlogAuthorPageView(LoginRequiredMixin, View):\n template_name = 'blog/blog_authors.html'\n context_object_name = 'user'\n\n def get(self, request, *args, **kwargs):\n user_profile = get_object_or_404(User, id=self.kwargs['pk'])\n user_blog = Blog.objects.filter(\n author=user_profile.id).order_by('-publish_date')\n context = {\n 'page_user': user_profile,\n 'user_blog': user_blog,\n }\n return render(request, 'blog/blog_authors.html', context)\n\n\n\ndef like(request, slug):\n \"\"\"Checks to see if the use has liked the blog\n If True, then delete the like if False then create the like\n (Logic by Mat @ JustDjango) Understood and implemented.\"\"\"\n\n blog = get_object_or_404(Blog, slug=slug)\n like_qs = Like.objects.filter(user=request.user, blog=blog)\n\n if like_qs.exists():\n # unlike the post\n like_qs[0].delete()\n return redirect('blog:details', slug=slug)\n\n Like.objects.create(user=request.user, blog=blog)\n return redirect('blog:details', slug=slug)\n", "sub_path": "blog/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 5410, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "django.views.generic.ListView", "line_number": 13, "usage_type": "name"}, {"api_name": "models.Blog", "line_number": 14, "usage_type": "name"}, {"api_name": "models.Blog.objects.filter", "line_number": 20, "usage_type": "call"}, {"api_name": "models.Blog.objects", "line_number": 20, "usage_type": "attribute"}, {"api_name": "models.Blog", "line_number": 20, "usage_type": "name"}, {"api_name": "models.Category.objects.all", "line_number": 21, "usage_type": "call"}, {"api_name": "models.Category.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "models.Category", "line_number": 21, "usage_type": "name"}, {"api_name": "django.views.View", "line_number": 28, "usage_type": "name"}, {"api_name": "models.Blog.objects.all", "line_number": 32, "usage_type": "call"}, {"api_name": "models.Blog.objects", "line_number": 32, "usage_type": "attribute"}, {"api_name": "models.Blog", "line_number": 32, "usage_type": "name"}, {"api_name": "models.Category.objects.all", "line_number": 33, "usage_type": "call"}, {"api_name": "models.Category.objects", "line_number": 33, "usage_type": "attribute"}, {"api_name": "models.Category", "line_number": 33, "usage_type": "name"}, {"api_name": "django.contrib.messages.error", "line_number": 39, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 39, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 40, "usage_type": "call"}, {"api_name": "django.shortcuts.reverse", "line_number": 40, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 41, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 49, "usage_type": "call"}, {"api_name": "models.Blog.objects.filter", "line_number": 53, "usage_type": "call"}, {"api_name": "models.Blog.objects", "line_number": 53, "usage_type": "attribute"}, {"api_name": "models.Blog", "line_number": 53, "usage_type": "name"}, {"api_name": "models.Category.objects.all", "line_number": 54, "usage_type": "call"}, {"api_name": "models.Category.objects", "line_number": 54, "usage_type": "attribute"}, {"api_name": "models.Category", "line_number": 54, "usage_type": "name"}, {"api_name": "models.Blog.objects.all", "line_number": 55, "usage_type": "call"}, {"api_name": "models.Blog.objects", "line_number": 55, "usage_type": "attribute"}, {"api_name": "models.Blog", "line_number": 55, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 62, "usage_type": "call"}, {"api_name": "django.views.generic.DetailView", "line_number": 65, "usage_type": "name"}, {"api_name": "models.Blog", "line_number": 66, "usage_type": "name"}, {"api_name": "forms.BlogCommentForm", "line_number": 70, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 77, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 78, "usage_type": "call"}, {"api_name": "forms.BlogCommentForm", "line_number": 83, "usage_type": "call"}, {"api_name": "models.BlogView.objects.get_or_create", "line_number": 92, "usage_type": "call"}, {"api_name": "models.BlogView.objects", "line_number": 92, "usage_type": "attribute"}, {"api_name": "models.BlogView", "line_number": 92, "usage_type": "name"}, {"api_name": "django.contrib.auth.mixins.LoginRequiredMixin", "line_number": 96, "usage_type": "name"}, {"api_name": "django.views.generic.CreateView", "line_number": 96, "usage_type": "name"}, {"api_name": "forms.BlogForm", "line_number": 101, "usage_type": "name"}, {"api_name": "models.Blog", "line_number": 102, "usage_type": "name"}, {"api_name": "django.contrib.auth.mixins.LoginRequiredMixin", "line_number": 117, "usage_type": "name"}, {"api_name": "django.views.generic.UpdateView", "line_number": 117, "usage_type": "name"}, {"api_name": "forms.BlogForm", "line_number": 123, "usage_type": "name"}, {"api_name": "models.Blog", "line_number": 124, "usage_type": "name"}, {"api_name": "django.contrib.auth.mixins.LoginRequiredMixin", "line_number": 135, "usage_type": "name"}, {"api_name": "django.views.generic.DeleteView", "line_number": 135, "usage_type": "name"}, {"api_name": "models.Blog", "line_number": 136, "usage_type": "name"}, {"api_name": "django.contrib.auth.mixins.LoginRequiredMixin", "line_number": 140, "usage_type": "name"}, {"api_name": "django.views.View", "line_number": 140, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 145, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 145, "usage_type": "argument"}, {"api_name": "models.Blog.objects.filter", "line_number": 146, "usage_type": "call"}, {"api_name": "models.Blog.objects", "line_number": 146, "usage_type": "attribute"}, {"api_name": "models.Blog", "line_number": 146, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 152, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 161, "usage_type": "call"}, {"api_name": "models.Blog", "line_number": 161, "usage_type": "argument"}, {"api_name": "models.Like.objects.filter", "line_number": 162, "usage_type": "call"}, {"api_name": "models.Like.objects", "line_number": 162, "usage_type": "attribute"}, {"api_name": "models.Like", "line_number": 162, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 167, "usage_type": "call"}, {"api_name": "models.Like.objects.create", "line_number": 169, "usage_type": "call"}, {"api_name": "models.Like.objects", "line_number": 169, "usage_type": "attribute"}, {"api_name": "models.Like", "line_number": 169, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 170, "usage_type": "call"}]} +{"seq_id": "512295908", "text": "from datetime import datetime, timedelta\n\nfrom .string_manip import getlist\nfrom .time_util import get_relativedelta\nfrom .string_template_substitution import do_string_sub\n\ndef time_generator(config):\n \"\"\"! Generator used to read METplusConfig variables for time looping\n\n @param METplusConfig object to read\n @returns None if not enough information is available on config.\n Yields the next run time dictionary or None if something went wrong\n \"\"\"\n # determine INIT or VALID prefix\n prefix = get_time_prefix(config)\n if not prefix:\n yield None\n return\n\n # get clock time of when the run started\n clock_dt = datetime.strptime(\n config.getstr('config', 'CLOCK_TIME'),\n '%Y%m%d%H%M%S'\n )\n\n time_format = config.getraw('config', f'{prefix}_TIME_FMT', '')\n if not time_format:\n config.logger.error(f'Could not read {prefix}_TIME_FMT')\n yield None\n return\n\n # check for [INIT/VALID]_LIST and use that list if set\n if config.has_option('config', f'{prefix}_LIST'):\n time_list = getlist(config.getraw('config', f'{prefix}_LIST'))\n if not time_list:\n config.logger.error(f\"Could not read {prefix}_LIST\")\n yield None\n return\n\n for time_string in time_list:\n current_dt = _get_current_dt(time_string,\n time_format,\n clock_dt,\n config.logger)\n if not current_dt:\n yield None\n\n time_info = _create_time_input_dict(prefix, current_dt, clock_dt)\n yield time_info\n\n return\n\n # if list is not provided, use _BEG, _END, and _INCREMENT\n start_string = config.getraw('config', f'{prefix}_BEG')\n end_string = config.getraw('config', f'{prefix}_END', start_string)\n time_interval = get_relativedelta(\n config.getstr('config', f'{prefix}_INCREMENT', '60')\n )\n\n start_dt = _get_current_dt(start_string,\n time_format,\n clock_dt,\n config.logger)\n\n end_dt = _get_current_dt(end_string,\n time_format,\n clock_dt,\n config.logger)\n\n if not _validate_time_values(start_dt,\n end_dt,\n time_interval,\n prefix,\n config.logger):\n yield None\n return\n\n current_dt = start_dt\n while current_dt <= end_dt:\n time_info = _create_time_input_dict(prefix, current_dt, clock_dt)\n yield time_info\n\n current_dt += time_interval\n\ndef _validate_time_values(start_dt, end_dt, time_interval, prefix, logger):\n if not start_dt:\n logger.error(f\"Could not read {prefix}_BEG\")\n return False\n\n if not end_dt:\n logger.error(f\"Could not read {prefix}_END\")\n return False\n\n # check that time increment is at least 60 seconds\n if (start_dt + time_interval <\n start_dt + timedelta(seconds=60)):\n logger.error(f'{prefix}_INCREMENT must be greater than or '\n 'equal to 60 seconds')\n return False\n\n if start_dt > end_dt:\n logger.error(f\"{prefix}_BEG must come after {prefix}_END \")\n return False\n\n return True\n\ndef _create_time_input_dict(prefix, current_dt, clock_dt):\n return {\n 'loop_by': prefix.lower(),\n prefix.lower(): current_dt,\n 'now': clock_dt,\n }\n\ndef get_time_prefix(config):\n \"\"\"! Read the METplusConfig object and determine the prefix for the time\n looping variables.\n\n @param config METplusConfig object to read\n @returns string 'INIT' if looping by init time, 'VALID' if looping by\n valid time, or None if not enough information was found in the config\n \"\"\"\n loop_by = config.getstr('config', 'LOOP_BY', '').upper()\n if loop_by in ['INIT', 'RETRO']:\n return 'INIT'\n\n if loop_by in ['VALID', 'REALTIME']:\n return 'VALID'\n\n # check for legacy variable LOOP_BY_INIT if LOOP_BY is not set properly\n if config.has_option('config', 'LOOP_BY_INIT'):\n if config.getbool('config', 'LOOP_BY_INIT'):\n return 'INIT'\n\n return 'VALID'\n\n # report an error if time prefix could not be determined\n config.logger.error('MUST SET LOOP_BY to VALID, INIT, RETRO, or REALTIME')\n return None\n\ndef _get_current_dt(time_string, time_format, clock_dt, logger):\n \"\"\"! Use time format to get datetime object from time string, substituting\n values for today or now template tags if specified.\n\n @param time_string string value read from the config that\n may include now or today tags\n @param time_format format of time_string, i.e. %Y%m%d\n @param clock_dt datetime object for time when execution started\n @returns datetime object if successful, None if not\n \"\"\"\n subbed_time_string = do_string_sub(\n time_string,\n now=clock_dt,\n today=clock_dt.strftime('%Y%m%d')\n )\n try:\n current_dt = datetime.strptime(subbed_time_string, time_format)\n except ValueError:\n logger.error(\n f'Could not format time string ({time_string}) using '\n f'time format ({time_format})'\n )\n return None\n\n return current_dt\n", "sub_path": "metplus/util/time_looping.py", "file_name": "time_looping.py", "file_ext": "py", "file_size_in_byte": 5460, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 21, "usage_type": "name"}, {"api_name": "string_manip.getlist", "line_number": 34, "usage_type": "call"}, {"api_name": "time_util.get_relativedelta", "line_number": 56, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 96, "usage_type": "call"}, {"api_name": "string_template_substitution.do_string_sub", "line_number": 150, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 156, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 156, "usage_type": "name"}]} +{"seq_id": "322967318", "text": "import unittest\n\nfrom flask_inputs import Inputs\nfrom wtforms.validators import DataRequired, Email\n\n\nclass RawInputs(Inputs):\n raw = {\n 'name': [\n DataRequired('Name is required.')\n ],\n 'email': [\n Email('Email must be valid.')\n ]\n }\n\n\nclass RawTest(unittest.TestCase):\n def test_valid(self):\n valid_input = {'name': 'Valtteri', 'email': 'valtteri@test.com'}\n inputs = RawInputs(valid_input)\n\n self.assertTrue(inputs.validate())\n\n def test_invalid(self):\n invalid_input = {'email': 'valtteri@test.com'}\n inputs = RawInputs(invalid_input)\n\n self.assertFalse(inputs.validate())\n self.assertIn('Name is required.', inputs.errors['name'])\n", "sub_path": "tests/test_raw.py", "file_name": "test_raw.py", "file_ext": "py", "file_size_in_byte": 747, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "flask_inputs.Inputs", "line_number": 7, "usage_type": "name"}, {"api_name": "wtforms.validators.DataRequired", "line_number": 10, "usage_type": "call"}, {"api_name": "wtforms.validators.Email", "line_number": 13, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 18, "usage_type": "attribute"}]} +{"seq_id": "118851084", "text": "from rest_framework.response import Response\nfrom rest_framework import status, generics\nfrom rest_framework.decorators import api_view\n\nfrom hotels.models import Reservation_Has_Room, Reservation, Reservation_Payment\nfrom hotels.models import Unit, Room, Customer\nfrom hotels.serializers import listSerializer, createReservation\n\n\nclass listHotels(generics.ListAPIView):\n queryset = Unit.objects.all()\n serializer_class = listSerializer\n\n\nclass createReservationView(generics.ListCreateAPIView):\n queryset = Reservation.objects.all()\n serializer_class = createReservation\n\n def create(self, request, *args, **kwargs):\n fname = request.data.get('firstname')\n lname = request.data.get('lastname')\n email = request.data.get('email', None)\n\n cust = Customer.objects.create(firstName=fname, lastName=lname, email=email)\n\n cid = cust.customerId\n fake = request.data.copy()\n fake.update({'customerId': cid})\n fake.update({'price': 200})\n serializer = self.get_serializer(data=fake)\n serializer.is_valid(raise_exception=True)\n\n # save survey to the user\n res = serializer.save()\n headers = self.get_success_headers(serializer.data)\n\n fromDate = request.data.get('dateFrom')\n toDate = request.data.get('dateTo')\n roomId = request.data.get('roomId', -1)\n try:\n room = Room.objects.get(number=roomId)\n except Room.DoesNotExist:\n return Response(status=status.HTTP_400_BAD_REQUEST)\n\n Reservation_Has_Room.objects.create(\n reservationId=res, number=room, fromDate=fromDate, toDate=toDate)\n\n return Response(\n serializer.data, status=status.HTTP_201_CREATED, headers=headers)\n\n\n@api_view(['POST'])\ndef check_room(request):\n people = request.POST.get('people', 1)\n fromDate = request.POST.get('dateFrom', None)\n toDate = request.POST.get('dateTo', None)\n\n pack1 = Reservation_Has_Room.objects.filter(\n number__beds__gte=people, toDate__lte=fromDate)\n pack2 = Reservation_Has_Room.objects.filter(\n number__beds__gte=people, toDate__gte=toDate)\n\n available = pack1 | pack2\n\n return Response()\n if available:\n return Response(available.values_list('number'))\n else:\n return Response(stats=status.HTTP_404_NOT_FOUND)\n\n\n@api_view(['POST'])\ndef check_payment(request):\n res_id = request.POST.get('reservationId', -1)\n\n try:\n reserv = Reservation.objects.get(reservationId=res_id)\n except Reservation.DoesNotExist:\n return Response(status=status.HTTP_400_BAD_REQUEST)\n\n if Reservation_Payment.objects.filter(reservationId=reserv).exists():\n return Response({'payed': True})\n else:\n return Response({'payed': False})\n", "sub_path": "backend/hotels/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2792, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "rest_framework.generics.ListAPIView", "line_number": 10, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 10, "usage_type": "name"}, {"api_name": "hotels.models.Unit.objects.all", "line_number": 11, "usage_type": "call"}, {"api_name": "hotels.models.Unit.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "hotels.models.Unit", "line_number": 11, "usage_type": "name"}, {"api_name": "hotels.serializers.listSerializer", "line_number": 12, "usage_type": "name"}, {"api_name": "rest_framework.generics.ListCreateAPIView", "line_number": 15, "usage_type": "attribute"}, {"api_name": "rest_framework.generics", "line_number": 15, "usage_type": "name"}, {"api_name": "hotels.models.Reservation.objects.all", "line_number": 16, "usage_type": "call"}, {"api_name": "hotels.models.Reservation.objects", "line_number": 16, "usage_type": "attribute"}, {"api_name": "hotels.models.Reservation", "line_number": 16, "usage_type": "name"}, {"api_name": "hotels.serializers.createReservation", "line_number": 17, "usage_type": "name"}, {"api_name": "hotels.models.Customer.objects.create", "line_number": 24, "usage_type": "call"}, {"api_name": "hotels.models.Customer.objects", "line_number": 24, "usage_type": "attribute"}, {"api_name": "hotels.models.Customer", "line_number": 24, "usage_type": "name"}, {"api_name": "hotels.models.Room.objects.get", "line_number": 41, "usage_type": "call"}, {"api_name": "hotels.models.Room.objects", "line_number": 41, "usage_type": "attribute"}, {"api_name": "hotels.models.Room", "line_number": 41, "usage_type": "name"}, {"api_name": "hotels.models.Room.DoesNotExist", "line_number": 42, "usage_type": "attribute"}, {"api_name": "hotels.models.Room", "line_number": 42, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 43, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 43, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 43, "usage_type": "name"}, {"api_name": "hotels.models.Reservation_Has_Room.objects.create", "line_number": 45, "usage_type": "call"}, {"api_name": "hotels.models.Reservation_Has_Room.objects", "line_number": 45, "usage_type": "attribute"}, {"api_name": "hotels.models.Reservation_Has_Room", "line_number": 45, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 48, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 49, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 49, "usage_type": "name"}, {"api_name": "hotels.models.Reservation_Has_Room.objects.filter", "line_number": 58, "usage_type": "call"}, {"api_name": "hotels.models.Reservation_Has_Room.objects", "line_number": 58, "usage_type": "attribute"}, {"api_name": "hotels.models.Reservation_Has_Room", "line_number": 58, "usage_type": "name"}, {"api_name": "hotels.models.Reservation_Has_Room.objects.filter", "line_number": 60, "usage_type": "call"}, {"api_name": "hotels.models.Reservation_Has_Room.objects", "line_number": 60, "usage_type": "attribute"}, {"api_name": "hotels.models.Reservation_Has_Room", "line_number": 60, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 65, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 67, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 69, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 69, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 69, "usage_type": "name"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 52, "usage_type": "call"}, {"api_name": "hotels.models.Reservation.objects.get", "line_number": 77, "usage_type": "call"}, {"api_name": "hotels.models.Reservation.objects", "line_number": 77, "usage_type": "attribute"}, {"api_name": "hotels.models.Reservation", "line_number": 77, "usage_type": "name"}, {"api_name": "hotels.models.Reservation.DoesNotExist", "line_number": 78, "usage_type": "attribute"}, {"api_name": "hotels.models.Reservation", "line_number": 78, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 79, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 79, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 79, "usage_type": "name"}, {"api_name": "hotels.models.Reservation_Payment.objects.filter", "line_number": 81, "usage_type": "call"}, {"api_name": "hotels.models.Reservation_Payment.objects", "line_number": 81, "usage_type": "attribute"}, {"api_name": "hotels.models.Reservation_Payment", "line_number": 81, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 82, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 84, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 72, "usage_type": "call"}]} +{"seq_id": "648119720", "text": "import asyncio\nimport logging\n\nimport GDPRClient\nimport apikeys\nimport commandRegistry\nimport config\nfrom helpers import managementHelpers\nfrom models import Session, TagReactables\nfrom rolemessages import TMHCRoles, TestRoles\n\nlogger = logging.getLogger(__name__)\n\n# Only log debug messages in debug mode\nif config.DEBUG:\n logging.basicConfig(level=logging.DEBUG)\nelse:\n logging.basicConfig(level=logging.INFO)\n\ncommands = {}\ncommand = commandRegistry.command\nreaction = commandRegistry.reaction\nrestrictions = commandRegistry.restrictions\nhelp_text = commandRegistry.help_text\ntag_reactables = commandRegistry.tag_reactables\n\n\n@help_text(\"Ping\")\n@command(\"ping\")\nasync def ping(command_data, metadata, send_reply):\n return await send_reply(\"Pong\")\n\n\n@restrictions(config.servers.get(\"TMHCRND\"))\n@help_text(\"Vet a member for the server. \\n\\\n approve: vote to approve the member (requires at least two approvals).\\n\\\n deny: vote to deny the member (only one deny required, but if a deny occurs \\\n after an approve one vetter must change their vote in order to approve or deny)\\n\\\n respond : send a PM to the member in question.\")\n@command(\"vet\")\nasync def vet(command_data, metadata, send_reply):\n if command_data[1][0] == \"approve\":\n pass\n elif command_data[1][0] == \"deny\":\n pass\n elif command_data[1][0] == \"respond\":\n pass\n\n\n@restrictions(config.servers.get(\"TMHC\"), config.servers.get(\"Test\"))\n@command(\"regenroles\")\n@help_text(\"Regenerate the roles text\")\nasync def regenroles(command_data, metadata, send_reply):\n if metadata[\"server\"].id == config.servers[\"TMHC\"]:\n rolearray = TMHCRoles.rolearray\n elif metadata[\"server\"].id == config.servers[\"Test\"]:\n rolearray = TestRoles.rolearray\n else:\n return await send_reply(\"An error occurred!\")\n\n guild = metadata[\"message\"].guild\n channel = guild.get_channel(int(config.roles_channel.get(metadata[\"server\"].id)))\n\n await managementHelpers.clearChannel(channel)\n\n for element in rolearray:\n if isinstance(element, str):\n await send_reply(element)\n await asyncio.sleep(1)\n elif isinstance(element, tuple):\n message = await send_reply(element[1])\n role = guild.get_role(int(element[0]))\n\n if role is None:\n return await send_reply(\"An error occurred!\")\n\n await do_add_role_reactable(message, role.id, metadata)\n await asyncio.sleep(1)\n\n\n@restrictions(config.servers.get(\"TMHC\"), config.servers.get(\"Test\"))\n@command(\"addrolereactable\")\n@help_text(\"Allow reactions on a message to assign a role. \"\n \"Usage: 'addrolereactable ' in the channel containing the message.\")\nasync def addrolereactable(command_data, metadata, send_reply):\n if len(command_data[1]) < 2:\n return await send_reply(\"Please specify a messageid and roleid\")\n\n messageid = command_data[1][0]\n roleid = command_data[1][1]\n guild = metadata.get(\"client\").get_guild(metadata.get(\"server\").id)\n role = guild.get_role(int(roleid))\n\n if role is None:\n return await send_reply(\"No role with that id was found\")\n\n try:\n message = await metadata[\"message\"].channel.fetch_message(messageid)\n except Exception as e:\n logger.exception(e)\n return await send_reply(\"No message with that id was found\")\n\n await do_add_role_reactable(message, roleid, metadata)\n\n await metadata[\"message\"].delete() # delete the invoking message\n return await send_reply(\"Added role react for \" + role.name, delete_after=5)\n\n\nasync def do_add_role_reactable(message, roleid, metadata):\n emoji = \"🔼\"\n await message.add_reaction(emoji)\n\n session = Session()\n instance = session.query(TagReactables).filter_by(message_id=message.id).first()\n\n if instance:\n instance.function_name = \"toggle_role\"\n instance.function_args = roleid\n else:\n instance = TagReactables(message_id=message.id, function_name=\"toggle_role\", function_args=roleid)\n session.add(instance)\n\n try:\n session.commit()\n except Exception as e:\n session.rollback()\n logger.error(\"Unable to commit to database: {}\", e)\n\n\n@restrictions(config.servers.get(\"TMHC\"), config.servers.get(\"Test\"))\n@tag_reactables()\nasync def accept_coc(args, event_type, metadata):\n if event_type != \"REACTION_ADD\":\n return\n\n roles = []\n guild = metadata.get(\"client\").get_guild(metadata.get(\"server\").id)\n\n for role in config.coc_roles[metadata.get(\"server\").id]:\n roles.append(guild.get_role(role))\n\n try:\n user = guild.get_member(metadata.get(\"user\").id)\n\n # ensure user is not already a member and does not already have the coc role\n if guild.get_role(config.member_role[metadata.get(\"server\").id]) not in user.roles and roles[0] not in user.roles:\n await guild.get_member(user.id).add_roles(*roles)\n except Exception as e:\n logger.exception(e)\n\n\n@restrictions(config.servers.get(\"TMHC\"), config.servers.get(\"Test\"))\n@tag_reactables()\nasync def toggle_role(args, event_type, metadata):\n guild = metadata.get(\"client\").get_guild(metadata.get(\"server\").id)\n role = guild.get_role(int(args))\n\n if event_type == \"REACTION_ADD\":\n try:\n userid = metadata.get(\"user\").id\n await guild.get_member(userid).add_roles(role)\n except Exception as e:\n logger.exception(e)\n elif event_type == \"REACTION_REMOVE\":\n try:\n userid = metadata.get(\"user\").id\n await guild.get_member(userid).remove_roles(role)\n except Exception as e:\n logger.exception(e)\n\n\n@restrictions(config.servers.get(\"TMHC\"), config.servers.get(\"Test\"))\n@command(\"gdpr\")\n@help_text(\"Compile GDPR data on a user. Usage: 'gdpr '.\")\nasync def get_gdpr(command_data, metadata, send_reply):\n await send_reply(\"Booting up worker...\")\n client = GDPRClient.GDPRClient()\n loop = asyncio.get_event_loop()\n loop.create_task(client.start(apikeys.workerkey))\n await client.wait_until_ready()\n userid = command_data[1][0]\n guildid = metadata[\"message\"].guild.id\n\n await client.getGDPR(userid, guildid, send_reply)\n await client.logout()\n\n\n@restrictions(config.servers.get(\"TMHC\"), config.servers.get(\"Test\"))\n@command(\"gdprdelete\")\n@help_text(\"Delete all messages sent by a user. Usage: 'gdprdelete '.\")\nasync def delete_gdpr(command_data, metadata, send_reply):\n await send_reply(\"Booting up worker...\")\n client = GDPRClient.GDPRClient()\n loop = asyncio.get_event_loop()\n loop.create_task(client.start(apikeys.workerkey))\n await client.wait_until_ready()\n userid = command_data[1][0]\n guildid = metadata[\"message\"].guild.id\n\n await client.deleteGDPR(userid, guildid, send_reply)\n await client.logout()\n\n", "sub_path": "src/commands/managementCommands.py", "file_name": "managementCommands.py", "file_ext": "py", "file_size_in_byte": 6919, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "logging.getLogger", "line_number": 12, "usage_type": "call"}, {"api_name": "config.DEBUG", "line_number": 15, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 16, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 16, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 18, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 18, "usage_type": "attribute"}, {"api_name": "commandRegistry.command", "line_number": 21, "usage_type": "attribute"}, {"api_name": "commandRegistry.reaction", "line_number": 22, "usage_type": "attribute"}, {"api_name": "commandRegistry.restrictions", "line_number": 23, "usage_type": "attribute"}, {"api_name": "commandRegistry.help_text", "line_number": 24, "usage_type": "attribute"}, {"api_name": "commandRegistry.tag_reactables", "line_number": 25, "usage_type": "attribute"}, {"api_name": "config.servers.get", "line_number": 34, "usage_type": "call"}, {"api_name": "config.servers", "line_number": 34, "usage_type": "attribute"}, {"api_name": "config.servers", "line_number": 54, "usage_type": "attribute"}, {"api_name": "rolemessages.TMHCRoles.rolearray", "line_number": 55, "usage_type": "attribute"}, {"api_name": "rolemessages.TMHCRoles", "line_number": 55, "usage_type": "name"}, {"api_name": "config.servers", "line_number": 56, "usage_type": "attribute"}, {"api_name": "rolemessages.TestRoles.rolearray", "line_number": 57, "usage_type": "attribute"}, {"api_name": "rolemessages.TestRoles", "line_number": 57, "usage_type": "name"}, {"api_name": "config.roles_channel.get", "line_number": 62, "usage_type": "call"}, {"api_name": "config.roles_channel", "line_number": 62, "usage_type": "attribute"}, {"api_name": "helpers.managementHelpers.clearChannel", "line_number": 64, "usage_type": "call"}, {"api_name": "helpers.managementHelpers", "line_number": 64, "usage_type": "name"}, {"api_name": "asyncio.sleep", "line_number": 69, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 78, "usage_type": "call"}, {"api_name": "config.servers.get", "line_number": 50, "usage_type": "call"}, {"api_name": "config.servers", "line_number": 50, "usage_type": "attribute"}, {"api_name": "config.servers.get", "line_number": 81, "usage_type": "call"}, {"api_name": "config.servers", "line_number": 81, "usage_type": "attribute"}, {"api_name": "models.Session", "line_number": 113, "usage_type": "call"}, {"api_name": "models.TagReactables", "line_number": 114, "usage_type": "argument"}, {"api_name": "models.TagReactables", "line_number": 120, "usage_type": "call"}, {"api_name": "config.coc_roles", "line_number": 139, "usage_type": "attribute"}, {"api_name": "config.member_role", "line_number": 146, "usage_type": "attribute"}, {"api_name": "config.servers.get", "line_number": 130, "usage_type": "call"}, {"api_name": "config.servers", "line_number": 130, "usage_type": "attribute"}, {"api_name": "config.servers.get", "line_number": 152, "usage_type": "call"}, {"api_name": "config.servers", "line_number": 152, "usage_type": "attribute"}, {"api_name": "GDPRClient.GDPRClient", "line_number": 177, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 178, "usage_type": "call"}, {"api_name": "apikeys.workerkey", "line_number": 179, "usage_type": "attribute"}, {"api_name": "config.servers.get", "line_number": 172, "usage_type": "call"}, {"api_name": "config.servers", "line_number": 172, "usage_type": "attribute"}, {"api_name": "GDPRClient.GDPRClient", "line_number": 193, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 194, "usage_type": "call"}, {"api_name": "apikeys.workerkey", "line_number": 195, "usage_type": "attribute"}, {"api_name": "config.servers.get", "line_number": 188, "usage_type": "call"}, {"api_name": "config.servers", "line_number": 188, "usage_type": "attribute"}]} +{"seq_id": "97144035", "text": "from django.core.management.base import BaseCommand\nfrom django.db import connection\nfrom django.core import management\nfrom decouple import config\n\n\n# Resets existing database and migrate\nclass Command(BaseCommand):\n help = 'Resets application to default'\n\n def handle(self, *args, **options):\n with connection.cursor() as cursor:\n cursor.execute(\"DROP DATABASE IF EXISTS \" + config('DB_NAME', default='') + \";\")\n cursor.execute(\n \"CREATE DATABASE \" + config('DB_NAME', default='') + \" /*!40100 COLLATE 'utf8_general_ci' */;\")\n cursor.execute(\"USE \" + config('DB_NAME', default='') + \";\")\n\n management.call_command('clear_migrations')\n management.call_command('makemigrations')\n management.call_command('migrate')\n", "sub_path": "movies/management/commands/reset.py", "file_name": "reset.py", "file_ext": "py", "file_size_in_byte": 797, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "django.core.management.base.BaseCommand", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.connection.cursor", "line_number": 12, "usage_type": "call"}, {"api_name": "django.db.connection", "line_number": 12, "usage_type": "name"}, {"api_name": "decouple.config", "line_number": 13, "usage_type": "call"}, {"api_name": "decouple.config", "line_number": 15, "usage_type": "call"}, {"api_name": "decouple.config", "line_number": 16, "usage_type": "call"}, {"api_name": "django.core.management.call_command", "line_number": 18, "usage_type": "call"}, {"api_name": "django.core.management", "line_number": 18, "usage_type": "name"}, {"api_name": "django.core.management.call_command", "line_number": 19, "usage_type": "call"}, {"api_name": "django.core.management", "line_number": 19, "usage_type": "name"}, {"api_name": "django.core.management.call_command", "line_number": 20, "usage_type": "call"}, {"api_name": "django.core.management", "line_number": 20, "usage_type": "name"}]} +{"seq_id": "447374862", "text": "import sys\r\nimport numpy as np\r\nimport ctypes as ct\r\nimport pyopencl as cl\r\nimport pyopencl.array\r\nimport time\r\nimport argparse\r\nFUT_BLOCK_DIM = \"16\"\r\ncl_group_size = np.int32(512)\r\nsynchronous = False\r\nfut_opencl_src = \"\"\"typedef char int8_t;\r\ntypedef short int16_t;\r\ntypedef int int32_t;\r\ntypedef long int64_t;\r\ntypedef uchar uint8_t;\r\ntypedef ushort uint16_t;\r\ntypedef uint uint32_t;\r\ntypedef ulong uint64_t;\r\nstatic inline int8_t add8(int8_t x, int8_t y)\r\n{\r\n return x + y;\r\n}\r\nstatic inline int16_t add16(int16_t x, int16_t y)\r\n{\r\n return x + y;\r\n}\r\nstatic inline int32_t add32(int32_t x, int32_t y)\r\n{\r\n return x + y;\r\n}\r\nstatic inline int64_t add64(int64_t x, int64_t y)\r\n{\r\n return x + y;\r\n}\r\nstatic inline int8_t sub8(int8_t x, int8_t y)\r\n{\r\n return x - y;\r\n}\r\nstatic inline int16_t sub16(int16_t x, int16_t y)\r\n{\r\n return x - y;\r\n}\r\nstatic inline int32_t sub32(int32_t x, int32_t y)\r\n{\r\n return x - y;\r\n}\r\nstatic inline int64_t sub64(int64_t x, int64_t y)\r\n{\r\n return x - y;\r\n}\r\nstatic inline int8_t mul8(int8_t x, int8_t y)\r\n{\r\n return x * y;\r\n}\r\nstatic inline int16_t mul16(int16_t x, int16_t y)\r\n{\r\n return x * y;\r\n}\r\nstatic inline int32_t mul32(int32_t x, int32_t y)\r\n{\r\n return x * y;\r\n}\r\nstatic inline int64_t mul64(int64_t x, int64_t y)\r\n{\r\n return x * y;\r\n}\r\nstatic inline uint8_t udiv8(uint8_t x, uint8_t y)\r\n{\r\n return x / y;\r\n}\r\nstatic inline uint16_t udiv16(uint16_t x, uint16_t y)\r\n{\r\n return x / y;\r\n}\r\nstatic inline uint32_t udiv32(uint32_t x, uint32_t y)\r\n{\r\n return x / y;\r\n}\r\nstatic inline uint64_t udiv64(uint64_t x, uint64_t y)\r\n{\r\n return x / y;\r\n}\r\nstatic inline uint8_t umod8(uint8_t x, uint8_t y)\r\n{\r\n return x % y;\r\n}\r\nstatic inline uint16_t umod16(uint16_t x, uint16_t y)\r\n{\r\n return x % y;\r\n}\r\nstatic inline uint32_t umod32(uint32_t x, uint32_t y)\r\n{\r\n return x % y;\r\n}\r\nstatic inline uint64_t umod64(uint64_t x, uint64_t y)\r\n{\r\n return x % y;\r\n}\r\nstatic inline int8_t sdiv8(int8_t x, int8_t y)\r\n{\r\n int8_t q = x / y;\r\n int8_t r = x % y;\r\n \r\n return q - ((r != 0 && r < 0 != y < 0) ? 1 : 0);\r\n}\r\nstatic inline int16_t sdiv16(int16_t x, int16_t y)\r\n{\r\n int16_t q = x / y;\r\n int16_t r = x % y;\r\n \r\n return q - ((r != 0 && r < 0 != y < 0) ? 1 : 0);\r\n}\r\nstatic inline int32_t sdiv32(int32_t x, int32_t y)\r\n{\r\n int32_t q = x / y;\r\n int32_t r = x % y;\r\n \r\n return q - ((r != 0 && r < 0 != y < 0) ? 1 : 0);\r\n}\r\nstatic inline int64_t sdiv64(int64_t x, int64_t y)\r\n{\r\n int64_t q = x / y;\r\n int64_t r = x % y;\r\n \r\n return q - ((r != 0 && r < 0 != y < 0) ? 1 : 0);\r\n}\r\nstatic inline int8_t smod8(int8_t x, int8_t y)\r\n{\r\n int8_t r = x % y;\r\n \r\n return r + (r == 0 || (x > 0 && y > 0) || (x < 0 && y < 0) ? 0 : y);\r\n}\r\nstatic inline int16_t smod16(int16_t x, int16_t y)\r\n{\r\n int16_t r = x % y;\r\n \r\n return r + (r == 0 || (x > 0 && y > 0) || (x < 0 && y < 0) ? 0 : y);\r\n}\r\nstatic inline int32_t smod32(int32_t x, int32_t y)\r\n{\r\n int32_t r = x % y;\r\n \r\n return r + (r == 0 || (x > 0 && y > 0) || (x < 0 && y < 0) ? 0 : y);\r\n}\r\nstatic inline int64_t smod64(int64_t x, int64_t y)\r\n{\r\n int64_t r = x % y;\r\n \r\n return r + (r == 0 || (x > 0 && y > 0) || (x < 0 && y < 0) ? 0 : y);\r\n}\r\nstatic inline int8_t squot8(int8_t x, int8_t y)\r\n{\r\n return x / y;\r\n}\r\nstatic inline int16_t squot16(int16_t x, int16_t y)\r\n{\r\n return x / y;\r\n}\r\nstatic inline int32_t squot32(int32_t x, int32_t y)\r\n{\r\n return x / y;\r\n}\r\nstatic inline int64_t squot64(int64_t x, int64_t y)\r\n{\r\n return x / y;\r\n}\r\nstatic inline int8_t srem8(int8_t x, int8_t y)\r\n{\r\n return x % y;\r\n}\r\nstatic inline int16_t srem16(int16_t x, int16_t y)\r\n{\r\n return x % y;\r\n}\r\nstatic inline int32_t srem32(int32_t x, int32_t y)\r\n{\r\n return x % y;\r\n}\r\nstatic inline int64_t srem64(int64_t x, int64_t y)\r\n{\r\n return x % y;\r\n}\r\nstatic inline uint8_t shl8(uint8_t x, uint8_t y)\r\n{\r\n return x << y;\r\n}\r\nstatic inline uint16_t shl16(uint16_t x, uint16_t y)\r\n{\r\n return x << y;\r\n}\r\nstatic inline uint32_t shl32(uint32_t x, uint32_t y)\r\n{\r\n return x << y;\r\n}\r\nstatic inline uint64_t shl64(uint64_t x, uint64_t y)\r\n{\r\n return x << y;\r\n}\r\nstatic inline uint8_t lshr8(uint8_t x, uint8_t y)\r\n{\r\n return x >> y;\r\n}\r\nstatic inline uint16_t lshr16(uint16_t x, uint16_t y)\r\n{\r\n return x >> y;\r\n}\r\nstatic inline uint32_t lshr32(uint32_t x, uint32_t y)\r\n{\r\n return x >> y;\r\n}\r\nstatic inline uint64_t lshr64(uint64_t x, uint64_t y)\r\n{\r\n return x >> y;\r\n}\r\nstatic inline int8_t ashr8(int8_t x, int8_t y)\r\n{\r\n return x >> y;\r\n}\r\nstatic inline int16_t ashr16(int16_t x, int16_t y)\r\n{\r\n return x >> y;\r\n}\r\nstatic inline int32_t ashr32(int32_t x, int32_t y)\r\n{\r\n return x >> y;\r\n}\r\nstatic inline int64_t ashr64(int64_t x, int64_t y)\r\n{\r\n return x >> y;\r\n}\r\nstatic inline uint8_t and8(uint8_t x, uint8_t y)\r\n{\r\n return x & y;\r\n}\r\nstatic inline uint16_t and16(uint16_t x, uint16_t y)\r\n{\r\n return x & y;\r\n}\r\nstatic inline uint32_t and32(uint32_t x, uint32_t y)\r\n{\r\n return x & y;\r\n}\r\nstatic inline uint64_t and64(uint64_t x, uint64_t y)\r\n{\r\n return x & y;\r\n}\r\nstatic inline uint8_t or8(uint8_t x, uint8_t y)\r\n{\r\n return x | y;\r\n}\r\nstatic inline uint16_t or16(uint16_t x, uint16_t y)\r\n{\r\n return x | y;\r\n}\r\nstatic inline uint32_t or32(uint32_t x, uint32_t y)\r\n{\r\n return x | y;\r\n}\r\nstatic inline uint64_t or64(uint64_t x, uint64_t y)\r\n{\r\n return x | y;\r\n}\r\nstatic inline uint8_t xor8(uint8_t x, uint8_t y)\r\n{\r\n return x ^ y;\r\n}\r\nstatic inline uint16_t xor16(uint16_t x, uint16_t y)\r\n{\r\n return x ^ y;\r\n}\r\nstatic inline uint32_t xor32(uint32_t x, uint32_t y)\r\n{\r\n return x ^ y;\r\n}\r\nstatic inline uint64_t xor64(uint64_t x, uint64_t y)\r\n{\r\n return x ^ y;\r\n}\r\nstatic inline char ult8(uint8_t x, uint8_t y)\r\n{\r\n return x < y;\r\n}\r\nstatic inline char ult16(uint16_t x, uint16_t y)\r\n{\r\n return x < y;\r\n}\r\nstatic inline char ult32(uint32_t x, uint32_t y)\r\n{\r\n return x < y;\r\n}\r\nstatic inline char ult64(uint64_t x, uint64_t y)\r\n{\r\n return x < y;\r\n}\r\nstatic inline char ule8(uint8_t x, uint8_t y)\r\n{\r\n return x <= y;\r\n}\r\nstatic inline char ule16(uint16_t x, uint16_t y)\r\n{\r\n return x <= y;\r\n}\r\nstatic inline char ule32(uint32_t x, uint32_t y)\r\n{\r\n return x <= y;\r\n}\r\nstatic inline char ule64(uint64_t x, uint64_t y)\r\n{\r\n return x <= y;\r\n}\r\nstatic inline char slt8(int8_t x, int8_t y)\r\n{\r\n return x < y;\r\n}\r\nstatic inline char slt16(int16_t x, int16_t y)\r\n{\r\n return x < y;\r\n}\r\nstatic inline char slt32(int32_t x, int32_t y)\r\n{\r\n return x < y;\r\n}\r\nstatic inline char slt64(int64_t x, int64_t y)\r\n{\r\n return x < y;\r\n}\r\nstatic inline char sle8(int8_t x, int8_t y)\r\n{\r\n return x <= y;\r\n}\r\nstatic inline char sle16(int16_t x, int16_t y)\r\n{\r\n return x <= y;\r\n}\r\nstatic inline char sle32(int32_t x, int32_t y)\r\n{\r\n return x <= y;\r\n}\r\nstatic inline char sle64(int64_t x, int64_t y)\r\n{\r\n return x <= y;\r\n}\r\nstatic inline int8_t pow8(int8_t x, int8_t y)\r\n{\r\n int8_t res = 1, rem = y;\r\n \r\n while (rem != 0) {\r\n if (rem & 1)\r\n res *= x;\r\n rem >>= 1;\r\n x *= x;\r\n }\r\n return res;\r\n}\r\nstatic inline int16_t pow16(int16_t x, int16_t y)\r\n{\r\n int16_t res = 1, rem = y;\r\n \r\n while (rem != 0) {\r\n if (rem & 1)\r\n res *= x;\r\n rem >>= 1;\r\n x *= x;\r\n }\r\n return res;\r\n}\r\nstatic inline int32_t pow32(int32_t x, int32_t y)\r\n{\r\n int32_t res = 1, rem = y;\r\n \r\n while (rem != 0) {\r\n if (rem & 1)\r\n res *= x;\r\n rem >>= 1;\r\n x *= x;\r\n }\r\n return res;\r\n}\r\nstatic inline int64_t pow64(int64_t x, int64_t y)\r\n{\r\n int64_t res = 1, rem = y;\r\n \r\n while (rem != 0) {\r\n if (rem & 1)\r\n res *= x;\r\n rem >>= 1;\r\n x *= x;\r\n }\r\n return res;\r\n}\r\nstatic inline int8_t sext_i8_i8(int8_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline int16_t sext_i8_i16(int8_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline int32_t sext_i8_i32(int8_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline int64_t sext_i8_i64(int8_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline int8_t sext_i16_i8(int16_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline int16_t sext_i16_i16(int16_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline int32_t sext_i16_i32(int16_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline int64_t sext_i16_i64(int16_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline int8_t sext_i32_i8(int32_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline int16_t sext_i32_i16(int32_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline int32_t sext_i32_i32(int32_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline int64_t sext_i32_i64(int32_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline int8_t sext_i64_i8(int64_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline int16_t sext_i64_i16(int64_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline int32_t sext_i64_i32(int64_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline int64_t sext_i64_i64(int64_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint8_t zext_i8_i8(uint8_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint16_t zext_i8_i16(uint8_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint32_t zext_i8_i32(uint8_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint64_t zext_i8_i64(uint8_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint8_t zext_i16_i8(uint16_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint16_t zext_i16_i16(uint16_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint32_t zext_i16_i32(uint16_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint64_t zext_i16_i64(uint16_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint8_t zext_i32_i8(uint32_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint16_t zext_i32_i16(uint32_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint32_t zext_i32_i32(uint32_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint64_t zext_i32_i64(uint32_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint8_t zext_i64_i8(uint64_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint16_t zext_i64_i16(uint64_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint32_t zext_i64_i32(uint64_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint64_t zext_i64_i64(uint64_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline float fdiv32(float x, float y)\r\n{\r\n return x / y;\r\n}\r\nstatic inline float fadd32(float x, float y)\r\n{\r\n return x + y;\r\n}\r\nstatic inline float fsub32(float x, float y)\r\n{\r\n return x - y;\r\n}\r\nstatic inline float fmul32(float x, float y)\r\n{\r\n return x * y;\r\n}\r\nstatic inline float fpow32(float x, float y)\r\n{\r\n return pow(x, y);\r\n}\r\nstatic inline char cmplt32(float x, float y)\r\n{\r\n return x < y;\r\n}\r\nstatic inline char cmple32(float x, float y)\r\n{\r\n return x <= y;\r\n}\r\nstatic inline float sitofp_i8_f32(int8_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline float sitofp_i16_f32(int16_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline float sitofp_i32_f32(int32_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline float sitofp_i64_f32(int64_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline float uitofp_i8_f32(uint8_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline float uitofp_i16_f32(uint16_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline float uitofp_i32_f32(uint32_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline float uitofp_i64_f32(uint64_t x)\r\n{\r\n return x;\r\n}\r\nstatic inline int8_t fptosi_f32_i8(float x)\r\n{\r\n return x;\r\n}\r\nstatic inline int16_t fptosi_f32_i16(float x)\r\n{\r\n return x;\r\n}\r\nstatic inline int32_t fptosi_f32_i32(float x)\r\n{\r\n return x;\r\n}\r\nstatic inline int64_t fptosi_f32_i64(float x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint8_t fptoui_f32_i8(float x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint16_t fptoui_f32_i16(float x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint32_t fptoui_f32_i32(float x)\r\n{\r\n return x;\r\n}\r\nstatic inline uint64_t fptoui_f32_i64(float x)\r\n{\r\n return x;\r\n}\r\n__kernel void map_kernel_930(float view_737, float y_745, float res_741,\r\n int32_t width_734, __global unsigned char *mem_940,\r\n __global unsigned char *mem_942)\r\n{\r\n const uint kernel_thread_index_930 = get_global_id(0);\r\n \r\n if (kernel_thread_index_930 >= width_734)\r\n return;\r\n \r\n int32_t i_931;\r\n \r\n // compute thread index\r\n {\r\n i_931 = kernel_thread_index_930;\r\n }\r\n // read kernel parameters\r\n { }\r\n \r\n float x_933 = sitofp_i32_f32(i_931);\r\n float x_934 = x_933 * res_741;\r\n float y_935 = x_934 / y_745;\r\n float res_936 = view_737 + y_935;\r\n float x_937 = res_936 * res_936;\r\n \r\n // write kernel result\r\n {\r\n *(__global float *) &mem_940[i_931 * 4] = res_936;\r\n *(__global float *) &mem_942[i_931 * 4] = x_937;\r\n }\r\n}\r\n__kernel void map_kernel_881(__global unsigned char *mem_940, __global\r\n unsigned char *res_mem_948, int32_t limit_736,\r\n __global unsigned char *mem_944, __global\r\n unsigned char *mem_946, int32_t width_734,\r\n float res_742, __global unsigned char *mem_942,\r\n float view_738, float y_746,\r\n int32_t nesting_size_879, char x_747,\r\n int32_t height_735, __global\r\n unsigned char *mem_952)\r\n{\r\n const uint kernel_thread_index_881 = get_global_id(0);\r\n \r\n if (kernel_thread_index_881 >= width_734 * height_735)\r\n return;\r\n \r\n int32_t i_882;\r\n int32_t i_883;\r\n float res_884;\r\n float x_885;\r\n \r\n // compute thread index\r\n {\r\n i_882 = squot32(kernel_thread_index_881, height_735);\r\n i_883 = kernel_thread_index_881 - squot32(kernel_thread_index_881,\r\n height_735) * height_735;\r\n }\r\n // read kernel parameters\r\n {\r\n res_884 = *(__global float *) &mem_940[i_882 * 4];\r\n x_885 = *(__global float *) &mem_942[i_882 * 4];\r\n }\r\n \r\n float x_887 = sitofp_i32_f32(i_883);\r\n float x_888 = x_887 * res_742;\r\n float y_889 = x_888 / y_746;\r\n float res_890 = view_738 + y_889;\r\n float y_891 = res_890 * res_890;\r\n float res_892 = x_885 + y_891;\r\n char y_893 = res_892 < 4.0F;\r\n char loop_cond_894 = x_747 && y_893;\r\n char nameless_914;\r\n float c_915;\r\n float c_916;\r\n int32_t i_917;\r\n char loop_while_895;\r\n float c_896;\r\n float c_897;\r\n int32_t i_898;\r\n \r\n loop_while_895 = loop_cond_894;\r\n c_896 = res_884;\r\n c_897 = res_890;\r\n i_898 = 0;\r\n while (loop_while_895) {\r\n float x_899 = c_896 * c_896;\r\n float y_900 = c_897 * c_897;\r\n float res_901 = x_899 - y_900;\r\n float x_902 = c_896 * c_897;\r\n float y_903 = c_897 * c_896;\r\n float res_904 = x_902 + y_903;\r\n float res_905 = res_884 + res_901;\r\n float res_906 = res_890 + res_904;\r\n int32_t res_907 = i_898 + 1;\r\n char x_908 = slt32(res_907, limit_736);\r\n float x_909 = res_905 * res_905;\r\n float y_910 = res_906 * res_906;\r\n float res_911 = x_909 + y_910;\r\n char y_912 = res_911 < 4.0F;\r\n char loop_cond_913 = x_908 && y_912;\r\n char loop_while_tmp_967 = loop_cond_913;\r\n float c_tmp_968 = res_905;\r\n float c_tmp_969 = res_906;\r\n int32_t i_tmp_970;\r\n \r\n i_tmp_970 = res_907;\r\n loop_while_895 = loop_while_tmp_967;\r\n c_896 = c_tmp_968;\r\n c_897 = c_tmp_969;\r\n i_898 = i_tmp_970;\r\n }\r\n nameless_914 = loop_while_895;\r\n c_915 = c_896;\r\n c_916 = c_897;\r\n i_917 = i_898;\r\n \r\n char cond_918 = limit_736 == i_917;\r\n int32_t trunc_arg_919 = 3 * i_917;\r\n int8_t res_920 = sext_i32_i8(trunc_arg_919);\r\n int32_t trunc_arg_921 = 5 * i_917;\r\n int8_t res_922 = sext_i32_i8(trunc_arg_921);\r\n int32_t trunc_arg_923 = 7 * i_917;\r\n int8_t res_924 = sext_i32_i8(trunc_arg_923);\r\n \r\n if (cond_918) {\r\n *(__global int8_t *) &mem_944[kernel_thread_index_881] = 0;\r\n *(__global int8_t *) &mem_944[nesting_size_879 +\r\n kernel_thread_index_881] = 0;\r\n *(__global int8_t *) &mem_944[2 * nesting_size_879 +\r\n kernel_thread_index_881] = 0;\r\n for (int i_971 = 0; i_971 < 3; i_971++) {\r\n *(__global int8_t *) &res_mem_948[nesting_size_879 * i_971 +\r\n kernel_thread_index_881] =\r\n *(__global int8_t *) &mem_944[nesting_size_879 * i_971 +\r\n kernel_thread_index_881];\r\n }\r\n } else {\r\n *(__global int8_t *) &mem_946[kernel_thread_index_881] = res_920;\r\n *(__global int8_t *) &mem_946[nesting_size_879 +\r\n kernel_thread_index_881] = res_922;\r\n *(__global int8_t *) &mem_946[2 * nesting_size_879 +\r\n kernel_thread_index_881] = res_924;\r\n for (int i_972 = 0; i_972 < 3; i_972++) {\r\n *(__global int8_t *) &res_mem_948[nesting_size_879 * i_972 +\r\n kernel_thread_index_881] =\r\n *(__global int8_t *) &mem_946[nesting_size_879 * i_972 +\r\n kernel_thread_index_881];\r\n }\r\n }\r\n // write kernel result\r\n {\r\n for (int i_973 = 0; i_973 < 3; i_973++) {\r\n *(__global int8_t *) &mem_952[3 * (height_735 * i_882) +\r\n (height_735 * i_973 + i_883)] =\r\n *(__global int8_t *) &res_mem_948[nesting_size_879 * i_973 +\r\n kernel_thread_index_881];\r\n }\r\n }\r\n}\r\n__kernel void fut_kernel_map_transpose_i8(__global int8_t *odata,\r\n uint odata_offset, __global\r\n int8_t *idata, uint idata_offset,\r\n uint width, uint height,\r\n uint total_size, __local\r\n int8_t *block)\r\n{\r\n uint x_index;\r\n uint y_index;\r\n uint our_array_offset;\r\n \r\n // Adjust the input and output arrays with the basic offset.\r\r\n odata += odata_offset / sizeof(int8_t);\r\n idata += idata_offset / sizeof(int8_t);\r\n // Adjust the input and output arrays for the third dimension.\r\r\n our_array_offset = get_global_id(2) * width * height;\r\n odata += our_array_offset;\r\n idata += our_array_offset;\r\n // read the matrix tile into shared memory\r\r\n x_index = get_global_id(0);\r\n y_index = get_global_id(1);\r\n \r\n uint index_in = y_index * width + x_index;\r\n \r\n if ((x_index < width && y_index < height) && index_in < total_size)\r\n block[get_local_id(1) * (FUT_BLOCK_DIM + 1) + get_local_id(0)] =\r\n idata[index_in];\r\n barrier(CLK_LOCAL_MEM_FENCE);\r\n // Write the transposed matrix tile to global memory.\r\r\n x_index = get_group_id(1) * FUT_BLOCK_DIM + get_local_id(0);\r\n y_index = get_group_id(0) * FUT_BLOCK_DIM + get_local_id(1);\r\n \r\n uint index_out = y_index * height + x_index;\r\n \r\n if ((x_index < height && y_index < width) && index_out < total_size)\r\n odata[index_out] = block[get_local_id(0) * (FUT_BLOCK_DIM + 1) +\r\n get_local_id(1)];\r\n}\r\n\"\"\"\r\n# Hacky parser/reader for values written in Futhark syntax. Used for\r\n# reading stdin when compiling standalone programs with the Python\r\n# code generator.\r\n\r\nlookahead_buffer = []\r\n\r\ndef reset_lookahead():\r\n global lookahead_buffer\r\n lookahead_buffer = []\r\n\r\ndef get_char(f):\r\n global lookahead_buffer\r\n if len(lookahead_buffer) == 0:\r\n return f.read(1)\r\n else:\r\n c = lookahead_buffer[0]\r\n lookahead_buffer = lookahead_buffer[1:]\r\n return c\r\n\r\ndef unget_char(f, c):\r\n global lookahead_buffer\r\n lookahead_buffer = [c] + lookahead_buffer\r\n\r\ndef peek_char(f):\r\n c = get_char(f)\r\n if c:\r\n unget_char(f, c)\r\n return c\r\n\r\ndef skip_spaces(f):\r\n c = get_char(f)\r\n while c != None:\r\n if c.isspace():\r\n c = get_char(f)\r\n elif c == '-':\r\n # May be line comment.\r\n if peek_char(f) == '-':\r\n # Yes, line comment. Skip to end of line.\r\n while (c != '\\n' and c != None):\r\n c = get_char(f)\r\n else:\r\n break\r\n else:\r\n break\r\n if c:\r\n unget_char(f, c)\r\n\r\ndef parse_specific_char(f, expected):\r\n got = get_char(f)\r\n if got != expected:\r\n unget_char(f, got)\r\n raise ValueError\r\n return True\r\n\r\ndef parse_specific_string(f, s):\r\n for c in s:\r\n parse_specific_char(f, c)\r\n return True\r\n\r\ndef optional(p, *args):\r\n try:\r\n return p(*args)\r\n except ValueError:\r\n return None\r\n\r\ndef sepBy(p, sep, *args):\r\n elems = []\r\n x = optional(p, *args)\r\n if x != None:\r\n elems += [x]\r\n while optional(sep, *args) != None:\r\n x = p(*args)\r\n elems += [x]\r\n return elems\r\n\r\ndef parse_int(f):\r\n s = ''\r\n c = get_char(f)\r\n while c != None:\r\n if c.isdigit():\r\n s += c\r\n c = get_char(f)\r\n else:\r\n unget_char(f, c)\r\n break\r\n optional(read_int_trailer, f)\r\n return s\r\n\r\ndef parse_int_signed(f):\r\n s = ''\r\n c = get_char(f)\r\n\r\n if c == '-' and peek_char(f).isdigit():\r\n s = c + parse_int(f)\r\n else:\r\n unget_char(f, c)\r\n s = parse_int(f)\r\n\r\n return s\r\n\r\ndef read_int_trailer(f):\r\n parse_specific_char(f, 'i')\r\n while peek_char(f).isdigit():\r\n get_char(f)\r\n\r\ndef read_comma(f):\r\n skip_spaces(f)\r\n parse_specific_char(f, ',')\r\n return ','\r\n\r\ndef read_int(f):\r\n skip_spaces(f)\r\n return int(parse_int_signed(f))\r\n\r\ndef read_char(f):\r\n skip_spaces(f)\r\n parse_specific_char(f, '\\'')\r\n c = get_char(f)\r\n parse_specific_char(f, '\\'')\r\n return c\r\n\r\ndef read_double(f):\r\n skip_spaces(f)\r\n c = get_char(f)\r\n if (c == '-'):\r\n sign = '-'\r\n else:\r\n unget_char(f,c)\r\n sign = ''\r\n bef = optional(parse_int, f)\r\n if bef == None:\r\n bef = '0'\r\n parse_specific_char(f, '.')\r\n aft = parse_int(f)\r\n elif optional(parse_specific_char, f, '.'):\r\n aft = parse_int(f)\r\n else:\r\n aft = '0'\r\n if (optional(parse_specific_char, f, 'E') or\r\n optional(parse_specific_char, f, 'e')):\r\n expt = parse_int_signed(f)\r\n else:\r\n expt = '0'\r\n optional(read_float_trailer, f)\r\n return float(sign + bef + '.' + aft + 'E' + expt)\r\n\r\ndef read_float(f):\r\n return read_double(f)\r\n\r\ndef read_float_trailer(f):\r\n parse_specific_char(f, 'f')\r\n while peek_char(f).isdigit():\r\n get_char(f)\r\n\r\ndef read_bool(f):\r\n skip_spaces(f)\r\n if peek_char(f) == 'T':\r\n parse_specific_string(f, 'True')\r\n return True\r\n elif peek_char(f) == 'F':\r\n parse_specific_string(f, 'False')\r\n return False\r\n else:\r\n raise ValueError\r\n\r\ndef read_array_elems(f, elem_reader):\r\n skip_spaces(f)\r\n parse_specific_char(f, '[')\r\n xs = sepBy(elem_reader, read_comma, f)\r\n skip_spaces(f)\r\n parse_specific_char(f, ']')\r\n return xs\r\n\r\ndef read_array_helper(f, elem_reader, rank):\r\n def nested_row_reader(_):\r\n return read_array_helper(f, elem_reader, rank-1)\r\n if rank == 1:\r\n row_reader = elem_reader\r\n else:\r\n row_reader = nested_row_reader\r\n return read_array_elems(f, row_reader)\r\n\r\ndef expected_array_dims(l, rank):\r\n if rank > 1:\r\n n = len(l)\r\n if n == 0:\r\n elem = []\r\n else:\r\n elem = l[0]\r\n return [n] + expected_array_dims(elem, rank-1)\r\n else:\r\n return [len(l)]\r\n\r\ndef verify_array_dims(l, dims):\r\n if dims[0] != len(l):\r\n raise ValueError\r\n if len(dims) > 1:\r\n for x in l:\r\n verify_array_dims(x, dims[1:])\r\n\r\ndef read_double_signed(f):\r\n\r\n skip_spaces(f)\r\n c = get_char(f)\r\n\r\n if c == '-' and peek_char(f).isdigit():\r\n v = -1 * read_double(f)\r\n else:\r\n unget_char(f, c)\r\n v = read_double(f)\r\n\r\n return v\r\n\r\ndef read_array(f, elem_reader, rank, bt):\r\n elems = read_array_helper(f, elem_reader, rank)\r\n dims = expected_array_dims(elems, rank)\r\n verify_array_dims(elems, dims)\r\n return np.array(elems, dtype=bt)\r\n# Scalar functions.\r\n\r\nimport numpy as np\r\n\r\ndef signed(x):\r\n if type(x) == np.uint8:\r\n return np.int8(x)\r\n elif type(x) == np.uint16:\r\n return np.int16(x)\r\n elif type(x) == np.uint32:\r\n return np.int32(x)\r\n else:\r\n return np.int64(x)\r\n\r\ndef unsigned(x):\r\n if type(x) == np.int8:\r\n return np.uint8(x)\r\n elif type(x) == np.int16:\r\n return np.uint16(x)\r\n elif type(x) == np.int32:\r\n return np.uint32(x)\r\n else:\r\n return np.uint64(x)\r\n\r\ndef shlN(x,y):\r\n return x << y\r\n\r\ndef ashrN(x,y):\r\n return x >> y\r\n\r\ndef sdivN(x,y):\r\n return x / y\r\n\r\ndef smodN(x,y):\r\n return x % y\r\n\r\ndef udivN(x,y):\r\n return signed(unsigned(x) / unsigned(y))\r\n\r\ndef umodN(x,y):\r\n return signed(unsigned(x) % unsigned(y))\r\n\r\ndef squotN(x,y):\r\n return np.int32(float(x) / float(y))\r\n\r\ndef sremN(x,y):\r\n return np.fmod(x,y)\r\n\r\ndef powN(x,y):\r\n return x ** y\r\n\r\ndef fpowN(x,y):\r\n return x ** y\r\n\r\ndef sleN(x,y):\r\n return x <= y\r\n\r\ndef sltN(x,y):\r\n return x < y\r\n\r\ndef uleN(x,y):\r\n return unsigned(x) <= unsigned(y)\r\n\r\ndef ultN(x,y):\r\n return unsigned(x) < unsigned(y)\r\n\r\ndef lshr8(x,y):\r\n return np.int8(np.uint8(x) >> np.uint8(y))\r\n\r\ndef lshr16(x,y):\r\n return np.int16(np.uint16(x) >> np.uint16(y))\r\n\r\ndef lshr32(x,y):\r\n return np.int32(np.uint32(x) >> np.uint32(y))\r\n\r\ndef lshr64(x,y):\r\n return np.int64(np.uint64(x) >> np.uint64(y))\r\n\r\ndef sext_T_i8(x):\r\n return np.int8(x)\r\n\r\ndef sext_T_i16(x):\r\n return np.int16(x)\r\n\r\ndef sext_T_i32(x):\r\n return np.int32(x)\r\n\r\ndef sext_T_i64(x):\r\n return np.int32(x)\r\n\r\ndef zext_i8_i8(x):\r\n return np.int8(np.uint8(x))\r\n\r\ndef zext_i8_i16(x):\r\n return np.int16(np.uint8(x))\r\n\r\ndef zext_i8_i32(x):\r\n return np.int32(np.uint8(x))\r\n\r\ndef zext_i8_i64(x):\r\n return np.int64(np.uint8(x))\r\n\r\ndef zext_i16_i8(x):\r\n return np.int8(np.uint16(x))\r\n\r\ndef zext_i16_i16(x):\r\n return np.int16(np.uint16(x))\r\n\r\ndef zext_i16_i32(x):\r\n return np.int32(np.uint16(x))\r\n\r\ndef zext_i16_i64(x):\r\n return np.int64(np.uint16(x))\r\n\r\ndef zext_i32_i8(x):\r\n return np.int8(np.uint32(x))\r\n\r\ndef zext_i32_i16(x):\r\n return np.int16(np.uint32(x))\r\n\r\ndef zext_i32_i32(x):\r\n return np.int32(np.uint32(x))\r\n\r\ndef zext_i32_i64(x):\r\n return np.int64(np.uint32(x))\r\n\r\ndef zext_i64_i8(x):\r\n return np.int8(np.uint64(x))\r\n\r\ndef zext_i64_i16(x):\r\n return np.int16(np.uint64(x))\r\n\r\ndef zext_i64_i32(x):\r\n return np.int32(np.uint64(x))\r\n\r\ndef zext_i64_i64(x):\r\n return np.int64(np.uint64(x))\r\n\r\nshl8 = shl16 = shl32 = shl64 = shlN\r\nashr8 = ashr16 = ashr32 = ashr64 = ashrN\r\nsdiv8 = sdiv16 = sdiv32 = sdiv64 = sdivN\r\nsmod8 = smod16 = smod32 = smod64 = smodN\r\nudiv8 = udiv16 = udiv32 = udiv64 = udivN\r\numod8 = umod16 = umod32 = umod64 = umodN\r\nsquot8 = squot16 = squot32 = squot64 = squotN\r\nsrem8 = srem16 = srem32 = srem64 = sremN\r\npow8 = pow16 = pow32 = pow64 = powN\r\nfpow32 = fpow64 = fpowN\r\nsle8 = sle16 = sle32 = sle64 = sleN\r\nslt8 = slt16 = slt32 = slt64 = sltN\r\nule8 = ule16 = ule32 = ule64 = uleN\r\nult8 = ult16 = ult32 = ult64 = ultN\r\nsext_i8_i8 = sext_i16_i8 = sext_i32_i8 = sext_i64_i8 = sext_T_i8\r\nsext_i8_i16 = sext_i16_i16 = sext_i32_i16 = sext_i64_i16 = sext_T_i16\r\nsext_i8_i32 = sext_i16_i32 = sext_i32_i32 = sext_i64_i32 = sext_T_i32\r\nsext_i8_i64 = sext_i16_i64 = sext_i32_i64 = sext_i64_i64 = sext_T_i64\r\n\r\ndef ssignum(x):\r\n return np.sign(x)\r\n\r\ndef usignum(x):\r\n if x < 0:\r\n return ssignum(-x)\r\n else:\r\n return ssignum(x)\r\n\r\ndef sitofp_T_f32(x):\r\n return np.float32(x)\r\nsitofp_i8_f32 = sitofp_i16_f32 = sitofp_i32_f32 = sitofp_i64_f32 = sitofp_T_f32\r\n\r\ndef sitofp_T_f64(x):\r\n return np.float64(x)\r\nsitofp_i8_f64 = sitofp_i16_f64 = sitofp_i32_f64 = sitofp_i64_f64 = sitofp_T_f64\r\n\r\ndef uitofp_T_f32(x):\r\n return np.float32(unsigned(x))\r\nuitofp_i8_f32 = uitofp_i16_f32 = uitofp_i32_f32 = uitofp_i64_f32 = uitofp_T_f32\r\n\r\ndef uitofp_T_f64(x):\r\n return np.float64(unsigned(x))\r\nuitofp_i8_f64 = uitofp_i16_f64 = uitofp_i32_f64 = uitofp_i64_f64 = uitofp_T_f64\r\n\r\ndef fptosi_T_i8(x):\r\n return np.int8(np.trunc(x))\r\nfptosi_f32_i8 = fptosi_f64_i8 = fptosi_T_i8\r\n\r\ndef fptosi_T_i16(x):\r\n return np.int16(np.trunc(x))\r\nfptosi_f32_i16 = fptosi_f64_i16 = fptosi_T_i16\r\n\r\ndef fptosi_T_i32(x):\r\n return np.int32(np.trunc(x))\r\nfptosi_f32_i32 = fptosi_f64_i32 = fptosi_T_i32\r\n\r\ndef fptosi_T_i64(x):\r\n return np.int64(np.trunc(x))\r\nfptosi_f32_i64 = fptosi_f64_i64 = fptosi_T_i64\r\n\r\ndef fptoui_T_i8(x):\r\n return np.uint8(np.trunc(x))\r\nfptoui_f32_i8 = fptoui_f64_i8 = fptoui_T_i8\r\n\r\ndef fptoui_T_i16(x):\r\n return np.uint16(np.trunc(x))\r\nfptoui_f32_i16 = fptoui_f64_i16 = fptoui_T_i16\r\n\r\ndef fptoui_T_i32(x):\r\n return np.uint32(np.trunc(x))\r\nfptoui_f32_i32 = fptoui_f64_i32 = fptoui_T_i32\r\n\r\ndef fptoui_T_i64(x):\r\n return np.uint64(np.trunc(x))\r\nfptoui_f32_i64 = fptoui_f64_i64 = fptoui_T_i64\r\n\r\ndef fpconv_f32_f64(x):\r\n return np.float64(x)\r\n\r\ndef fpconv_f64_f32(x):\r\n return np.float32(x)\r\n\r\ndef futhark_log64(x):\r\n return np.float64(np.log(x))\r\n\r\ndef futhark_sqrt64(x):\r\n return np.sqrt(x)\r\n\r\ndef futhark_exp64(x):\r\n return np.exp(x)\r\n\r\ndef futhark_cos64(x):\r\n return np.cos(x)\r\n\r\ndef futhark_sin64(x):\r\n return np.sin(x)\r\n\r\ndef futhark_atan2_64(x, y):\r\n return np.arctan2(x, y)\r\n\r\ndef futhark_isnan64(x):\r\n return np.isnan(x)\r\n\r\ndef futhark_isinf64(x):\r\n return np.isinf(x)\r\n\r\ndef futhark_log32(x):\r\n return np.float32(np.log(x))\r\n\r\ndef futhark_sqrt32(x):\r\n return np.float32(np.sqrt(x))\r\n\r\ndef futhark_exp32(x):\r\n return np.exp(x)\r\n\r\ndef futhark_cos32(x):\r\n return np.cos(x)\r\n\r\ndef futhark_sin32(x):\r\n return np.sin(x)\r\n\r\ndef futhark_atan2_32(x, y):\r\n return np.arctan2(x, y)\r\n\r\ndef futhark_isnan32(x):\r\n return np.isnan(x)\r\n\r\ndef futhark_isinf32(x):\r\n return np.isinf(x)\r\nclass mandelbrot:\r\n def __init__(self):\r\n self.ctx = cl.create_some_context(interactive=False)\r\r\n self.queue = cl.CommandQueue(self.ctx)\r\r\n # XXX: Assuming just a single device here.\r\r\n platform_name = self.ctx.get_info(cl.context_info.DEVICES)[0].platform.name\r\r\n device_type = self.ctx.get_info(cl.context_info.DEVICES)[0].type\r\r\n lockstep_width = 1\r\r\n if ((platform_name == \"NVIDIA CUDA\") and (device_type == cl.device_type.GPU)):\r\n lockstep_width = np.int32(32)\r\n if ((platform_name == \"AMD Accelerated Parallel Processing\") and (device_type == cl.device_type.GPU)):\r\n lockstep_width = np.int32(64)\r\r\n if (len(fut_opencl_src) >= 0):\r\r\n program = cl.Program(self.ctx, fut_opencl_src).build([\"-DFUT_BLOCK_DIM={}\".format(FUT_BLOCK_DIM), \"-DLOCKSTEP_WIDTH={}\".format(lockstep_width)])\r\r\n \r\r\n self.map_kernel_930_var = program.map_kernel_930\r\n self.map_kernel_881_var = program.map_kernel_881\r\n self.fut_kernel_map_transpose_i8_var = program.fut_kernel_map_transpose_i8\r\n def futhark_main(self, width_734, height_735, limit_736, view_737, view_738,\r\n view_739, view_740):\r\n res_741 = (view_739 - view_737)\r\n res_742 = (view_740 - view_738)\r\n y_745 = sitofp_i32_f32(width_734)\r\n y_746 = sitofp_i32_f32(height_735)\r\n x_747 = slt32(np.int32(0), limit_736)\r\n bytes_939 = (np.int32(4) * width_734)\r\n mem_940 = cl.Buffer(self.ctx, cl.mem_flags.READ_WRITE,\r\n long(long(bytes_939) if (bytes_939 > np.int32(0)) else np.int32(1)))\r\n mem_942 = cl.Buffer(self.ctx, cl.mem_flags.READ_WRITE,\r\n long(long(bytes_939) if (bytes_939 > np.int32(0)) else np.int32(1)))\r\n group_size_965 = np.int32(512)\r\n num_groups_966 = squot32(((width_734 + group_size_965) - np.int32(1)),\r\n group_size_965)\r\n if ((np.int32(1) * (num_groups_966 * group_size_965)) != np.int32(0)):\r\n self.map_kernel_930_var.set_args(np.float32(view_737), np.float32(y_745),\r\n np.float32(res_741), np.int32(width_734),\r\n mem_940, mem_942)\r\n cl.enqueue_nd_range_kernel(self.queue, self.map_kernel_930_var,\r\n (long((num_groups_966 * group_size_965)),),\r\n (long(group_size_965),))\r\n if synchronous:\r\n self.queue.finish()\r\n nesting_size_879 = (height_735 * width_734)\r\n x_951 = (width_734 * np.int32(3))\r\n bytes_949 = (x_951 * height_735)\r\n mem_952 = cl.Buffer(self.ctx, cl.mem_flags.READ_WRITE,\r\n long(long(bytes_949) if (bytes_949 > np.int32(0)) else np.int32(1)))\r\n total_size_960 = (nesting_size_879 * np.int32(3))\r\n res_mem_948 = cl.Buffer(self.ctx, cl.mem_flags.READ_WRITE,\r\n long(long(total_size_960) if (total_size_960 > np.int32(0)) else np.int32(1)))\r\n total_size_961 = (nesting_size_879 * np.int32(3))\r\n mem_944 = cl.Buffer(self.ctx, cl.mem_flags.READ_WRITE,\r\n long(long(total_size_961) if (total_size_961 > np.int32(0)) else np.int32(1)))\r\n total_size_962 = (nesting_size_879 * np.int32(3))\r\n mem_946 = cl.Buffer(self.ctx, cl.mem_flags.READ_WRITE,\r\n long(long(total_size_962) if (total_size_962 > np.int32(0)) else np.int32(1)))\r\n group_size_974 = np.int32(512)\r\n num_groups_975 = squot32((((width_734 * height_735) + group_size_974) - np.int32(1)),\r\n group_size_974)\r\n if ((np.int32(1) * (num_groups_975 * group_size_974)) != np.int32(0)):\r\n self.map_kernel_881_var.set_args(mem_940, res_mem_948,\r\n np.int32(limit_736), mem_944, mem_946,\r\n np.int32(width_734), np.float32(res_742),\r\n mem_942, np.float32(view_738),\r\n np.float32(y_746),\r\n np.int32(nesting_size_879),\r\n np.byte(x_747), np.int32(height_735),\r\n mem_952)\r\n cl.enqueue_nd_range_kernel(self.queue, self.map_kernel_881_var,\r\n (long((num_groups_975 * group_size_974)),),\r\n (long(group_size_974),))\r\n if synchronous:\r\n self.queue.finish()\r\n x_955 = (width_734 * height_735)\r\n bytes_953 = (x_955 * np.int32(3))\r\n mem_956 = cl.Buffer(self.ctx, cl.mem_flags.READ_WRITE,\r\n long(long(bytes_953) if (bytes_953 > np.int32(0)) else np.int32(1)))\r\n if ((((np.int32(1) * (height_735 + srem32((np.int32(16) - srem32(height_735,\r\n np.int32(16))),\r\n np.int32(16)))) * (np.int32(3) + srem32((np.int32(16) - srem32(np.int32(3),\r\n np.int32(16))),\r\n np.int32(16)))) * width_734) != np.int32(0)):\r\n self.fut_kernel_map_transpose_i8_var.set_args(mem_956,\r\n np.int32(np.int32(0)),\r\n mem_952,\r\n np.int32(np.int32(0)),\r\n np.int32(height_735),\r\n np.int32(np.int32(3)),\r\n np.int32(((width_734 * height_735) * np.int32(3))),\r\n cl.LocalMemory(long((((np.int32(16) + np.int32(1)) * np.int32(16)) * np.int32(1)))))\r\n cl.enqueue_nd_range_kernel(self.queue,\r\n self.fut_kernel_map_transpose_i8_var,\r\n (long((height_735 + srem32((np.int32(16) - srem32(height_735,\r\n np.int32(16))),\r\n np.int32(16)))),\r\n long((np.int32(3) + srem32((np.int32(16) - srem32(np.int32(3),\r\n np.int32(16))),\r\n np.int32(16)))),\r\n long(width_734)), (long(np.int32(16)),\r\n long(np.int32(16)),\r\n long(np.int32(1))))\r\n if synchronous:\r\n self.queue.finish()\r\n out_mem_963 = mem_956\r\n out_memsize_964 = bytes_953\r\n return (out_memsize_964, out_mem_963)\r\n def main(self, width_734_ext, height_735_ext, limit_736_ext, view_737_ext,\r\n view_738_ext, view_739_ext, view_740_ext):\r\n width_734 = np.int32(width_734_ext)\r\n height_735 = np.int32(height_735_ext)\r\n limit_736 = np.int32(limit_736_ext)\r\n view_737 = np.float32(view_737_ext)\r\n view_738 = np.float32(view_738_ext)\r\n view_739 = np.float32(view_739_ext)\r\n view_740 = np.float32(view_740_ext)\r\n (out_memsize_964, out_mem_963) = self.futhark_main(width_734, height_735,\r\n limit_736, view_737,\r\n view_738, view_739,\r\n view_740)\r\n return cl.array.Array(self.queue, (width_734, height_735, np.int32(3)),\r\n ct.c_int8, data=out_mem_963)", "sub_path": "examples/mandelbrot-explorer/mandelbrot.py", "file_name": "mandelbrot.py", "file_ext": "py", "file_size_in_byte": 37651, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "numpy.int32", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1029, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 1035, "usage_type": "attribute"}, {"api_name": "numpy.int8", "line_number": 1036, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 1037, "usage_type": "attribute"}, {"api_name": "numpy.int16", "line_number": 1038, "usage_type": "call"}, {"api_name": "numpy.uint32", "line_number": 1039, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 1040, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 1042, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 1045, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 1046, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 1047, "usage_type": "attribute"}, {"api_name": "numpy.uint16", "line_number": 1048, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1049, "usage_type": "attribute"}, {"api_name": "numpy.uint32", "line_number": 1050, "usage_type": "call"}, {"api_name": "numpy.uint64", "line_number": 1052, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1073, "usage_type": "call"}, {"api_name": "numpy.fmod", "line_number": 1076, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 1097, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 1097, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 1100, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 1100, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1103, "usage_type": "call"}, {"api_name": "numpy.uint32", "line_number": 1103, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 1106, "usage_type": "call"}, {"api_name": "numpy.uint64", "line_number": 1106, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 1109, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 1112, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1115, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1118, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 1121, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 1121, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 1124, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 1124, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1127, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 1127, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 1130, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 1130, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 1133, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 1133, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 1136, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 1136, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1139, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 1139, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 1142, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 1142, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 1145, "usage_type": "call"}, {"api_name": "numpy.uint32", "line_number": 1145, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 1148, "usage_type": "call"}, {"api_name": "numpy.uint32", "line_number": 1148, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1151, "usage_type": "call"}, {"api_name": "numpy.uint32", "line_number": 1151, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 1154, "usage_type": "call"}, {"api_name": "numpy.uint32", "line_number": 1154, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 1157, "usage_type": "call"}, {"api_name": "numpy.uint64", "line_number": 1157, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 1160, "usage_type": "call"}, {"api_name": "numpy.uint64", "line_number": 1160, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1163, "usage_type": "call"}, {"api_name": "numpy.uint64", "line_number": 1163, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 1166, "usage_type": "call"}, {"api_name": "numpy.uint64", "line_number": 1166, "usage_type": "call"}, {"api_name": "numpy.sign", "line_number": 1188, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1197, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 1201, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1205, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 1209, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 1213, "usage_type": "call"}, {"api_name": "numpy.trunc", "line_number": 1213, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 1217, "usage_type": "call"}, {"api_name": "numpy.trunc", "line_number": 1217, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1221, "usage_type": "call"}, {"api_name": "numpy.trunc", "line_number": 1221, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 1225, "usage_type": "call"}, {"api_name": "numpy.trunc", "line_number": 1225, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 1229, "usage_type": "call"}, {"api_name": "numpy.trunc", "line_number": 1229, "usage_type": "call"}, {"api_name": "numpy.uint16", "line_number": 1233, "usage_type": "call"}, {"api_name": "numpy.trunc", "line_number": 1233, "usage_type": "call"}, {"api_name": "numpy.uint32", "line_number": 1237, "usage_type": "call"}, {"api_name": "numpy.trunc", "line_number": 1237, "usage_type": "call"}, {"api_name": "numpy.uint64", "line_number": 1241, "usage_type": "call"}, {"api_name": "numpy.trunc", "line_number": 1241, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 1245, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1248, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 1251, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 1251, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 1254, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 1257, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 1260, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 1263, "usage_type": "call"}, {"api_name": "numpy.arctan2", "line_number": 1266, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 1269, "usage_type": "call"}, {"api_name": "numpy.isinf", "line_number": 1272, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1275, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 1275, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1278, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 1278, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 1281, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 1284, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 1287, "usage_type": "call"}, {"api_name": "numpy.arctan2", "line_number": 1290, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 1293, "usage_type": "call"}, {"api_name": "numpy.isinf", "line_number": 1296, "usage_type": "call"}, {"api_name": "pyopencl.create_some_context", "line_number": 1299, "usage_type": "call"}, {"api_name": "pyopencl.CommandQueue", "line_number": 1301, "usage_type": "call"}, {"api_name": "pyopencl.context_info", "line_number": 1305, "usage_type": "attribute"}, {"api_name": "pyopencl.context_info", "line_number": 1307, "usage_type": "attribute"}, {"api_name": "pyopencl.device_type", "line_number": 1311, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 1312, "usage_type": "call"}, {"api_name": "pyopencl.device_type", "line_number": 1313, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 1314, "usage_type": "call"}, {"api_name": "pyopencl.Program", "line_number": 1318, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1331, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1332, "usage_type": "call"}, {"api_name": "pyopencl.Buffer", "line_number": 1333, "usage_type": "call"}, {"api_name": "pyopencl.mem_flags", "line_number": 1333, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 1334, "usage_type": "call"}, {"api_name": "pyopencl.Buffer", "line_number": 1335, "usage_type": "call"}, {"api_name": "pyopencl.mem_flags", "line_number": 1335, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 1336, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1337, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1338, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1340, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1341, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1342, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1342, "usage_type": "call"}, {"api_name": "pyopencl.enqueue_nd_range_kernel", "line_number": 1344, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1350, "usage_type": "call"}, {"api_name": "pyopencl.Buffer", "line_number": 1352, "usage_type": "call"}, {"api_name": "pyopencl.mem_flags", "line_number": 1352, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 1353, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1354, "usage_type": "call"}, {"api_name": "pyopencl.Buffer", "line_number": 1355, "usage_type": "call"}, {"api_name": "pyopencl.mem_flags", "line_number": 1355, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 1356, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1357, "usage_type": "call"}, {"api_name": "pyopencl.Buffer", "line_number": 1358, "usage_type": "call"}, {"api_name": "pyopencl.mem_flags", "line_number": 1358, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 1359, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1360, "usage_type": "call"}, {"api_name": "pyopencl.Buffer", "line_number": 1361, "usage_type": "call"}, {"api_name": "pyopencl.mem_flags", "line_number": 1361, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 1362, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1363, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1364, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1366, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1368, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1369, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1369, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1370, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1371, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1372, "usage_type": "call"}, {"api_name": "numpy.byte", "line_number": 1373, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1373, "usage_type": "call"}, {"api_name": "pyopencl.enqueue_nd_range_kernel", "line_number": 1375, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1381, "usage_type": "call"}, {"api_name": "pyopencl.Buffer", "line_number": 1382, "usage_type": "call"}, {"api_name": "pyopencl.mem_flags", "line_number": 1382, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 1383, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1384, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1385, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1386, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1387, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1388, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1390, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1392, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1393, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1394, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1395, "usage_type": "call"}, {"api_name": "pyopencl.LocalMemory", "line_number": 1396, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1396, "usage_type": "call"}, {"api_name": "pyopencl.enqueue_nd_range_kernel", "line_number": 1397, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1399, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1400, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1401, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1402, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1403, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1404, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1405, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1406, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1407, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1415, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1416, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1417, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1418, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1419, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1420, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1421, "usage_type": "call"}, {"api_name": "pyopencl.array.Array", "line_number": 1426, "usage_type": "call"}, {"api_name": "pyopencl.array", "line_number": 1426, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 1426, "usage_type": "call"}, {"api_name": "ctypes.c_int8", "line_number": 1427, "usage_type": "attribute"}]} +{"seq_id": "318644389", "text": "import pygame\r\nimport pyganim\r\n\r\nimport os\r\nimport sys\r\n\r\nimport random\r\n\r\nimport time\r\n\r\nimport pymorphy2\r\n\r\nos.environ['SDL_VIDEO_CENTERED'] = '1'\r\npygame.init()\r\n\r\nsize = WIDTH, HEIGHT = 800, 600\r\nscreen = pygame.display.set_mode(size)\r\npygame.display.set_caption('Танки')\r\n\r\nclock = pygame.time.Clock()\r\nFPS = 50\r\nshoot_right = False\r\nshoot_left = False\r\nshoot_up = False\r\nshoot_down = False\r\n\r\n\r\ndef terminate():\r\n pygame.quit()\r\n sys.exit()\r\n\r\n\r\ndef load_image(name, colorkey=None):\r\n fullname = os.path.join('IMAGE', name)\r\n image = pygame.image.load(fullname).convert()\r\n if colorkey is not None:\r\n if colorkey == -1:\r\n colorkey = image.get_at((0, 0))\r\n image.set_colorkey(colorkey)\r\n else:\r\n image = image.convert_alpha()\r\n return image\r\n\r\n\r\ndef start_screen():\r\n intro_text = [\"Игра Танки\",\r\n \"Добро пожаловать в игру!\",\r\n \"Для начала нажмите на пробел\"]\r\n\r\n fon = pygame.transform.scale(load_image('fon1.jpg'), (WIDTH, HEIGHT))\r\n screen.blit(fon, (0, 0))\r\n font = pygame.font.Font(None, 72)\r\n text_coord = 50\r\n for line in intro_text:\r\n string_rendered = font.render(line, 1, pygame.Color('#2F4F4F'))\r\n intro_rect = string_rendered.get_rect()\r\n text_coord += 10\r\n intro_rect.top = text_coord\r\n intro_rect.x = 10\r\n text_coord += intro_rect.height\r\n screen.blit(string_rendered, intro_rect)\r\n\r\n\r\ndef end_screen(timeee):\r\n morph = pymorphy2.MorphAnalyzer()\r\n sec = morph.parse('секунда')[0]\r\n sec = sec.make_agree_with_number(timeee).word\r\n \r\n intro_text = [\"Поздравляем!\",\r\n \"Вы победили\",\r\n \"Время прохождения:\",\r\n str(round(timeee, 2)) + \" \" + sec]\r\n\r\n fon = pygame.transform.scale(load_image('fon3.jpg'), (WIDTH, HEIGHT))\r\n screen.blit(fon, (0, 0))\r\n font = pygame.font.Font(None, 72)\r\n text_coord = 100\r\n for line in intro_text:\r\n string_rendered = font.render(line, 1, pygame.Color('white'))\r\n intro_rect = string_rendered.get_rect()\r\n text_coord += 10\r\n intro_rect.top = text_coord\r\n intro_rect.x = 10\r\n text_coord += intro_rect.height\r\n screen.blit(string_rendered, intro_rect)\r\n \r\n \r\ndef not_good_end(timeee):\r\n morph = pymorphy2.MorphAnalyzer()\r\n sec = morph.parse('секунда')[0]\r\n sec = sec.make_agree_with_number(timeee).word\r\n intro_text = [\"К сожалению, вы проиграли\",\r\n \"Время прохождения:\",\r\n str(round(timeee, 2)) + \" \" + sec]\r\n\r\n fon = pygame.transform.scale(load_image('fon3.jpg'), (WIDTH, HEIGHT))\r\n screen.blit(fon, (0, 0))\r\n font = pygame.font.Font(None, 72)\r\n text_coord = 100\r\n for line in intro_text:\r\n string_rendered = font.render(line, 1, pygame.Color('#800000'))\r\n intro_rect = string_rendered.get_rect()\r\n text_coord += 10\r\n intro_rect.top = text_coord\r\n intro_rect.x = 10\r\n text_coord += intro_rect.height\r\n screen.blit(string_rendered, intro_rect)\r\n\r\n\r\ndef load_level(filename):\r\n filename = \"TEXT/\" + filename\r\n with open(filename, 'r') as mapFile:\r\n level_map = [line.strip() for line in mapFile]\r\n max_width = max(map(len, level_map))\r\n return list(map(lambda x: x.ljust(max_width, '.'), level_map))\r\n\r\n\r\nbricks = [load_image('brick1.png'), load_image('brick2.png'), load_image('brick3.png'),\r\n load_image('brick4.png'), load_image('brick5.png'), load_image('brick6.png'),\r\n load_image('brick7.png'), load_image('brick8.png'), load_image('brick9.png')]\r\nbrick = random.choice(bricks)\r\ntile_images = {'wall': brick, 'empty': load_image('empty.png'), 'water': load_image('water.png'),\r\n 'flagok': load_image('flagok.png', -1)}\r\nplayer_image = load_image('blue_tank.png', -1)\r\nenemy_image = load_image('en_tank.png', -1)\r\ntile_width = tile_height = 50\r\n\r\n\r\nclass Tile(pygame.sprite.Sprite):\r\n def __init__(self, tile_type, pos_x, pos_y):\r\n super().__init__(tiles_group, all_sprites)\r\n self.image = tile_images[tile_type]\r\n self.rect = self.image.get_rect().move(tile_width * pos_x, tile_height * pos_y)\r\n\r\n\r\nall_sprites = pygame.sprite.Group()\r\ntiles_group = pygame.sprite.Group()\r\nplayer_group = pygame.sprite.Group()\r\nbricks_group = pygame.sprite.Group()\r\nenemy_group = pygame.sprite.Group()\r\n\r\n\r\ndef generate_level(level):\r\n new_player, x, y = None, None, None\r\n for y in range(len(level)):\r\n for x in range(len(level[y])):\r\n if level[y][x] == '.':\r\n Tile('empty', x, y)\r\n elif level[y][x] == '#':\r\n a = Tile('wall', x, y)\r\n bricks_group.add(a)\r\n elif level[y][x] == '?':\r\n Tile('water', x, y)\r\n elif level[y][x] == '@':\r\n Tile('empty', x, y)\r\n new_player = Player(x, y)\r\n player_group.add(new_player)\r\n elif level[y][x] == 'e':\r\n Tile('empty', x, y)\r\n enem = Enemy(x, y)\r\n enemy_group.add(enem)\r\n elif level[y][x] == 'f':\r\n Tile('flagok', x, y)\r\n flag_x = x\r\n flag_y = y\r\n return [new_player, x, y], flag_x, flag_y\r\n\r\n\r\nclass Player(pygame.sprite.Sprite):\r\n def __init__(self, pos_x, pos_y):\r\n super().__init__(all_sprites)\r\n self.image = player_image\r\n self.x = pos_x\r\n self.y = pos_y\r\n self.rect = self.image.get_rect().move(tile_width * pos_x, tile_height * pos_y)\r\n\r\n def get_coord(self):\r\n return self.x, self.y\r\n\r\n def move_player_down(self, x, y):\r\n COLOR = \"#888888\"\r\n ANIMATION_DOWN = [('%s/IMAGE/tank/down1.png' % ICON_DIR),\r\n ('%s/IMAGE/tank/down3.png' % ICON_DIR)]\r\n boltAnim = []\r\n for anim in ANIMATION_DOWN:\r\n boltAnim.append((anim, 0.1))\r\n self.boltAnimDown = pyganim.PygAnimation(boltAnim)\r\n self.boltAnimDown.play()\r\n self.image.fill(pygame.Color(COLOR))\r\n self.boltAnimDown.blit(self.image, (0, 0))\r\n new = Player(self.x + x, self.y + y)\r\n collide = pygame.sprite.spritecollide(new, bricks_group, False)\r\n collide_with_en = pygame.sprite.spritecollide(new, enemy_group, False)\r\n if collide or collide_with_en or self.rect.bottom > 600:\r\n y = 0\r\n self.x += x\r\n self.y += y\r\n self.rect = self.image.get_rect().move(tile_width * self.x, tile_height * self.y)\r\n\r\n def move_player_up(self, x, y):\r\n COLOR = \"#888888\"\r\n ANIMATION_UP = [('%s/IMAGE/tank/up1.png' % ICON_DIR),\r\n ('%s/IMAGE/tank/up3.png' % ICON_DIR)]\r\n boltAnim = []\r\n for anim in ANIMATION_UP:\r\n boltAnim.append((anim, 0.1))\r\n self.boltAnimUp = pyganim.PygAnimation(boltAnim)\r\n self.boltAnimUp.play()\r\n self.image.fill(pygame.Color(COLOR))\r\n self.boltAnimUp.blit(self.image, (0, 0))\r\n new = Player(self.x + x, self.y + y)\r\n collide = pygame.sprite.spritecollide(new, bricks_group, False)\r\n collide_with_en = pygame.sprite.spritecollide(new, enemy_group, False)\r\n if collide or collide_with_en or self.rect.top < 0:\r\n y = 0\r\n self.x += x\r\n self.y += y\r\n self.rect = self.image.get_rect().move(tile_width * self.x, tile_height * self.y)\r\n\r\n def move_player_left(self, x, y):\r\n COLOR = \"#888888\"\r\n ANIMATION_LEFT = [('%s/IMAGE/tank/l1.png' % ICON_DIR),\r\n ('%s/IMAGE/tank/l3.png' % ICON_DIR)]\r\n boltAnim = []\r\n for anim in ANIMATION_LEFT:\r\n boltAnim.append((anim, 0.1))\r\n self.boltAnimLeft = pyganim.PygAnimation(boltAnim)\r\n self.boltAnimLeft.play()\r\n self.image.fill(pygame.Color(COLOR))\r\n self.boltAnimLeft.blit(self.image, (0, 0))\r\n new = Player(self.x + x, self.y + y)\r\n collide = pygame.sprite.spritecollide(new, bricks_group, False)\r\n collide_with_en = pygame.sprite.spritecollide(new, enemy_group, False)\r\n if collide or collide_with_en or self.rect.left < 0:\r\n x = 0\r\n self.x += x\r\n self.y += y\r\n self.rect = self.image.get_rect().move(tile_width * self.x, tile_height * self.y)\r\n\r\n def move_player_right(self, x, y):\r\n COLOR = \"#888888\"\r\n ANIMATION_RIGHT = [('%s/IMAGE/tank/r1.png' % ICON_DIR),\r\n ('%s/IMAGE/tank/r3.png' % ICON_DIR)]\r\n boltAnim = []\r\n for anim in ANIMATION_RIGHT:\r\n boltAnim.append((anim, 0.1))\r\n self.boltAnimRight = pyganim.PygAnimation(boltAnim)\r\n self.boltAnimRight.play()\r\n self.image.fill(pygame.Color(COLOR))\r\n self.boltAnimRight.blit(self.image, (0, 0))\r\n if self.rect.right > WIDTH:\r\n x = 0\r\n new = Player(self.x + x, self.y + y)\r\n collide = pygame.sprite.spritecollide(new, bricks_group, False)\r\n collide_with_en = pygame.sprite.spritecollide(new, enemy_group, False)\r\n if collide or collide_with_en or self.rect.left > 800:\r\n x = 0\r\n self.x += x\r\n self.y += y\r\n self.rect = self.image.get_rect().move(tile_width * self.x, tile_height * self.y)\r\n\r\n def shoot(self):\r\n bullet = Bullet(self.rect.centerx, self.rect.top)\r\n bullets.add(bullet)\r\n\r\n\r\nclass Enemy(pygame.sprite.Sprite):\r\n def __init__(self, pos_x, pos_y):\r\n super().__init__(all_sprites)\r\n self.image = enemy_image\r\n self.x = pos_x\r\n self.y = pos_y\r\n self.kill = False\r\n self.rect = self.image.get_rect().move(tile_width * pos_x, tile_height * pos_y)\r\n\r\n def get_coord(self):\r\n return (self.x, self.y)\r\n\r\n def move_player_down(self):\r\n COLOR = \"#888888\"\r\n ANIMATION_DOWN = [('%s/IMAGE/tank/end1.png' % ICON_DIR),\r\n ('%s/IMAGE/tank/end3.png' % ICON_DIR)]\r\n boltAnim = []\r\n for anim in ANIMATION_DOWN:\r\n boltAnim.append((anim, 0.1))\r\n self.boltAnimDown = pyganim.PygAnimation(boltAnim)\r\n self.boltAnimDown.play()\r\n self.image.fill(pygame.Color(COLOR))\r\n self.boltAnimDown.blit(self.image, (0, 0))\r\n\r\n def move_player_up(self):\r\n COLOR = \"#888888\"\r\n ANIMATION_UP = [('%s/IMAGE/tank/enu1.png' % ICON_DIR),\r\n ('%s/IMAGE/tank/enu3.png' % ICON_DIR)]\r\n boltAnim = []\r\n for anim in ANIMATION_UP:\r\n boltAnim.append((anim, 0.1))\r\n self.boltAnimUp = pyganim.PygAnimation(boltAnim)\r\n self.boltAnimUp.play()\r\n self.image.fill(pygame.Color(COLOR))\r\n self.boltAnimUp.blit(self.image, (0, 0))\r\n\r\n\r\n def move_player_left(self):\r\n COLOR = \"#888888\"\r\n ANIMATION_LEFT = [('%s/IMAGE/tank/enl1.png' % ICON_DIR),\r\n ('%s/IMAGE/tank/enl3.png' % ICON_DIR)]\r\n boltAnim = []\r\n for anim in ANIMATION_LEFT:\r\n boltAnim.append((anim, 0.1))\r\n self.boltAnimLeft = pyganim.PygAnimation(boltAnim)\r\n self.boltAnimLeft.play()\r\n self.image.fill(pygame.Color(COLOR))\r\n self.boltAnimLeft.blit(self.image, (0, 0))\r\n\r\n\r\n def move_player_right(self):\r\n COLOR = \"#888888\"\r\n ANIMATION_RIGHT = [('%s/IMAGE/tank/enr1.png' % ICON_DIR),\r\n ('%s/IMAGE/tank/enr3.png' % ICON_DIR)]\r\n boltAnim = []\r\n for anim in ANIMATION_RIGHT:\r\n boltAnim.append((anim, 0.1))\r\n self.boltAnimRight = pyganim.PygAnimation(boltAnim)\r\n self.boltAnimRight.play()\r\n self.image.fill(pygame.Color(COLOR))\r\n self.boltAnimRight.blit(self.image, (0, 0))\r\n\r\n\r\n def move(self, x, y):\r\n new = Enemy(self.x + x, self.y + y)\r\n collide = pygame.sprite.spritecollide(new, bricks_group, False)\r\n collide_with_en = pygame.sprite.spritecollide(new, player_group, False)\r\n if not (collide or collide_with_en):\r\n self.x += x\r\n self.y += y\r\n if self.rect.right > 800:\r\n self.x -= 0.5\r\n if self.rect.left < 0:\r\n self.x += 0.5\r\n if self.rect.top < 0:\r\n self.y += 0.5\r\n if self.rect.bottom > 600:\r\n self.y -= 0.5\r\n self.rect = self.image.get_rect().move(tile_width * self.x, tile_height * self.y)\r\n\r\n def shoot(self):\r\n bullet = Bullet(self.rect.centerx, self.rect.top)\r\n bullets.add(bullet)\r\n\r\n\r\nclass Bullet(pygame.sprite.Sprite):\r\n def __init__(self, x, y):\r\n BLUE = (255, 255, 0)\r\n pygame.sprite.Sprite.__init__(self)\r\n self.image = pygame.Surface((2, 2))\r\n self.image.fill(BLUE)\r\n self.rect = self.image.get_rect()\r\n if shoot_up:\r\n self.rect.bottom = y\r\n self.rect.centerx = x\r\n if shoot_down:\r\n self.rect.bottom = y + 40\r\n self.rect.centerx = x\r\n if shoot_left:\r\n self.rect.bottom = y + 20\r\n self.rect.centerx = x - 20\r\n if shoot_right:\r\n self.rect.bottom = y + 20\r\n self.rect.centerx = x + 20\r\n\r\n def update(self):\r\n if shoot_up:\r\n self.rect.y -= 10\r\n if self.rect.bottom < 0:\r\n self.kill()\r\n hit = pygame.sprite.groupcollide(bullets, bricks_group, False, False)\r\n if hit:\r\n self.kill()\r\n elif shoot_down:\r\n self.rect.y += 10\r\n if self.rect.bottom > 600:\r\n self.kill()\r\n hit = pygame.sprite.groupcollide(bullets, bricks_group, False, False)\r\n if hit:\r\n self.kill()\r\n elif shoot_left:\r\n self.rect.x -= 10\r\n if self.rect.right < 0:\r\n self.kill()\r\n hit = pygame.sprite.groupcollide(bullets, bricks_group, False, False)\r\n if hit:\r\n self.kill()\r\n elif shoot_right:\r\n self.rect.x += 10\r\n if self.rect.left > 800:\r\n self.kill()\r\n hit = pygame.sprite.groupcollide(bullets, bricks_group, False, False)\r\n if hit:\r\n self.kill()\r\n\r\nICON_DIR = os.path.dirname(__file__) # путь Рє каталогу СЃ файлами\r\nANIMATION_RIGHT = [('%s/IMAGE/tank/down1.png' % ICON_DIR),\r\n ('%s/IMAGE/tank/down3.png' % ICON_DIR)]\r\n\r\n\r\ndef move(st, x, y):\r\n x1, y1 = player.get_coord()\r\n for i in range(len(level)):\r\n level[i] = level[i].replace('@', '.')\r\n\r\n if st.lower() == 'down':\r\n player.move_player_down(x, y)\r\n if st.lower() == 'up':\r\n player.move_player_up(x, y)\r\n if st.lower() == 'left':\r\n player.move_player_left(x, y)\r\n if st.lower() == 'right':\r\n player.move_player_right(x, y)\r\n\r\n\r\nlevels = [load_level('map1.txt'), load_level('map2.txt'), load_level('map3.txt'), load_level('map4.txt'),\r\n load_level('map5.txt'), load_level('map6.txt')]\r\nlevels_for_game = random.sample(levels, 3)\r\n\r\n\r\ndef do_level():\r\n if levels_for_game:\r\n all_sprites.empty()\r\n tiles_group.empty()\r\n player_group.empty()\r\n bricks_group.empty()\r\n enemy_group.empty()\r\n player = None\r\n l = levels_for_game[0]\r\n s, flag1, flag2 = generate_level(l)\r\n levels_for_game.remove(l)\r\n return l, s, flag1, flag2\r\n\r\n\r\ndef random_move():\r\n napr = ['up', 'down', 'left', 'right']\r\n for enem in enemy_group:\r\n random_napr = random.choice(napr)\r\n if random_napr == 'up':\r\n enem.move_player_down()\r\n enem.move(0, 0.5)\r\n elif random_napr == 'down':\r\n enem.move_player_up()\r\n enem.move(0, -0.5)\r\n elif random_napr == 'left':\r\n enem.move_player_left()\r\n enem.move(-0.5, 0)\r\n else:\r\n enem.move_player_right()\r\n enem.move(0.5, 0)\r\n enem.shoot()\r\n\r\n\r\ntok = time.time()\r\nrunning = True\r\nwhile running:\r\n for event in pygame.event.get():\r\n if event.type == pygame.QUIT:\r\n terminate()\r\n elif event.type == pygame.KEYDOWN and event.key == pygame.K_ESCAPE:\r\n terminate()\r\n elif event.type == pygame.MOUSEBUTTONDOWN:\r\n running = False\r\n elif event.type == pygame.KEYDOWN and event.key == pygame.K_SPACE:\r\n running = False\r\n start_screen()\r\n pygame.display.flip()\r\n clock.tick(FPS)\r\n\r\n\r\nfoo = True\r\n\r\nwhile levels_for_game and foo:\r\n level, s, flag_x, flag_y = do_level()\r\n player, level_x, level_y = s[0], s[1], s[2]\r\n screen.fill((0, 0, 0))\r\n bullets = pygame.sprite.Group()\r\n running = True\r\n while running:\r\n for event in pygame.event.get():\r\n if event.type == pygame.QUIT:\r\n terminate()\r\n if event.type == pygame.KEYDOWN:\r\n random_move()\r\n if event.key == pygame.K_DOWN:\r\n move(\"down\", 0, 0.5)\r\n shoot_right = False\r\n shoot_left = False\r\n shoot_up = False\r\n shoot_down = True\r\n if event.key == pygame.K_UP:\r\n move(\"up\", 0, -0.5)\r\n shoot_right = False\r\n shoot_left = False\r\n shoot_up = True\r\n shoot_down = False\r\n if event.key == pygame.K_LEFT:\r\n move(\"left\", -0.5, 0)\r\n shoot_right = False\r\n shoot_left = True\r\n shoot_up = False\r\n shoot_down = False\r\n if event.key == pygame.K_RIGHT:\r\n move(\"right\", 0.5, 0)\r\n shoot_right = True\r\n shoot_left = False\r\n shoot_up = False\r\n shoot_down = False\r\n if event.key == pygame.K_SPACE:\r\n player.shoot()\r\n if player.get_coord()[0] == flag_x and player.get_coord()[1] == flag_y:\r\n running = False\r\n break\r\n collide = pygame.sprite.spritecollide(player, bullets, False)\r\n if collide:\r\n running = False\r\n foo = False\r\n enemy_group.update()\r\n bullets.update()\r\n tiles_group.draw(screen)\r\n enemy_group.draw(screen)\r\n player_group.draw(screen)\r\n bullets.draw(screen)\r\n pygame.display.flip()\r\n clock.tick(FPS)\r\n \r\ntik = time.time()\r\ntimeall = tik - tok\r\n\r\nif not levels_for_game or not foo:\r\n run = True\r\n while run:\r\n for event in pygame.event.get():\r\n if event.type == pygame.QUIT:\r\n terminate()\r\n elif event.type == pygame.KEYDOWN and event.key == pygame.K_ESCAPE:\r\n terminate()\r\n elif event.type == pygame.MOUSEBUTTONDOWN:\r\n run = False\r\n elif event.type == pygame.KEYDOWN and event.key == pygame.K_SPACE:\r\n run = False\r\n if foo:\r\n end_screen(timeall)\r\n else:\r\n not_good_end(timeall)\r\n pygame.display.flip()\r\n clock.tick(FPS)\r\npygame.quit()\r\nsys.exit()", "sub_path": "tanksss.py", "file_name": "tanksss.py", "file_ext": "py", "file_size_in_byte": 19496, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "os.environ", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pygame.init", "line_number": 14, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 17, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 17, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 18, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 20, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 20, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 29, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 35, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 35, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 50, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 50, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 52, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 52, "usage_type": "attribute"}, {"api_name": "pygame.Color", "line_number": 55, "usage_type": "call"}, {"api_name": "pymorphy2.MorphAnalyzer", "line_number": 65, "usage_type": "call"}, {"api_name": "pygame.transform.scale", "line_number": 74, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 74, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 76, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 76, "usage_type": "attribute"}, {"api_name": "pygame.Color", "line_number": 79, "usage_type": "call"}, {"api_name": "pymorphy2.MorphAnalyzer", "line_number": 89, "usage_type": "call"}, {"api_name": "pygame.transform.scale", "line_number": 96, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 96, "usage_type": "attribute"}, {"api_name": "pygame.font.Font", "line_number": 98, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 98, "usage_type": "attribute"}, {"api_name": "pygame.Color", "line_number": 101, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 121, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 129, "usage_type": "attribute"}, {"api_name": "pygame.sprite.Group", "line_number": 136, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 136, "usage_type": "attribute"}, {"api_name": "pygame.sprite.Group", "line_number": 137, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 137, "usage_type": "attribute"}, {"api_name": "pygame.sprite.Group", "line_number": 138, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 138, "usage_type": "attribute"}, {"api_name": "pygame.sprite.Group", "line_number": 139, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 139, "usage_type": "attribute"}, {"api_name": "pygame.sprite.Group", "line_number": 140, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 140, "usage_type": "attribute"}, {"api_name": "pygame.sprite", "line_number": 169, "usage_type": "attribute"}, {"api_name": "pyganim.PygAnimation", "line_number": 187, "usage_type": "call"}, {"api_name": "pygame.Color", "line_number": 189, "usage_type": "call"}, {"api_name": "pygame.sprite.spritecollide", "line_number": 192, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 192, "usage_type": "attribute"}, {"api_name": "pygame.sprite.spritecollide", "line_number": 193, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 193, "usage_type": "attribute"}, {"api_name": "pyganim.PygAnimation", "line_number": 207, "usage_type": "call"}, {"api_name": "pygame.Color", "line_number": 209, "usage_type": "call"}, {"api_name": "pygame.sprite.spritecollide", "line_number": 212, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 212, "usage_type": "attribute"}, {"api_name": "pygame.sprite.spritecollide", "line_number": 213, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 213, "usage_type": "attribute"}, {"api_name": "pyganim.PygAnimation", "line_number": 227, "usage_type": "call"}, {"api_name": "pygame.Color", "line_number": 229, "usage_type": "call"}, {"api_name": "pygame.sprite.spritecollide", "line_number": 232, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 232, "usage_type": "attribute"}, {"api_name": "pygame.sprite.spritecollide", "line_number": 233, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 233, "usage_type": "attribute"}, {"api_name": "pyganim.PygAnimation", "line_number": 247, "usage_type": "call"}, {"api_name": "pygame.Color", "line_number": 249, "usage_type": "call"}, {"api_name": "pygame.sprite.spritecollide", "line_number": 254, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 254, "usage_type": "attribute"}, {"api_name": "pygame.sprite.spritecollide", "line_number": 255, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 255, "usage_type": "attribute"}, {"api_name": "pygame.sprite", "line_number": 267, "usage_type": "attribute"}, {"api_name": "pyganim.PygAnimation", "line_number": 286, "usage_type": "call"}, {"api_name": "pygame.Color", "line_number": 288, "usage_type": "call"}, {"api_name": "pyganim.PygAnimation", "line_number": 298, "usage_type": "call"}, {"api_name": "pygame.Color", "line_number": 300, "usage_type": "call"}, {"api_name": "pyganim.PygAnimation", "line_number": 311, "usage_type": "call"}, {"api_name": "pygame.Color", "line_number": 313, "usage_type": "call"}, {"api_name": "pyganim.PygAnimation", "line_number": 324, "usage_type": "call"}, {"api_name": "pygame.Color", "line_number": 326, "usage_type": "call"}, {"api_name": "pygame.sprite.spritecollide", "line_number": 332, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 332, "usage_type": "attribute"}, {"api_name": "pygame.sprite.spritecollide", "line_number": 333, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 333, "usage_type": "attribute"}, {"api_name": "pygame.sprite", "line_number": 352, "usage_type": "attribute"}, {"api_name": "pygame.sprite.Sprite.__init__", "line_number": 355, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 355, "usage_type": "attribute"}, {"api_name": "pygame.Surface", "line_number": 356, "usage_type": "call"}, {"api_name": "pygame.sprite.groupcollide", "line_number": 377, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 377, "usage_type": "attribute"}, {"api_name": "pygame.sprite.groupcollide", "line_number": 384, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 384, "usage_type": "attribute"}, {"api_name": "pygame.sprite.groupcollide", "line_number": 391, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 391, "usage_type": "attribute"}, {"api_name": "pygame.sprite.groupcollide", "line_number": 398, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 398, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 402, "usage_type": "call"}, {"api_name": "os.path", "line_number": 402, "usage_type": "attribute"}, {"api_name": "random.sample", "line_number": 424, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 444, "usage_type": "call"}, {"api_name": "time.time", "line_number": 460, "usage_type": "call"}, {"api_name": "pygame.event.get", "line_number": 463, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 463, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 464, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 466, "usage_type": "attribute"}, {"api_name": "pygame.K_ESCAPE", "line_number": 466, "usage_type": "attribute"}, {"api_name": "pygame.MOUSEBUTTONDOWN", "line_number": 468, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 470, "usage_type": "attribute"}, {"api_name": "pygame.K_SPACE", "line_number": 470, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 473, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 473, "usage_type": "attribute"}, {"api_name": "pygame.sprite.Group", "line_number": 483, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 483, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 486, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 486, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 487, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 489, "usage_type": "attribute"}, {"api_name": "pygame.K_DOWN", "line_number": 491, "usage_type": "attribute"}, {"api_name": "pygame.K_UP", "line_number": 497, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 503, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 509, "usage_type": "attribute"}, {"api_name": "pygame.K_SPACE", "line_number": 515, "usage_type": "attribute"}, {"api_name": "pygame.sprite.spritecollide", "line_number": 520, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 520, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 530, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 530, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 533, "usage_type": "call"}, {"api_name": "pygame.event.get", "line_number": 539, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 539, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 540, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 542, "usage_type": "attribute"}, {"api_name": "pygame.K_ESCAPE", "line_number": 542, "usage_type": "attribute"}, {"api_name": "pygame.MOUSEBUTTONDOWN", "line_number": 544, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 546, "usage_type": "attribute"}, {"api_name": "pygame.K_SPACE", "line_number": 546, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 552, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 552, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 554, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 555, "usage_type": "call"}]} +{"seq_id": "12332823", "text": "import io, math\nfrom typing import Tuple, Iterable, Optional, Union\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n\ndef info(arr, header=None):\n if header is None:\n header = \"=\"*30\n print(header)\n print(\"shape: \", arr.shape)\n print(\"dtype: \", arr.dtype)\n print(\"min, max: \", min(np.ravel(arr)), max(np.ravel(arr)))\n\ndef get_fig(n_total: int, nrows: int=None, factor=3.0) -> Tuple[plt.Figure, plt.Axes]:\n \"\"\"Create a tuple of plt.Figure and plt.Axes with total number of subplots `n_total` with `nrows` number of rows.\n By default, nrows and ncols are sqrt of n_total.\n\n :param n_total: total number of subplots\n :param nrows: number of rows in this Figure\n :param factor: scaling factor that is multipled to both to the row and column sizes\n :return: Tuple[Figure, flatten list of Axes]\n \"\"\"\n if nrows is None:\n nrows = math.ceil(n_total ** .5)\n\n ncols = math.ceil(n_total / nrows)\n f, axes = plt.subplots(nrows=nrows, ncols=ncols, figsize=(factor * ncols, factor * nrows))\n axes = axes.flatten()\n return f, axes\n\n\ndef show_npimgs(npimgs: Iterable[np.ndarray], *,\n titles: Iterable[Union[str, int]]=None,\n nrows: int=None,\n factor=3.0,\n cmap:str = None,\n title: Optional[str] = None,\n set_axis_off: bool=True) -> plt.Axes:\n n_imgs = len(npimgs)\n f, axes = get_fig(n_imgs, nrows=nrows, factor=factor)\n\n for i, ax in enumerate(axes):\n if i < n_imgs:\n ax.imshow(npimgs[i], cmap=cmap)\n\n if titles is not None:\n ax.set_title(titles[i])\n if set_axis_off:\n ax.set_axis_off()\n else:\n f.delaxes(ax)\n if title is not None:\n f.suptitle(title)\n return axes\n\ndef plt_figure_to_np(fig, dpi=30):\n io_buf = io.BytesIO()\n fig.savefig(io_buf, format='raw', dpi=dpi)\n io_buf.seek(0)\n img_arr = np.reshape(np.frombuffer(io_buf.getvalue(), dtype=np.uint8),\n newshape=(int(fig.bbox.bounds[3]), int(fig.bbox.bounds[2]), -1))\n io_buf.close()\n return img_arr", "sub_path": "tilemani/utils/np.py", "file_name": "np.py", "file_ext": "py", "file_size_in_byte": 2145, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "numpy.ravel", "line_number": 13, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 25, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 15, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.Figure", "line_number": 15, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.Axes", "line_number": 15, "usage_type": "attribute"}, {"api_name": "typing.Iterable", "line_number": 33, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 33, "usage_type": "attribute"}, {"api_name": "typing.Iterable", "line_number": 34, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 34, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.Axes", "line_number": 39, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.frombuffer", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 61, "usage_type": "attribute"}]} +{"seq_id": "308006911", "text": "# Licensed to the Apache Software Foundation (ASF) under one\n# or more contributor license agreements. See the NOTICE file\n# distributed with this work for additional information\n# regarding copyright ownership. The ASF licenses this file\n# to you under the Apache License, Version 2.0 (the\n# \"License\"); you may not use this file except in compliance\n# with the License. You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing,\n# software distributed under the License is distributed on an\n# \"AS IS\" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY\n# KIND, either express or implied. See the License for the\n# specific language governing permissions and limitations\n# under the License.\n\nimport mxnet as mx\nfrom mxnet.gluon import nn\nfrom mxnet import amp\nimport numpy as np\nimport pytest\n\n\n@pytest.fixture\ndef np_shape_array():\n flags = mx.npx.is_np_shape(), mx.npx.is_np_array(), mx.npx.is_np_default_dtype()\n mx.npx.set_np()\n yield\n mx.npx.set_np(*flags)\n\n\n@pytest.fixture(scope='module')\ndef amp_init():\n amp.init()\n\n\ndef test_npi_concatenate_multicast(np_shape_array, amp_init):\n class Foo(nn.HybridBlock):\n def __init__(self, **kwargs):\n super().__init__(**kwargs)\n self.dense0 = nn.Dense(16, in_units=8)\n\n def forward(self, x):\n y = self.dense0(x)\n return mx.np.concatenate([y, x], axis=-1)\n\n foo = Foo()\n foo.initialize(ctx=mx.gpu())\n\n data = mx.np.ones((32, 8), ctx=mx.gpu())\n out = foo(data)\n assert out.dtype == np.float32\n", "sub_path": "tests/python/gpu/test_amp_init.py", "file_name": "test_amp_init.py", "file_ext": "py", "file_size_in_byte": 1623, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "mxnet.npx.is_np_shape", "line_number": 27, "usage_type": "call"}, {"api_name": "mxnet.npx", "line_number": 27, "usage_type": "attribute"}, {"api_name": "mxnet.npx.is_np_array", "line_number": 27, "usage_type": "call"}, {"api_name": "mxnet.npx.is_np_default_dtype", "line_number": 27, "usage_type": "call"}, {"api_name": "mxnet.npx.set_np", "line_number": 28, "usage_type": "call"}, {"api_name": "mxnet.npx", "line_number": 28, "usage_type": "attribute"}, {"api_name": "mxnet.npx.set_np", "line_number": 30, "usage_type": "call"}, {"api_name": "mxnet.npx", "line_number": 30, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 25, "usage_type": "attribute"}, {"api_name": "mxnet.amp.init", "line_number": 35, "usage_type": "call"}, {"api_name": "mxnet.amp", "line_number": 35, "usage_type": "name"}, {"api_name": "pytest.fixture", "line_number": 33, "usage_type": "call"}, {"api_name": "mxnet.gluon.nn.HybridBlock", "line_number": 39, "usage_type": "attribute"}, {"api_name": "mxnet.gluon.nn", "line_number": 39, "usage_type": "name"}, {"api_name": "mxnet.gluon.nn.Dense", "line_number": 42, "usage_type": "call"}, {"api_name": "mxnet.gluon.nn", "line_number": 42, "usage_type": "name"}, {"api_name": "mxnet.np.concatenate", "line_number": 46, "usage_type": "call"}, {"api_name": "mxnet.np", "line_number": 46, "usage_type": "attribute"}, {"api_name": "mxnet.gpu", "line_number": 49, "usage_type": "call"}, {"api_name": "mxnet.np.ones", "line_number": 51, "usage_type": "call"}, {"api_name": "mxnet.np", "line_number": 51, "usage_type": "attribute"}, {"api_name": "mxnet.gpu", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 53, "usage_type": "attribute"}]} +{"seq_id": "269208501", "text": "import importlib\nimport re\nimport time\nimport pymysql\nfrom handler import db_con\n\nfrom sub_module import csvX\nfrom fuzzywuzzy import fuzz\n\nimportlib.reload(fuzz)\nimportlib.reload(csvX)\n\nstartALL = time.clock()\nwd = 'd://Dropbox//Dropbox//Develop//python//leha//data//geol_app//'\ndata = csvX.read(wd + 'pi.csv', 'id;pi')\ndb_dict_pi = db_con.db_connection()\ncur_dict_pi = db_dict_pi.cursor(pymysql.cursors.DictCursor)\n\npi_item_mas = []\nnone_pi = set()\n\nfor obj in data[1:]:\n\n if obj['pi'] is not [None, 0, '']:\n obj['pi'] = (obj['pi']).replace('(', ' (')\n obj['pi'] = re.sub(r'([\\s][\\s]+)', ' ', obj['pi'])\n pi_item_mas = re.split(r',(?![^()]+\\))', obj['pi'].lower())\n obj['pi_mas'] = pi_item_mas\n\n for pi in pi_item_mas:\n pi = re.sub(r'\\(.+?\\)', '', pi)\n pi = pi.strip()\n\n piobj_in_pidict = cur_dict_pi.execute(\n \"SELECT * FROM dic_pi WHERE pi = %s\", pi)\n\n if piobj_in_pidict is 0:\n none_pi.add(pi)\n\ncsvX.simplewrite(wd + '!NOT_FOUD_PI_test!.csv', none_pi, 'not_found_pi')", "sub_path": "db/import/prepare_import.py", "file_name": "prepare_import.py", "file_ext": "py", "file_size_in_byte": 1083, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "importlib.reload", "line_number": 10, "usage_type": "call"}, {"api_name": "fuzzywuzzy.fuzz", "line_number": 10, "usage_type": "argument"}, {"api_name": "importlib.reload", "line_number": 11, "usage_type": "call"}, {"api_name": "sub_module.csvX", "line_number": 11, "usage_type": "argument"}, {"api_name": "time.clock", "line_number": 13, "usage_type": "call"}, {"api_name": "sub_module.csvX.read", "line_number": 15, "usage_type": "call"}, {"api_name": "sub_module.csvX", "line_number": 15, "usage_type": "name"}, {"api_name": "handler.db_con.db_connection", "line_number": 16, "usage_type": "call"}, {"api_name": "handler.db_con", "line_number": 16, "usage_type": "name"}, {"api_name": "pymysql.cursors", "line_number": 17, "usage_type": "attribute"}, {"api_name": "re.sub", "line_number": 26, "usage_type": "call"}, {"api_name": "re.split", "line_number": 27, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 31, "usage_type": "call"}, {"api_name": "sub_module.csvX.simplewrite", "line_number": 40, "usage_type": "call"}, {"api_name": "sub_module.csvX", "line_number": 40, "usage_type": "name"}]} +{"seq_id": "137279716", "text": "#!/usr/bin/env python2.7\n\n\"\"\"\nUniversal Command Line Environment for Continous Delivery Pipeline on Gitlab-CI.\nUsage:\n cdp build [(-v | --verbose | -q | --quiet)] [(-d | --dry-run)] [--sleep=]\n (--docker-image=)\n (--command=)\n [--simulate-merge-on=]\n [--volume-from=]\n cdp maven [(-v | --verbose | -q | --quiet)] [(-d | --dry-run)] [--sleep=]\n (--docker-version=)\n (--goals=|--deploy=)\n [--maven-release-plugin=]\n [--simulate-merge-on=]\n [--volume-from=]\n cdp sonar [(-v | --verbose | -q | --quiet)] [(-d | --dry-run)] [--sleep=]\n (--preview | --publish)\n (--codeclimate | --sast)\n [--simulate-merge-on=]\n cdp docker [(-v | --verbose | -q | --quiet)] [(-d | --dry-run)] [--sleep=]\n [--use-docker | --use-docker-compose]\n [--image-tag-branch-name] [--image-tag-latest] [--image-tag-sha1]\n [--use-gitlab-registry | --use-aws-ecr | --use-custom-registry]\n cdp artifactory [(-v | --verbose | -q | --quiet)] [(-d | --dry-run)] [--sleep=]\n [--image-tag-branch-name] [--image-tag-latest] [--image-tag-sha1]\n (--put= | --delete=)\n cdp k8s [(-v | --verbose | -q | --quiet)] [(-d | --dry-run)] [--sleep=]\n [--image-tag-branch-name | --image-tag-latest | --image-tag-sha1]\n (--use-gitlab-registry | --use-aws-ecr | --use-custom-registry)\n [--values=]\n [--delete-labels=]\n [--namespace-project-branch-name | --namespace-project-name]\n [--create-default-helm] [--deploy-spec-dir=]\n [--timeout=]\n cdp validator [(-v | --verbose | -q | --quiet)] [(-d | --dry-run)] [--sleep=]\n [--path=]\n [--block-provider | --block | --block-json]\n [--namespace-project-branch-name | --namespace-project-name]\n cdp (-h | --help | --version)\nOptions:\n -h, --help Show this screen and exit.\n -v, --verbose Make more noise.\n -q, --quiet Make less noise.\n -d, --dry-run Simulate execution.\n --sleep= Time to sleep int the end (for debbuging) in seconds [default: 0].\n --docker-image= Specify docker image name for build project.\n --command= Command to run in the docker image.\n --simulate-merge-on= Build docker image with the merge current branch on specify branch (no commit).\n --volume-from= Volume type of sources - docker or k8s [default: k8s]\n --docker-version= Specify maven docker version [default: 3.5-jdk-8].\n --goals= Goals and args to pass maven command.\n --deploy= 'release' or 'snapshot' - Maven command to deploy artifact.\n --maven-release-plugin= Specify maven-release-plugin version [default: 2.5.3].\n --preview Run issues mode (Preview).\n --publish Run publish mode (Analyse).\n --codeclimate Codeclimate mode.\n --sast Static Application Security Testing mode.\n --use-docker Use docker to build / push image [default].\n --use-docker-compose Use docker-compose to build / push image.\n --image-tag-branch-name Tag docker image with branch name or use it [default].\n --image-tag-latest Tag docker image with 'latest' or use it.\n --image-tag-sha1 Tag docker image with commit sha1 or use it.\n --use-gitlab-registry Use gitlab registry for pull/push docker image [default].\n --use-aws-ecr Use AWS ECR from k8s configuration for pull/push docker image.\n --use-custom-registry Use custom registry for pull/push docker image.\n --put= Put file to artifactory.\n --delete= Delete file in artifactory.\n --values= Specify values in a YAML file (can specify multiple separate by comma). The priority will be given to the last (right-most) file specified.\n --delete-labels= Add namespace labels (deletable=true deletionTimestamp=now + minutes) for external cleanup.\n --namespace-project-branch-name Use project and branch name to create k8s namespace or choice environment host [default].\n --namespace-project-name Use project name to create k8s namespace or choice environment host.\n --create-default-helm Create default helm for simple project (One docker image).\n --deploy-spec-dir= k8s deployment files [default: charts].\n --timeout= Time in seconds to wait for any individual kubernetes operation [default: 300].\n --path= Path to validate [default: configurations].\n --block-provider Valid BlockProviderConfig interface [default].\n --block Valid BlockConfig interface.\n --block-json Valid BlockJSON interface.\n\"\"\"\n\nimport ConfigParser\nimport sys, os\nimport logging, verboselogs\nimport time, datetime\nimport yaml\nfrom Context import Context\nfrom clicommand import CLICommand\nfrom cdpcli import __version__\nfrom docopt import docopt, DocoptExit\nfrom PropertiesParser import PropertiesParser\n\nLOG = verboselogs.VerboseLogger('clidriver')\nLOG.addHandler(logging.StreamHandler())\nLOG.setLevel(logging.INFO)\n\ndef main():\n opt = docopt(__doc__, sys.argv[1:], version=__version__)\n if opt['--verbose']:\n LOG.setLevel(logging.VERBOSE)\n elif opt['--quiet']:\n LOG.setLevel(logging.WARNING)\n\n driver = CLIDriver(cmd = CLICommand(opt['--dry-run']), opt = opt)\n return driver.main()\n\nclass CLIDriver(object):\n\n def __init__(self, cmd=None, opt=None):\n if cmd is None:\n raise ValueError('TODO')\n else:\n self._cmd = cmd\n\n if opt is None:\n raise ValueError('TODO')\n else:\n self._context = Context(opt, cmd)\n LOG.verbose('Context : %s', self._context.__dict__)\n\n # Default value of DOCKER_HOST env var if not set\n if os.getenv('DOCKER_HOST', None) is None:\n os.environ['DOCKER_HOST'] = 'unix:///var/run/docker.sock'\n\n LOG.verbose('DOCKER_HOST : %s', os.getenv('DOCKER_HOST',''))\n\n def main(self, args=None):\n try:\n if self._context.opt['--verbose']:\n self._cmd.run_command('env')\n\n if self._context.opt['build']:\n self.__build()\n\n if self._context.opt['maven']:\n self.__maven()\n\n if self._context.opt['sonar']:\n self.__sonar()\n\n if self._context.opt['docker']:\n self.__docker()\n\n if self._context.opt['artifactory']:\n self.__artifactory()\n\n if self._context.opt['k8s']:\n self.__k8s()\n\n if self._context.opt['validator']:\n self.__validator()\n\n finally:\n if self._context.opt['--sleep'] != \"0\":\n self._cmd.run_command('sleep %s' % self._context.opt['--sleep'])\n\n\n def __build(self):\n self.__create_ssh_key()\n self.__simulate_merge_on()\n self._cmd.run_command('docker pull %s' % (self._context.opt['--docker-image']))\n\n command_run_image = 'docker run $(env | grep \"\\(^CI\\|^CDP\\|^AWS\\|^GITLAB\\)\" | cut -f1 -d= | sed \\'s/^/-e /\\') --rm -v /var/run/docker.sock:/var/run/docker.sock -e DOCKER_HOST=unix:///var/run/docker.sock'\n\n if self._context.opt['--volume-from'] == 'k8s':\n command_run_image = '%s --volumes-from $(docker ps -aqf \"name=k8s_build_${HOSTNAME}\")' % command_run_image\n else:\n command_run_image = '%s --volumes-from $(docker ps -aqf \"name=${HOSTNAME}-build\")' % command_run_image\n\n command = self._context.opt['--command']\n\n command_run_image = '%s -w ${PWD}' % command_run_image\n command_run_image = '%s %s /bin/sh -c \\'%s\\'' % (command_run_image, self._context.opt['--docker-image'], command)\n\n self._cmd.run_command(command_run_image)\n\n def __maven(self):\n self.__create_ssh_key()\n self.__simulate_merge_on()\n self._cmd.run_command('docker pull maven:%s' % (self._context.opt['--docker-version']))\n\n settings = 'maven-settings.xml'\n\n self._cmd.run_command('cp /cdp/maven/settings.xml %s' % settings)\n\n command_run_image = 'docker run $(env | grep \"\\(^CI\\|^CDP\\|^AWS\\|^GITLAB\\)\" | cut -f1 -d= | sed \\'s/^/-e /\\') --rm -v /var/run/docker.sock:/var/run/docker.sock -e DOCKER_HOST=unix:///var/run/docker.sock'\n\n if self._context.opt['--volume-from'] == 'k8s':\n command_run_image = '%s --volumes-from $(docker ps -aqf \"name=k8s_build_${HOSTNAME}\")' % command_run_image\n else:\n command_run_image = '%s --volumes-from $(docker ps -aqf \"name=${HOSTNAME}-build\")' % command_run_image\n\n command = self._context.opt['--goals']\n\n if self._context.opt['--deploy']:\n if self._context.opt['--deploy'] == 'release':\n command = '--batch-mode org.apache.maven.plugins:maven-release-plugin:%s:prepare org.apache.maven.plugins:maven-release-plugin:%s:perform -Dresume=false -DautoVersionSubmodules=true -DdryRun=false -DscmCommentPrefix=\"[ci skip]\"' % (self._context.opt['--maven-release-plugin'], self._context.opt['--maven-release-plugin'])\n arguments = '-DskipTest -DskipITs -DaltDeploymentRepository=release::default::%s/%s' % (os.environ['CDP_REPOSITORY_URL'], os.environ['CDP_REPOSITORY_MAVEN_RELEASE'])\n\n if os.getenv('MAVEN_OPTS', None) is not None:\n arguments = '%s %s' % (arguments, os.environ['MAVEN_OPTS'])\n\n command = '%s -DreleaseProfiles=release -Darguments=\"%s\"' % (command, arguments)\n else:\n command = 'deploy -DskipTest -DskipITs -DaltDeploymentRepository=snapshot::default::%s/%s' % (os.environ['CDP_REPOSITORY_URL'], os.environ['CDP_REPOSITORY_MAVEN_SNAPSHOT'])\n\n\n if os.getenv('MAVEN_OPTS', None) is not None:\n command = '%s %s' % (command, os.environ['MAVEN_OPTS'])\n\n command = 'mvn %s %s' % (command, '-s %s' % settings)\n\n if os.getenv('CDP_SSH_PRIVATE_KEY', None) is not None:\n command = 'mkdir -p ~/.ssh && echo \"$CDP_SSH_PRIVATE_KEY\" | tr -d \\'\\r\\' > ~/.ssh/id_rsa && chmod 600 ~/.ssh/id_rsa && %s' % (command)\n\n command_run_image = '%s -w ${PWD}' % command_run_image\n command_run_image = '%s maven:%s /bin/sh -c \\'%s\\'' % (command_run_image, self._context.opt['--docker-version'], command)\n\n self._cmd.run_command(command_run_image)\n\n\n def __sonar(self):\n self.__simulate_merge_on()\n\n sonar_file = 'sonar-project.properties'\n project_key = None\n sources = None\n\n command = 'sonar-scanner -Dsonar.login=%s -Dsonar.host.url=%s -Dsonar.gitlab.user_token=%s -Dsonar.gitlab.commit_sha=%s -Dsonar.gitlab.ref_name=%s -Dsonar.gitlab.project_id=%s -Dsonar.branch.name=%s' % (os.environ['CDP_SONAR_LOGIN'],\n os.environ['CDP_SONAR_URL'], os.environ['GITLAB_USER_TOKEN'], os.environ['CI_COMMIT_SHA'], os.environ['CI_COMMIT_REF_NAME'], os.environ['CI_PROJECT_PATH'], self.__getTagBranchName())\n\n # Check if mandatory properties are setted\n if os.path.isfile(sonar_file):\n LOG.verbose('Read : %s', sonar_file)\n cfg = PropertiesParser()\n cfg.read(sonar_file)\n project_key = cfg.get('sonar.projectKey')\n sources = cfg.get('sonar.sources')\n\n # Set property if not setted\n if not (project_key and project_key.strip()):\n command = \"%s -Dsonar.projectKey=%s\" % (command, os.environ['CI_PROJECT_PATH'].replace('/', '_'))\n\n # Set property if not setted\n if not (sources and sources.strip()):\n command = \"%s -Dsonar.sources=.\" % command\n\n if self._context.opt['--sast']:\n command = \"%s -Dsonar.gitlab.json_mode=SAST\" % command\n else:\n command = \"%s -Dsonar.gitlab.json_mode=CODECLIMATE\" % command\n\n if self._context.opt['--preview']:\n command = \"%s -Dsonar.analysis.mode=preview\" % command\n\n self._cmd.run_command(command)\n\n def __docker(self):\n # Login to the docker registry\n self._cmd.run_command(self._context.login)\n\n if self._context.opt['--use-aws-ecr']:\n try:\n self._cmd.run_command('aws ecr list-images --repository-name %s --max-items 0' % (self._context.repository))\n except ValueError:\n LOG.warning('AWS ECR repository doesn\\'t exist. Creating this one.')\n self._cmd.run_command('aws ecr create-repository --repository-name %s' % (self._context.repository))\n\n # Tag and push docker image\n if not (self._context.opt['--image-tag-branch-name'] or self._context.opt['--image-tag-latest'] or self._context.opt['--image-tag-sha1']) or self._context.opt['--image-tag-branch-name']:\n # Default if none option selected\n self.__buildTagAndPushOnDockerRegistry(self.__getTagBranchName())\n if self._context.opt['--image-tag-latest']:\n self.__buildTagAndPushOnDockerRegistry(self.__getTagLatest())\n if self._context.opt['--image-tag-sha1']:\n self.__buildTagAndPushOnDockerRegistry(self.__getTagSha1())\n\n def __artifactory(self):\n if self._context.opt['--put']:\n upload_file = self._context.opt['--put']\n http_verb = 'PUT'\n elif self._context.opt['--delete']:\n upload_file = self._context.opt['--delete']\n http_verb = 'DELETE'\n else:\n raise ValueError('Incorrect option with artifactory command.')\n\n # Tag and push docker image\n if not (self._context.opt['--image-tag-branch-name'] or self._context.opt['--image-tag-latest'] or self._context.opt['--image-tag-sha1']) or self._context.opt['--image-tag-branch-name']:\n # Default if none option selected\n self.__callArtifactoryFile(self.__getTagBranchName(), upload_file, http_verb)\n if self._context.opt['--image-tag-latest']:\n self.__callArtifactoryFile(self.__getTagLatest(), upload_file, http_verb)\n if self._context.opt['--image-tag-sha1']:\n self.__callArtifactoryFile(self.__getTagSha1(), upload_file, http_verb)\n\n def __k8s(self):\n # Need to create default helm charts\n if self._context.opt['--create-default-helm']:\n # Check that the chart dir no exists\n if os.path.isdir('%s/templates' % self._context.opt['--deploy-spec-dir']):\n raise ValueError('Directory %s/templates already exists, while --deploy-spec-dir has been selected.' % self._context.opt['--deploy-spec-dir'])\n elif os.path.isfile('%s/values.yaml' % self._context.opt['--deploy-spec-dir']):\n raise ValueError('File %s/values.yaml already exists, while --deploy-spec-dir has been selected.' % self._context.opt['--deploy-spec-dir'])\n elif os.path.isfile('%s/Chart.yaml' % self._context.opt['--deploy-spec-dir']):\n raise ValueError('File %s/Chart.yaml already exists, while --deploy-spec-dir has been selected.' % self._context.opt['--deploy-spec-dir'])\n else:\n os.makedirs('%s/templates' % self._context.opt['--deploy-spec-dir'])\n self._cmd.run_command('cp -R /cdp/k8s/charts/* %s/' % self._context.opt['--deploy-spec-dir'])\n with open('%s/Chart.yaml' % self._context.opt['--deploy-spec-dir'], 'w') as outfile:\n data = dict(\n apiVersion = 'v1',\n description = 'A Helm chart for Kubernetes',\n name = os.environ['CI_PROJECT_NAME'],\n version = '0.1.0'\n )\n yaml.dump(data, outfile, default_flow_style=False)\n\n namespace = self.__getNamespace()\n host = self.__getHost()\n\n if self._context.opt['--image-tag-latest']:\n tag = self.__getTagLatest()\n elif self._context.opt['--image-tag-sha1']:\n tag = self.__getTagSha1()\n else :\n tag = self.__getTagBranchName()\n\n # Need to add secret file for docker registry\n if not self._context.opt['--use-aws-ecr']:\n # Copy secret file on k8s deploy dir\n self._cmd.run_command('cp /cdp/k8s/secret/cdp-secret.yaml %s/templates/' % self._context.opt['--deploy-spec-dir'])\n secretParams = '--set image.credentials.username=%s --set image.credentials.password=%s' % (self._context.registry_user, self._context.registry_token_ro)\n else:\n secretParams = ''\n\n if self._context.opt['--values']:\n valuesFiles = self._context.opt['--values'].strip().split(',')\n values = '--values %s/' % self._context.opt['--deploy-spec-dir'] + (' --values %s/' % self._context.opt['--deploy-spec-dir']).join(valuesFiles)\n else:\n values = ''\n\n # Instal or Upgrade environnement\n self._cmd.run_command('helm upgrade %s %s --timeout %s --set namespace=%s --set ingress.host=%s --set image.commit.sha=sha-%s --set image.registry=%s --set image.repository=%s --set image.tag=%s %s %s --debug -i --namespace=%s'\n % (namespace, self._context.opt['--deploy-spec-dir'], self._context.opt['--timeout'], namespace, host, os.environ['CI_COMMIT_SHA'][:8], self._context.registry, self._context.repository, tag, secretParams, values, namespace))\n\n if self._context.opt['--delete-labels']:\n now = datetime.datetime.now()\n date_format = '%Y-%m-%dT%H%M%S'\n self._cmd.run_command('kubectl label namespace %s deletable=true creationTimestamp=%s deletionTimestamp=%s --namespace=%s --overwrite'\n % (namespace, now.strftime(date_format), (now + datetime.timedelta(minutes = int(self._context.opt['--delete-labels']))).strftime(date_format) , namespace))\n\n ressources = self._cmd.run_command('kubectl get deployments -n %s -o name' % (namespace))\n if ressources is not None:\n ressources = ressources.strip().split('\\n')\n\n # Patch\n for ressource in ressources:\n if not self._context.opt['--use-aws-ecr']:\n # Patch secret on deployment (Only deployment imagePullSecrets patch is possible. It's forbidden for pods)\n # Forbidden: pod updates may not change fields other than `containers[*].image` or `spec.activeDeadlineSeconds` or `spec.tolerations` (only additions to existing tolerations)\n self._cmd.run_command('kubectl patch %s -p \\'{\"spec\":{\"template\":{\"spec\":{\"imagePullSecrets\": [{\"name\": \"cdp-%s\"}]}}}}\\' -n %s'\n % (ressource.replace('/', ' '), self._context.registry, namespace))\n\n # Rollout\n for ressource in ressources:\n # Issue on --request-timeout option ? https://github.com/kubernetes/kubernetes/issues/51952\n self._cmd.run_command('timeout %s kubectl rollout status %s -n %s' % (self._context.opt['--timeout'], ressource, namespace))\n\n\n def __buildTagAndPushOnDockerRegistry(self, tag):\n if self._context.opt['--use-docker-compose']:\n os.environ['CDP_TAG'] = tag\n os.environ['CDP_REGISTRY'] = self.__getImageName()\n self._cmd.run_command('docker-compose build')\n self._cmd.run_command('docker-compose push')\n else:\n image_tag = self.__getImageTag(self.__getImageName(), tag)\n # Tag docker image\n self._cmd.run_command('docker build -t %s .' % (image_tag))\n # Push docker image\n self._cmd.run_command('docker push %s' % (image_tag))\n\n\n def __callArtifactoryFile(self, tag, upload_file, http_verb):\n if http_verb is 'PUT':\n self._cmd.run_command('curl --fail -X PUT %s/%s/%s/ -H X-JFrog-Art-Api:%s -T %s' % (os.environ['CDP_ARTIFACTORY_PATH'], self._context.repository, tag, os.environ['CDP_ARTIFACTORY_TOKEN'], upload_file))\n elif http_verb is 'DELETE':\n self._cmd.run_command('curl --fail -X DELETE %s/%s/%s/%s -H X-JFrog-Art-Api:%s' % (os.environ['CDP_ARTIFACTORY_PATH'], self._context.repository, tag, upload_file, os.environ['CDP_ARTIFACTORY_TOKEN']))\n\n def __validator(self):\n if self._context.opt['--block']:\n schema = 'BlockConfig'\n elif self._context.opt['--block-json']:\n schema = 'BlockJSON'\n else :\n schema = 'BlockProviderConfig'\n\n url = 'http://%s/%s' % (self.__getHost(), self._context.opt['--path'])\n\n self._cmd.run_command('validator-cli --url %s --schema %s' % (url, schema))\n\n\n def __getImageName(self):\n # Configure docker registry\n image_name = '%s/%s' % (self._context.registry, self._context.repository)\n LOG.verbose('Image name : %s', image_name)\n return image_name\n\n def __getImageTag(self, image_name, tag):\n return '%s:%s' % (image_name, tag)\n\n def __getTagBranchName(self):\n return os.environ['CI_COMMIT_REF_NAME']\n\n def __getTagLatest(self):\n return 'latest'\n\n def __getTagSha1(self):\n return os.environ['CI_COMMIT_SHA']\n\n def __getNamespace(self):\n # Get k8s namespace\n if self._context.opt['--namespace-project-name']:\n namespace = os.environ['CI_PROJECT_NAME']\n else:\n namespace = '%s-%s' % (os.environ['CI_PROJECT_NAME'], os.environ['CI_COMMIT_REF_NAME']) # Get deployment host\n\n return namespace.replace('_', '-')\n\n def __getHost(self):\n # Get k8s namespace\n if self._context.opt['--namespace-project-name']:\n return '%s.%s' % (os.environ['CI_PROJECT_NAME'], os.environ['DNS_SUBDOMAIN'])\n else:\n return '%s.%s.%s' % (os.getenv('CI_ENVIRONMENT_SLUG', os.environ['CI_COMMIT_REF_NAME']), os.environ['CI_PROJECT_NAME'], os.environ['DNS_SUBDOMAIN'])\n\n def __simulate_merge_on(self):\n if self._context.opt['--simulate-merge-on']:\n LOG.notice('Build docker image with the merge current branch on %s branch', self._context.opt['--simulate-merge-on'])\n\n # Merge branch on selected branch\n self._cmd.run_command('git config --global user.email \\\"%s\\\"' % os.environ['GITLAB_USER_EMAIL'])\n self._cmd.run_command('git config --global user.name \\\"%s\\\"' % os.environ['GITLAB_USER_ID'])\n self._cmd.run_command('git checkout %s' % self._context.opt['--simulate-merge-on'])\n self._cmd.run_command('git reset --hard origin/%s' % self._context.opt['--simulate-merge-on'])\n self._cmd.run_command('git merge %s --no-commit --no-ff' % os.environ['CI_COMMIT_SHA'])\n\n # TODO Exception process\n else:\n LOG.notice('Build docker image with the current branch : %s', os.environ['CI_COMMIT_REF_NAME'])\n\n def __create_ssh_key(self):\n if os.getenv('CDP_SSH_PRIVATE_KEY', None) is not None:\n self._cmd.run_command('mkdir -p ~/.ssh && echo \"$CDP_SSH_PRIVATE_KEY\" | tr -d \\'\\r\\' > ~/.ssh/id_rsa && chmod 600 ~/.ssh/id_rsa')\n", "sub_path": "cdpcli/clidriver.py", "file_name": "clidriver.py", "file_ext": "py", "file_size_in_byte": 23824, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "verboselogs.VerboseLogger", "line_number": 93, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 94, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 95, "usage_type": "attribute"}, {"api_name": "docopt.docopt", "line_number": 98, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 98, "usage_type": "attribute"}, {"api_name": "cdpcli.__version__", "line_number": 98, "usage_type": "name"}, {"api_name": "logging.VERBOSE", "line_number": 100, "usage_type": "attribute"}, {"api_name": "logging.WARNING", "line_number": 102, "usage_type": "attribute"}, {"api_name": "clicommand.CLICommand", "line_number": 104, "usage_type": "call"}, {"api_name": "Context.Context", "line_number": 118, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 122, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 123, "usage_type": "attribute"}, {"api_name": "os.getenv", "line_number": 125, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 198, "usage_type": "attribute"}, {"api_name": "os.getenv", "line_number": 200, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 201, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 205, "usage_type": "attribute"}, {"api_name": "os.getenv", "line_number": 208, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 209, "usage_type": "attribute"}, {"api_name": "os.getenv", "line_number": 213, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 229, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 230, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 233, "usage_type": "call"}, {"api_name": "os.path", "line_number": 233, "usage_type": "attribute"}, {"api_name": "PropertiesParser.PropertiesParser", "line_number": 235, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 242, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 301, "usage_type": "call"}, {"api_name": "os.path", "line_number": 301, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 303, "usage_type": "call"}, {"api_name": "os.path", "line_number": 303, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 305, "usage_type": "call"}, {"api_name": "os.path", "line_number": 305, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 308, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 314, "usage_type": "attribute"}, {"api_name": "yaml.dump", "line_number": 317, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 345, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 348, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 348, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 351, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 373, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 374, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 387, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 389, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 414, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 420, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 425, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 427, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 434, "usage_type": "attribute"}, {"api_name": "os.getenv", "line_number": 436, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 436, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 443, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 444, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 447, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 451, "usage_type": "attribute"}, {"api_name": "os.getenv", "line_number": 454, "usage_type": "call"}]} +{"seq_id": "227960758", "text": "'''Cross process log sink\n\nThe multiprocess executor processses execution steps in child processes and recieves back the system\nevents. Dagstermill notebooks are executed in a subprocess (opened by jupyter_client.manager.KernelManager).\n\nGiven these setups, we need a way to ship logs from the subprocess back to the parent process\n(executing the solid compute logic), but since we don't manage the subprocess directly in some cases, we can't use a\nQueue/QueueHandler-based scheme. (Socket-based schemes might work, but getting these to play nicely with\nDocker and the wide range of possible user networking setups sounds like a headache.)\n\nThere are a variety of persistent file-based queue projects in the Python ecosystem, but they are\nwritten for multithreaded cases, not multiprocessing cases; Celery is too heavy for our use\ncase; and obviously we want to avoid implementing our own locking logic (perhaps on top of\nfilelock).\n\nInstead we use a sqlite database as a queue. The elements of this setup are exhibited cleanly\nin dagster_tests/core/test_xproc_log_sink.py\n\n1. Parent process opens a NamedTemporaryFile for use as a sqlite db.\n2. Parent process initializes the db table by calling init_db. We rely on a\n sqlite autoincrement primary key to preserve the order of log messages.\n3. Parent process starts a watcher thread that spins up a JsonSqlite3LogWatcher, which polls the\n sqlite db table for new logs until its is_done flag (a threading.Event) is set. New log messages\n are rehydrated using logging.makeLogRecord and handed directly to the logging.Handlers configured\n on the pipeline DagsterLogManager.\n4. Child process is started to execute.\n5. Child process creates a DagsterLogManager with a single-handler (JsonSqlite3Handler) logger.\n This handler writes the internal dict representation of logging.LogRecords generated by the\n dagster logging machinery to the sqlite table as json.\n6. Child process returns. Parent proess sets is_done flag.\n7. Parent process waits for watcher process to complete logging.\n\n'''\nimport copy\nimport json\nimport logging\nimport sqlite3\nimport sys\nimport threading\nimport time\n\nfrom dagster import check, seven\nfrom dagster.core.log_manager import DagsterLogManager\n\nCREATE_LOG_TABLE_STATEMENT = '''create table if not exists logs (\n timestamp integer primary key asc,\n json_log text\n)\n'''\n\nINSERT_LOG_RECORD_STATEMENT = '''insert into logs (json_log) values (\n '{json_log}'\n)\n'''\n\nRETRIEVE_LOG_RECORDS_STATEMENT = '''select * from logs\n where timestamp >= {timestamp}\n order by timestamp asc\n'''\n\nif sys.version_info < (3,): # pragma: nocover\n EVENT_TYPE = threading._Event # pylint: disable=no-member, protected-access\nelse:\n EVENT_TYPE = threading.Event\n\n\ndef init_db(sqlite_db_path):\n with sqlite3.connect(sqlite_db_path) as con:\n con.execute(CREATE_LOG_TABLE_STATEMENT)\n con.close()\n\n\nclass JsonSqlite3Handler(logging.Handler):\n def __init__(self, sqlite_db_path, log_msg_only=False):\n check.str_param(sqlite_db_path, 'sqlite_db_path')\n\n self.sqlite_db_path = sqlite_db_path\n self.log_msg_only = log_msg_only\n\n super(JsonSqlite3Handler, self).__init__()\n\n def connect(self):\n return sqlite3.connect(self.sqlite_db_path)\n\n def emit(self, record):\n try:\n log_dict = copy.copy(record.__dict__)\n\n if self.log_msg_only and log_dict.get('dagster_meta', {}).get('dagster_event'):\n return\n\n # while it may seem reasonable hold this connection open, that caused a\n # difficult to debug race-condition-esque problem on py2 where the connect call\n # would sporadically lock up despite setting the timeout argument.\n with self.connect() as con:\n con.execute(INSERT_LOG_RECORD_STATEMENT.format(json_log=seven.json.dumps(log_dict)))\n con.close()\n\n except Exception as e: # pylint: disable=W0703\n logging.critical('Error during logging!')\n logging.exception(str(e))\n\n\nclass JsonSqlite3LogWatcher(object):\n def __init__(self, sqlite_db_path, log_manager, is_done):\n check.str_param(sqlite_db_path, 'sqlite_db_path')\n check.inst_param(log_manager, 'log_manager', DagsterLogManager)\n check.inst_param(is_done, 'is_done', EVENT_TYPE)\n\n self.sqlite_db_path = sqlite_db_path\n self.next_timestamp = 0\n self.log_manager = log_manager\n self.is_done = is_done\n\n def connect(self):\n return sqlite3.connect(self.sqlite_db_path)\n\n def watch(self):\n last_pass = False\n while True:\n with self.connect() as conn:\n res = (\n conn.cursor()\n .execute(RETRIEVE_LOG_RECORDS_STATEMENT.format(timestamp=self.next_timestamp))\n .fetchall()\n )\n if res:\n self.next_timestamp = res[-1][0] + 1\n json_records = [r[1] for r in res]\n for json_record in json_records:\n record = logging.makeLogRecord(json.loads(json_record))\n\n for logger in self.log_manager.loggers:\n for handler in logger.handlers:\n # Because we're rehydrating the LogMessage, rather than passing\n # through Logger._log again (which would obscure the original metadata)\n # we need to filter for log level here\n if handler.level <= record.levelno:\n handler.handle(record)\n conn.close()\n time.sleep(0.5) # 500 ms\n if last_pass:\n break\n if self.is_done.is_set():\n last_pass = True\n\n\ndef construct_sqlite_logger(sqlite_db_path, log_msg_only=False):\n logger = logging.Logger('xproc_sqlite')\n logger.addHandler(JsonSqlite3Handler(sqlite_db_path, log_msg_only))\n logger.setLevel(10)\n return logger\n", "sub_path": "python_modules/dagster/dagster/loggers/xproc_log_sink.py", "file_name": "xproc_log_sink.py", "file_ext": "py", "file_size_in_byte": 6072, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "sys.version_info", "line_number": 61, "usage_type": "attribute"}, {"api_name": "threading._Event", "line_number": 62, "usage_type": "attribute"}, {"api_name": "threading.Event", "line_number": 64, "usage_type": "attribute"}, {"api_name": "sqlite3.connect", "line_number": 68, "usage_type": "call"}, {"api_name": "logging.Handler", "line_number": 73, "usage_type": "attribute"}, {"api_name": "dagster.check.str_param", "line_number": 75, "usage_type": "call"}, {"api_name": "dagster.check", "line_number": 75, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 83, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 87, "usage_type": "call"}, {"api_name": "dagster.seven.json.dumps", "line_number": 96, "usage_type": "call"}, {"api_name": "dagster.seven.json", "line_number": 96, "usage_type": "attribute"}, {"api_name": "dagster.seven", "line_number": 96, "usage_type": "name"}, {"api_name": "logging.critical", "line_number": 100, "usage_type": "call"}, {"api_name": "logging.exception", "line_number": 101, "usage_type": "call"}, {"api_name": "dagster.check.str_param", "line_number": 106, "usage_type": "call"}, {"api_name": "dagster.check", "line_number": 106, "usage_type": "name"}, {"api_name": "dagster.check.inst_param", "line_number": 107, "usage_type": "call"}, {"api_name": "dagster.core.log_manager.DagsterLogManager", "line_number": 107, "usage_type": "argument"}, {"api_name": "dagster.check", "line_number": 107, "usage_type": "name"}, {"api_name": "dagster.check.inst_param", "line_number": 108, "usage_type": "call"}, {"api_name": "dagster.check", "line_number": 108, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 116, "usage_type": "call"}, {"api_name": "logging.makeLogRecord", "line_number": 131, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 131, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 141, "usage_type": "call"}, {"api_name": "logging.Logger", "line_number": 149, "usage_type": "call"}]} +{"seq_id": "376951127", "text": "#! /usr/bin/env python3\n\nimport fire\nimport sys, os\nimport yaml\nimport pathlib\nimport jsonpath_ng\nimport copy\nfrom collections import defaultdict\n\nimport logging\nlogging.getLogger().setLevel(logging.INFO)\n\nimport subprocess\n\nimport k8s_quantity\n\nARTIFACT_DIR = pathlib.Path(os.environ.get(\"ARTIFACT_DIR\", \".\"))\n\ndef run(command, capture_stdout=False, capture_stderr=False, check=True, protect_shell=True, cwd=None):\n logging.info(f\"run: {command}\")\n args = {}\n\n args[\"cwd\"] = cwd\n if capture_stdout: args[\"stdout\"] = subprocess.PIPE\n if capture_stderr: args[\"stderr\"] = subprocess.PIPE\n if check: args[\"check\"] = True\n\n if protect_shell:\n command = f\"set -o errexit;set -o pipefail;set -o nounset;set -o errtrace;{command}\"\n\n proc = subprocess.run(command, shell=True, **args)\n\n if capture_stdout: proc.stdout = proc.stdout.decode(\"utf8\")\n if capture_stderr: proc.stderr = proc.stderr.decode(\"utf8\")\n\n return proc\n\n\nwith open(pathlib.Path(__file__).parent / \"base_appwrapper.yaml\") as f:\n base_appwrapper = yaml.safe_load(f)\n\nwith open(pathlib.Path(__file__).parent / \"config.yaml\") as f:\n main_config = yaml.safe_load(f)\n\ndef get_config(jsonpath, config=main_config):\n return jsonpath_ng.parse(jsonpath).find(config)[0].value\n\ndef set_config(config, jsonpath, value):\n get_config(jsonpath, config=config) # will raise an exception if the jsonpath does not exist\n jsonpath_ng.parse(jsonpath).update(config, value)\n\ndef main(namespace=None, dry_run=True, job_mode=True):\n if namespace is None:\n logging.info(\"Getting the current project name ...\")\n\n namespace = run(\"oc project --short\", capture_stdout=True).stdout.strip() \\\n if not dry_run else \"\"\n\n logging.info(f\"Using namespace '{namespace}' to deploy the workload.\")\n\n if job_mode:\n logging.info(\"Running in Job mode\")\n\n base_name = get_config(\"base_name\")\n\n set_config(base_appwrapper, \"metadata.name\", base_name)\n set_config(base_appwrapper, \"metadata.namespace\", namespace)\n set_config(base_appwrapper, \"spec.priority\", get_config(\"defaults.priority\"))\n\n job_template = get_config(\"job_template\")\n\n summary = []\n total_aw_count = 0\n total_aws_configs = len(get_config(\"aw_resources\"))\n for aw_index, aw_resource in enumerate(get_config(\"aw_resources\")):\n aw_count = get_config(\"count\", aw_resource)\n aw_pod_count = get_config(\"pods\", aw_resource)\n aw_pod_runtime = get_config(\"runtime\", aw_resource)\n aw_pod_resources = get_config(\"resources\", aw_resource)\n\n all_aw_total_resources = {}\n aw_total_resources = {}\n summary_aw = []\n summary_all_aw = []\n\n total_aw_count += aw_count\n\n for res_name, res_quantity in aw_pod_resources.items():\n aw_total = float(k8s_quantity.parse_quantity(res_quantity)) * aw_pod_count\n aw_total_resources[res_name] = aw_total\n all_aw_total_resources[res_name] = aw_total * aw_count\n aw_pod_resources[res_name] = str(res_quantity)\n\n # https://stackoverflow.com/questions/1094841/get-human-readable-version-of-file-size\n def sizeof_fmt(num, suffix=\"\"):\n for unit in [\"\", \"Ki\", \"Mi\", \"Gi\", \"Ti\", \"Pi\", \"Ei\", \"Zi\"]:\n if abs(num) < 1024.0:\n return f\"{num:3.1f}{unit}{suffix}\"\n num /= 1024.0\n return f\"{num:.1f}Yi{suffix}\"\n\n summary += [f\"\"\"\n# AppWrapper #{aw_index}:\n# - {aw_count} AppWrapper resource{'s' if aw_count > 1 else ''}, each creating:\n# - {aw_pod_count} Pod{'s' if aw_pod_count > 1 else ''}\n# - running for {aw_pod_runtime} seconds\n# - with {', '.join(f'{k}:{v}' for k, v in aw_pod_resources.items())} per pod (x{aw_pod_count})\n# - with {', '.join(f'{k}:{sizeof_fmt(v)}' for k, v in aw_total_resources.items())} per AppWrapper (x{aw_count})\n# - with {', '.join(f'{k}:{sizeof_fmt(v)}' for k, v in all_aw_total_resources.items())} for all these AppWrappers\n\"\"\"]\n\n for aw_count_index in range(aw_count):\n appwrapper = copy.deepcopy(base_appwrapper)\n appwrapper_name = f\"aw{aw_index:03d}-{aw_count_index:03d}-{aw_pod_runtime}s\"\n\n set_config(appwrapper, \"metadata.name\", appwrapper_name)\n\n job = copy.deepcopy(job_template)\n job_name = f\"{appwrapper_name}-job\"\n set_config(job, \"metadata.name\", job_name)\n set_config(job, \"metadata.namespace\", namespace)\n set_config(job, \"spec.template.spec.containers[0].env[0].value\", str(aw_pod_runtime))\n set_config(job, \"spec.template.spec.containers[0].resources.limits\", copy.deepcopy(aw_pod_resources))\n set_config(job, \"spec.template.spec.containers[0].resources.requests\", copy.deepcopy(aw_pod_resources))\n\n aw_genericitems = [dict(\n replicas = aw_pod_count,\n completionstatus = \"Complete\",\n custompodresources=[dict(\n replicas=1,\n requests=copy.deepcopy(aw_pod_resources),\n limits=copy.deepcopy(aw_pod_resources),\n )],\n generictemplate = job,\n )]\n set_config(appwrapper, \"spec.resources.GenericItems\", aw_genericitems)\n\n print(f\"\"\"---\n# AppWrapper config #{aw_index}, replica #{aw_count_index}:\n# - {aw_pod_count} Pods with {aw_pod_resources}\n# - running for {aw_pod_runtime} seconds\n---\n\"\"\")\n\n if job_mode:\n src_file = ARTIFACT_DIR / f\"job_{appwrapper_name}.yaml\"\n with open(src_file, \"w\") as f:\n for item in appwrapper[\"spec\"][\"resources\"][\"GenericItems\"]:\n replica = item[\"replicas\"] # currently ignored\n job = item[\"generictemplate\"]\n\n yaml.dump(job, f)\n print(\"---\", file=f)\n else:\n src_file = ARTIFACT_DIR / f\"{appwrapper_name}.yaml\"\n with open(src_file, \"w\") as f:\n yaml.dump(appwrapper, f)\n\n command = f\"oc apply -f {src_file}\"\n if dry_run:\n logging.info(f\"DRY_RUN: {command}\")\n else:\n run(command)\n\n print(f\"---\\n# Summary: {total_aw_count} AppWrappers over {total_aws_configs} configurations\")\n print(\"\\n\".join(summary))\n\n\nif __name__ == \"__main__\":\n try:\n # Print help rather than opening a pager\n fire.core.Display = lambda lines, out: print(*lines, file=out)\n\n fire.Fire(main)\n except subprocess.CalledProcessError as e:\n logging.error(f\"Command '{e.cmd}' failed --> {e.returncode}\")\n sys.exit(1)\n except KeyboardInterrupt:\n print() # empty line after ^C\n logging.error(f\"Interrupted.\")\n sys.exit(1)\n", "sub_path": "subprojects/mcad-workload-generator/generator.py", "file_name": "generator.py", "file_ext": "py", "file_size_in_byte": 6831, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "logging.getLogger", "line_number": 12, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 18, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 18, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 18, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 21, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 25, "usage_type": "attribute"}, {"api_name": "subprocess.PIPE", "line_number": 26, "usage_type": "attribute"}, {"api_name": "subprocess.run", "line_number": 32, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 40, "usage_type": "call"}, {"api_name": "yaml.safe_load", "line_number": 41, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 43, "usage_type": "call"}, {"api_name": "yaml.safe_load", "line_number": 44, "usage_type": "call"}, {"api_name": "jsonpath_ng.parse", "line_number": 47, "usage_type": "call"}, {"api_name": "jsonpath_ng.parse", "line_number": 51, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 55, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 60, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 63, "usage_type": "call"}, {"api_name": "k8s_quantity.parse_quantity", "line_number": 90, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 114, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 119, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 124, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 125, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 132, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 133, "usage_type": "call"}, {"api_name": "yaml.dump", "line_number": 153, "usage_type": "call"}, {"api_name": "yaml.dump", "line_number": 158, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 162, "usage_type": "call"}, {"api_name": "fire.core", "line_number": 173, "usage_type": "attribute"}, {"api_name": "fire.Fire", "line_number": 175, "usage_type": "call"}, {"api_name": "subprocess.CalledProcessError", "line_number": 176, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 177, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 178, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 181, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 182, "usage_type": "call"}]} +{"seq_id": "279035468", "text": "#!/usr/bin/env python\n\nfrom __future__ import unicode_literals\nfrom flask import current_app, request, jsonify\nimport xmltodict\n\nfrom . import api_1_0\nfrom .errors import bad_request\nfrom .authentication import auth, verify_password\n\nfrom .smstools import *\n\n@api_1_0.route('/monitoring', methods=['GET'])\ndef monitoring_view():\n return jsonify({'monitoring': 'ok'})\n\n@api_1_0.route('/sms//', methods=['GET'])\n@auth.login_required\ndef list_some_sms(kind):\n return list_some_sms(kind)\n\n@api_1_0.route('/sms//', methods=['GET'])\n@auth.login_required\ndef get_some_sms_view(kind, message_id):\n return get_some_sms(kind, message_id)\n\n@api_1_0.route('/sms//', methods=['DELETE'])\n@auth.login_required\ndef delete_sms_view(kind, message_id):\n return delete_some_sms(kind, message_id)\n\n@api_1_0.route('/sms/outgoing', methods=['GET', 'POST'])\n@auth.login_required\ndef outgoing_view():\n required_fields = ( 'mobiles', 'text' )\n\n if request.method == 'POST':\n request_object = request.json\n elif request.method == 'GET':\n request_object = {}\n mobiles = request.args.get('mobiles')\n text = request.args.get('text')\n if mobiles:\n request_object['mobiles'] = mobiles.replace(' ', '+').split(',')\n if text:\n request_object['text'] = text\n\n # Check input data\n if type(request_object) is not dict:\n return bad_request('Wrong JSON object')\n for required_field in required_fields:\n if required_field not in request_object:\n return bad_request('Missing required: {0}'.format(required_field))\n if type(request_object['mobiles']) is not list:\n return bad_request('mobiles is not array')\n if len(request_object['mobiles']) == 0:\n return bad_request('mobiles array is empty')\n\n try:\n unicode_str = unicode()\n except NameError:\n unicode_str = str()\n\n for mobile in request_object['mobiles']:\n if type(mobile) is not type(unicode_str):\n return bad_request('mobiles is not unicode')\n\n if type(request_object['text']) is not type(unicode_str):\n return bad_request('text is not unicode')\n\n queue = request_object.get('queue', current_app.config.get('DEFAULTQUEUE'))\n data = {\n 'mobiles': request_object['mobiles'],\n 'text': request_object['text'],\n 'queue' : queue\n }\n\n result = send_sms(data)\n return jsonify(result)\n\n\ndef response_xml(code, description):\n import xml.etree.cElementTree as ET\n response = ET.Element(\"Response\")\n ET.SubElement(response, \"Code\").text = str(code)\n ET.SubElement(response, \"CodeDescription\").text = description\n return ET.tostring(response, encoding=\"ISO-8859-1\")\n\n# xml_interface\n@api_1_0.route('/', methods=['GET', 'POST'])\ndef outgoing_view_xml():\n\n required_fields = ( 'mobiles', 'text' )\n\n if request.method == 'POST':\n request_object = {}\n #print(request.data)\n try:\n xmldata = xmltodict.parse(request.data)\n except:\n return response_xml(4000, \"ERR - BAD XML\")\n\n #print(xmldata)\n\n if not verify_password(xmldata['Request']['AccountLogin']['#text'], xmldata['Request']['AccountPass']):\n return response_xml(4001, \"ERR - INVALID CREDENTIALS\")\n recipients = xmldata['Request']['Message']['Recipients']['Recipient']\n mobiles = []\n if isinstance(recipients, list):\n for recipient in recipients:\n mobiles.append('+' + recipient['#text'])\n else:\n mobiles.append('+' + recipients['#text'])\n text = xmldata['Request']['Message']['Text']['#text']\n if mobiles:\n request_object['mobiles'] = mobiles\n if text:\n request_object['text'] = bytes.fromhex(text.replace('20AC', '80')).decode('cp1252')\n #print(request_object)\n else:\n return response_xml(4000, \"ERR - BAD XML\")\n\n # Check input data\n for required_field in required_fields:\n if required_field not in request_object:\n return response_xml(4000, 'Missing required: {0}'.format(required_field))\n if len(request_object['mobiles']) == 0:\n return response_xml(4002, 'ERR – INVALID RECIPIENTS')\n\n try:\n unicode_str = unicode()\n except NameError:\n unicode_str = str()\n\n for mobile in request_object['mobiles']:\n if type(mobile) is not type(unicode_str):\n return response_xml(4005, 'mobiles is not unicode')\n\n if type(request_object['text']) is not type(unicode_str):\n return response_xml(4005, 'text is not unicode')\n\n queue = request_object.get('queue', current_app.config.get('DEFAULTQUEUE'))\n data = {\n 'mobiles': request_object['mobiles'],\n 'text': request_object['text'],\n 'queue' : queue\n }\n\n result = send_sms(data)\n ok=0\n ids=[]\n for m in result['mobiles']:\n if result['mobiles'][m]['response'] == 'Ok':\n ok += 1\n ids.append(result['mobiles'][m]['message_id'])\n\n if ok == len(result['mobiles']):\n return response_xml(2001, 'OK - QUEUED')\n\n return response_xml(5000, 'ERR - NOT RECIPIENTS OK')\n", "sub_path": "app/api_1_0/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 5205, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "flask.jsonify", "line_number": 15, "usage_type": "call"}, {"api_name": "authentication.auth.login_required", "line_number": 18, "usage_type": "attribute"}, {"api_name": "authentication.auth", "line_number": 18, "usage_type": "name"}, {"api_name": "authentication.auth.login_required", "line_number": 23, "usage_type": "attribute"}, {"api_name": "authentication.auth", "line_number": 23, "usage_type": "name"}, {"api_name": "authentication.auth.login_required", "line_number": 28, "usage_type": "attribute"}, {"api_name": "authentication.auth", "line_number": 28, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 37, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 37, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 38, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 38, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 39, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 39, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 41, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 41, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 41, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 42, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 42, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 42, "usage_type": "name"}, {"api_name": "errors.bad_request", "line_number": 50, "usage_type": "call"}, {"api_name": "errors.bad_request", "line_number": 53, "usage_type": "call"}, {"api_name": "errors.bad_request", "line_number": 55, "usage_type": "call"}, {"api_name": "errors.bad_request", "line_number": 57, "usage_type": "call"}, {"api_name": "errors.bad_request", "line_number": 66, "usage_type": "call"}, {"api_name": "errors.bad_request", "line_number": 69, "usage_type": "call"}, {"api_name": "flask.current_app.config.get", "line_number": 71, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 71, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 71, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 79, "usage_type": "call"}, {"api_name": "authentication.auth.login_required", "line_number": 33, "usage_type": "attribute"}, {"api_name": "authentication.auth", "line_number": 33, "usage_type": "name"}, {"api_name": "xml.etree.cElementTree.Element", "line_number": 84, "usage_type": "call"}, {"api_name": "xml.etree.cElementTree", "line_number": 84, "usage_type": "name"}, {"api_name": "xml.etree.cElementTree.SubElement", "line_number": 85, "usage_type": "call"}, {"api_name": "xml.etree.cElementTree", "line_number": 85, "usage_type": "name"}, {"api_name": "xml.etree.cElementTree.SubElement", "line_number": 86, "usage_type": "call"}, {"api_name": "xml.etree.cElementTree", "line_number": 86, "usage_type": "name"}, {"api_name": "xml.etree.cElementTree.tostring", "line_number": 87, "usage_type": "call"}, {"api_name": "xml.etree.cElementTree", "line_number": 87, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 95, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 95, "usage_type": "name"}, {"api_name": "xmltodict.parse", "line_number": 99, "usage_type": "call"}, {"api_name": "flask.request.data", "line_number": 99, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 99, "usage_type": "name"}, {"api_name": "authentication.verify_password", "line_number": 105, "usage_type": "call"}, {"api_name": "flask.current_app.config.get", "line_number": 142, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 142, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 142, "usage_type": "name"}]} +{"seq_id": "428501682", "text": "# Copyright 2016 Twitter. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n#!/usr/bin/env python2.7\n''' main.py '''\nimport logging\nimport tornado.ioloop\nimport tornado.web\n\nfrom tornado.httpclient import AsyncHTTPClient\nfrom tornado.options import define, options, parse_command_line\n\nfrom heron.shell.src.python import handlers\n\ndefault_handlers = [\n (r\"^/jmap/([0-9]+$)\", handlers.JmapHandler),\n (r\"^/histo/([0-9]+$)\", handlers.MemoryHistogramHandler),\n (r\"^/pmap/([0-9]+$)\", handlers.PmapHandler),\n (r\"^/jstack/([0-9]+$)\", handlers.JstackHandler),\n (r\"^/pid/(.*)\", handlers.PidHandler),\n (r\"^/browse/(.*)\", handlers.BrowseHandler),\n (r\"^/file/(.*)\", handlers.FileHandler),\n (r\"^/filedata/(.*)\", handlers.FileDataHandler),\n (r\"^/filestats/(.*)\", handlers.FileStatsHandler),\n (r\"^/download/(.*)\", handlers.DownloadHandler),\n (r\"^/killexecutor\", handlers.KillExecutorHandler),\n]\n\n# pylint: disable=dangerous-default-value\ndef run(url_to_handlers=default_handlers):\n define(\"port\", default=9999, help=\"Runs on the given port\", type=int)\n define(\"secret\", default='', help=\"Shared secret for /killexecutor\", type=str)\n parse_command_line()\n\n logger = logging.getLogger(__file__)\n logger.info(\"Starting Heron Shell\")\n logger.info(\"Shared secret for /killexecutor: %s\", options.secret)\n\n AsyncHTTPClient.configure(None, defaults=dict(request_timeout=120.0))\n app = tornado.web.Application(url_to_handlers)\n app.listen(options.port)\n tornado.ioloop.IOLoop.instance().start()\n\nif __name__ == '__main__':\n run()\n", "sub_path": "heron/shell/src/python/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2067, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "heron.shell.src.python.handlers.JmapHandler", "line_number": 27, "usage_type": "attribute"}, {"api_name": "heron.shell.src.python.handlers", "line_number": 27, "usage_type": "name"}, {"api_name": "heron.shell.src.python.handlers.MemoryHistogramHandler", "line_number": 28, "usage_type": "attribute"}, {"api_name": "heron.shell.src.python.handlers", "line_number": 28, "usage_type": "name"}, {"api_name": "heron.shell.src.python.handlers.PmapHandler", "line_number": 29, "usage_type": "attribute"}, {"api_name": "heron.shell.src.python.handlers", "line_number": 29, "usage_type": "name"}, {"api_name": "heron.shell.src.python.handlers.JstackHandler", "line_number": 30, "usage_type": "attribute"}, {"api_name": "heron.shell.src.python.handlers", "line_number": 30, "usage_type": "name"}, {"api_name": "heron.shell.src.python.handlers.PidHandler", "line_number": 31, "usage_type": "attribute"}, {"api_name": "heron.shell.src.python.handlers", "line_number": 31, "usage_type": "name"}, {"api_name": "heron.shell.src.python.handlers.BrowseHandler", "line_number": 32, "usage_type": "attribute"}, {"api_name": "heron.shell.src.python.handlers", "line_number": 32, "usage_type": "name"}, {"api_name": "heron.shell.src.python.handlers.FileHandler", "line_number": 33, "usage_type": "attribute"}, {"api_name": "heron.shell.src.python.handlers", "line_number": 33, "usage_type": "name"}, {"api_name": "heron.shell.src.python.handlers.FileDataHandler", "line_number": 34, "usage_type": "attribute"}, {"api_name": "heron.shell.src.python.handlers", "line_number": 34, "usage_type": "name"}, {"api_name": "heron.shell.src.python.handlers.FileStatsHandler", "line_number": 35, "usage_type": "attribute"}, {"api_name": "heron.shell.src.python.handlers", "line_number": 35, "usage_type": "name"}, {"api_name": "heron.shell.src.python.handlers.DownloadHandler", "line_number": 36, "usage_type": "attribute"}, {"api_name": "heron.shell.src.python.handlers", "line_number": 36, "usage_type": "name"}, {"api_name": "heron.shell.src.python.handlers.KillExecutorHandler", "line_number": 37, "usage_type": "attribute"}, {"api_name": "heron.shell.src.python.handlers", "line_number": 37, "usage_type": "name"}, {"api_name": "tornado.options.define", "line_number": 42, "usage_type": "call"}, {"api_name": "tornado.options.define", "line_number": 43, "usage_type": "call"}, {"api_name": "tornado.options.parse_command_line", "line_number": 44, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 46, "usage_type": "call"}, {"api_name": "tornado.options.options.secret", "line_number": 48, "usage_type": "attribute"}, {"api_name": "tornado.options.options", "line_number": 48, "usage_type": "name"}, {"api_name": "tornado.httpclient.AsyncHTTPClient.configure", "line_number": 50, "usage_type": "call"}, {"api_name": "tornado.httpclient.AsyncHTTPClient", "line_number": 50, "usage_type": "name"}, {"api_name": "tornado.ioloop.web.Application", "line_number": 51, "usage_type": "call"}, {"api_name": "tornado.ioloop.web", "line_number": 51, "usage_type": "attribute"}, {"api_name": "tornado.ioloop", "line_number": 51, "usage_type": "name"}, {"api_name": "tornado.options.options.port", "line_number": 52, "usage_type": "attribute"}, {"api_name": "tornado.options.options", "line_number": 52, "usage_type": "name"}, {"api_name": "tornado.ioloop.ioloop.IOLoop.instance", "line_number": 53, "usage_type": "call"}, {"api_name": "tornado.ioloop.ioloop", "line_number": 53, "usage_type": "attribute"}, {"api_name": "tornado.ioloop", "line_number": 53, "usage_type": "name"}]} +{"seq_id": "125313119", "text": "import apache_beam as beam\r\n\r\nimport config\r\nimport argparse\r\nimport json\r\nfrom apache_beam.options.pipeline_options import PipelineOptions\r\nfrom apache_beam.options.pipeline_options import GoogleCloudOptions\r\nfrom apache_beam.options.pipeline_options import StandardOptions\r\nfrom apache_beam.options.pipeline_options import SetupOptions\r\nfrom apache_beam.io.gcp.internal.clients import bigquery\r\n\r\n\r\noptions = PipelineOptions()\r\n\r\ngoogle_cloud_options = options.view_as(GoogleCloudOptions)\r\ngoogle_cloud_options.project = config.PROJECT_ID\r\ngoogle_cloud_options.staging_location = config.STAGING\r\ngoogle_cloud_options.temp_location = config.TEMP\r\noptions.view_as(StandardOptions).runner = 'DataflowRunner'\r\noptions.view_as(StandardOptions).streaming = True\r\noptions.view_as(SetupOptions)\r\n\r\n\r\n\r\ndef compute_sentiment(line):\r\n import textblob\r\n from textblob import TextBlob\r\n templist = line.split('-=-')\r\n for j, item in enumerate(templist):\r\n templist[j] = item.replace(',', '')\r\n tweet = templist[1]\r\n sent = TextBlob(tweet).sentiment.polarity\r\n templist.append(str(sent))\r\n\r\n diction = dict(zip(['Username', 'Tweet', 'Time', 'Followers', 'Location', 'Source', 'Sentiment'], templist))\r\n\r\n return diction\r\n \r\ndef run(argv=None):\r\n parser = argparse.ArgumentParser()\r\n parser.add_argument('--requirements_file', required=True)\r\n parser.add_argument('--input_topic',\r\n help=('Input PubSub topic of the form '\r\n '\"projects//topics/\".'),required=True)\r\n parser.add_argument(\r\n '--output_table', required=True,\r\n help=\r\n ('Output BigQuery table for results specified as: PROJECT:DATASET.TABLE '))\r\n known_args, pipeline_args=parser.parse_known_args(argv)\r\n\r\n with beam.Pipeline(options=options, argv=pipeline_args) as p:\r\n # Read the pubsub topic into a PCollection.\r\n lines = (p | beam.io.ReadStringsFromPubSub(topic=known_args.input_topic)\r\n | beam.Map(compute_sentiment)\r\n | beam.io.WriteToBigQuery(known_args.output_table,\r\n schema='Username:STRING, Tweet:STRING, Time:TIMESTAMP, Followers:INTEGER, Location:STRING, Source:STRING, Sentiment:FLOAT',\r\n create_disposition=beam.io.BigQueryDisposition.CREATE_IF_NEEDED,\r\n write_disposition=beam.io.BigQueryDisposition.WRITE_APPEND))\r\n\r\nif __name__ == '__main__':\r\n run()\r\n\r\n\r\n", "sub_path": "twitterStreamingtoBigQuery-Code/stream-tweets-dataflow.py", "file_name": "stream-tweets-dataflow.py", "file_ext": "py", "file_size_in_byte": 2441, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "apache_beam.options.pipeline_options.PipelineOptions", "line_number": 13, "usage_type": "call"}, {"api_name": "apache_beam.options.pipeline_options.GoogleCloudOptions", "line_number": 15, "usage_type": "argument"}, {"api_name": "config.PROJECT_ID", "line_number": 16, "usage_type": "attribute"}, {"api_name": "config.STAGING", "line_number": 17, "usage_type": "attribute"}, {"api_name": "config.TEMP", "line_number": 18, "usage_type": "attribute"}, {"api_name": "apache_beam.options.pipeline_options.StandardOptions", "line_number": 19, "usage_type": "argument"}, {"api_name": "apache_beam.options.pipeline_options.StandardOptions", "line_number": 20, "usage_type": "argument"}, {"api_name": "apache_beam.options.pipeline_options.SetupOptions", "line_number": 21, "usage_type": "argument"}, {"api_name": "textblob.TextBlob", "line_number": 32, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 40, "usage_type": "call"}, {"api_name": "apache_beam.Pipeline", "line_number": 51, "usage_type": "call"}, {"api_name": "apache_beam.io.ReadStringsFromPubSub", "line_number": 53, "usage_type": "call"}, {"api_name": "apache_beam.io", "line_number": 53, "usage_type": "attribute"}, {"api_name": "apache_beam.Map", "line_number": 54, "usage_type": "call"}, {"api_name": "apache_beam.io.WriteToBigQuery", "line_number": 55, "usage_type": "call"}, {"api_name": "apache_beam.io", "line_number": 55, "usage_type": "attribute"}, {"api_name": "apache_beam.io", "line_number": 57, "usage_type": "attribute"}, {"api_name": "apache_beam.io", "line_number": 58, "usage_type": "attribute"}]} +{"seq_id": "596259188", "text": "import requests\nfrom bs4 import BeautifulSoup\nfrom selenium import webdriver\nimport time\n\n# headless 사용중 주의할점:\n# 웹크롤링 중 headless 사요시 user agent값이 HeadlessChrome으로 요청하는걸 볼수잇다\n# headless 사요시엔 user agent를 필요시 수정.\n# Chrome: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/94.0.4606.71 Safari/537.36\n# headless: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) HeadlessChrome/94.0.4606.71 Safari/537.36\n\n# headless를 활용하면 브라우저 창 안켜진 상태로 진행가능하다\noptions = webdriver.ChromeOptions()\noptions.headless = True\noptions.add_argument(\"window-size=1920x1080\")\noptions.add_argument(\"User-Agent=Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/94.0.4606.71 Safari/537.36\")\n# options.add_argument('lang=zh_CN.UTF-8') # 크롬 브라우저 언어 설정\n\nbrowser = webdriver.Chrome(\"D:\\Downloads\\coding\\chromedriver.exe\",options=options)\n\nurl = \"https://www.whatismybrowser.com/detect/what-is-my-user-agent\"\nbrowser.get(url)\n\ndetected_value = browser.find_element_by_id(\"detected_value\")\nprint(detected_value.text)\n\nbrowser.quit()", "sub_path": "nadocoding_study/webscraping_basic/18_headless_chrome_useragent.py", "file_name": "18_headless_chrome_useragent.py", "file_ext": "py", "file_size_in_byte": 1246, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "selenium.webdriver.ChromeOptions", "line_number": 13, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 13, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 19, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 19, "usage_type": "name"}]} +{"seq_id": "33912026", "text": "import pymongo\nimport MeCab\n\ndb_name = 'language_analysis'\ntalk_collection_name = 'line_talk'\nconversation_collection_name = 'line_conversation'\nclient = pymongo.MongoClient()\ntalk_collection = client[db_name][talk_collection_name]\nconversation_collection = client[db_name][conversation_collection_name]\nconversation_collection.ensure_index([\n ('conversation', pymongo.ASCENDING)\n])\ntalks = talk_collection.find({})\n\nMINUTE = 60\nCONVERSATION_SECOND_THRESHOLD = 30 * MINUTE\n\nex_talk = talks[0]\nconversation = []\nwakati_conversation = []\nconversations = []\nfor talk in talks:\n talk_second_diff = (talk['created_at'] - ex_talk['created_at']).seconds\n # print(talk_second_diff, talk['content'])\n if talk_second_diff < CONVERSATION_SECOND_THRESHOLD:\n conversation.append(talk['content'])\n wakati_conversation.append(talk['wakati_content'])\n else:\n conversation_collection.update({\n 'conversation': conversation,\n },\n {\n 'conversation': conversation,\n 'wakati_conversation': wakati_conversation,\n }, upsert=True)\n conversations.append(conversation)\n conversation = []\n wakati_conversation = []\n conversation.append(talk['content'])\n wakati_conversation.append(talk['wakati_content'])\n ex_talk = talk\n\n# for i, conversation in enumerate(conversations):\n# if len(conversation) > 1:\n# print(conversation)\n\n", "sub_path": "lang_analysis/extract_conversation.py", "file_name": "extract_conversation.py", "file_ext": "py", "file_size_in_byte": 1437, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pymongo.MongoClient", "line_number": 7, "usage_type": "call"}, {"api_name": "pymongo.ASCENDING", "line_number": 11, "usage_type": "attribute"}]} +{"seq_id": "337952624", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 3.8 (3413)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.linux-x86_64/egg/sovereign/utils/eds.py\n# Compiled at: 2020-04-29 02:35:50\n# Size of source mod 2**32: 3234 bytes\nimport random\nfrom copy import deepcopy\nfrom starlette.exceptions import HTTPException\nfrom sovereign import config\nfrom sovereign.schemas import DiscoveryRequest\nfrom sovereign.utils.templates import resolve\npriority_mapping = config.eds_priority_matrix\ntotal_regions = len(config.regions)\n\ndef _upstream_kwargs(upstream, proxy_region=None, resolve_dns=True, default_region=None, hard_fail=config.dns_hard_fail) -> dict:\n try:\n ip_addresses = resolve(upstream['address']) if resolve_dns else [upstream['address']]\n except HTTPException:\n if hard_fail:\n raise\n ip_addresses = [\n upstream['address']]\n else:\n return {'addrs':ip_addresses, \n 'port':upstream['port'], \n 'region':default_region or upstream.get('region', 'unknown'), \n 'zone':proxy_region}\n\n\ndef total_zones(endpoints: list) -> int:\n \"\"\"\n Returns the true unique number of zones, taking into account\n that multiple endpoints can have the same zone name.\n\n - us-west-1\n - us-west-1 == 2 zones\n - us-east-1\n\n - us-west-1\n - us-west-2 == 3 zones\n - us-east-2\n \"\"\"\n zones = {e['locality']['zone'] for e in endpoints}\n return len(zones)\n\n\ndef locality_lb_endpoints(upstreams, request: DiscoveryRequest=None, resolve_dns=True):\n if request is None:\n proxy_region = None\n else:\n proxy_region = request.node.locality.zone\n kw_args = [_upstream_kwargs(u, proxy_region, resolve_dns) for u in upstreams]\n ret = [lb_endpoints(**kw) for kw in kw_args]\n if total_zones(ret) == 1:\n return ret\n else:\n upstreams_copy = deepcopy(upstreams)\n while True:\n if total_zones(ret) < total_regions:\n region = f\"zone-padding-{total_zones(ret)}\"\n try:\n upstream = upstreams_copy.pop()\n except IndexError:\n random.seed(128)\n upstream = random.choice(upstreams)\n else:\n kw_args = _upstream_kwargs(upstream, proxy_region, resolve_dns, region)\n ret.append(lb_endpoints(**kw_args))\n\n return ret\n\n\ndef lb_endpoints(addrs: list, port: int, region: str, zone: str=None) -> dict:\n \"\"\"\n Creates an envoy endpoint.LbEndpoints proto\n\n :param addrs: The IP addresses or hostname(s) of the upstream.\n :param port: The port that the upstream should be accessed on.\n :param region: The region of the upstream.\n :param zone: The region of the proxy asking for the endpoint configuration.\n \"\"\"\n node_priorities = priority_mapping.get(zone, {})\n priority = node_priorities.get(region, 10)\n return {'priority':priority, \n 'locality':{'zone': region}, \n 'lb_endpoints':[{'endpoint': {'address': {'socket_address': {'address':addr, \n 'port_value':port}}}} for addr in addrs]}", "sub_path": "pycfiles/sovereign-0.7.2-py3.8/eds.cpython-38.py", "file_name": "eds.cpython-38.py", "file_ext": "py", "file_size_in_byte": 3206, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "sovereign.config.eds_priority_matrix", "line_number": 14, "usage_type": "attribute"}, {"api_name": "sovereign.config", "line_number": 14, "usage_type": "name"}, {"api_name": "sovereign.config.regions", "line_number": 15, "usage_type": "attribute"}, {"api_name": "sovereign.config", "line_number": 15, "usage_type": "name"}, {"api_name": "sovereign.config.dns_hard_fail", "line_number": 17, "usage_type": "attribute"}, {"api_name": "sovereign.config", "line_number": 17, "usage_type": "name"}, {"api_name": "sovereign.utils.templates.resolve", "line_number": 19, "usage_type": "call"}, {"api_name": "starlette.exceptions.HTTPException", "line_number": 20, "usage_type": "name"}, {"api_name": "sovereign.schemas.DiscoveryRequest", "line_number": 49, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 59, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 66, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 67, "usage_type": "call"}]} +{"seq_id": "86473252", "text": "# coding=UTF-8\n\n\"\"\"\n=========================================\nRedis数据库类\n=========================================\n\n:Author: glen\n:Date: 2017.9.3\n:Tags: redis database\n:abstract: 连接redis数据库,并进行基本操作。\n\n**类**\n==================\nRedis\n 连接Redis数据库\n\n**使用方法**\n==================\n\n\n**示范代码**\n==================\n::\n\n\"\"\"\n\nimport redis\n\n\nclass Redis:\n \"\"\" 连接Redis数据库\n\n :param str host: 数据库主机,默认是'localhost'\n :param int port: 数据库端口,默认是6379\n :param str password: 数据库登录密码\n :return: 无返回值\n \"\"\"\n def __init__(self, host='106.14.237.43', port=6379, password='z1Yh2900'):\n self._r = redis.Redis(host=host, port=port, password=password)\n\n def __len__(self):\n \"\"\" 返回redis数据库的记��数\n\n :return: 返回redis数据库的记录数\n \"\"\"\n return self._r.dbsize()\n\n def set(self,name,value):\n \"\"\" 添加记录到Redis数据库中\n\n :param str name: 记录的名称\n :param str,tuple,list,set,dict value: 记录的值\n :return: 无返回值\n \"\"\"\n if isinstance(value,str):\n self._r.set(name=name, value=value)\n elif isinstance(value,(tuple,list)):\n for item in value:\n self._r.rpush(name, item)\n elif isinstance(value,set):\n for item in value:\n self._r.sadd(name, item)\n elif isinstance(value,dict):\n for key,kvalue in value.items():\n self._r.hmset(name,{key:kvalue})\n else:\n raise Exception\n\n def get(self,name):\n \"\"\" 获取记录\n\n :param str name: 记录的名称\n :return: 返回记录的值\n \"\"\"\n if bytes.decode(self.type(name)) == 'str':\n return self._r.get(name=name)\n elif bytes.decode(self.type(name)) == 'list':\n return [bytes.decode(item) for item in self._r.lrange(name=name,start=0,end=self._r.llen(name=name))]\n elif bytes.decode(self.type(name)) == 'set':\n return {bytes.decode(item) for item in self._r.smembers(name=name)}\n elif bytes.decode(self.type(name)) == 'hash':\n return {bytes.decode(key):bytes.decode(self._r.hgetall(name)[key]) for key in self._r.hgetall(name)}\n else:\n raise Exception\n\n def clear_all(self):\n \"\"\" 清除Redis数据库所有记录\n\n :return: 无返回值\n \"\"\"\n self._r.flushdb()\n\n def delete(self,name):\n \"\"\" 删除某个记录\n\n :param name:\n :return:\n \"\"\"\n self._r.delete(name)\n\n def type(self,name):\n \"\"\" 返回某个记录的类型\n\n :param str name: 记录的名称\n :return: 返回某个记录的类型\n \"\"\"\n return self._r.type(name=name)\n\nif __name__ == '__main__':\n db = Redis()\n\n core_dbs = db.get('core_database')\n for item in core_dbs:\n print(db.get(item)['label'])\n\n '''\n #db.set('core_database',['ceic'])\n #db.set('scraper_database',['airquality'])\n db.set('ceic',{'label':'中国宏观经济数据平台(CEIC)',\n 'intro':'中国数据库包含超过30万条宏观经济、行业及区域的时间序列数据,CEIC中国数据库已经成为分析中国经济的最佳工具。',\n 'author':'system',\n 'group':'core_database',\n 'source':'CEIC中国数据库',\n 'link':'../core_database/querier_ceic.ipynb?dashboard'})\n db.set('airquality', {'label': '中国城市空气质量日报数据库',\n 'intro': '中国城市空气质量日报数据包含2014年1月1日以来每天中国城市空气质量日报的数据。',\n 'author': 'admin',\n 'group': 'scraper_database',\n 'source':'中华人民共和国环境保护部数据中心',\n 'link':'../scraper_database/querier_airquality.ipynb?dashboard'})'''\n\n '''\n print(len(db))\n db.clear_all()\n\n db.set('car',['hello','world'])\n print(db.get('car'))\n\n db.set('check',{'not','good'})\n print(db.get('check'))\n\n db.set('mdic',{'name':'Alice','age':24})\n print(db.get('mdic'))'''\n\n\n", "sub_path": "lib/base/database/class_redis.py", "file_name": "class_redis.py", "file_ext": "py", "file_size_in_byte": 4332, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "redis.Redis", "line_number": 40, "usage_type": "call"}]} +{"seq_id": "255519810", "text": "import nltk\nfrom nltk.tokenize.punkt import PunktSentenceTokenizer, PunktParameters\nimport re\n\nvocabulary = set()\n\ndef outputVocabulary():\n f = open('holmes.vocab', 'w')\n for item in vocabulary:\n f.write(item)\n f.write(\"\\n\")\n\ndef getVocabulary(file):\n f = open(file, 'r')\n for line in f:\n line = re.sub('\\[', '', line)\n line = re.sub('\\]', '', line)\n line = re.sub('Mr\\.', 'Mr', line)\n line = re.sub('Mrs\\.', 'Mrs', line)\n line = re.sub('Dr\\.', 'Dr', line)\n line = re.sub(',', '', line)\n tokens = nltk.word_tokenize(line)\n removed_item = tokens.pop(0)\n for t in tokens:\n vocabulary.add(t.lower())\n \n\ndef main():\n getVocabulary('dev_questions.txt')\n getVocabulary('test/test_questions.txt')\n outputVocabulary()\n \nmain()\n\n\n\n\n", "sub_path": "obsolete/getVocab.py", "file_name": "getVocab.py", "file_ext": "py", "file_size_in_byte": 838, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "re.sub", "line_number": 16, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 17, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 18, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 19, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 20, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 21, "usage_type": "call"}, {"api_name": "nltk.word_tokenize", "line_number": 22, "usage_type": "call"}]} +{"seq_id": "227902816", "text": "import numpy as np\nimport math\nfrom Helper import *\nfrom sympy.solvers import solve\nfrom sympy import Symbol\n\n# ToDo - remeasure\nROOM_X = 640\nROOM_Y = 480\nROOM_Z = 220\n\nclass Line:\n \"\"\"\"A line class representing the connection between a camera and a\n balloon\"\"\"\n\n def outOfRoom(self, balloon):\n \"\"\"returns True if the approximated point is outside the room and False otherwise\"\"\"\n if balloon[0] > 1.1 * ROOM_X or balloon[0] < -0.1 * ROOM_X:\n return True\n elif balloon[1] > 1.1 * ROOM_Y or balloon[1] < -0.1 * ROOM_Y:\n return True\n elif balloon[2] > 1.1 * ROOM_Z or balloon[2] < -0.1 * ROOM_Z:\n return True\n return False\n\n def __init__(self, start, dir):\n self.start = start\n self.dir = dir\n self.connected = False\n\n def distanceToLine(self, line):\n \"\"\"the function gets a line in space and returns the distance\n between the lines\"\"\"\n\n a = self.findClosest(line)\n b = line.findClosest(self)\n balloon = np.array([(a[0] + b[0]) / 2, (a[1] + b[1]) / 2, (a[2] + b[2]) / 2])\n if self.outOfRoom(balloon):\n return 1000\n else:\n parallel = np.cross(self.dir, line.dir)\n a, b, c = parallel[0], parallel[1], parallel[2]\n d = -a * line.start[0] - b * line.start[1] - c * line.start[2]\n x, y, z = self.start[0], self.start[1], self.start[2]\n dist = abs(a * x + b * y + c * z + d) / math.sqrt(a * a + b * b\n + c * c)\n return dist\n\n def __eq__(self, other):\n a = self.start[0] == other.start[0]\n b = self.start[1] == other.start[1]\n c = self.start[2] == other.start[2]\n d = self.dir[0] == other.dir[0]\n e = self.dir[1] == other.dir[1]\n f = self.dir[2] == other.dir[2]\n return a and b and c and d and e and f\n\n def getClosestLine(self, lines, forbidden):\n \"\"\"given a list of line, finds the one whose closest to our line\"\"\"\n new = []\n for line in lines:\n new.append(line)\n lines = new\n for line in forbidden:\n for l in lines:\n if line.__eq__(l):\n lines.remove(line)\n if len(lines) == 0:\n return None\n line = lines[0]\n dist = self.distanceToLine(line)\n for l in lines:\n if self.distanceToLine(l) < dist:\n line = l\n dist = self.distanceToLine(l)\n return line\n\n def distanceToPoint(self, point):\n \"\"\"the function gets a point in space and returns the distance\n between the line to the point\"\"\"\n\n t = Symbol('t')\n vec = self.start + t * self.dir\n t = solve(vec.dot(point - vec), t)\n t = t[0]\n vec = self.start + t * self.dir\n return np.linalg.norm(point - vec)\n\n def intersect(self, l):\n \"\"\"the function gets a line and returns the point in space who'se\n the sum of it's distances from both of the lines is minimal\"\"\"\n\n a = self.findClosest(l)\n b = l.findClosest(self)\n c = (a[0] - b[0]) ** 2 + (a[1] - b[1]) ** 2 + (a[2] - b[2]) ** 2\n r = np.array([(a[0] + b[0]) / 2, (a[1] + b[1]) / 2, (a[2] + b[2]) / 2])\n return r\n\n def findClosest(self, l):\n \"\"\"the function gets a line and finds the point in our line that is\n closest to the given line\"\"\"\n\n parallel = np.cross(self.dir, l.dir)\n ver = np.cross(l.dir, parallel)\n a, b, c = ver[0], ver[1], ver[2]\n d = -a * l.start[0] - b * l.start[1] - c * l.start[2]\n t = Symbol('t')\n x, y, z = self.start[0], self.start[1], self.start[2]\n u, v, w = self.dir[0], self.dir[1], self.dir[2]\n t = solve(a * (x + t * u) + b * (y + t * v) + c * (z + t * w) + d, t)\n t = t[0]\n return np.array([x + t * u, y + t * v, z + t * w])", "sub_path": "CarCodes/Blochs code/Line.py", "file_name": "Line.py", "file_ext": "py", "file_size_in_byte": 3964, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "numpy.array", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.cross", "line_number": 41, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 45, "usage_type": "call"}, {"api_name": "sympy.Symbol", "line_number": 82, "usage_type": "call"}, {"api_name": "sympy.solvers.solve", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 87, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.cross", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.cross", "line_number": 104, "usage_type": "call"}, {"api_name": "sympy.Symbol", "line_number": 107, "usage_type": "call"}, {"api_name": "sympy.solvers.solve", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 112, "usage_type": "call"}]} +{"seq_id": "194791519", "text": "\"\"\" Development of TRENDY formulation of the FeedbackAnalysis, where the\nadvantage of the availability of S1 and S3 datasets is taken.\n\"\"\"\n\nfrom core import FeedbackAnalysis\n\nimport pandas as pd\nimport xarray as xr\nimport numpy as np\nfrom scipy import signal\nfrom statsmodels import api as sm\n\n\ndef deseasonalise(x):\n \"\"\"\n \"\"\"\n\n fs = 12\n fc = 365/667\n\n w = fc / (fs / 2) # Normalize the frequency.\n b, a = signal.butter(5, w, 'low')\n\n return signal.filtfilt(b, a, x)\n\n\ntrendy_duptake['S1']['month']['VISIT'].Earth_Land\n\n\nfrom importlib import reload\nreload(FeedbackAnalysis);\n\ndf = FeedbackAnalysis.TRENDY(co2['month'], temp['month'], trendy_duptake, 'Earth_Land')\ndf.regstats()['S1']['VISIT']\n\n\nDIR = './../../../'\nOUTPUT_DIR = DIR + 'output/'\n\nco2 = {\n \"year\": pd.read_csv(DIR + f\"data/CO2/co2_year.csv\", index_col=[\"Year\"]).CO2,\n \"month\": pd.read_csv(DIR + f\"data/CO2/co2_month.csv\", index_col=[\"Year\", \"Month\"]).CO2\n}\n\ntemp = {\n \"year\": xr.open_dataset(OUTPUT_DIR + f'TEMP/spatial/output_all/HadCRUT/year.nc'),\n \"month\": xr.open_dataset(OUTPUT_DIR + f'TEMP/spatial/output_all/HadCRUT/month.nc')\n}\n\ntrendy_models = ['VISIT', 'OCN', 'JSBACH', 'CLASS-CTEM', 'CABLE-POP']\ntrendy_uptake = {\n \"S1\": {\n \"year\": {model_name : xr.open_dataset(OUTPUT_DIR + f'TRENDY/spatial/output_all/{model_name}_S1_nbp/year.nc')\n for model_name in trendy_models},\n \"month\": {model_name : xr.open_dataset(OUTPUT_DIR + f'TRENDY/spatial/output_all/{model_name}_S1_nbp/month.nc')\n for model_name in trendy_models if model_name != \"LPJ-GUESS\"}\n },\n \"S3\": {\n \"year\": {model_name : xr.open_dataset(OUTPUT_DIR + f'TRENDY/spatial/output_all/{model_name}_S3_nbp/year.nc')\n for model_name in trendy_models},\n \"month\": {model_name : xr.open_dataset(OUTPUT_DIR + f'TRENDY/spatial/output_all/{model_name}_S3_nbp/month.nc')\n for model_name in trendy_models if model_name != \"LPJ-GUESS\"}\n }\n}\n\ntrendy_duptake = {'S1': {'month': {}}, 'S3': {'month': {}}}\nfor sim in trendy_uptake:\n df_models = trendy_uptake[sim]['month']\n for model in trendy_uptake[sim]['month']:\n trendy_duptake[sim]['month'][model] = xr.Dataset(\n {key: (('time'), deseasonalise(df_models[model][key].values)) for\n key in ['Earth_Land', 'South_Land', 'North_Land', 'Tropical_Land']},\n coords={'time': (('time'), df_models[model].time.values)}\n )\n\ntemp_zero = {}\nfor timeres in temp:\n temp_zero[timeres] = xr.Dataset(\n {key: (('time'), np.zeros(len(temp[timeres][key]))) for key in ['Earth', 'South', 'Tropical', 'North']},\n coords={'time': (('time'), temp[timeres].time)}\n )\n\nphi, rho = 0.015 / 2.12, 1.93\n\n\ndef feedback_regression(timeres, variable):\n uptake = trendy_duptake if timeres == \"month\" else trendy_uptake\n start, end = 1960, 2017\n reg_models = {}\n model_stats = {}\n for simulation in ['S1', 'S3']:\n model_names = uptake[simulation][timeres].keys()\n\n sim_reg_models = {}\n sim_model_stats = {\n 'r_squared': [],\n 't_values_beta': [],\n 't_values_gamma': [],\n 'p_values_beta': [],\n 'p_values_gamma': [],\n 'mse_total': [],\n 'nobs': []\n }\n\n for model_name in model_names:\n C = co2[timeres].loc[start:end]\n T = temp[timeres].sel(time=slice(str(start), str(end)))\n if simulation == 'S1':\n U = uptake['S1'][timeres][model_name]\n elif simulation == 'S3':\n U = uptake['S3'][timeres][model_name] - uptake['S1'][timeres][model_name]\n U = U.sel(time=slice(str(start), str(end)))\n\n df = pd.DataFrame(data = {\n \"C\": C,\n \"U\": U[variable],\n \"T\": T['Earth']\n }\n )\n\n if simulation == 'S1':\n x_var = 'C'\n other_var = 'T'\n elif simulation == 'S3':\n x_var = 'T'\n other_var = 'C'\n X = sm.add_constant(df[x_var])\n Y = df[\"U\"]\n reg_model = sm.OLS(Y, X).fit()\n reg_model_params = reg_model.params\n reg_model_params[other_var] = 0\n sim_reg_models[model_name] = (reg_model_params\n .loc[[x_var, other_var]]\n .sort_index()\n )\n sim_model_stats['r_squared'].append(reg_model.rsquared)\n if simulation == 'S1':\n sim_model_stats['t_values_beta'].append(reg_model.tvalues.loc['C'])\n sim_model_stats['p_values_beta'].append(reg_model.pvalues.loc['C'])\n sim_model_stats['t_values_gamma'].append(np.nan)\n sim_model_stats['p_values_gamma'].append(np.nan)\n elif simulation == 'S3':\n sim_model_stats['t_values_beta'].append(np.nan)\n sim_model_stats['p_values_beta'].append(np.nan)\n sim_model_stats['t_values_gamma'].append(reg_model.tvalues.loc['T'])\n sim_model_stats['p_values_gamma'].append(reg_model.pvalues.loc['T'])\n sim_model_stats['mse_total'].append(reg_model.mse_total)\n sim_model_stats['nobs'].append(reg_model.nobs)\n\n df = pd.DataFrame(sim_reg_models)\n df.index = ['beta', 'gamma']\n reg_models[simulation] = df\n reg_models[simulation].loc['beta'] /= 2.12\n reg_models[simulation].loc['u_gamma'] = reg_models[simulation].loc['gamma'] * phi / rho\n\n model_stats[simulation] = pd.DataFrame(sim_model_stats, index=model_names)\n\n return reg_models, model_stats\n\n\nfeedback_regression('month', 'Earth_Land')[0]\n", "sub_path": "scripts/devs/feedbacks/TRENDY_FA.py", "file_name": "TRENDY_FA.py", "file_ext": "py", "file_size_in_byte": 5911, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "scipy.signal.butter", "line_number": 22, "usage_type": "call"}, {"api_name": "scipy.signal", "line_number": 22, "usage_type": "name"}, {"api_name": "scipy.signal.filtfilt", "line_number": 24, "usage_type": "call"}, {"api_name": "scipy.signal", "line_number": 24, "usage_type": "name"}, {"api_name": "importlib.reload", "line_number": 31, "usage_type": "call"}, {"api_name": "core.FeedbackAnalysis", "line_number": 31, "usage_type": "argument"}, {"api_name": "core.FeedbackAnalysis.TRENDY", "line_number": 33, "usage_type": "call"}, {"api_name": "core.FeedbackAnalysis", "line_number": 33, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 41, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 42, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 46, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 47, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 53, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 55, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 59, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 61, "usage_type": "call"}, {"api_name": "xarray.Dataset", "line_number": 70, "usage_type": "call"}, {"api_name": "xarray.Dataset", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 79, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 114, "usage_type": "call"}, {"api_name": "statsmodels.api.add_constant", "line_number": 127, "usage_type": "call"}, {"api_name": "statsmodels.api", "line_number": 127, "usage_type": "name"}, {"api_name": "statsmodels.api.OLS", "line_number": 129, "usage_type": "call"}, {"api_name": "statsmodels.api", "line_number": 129, "usage_type": "name"}, {"api_name": "numpy.nan", "line_number": 140, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 141, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 143, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 144, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 150, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 156, "usage_type": "call"}]} +{"seq_id": "20463082", "text": "# Copyright 2013 - 2016 Mirantis, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\nimport functools\nimport inspect\nfrom time import sleep\n\nfrom devops.error import DevopsException\nfrom devops import logger\n\n\ndef retry(exception, count=10, delay=1):\n \"\"\"Retry decorator\n\n Retries to run decorated method with the same parameters in case of\n thrown :exception:\n\n :type exception: class\n :param exception: exception class\n :type count: int\n :param count: retry count\n :type delay: int\n :param delay: delay between retries in seconds\n :rtype: function\n \"\"\"\n def decorator(func):\n if inspect.ismethod(func):\n full_name = '{}:{}.{}'.format(\n inspect.getmodule(func.im_class).__name__,\n func.im_class.__name__,\n func.__name__)\n elif inspect.isfunction(func):\n full_name = '{}.{}'.format(\n inspect.getmodule(func).__name__,\n func.__name__)\n else:\n raise DevopsException(\n 'Wrong func parameter type {!r}'.format(func))\n\n @functools.wraps(func)\n def wrapper(*args, **kwargs):\n i = 0\n while True:\n try:\n return func(*args, **kwargs)\n except exception as e:\n i += 1\n if i >= count:\n raise\n\n logger.debug(\n 'Exception {!r} while running {!r}. '\n 'Waiting {} seconds.'.format(e, func.__name__, delay),\n exc_info=True) # logs traceback\n sleep(delay)\n\n arg_str = ', '.join((\n ', '.join(map(repr, args)),\n ', '.join('{}={!r}'.format(k, v) for k, v in kwargs),\n ))\n logger.debug('Retrying {}({})'.format(full_name, arg_str))\n\n return wrapper\n\n return decorator\n", "sub_path": "devops/helpers/retry.py", "file_name": "retry.py", "file_ext": "py", "file_size_in_byte": 2525, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "inspect.ismethod", "line_number": 38, "usage_type": "call"}, {"api_name": "inspect.getmodule", "line_number": 40, "usage_type": "call"}, {"api_name": "inspect.isfunction", "line_number": 43, "usage_type": "call"}, {"api_name": "inspect.getmodule", "line_number": 45, "usage_type": "call"}, {"api_name": "devops.error.DevopsException", "line_number": 48, "usage_type": "call"}, {"api_name": "devops.logger.debug", "line_number": 62, "usage_type": "call"}, {"api_name": "devops.logger", "line_number": 62, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 66, "usage_type": "call"}, {"api_name": "devops.logger.debug", "line_number": 72, "usage_type": "call"}, {"api_name": "devops.logger", "line_number": 72, "usage_type": "name"}, {"api_name": "functools.wraps", "line_number": 51, "usage_type": "call"}]} +{"seq_id": "423061826", "text": "import matplotlib.pyplot as plt\nimport numpy as np\nimport os\nimport PIL\nimport tensorflow as tf\nfrom tensorflow import keras\nfrom tensorflow.keras import layers\nfrom tensorflow.keras.models import Sequential, save_model, load_model\nimport pathlib\nimport tensorflowjs as tfjs\n\n#get dataset\ndataset_url = \"https://storage.googleapis.com/download.tensorflow.org/example_images/flower_photos.tgz\"\ndata_dir = tf.keras.utils.get_file('flower_photos', origin=dataset_url, untar=True)\ndata_dir = pathlib.Path(data_dir)\n\n#Parameters for loader\nbatch_size = 32\nimg_height = 180\nimg_width = 180\n\n#Loader for training\ntrain_ds = tf.keras.preprocessing.image_dataset_from_directory(\n data_dir,\n validation_split=0.2,\n subset=\"training\",\n seed=123,\n image_size=(img_height, img_width),\n batch_size=batch_size)\n\n#loader for validation\nval_ds = tf.keras.preprocessing.image_dataset_from_directory(\n data_dir,\n validation_split=0.2,\n subset=\"validation\",\n seed=123,\n image_size=(img_height, img_width),\n batch_size=batch_size)\n\n#performance related functions\nAUTOTUNE = tf.data.AUTOTUNE\ntrain_ds = train_ds.cache().shuffle(1000).prefetch(buffer_size=AUTOTUNE)\nval_ds = val_ds.cache().prefetch(buffer_size=AUTOTUNE)\n\n#normalization_layer sets pixl values to be between [0,1] rather than [0,255]\nnormalization_layer = layers.experimental.preprocessing.Rescaling(1./255)\nnum_classes = 5\n\n#model is structure of network. Data augmentation is series of layers that goes into model\ndata_augmentation = keras.Sequential(\n [\n layers.experimental.preprocessing.RandomFlip(\"horizontal\", \n input_shape=(img_height, \n img_width,\n 3)),\n layers.experimental.preprocessing.RandomRotation(0.1),\n layers.experimental.preprocessing.RandomZoom(0.1),\n ]\n)\n\nmodel = Sequential([\n data_augmentation,\n layers.experimental.preprocessing.Rescaling(1./255),\n layers.Conv2D(16, 3, padding='same', activation='relu'),\n layers.MaxPooling2D(),\n layers.Conv2D(32, 3, padding='same', activation='relu'),\n layers.MaxPooling2D(),\n layers.Conv2D(64, 3, padding='same', activation='relu'),\n layers.MaxPooling2D(),\n layers.Dropout(0.2),\n layers.Flatten(),\n layers.Dense(128, activation='relu'),\n layers.Dense(num_classes)\n])\n\n\nmodel.compile(optimizer='adam',\n loss=tf.keras.losses.SparseCategoricalCrossentropy(from_logits=True),\n metrics=['accuracy'])\n\n#train data for epochs num of epochs\nepochs=15\nhistory = model.fit(\n train_ds,\n validation_data=val_ds,\n epochs=epochs\n)\n#tfjs.converters.save_keras_model(model, 'flower-learner/saved_model_js')\nmodel.save('saved_model/my_model')", "sub_path": "flower-model.py", "file_name": "flower-model.py", "file_ext": "py", "file_size_in_byte": 2764, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "tensorflow.keras.utils.get_file", "line_number": 14, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 15, "usage_type": "call"}, {"api_name": "tensorflow.keras.preprocessing.image_dataset_from_directory", "line_number": 23, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 23, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.preprocessing.image_dataset_from_directory", "line_number": 32, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 32, "usage_type": "attribute"}, {"api_name": "tensorflow.data", "line_number": 41, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.experimental.preprocessing.Rescaling", "line_number": 46, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.experimental", "line_number": 46, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers", "line_number": 46, "usage_type": "name"}, {"api_name": "tensorflow.keras.Sequential", "line_number": 50, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 50, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.experimental.preprocessing.RandomFlip", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.experimental", "line_number": 52, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers", "line_number": 52, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.experimental.preprocessing.RandomRotation", "line_number": 56, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.experimental", "line_number": 56, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers", "line_number": 56, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.experimental.preprocessing.RandomZoom", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.experimental", "line_number": 57, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers", "line_number": 57, "usage_type": "name"}, {"api_name": "tensorflow.keras.models.Sequential", "line_number": 61, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.experimental.preprocessing.Rescaling", "line_number": 63, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.experimental", "line_number": 63, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers", "line_number": 63, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Conv2D", "line_number": 64, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 64, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.MaxPooling2D", "line_number": 65, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 65, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Conv2D", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 66, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.MaxPooling2D", "line_number": 67, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 67, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Conv2D", "line_number": 68, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 68, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.MaxPooling2D", "line_number": 69, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 69, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Dropout", "line_number": 70, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 70, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Flatten", "line_number": 71, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 71, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 72, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 72, "usage_type": "name"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers", "line_number": 73, "usage_type": "name"}, {"api_name": "tensorflow.keras.losses.SparseCategoricalCrossentropy", "line_number": 78, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 78, "usage_type": "attribute"}]} +{"seq_id": "150691994", "text": "from collections import OrderedDict\n\nimport jenkinsapi.jenkins\nfrom jenkins import Jenkins\n\nclass JenkinsHelper:\n def __init__(self, url, username, password, key, configuration):\n self.configuration = configuration\n self.configuration_prefix = configuration.jenkins_configuration_prefix()\n self.grouped_components = configuration.jenkins_grouped_components()\n self.pullrequest_job = configuration.pullrequest_job()\n\n self.jenkins_configurations = Jenkins(url, username, password)\n self.build_components_jobs_matrix()\n self.sort_components()\n\n self.jenkins_builds = jenkinsapi.jenkins.Jenkins(url, username, password)\n\n# Getters\n def get_jobs(self):\n return self.jobs\n\n def get_components(self):\n return self.components\n\n def get_pull_request_builds(self):\n job = self.jenkins_builds.get_job(self.pullrequest_job)\n\n self.pullrequest_builds = {}\n for build_id in job.get_build_ids():\n build = job[build_id]\n self.pullrequest_builds[build.buildno] = {}\n self.pullrequest_builds[build.buildno]['status'] = build.get_status()\n self.pullrequest_builds[build.buildno]['url'] = build.baseurl\n self.pullrequest_builds[build.buildno]['name'] = build.name\n\n revision = build.get_revision_branch()[0]\n self.pullrequest_builds[build.buildno]['revision'] = revision['SHA1']\n self.pullrequest_builds[build.buildno]['revision_name'] = revision['name']\n\n return self.pullrequest_builds\n\n# Helper methods\n def initial_jobs_info(self):\n wanted_jobs, wanted_ids = self.configuration.jenkins_jobs()\n jobs = self.wanted_jobs(wanted_jobs)\n return self.add_human_name_to_job(jobs, wanted_ids)\n\n def sort_components(self):\n self.components = OrderedDict(sorted(self.components.items(), key=lambda x: self.sorting_modification(x)))\n\n def build_components_jobs_matrix(self):\n self.jobs = self.initial_jobs_info()\n\n self.components = {}\n for job in self.jobs:\n groups = {}\n job_raw_components = self.jenkins_configurations.get_job_info(job['name'])[\"activeConfigurations\"]\n\n job['components'] = {}\n for raw_component in job_raw_components:\n self.process_component(raw_component, job, groups)\n\n for name, group in groups.iteritems():\n job['components'][name] = group\n self.add_group_to_components(name, group)\n\n def add_group_to_components(self, name, group):\n self.components[name] = {}\n self.components[name]['name'] = group['name']\n self.components[name]['global_class'] = 'group'\n self.components[name]['type'] = 'group';\n\n def process_component(self, raw_component, job, groups):\n name = raw_component['name'].replace(self.configuration_prefix, '')\n\n if name not in self.components:\n self.components[name] = {}\n self.components[name]['name'] = name\n\n job['components'][name] = {};\n job['components'][name]['name'] = name;\n job['components'][name]['color'] = raw_component['color']\n job['components'][name]['href'] = raw_component['url']\n\n # Manage grouped components\n grouped_component = self.has_to_be_grouped(name, self.grouped_components)\n if grouped_component:\n self.components[name]['global_class'] = grouped_component + ' hide grouped'\n\n # Create component group entry\n group_name = grouped_component + '_grouped'\n if not group_name in groups:\n groups[group_name] = {'name': grouped_component, 'color': ''}\n\n groups[group_name]['color'] = self.logical_color_conjunction(\n groups[group_name]['color'],\n raw_component['color'])\n\n# Second level helper methods\n def wanted_jobs(self, wanted_jobs):\n jobs = self.jenkins_configurations.get_jobs()\n\n return [ job for job in jobs if job['name'] in wanted_jobs ]\n\n def add_human_name_to_job(self, jobs, wanted_jobs_ids):\n for i in range(len(jobs)):\n jobs[i]['short_name'] = wanted_jobs_ids[i]\n\n return jobs\n\n\n def has_to_be_grouped(self, name, grouped_configuration):\n for keyword in grouped_configuration:\n if name.find(keyword) == 0:\n return keyword\n\n return False\n\n\n def logical_color_conjunction(self, color1, color2):\n if (color1 == 'red' or color2 == 'red'):\n return 'red'\n if (color1 == 'yellow' or color2 == 'yellow'):\n return 'yellow'\n if (color1 == 'blue' or color2 == 'blue'):\n return 'blue'\n\n return 'white'\n\n def sorting_modification(self, value):\n if ('type' in value[1]) and (value[1]['type'] == 'group'):\n return value[1]['name']\n\n return value[0] + \"_\"\n\n", "sub_path": "jenkins_parser.py", "file_name": "jenkins_parser.py", "file_ext": "py", "file_size_in_byte": 5030, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "jenkins.Jenkins", "line_number": 13, "usage_type": "call"}, {"api_name": "jenkinsapi.jenkins.jenkins.Jenkins", "line_number": 17, "usage_type": "call"}, {"api_name": "jenkinsapi.jenkins.jenkins", "line_number": 17, "usage_type": "attribute"}, {"api_name": "jenkinsapi.jenkins", "line_number": 17, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 50, "usage_type": "call"}]} +{"seq_id": "432188732", "text": "import csv\nfrom datetime import datetime\nimport re\nimport glob\n\ndef removeEmoji(inputString):\n return inputString.encode('ascii', 'ignore').decode('ascii')\n\nclass tweet:\n tString = \" \"\n likes = 0\n date = 0 \n retweets = 0 \n links = []\n label = 0\n\ndef __repr__(self):\n return (tString)\n\n#all tweets stored in this dictionary\n#each key is the tweet ID\n#each value is the tweet class above\ntweetDictionary = {}\n\ndef loadTweets(subFile):\n for folder in glob.glob(\"TweetsFolder/\"):\n for file in glob.glob(folder + \"*.csv\"):\n with open(file , encoding='cp437') as csvfile:\n readCSV = csv.reader(csvfile, delimiter=',')\n firstline = True\n for row in readCSV:\n #print(row[7])\n if firstline and row[0] == \"tweets\":\n firstline = False\n continue\n else:\n firstline = False \n #currently the csv format os\n #0 - tweet 1 - id 2 - len\n #3 - mm/dd/yyyy 4 - source 5 - likes\n #6 - retweets 7 - label\n newTweet = tweet()\n #need to filter the tweet of caps and grammar and stuff \n tweetString = removeEmoji(row[0])\n link = re.search(r'http\\S+', tweetString)\n if link != None:\n #print(link.group(0))\n newTweet.links.append(link.group(0))\n tweetString = re.sub(r'http\\S+', '', tweetString)\n tweetString = re.sub(r'\\r\\n', \" \", tweetString)\n tweetString = re.sub(r'\\'' , \"\" , tweetString)\n tweetString = re.sub(r\"[^a-zA-Z@_0-9]+\", \" \", tweetString)\n tweetString = re.sub(r'[\" \"]+', \" \", tweetString)\n tweetString = tweetString.lower()\n newTweet.tString = tweetString\n newTweet.likes = row[5]\n newTweet.retweets = row[6]\n if (subFile == \"HashtagTweets\"):\n newTweet.label = row[7]\n date = (row[3].split(\" \"))\n YMD = date[0]\n date = datetime.strptime(YMD, '%Y-%m-%d')\n newTweet.date = date\n id = row[1]\n #duplicate checking \n if id not in tweetDictionary:\n tweetDictionary[id] = newTweet\n return (tweetDictionary)\n\n\ndef main():\n print(\"Loading Tweets\")\n dict = loadTweets(\"*\")\n print(len(dict.keys()))\n print(\"Finished Loading Tweets\")\n return\n \nif __name__ == \"__main__\":\n main()\n", "sub_path": "src/tweetLoader.py", "file_name": "tweetLoader.py", "file_ext": "py", "file_size_in_byte": 2802, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "glob.glob", "line_number": 26, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 27, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 29, "usage_type": "call"}, {"api_name": "re.search", "line_number": 45, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 49, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 50, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 51, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 52, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 53, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 62, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 62, "usage_type": "name"}]} +{"seq_id": "5855938", "text": "import os\nimport sys\nimport time\nimport argparse\nimport operator\nimport collections\n\n\nfrom neopixel import *\nfrom emotion_API import Emotion_API\n\n# From https://github.com/jgarff/rpi_ws281x/blob/master/python/examples/strandtest.py\n# LED strip configuration:\nLED_COUNT = 60 # Number of LED pixels.\nLED_PIN = 18 # GPIO pin connected to the pixels (18 uses PWM!).\n# LED_PIN = 10 # GPIO pin connected to the pixels (10 uses SPI /dev/spidev0.0).\nLED_FREQ_HZ = 800000 # LED signal frequency in hertz (usually 800khz)\nLED_DMA = 5 # DMA channel to use for generating signal (try 5)\nLED_BRIGHTNESS = 50 # Set to 0 for darkest and 255 for brightest\n# True to invert the signal (when using NPN transistor level shift)\nLED_INVERT = False\nLED_CHANNEL = 0 # set to '1' for GPIOs 13, 19, 41, 45 or 53\nLED_STRIP = ws.WS2812_STRIP # Strip type and colour ordering\n\n\ndef display_color(strip, color):\n \"\"\"Display `color`.\"\"\"\n for i in range(strip.numPixels()):\n strip.setPixelColor(i, color)\n strip.show()\n time.sleep(10)\n\n\ndef color_wipe(strip, color, wait_ms=50):\n # From https://github.com/jgarff/rpi_ws281x/blob/master/python/examples/strandtest.py\n \"\"\"Wipe color across display a pixel at a time.\"\"\"\n for i in range(strip.numPixels()):\n strip.setPixelColor(i, color)\n strip.show()\n time.sleep(wait_ms / 1000.0)\n\n\ndef get_emotion_scores(emo, filename='image.jpg'):\n # Detect emotions\n results, _ = emo.get_emotions(filename)\n\n # Take first face detected\n try:\n scores = results[0]['scores']\n except IndexError:\n print(\"No faces found.\")\n return None, None\n except TypeError:\n print(\"Make sure your API key is correct.\")\n sys.exit()\n\n # Get most likely emotion\n top_emotion = max(scores, key=lambda key: scores[key])\n print(\"Top emotion: {}\".format(top_emotion))\n return scores, top_emotion\n\n\ndef get_colors(scores, top_emotion):\n \"\"\"Map colors to emotion `scores`\"\"\"\n red = green = blue = 0\n # For warm and cold emotions, return solid color\n if top_emotion == 'anger':\n color = (255, 0, 0) # red\n return color\n elif top_emotion == 'fear':\n color = (255, 255, 0) # yellow\n return color\n elif top_emotion in ['sadness', 'contempt']:\n color = (0, 0, 255) # blue\n return color\n for e in scores.keys():\n if e in ['anger', 'fear', 'happiness']:\n red += scores[e]\n if e in ['disgust', 'surprise', 'contempt', 'happiness']:\n green += scores[e]\n if e in ['neutral', 'sadness']:\n blue += scores[e]\n color = [int(c * 255) for c in [red, green, blue]]\n print(\"Red: {}, Green: {}, Blue: {}\".format(*color))\n return color\n\n\ndef main(single=False, delay=10):\n # Initialize emotion API\n emo = Emotion_API()\n if single:\n photo = None\n while photo is None:\n os.system('sudo fswebcam --no-banner image.jpg')\n while not os.path.exists('image.jpg'):\n os.system('sudo fswebcam --no-banner image.jpg')\n time.sleep(3)\n scores, top_emotion = get_emotion_scores(emo)\n if scores == None: # No emotions detected\n continue\n _color = get_colors(scores, top_emotion)\n color = Color(*_color)\n print(\"Displaying {}\".format(_color))\n display_color(strip, color)\n photo = True\n input(\"Press Enter to exit...\")\n else: # looping\n while True:\n os.system('sudo fswebcam --no-banner image.jpg')\n while not os.path.exists('image.jpg'):\n os.system('sudo fswebcam --no-banner image.jpg')\n time.sleep(3)\n # Initialize emotion API\n emo = Emotion_API()\n scores, top_emotion = get_emotion_scores(emo)\n if scores == None:\n continue\n _color = get_colors(scores, top_emotion)\n color = Color(*_color)\n display_color(strip, color)\n time.sleep(delay)\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument(\"-c\", \"--color\", type=str,\n help=\"display single RGB or hex color (eg, 255x255x255 or #FFFFFF)\")\n parser.add_argument(\"-s\", \"--single\",\n help=\"update lights with a single photo\", action=\"store_true\")\n parser.add_argument(\"-d\", \"--delay\", type=int,\n help=\"time delay between emotion sensing (in seconds)\")\n args = parser.parse_args()\n\n # Create NeoPixel object with appropriate configuration\n strip = Adafruit_NeoPixel(LED_COUNT, LED_PIN, LED_FREQ_HZ,\n LED_DMA, LED_INVERT, LED_BRIGHTNESS, LED_CHANNEL, LED_STRIP)\n\n # Initialize the library\n strip.begin()\n\n if args.color: # Display single color\n if '#' in args.color:\n # Convert hex to Color\n color = args.color.lstrip('#')\n color = Color(tuple(int(h[i:i + 2], 16) for i in (0, 2, 4)))\n else:\n # Extract RGB int values from string\n _color = args.color.split('x')\n _color = [int(x) for x in _color]\n print(\"Displaying {}\".format(_color))\n color = Color(*tuple(_color))\n while True:\n display_color(strip, color)\n if args.delay:\n main(single=args.single, delay=args.delay)\n else:\n main(single=args.single)\n", "sub_path": "moodlight.py", "file_name": "moodlight.py", "file_ext": "py", "file_size_in_byte": 5521, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "time.sleep", "line_number": 31, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 40, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 55, "usage_type": "call"}, {"api_name": "emotion_API.Emotion_API", "line_number": 90, "usage_type": "call"}, {"api_name": "os.system", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 96, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 97, "usage_type": "call"}, {"api_name": "os.system", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 110, "usage_type": "call"}, {"api_name": "os.path", "line_number": 110, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 111, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 112, "usage_type": "call"}, {"api_name": "emotion_API.Emotion_API", "line_number": 114, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 121, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 125, "usage_type": "call"}]} +{"seq_id": "250354244", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Sun May 19 14:28:59 2019\r\n\r\n@author: Andrew\r\n\"\"\"\r\n\r\n######################\r\n## import libraries ##\r\n######################\r\nimport numpy as np\r\nimport pandas as pd\r\nfrom sklearn import datasets\r\nimport statsmodels.formula.api as sm\r\nfrom numpy.linalg import inv\r\nimport matplotlib.pyplot as plt \r\n\r\n###################################\r\n## import and clean iris dataset ##\r\n###################################\r\niris = datasets.load_iris()\r\nLR_df = pd.DataFrame()\r\nLR_df['S_Length'] = iris['data'][:,0]\r\nLR_df['Intercept']=np.full(iris['data'].shape[0], 1)\r\nLR_df['S_Width'] = iris['data'][:,1]\r\nLR_df['P_Length'] = iris['data'][:,2]\r\nLR_df['P_Width'] = iris['data'][:,3]\r\nLR_df['Species'] = iris['target']\r\nLR_df['Species'] = LR_df['Species'].apply(str)\r\nLR_df.loc[LR_df['Species']==str(0), \"Species\"] = str(iris['target_names'][0])\r\nLR_df.loc[LR_df['Species']==str(1), \"Species\"] = str(iris['target_names'][1])\r\nLR_df.loc[LR_df['Species']==str(2), \"Species\"] = str(iris['target_names'][2])\r\nLR_df['Species_setosa']=0\r\nLR_df.loc[LR_df['Species']=='setosa', 'Species_setosa']=1\r\nLR_df['Species_versicolor']=0\r\nLR_df.loc[LR_df['Species']=='versicolor', 'Species_versicolor']=1\r\nLR_df = LR_df.drop('Species', axis=1)\r\n\r\n#######################################\r\n## creat arrays for Gradient Descent ##\r\n#######################################\r\nY = np.array(LR_df['S_Length']).reshape((len(LR_df['S_Length']), 1))\r\nX = np.array(LR_df[['Intercept', 'S_Width', 'P_Length', 'P_Width', 'Species_setosa', 'Species_versicolor']])\r\n\r\n#################################\r\n## initialize parameter matrix ##\r\n#################################\r\nk = X.shape[1]\r\nnp.random.seed(10815657)\r\nnudge=0.01\r\nBeta = np.random.uniform(low=-1*nudge, high=1*nudge, size=k).reshape(k, 1)\r\n\r\n####################\r\n## Newton Rhapson ##\r\n####################\r\nm = 5\r\nJ = pd.DataFrame()\r\nJ['iterative_step'] = range(0,m+1)\r\nJ['cost'] = np.full(m+1, None)\r\nJ.loc[0, 'cost'] = np.asscalar(np.dot((Y - np.dot(X, Beta)).T, (Y - np.dot(X, Beta))))\r\n\r\ninv_J2_partial_Beta2 = inv(2*np.dot(X.T, X))\r\nfor i in range(1, m+1): \r\n J_partial_Beta = (-2*np.dot(X.T, Y)) + (2*np.dot(np.dot(X.T, X), Beta))\r\n Beta = Beta - np.dot(inv_J2_partial_Beta2, J_partial_Beta)\r\n J.loc[i, 'cost'] = np.asscalar(np.dot((Y - np.dot(X, Beta)).T, (Y - np.dot(X, Beta))))\r\n del J_partial_Beta \r\n\r\nplt.plot(J['iterative_step'], J['cost'])\r\nplt.title('Newton Rhapson') \r\nplt.xlabel('Iterative Step') \r\nplt.ylabel('Cost') \r\nprint(Beta)\r\n\r\n## built in package\r\nresults = sm.ols(formula=\"S_Length ~ S_Width + P_Length + P_Width + Species_setosa + Species_versicolor\", data=LR_df).fit()\r\nprint(results.params)\r\n\r\n\r\n", "sub_path": "python code/OLS Regression - Newton Rhapson.py", "file_name": "OLS Regression - Newton Rhapson.py", "file_ext": "py", "file_size_in_byte": 2700, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "sklearn.datasets.load_iris", "line_number": 21, "usage_type": "call"}, {"api_name": "sklearn.datasets", "line_number": 21, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 49, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 51, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.full", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.asscalar", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.asscalar", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "statsmodels.formula.api.ols", "line_number": 76, "usage_type": "call"}, {"api_name": "statsmodels.formula.api", "line_number": 76, "usage_type": "name"}]} +{"seq_id": "12949177", "text": "'''\n\nConsider quadratic Diophantine equations of the form:\n\nx2 – Dy2 = 1\n\nFor example, when D=13, the minimal solution in x is 649^2 – 13×180^2 = 1.\n\nIt can be assumed that there are no solutions in positive integers when D is square.\n\nBy finding minimal solutions in x for D = {2, 3, 5, 6, 7}, we obtain the following:\n\n3^2 – 2×2^2 = 1\n2^2 – 3×1^2 = 1\n9^2 – 5×4^2 = 1\n5^2 – 6×2^2 = 1\n8^2 – 7×3^2 = 1\n\nHence, by considering minimal solutions in x for D ≤ 7, the largest x is obtained when D=5.\n\nFind the value of D ≤ 1000 in minimal solutions of x for which the largest value of x is obtained.\n\n'''\n\nimport fractions as fr\nimport numpy as np\n\ndef period(n):\n\n\tdef continuedFraction(number,multiplier,rest):\n\t\tdivisor = (number-rest**2)/multiplier\n\t\trest -= divisor\n\t\ta = 1\n\t\twhile rest > 0 or (rest-divisor)**2 < number:\n\t\t\trest = rest - divisor\n\t\t\ta += 1\n\t\trest = -rest\n\t\tcontFractions.append((a,divisor,rest))\n\t\treturn a,divisor,rest\n\n\ti = 0\n\twhile (i+1)**2 <= n:\n\t\ti+=1\n\tif i**2 == n:\n\t\treturn 0\n\telse:\n\t\tcontFractions = [i]\n\t\tmultiplier = 1\n\t\ta = 0\n\t\trest = i\n\t\tperiod = 0\n\t\twhile True:\n\t\t\ta,multiplier,rest = continuedFraction(n,multiplier,rest)\n\t\t\tif period != 0 and (a,multiplier,rest) == contFractions[1]:\n\t\t\t\treturn contFractions[:-1]\n\t\t\tperiod += 1\n\n\ndef solutions(frac):\n\n\tdef continuedFraction(i,frac):\n\t\tlength = len(frac)\n\t\tif i >= length:\n\t\t\t\tj = i%length\n\t\telse:\n\t\t\tj = i\n\t\tf = fr.Fraction(1,frac[j][0])\n\t\ti-=1\n\t\twhile i >=0:\n\t\t\tif i >= length:\n\t\t\t\tj = i%length\n\t\t\telse:\n\t\t\t\tj = i\n\t\t\tf = fr.Fraction(1,frac[j][0]+f)\n\t\t\ti-=1\n\t\treturn f\n\n\tbase = frac[0]\n\tfrac = frac[1:]\n\ti = 0\n\twhile True:\n\t\tyield base+continuedFraction(i,frac)\n\t\ti+=1\n\n\ndef main():\n\tmaxNumerator = 1\n\tmaxD = 1\n\tfor D in range(2,1001):\n\t\tfrac = period(D)\n\t\tif frac == 0:\n\t\t\tcontinue\n\t\tfor fraction in solutions(frac):\n\t\t\tif fraction.numerator**2 - D*fraction.denominator**2 == 1:\n\t\t\t\tprint(D, fraction.numerator,fraction.denominator)\n\t\t\t\tif fraction.numerator > maxNumerator:\n\t\t\t\t\tmaxNumerator = fraction.numerator\n\t\t\t\t\tmaxD = D\n\t\t\t\tbreak\n\tprint(maxD)\n\n\nif __name__ == '__main__':\n\tmain()\n\n", "sub_path": "Problem66.py", "file_name": "Problem66.py", "file_ext": "py", "file_size_in_byte": 2096, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "fractions.Fraction", "line_number": 67, "usage_type": "call"}, {"api_name": "fractions.Fraction", "line_number": 74, "usage_type": "call"}]} +{"seq_id": "388068210", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nimport ckeditor.fields\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('blog', '0012_slider_activo'),\n ]\n\n operations = [\n migrations.CreateModel(\n name='Semblanza',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('imagen', models.ImageField(help_text=b'Imagen del maestro.', upload_to=b'semblanzas')),\n ('nombre', models.CharField(max_length=100)),\n ('contenido', ckeditor.fields.RichTextField()),\n ],\n options={\n },\n bases=(models.Model,),\n ),\n migrations.AlterField(\n model_name='entrada',\n name='imagen',\n field=models.ImageField(help_text=b'Imagen de la noticia', upload_to=b'imagenes'),\n preserve_default=True,\n ),\n migrations.AlterField(\n model_name='entrada',\n name='resumen',\n field=models.CharField(help_text=b'Resumen de la noticia maximo 170 caracteres', max_length=170),\n preserve_default=True,\n ),\n migrations.AlterField(\n model_name='entrada',\n name='titulo',\n field=models.CharField(help_text=b'Titulo de la noticia maximo 100 caracteres', max_length=100),\n preserve_default=True,\n ),\n migrations.AlterField(\n model_name='slider',\n name='activo',\n field=models.BooleanField(default=True, help_text=b'Mostrar o no la imagen con texto'),\n preserve_default=True,\n ),\n migrations.AlterField(\n model_name='slider',\n name='imagen',\n field=models.ImageField(help_text=b'Imagen que sera mostrada', upload_to=b'noticias'),\n preserve_default=True,\n ),\n migrations.AlterField(\n model_name='slider',\n name='titulo',\n field=models.CharField(help_text=b'Titulo que sera mostrado en la imagen, maximo 50 caracteres', max_length=50),\n preserve_default=True,\n ),\n ]\n", "sub_path": "blog/migrations/0013_auto_20160406_2145.py", "file_name": "0013_auto_20160406_2145.py", "file_ext": "py", "file_size_in_byte": 2256, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 15, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.ImageField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "ckeditor.fields.fields.RichTextField", "line_number": 21, "usage_type": "call"}, {"api_name": "ckeditor.fields.fields", "line_number": 21, "usage_type": "attribute"}, {"api_name": "ckeditor.fields", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 27, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 27, "usage_type": "name"}, {"api_name": "django.db.models.ImageField", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 33, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 36, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 36, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 39, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 39, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 42, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 42, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 45, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 45, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 48, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 48, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 51, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 51, "usage_type": "name"}, {"api_name": "django.db.models.ImageField", "line_number": 54, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 54, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 57, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 57, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 60, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 60, "usage_type": "name"}]} +{"seq_id": "450452933", "text": "import logging\nimport datetime\n\nfrom umsgpack import unpackb\n\nfrom .backends.hbase import HBaseBase\nfrom .backends.elasticsearch import ESMessage, ESPersona\nfrom .config import HBASE_HOSTNAME, HBASE_PORT, LOGGING\n\n\nlogger = logging.getLogger(__name__)\nlogging.config.dictConfig(LOGGING)\n\n\nclass ModelBase:\n \"\"\"\n Define some common methods to be used in all Models.\n \"\"\"\n\n def save(self):\n try:\n # Raise exception if trying to double save a message\n self.save_to_hbase()\n self.save_to_es()\n except:\n self.delete_from_hbase()\n self.delete_from_es()\n logger.error(exc)\n else:\n logger.info(f'{self} sucessfully saved')\n\n\nclass Message(ModelBase, HBaseBase, ESMessage):\n hbase_table = 'message'\n hbase_family = 'message'\n hbase_row_key = 'hash'\n hbase_fields = (\n 'persona_sender',\n 'persona_pubkey',\n 'persona_nickname',\n 'type',\n 'hash',\n 'signature',\n 'timestamp',\n 'dossier_hash',\n 'body_hash',\n 'acl',\n 'objects',\n 'message',\n 'created_at',\n )\n es_fields = (\n 'persona_sender',\n 'persona_nickname',\n 'hash',\n 'dossier_hash',\n 'created_at',\n )\n es_id = 'hash'\n\n\n def __repr__(self):\n return f\"\"\n\n def to_dict(self):\n \"\"\"\n The MessageEnvelope structure used accross the infrastructure\n have fields in camelCase. Since, this is a Pythonic library, the\n to_dict() results returns fields in the pythonic_way.\n\n The to_dict() method is highly used by Reader to return the Message\n in JSON format. Here, we change all the field names to be complient\n with the MessageEnvelope structure.\n \"\"\"\n\n result = super().to_dict()\n\n persona_sender = result.get('persona_sender')\n if persona_sender:\n result['sender'] = persona_sender\n result.pop('persona_sender')\n\n type = result.get('type')\n if type:\n result['messageType'] = type\n result.pop('type')\n\n hash = result.get('hash')\n if hash:\n result['messageHash'] = hash\n result.pop('hash')\n\n signature = result.get('signature')\n if signature:\n result['messageSign'] = signature\n result.pop('signature')\n\n dossier_hash = result.get('dossier_hash')\n if dossier_hash:\n result['dossierHash'] = dossier_hash\n result.pop('dossier_hash')\n\n body_hash = result.get('body_hash')\n if body_hash:\n result['bodyHash'] = body_hash\n result.pop('body_hash')\n\n acl = result.get('acl')\n if acl:\n unpacked_acl = unpackb(acl)\n acls = unpacked_acl if (isinstance(unpacked_acl, list) and len(unpacked_acl) > 0) else None\n if acls:\n for item in acls:\n persona = Persona.get(address=item['reader'])\n item['reader'] = {\n 'address': persona.address,\n 'nickname': persona.nickname,\n 'pubkey': persona.pubkey,\n }\n\n result['ACL'] = unpacked_acl\n result.pop('acl')\n\n objects = result.get('objects')\n if objects:\n result['objects'] = unpackb(objects)\n\n return result\n\n @classmethod\n def get(cls, hash=None, container_hash=None):\n if hash:\n return cls.get_from_hbase(hash)\n\n if container_hash:\n hash = cls.get_hash_from_es(container_hash=container_hash)\n return cls.get_from_hbase(hash)\n\n @classmethod\n def filter(cls, created_at=None, persona_sender=None):\n if created_at:\n hashes = cls.filter_from_es(created_at=created_at)\n\n if persona_sender:\n hashes = cls.filter_from_es(persona_sender=persona_sender)\n\n messages = [cls.get(hash=hash) for hash in hashes if hashes]\n return messages\n\n\nclass Persona(ModelBase, HBaseBase, ESPersona):\n hbase_table = 'persona'\n hbase_family = 'persona'\n hbase_row_key = 'address'\n hbase_fields = (\n 'address',\n 'pubkey',\n 'nickname',\n 'created_at',\n )\n es_fields = (\n 'address',\n 'pubkey',\n 'nickname',\n 'created_at',\n )\n es_id = 'address'\n\n def __repr__(self):\n return f\"\"\n\n @classmethod\n def get(cls, address=None, nickname=None, pubkey=None):\n if address:\n return cls.get_from_hbase(address)\n\n if pubkey:\n address = cls.get_address_from_es(pubkey=pubkey)\n return cls.get_from_hbase(address)\n\n if nickname:\n address = cls.get_address_from_es(nickname=nickname)\n return cls.get_from_hbase(address)\n", "sub_path": "himalaya_models/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 4960, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "logging.config.dictConfig", "line_number": 12, "usage_type": "call"}, {"api_name": "config.LOGGING", "line_number": 12, "usage_type": "argument"}, {"api_name": "logging.config", "line_number": 12, "usage_type": "attribute"}, {"api_name": "backends.hbase.HBaseBase", "line_number": 33, "usage_type": "name"}, {"api_name": "backends.elasticsearch.ESMessage", "line_number": 33, "usage_type": "name"}, {"api_name": "umsgpack.unpackb", "line_number": 110, "usage_type": "call"}, {"api_name": "umsgpack.unpackb", "line_number": 126, "usage_type": "call"}, {"api_name": "backends.hbase.HBaseBase", "line_number": 151, "usage_type": "name"}, {"api_name": "backends.elasticsearch.ESPersona", "line_number": 151, "usage_type": "name"}]} +{"seq_id": "110407841", "text": "import pytest\nimport sys, os\nimport xarray as xr\nimport datetime\nimport logging\nimport multiprocessing\n\nsys.path.append(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))\nimport process\nfrom process._common import ProcessArgumentInvalid, ProcessArgumentRequired\n\n\n@pytest.fixture\ndef generate_data():\n def _construct(\n data = [[[[0.1, 0.15], [0.15, 0.2]], [[0.05, 0.1], [-0.9, 0.05]]]],\n dims = ('t','y', 'x', 'band')\n ):\n\n xrdata = xr.DataArray(\n data,\n dims=dims\n )\n\n return xrdata\n return _construct\n\n\n@pytest.fixture\ndef execute_reduce_process(generate_data):\n logger = multiprocessing.log_to_stderr()\n logger.setLevel(logging.DEBUG)\n def wrapped(data_arguments={}, dimension=\"band\", reducer=None, target_dimension=None, binary=None, logger=logger):\n arguments = {}\n if data_arguments is not None: arguments[\"data\"] = generate_data(**data_arguments)\n if dimension is not None: arguments[\"dimension\"] = dimension\n if reducer is not None: arguments[\"reducer\"] = reducer\n if target_dimension is not None: arguments[\"target_dimension\"] = target_dimension\n if binary is not None: arguments[\"binary\"] = binary\n\n return process.reduce.reduceEOTask(None, \"\" , logger, {}, \"node1\").process(arguments)\n return wrapped\n\n\n###################################\n# tests:\n###################################\n\ndef test_no_reducer(execute_reduce_process, generate_data):\n \"\"\"\n Test reduce process without reducer\n \"\"\"\n with pytest.raises(ProcessArgumentInvalid) as ex:\n result = execute_reduce_process()\n assert ex.value.args[0] == \"The argument 'dimension' in process 'reduce' is invalid: Dimension 'band' has more than one value, but reducer is not specified.\"\n\n expected_result = generate_data = generate_data(data = [[[0.1, 0.15], [0.15, 0.2]], [[0.05, 0.1], [-0.9, 0.05]]], dims = ('y','x','band'))\n result = execute_reduce_process(dimension=\"t\")\n xr.testing.assert_allclose(result, expected_result)\n\n\ndef test_recursiver_reducer(execute_reduce_process, generate_data):\n \"\"\"\n Test reduce process with a recursive reducer, which applies min to all dimensions, apart from the last one\n \"\"\"\n reducer = {\n \"callback\": {\n \"p1\": {\n \"process_id\": \"reduce\",\n \"arguments\": {\n \"data\": {\"from_argument\": \"data\"},\n \"dimension\": \"x\",\n \"reducer\": {\n \"callback\": {\n \"p1\": {\n \"process_id\": \"reduce\",\n \"arguments\": {\n \"data\": {\"from_argument\": \"data\"},\n \"dimension\": \"band\",\n \"reducer\": {\n \"callback\": {\n \"min\": {\n \"process_id\": \"min\",\n \"arguments\": {\n \"data\": {\"from_argument\": \"data\"}\n },\n \"result\": True\n }\n }\n }\n }\n },\n \"min\": {\n \"process_id\": \"min\",\n \"arguments\": {\n \"data\": {\"from_node\": \"p1\"}\n },\n \"result\": True\n }\n }\n }\n }\n },\n \"min\": {\n \"process_id\": \"min\",\n \"arguments\": {\n \"data\": {\"from_node\": \"p1\"}\n },\n \"result\": True\n }\n }\n }\n\n result = execute_reduce_process(reducer=reducer, dimension=\"y\")\n expected_result = generate_data(data = [-0.9], dims = ('t'))\n xr.testing.assert_allclose(result, expected_result)\n\n\ndef test_reducer_sum_of_min_and_mean(execute_reduce_process, generate_data):\n \"\"\"\n Test reduce process with a reducer, which takes min and mean of bands and sums it up\n \"\"\"\n reducer = {\n \"callback\": {\n \"min\": {\n \"process_id\": \"min\",\n \"arguments\": {\n \"data\": {\"from_argument\": \"data\"}\n },\n },\n \"mean\": {\n \"process_id\": \"mean\",\n \"arguments\": {\n \"data\": {\"from_argument\": \"data\"}\n },\n },\n \"sum\": {\n \"process_id\": \"sum\",\n \"arguments\": {\n \"data\": [{\"from_node\": \"min\"},{\"from_node\": \"mean\"}]\n },\n \"result\": True\n }\n }\n }\n\n result = execute_reduce_process(reducer=reducer, dimension=\"band\")\n expected_result = generate_data(data = [[[0.225, 0.325], [0.125, -1.325]]], dims = ('t','y','x'))\n xr.testing.assert_allclose(result, expected_result)\n\n\ndef test_min_time_dim(execute_reduce_process, generate_data):\n \"\"\"\n Test reduce process with a reducer, which applies min to the temporal dimension\n \"\"\"\n reducer = {\n \"callback\": {\n \"min\": {\n \"process_id\": \"min\",\n \"arguments\": {\n \"data\": {\"from_argument\": \"data\"}\n },\n \"result\": True\n }\n }\n }\n\n data_arguments = {\"data\": [[[[0.1, 0.15], [0.15, 0.2]], [[0.05, 0.1], [-0.9, 0.05]]],[[[0.7, 0.05], [-0.009, -0.2]], [[0.05, 0.1], [-0.9, 0.07]]]]}\n result = execute_reduce_process(reducer=reducer, dimension=\"t\", data_arguments=data_arguments)\n expected_result = generate_data(data=[[[0.1, 0.05], [-0.009, -0.2]], [[0.05, 0.1], [-0.9, 0.05]]], dims=('y','x','band'))\n xr.testing.assert_allclose(result, expected_result)\n", "sub_path": "workers/tests/test_reduce.py", "file_name": "test_reduce.py", "file_ext": "py", "file_size_in_byte": 5930, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "sys.path.append", "line_number": 8, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 8, "usage_type": "call"}, {"api_name": "xarray.DataArray", "line_number": 20, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 13, "usage_type": "attribute"}, {"api_name": "multiprocessing.log_to_stderr", "line_number": 31, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 32, "usage_type": "attribute"}, {"api_name": "process.reduce.reduceEOTask", "line_number": 41, "usage_type": "call"}, {"api_name": "process.reduce", "line_number": 41, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 29, "usage_type": "attribute"}, {"api_name": "pytest.raises", "line_number": 53, "usage_type": "call"}, {"api_name": "process._common.ProcessArgumentInvalid", "line_number": 53, "usage_type": "argument"}, {"api_name": "xarray.testing.assert_allclose", "line_number": 59, "usage_type": "call"}, {"api_name": "xarray.testing", "line_number": 59, "usage_type": "attribute"}, {"api_name": "xarray.testing.assert_allclose", "line_number": 116, "usage_type": "call"}, {"api_name": "xarray.testing", "line_number": 116, "usage_type": "attribute"}, {"api_name": "xarray.testing.assert_allclose", "line_number": 149, "usage_type": "call"}, {"api_name": "xarray.testing", "line_number": 149, "usage_type": "attribute"}, {"api_name": "xarray.testing.assert_allclose", "line_number": 171, "usage_type": "call"}, {"api_name": "xarray.testing", "line_number": 171, "usage_type": "attribute"}]} +{"seq_id": "73217937", "text": "def main_function(path,c):\n \n from PIL import Image, ImageFilter\n from matplotlib import pyplot as plt\n \"\"\"\n #REMOVES SHADOWs AND NOMALIZES THE IMAGE\n \"\"\"\n def remove_shadow(path):\n import cv2\n import numpy as np\n from matplotlib import pyplot as plt\n \n img = cv2.imread(path, -1)\n \n rgb_planes = cv2.split(img)\n \n result_planes = []\n result_norm_planes = []\n for plane in rgb_planes:\n dilated_img = cv2.dilate(plane, np.ones((7,7), np.uint8))\n bg_img = cv2.medianBlur(dilated_img, 21)\n diff_img = 255 - cv2.absdiff(plane, bg_img)\n #norm_img = cv2.normalize(diff_img, alpha=0, beta=255, norm_type=cv2.NORM_MINMAX, dtype=cv2.CV_8UC1) \n norm_img = cv2.normalize(diff_img, None)\n result_planes.append(diff_img)\n result_norm_planes.append(norm_img)\n \n result = cv2.merge(result_planes)\n result_norm = cv2.merge(result_norm_planes)\n \n cv2.imwrite('shadows_out.png', result)\n ###############################################\n \n img = cv2.imread('shadows_out.png',0)\n img = cv2.medianBlur(img,5)\n \n ret,th1 = cv2.threshold(img,210,255,cv2.THRESH_BINARY)\n #th2 = cv2.adaptiveThreshold(img,255,cv2.ADAPTIVE_THRESH_MEAN_C,\\cv2.THRESH_BINARY,11,2)\n #th3 = cv2.adaptiveThreshold(img,255,cv2.ADAPTIVE_THRESH_GAUSSIAN_C,\\cv2.THRESH_BINARY,11,2)\n #images = [img, th1, th2, th3]\n \n cv2.imwrite('formated.png', th1)\n \"\"\"\n #CONVERT THE IMAGE IN RESULATION 28x28\n \"\"\"\n def imageprepare(argv):\n import os\n \"\"\"\n This function returns the pixel values.\n The imput is a png file location.\n \"\"\"\n im = Image.open(argv).convert('L')\n width = float(im.size[0])\n height = float(im.size[1])\n newImage = Image.new('L', (28, 28), (255)) # creates white canvas of 28x28 pixels\n \n if width > height: # check which dimension is bigger\n # Width is bigger. Width becomes 20 pixels.\n nheight = int(round((20.0 / width * height), 0)) # resize height according to ratio width\n if (nheight == 0): # rare case but minimum is 1 pixel\n nheight = 1\n # resize and sharpen\n img = im.resize((20, nheight), Image.ANTIALIAS)#.filter(ImageFilter.SHARPEN)\n wtop = int(round(((28 - nheight) / 2), 0)) # calculate horizontal position\n newImage.paste(img, (4, wtop)) # paste resized image on white canvas\n else:\n # Height is bigger. Heigth becomes 20 pixels.\n nwidth = int(round((20.0 / height * width), 0)) # resize width according to ratio height\n if (nwidth == 0): # rare case but minimum is 1 pixel\n nwidth = 1\n # resize and sharpen\n img = im.resize((nwidth, 20), Image.ANTIALIAS)#.filter(ImageFilter.SHARPEN)\n wleft = int(round(((28 - nwidth) / 2), 0)) # caculate vertical pozition\n newImage.paste(img, (wleft, 4)) # paste resized image on white canvas\n \n c1=str(c)\n newImage.save(\"sample.png\")\n outPath = \"D:\\edited\"\n fullpath = os.path.join(outPath, c1+\".png\")\n newImage.save(fullpath,\"PNG\")\n \n tv = list(newImage.getdata()) # get pixel values\n \n # normalize pixels to 0 and 1. 0 is pure white, 1 is pure black.\n tva = [(255 - x) * 1.0 / 255.0 for x in tv]\n #print(tva)\n return tva\n \"\"\"\n #CROP ANY UNNCESSARY WHITE SPACE\n \"\"\"\n def crop_image(imageFile):\n image = Image.open(imageFile).convert('L')\n print('processing ' + imageFile)\n rgbim = image.convert('RGB')\n \n w, h = rgbim.size\n #print((w,h))\n first_pixel=(256,256)\n last_pixel=(-1, -1)\n \n mpixel=255\n \n for i in range(w):\n for j in range(h):\n px = rgbim.getpixel((i, j))\n mpixel = min(mpixel, px[0])\n if px[0]<90: \n first_pixel=(min(first_pixel[0], i), min(first_pixel[1], j))\n last_pixel=(max(last_pixel[0], i), max(last_pixel[1], j))\n \n threshold=mpixel+80\n #print(threshold) \n #print(first_pixel)\n #print(last_pixel)\n nw = last_pixel[0]-first_pixel[0]\n nh = last_pixel[1]-first_pixel[1]\n nrgbim = Image.new('RGB', (nw,nh))\n \n for i in range(nw):\n for j in range(nh):\n if rgbim.getpixel((i+first_pixel[0], j+first_pixel[1]))[0] len(data) - 1:\n for c in range(filter_size):\n temp.append(0)\n else:\n if j + z - indexer < 0 or j + indexer > len(data[0]) - 1:\n temp.append(0)\n else:\n for k in range(filter_size):\n temp.append(data[i + z - indexer][j + k - indexer])\n \n temp.sort()\n data_final[i][j] = temp[len(temp) // 2]\n temp = []\n return data_final\n \n def create_array(filename):\n crop_image(filename)\n #medfilter(path)\n x=[imageprepare('./cropped_processed_image.png')]\n newArr=[[0 for d in range(28)] for y in range(28)]\n k = 0\n for i in range(28):\n for j in range(28):\n newArr[i][j]=x[0][k]\n k=k+1\n return newArr\n \"\"\"\n \n ###########################################################\n #from keras.models import Sequential\n #from keras.layers import Dense\n #from keras.models import model_from_json\n import numpy\n #import os\n from tensorflow import keras\n #from tensorflow.keras import layers\n #from keras.models import load_model\n \"\"\"LOADING THE TRAINNED MODEL AS new_model\"\"\"\n new_model = keras.models.load_model('path_to_my_model.h5')\n new_model.summary()\n \n \n import numpy \n \n \n \"\"\"\n img = Image.open('shadows_out.png').convert(\"L\")\n \n arr = numpy.array(img)\n removed_noise = median_filter(arr, 1) \n img = Image.fromarray(removed_noise)\n #img.show()\n img = img.convert(\"L\")\n img.save(\"noiseless_image.png\")\n \"\"\"\n # crop_image(path) \n \n remove_shadow(path)\n #from matplotlib import pyplot as plt\n #plt.imshow(\"formated.png\",cmap='gray')\n \n crop_image( 'formated.png')\n #from matplotlib import pyplot as plt\n #plt.imshow(\"cropped_processed_image.png\",cmap='gray')\n \n \n \n x=[imageprepare('cropped_processed_image.png')]\n newArr=[[0 for d in range(28)] for y in range(28)]\n k = 0\n for i in range(28):\n for j in range(28):\n newArr[i][j]=x[0][k]\n k=k+1\n \n #from matplotlib import pyplot as plt\n #plt.imshow(newArr,cmap='binary')\n \n import numpy as np\n newArr= np.asarray(newArr)\n from numpy import newaxis\n newArr = newArr[...,newaxis]\n newArr.shape\n newArr = newArr.reshape((1,newArr.shape[0], newArr.shape[1], 1))\n newArr.shape\n \n predictions = new_model.predict(newArr)\n print(predictions)\n print(np.argmax(predictions))\n return np.argmax(predictions)", "sub_path": "src/load_model_2.py", "file_name": "load_model_2.py", "file_ext": "py", "file_size_in_byte": 8218, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "cv2.imread", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.split", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.dilate", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 20, "usage_type": "attribute"}, {"api_name": "cv2.medianBlur", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.absdiff", "line_number": 22, "usage_type": "call"}, {"api_name": "cv2.normalize", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.merge", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.merge", "line_number": 29, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 31, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.medianBlur", "line_number": 35, "usage_type": "call"}, {"api_name": "cv2.threshold", "line_number": 37, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 37, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 42, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 52, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 52, "usage_type": "name"}, {"api_name": "PIL.Image.new", "line_number": 55, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 55, "usage_type": "name"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 63, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 63, "usage_type": "name"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 72, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 72, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path", "line_number": 79, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 92, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 92, "usage_type": "name"}, {"api_name": "PIL.Image.new", "line_number": 117, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 117, "usage_type": "name"}, {"api_name": "tensorflow.keras.models.load_model", "line_number": 183, "usage_type": "call"}, {"api_name": "tensorflow.keras.models", "line_number": 183, "usage_type": "attribute"}, {"api_name": "tensorflow.keras", "line_number": 183, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 226, "usage_type": "name"}, {"api_name": "numpy.argmax", "line_number": 233, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 234, "usage_type": "call"}]} +{"seq_id": "300181255", "text": "import cv2\r\nimport numpy as np\r\nface_cascade = cv2.CascadeClassifier('haarcascade_frontalface_default.xml')\r\n\r\nvideocapture = cv2.VideoCapture(0)\r\nscale_factor = 1.3\r\n\r\nwhile True:\r\n ret,pic = videocapture.read()\r\n gray = cv2.cvtColor(pic, cv2.COLOR_BGR2GRAY)\r\n faces = face_cascade.detectMultiScale(gray, scale_factor, 5)\r\n \r\n for (x,y,w,h) in faces:\r\n cv2.rectangle(pic, (x, y),(x+w,y+h),(255,0,0),2)\r\n\r\n \r\n print(\"Number of faces {}\".format(len(faces)))\r\n cv2.imshow('face',pic)\r\n if cv2.waitKey(30) & 0xFF == ord('q'):\r\n break\r\nvideocapture.release()\r\ncv2.destroyAllWindows()\r\n", "sub_path": "faceDetection/face.py", "file_name": "face.py", "file_ext": "py", "file_size_in_byte": 622, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "cv2.CascadeClassifier", "line_number": 3, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 5, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 10, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 22, "usage_type": "call"}]} +{"seq_id": "512408035", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Sat Jan 2 14:19:40 2021\r\n\r\n@author: pauli\r\n\"\"\"\r\n\r\nimport pandas as pd\r\nimport numpy as np\r\nimport tweepy as tw\r\nimport tensorflow as tf\r\nimport sys\r\nfrom sklearn.model_selection import train_test_split\r\nimport matplotlib.pyplot as plt\r\nimport texttable as tt\r\nfrom wordcloud import WordCloud\r\n\r\n#działanie na danych (wczytywanie / podgląd / wstępna analiza)\r\ndf = pd.read_csv('./train.csv', index_col=0)\r\nprint()\r\nprint('Podgląd danych treningowych')\r\nprint()\r\nprint(df.head())\r\n\r\n\r\nprint()\r\nprint('Dystrybucja danych między klasami')\r\nprint()\r\nprint(df.label.value_counts())\r\n\r\nprint()\r\n\r\nlengths = []\r\nfor tweet in df['tweet']:\r\n lengths.append(len(tweet))\r\nprint('Maksymalna długości tweeta: ', max(lengths))\r\ndel lengths\r\n\r\n#mapa najczęscniej występujacych slow\r\nwords = ' '.join([tweet for tweet in df['tweet']])\r\nwordCloud = WordCloud(width=600, height=400).generate(words)\r\nplt.imshow(wordCloud)\r\nplt.savefig('./wyniki/mapa_slow.png')\r\nplt.show()\r\n\r\nprint()\r\nprint('train_test_split')\r\nprint()\r\nTrain, Test = train_test_split(df, test_size=0.3, random_state=101)\r\nprint(Train.head())\r\n\r\n#Konwertowanie ramki danych pandy na zestaw danych tensorflow...\r\ndef df_to_dataset(dataframe, batch_size=32):\r\n dataframe = dataframe.copy()\r\n labels = dataframe.pop('label')\r\n ds = tf.data.Dataset.from_tensor_slices((dataframe['tweet'], labels))\r\n ds = ds.batch(batch_size)\r\n ds = ds.prefetch(batch_size)\r\n return ds\r\n\r\ntrain_ds = df_to_dataset(Train)\r\ntest_ds = df_to_dataset(Test)\r\n\r\nprint()\r\nprint('Testowanie, czy zestaw danych został utworzony poprawnie')\r\nprint()\r\nfor text_batch, label_batch in train_ds.take(1):\r\n for i in range(3):\r\n print(\"Tweet\", text_batch.numpy()[i])\r\n print(\"Label\", label_batch.numpy()[i])\r\n\r\n\r\n#Tworzenie warstwy wektoryzacji tekstu \r\nvectorize_layer = tf.keras.layers.experimental.preprocessing.TextVectorization(max_tokens=10000, output_mode='int',\r\n output_sequence_length=274)\r\n#Tworzenie zbioru danych zawierającego tylko tekst pasujący do warstwy wektoryzacji kodowania\r\ntrain_text = train_ds.map(lambda x, y: x)\r\n\r\n#Dopasowanie warstwy\r\nvectorize_layer.adapt(train_text)\r\n\r\nprint()\r\nprint('Testowanie wektoryzowanego tekstu')\r\nprint()\r\ntext_batch, label_batch = next(iter(train_ds))\r\nfirst_review, first_label = text_batch[0], label_batch[0]\r\nprint(first_review)\r\ntext = tf.expand_dims(first_review, -1)\r\nprint(vectorize_layer(text))\r\n\r\ndef vectorize_text(text, label):\r\n text = tf.expand_dims(text[0], -1)\r\n return vectorize_layer(text), label\r\n\r\n#Wstępne przetwarzanie zbiorów danych\r\nAUTOTUNE = tf.data.experimental.AUTOTUNE\r\ntrain_ds = train_ds.cache().prefetch(buffer_size=AUTOTUNE)\r\ntest_ds = test_ds.cache().prefetch(buffer_size=AUTOTUNE)\r\n\r\n#Przygotowanie modelu\r\nmodel = tf.keras.Sequential([\r\n vectorize_layer,\r\n tf.keras.layers.Embedding(10000, 20),\r\n tf.keras.layers.Dense(256),\r\n tf.keras.layers.GlobalAveragePooling1D(),\r\n tf.keras.layers.Dropout(0.3),\r\n tf.keras.layers.Dense(1)])\r\n\r\n#Dodanie punktu kontrolnego, aby zapisać model z najlepszymi wynikami i najmniejszym dopasowaniem\r\n#jeżeli model tworzony jest na danych w innym języku niż angielski, należy zmienić w nazwie 'pliki_weryfikacyjne' skrót językowy\r\ncheckpoint_val_acc = tf.keras.callbacks.ModelCheckpoint(\r\n 'pliki_weryfikacyjne_en.tf', monitor='val_binary_accuracy', verbose=1, save_best_only=True,\r\n save_weights_only=False, save_freq='epoch')\r\n\r\n#Kompilowanie modelu\r\nmodel.compile(loss=tf.keras.losses.BinaryCrossentropy(from_logits=True),\r\n optimizer='adam',\r\n metrics=tf.metrics.BinaryAccuracy(threshold=0.0))\r\n\r\nprint()\r\nprint('Trenowanie modelu')\r\nprint()\r\nepochs = 10\r\nhistory = model.fit(train_ds, validation_data=test_ds, epochs=epochs, callbacks=[checkpoint_val_acc])\r\n\r\n\r\n#Importowanie modelu testowego\r\ng = df['tweet'][5000:7000]\r\nlabel = df['label'][5000:7000]\r\nlabels = list(label)\r\ntweets = list(g)\r\ntweets\r\n\r\n#Tworzenie funkcji, która konwertuje listę tweetów do TF Dataset\r\ndef sample_list_to_dataset(dataframe, batch_size=20):\r\n ds = tf.data.Dataset.from_tensor_slices(dataframe)\r\n ds = ds.batch(batch_size)\r\n ds = ds.prefetch(batch_size)\r\n return ds\r\n\r\n#Działanie na modelu\r\n#wczytywany jest model o danym języku - tu też należy zmienić w przypadku języka innego niż angielki\r\nmodel_2 = tf.keras.models.load_model('pliki_weryfikacyjne_en.tf')\r\npredict_ds = sample_list_to_dataset(tweets)\r\nscore = model_2.predict(predict_ds)\r\n\r\n#Otrzymane wartoci\r\npredicted_1 = 0\r\nfor i in score:\r\n if i > 0.5:\r\n predicted_1+=1 \r\nactual_1 = 0\r\nfor i in labels:\r\n if i > 0:\r\n actual_1+=1\r\n \r\n#tworzenie wyników w formie tabeli\r\nprint()\r\ntable = tt.Texttable()\r\ntable.set_cols_align([\"c\", \"c\"])\r\ntable.set_cols_valign([\"t\", \"i\"])\r\ntable.add_rows([[\"Cecha\", \"Wartosc\"],\r\n [\"Przewidywane wartosci negatywne\",predicted_1],\r\n [\"Aktualne wartosci negatywne\", actual_1],\r\n [\"Przewidywane wartosci pozytywne\",(len(score)-predicted_1)],\r\n [\"Aktualne wartosci pozytywne\", (len(score)-actual_1)],\r\n [\"Iloć elementów\", len(score)],\r\n [\"Dokładnosc w %\",(((len(score)-abs(actual_1-predicted_1))/len(score))*100)]])\r\nprint (table.draw() + \"\\n\")\r\n \r\n\r\n#tworzenie wyników w formie graficznej\r\ndf = pd.DataFrame({'Wartosc': score[:,0]})\r\ndef getTextAnalysis(a):\r\n if a >= 0.5:\r\n return \"Negatywne\"\r\n else:\r\n return \"Pozytywne\"\r\ndf['Charakter'] = df['Wartosc'].apply(getTextAnalysis)\r\nlabels = df.groupby('Charakter').count().index.values\r\nvalues = df.groupby('Charakter').size().values\r\nplt.bar(labels, values)\r\nplt.suptitle('Ilosc negatywnych i pozytywnych Tweetów')\r\nplt.savefig('./wyniki/pozytywy_negatywy.png')\r\nplt.show()\r\n", "sub_path": "trening.py", "file_name": "trening.py", "file_ext": "py", "file_size_in_byte": 5949, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pandas.read_csv", "line_number": 19, "usage_type": "call"}, {"api_name": "wordcloud.WordCloud", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 49, "usage_type": "call"}, {"api_name": "tensorflow.data.Dataset.from_tensor_slices", "line_number": 56, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 56, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.experimental.preprocessing.TextVectorization", "line_number": 74, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 74, "usage_type": "attribute"}, {"api_name": "tensorflow.expand_dims", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 92, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 96, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.Sequential", "line_number": 101, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 101, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Embedding", "line_number": 103, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 103, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 104, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 104, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.GlobalAveragePooling1D", "line_number": 105, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 105, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dropout", "line_number": 106, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 106, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 107, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 107, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.callbacks.ModelCheckpoint", "line_number": 111, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 111, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.losses.BinaryCrossentropy", "line_number": 116, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 116, "usage_type": "attribute"}, {"api_name": "tensorflow.metrics.BinaryAccuracy", "line_number": 118, "usage_type": "call"}, {"api_name": "tensorflow.metrics", "line_number": 118, "usage_type": "attribute"}, {"api_name": "tensorflow.data.Dataset.from_tensor_slices", "line_number": 136, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 136, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.models.load_model", "line_number": 143, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 143, "usage_type": "attribute"}, {"api_name": "texttable.Texttable", "line_number": 159, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 173, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 182, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.suptitle", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 184, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 185, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 185, "usage_type": "name"}]} +{"seq_id": "382868639", "text": "# 开发者:朱梦婕\n# 开发日期:2020年6月11日\n# 开发框架:keras\n#----------------------------------------------------------#\n# ---------------------- 代码布局: ----------------------\n# 1、导入 Keras, matplotlib, numpy和os的包\n# 2、读取手写体数据及与图像预处理\n# 3、构建自编码器模型\n# 4、模型可视化\n# 5、训练\n# 6、查看解码效果\n# 7、训练过程可视化\n# ---------------------- 代码布局: ----------------------\n\n# -------------------------- 1、导入需要包 -------------------------------\nimport keras\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom keras.datasets import mnist\nfrom keras.models import Model\nfrom keras.layers import Input, add\nfrom keras.layers.convolutional import Conv2D, MaxPooling2D, UpSampling2D, ZeroPadding2D\nimport os\nos.environ['CUDA_VISIBLE_DEVICES']='3,4'\n# -------------------------- 1、导入需要包 -------------------------------\n\n# --------------------- 2、读取手写体数据及与图像预处理 ---------------------\n\n# 导入mnist数据\n# (X_train, _), (X_test, _) = mnist.load_data() 服务器无法访问\n# 本地读取数据\n# D:\\\\keras_datasets\\\\mnist.npz(本地路径)\npath = 'E:\\\\study\\\\kedata\\\\mnist.npz'\nf = np.load(path)\n#### 以npz结尾的数据集是压缩文件,里面还有其他的文件\n#### 使用:f.files 命令进行查看,输出结果为 ['x_test', 'x_train', 'y_train', 'y_test']\n# 60000个训练,10000个测试\n# 训练数据\nX_train=f['x_train']\n# 测试数据\nX_test=f['x_test']\nf.close()\n# 数据放到本地路径\n\n# 数据格式进行转换\nX_train = X_train.reshape(X_train.shape[0], 28, 28, 1)\nX_test = X_test.reshape(X_test.shape[0], 28, 28, 1)\n\n# 数据预处理\n# 归一化\nX_train = X_train.astype(\"float32\")/255.\nX_test = X_test.astype(\"float32\")/255.\n# 输出X_train和X_test维度\nprint('X_train shape:', X_train.shape)\nprint(X_train.shape[0], 'train samples')\nprint(X_test.shape[0], 'test samples')\n\n\n##### --------- 输出语句结果 --------\n# X_train shape: (60000, 28, 28, 1)\n# 60000 train samples\n# 10000 test samples\n##### --------- 输出语句结果 --------\n\n# --------------------- 2、读取手写体数据及与图像预处理 ---------------------\n\n\n# --------------------- 3、构建卷积自编码器模型 ---------------------\n\n# 输入维度为 1*28*28\ninputs = Input(shape=(28, 28,1))\n\nclass CNNAutoEncoder(keras.Model):\n def __init__(self):\n super(CNNAutoEncoder, self).__init__(name='CNNAutoEncoder_C')\n self.conv1_1 = Conv2D(16, (3, 3), activation='relu', padding='same') # 1*28*28 --> 16*28*28\n self.pool1 = MaxPooling2D((2, 2), padding='same') # 16*28*28 --> 16*14*14\n self.conv1_2 = Conv2D(8, (3, 3), activation='relu', padding='same') # 16*14*14 --> 8*14*14\n self.pool2 = MaxPooling2D((2, 2), padding='same') # 8*14*14 --> 8*7*7\n self.conv1_3 = Conv2D(8, (3, 3), activation='relu', padding='same') # 8*7*7 --> 8*7*7\n self.pool3 = MaxPooling2D((2, 2), padding='same') # 8*7*7 --> 8*4*4\n\n self.conv2_1 = Conv2D(8, (3, 3), activation='relu', padding='same') # 8*4*4 --> 8*4*4\n self.up1 = UpSampling2D((2, 2)) # 8*4*4 --> 8*8*8\n self.conv2_2 = Conv2D(8, (3, 3), activation='relu', padding='same') # 8*8*8 --> 8*8*8\n self.up2 = UpSampling2D((2, 2)) # 8*8*8 --> 8*16*16\n self.conv2_3 = Conv2D(16, (3, 3), activation='relu') # 8*16*16 --> 16*14*14 (not same)\n self.up3 = UpSampling2D((2, 2)) # 16*14*14 --> 16*28*28\n self.conv2_4 = Conv2D(1, (3, 3), activation='sigmoid', padding='same') # 16*28*28 --> 1*8*8\n\n def call(self, inputs, mask=None):\n # 编码器\n x = self.conv1_1(inputs)\n x = self.pool1(x)\n x = self.conv1_2(x)\n x = self.pool2(x)\n x = self.conv1_3(x)\n x = self.pool3(x)\n\n # 解码器\n x = self.conv2_1(x)\n x = self.up1(x)\n x = self.conv2_2(x)\n x = self.up2(x)\n x = self.conv2_3(x)\n x = self.up3(x)\n x = self.conv2_4(x)\n return x\n\nmodel = CNNAutoEncoder()\nmodel.compile(optimizer='adadelta', loss='binary_crossentropy')\n\n# --------------------- 3、构建卷积自编码器模型 ---------------------\n\n# --------------------- 4、模型可视化 ---------------------\n\nfrom IPython.display import SVG\nfrom keras.utils.vis_utils import model_to_dot\n\nSVG(model_to_dot(model).create(prog='dot', format='svg'))\n\n# --------------------- 4、模型可视化 ---------------------\n\n# --------------------- 5、训练 ---------------------\n\n# 设定peochs和batch_size大小\nepochs = 3\nbatch_size = 128\n\nhistory = model.fit(X_train, X_train,\n batch_size=batch_size,\n epochs=epochs, verbose=1,\n validation_data=(X_test, X_test)\n )\n\n# --------------------- 5、训练 ---------------------\n\n# --------------------- 6、查看解码效果 ---------------------\n\n# decoded_imgs 为输出层的结果\ndecoded_imgs = model.predict(X_test)\n\nn = 10\nplt.figure(figsize=(20, 6))\nfor i in range(n):\n # 原图\n ax = plt.subplot(3, n, i + 1)\n plt.imshow(X_test[i].reshape(28, 28))\n plt.gray()\n ax.get_xaxis().set_visible(False)\n ax.get_yaxis().set_visible(False)\n\n # 解码效果图\n ax = plt.subplot(3, n, i + n + 1)\n plt.imshow(decoded_imgs[i].reshape(28, 28))\n plt.gray()\n ax.get_xaxis().set_visible(False)\n ax.get_yaxis().set_visible(False)\n\nplt.show()\n\n# --------------------- 6、查看解码效果 ---------------------\n\n\n# --------------------- 7、训练过程可视化 ---------------------\n\nprint(history.history.keys())\n\nplt.plot(history.history['loss'])\nplt.plot(history.history['val_loss'])\nplt.title('model loss')\nplt.ylabel('loss')\nplt.xlabel('epoch')\nplt.legend(['train', 'validation'], loc='upper right')\nplt.show()\n\n# --------------------- 7、训练过程可视化 ---------------------", "sub_path": "zhumengjie/3.AutoEncoder/Keras/6.11 CNNAutoEncoder-Class.py", "file_name": "6.11 CNNAutoEncoder-Class.py", "file_ext": "py", "file_size_in_byte": 5986, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "os.environ", "line_number": 24, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 34, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 71, "usage_type": "call"}, {"api_name": "keras.Model", "line_number": 73, "usage_type": "attribute"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 76, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.MaxPooling2D", "line_number": 77, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 78, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.MaxPooling2D", "line_number": 79, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 80, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.MaxPooling2D", "line_number": 81, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 83, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.UpSampling2D", "line_number": 84, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 85, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.UpSampling2D", "line_number": 86, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 87, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.UpSampling2D", "line_number": 88, "usage_type": "call"}, {"api_name": "keras.layers.convolutional.Conv2D", "line_number": 89, "usage_type": "call"}, {"api_name": "IPython.display.SVG", "line_number": 120, "usage_type": "call"}, {"api_name": "keras.utils.vis_utils.model_to_dot", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 144, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 147, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 148, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 148, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gray", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 149, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 155, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gray", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 156, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 160, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 169, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 169, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 170, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 170, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 171, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 171, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 172, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 172, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 173, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 173, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 174, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 174, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 175, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 175, "usage_type": "name"}]} +{"seq_id": "161980139", "text": "from django.shortcuts import render\n\n# Create your views here.\n#from django.http import HttpResponse, JsonResponse # request của RFW extend HttpRequest\n#from django.views.decorators.csrf import csrf_exempt\n#from rest_framework.parsers import JSONParser\nfrom .models import Article\nfrom .serializers import ArticleSerializer\n\nfrom rest_framework import viewsets\nfrom rest_framework.decorators import api_view\nfrom rest_framework.response import Response\nfrom rest_framework import status\nfrom django.shortcuts import get_object_or_404\n\n#@csrf_exempt\n@api_view(['GET', 'POST'])\ndef article_list(request):\n \"\"\"\n List all code articles, or create a new Article.\n \"\"\"\n if request.method == 'GET':\n articles = Article.objects.all()\n serializer = ArticleSerializer(articles, many=True)\n return Response(serializer.data) #Response(serializer.data) k cần safe\n \n elif request.method == 'POST':\n #data = JSONParser().parse(request)\n serializer = ArticleSerializer(data=request.data)\n if serializer.is_valid():\n serializer.save()\n return Response(serializer.data, status=status.HTTP_201_CREATED)\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n#@csrf_exempt\n@api_view(['GET', 'PUT', 'DELETE'])\ndef article_detail(request, pk):\n \"\"\"\n Retrieve, update or delete article.\n \"\"\"\n try:\n article = Article.objects.get(pk=pk)\n except Article.DoesNotExist:\n return Response(status=status.HTTP_404_NOT_FOUND)\n \n if request.method == 'GET':\n serializer = ArticleSerializer(article)\n return Response(serializer.data)\n \n elif request.method == 'PUT':\n #data = JSONParser().parse(request)\n serializer = ArticleSerializer(article, data=request.data)\n if serializer.is_valid():\n serializer.save()\n return Response(serializer.data)\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n \n elif request.method == 'DELETE':\n article.delete()\n return Response(status=status.HTTP_204_NO_CONTENT)\n\n\nclass ArticleViewSet(viewsets.ViewSet):\n \n def list(self, request):\n articles = Article.objects.all()\n serializer = ArticleSerializer(articles, many=True)\n return Response(serializer.data)\n \n \n def create(self, request):\n serializer = ArticleSerializer(data=request.data)\n if serializer.is_valid():\n serializer.save()\n return Response(serializer.data, status=status.HTTP_201_CREATED)\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n \n \n def retrieve(self, request, pk=None):\n queryset = Article.objects.all()\n article = get_object_or_404(queryset, pk=pk)\n serializer = ArticleSerializer(article)\n return Response(serializer.data)", "sub_path": "api_ibroker/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2869, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "models.Article.objects.all", "line_number": 23, "usage_type": "call"}, {"api_name": "models.Article.objects", "line_number": 23, "usage_type": "attribute"}, {"api_name": "models.Article", "line_number": 23, "usage_type": "name"}, {"api_name": "serializers.ArticleSerializer", "line_number": 24, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 25, "usage_type": "call"}, {"api_name": "serializers.ArticleSerializer", "line_number": 29, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 32, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 32, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 32, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 33, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 33, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 33, "usage_type": "name"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 17, "usage_type": "call"}, {"api_name": "models.Article.objects.get", "line_number": 42, "usage_type": "call"}, {"api_name": "models.Article.objects", "line_number": 42, "usage_type": "attribute"}, {"api_name": "models.Article", "line_number": 42, "usage_type": "name"}, {"api_name": "models.Article.DoesNotExist", "line_number": 43, "usage_type": "attribute"}, {"api_name": "models.Article", "line_number": 43, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 44, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 44, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 44, "usage_type": "name"}, {"api_name": "serializers.ArticleSerializer", "line_number": 47, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 48, "usage_type": "call"}, {"api_name": "serializers.ArticleSerializer", "line_number": 52, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 55, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 56, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 56, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 56, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 60, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_204_NO_CONTENT", "line_number": 60, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 60, "usage_type": "name"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 36, "usage_type": "call"}, {"api_name": "rest_framework.viewsets.ViewSet", "line_number": 63, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 63, "usage_type": "name"}, {"api_name": "models.Article.objects.all", "line_number": 66, "usage_type": "call"}, {"api_name": "models.Article.objects", "line_number": 66, "usage_type": "attribute"}, {"api_name": "models.Article", "line_number": 66, "usage_type": "name"}, {"api_name": "serializers.ArticleSerializer", "line_number": 67, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 68, "usage_type": "call"}, {"api_name": "serializers.ArticleSerializer", "line_number": 72, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 75, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_201_CREATED", "line_number": 75, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 75, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 76, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_400_BAD_REQUEST", "line_number": 76, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 76, "usage_type": "name"}, {"api_name": "models.Article.objects.all", "line_number": 80, "usage_type": "call"}, {"api_name": "models.Article.objects", "line_number": 80, "usage_type": "attribute"}, {"api_name": "models.Article", "line_number": 80, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 81, "usage_type": "call"}, {"api_name": "serializers.ArticleSerializer", "line_number": 82, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 83, "usage_type": "call"}]} +{"seq_id": "48040399", "text": "from VaeEncoderDecoder_808_9 import VaeEncoderDecoder\nimport pypianoroll as ppr\nfrom pypianoroll import Track,Multitrack\nfrom utils import numpy_drums_save_to_midi\nfrom DrumsDataset import EmbeddingsDataset\nfrom DrumReducerExpander import DrumReducerExpander\n\n\n\nimport numpy as np\ndecoderVAE = VaeEncoderDecoder()\n\n\ndata=np.load('/home/ftamagna/Documents/_AcademiaSinica/dataset/drumGeneration/bigsupervised.npz')\n\ntr=data['track_array'][0]\n\nprint(tr.shape)\ntr_emb=decoderVAE.encode_to_embeddings(tr)\n\n\nmu=np.zeros((1000,32,1))\n# print(mu)\nsigma=np.ones((1000,32,1))\n\nlol=np.concatenate((mu,sigma),axis=2)\nprint(lol.shape,\"lol\")\n\ndecoder= DrumReducerExpander(drumpitches=9)\n\n# dataset=EmbeddingsDataset(lol)\n#\n# print(dataset)\n\n\n\ndrums_reduced=decoderVAE.decode_to_reduced_drums(lol)\n\n\n# print(drums_reduced)\nprint(drums_reduced.shape)\ndrums_reduced=drums_reduced>0.75\nfilepath='/home/ftamagna/Documents/_AcademiaSinica/dataset/temp/'\n\n# np.save(filepath+\"generated_with_normal\",drums_reduced)\n\nexpanded_drums=decoder.decode(drums_reduced)\nexpanded_drums=decoder.decode_808(expanded_drums)\n#\nfor i in range(len(expanded_drums)):\n numpy_drums_save_to_midi(np.concatenate((expanded_drums[i],expanded_drums[i])),filepath,str(i))\n\n\n\n", "sub_path": "drumGeneration/ScriptCheckVAEBasicGeneration_808_9_reconstruction.py", "file_name": "ScriptCheckVAEBasicGeneration_808_9_reconstruction.py", "file_ext": "py", "file_size_in_byte": 1232, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "VaeEncoderDecoder_808_9.VaeEncoderDecoder", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 26, "usage_type": "call"}, {"api_name": "DrumReducerExpander.DrumReducerExpander", "line_number": 29, "usage_type": "call"}, {"api_name": "utils.numpy_drums_save_to_midi", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 51, "usage_type": "call"}]} +{"seq_id": "72605772", "text": "import os\n\n# Librerias Django\nfrom django.db import models\nfrom django.utils.translation import gettext as _\nfrom django.conf import settings\nfrom django.conf.urls.static import static\nfrom django.utils.safestring import mark_safe\n\n\n# Librerias de terceros\n# Librerias locales\nfrom config.models import(TipoArticulo, TipoAtributo, CondicionVenta,\n UnidadMedida, FormaPago, DocumentoStatus, Pais)\nfrom empresa.models import (Empresa, Sucursal, Impuesto, Descuento)\nfrom cliente.models import(Cliente)\nfrom inventario.models import(Articulo, ArticuloCategoria, ArticuloVariante)\n# Validacion\nfrom .validators import *\nfrom empresa.validators import val_empresa\n\n# Create your models here.\n\n\"\"\" Modelo de Caja \"\"\"\n\n\nclass Caja(models.Model):\n sucursal = models.ForeignKey(Sucursal, on_delete=models.CASCADE)\n codigo = models.CharField(_('Código de Caja'), max_length=5, db_index=True, help_text=_(\n \"Codigo de la Caja Terminal o punto de venta\"))\n descripcion = models.CharField(\n _('Descripcion de Caja'), null=True, blank=True, max_length=20, db_index=True)\n activo = models.BooleanField(_('Activo'))\n\n # Metadata\n class Meta:\n db_table = 'empresa_sucursal_caja'\n verbose_name = \"Caja\"\n verbose_name_plural = \"Cajas\"\n unique_together = ('sucursal', 'codigo',)\n # Métodos\n\n def __str__(self):\n return '%s' % (self.codigo).upper()\n\n\n\"\"\" Modelo de Factura \"\"\"\n\n\nclass Factura(models.Model):\n \"\"\"\n Model Factura.\n\n - **parameters**:\n :empresa: Id Empresa que pertenece la Factura.\n :numero: numero correlactivo.\n :numero_control: Numero Control.\n \"\"\"\n empresa = models.ForeignKey(\n Empresa, on_delete=models.CASCADE, verbose_name=_(\"Empresa\"), db_index=True)\n numero = models.CharField(\n _('Nro. Factura'), max_length=20, db_index=True)\n numero_control = models.CharField(\n _('Nro. Control'), max_length=20, db_index=True, blank=True, null=True)\n cliente = models.ForeignKey(\n Cliente, on_delete=models.DO_NOTHING, verbose_name=\"Cliente\", default='0')\n # fecha=models.DateField('Fecha de Emision',auto_now=True)\n fecha = models.DateField('Fecha de Emisión')\n condicion_venta = models.ForeignKey(\n CondicionVenta, on_delete=models.CASCADE, verbose_name=_(\"Condicion de Venta\"), default='0')\n forma_pago = models.ForeignKey(\n FormaPago, on_delete=models.CASCADE, verbose_name=\"Forma de Pago\", default='0')\n status = models.ForeignKey(\n DocumentoStatus, on_delete=models.CASCADE, verbose_name=\"Status\", default='1')\n fecha_ven = models.DateField('Fecha de Vencimineto', auto_now=True)\n sucursal = models.ForeignKey(\n Sucursal, on_delete=models.DO_NOTHING, default=1, verbose_name=_(\"Sucursal\"))\n caja = models.ForeignKey(\n Caja, on_delete=models.DO_NOTHING, default=1, verbose_name=_(\"Caja\"))\n notas = models.TextField('Descripción', blank=True, null=True)\n base_imponible = models.FloatField(\"Base Imponible\", validators=[\n base_imponible_validation])\n impuesto = models.FloatField(\"Impuesto\", validators=[impuesto_validation])\n total = models.FloatField(\"Total\", validators=[total_validation])\n # https://tribunet.hacienda.go.cr/docs/esquemas/2016/v4.2/FacturaElectronica_V.4.2.xsd\n # PlazoCredito maxLength10\n # Metadata\n\n class Meta:\n db_table = 'factura'\n verbose_name = \"Factura de Venta\"\n verbose_name_plural = \"Facturas de Ventas\"\n unique_together = [('empresa', 'numero'),\n ('empresa', 'numero_control')]\n\n # Métodos\n def __str__(self):\n return '%s' % (self.numero)\n\n def cliente_link(self):\n display_text = \", \".join([\n \"{}\".format(\n reverse('admin:{}_{}_change'.format('ventas', 'cliente'),\n args=(self.cliente.id,)),\n self.cliente)\n\n ])\n if display_text:\n return mark_safe(display_text)\n return \"-\"\n cliente_link.allow_tags = True\n cliente_link.short_description = 'Cliente'\n\n def factura_link_xml(self):\n return mark_safe('XML' % ('/ventas/signed/', self.numero, self.numero))\n\n factura_link_xml.allow_tags = True\n factura_link_xml.short_description = 'XML'\n\n def count_len_numero(self):\n return len(self.numero)\n count_len_numero.allow_tags = True\n count_len_numero.short_description = 'Len Numero'\n\n def save_model(self, request, obj, form, change):\n super().save_model(request, obj, form, change)\n\n def save(self, *args, **kwargs):\n directory = 'ventas/signed/'\n if not os.path.exists(directory):\n os.makedirs(directory)\n f = open(directory+'factura'+str(self.numero)+'.xml', 'w+')\n text = ''\n text += ''\n text += ''+self.numero+''\n # empresa = Empresa.objects.get(id=1)\n\n text += ''+self.numero+''\n text += ''+str(self.fecha)+''\n ''''''''''''''''''''''''''''''''''''''''''''''''''''''\n text += ''\n text += ''+self.empresa.razon_social+''\n text += ''\n text += '' + \\\n str(self.empresa.denominacion_social.get_id())+''\n text += ''+str(self.empresa.rif)+''\n text += ''\n if self.empresa.nombre_comercial is not None:\n text += ''+self.empresa.nombre_comercial+''\n else:\n text += ''\n text += ''\n text += ' '+str(self.empresa.estado.id)+''\n text += ' '+str(self.empresa.municipio.id)+''\n text += ' '+str(self.empresa.parroquia.id)+''\n if self.empresa.barrio is not None:\n text += ' '+str(self.empresa.barrio.id)+''\n else:\n text += ' '\n\n text += ' '+str(self.empresa.direccion)+''\n text += ''\n text += ''\n text += ' ' + \\\n str(self.empresa.pais.telefono_codigo)+''\n text += ' '+str(self.empresa.telefono)+''\n text += ''\n\n if self.empresa.fax is not None:\n text += ''\n text += ' ' + \\\n str(self.empresa.pais.telefono_codigo)+''\n text += ' '+str(self.empresa.fax)+''\n text += ''\n if self.empresa.correo is not None:\n text += '' + \\\n str(self.empresa.correo)+''\n text += ''\n\n ''''''''''''''''''''''''''''''''''''\n text += ' '\n text += ' '+self.cliente.nombre+''\n text += ''\n text += ' ' + \\\n str(self.cliente.tipo_identificacion.codigo)+''\n text += ' '+self.cliente.nro_identificacion+''\n text += ''\n text += ''\n text += ''\n text += ' '+str(self.cliente.estado.id)+''\n text += ' '+str(self.cliente.municipio.id)+''\n text += ' '+str(self.cliente.parroquia.id)+''\n if self.cliente.barrio is not None:\n text += ' '+str(self.cliente.barrio.id)+''\n\n text += ' ' + \\\n str(self.cliente.direccion)+''\n text += ''\n\n if self.cliente.telefono is not None:\n text += ''\n text += ' ' + \\\n str(self.cliente.pais.telefono_codigo)+''\n text += ' ' + \\\n str(self.cliente.telefono)+''\n text += ''\n\n if self.cliente.fax is not None:\n text += ''\n text += '' + \\\n str(self.cliente.pais.telefono_codigo)+''\n text += ''+str(self.cliente.fax)+''\n text += ''\n text += '' + \\\n str(self.cliente.correo)+''\n text += ''\n\n text += '' + \\\n str(self.condicion_venta.get_id())+''\n text += 'xxxxxxxxxxxxxx'\n text += ''+str(self.forma_pago.get_id())+''\n\n items = self.facturaitems_set.all()\n text += ''\n i = sumimpuesto = sumdescuento = sumexento = sumgravada = 0\n summercagravada = sumservgravada = sumservexenta = summercaexenta = 0\n\n for item in items:\n montototal = item.cantidad * item.precio_unitario\n i += 1\n text += ''\n text += ' '+str(i)+''\n text += ' '\n text += ' 04'\n text += ' ' + str(item.articulo.id)+''\n text += ' '\n text += ' ' + str(item.cantidad)+''\n text += ' ' + \\\n str(item.unidad_medida)+''\n text += ' '\n text += ' '+str(item.articulo)+''\n text += ' ' + \\\n str(item.precio_unitario)+''\n text += ' '+str(montototal)+''\n\n if item.descuento > 0:\n sumdescuento = sumdescuento+item.descuento\n text += '' + \\\n str(item.descuento)+''\n text += '' + \\\n str(item.valor_descuento.nombre) + \\\n ' '\n\n text += ' ' + \\\n str(montototal-item.descuento)+''\n\n if item.impuesto > 0:\n sumimpuesto = float(sumimpuesto) + float(item.impuesto)\n sumgravada += montototal\n text += ''+str(item.impuesto)+''\n text += ' ' + \\\n str(float(item.subtotal+item.impuesto)) + \\\n ''\n if item.articulo.tipo.id == 1:\n summercagravada += montototal\n elif item.articulo.tipo.id == 2:\n sumservgravada += montototal\n else:\n sumexento = sumexento+montototal\n if item.articulo.tipo.id == 1:\n summercaexenta += montototal\n elif item.articulo.tipo.id == 2:\n sumservexenta += montototal\n text += ' ' + \\\n str(item.subtotal)+''\n text += ''\n\n text += ''\n ''''''''''''''''''''''''''''''''''''\n text += ''\n text += '' + \\\n str(self.empresa.moneda.codigo_iso)+''\n text += '576.74000'\n text += '' + \\\n str(sumservgravada)+''\n text += ''+str(sumservexenta)+''\n text += '' + \\\n str(summercagravada)+''\n text += '' + \\\n str(summercaexenta)+''\n text += ''+str(sumgravada)+''\n text += ''+str(sumexento)+''\n text += ''+str(float(sumgravada+sumexento))+''\n text += ''+str(sumdescuento)+''\n text += '' + \\\n str(float(sumgravada+sumexento-sumdescuento))+''\n text += ''+str(float(sumimpuesto))+''\n text += '' + \\\n str(float(sumgravada+sumexento+sumimpuesto-sumdescuento)) + \\\n ''\n text += ''\n text += ''\n text += ' DGT-R-48-2016'\n text += ' 20-02-2017 13:22:22'\n text += ''\n text += ''\n text += 'BNCR $ 200-'\n text += ''\n text += ''\n\n f.write(text)\n f.close()\n super(Factura, self).save(*args, **kwargs)\n\n\n\"\"\" Modelo de FacturaItems \"\"\"\n\n\nclass FacturaItems(models.Model):\n # empresa=models.ForeignKey(Empresa,on_delete=models.CASCADE,verbose_name=_(\"Empresa\"),db_index=True)\n numero = models.ForeignKey(Factura, on_delete=models.CASCADE, verbose_name=_(\n \"Nro. de Factura\"), db_index=True)\n articulo = models.ForeignKey(\n Articulo, on_delete=models.CASCADE, verbose_name=_(\"Artículo\"))\n cantidad = models.FloatField(\"Cantidad\")\n unidad_medida = models.ForeignKey(\n UnidadMedida, on_delete=models.DO_NOTHING, verbose_name=_(\"Unidad de Medida\"), default=1)\n precio_unitario = models.FloatField(_(\"Precio Unitario\"))\n valor_impuesto = models.ForeignKey(\n Impuesto, on_delete=models.CASCADE, verbose_name=_(\"Impuesto\"), blank=True, null=True, db_index=True)\n impuesto = models.FloatField(_(\"Monto del Impuesto\"), default=0)\n valor_descuento = models.ForeignKey(\n Descuento, on_delete=models.CASCADE, verbose_name=_(\"Descuento\"), blank=True, null=True, db_index=True)\n descuento = models.FloatField(_(\"Monto del Descuento\"), default=0)\n subtotal = models.FloatField(_(\"Subtotal\"), validators=[items_validation])\n # NaturalezaDescuento\n # MontoDescuento\n\n def __str__(self):\n return '%s' % (self.articulo)\n\n def sum_impuesto(self, request):\n return FacturaItems.objects.aggregate(total=Sum('impuesto'))\n\n class Meta:\n db_table = 'factura_items'\n verbose_name = \"Items Factura de Venta\"\n verbose_name_plural = \"Items Facturas de Ventas\"\n\n\nclass FacturaCaja(models.Model):\n empresa = models.ForeignKey(Empresa, on_delete=models.CASCADE, verbose_name=_(\n \"Empresa/Sucursal\"), db_index=True, default=1)\n TIPO_DOCUMENTO = (\n ('01', 'Factura'),\n ('02', 'Nota de Debito'),\n ('03', 'Nota de Credito'),\n ('04', 'Tiquete'),\n ('05', 'Nota de despacho'),\n ('06', 'Contrato'),\n ('07', 'Procedimiento'),\n ('08', 'Comprobante emitido en contingencia'),\n ('99', 'Otros'),\n\n )\n sucursal = models.ForeignKey(\n Sucursal, on_delete=models.CASCADE, verbose_name=_(\"Sucursal\"), db_index=True)\n documento = models.CharField(_('Tipo de Documento'), db_index=True,\n max_length=2, choices=TIPO_DOCUMENTO, help_text=_('Tipo de Documento'))\n numero = models.CharField(\n _('Nro. Documento'), max_length=10, db_index=True)\n\n def __str__(self):\n return '%s %s %s' % (self.sucursal, self.documento, self.numero)\n\n class Meta:\n db_table = 'factura_caja'\n verbose_name = \"Factura en caja\"\n", "sub_path": "ventas/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 16150, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "django.db.models.Model", "line_number": 27, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 27, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 28, "usage_type": "call"}, {"api_name": "empresa.models.Sucursal", "line_number": 28, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 28, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 28, "usage_type": "attribute"}, {"api_name": "django.db.models.CharField", "line_number": 29, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 29, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 29, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 31, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 31, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 32, "usage_type": "call"}, {"api_name": "django.db.models.BooleanField", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 33, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 50, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 50, "usage_type": "name"}, {"api_name": "empresa.models", "line_number": 59, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 59, "usage_type": "call"}, {"api_name": "empresa.models.Empresa", "line_number": 60, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 59, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 60, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 60, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 60, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 61, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 61, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 62, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 63, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 63, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 64, "usage_type": "call"}, {"api_name": "cliente.models", "line_number": 65, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 65, "usage_type": "call"}, {"api_name": "cliente.models.Cliente", "line_number": 66, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 65, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 66, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 66, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 68, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 68, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 69, "usage_type": "call"}, {"api_name": "config.models.CondicionVenta", "line_number": 70, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 69, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 70, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 70, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 70, "usage_type": "call"}, {"api_name": "django.db.models.ForeignKey", "line_number": 71, "usage_type": "call"}, {"api_name": "config.models.FormaPago", "line_number": 72, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 71, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 72, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 72, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 73, "usage_type": "call"}, {"api_name": "config.models.DocumentoStatus", "line_number": 74, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 73, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 74, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 74, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 75, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 75, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 76, "usage_type": "call"}, {"api_name": "empresa.models.Sucursal", "line_number": 77, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 76, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 77, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 77, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 77, "usage_type": "call"}, {"api_name": "django.db.models.ForeignKey", "line_number": 78, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 78, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 79, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 79, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 79, "usage_type": "call"}, {"api_name": "django.db.models.TextField", "line_number": 80, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 80, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 81, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 81, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 83, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 83, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 84, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 84, "usage_type": "name"}, {"api_name": "django.utils.safestring.mark_safe", "line_number": 109, "usage_type": "call"}, {"api_name": "django.utils.safestring.mark_safe", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 130, "usage_type": "call"}, {"api_name": "os.path", "line_number": 130, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 131, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 319, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 319, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 321, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 321, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 321, "usage_type": "attribute"}, {"api_name": "django.utils.translation.gettext", "line_number": 321, "usage_type": "call"}, {"api_name": "django.db.models.ForeignKey", "line_number": 323, "usage_type": "call"}, {"api_name": "inventario.models.Articulo", "line_number": 324, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 323, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 324, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 324, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 324, "usage_type": "call"}, {"api_name": "django.db.models.FloatField", "line_number": 325, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 325, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 326, "usage_type": "call"}, {"api_name": "config.models.UnidadMedida", "line_number": 327, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 326, "usage_type": "name"}, {"api_name": "django.db.models.DO_NOTHING", "line_number": 327, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 327, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 327, "usage_type": "call"}, {"api_name": "django.db.models.FloatField", "line_number": 328, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 328, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 328, "usage_type": "call"}, {"api_name": "django.db.models.ForeignKey", "line_number": 329, "usage_type": "call"}, {"api_name": "empresa.models.Impuesto", "line_number": 330, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 329, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 330, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 330, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 330, "usage_type": "call"}, {"api_name": "django.db.models.FloatField", "line_number": 331, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 331, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 331, "usage_type": "call"}, {"api_name": "django.db.models.ForeignKey", "line_number": 332, "usage_type": "call"}, {"api_name": "empresa.models.Descuento", "line_number": 333, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 332, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 333, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 333, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 333, "usage_type": "call"}, {"api_name": "django.db.models.FloatField", "line_number": 334, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 334, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 334, "usage_type": "call"}, {"api_name": "django.db.models.FloatField", "line_number": 335, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 335, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 335, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 351, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 351, "usage_type": "name"}, {"api_name": "empresa.models", "line_number": 352, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 352, "usage_type": "call"}, {"api_name": "empresa.models.Empresa", "line_number": 352, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 352, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 352, "usage_type": "attribute"}, {"api_name": "django.utils.translation.gettext", "line_number": 352, "usage_type": "call"}, {"api_name": "django.db.models.ForeignKey", "line_number": 366, "usage_type": "call"}, {"api_name": "empresa.models.Sucursal", "line_number": 367, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 366, "usage_type": "name"}, {"api_name": "django.db.models.CASCADE", "line_number": 367, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 367, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 367, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 368, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 368, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 368, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext", "line_number": 369, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 370, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 370, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext", "line_number": 371, "usage_type": "call"}]} +{"seq_id": "294712744", "text": "\"\"\"deleted op\n\nRevision ID: d354fd98ff7b\nRevises: ff0271c6d3b7\nCreate Date: 2019-12-18 20:20:08.638048\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\n\n\n# revision identifiers, used by Alembic.\nrevision = 'd354fd98ff7b'\ndown_revision = 'ff0271c6d3b7'\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n # ### commands auto generated by Alembic - please adjust! ###\n with op.batch_alter_table('transactions') as batch_op:\n batch_op.drop_column('op')\n # ### end Alembic commands ###\n\n\ndef downgrade():\n # ### commands auto generated by Alembic - please adjust! ###\n op.add_column('transactions', sa.Column('op', sa.SMALLINT(), nullable=True))\n # ### end Alembic commands ###\n", "sub_path": "migrations/versions/d354fd98ff7b_deleted_op.py", "file_name": "d354fd98ff7b_deleted_op.py", "file_ext": "py", "file_size_in_byte": 707, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "alembic.op.batch_alter_table", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 21, "usage_type": "name"}, {"api_name": "alembic.op.add_column", "line_number": 28, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 28, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 28, "usage_type": "call"}, {"api_name": "sqlalchemy.SMALLINT", "line_number": 28, "usage_type": "call"}]} +{"seq_id": "200440019", "text": "from . import content_api\nfrom flask import Response, json\n\n@content_api.errorhandler(500)\ndef server_error(error):\n \"\"\"\n Error 500 status code custom handler for content blueprint\n \"\"\"\n message = {'error':'An error occurred, the server could not process your request'}\n return Response(\n mimetype=\"application/json\",\n response=json.dumps(message),\n status=500\n )\n", "sub_path": "src/Views/ContentView/error.py", "file_name": "error.py", "file_ext": "py", "file_size_in_byte": 379, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "flask.Response", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.json.dumps", "line_number": 12, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 12, "usage_type": "name"}]} +{"seq_id": "359464259", "text": "import numpy as np\nimport scipy as sci\nimport matplotlib.pyplot as plt\n\n\nclass People:\n \"\"\"\n This class defines the people in the room.\n \"\"\"\n\n # define class attributes\n # Parameters\n SD = 350 # surface density is defined to be 350 kg/m^2\n\n # parameters in computing forces\n A = 2000\n B = 0.08\n k = 120000\n kappa = 240000\n tau = 0.5 # characteristic time\n v_des = 0.8 # desired speed\n\n def __init__(self, vec_r, vec_v, r_i):\n \"\"\"\n vec_r: position 2-vector\n vec_v: velocity 2-vector\n radius: radius of a people\n m: m is mass\n \"\"\"\n self.vec_r = vec_r\n self.vec_v = vec_v\n self.r_i = r_i\n self.m = self.SD * (\n np.pi * r_i ** 2\n ) # here self.SD access to a class attribute\n\n # the following 2 attributes are for visualizing\n # default circle styles\n self.styles = {\"edgecolor\": \"b\", \"fill\": True}\n # create a circle representation of itself\n self.circle = plt.Circle(self.vec_r, self.r_i, **self.styles)\n\n def move(self, vec_Fi):\n self.vec_r += self.vec_v\n # update its circle representation\n self.circle.center = (self.vec_r[0], self.vec_r[1])\n\n def _F_from_self(self, vec_ei):\n \"\"\"\n Compute the force from v_des\n ei: desired direction\n \"\"\"\n vec_F_des = self.m * (self.v_des * vec_ei - self.vec_v) / self.tau\n return vec_F_des\n\n g = lambda x: np.max([0, x]) # this is a class attribute\n\n def F_from_other(self, other):\n \"\"\"\n Compute the force from another people, i.e. f_ij\n other: another people\n \"\"\"\n\n r_ij = self.r_i + other.r_i\n d_ij = sci.linalg.norm(self.vec_r - other.vec_r)\n vec_n_ij = (self.vec_r - other.vec_r) / d_ij\n vec_t_ij = np.array([-vec_n_ij[1], vec_n_ij[0]])\n delta_v_ji = np.dot(other.vec_v - self.vec_v, vec_t_ij)\n\n vec_F_ij = (\n self.A * np.exp(r_ij - d_ij) / self.B + self.k * self.g(r_ij - d_ij)\n ) * vec_n_ij + self.kappa * g(r_ij - d_ij) * delta_v_ji * vec_t_ij\n return vec_F_ij\n\n def F_from_wall(self, wall):\n \"\"\"\n Compute force from wall\n \"\"\"\n\n d_iW = sci.linalg.norm(wall.b - self.vec_r[0]) # this computes the distance\n vec_n_iW = (self.vec_r - np.array([wall.b, self.vec_r[1]])) / d_iW\n vec_t_iW = np.array([-vec_n_iW[1], vec_n_iW[0]])\n\n vec_F_iW = (\n (\n self.A * np.exp(self.r_i - d_iW) / self.B\n + self.k * self.g(self.r_i - d_iW)\n )\n * vec_n_iW\n - self.kappa\n * self.g(self.r_i - d_iW)\n * np.dot(self.vec_v, vec_t_iW)\n * vec_t_iW\n )\n\n return vec_F_iW\n\n # migh have problem\n def draw(self):\n \"\"\"return its current circle representation\"\"\"\n return self.circle\n", "sub_path": "Final Project/escapist2/people_copy.py", "file_name": "people_copy.py", "file_ext": "py", "file_size_in_byte": 2925, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "numpy.pi", "line_number": 34, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.Circle", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 56, "usage_type": "call"}, {"api_name": "scipy.linalg.norm", "line_number": 65, "usage_type": "call"}, {"api_name": "scipy.linalg", "line_number": 65, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 71, "usage_type": "call"}, {"api_name": "scipy.linalg.norm", "line_number": 80, "usage_type": "call"}, {"api_name": "scipy.linalg", "line_number": 80, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 92, "usage_type": "call"}]} +{"seq_id": "275644395", "text": "#!/usr/bin/env python3\n\n\"\"\"Demonstration Script that extracts agent data from cache directory files.\n\nThis could be a modified to be a daemon\n\n\"\"\"\n\n# Standard libraries\nimport os\nimport time\nimport shutil\nfrom collections import defaultdict\nimport queue as Queue\nimport threading\nimport re\n\n# Infoset libraries\nfrom infoset.db import db\nfrom infoset.db import db_agent as agent\nfrom infoset.utils import log\nfrom infoset.cache import drain\n\n# Define a key global variable\nTHREAD_QUEUE = Queue.Queue()\n\n\nclass FillDB(threading.Thread):\n \"\"\"Threaded polling.\n\n Graciously modified from:\n http://www.ibm.com/developerworks/aix/library/au-threadingpython/\n\n \"\"\"\n\n def __init__(self, queue):\n \"\"\"Initialize the threads.\"\"\"\n threading.Thread.__init__(self)\n self.queue = queue\n\n def run(self):\n \"\"\"Update the database using threads.\"\"\"\n while True:\n # Get the data_dict\n data_dict = self.queue.get()\n uid = data_dict['uid']\n metadata = data_dict['metadata']\n config = data_dict['config']\n agents = data_dict['agents']\n datapoints = data_dict['datapoints']\n\n # Initialize other values\n max_timestamp = 0\n\n # Sort metadata by timestamp\n metadata.sort()\n\n # Process file for each timestamp\n for (timestamp, filepath) in metadata:\n # Read in data\n ingest = drain.Drain(filepath)\n\n # Make sure file is OK\n # Move it to a directory for further analysis\n # by administrators\n if ingest.valid() is False:\n log_message = (\n 'Cache ingest file %s is invalid. Moving.'\n '') % (filepath)\n log.log2warn(1054, log_message)\n shutil.move(\n filepath, config.ingest_failures_directory())\n continue\n\n # Update agent table if not there\n if ingest.uid() not in agents:\n _insert_agent(\n ingest.uid(),\n ingest.agent(),\n ingest.hostname(),\n config\n )\n # Append the new insertion to the list\n agents.append(ingest.uid())\n\n # Update datapoint metadata if not there\n for item in ingest.sources():\n did = item[1]\n if did not in datapoints:\n _insert_datapoint(item, config)\n # Append the new insertion to the list\n datapoints.append(did)\n\n # Create map of DIDs to database row index values\n mapping = _datapoints_by_did(config)\n\n # Update chartable data\n _update_chartable(mapping, ingest, config)\n _update_unchartable(mapping, ingest, config)\n\n # Get the max timestamp\n max_timestamp = max(timestamp, max_timestamp)\n\n # Purge source file\n ingest.purge()\n\n # Update the last time the agent was contacted\n _update_agent_last_update(uid, max_timestamp, config)\n\n # All done!\n self.queue.task_done()\n\n\ndef _update_chartable(mapping, ingest, config):\n \"\"\"Insert data into the database \"iset_data\" table.\n\n Args:\n mapping: Map of DIDs to database row index values\n ingest: Drain object\n config: Config object\n\n Returns:\n None\n\n \"\"\"\n # Initialize key variables\n data = ingest.chartable()\n data_list = []\n timestamp_tracker = {}\n\n # Update data\n for item in data:\n # Process each datapoint item found\n (_, did, tuple_value, timestamp) = item\n idx_datapoint = int(mapping[did][0])\n idx_agent = int(mapping[did][1])\n last_timestamp = int(mapping[did][2])\n value = float(tuple_value)\n\n # Only update with data collected after\n # the most recent update. Don't do anything more\n if timestamp > last_timestamp:\n data_list.append(\n (idx_datapoint, idx_agent, value, timestamp)\n )\n\n # Update DID's last updated timestamp\n if idx_datapoint in timestamp_tracker:\n timestamp_tracker[idx_datapoint] = max(\n timestamp, timestamp_tracker[idx_datapoint])\n else:\n timestamp_tracker[idx_datapoint] = timestamp\n\n # Update if there is data\n if bool(data_list) is True:\n # Prepare SQL query to read a record from the database.\n sql_insert = (\n 'REPLACE INTO iset_data '\n '(idx_datapoint, idx_agent, value, timestamp) VALUES '\n '(%s, %s, %s, %s)')\n\n # Do query and get results\n database = db.Database(config)\n database.modify(sql_insert, 1056, data_list=data_list)\n\n # Change the last updated timestamp\n for idx_datapoint, last_timestamp in timestamp_tracker.items():\n # Prepare SQL query to read a record from the database.\n sql_modify = (\n 'UPDATE iset_datapoint SET last_timestamp=%s '\n 'WHERE iset_datapoint.idx=%s'\n '') % (last_timestamp, idx_datapoint)\n database.modify(sql_modify, 1057)\n\n # Report success\n log_message = (\n 'Successful cache drain for UID %s at timestamp %s') % (\n ingest.uid(), ingest.timestamp())\n log.log2quiet(1058, log_message)\n\n\ndef _update_unchartable(mapping, ingest, config):\n \"\"\"Update unchartable data into the database \"iset_datapoint\" table.\n\n Args:\n mapping: Map of DIDs to database row index values\n ingest: Drain object\n config: Config object\n\n Returns:\n None\n\n \"\"\"\n # Initialize key variables\n data = ingest.other()\n data_list = []\n timestamp_tracker = {}\n\n # Update data\n for item in data:\n # Process each datapoint item found\n (_, did, tuple_value, timestamp) = item\n idx_datapoint = int(mapping[did][0])\n last_timestamp = int(mapping[did][2])\n value = ('%s') % (tuple_value)\n\n # Only update with data collected after\n # the most recent update. Don't do anything more\n if timestamp > last_timestamp:\n data_list.append(\n (idx_datapoint, value)\n )\n\n # Update DID's last updated timestamp\n if idx_datapoint in timestamp_tracker:\n timestamp_tracker[idx_datapoint] = max(\n timestamp, timestamp_tracker[idx_datapoint])\n else:\n timestamp_tracker[idx_datapoint] = timestamp\n\n # Update if there is data\n if bool(data_list) is True:\n for item in data_list:\n (idx_datapoint, value) = item\n fixed_value = str(value)[0:128]\n\n # Prepare SQL query to read a record from the database.\n sql_modify = (\n 'UPDATE iset_datapoint set uncharted_value=\"%s\" WHERE '\n 'idx=%s') % (fixed_value, idx_datapoint)\n\n # Do query and get results\n database = db.Database(config)\n database.modify(sql_modify, 1037)\n\n # Change the last updated timestamp\n for idx_datapoint, last_timestamp in timestamp_tracker.items():\n # Prepare SQL query to read a record from the database.\n sql_modify = (\n 'UPDATE iset_datapoint SET last_timestamp=%s '\n 'WHERE iset_datapoint.idx=%s'\n '') % (last_timestamp, idx_datapoint)\n database.modify(sql_modify, 1044)\n\n # Report success\n log_message = (\n 'Successful cache drain (Uncharted Data) '\n 'for UID %s at timestamp %s') % (\n ingest.uid(), ingest.timestamp())\n log.log2quiet(1045, log_message)\n\n\ndef _update_agent_last_update(uid, last_timestamp, config):\n \"\"\"Insert new datapoint into database.\n\n Args:\n uid: UID of agent\n last_timestamp: The last time a DID for the agent was updated\n in the database\n config: Config object\n\n Returns:\n None\n\n \"\"\"\n # Initialize key variables\n sql_modify = (\n 'UPDATE iset_agent SET iset_agent.last_timestamp=%s '\n 'WHERE iset_agent.id=\"%s\"'\n '') % (last_timestamp, uid)\n database = db.Database(config)\n database.modify(sql_modify, 1055)\n\n\ndef _insert_datapoint(metadata, config):\n \"\"\"Insert new datapoint into database.\n\n Args:\n metadata: Tuple of datapoint metadata.\n (uid, did, label, source, description)\n uid: Agent UID\n did: Datapoint ID\n label: Datapoint label created by agent\n source: Source of the data (subsystem being tracked)\n description: Description provided by agent config file (unused)\n base_type = SNMP base type (Counter32, Counter64, Gauge etc.)\n config: Configuration object\n\n Returns:\n None\n\n \"\"\"\n # Initialize key variables\n (uid, did, label, source, _, base_type) = metadata\n\n # Get agent index value\n agent_object = agent.Get(uid, config)\n idx_agent = agent_object.idx()\n\n # Prepare SQL query to read a record from the database.\n sql_query = (\n 'INSERT INTO iset_datapoint '\n '(id, idx_agent, agent_label, agent_source, base_type ) VALUES '\n '(\"%s\", %d, \"%s\", \"%s\", %d)'\n '') % (did, idx_agent, label, source, base_type)\n\n # Do query and get results\n database = db.Database(config)\n database.modify(sql_query, 1032)\n\n\ndef _insert_agent(uid, name, hostname, config):\n \"\"\"Insert new agent into database.\n\n Args:\n uid: Agent uid\n name: Agent name\n Hostname: Hostname the agent gets data from\n config: Configuration object\n\n Returns:\n None\n\n \"\"\"\n # Prepare SQL query to read a record from the database.\n sql_query = (\n 'INSERT INTO iset_agent (id, name, hostname) '\n 'VALUES (\"%s\", \"%s\", \"%s\")'\n '') % (uid, name, hostname)\n\n # Do query and get results\n database = db.Database(config)\n database.modify(sql_query, 1033)\n\n\ndef _datapoints(config):\n \"\"\"Create list of enabled datapoints.\n\n Args:\n config: Configuration object\n\n Returns:\n data: List of active datapoints\n\n \"\"\"\n # Initialize key variables\n data = []\n\n # Prepare SQL query to read a record from the database.\n sql_query = (\n 'SELECT iset_datapoint.id '\n 'FROM iset_datapoint WHERE (iset_datapoint.enabled=1)')\n\n # Do query and get results\n database = db.Database(config)\n query_results = database.query(sql_query, 1034)\n\n # Massage data\n for row in query_results:\n data.append(row[0])\n\n # Return\n return data\n\n\ndef _datapoints_by_did(config):\n \"\"\"Create dict of enabled datapoints and their corresponding indices.\n\n Args:\n config: Configuration object\n\n Returns:\n data: Dict keyed by datapoint ID,\n with a tuple as its value (idx, idx_agent)\n idx: Datapoint index\n idx_agent: Agent index\n last_timestamp: The last time the timestamp was updated\n\n \"\"\"\n # Initialize key variables\n data = {}\n\n # Prepare SQL query to read a record from the database.\n sql_query = (\n 'SELECT iset_datapoint.id, iset_datapoint.idx, '\n 'iset_datapoint.idx_agent, iset_datapoint.last_timestamp '\n 'FROM iset_datapoint WHERE (iset_datapoint.enabled=1)')\n\n # Do query and get results\n database = db.Database(config)\n query_results = database.query(sql_query, 1035)\n\n # Massage data\n for row in query_results:\n did = row[0]\n idx = row[1]\n idx_agent = row[2]\n last_timestamp = row[3]\n data[did] = (idx, idx_agent, last_timestamp)\n\n # Return\n return data\n\n\ndef _agents(config):\n \"\"\"Create list of active agent UIDs.\n\n Args:\n config: Configuration object\n\n Returns:\n data: List of active agents\n\n \"\"\"\n # Initialize key variables\n data = []\n\n # Prepare SQL query to read a record from the database.\n sql_query = (\n 'SELECT iset_agent.id '\n 'FROM iset_agent WHERE (iset_agent.enabled=1)')\n\n # Do query and get results\n database = db.Database(config)\n query_results = database.query(sql_query, 1036)\n\n # Massage data\n for row in query_results:\n data.append(row[0])\n\n # Return\n return data\n\n\ndef process(config):\n \"\"\"Method initializing the class.\n\n Args:\n config: Configuration object\n\n Returns:\n None\n\n \"\"\"\n # Initialize key variables\n threads_in_pool = config.ingest_threads()\n uid_metadata = defaultdict(lambda: defaultdict(dict))\n cache_dir = config.ingest_cache_directory()\n\n # Filenames must start with a numeric timestamp and #\n # end with a hex string. This will be tested later\n regex = re.compile(r'^\\d+_[0-9a-f]+.json')\n\n # Get a list of active agents and datapoints\n agents = _agents(config)\n datapoints = _datapoints(config)\n\n # Spawn a pool of threads, and pass them queue instance\n for _ in range(threads_in_pool):\n update_thread = FillDB(THREAD_QUEUE)\n update_thread.daemon = True\n update_thread.start()\n\n # Add files in cache directory to list\n all_filenames = [filename for filename in os.listdir(\n cache_dir) if os.path.isfile(\n os.path.join(cache_dir, filename))]\n\n ######################################################################\n # Create threads\n ######################################################################\n\n # Process only valid agent filenames\n for filename in all_filenames:\n # Add valid data to lists\n if bool(regex.match(filename)) is True:\n # Create a complete filepath\n filepath = os.path.join(cache_dir, filename)\n\n # Only read files that are 15 seconds or older\n # to prevent corruption caused by reading a file that could be\n # updating simultaneously\n if time.time() - os.path.getmtime(filepath) < 15:\n continue\n\n # Create a dict of UIDs, timestamps and filepaths\n (name, _) = filename.split('.')\n (tstamp, uid) = name.split('_')\n timestamp = int(tstamp)\n if uid in uid_metadata:\n uid_metadata[uid].append(\n (timestamp, filepath))\n else:\n uid_metadata[uid] = [(timestamp, filepath)]\n\n # Read each cache file\n for uid in uid_metadata.keys():\n\n ####################################################################\n #\n # Define variables that will be required for the threading\n # We have to initialize the dict during every loop to prevent\n # data corruption\n #\n ####################################################################\n data_dict = {}\n data_dict['uid'] = uid\n data_dict['metadata'] = uid_metadata[uid]\n data_dict['config'] = config\n data_dict['agents'] = agents\n data_dict['datapoints'] = datapoints\n THREAD_QUEUE.put(data_dict)\n\n # Wait on the queue until everything has been processed\n THREAD_QUEUE.join()\n\n # PYTHON BUG. Join can occur while threads are still shutting down.\n # This can create spurious \"Exception in thread (most likely raised\n # during interpreter shutdown)\" errors.\n # The \"time.sleep(1)\" adds a delay to make sure things really terminate\n # properly. This seems to be an issue on virtual machines in Dev only\n time.sleep(1)\n", "sub_path": "infoset/cache/cache.py", "file_name": "cache.py", "file_ext": "py", "file_size_in_byte": 15892, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "queue.Queue", "line_number": 25, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 28, "usage_type": "attribute"}, {"api_name": "threading.Thread.__init__", "line_number": 38, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 38, "usage_type": "attribute"}, {"api_name": "infoset.cache.drain.Drain", "line_number": 61, "usage_type": "call"}, {"api_name": "infoset.cache.drain", "line_number": 61, "usage_type": "name"}, {"api_name": "infoset.utils.log.log2warn", "line_number": 70, "usage_type": "call"}, {"api_name": "infoset.utils.log", "line_number": 70, "usage_type": "name"}, {"api_name": "shutil.move", "line_number": 71, "usage_type": "call"}, {"api_name": "infoset.db.db.Database", "line_number": 163, "usage_type": "call"}, {"api_name": "infoset.db.db", "line_number": 163, "usage_type": "name"}, {"api_name": "infoset.utils.log.log2quiet", "line_number": 179, "usage_type": "call"}, {"api_name": "infoset.utils.log", "line_number": 179, "usage_type": "name"}, {"api_name": "infoset.db.db.Database", "line_number": 233, "usage_type": "call"}, {"api_name": "infoset.db.db", "line_number": 233, "usage_type": "name"}, {"api_name": "infoset.utils.log.log2quiet", "line_number": 250, "usage_type": "call"}, {"api_name": "infoset.utils.log", "line_number": 250, "usage_type": "name"}, {"api_name": "infoset.db.db.Database", "line_number": 271, "usage_type": "call"}, {"api_name": "infoset.db.db", "line_number": 271, "usage_type": "name"}, {"api_name": "infoset.db.db_agent.Get", "line_number": 297, "usage_type": "call"}, {"api_name": "infoset.db.db_agent", "line_number": 297, "usage_type": "name"}, {"api_name": "infoset.db.db.Database", "line_number": 308, "usage_type": "call"}, {"api_name": "infoset.db.db", "line_number": 308, "usage_type": "name"}, {"api_name": "infoset.db.db.Database", "line_number": 332, "usage_type": "call"}, {"api_name": "infoset.db.db", "line_number": 332, "usage_type": "name"}, {"api_name": "infoset.db.db.Database", "line_number": 355, "usage_type": "call"}, {"api_name": "infoset.db.db", "line_number": 355, "usage_type": "name"}, {"api_name": "infoset.db.db.Database", "line_number": 390, "usage_type": "call"}, {"api_name": "infoset.db.db", "line_number": 390, "usage_type": "name"}, {"api_name": "infoset.db.db.Database", "line_number": 424, "usage_type": "call"}, {"api_name": "infoset.db.db", "line_number": 424, "usage_type": "name"}, {"api_name": "collections.defaultdict", "line_number": 447, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 452, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 465, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 466, "usage_type": "call"}, {"api_name": "os.path", "line_number": 466, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 467, "usage_type": "call"}, {"api_name": "os.path", "line_number": 467, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 478, "usage_type": "call"}, {"api_name": "os.path", "line_number": 478, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 483, "usage_type": "call"}, {"api_name": "os.path.getmtime", "line_number": 483, "usage_type": "call"}, {"api_name": "os.path", "line_number": 483, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 522, "usage_type": "call"}]} +{"seq_id": "503243608", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 2.7 (62211)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /apps/xio/core/request.py\n# Compiled at: 2018-12-16 13:59:21\nimport requests, json, base64\nfrom pprint import pprint\nfrom xio.core.lib.utils import sha1\nimport traceback, sys\nif sys.version_info.major == 2:\n from Cookie import SimpleCookie\n from urllib import unquote\nelse:\n from http.cookies import SimpleCookie\n from urllib.parse import unquote\n__ALLOWED_METHODS__ = [\n 'HEAD', 'GET', 'POST', 'PUT', 'DELETE', 'PATCH', 'OPTIONS', 'CONNECT']\n\ndef request(method, path, **kwargs):\n if '://' in path:\n import requests\n url = path\n h = getattr(requests, method.lower())\n params = kwargs.get('query') or {}\n headers = kwargs.get('headers') or {}\n data = kwargs.get('data') or None\n r = h(url, params=params, data=data, headers=headers)\n response = Response(r.status_code)\n response.content_type = r.headers['content-type']\n response.headers = r.headers\n response.content = r.json() if response.content_type == 'application/json' else r.text\n return response\n else:\n return Request(method, path, **kwargs)\n\n\nclass UnhandledRequest:\n \"\"\" redirect to default hander \"\"\"\n pass\n\n\nclass Request(object):\n PASS = UnhandledRequest\n\n def __init__(self, method, path, query=None, headers=None, data=None, context=None, debug=False, client=None, client_context=None, server=None, **kwargs):\n context = context or {}\n headers = headers or {}\n path = path[1:] if path and path[0] == '/' else path\n xmethod = headers.get('XIO-method', headers.get('xio_method')) if headers else None\n if xmethod:\n xmethod = xmethod.upper()\n method = 'POST'\n if not xmethod and method.upper() not in __ALLOWED_METHODS__:\n xmethod = method.upper()\n method = 'POST'\n if 'XIO-method' not in headers and 'xio_method' not in headers:\n headers['XIO-method'] = xmethod\n method = method.upper()\n for m in __ALLOWED_METHODS__:\n setattr(self, m, False)\n\n if method == 'GET' and path and '.' in path:\n p = path.split('/')\n last = p.pop()\n if last and last[0] == '.':\n newmethod = last[1:].upper()\n if newmethod not in __ALLOWED_METHODS__:\n xmethod = newmethod\n method = 'POST'\n else:\n method = newmethod\n xmethod = None\n path = ('/').join(p)\n data = query\n query = None\n headers['XIO-method'] = xmethod\n setattr(self, method.upper(), True)\n if xmethod:\n setattr(self, xmethod.upper(), True)\n self.method = method\n self.xmethod = xmethod\n self.path = path\n self.fullpath = self.path\n self.context = context or {}\n self.headers = headers\n self.debug = False\n self.query = query or {}\n self.data = data or {}\n self.input = self.data or self.query\n self.cookie = Cookie(self)\n self.response = Response(200)\n self._uid = None\n self._stack = []\n self.stat = None\n self._server = server\n self._client = client\n self._client_context = client_context\n self.init()\n return\n\n def init(self):\n self.server = self.server or self.context.get('root', self.context.get('app', self.context.get('resource', None)))\n self.client = ReqClient(self, self._client_context, peer=self._client)\n return\n\n def __repr__(self):\n return 'REQUEST %s %s' % (self.xmethod or self.method, repr(self.path))\n\n def send(self, target, *args, **kwargs):\n \"\"\"\n send this request to handler and/or resource\n \"\"\"\n if callable(target):\n func = target\n else:\n raise Exception('not implemented yet')\n try:\n resp = func(self, *args, **kwargs)\n except Exception as err:\n args = err.args[0].args if err.args and isinstance(err.args[0], Exception) else err.args\n if args and isinstance(args[0], int):\n self.response.status = args[0]\n resp = args[1] if len(args) > 1 else None\n else:\n traceback.print_exc()\n self.response.status = 500\n self.response.traceback = str(traceback.format_exc())\n resp = None\n\n return resp\n\n def _debug(self):\n return {'method': self.method, \n 'path': self.path, \n 'xmethod': self.xmethod, \n 'headers': self.headers, \n 'query': self.query, \n 'data': self.data, \n 'input': self.input, \n 'profile': self.profile, \n 'client': {'auth': {'scheme': self.client.auth.scheme, \n 'token': self.client.auth.token, \n 'data': self.client.peer.key.tokendata if self.client.auth.token else None}, \n 'id': self.client.id, \n 'context': self.client.context, \n 'peer': self.client.peer, \n '_peer': self.client._peer}, \n 'server': self.server, \n 'context': self.context}\n\n def service(self, name):\n server = self.context.get('root')\n if server:\n service = server.get('services/%s' % name)\n return service\n\n def require(self, key, value, content=None):\n if key == 'auth':\n if not self.client.id:\n self.response.headers['WWW-Authenticate'] = 'Basic realm=\"%s\"' % value\n raise Exception(401)\n else:\n if key == 'signature':\n signature = self.headers.get('XIO-Signature')\n if not signature:\n self.response.headers['WWW-Authenticate'] = '%s realm=\"%s\", charset=\"UTF-8\"' % (value, 'xio realm')\n self.response.status = 402\n raise Exception(402, content)\n return signature\n if key == 'scope':\n if value not in self.client.data.get('scope', []):\n raise Exception(401, 'scope not allowed')\n elif key == 'quota':\n statservice = self.service('stats')\n if statservice:\n content = content or []\n path = ('/').join(content)\n stat = statservice.get(path).content\n hourly = stat.get('hourly')\n assert hourly < value, Exception(429, 'QUOTA EXCEEDED')\n assert statservice.incr(path)\n else:\n raise Exception('unknow require rule : %s' % key)\n\n def uid(self):\n \"\"\"\n generate uniq identifier for stats, cache, ...\n warning about\n - userid and/or profile for cache\n - qurey sting key orders\n \"\"\"\n if not self._uid:\n import hashlib, json\n struid = '%s-%s' % (self.fullpath, json.dumps(self.input, sort_keys=True))\n uid = sha1(struid)\n self._uid = uid\n return self._uid\n\n def getQuotas(self):\n \"\"\"\n retreive quotas provided by node\n \"\"\"\n quotas = self.headers.get('XIO-quotas') or self.headers.get('xio_quotas')\n if quotas:\n infos = json.loads(quotas)\n quotas = {}\n if infos[0]:\n quotas['ttl'] = infos[0]\n if infos[1]:\n quotas['storage'] = infos[1]\n if infos[2]:\n quotas['request'] = infos[2]\n if infos[2]:\n quotas['items'] = infos[3]\n return quotas or {}\n\n def __getattr__(self, name):\n if name in self.__dict__:\n return self.__dict__[name]\n else:\n return\n\n\nclass Response:\n\n def __init__(self, status):\n self.status = status\n self.headers = {}\n self.content_type = 'text/plain'\n self.content = None\n self.ttl = 0\n self.traceback = None\n return\n\n def __repr__(self):\n return 'RESPONSE %s %s' % (self.status, self.content_type)\n\n\nclass Cookie:\n\n def __init__(self, req):\n self._req = req\n self._data = req.context.get('cookies', {})\n\n def set(self, key, value):\n import http.cookies\n cookie = http.cookies.SimpleCookie()\n cookie[key] = str(value)\n cookie[key]['path'] = '/'\n strcookie = cookie.output()\n valuecookie = strcookie.replace('Set-Cookie: ', '')\n self._req.response.headers['Set-Cookie'] = valuecookie\n\n def get(self, key):\n value = self._data.get(key)\n if value:\n return unquote(value)\n else:\n return\n\n\nclass Auth:\n scheme = None\n token = None\n\n def __init__(self, client):\n self.req = client.req\n authorization = self.req.headers.get('Authorization', self.req.headers.get('authorization'))\n if not authorization and client.context:\n authorization = client.context.get('authorization')\n if not authorization:\n token = self.req.cookie.get('XIO-AUTH')\n authorization = 'bearer ' + token if token else None\n if authorization:\n scheme, token = authorization.split(' ')\n self.scheme = scheme.lower()\n self.token = token\n return\n\n\nclass ReqClient:\n\n def __init__(self, req, context=None, peer=None):\n import xio\n self.req = req\n self.id = None\n self.peer = peer\n self._peer = peer\n self.context = context\n self.auth = Auth(self)\n self.data = dict()\n if self.auth.token:\n if self.auth.scheme == 'basic':\n login, password = base64.urlsafe_b64decode(self.auth.token).split(':')\n if login == 'seed':\n user = xio.user(seed=password)\n self.auth.scheme = 'bearer'\n self.auth.token = user.key.generateToken('xio/ethereum')\n try:\n self.peer = xio.user(token=self.auth.token)\n self.id = self.peer.id\n self.data = self.peer.key.tokendata\n except Exception as err:\n import traceback\n traceback.print_exc()\n print ('UNCOVERABLE TOKEN', err)\n self.id = None\n self.data = {}\n\n self.data.setdefault('scope', [])\n if self.id and xio.env.get('admin') == self.id:\n self.data['scope'].append('admin')\n self._feedback = req.context.get('feedback')\n self._wsendpoint = req.context.get('wsendpoint')\n self.send = self._send if self._feedback else None\n self.onreceive = self._onreceive if self._wsendpoint else None\n if context:\n if 'authorization' in context:\n req.headers['Authorization'] = context.pop('authorization')\n self.context = context\n req.headers['XIO-context'] = json.dumps(self.context)\n else:\n get_context = req.query.get('xio_context', {})\n self.context = req.headers.get('XIO-context', req.headers.get('xio_context', get_context))\n if self.context and self.context[0] == '{':\n self.context = json.loads(self.context)\n return\n\n def auth(self):\n return bool(self.token)\n\n def __bool__(self):\n return self.id != None\n\n def __nonzero__(self):\n return self.id != None\n\n def _ws_onreceive(self, msg):\n self._wsendpoint.send(msg)\n\n def _ws_send(self, msg):\n self._feedback(msg)", "sub_path": "pycfiles/xio-0.0.6-py3-none-any/request.py", "file_name": "request.py", "file_ext": "py", "file_size_in_byte": 11941, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "sys.version_info", "line_number": 11, "usage_type": "attribute"}, {"api_name": "traceback.print_exc", "line_number": 126, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 128, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 195, "usage_type": "call"}, {"api_name": "xio.core.lib.utils.sha1", "line_number": 196, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 206, "usage_type": "call"}, {"api_name": "http.cookies.cookies.SimpleCookie", "line_number": 248, "usage_type": "call"}, {"api_name": "http.cookies.cookies", "line_number": 248, "usage_type": "attribute"}, {"api_name": "http.cookies", "line_number": 248, "usage_type": "name"}, {"api_name": "urllib.parse.unquote", "line_number": 258, "usage_type": "call"}, {"api_name": "base64.urlsafe_b64decode", "line_number": 295, "usage_type": "call"}, {"api_name": "xio.user", "line_number": 297, "usage_type": "call"}, {"api_name": "xio.user", "line_number": 301, "usage_type": "call"}, {"api_name": "traceback.print_exc", "line_number": 306, "usage_type": "call"}, {"api_name": "xio.env.get", "line_number": 312, "usage_type": "call"}, {"api_name": "xio.env", "line_number": 312, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 322, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 327, "usage_type": "call"}]} +{"seq_id": "200111375", "text": "import cv2\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nfrom PIL import Image\nimport pytesseract\n\n\nimport urllib.request\nimport urllib.error\nimport pika\n\nconnection = pika.BlockingConnection(pika.URLParameters('amqp://cfcwgxer:LLYwXV4lQBYeTxELzkGZrXJyia6CvCDD@bee.rmq.cloudamqp.com/cfcwgxer'))\nchannel = connection.channel()\nchannel.queue_declare(queue=\"transcribed\")\n\n\nupperBlue = np.array([274,220,110])\nlowerBlue = np.array([234,165,10])\n\nlowerGrey = np.array([220,220,220])\nupperGrey = np.array([240,240,240])\n\ndef urlToImage(url):\n\t# download the image, convert it to a NumPy array, and then read\n\t# it into OpenCV format\n\tresp = urllib.request.urlopen(url)\n\timage = np.asarray(bytearray(resp.read()), dtype=\"uint8\")\n\timage = cv2.imdecode(image, cv2.IMREAD_COLOR)\n \n\treturn image\n\n\n\ndef cropMessage(index,mask,image):\n #Crops the nth message given a mask\n cnts = cv2.findContours(mask, cv2.RETR_EXTERNAL,\n cv2.CHAIN_APPROX_SIMPLE)\n mask = np.zeros_like(image)\n cv2.drawContours(mask, cnts[0], index, 255, -1)\n out = np.zeros_like(image)\n out[mask == 255] = image[mask == 255]\n\n\n a = np.where(mask == 255)\n\n (topx, topy) = (np.min(a[0]), np.min(a[1]))\n (bottomx, bottomy) = (np.max(a[0]), np.max(a[1]))\n return image[topx:bottomx+1, topy:bottomy+1],topx\n\ndef transformBlue(messageCrop):\n #blue image is converted to a more readable black and white\n messageCrop[np.where((messageCrop > [140,140,140]).all(axis = 2))] = [0,0,0]\n messageCrop[np.where((messageCrop != [0,0,0]).all(axis = 2))] = [255,255,255]\n return messageCrop\n\n\n\ndef getMessages(image, blueThreshold = (lowerBlue, upperBlue), greyThreshold = (lowerGrey, upperGrey)):\n #Returns an array of messages in the form (message image, message height, message sender)\n \n\n\n #get masks for blue and grey messages\n shapeMask1 = cv2.inRange(image, blueThreshold[0], blueThreshold[1])\n\n shapeMask2 = cv2.inRange(image, greyThreshold[0], greyThreshold[1])\n\n\n erode = cv2.erode(shapeMask1, None, iterations = 3)\n mask1 = cv2.dilate(erode, None, iterations = 13)\n\n erode = cv2.erode(shapeMask2, None, iterations = 3)\n mask2 = cv2.dilate(erode, None, iterations = 13)\n\n\n messages = []\n counter1 = 0\n counter2 = 0\n\n\n #add blue messages\n while True:\n try:\n \n crop,pos = cropMessage(counter1,mask1,image)\n messages.append((crop,pos,1))\n counter1+=1\n \n except Exception as e:\n \n break\n\n #add grey images\n while True:\n try:\n crop,pos = cropMessage(counter2,mask2,image)\n messages.append((crop,pos,2))\n counter2+=1\n except Exception:\n break\n\n #sort by message position\n messages.sort(key = lambda pair : pair[1])\n\n #transform blue messages\n for message in messages:\n if message[2] == 1:\n message = (transformBlue(message[0]),message[1],1) \n\n\n return messages\n \n\n#pytesseract transcriotion for a given numpy image\ndef imageToText(image):\n img = Image.fromarray(image)\n txt = pytesseract.image_to_string(img)\n return txt\n\n#main function - transcribes image from given url\ndef transcirbeUrl(url):\n\n transcription = \"\"\n image = urlToImage(url)\n\n messages = getMessages(image,(lowerBlue,upperBlue), (lowerGrey,upperGrey))\n for message in messages:\n text = imageToText(message[0])\n if len(text)>1:\n transcription += \"[\"+str(message[2])+\"]\"+ text + \"\\n\"\n return transcription\n\ndef publish(url):\n channel.basic_publish(exchange='',\n routing_key=\"transcribed\",\n body=url)\n\n\ndef filter(ch, method, properties, body):\n url = body.decode('UTF-8')\n print(\"transcribing \" + url)\n try:\n publish(transcirbeUrl(url))\n except Exception:\n print(\"error prasing link\")\n ch.basic_ack(delivery_tag=method.delivery_tag)\n\n\n\n\n\nchannel.basic_qos(prefetch_count=3)\nchannel.basic_consume(filter,\n queue='filtered',\n no_ack=False)\n\nchannel.start_consuming()\n\n\n\n", "sub_path": "Src/Python/transcriber.py", "file_name": "transcriber.py", "file_ext": "py", "file_size_in_byte": 3935, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pika.BlockingConnection", "line_number": 13, "usage_type": "call"}, {"api_name": "pika.URLParameters", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 22, "usage_type": "call"}, {"api_name": "urllib.request.request.urlopen", "line_number": 27, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 27, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 27, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.imdecode", "line_number": 29, "usage_type": "call"}, {"api_name": "cv2.IMREAD_COLOR", "line_number": 29, "usage_type": "attribute"}, {"api_name": "cv2.findContours", "line_number": 37, "usage_type": "call"}, {"api_name": "cv2.RETR_EXTERNAL", "line_number": 37, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_SIMPLE", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.zeros_like", "line_number": 39, "usage_type": "call"}, {"api_name": "cv2.drawContours", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 54, "usage_type": "call"}, {"api_name": "cv2.inRange", "line_number": 65, "usage_type": "call"}, {"api_name": "cv2.inRange", "line_number": 67, "usage_type": "call"}, {"api_name": "cv2.erode", "line_number": 70, "usage_type": "call"}, {"api_name": "cv2.dilate", "line_number": 71, "usage_type": "call"}, {"api_name": "cv2.erode", "line_number": 73, "usage_type": "call"}, {"api_name": "cv2.dilate", "line_number": 74, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 117, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 117, "usage_type": "name"}, {"api_name": "pytesseract.image_to_string", "line_number": 118, "usage_type": "call"}]} +{"seq_id": "368727221", "text": "import os\nimport platform\nimport sys\n\nfrom channels.generic.websocket import AsyncWebsocketConsumer\nfrom my_app.models import trade_record, close_record\nimport json\nimport asyncio\nimport logging\nfrom configobj import ConfigObj\nfrom Profile import profile\n\nlogger = logging.getLogger(\"intranet\")\nconf = ConfigObj(profile.CONFIG_INI_URL, encoding=\"utf-8\")\nloop_userlist = None\nloop_online = None\nsend_flag = False\nuser_list = {}\nuser_online = {}\nkeep_dict = {}\nmodel = \"normal\"\n\"\"\"\nuser_list——用户列表 {self1:\"dc1\",self2:\"dc2\",...}\nuser_online——在线用户列表 {self1:\"dc1\",self2:\"dc2\",...}\nkeep_dict——控制端 {\"keep\":self}\n model——模式\n normal:正常\n risk_off:风控模式下未登录\n risk_on:风控模式下已登陆\n \"\"\"\n\n\n# 对冲端内网\nclass Duichong(AsyncWebsocketConsumer):\n # 风控掉线时,报告给签约\n async def sendUser(self):\n global send_flag\n while True:\n if send_flag:\n await asyncio.sleep(3)\n global user_list\n data = json.dumps({\n \"purpose\": \"send_user\",\n \"user_list\": list(user_list.values())\n })\n await keep_dict[\"keep\"].my_send(data)\n logger.error(\"报告用户列表给签约:\" + str(user_list.values()))\n send_flag = False\n else:\n await asyncio.sleep(1)\n\n # 检测客户端在线的心跳\n async def check_online(self):\n global user_online, user_list\n while True:\n logger.error(\"用户:\" + str(user_list.values())) # {self1:\"dc1\",self2:\"dc2\",...}\n logger.error(\"在线用户:\" + str(user_online.values())) # [self1,self2,...]\n # 踢掉断网的客户端\n for user in user_list.keys():\n if user not in user_online.keys():\n logger.error(\"掉线:\" + str(user_list[user]))\n user.close()\n user_online = {} # 清空在线用户,等待下一次心跳\n await asyncio.sleep(float(conf[\"timeout\"][\"heart_interval\"])/1000)\n\n # 创建连接\n async def connect(self):\n global model, loop_userlist\n if model == \"risk_off\":\n pass\n else:\n await self.accept()\n\n # 接收消息\n async def receive(self, text_data):\n try:\n global user_list, keep_dict, model, flag\n message = json.loads(text_data)\n \"\"\"日志记录\"\"\"\n if message[\"purpose\"] == \"control\" or message[\"purpose\"] == \"model_choose\" or message[\"purpose\"] == \"all\" or message[\"purpose\"] == \"keep\":\n # logger.info(\"收到控制端数据\" + text_data)\n pass\n else:\n logger.info(\"收到客户端数据\" + text_data)\n\n \"\"\"客户端发来消息\"\"\"\n\n \"\"\"心跳\"\"\"\n if message[\"purpose\"] == \"keep\":\n global loop_online, user_online\n await self.my_send(({\n \"purpose\": \"keet\"\n }))\n if self not in user_online.keys():\n user_online[self] = message[\"name\"]\n # 开启协程,检测客户端在线(永久循环)\n if loop_online is None:\n loop_online = asyncio.get_event_loop()\n loop_online.create_task(self.check_online())\n # 登陆\n if message[\"purpose\"] == \"intranet_login_request\":\n # 判断内网当前模式\n if model == \"normal\": # 正常\n data = {\n \"purpose\": \"intranet_login_success\",\n \"model\": \"正常模式\"\n }\n await self.my_send(data)\n elif model == \"risk_off\": # 风控,且未登陆签约\n data = {\n \"purpose\": \"intranet_login_fail\",\n \"reason\": \"风控模式下对冲内网尚未连接签约,不允许连接\"\n }\n await self.my_send(data)\n elif model == \"risk_on\":\n user_list[self] = message[\"user_name\"] # 保存用户列表\n data = {\n \"purpose\": \"intranet_login_success\",\n \"model\": \"风控模式\"\n }\n await self.my_send(data) # 返回对冲登陆成功\n # 交易记录\n if message[\"purpose\"] == \"trade_record\":\n # 数组转数据库\n options_list = message[\"options\"]\n tmp_list = []\n for i in range(len(options_list)):\n tmp_list.append(\"!\".join(str(j) for j in options_list[i]))\n options = \"|\".join(str(j) for j in tmp_list)\n # 存\n p = trade_record(\n user_id=message[\"user_id\"],\n time=message[\"record_time\"],\n instrument_id=message[\"instrument_id\"],\n account_position=message[\"account_position\"],\n other_position=message[\"other_position\"],\n ratio=message[\"ratio\"],\n thero_position=message[\"thero_position\"],\n close_diff=message[\"close_diff\"],\n options=options,\n )\n p.save()\n logger.info(\"成功保存交易记录\")\n # 收盘记录\n if message[\"purpose\"] == \"position_broadcast\":\n datas = message[\"messages\"]\n for data in datas:\n opt_ids = data[\"opt_ids\"]\n tmp_list = []\n # 期权数组转数据库\n for j in range(len(opt_ids)):\n tmp = str(opt_ids[j][\"opt_id\"]) + \"!\" + str(opt_ids[j][\"theoretical_position\"]) + \"!\" + str(\n opt_ids[j][\"split_close\"])\n tmp_list.append(tmp)\n options = \"|\".join(str(j) for j in tmp_list)\n p = close_record(\n user_id=message[\"user_id\"],\n time=message[\"time\"],\n instrument_id=data[\"future_id\"],\n account_position=data[\"account_position\"],\n other_position=data[\"other_position\"],\n ratio=message[\"ratio\"],\n thero_position=data[\"all_theoretical_position\"],\n close_diff=data[\"all_split_close\"],\n options=options,\n )\n p.save()\n logger.info(\"成功保存收盘记录\")\n\n \"\"\"以下是控制端发来的消息\"\"\"\n # 建立连接\n if message[\"purpose\"] == \"control\":\n keep_dict[\"keep\"] = self\n keep_dict[\"process_port\"] = message[\"process_port\"]\n keep_dict[\"server_port\"] = message[\"server_port\"]\n data = {\n \"purpose\": \"control success\"\n }\n await self.my_send(data)\n # 模式选择\n if message[\"purpose\"] == \"model_choose\":\n if message[\"model_type\"] == \"risk_off\": # 风控:未登录\n model = \"risk_off\"\n elif message[\"model_type\"] == \"risk_on\": # 风控:登陆\n model = \"risk_on\"\n logger.info(\"内网模式改变:\" + model)\n except Exception as e:\n logger.error(\"异常:\" + str(e))\n\n async def disconnect(self, close_code):\n global loop_userlist, send_flag\n try:\n if self in list(user_list.keys()): # 风控模式下,有用户退出\n user_list.pop(self) # 踢出用户列表\n send_flag = True\n if loop_userlist is None: # 开启协程,延迟3秒发(永久循环)\n loop_userlist = asyncio.get_event_loop()\n loop_userlist.create_task(self.sendUser())\n elif self == keep_dict[\"keep\"]:\n # 检测到监控端掉线,内网自己关闭\n try:\n if platform.system() == 'Linux':\n # os.system(\"ps -ef | grep 'python manage.py runserver 0.0.0.0:\" + str(\n # keep_dict[\"server_port\"]) + \"' | grep -v grep | cut -c 9-15 | xargs kill -s 9\")\n pass\n else:\n os.system(\"taskkill /F /pid \" + str(keep_dict[\"process_port\"]) + \" -t\")\n except Exception as e:\n print(e)\n await self.close()\n except Exception as e:\n logger.error(\"异常:\" + str(e))\n\n async def my_send(self, data):\n \"\"\"发送数据\"\"\"\n logger.info(\"发送数据\" + str(data))\n if isinstance(data, dict):\n d = json.dumps(data)\n else:\n d = str(data)\n await self.send(d)\n", "sub_path": "app/my_project/consumers.py", "file_name": "consumers.py", "file_ext": "py", "file_size_in_byte": 9115, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "logging.getLogger", "line_number": 13, "usage_type": "call"}, {"api_name": "configobj.ConfigObj", "line_number": 14, "usage_type": "call"}, {"api_name": "Profile.profile.CONFIG_INI_URL", "line_number": 14, "usage_type": "attribute"}, {"api_name": "Profile.profile", "line_number": 14, "usage_type": "name"}, {"api_name": "channels.generic.websocket.AsyncWebsocketConsumer", "line_number": 34, "usage_type": "name"}, {"api_name": "asyncio.sleep", "line_number": 40, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 42, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 50, "usage_type": "call"}, {"api_name": "asyncio.sleep", "line_number": 64, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 78, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 98, "usage_type": "call"}, {"api_name": "my_app.models.trade_record", "line_number": 131, "usage_type": "call"}, {"api_name": "my_app.models.close_record", "line_number": 156, "usage_type": "call"}, {"api_name": "asyncio.get_event_loop", "line_number": 197, "usage_type": "call"}, {"api_name": "platform.system", "line_number": 202, "usage_type": "call"}, {"api_name": "os.system", "line_number": 207, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 218, "usage_type": "call"}]} +{"seq_id": "482049356", "text": "from datetime import datetime\n\nimport pandas as pd\n\nfrom cryptoz.exchanges import Exchange\n\n\nclass Binance(Exchange):\n def __init__(self, client):\n Exchange.__init__(self, client)\n\n @staticmethod\n def _ts_to_dt(ts):\n return datetime.utcfromtimestamp(ts / 1000)\n\n @staticmethod\n def _to_intern_pair(pair):\n \"\"\"BTCUSDT to BTC/USDT\"\"\"\n if '/' in pair:\n return pair\n supported_quotes = ['USDT', 'BTC', 'ETH', 'BNB']\n for quote in supported_quotes:\n if pair[-len(quote):] == quote:\n return pair[:-len(quote)] + '/' + quote\n return None\n\n @staticmethod\n def _to_exchange_pair(pair):\n \"\"\"BTC/USDT to BTCUSDT\"\"\"\n if '/' not in pair:\n return pair\n return ''.join(pair.split('/'))\n\n def get_pairs(self):\n pairs = set(map(lambda d: self._to_intern_pair(d['symbol']), self.client.get_all_tickers()))\n return list(filter(lambda s: s is not None, pairs))\n\n def get_ticker(self):\n ticker = {}\n for d in self.client.get_all_tickers():\n pair = self._to_intern_pair(d['symbol'])\n if pair is not None:\n ticker[pair] = float(d['price'])\n return ticker\n\n def _get_balances(self):\n balances = self.client.get_account()['balances']\n df = pd.DataFrame(balances)\n df.set_index('asset', drop=True, inplace=True)\n df = df.astype(float)\n df = df[(df['free'] > 0) & (df['locked'] == 0)]\n df.drop('locked', 1, inplace=True)\n return df['free'].to_dict()\n\n def get_balances(self, hide_small_assets=True):\n return Exchange._load_and_convert_balances(self, hide_small_assets)\n\n def _get_ohlc(self, pair, **kwargs):\n pair = self._to_exchange_pair(pair)\n \"\"\"Load OHLC data on a single pair\"\"\"\n candles = self.client.get_klines(symbol=pair, **kwargs)\n columns = ['date', 'O', 'H', 'L', 'C', '_', '_', 'V', '_', '_', '_', '_']\n\n df = pd.DataFrame(candles, columns=columns)\n df.set_index('date', drop=True, inplace=True)\n df.index = [self._ts_to_dt(i) for i in df.index]\n df.fillna(method='ffill', inplace=True) # fill gaps forwards\n df.fillna(method='bfill', inplace=True) # fill gaps backwards\n df = df.astype(float)\n df = df[['O', 'H', 'L', 'C', 'V']]\n df['M'] = (df['L'] + df['H'] + df['C']) / 3\n df = df.iloc[1:] # first entry can be dirty\n return df\n\n def get_ohlc(self, pairs, **kwargs):\n return Exchange._load_and_convert_ohlc(self, pairs, **kwargs)\n\n def _get_orderbook(self, pair, **kwargs):\n pair = self._to_exchange_pair(pair)\n orderbook = self.client.get_order_book(symbol=pair, **kwargs)\n\n rates, amounts, _ = zip(*orderbook['bids'])\n cum_bids = pd.Series(amounts, index=rates, dtype=float)\n cum_bids.index = cum_bids.index.astype(float)\n cum_bids = cum_bids.sort_index(ascending=False).cumsum().sort_index()\n cum_bids *= cum_bids.index\n\n rates, amounts, _ = zip(*orderbook['asks'])\n cum_asks = pd.Series(amounts, index=rates, dtype=float)\n cum_asks.index = cum_asks.index.astype(float)\n cum_asks = -cum_asks.sort_index().cumsum()\n cum_asks *= cum_asks.index\n\n return cum_bids.append(cum_asks).sort_index()\n\n def get_orderbooks(self, pairs, **kwargs):\n return Exchange._load_and_convert_orderbooks(self, pairs, **kwargs)\n", "sub_path": "cryptoz/exchanges/binance.py", "file_name": "binance.py", "file_ext": "py", "file_size_in_byte": 3493, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "cryptoz.exchanges.Exchange", "line_number": 8, "usage_type": "name"}, {"api_name": "cryptoz.exchanges.Exchange.__init__", "line_number": 10, "usage_type": "call"}, {"api_name": "cryptoz.exchanges.Exchange", "line_number": 10, "usage_type": "name"}, {"api_name": "datetime.datetime.utcfromtimestamp", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 14, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 48, "usage_type": "call"}, {"api_name": "cryptoz.exchanges.Exchange._load_and_convert_balances", "line_number": 56, "usage_type": "call"}, {"api_name": "cryptoz.exchanges.Exchange", "line_number": 56, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 64, "usage_type": "call"}, {"api_name": "cryptoz.exchanges.Exchange._load_and_convert_ohlc", "line_number": 76, "usage_type": "call"}, {"api_name": "cryptoz.exchanges.Exchange", "line_number": 76, "usage_type": "name"}, {"api_name": "pandas.Series", "line_number": 83, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 89, "usage_type": "call"}, {"api_name": "cryptoz.exchanges.Exchange._load_and_convert_orderbooks", "line_number": 97, "usage_type": "call"}, {"api_name": "cryptoz.exchanges.Exchange", "line_number": 97, "usage_type": "name"}]} +{"seq_id": "354245797", "text": "from cerberus import Validator\n\nfrom TopSpeedData_Worker.connection.oracle import connect_target, engine_target\nimport pandas as pd\n\nclass CustomValidator(Validator):\n\n def __init__(self,rule,target_table_name):#传不定长参数\n super(CustomValidator, self).__init__(rule)\n self.conn_target = connect_target()\n self.target_table_name = target_table_name\n self.tmp_data = pd.read_sql(\"select * from {}\".format(self.target_table_name),con = engine_target)\n\n\n def _validate_isduplication(self, isduplication, field, value):#判断一个字段是否有重复值\n \"\"\" Test the oddity of a value.\n The rule's arguments are validated against this schema:\n {'type': 'boolean'}\n \"\"\"\n tmp_data = self.tmp_data[(self.tmp_data['{}'.format(field)] == '{}'.format(value))]\n if tmp_data.count()[0] == 2 and isduplication == True:\n self._error(field, \"证券简称为{}的债券重复\".format(tmp_data.values[0][0]))\n\n\n\n# rule = {'vc_sname': {'isduplication': True}}\n# v = CustomValidator(rule, 't_tmp_bsc_investment_advice')\n# print(v.validate({'vc_sname': '银河次级'}))\n# print(v.errors)\n\n\n# schema = {'amount': {'odd': True, 'type': 'integer'}}\n# v = MyValidator(schema)\n# v.validate({'amount': 10})\n#print(v.errors)\n#v.validate({'amount': 9})\n#print(v.errors)\n", "sub_path": "worker_fd/validation/customvalidator.py", "file_name": "customvalidator.py", "file_ext": "py", "file_size_in_byte": 1342, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "cerberus.Validator", "line_number": 6, "usage_type": "name"}, {"api_name": "TopSpeedData_Worker.connection.oracle.connect_target", "line_number": 10, "usage_type": "call"}, {"api_name": "pandas.read_sql", "line_number": 12, "usage_type": "call"}, {"api_name": "TopSpeedData_Worker.connection.oracle.engine_target", "line_number": 12, "usage_type": "name"}]} +{"seq_id": "448839290", "text": "# Copyright 2019 Shift Cryptosecurity AG\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"BitBox02\"\"\"\n\n\nimport os\nimport sys\nimport time\nimport base64\nimport binascii\nfrom datetime import datetime\nimport hashlib\n\nimport ecdsa\nfrom noise.connection import NoiseConnection, Keypair\nimport hid\nimport semver\n\nfrom .usb import hid_send_frames, hid_read_frames\nfrom .devices import parse_device_version\n\ntry:\n from .generated import hww_pb2 as hww\nexcept ModuleNotFoundError:\n print(\"Run `make messages` to generate the protobuf messages\")\n sys.exit()\n\ntry:\n # Optional rlp dependency only needed to sign ethereum transactions.\n # pylint: disable=import-error\n import rlp\nexcept ModuleNotFoundError:\n pass\n\n\nHWW_CMD = 0x80 + 0x40 + 0x01\n\nERR_GENERIC = 103\n\nHARDENED = 0x80000000\n\n# values: uncompressed secp256k1 pubkey serialization.\nATTESTATION_PUBKEYS = [\n binascii.unhexlify(\n \"04074ff1273b36c24e80fe3d59e0e897a81732d3f8e9cd07e17e9fc06319cd16b\"\n \"25cf74255674477b3ac9cbac2d12f0dc27a662681fcbc12955b0bccdcbbdcfd01\"\n )\n]\n\nATTESTATION_PUBKEYS_MAP = {hashlib.sha256(val).digest(): val for val in ATTESTATION_PUBKEYS}\n\nOP_ATTESTATION = b\"a\"\nOP_UNLOCK = b\"u\"\nOP_I_CAN_HAS_HANDSHAEK = b\"h\"\nOP_I_CAN_HAS_PAIRIN_VERIFICASHUN = b\"v\"\n\nRESPONSE_SUCCESS = b\"\\x00\"\nRESPONSE_FAILURE = b\"\\x01\"\n\n\nclass Bitbox02Exception(Exception):\n def __init__(self, code, message):\n self.code = code\n self.message = message\n super().__init__()\n\n def __str__(self):\n return f\"error code: {self.code}, message: {self.message}\"\n\n\nclass AttestationException(Exception):\n pass\n\n\nclass BitBox02:\n \"\"\"Class to communicate with a BitBox02\"\"\"\n\n def __init__(self, device_info, show_pairing_callback, attestation_check_callback=None):\n self.debug = False\n serial_number = device_info[\"serial_number\"]\n self.version = parse_device_version(serial_number)\n if self.version is None:\n raise ValueError(f\"Could not parse version from {serial_number}\")\n self.device = hid.device()\n self.device.open_path(device_info[\"path\"])\n\n if self.version > semver.VersionInfo(1, 0, 0):\n if attestation_check_callback is not None:\n # Perform attestation\n attestation_check_callback(self._perform_attestation())\n\n # Invoke unlock workflow on the device.\n # In version <=1.0.0, the device did this automatically.\n self._query(OP_UNLOCK)\n\n if self._query(OP_I_CAN_HAS_HANDSHAEK) != RESPONSE_SUCCESS:\n raise Exception(\"Couldn't kick off handshake\")\n\n # init noise channel\n noise = NoiseConnection.from_name(b\"Noise_XX_25519_ChaChaPoly_SHA256\")\n noise.set_as_initiator()\n dummy_private_key = os.urandom(32)\n noise.set_keypair_from_private_bytes(Keypair.STATIC, dummy_private_key)\n noise.set_prologue(b\"Noise_XX_25519_ChaChaPoly_SHA256\")\n noise.start_handshake()\n noise.read_message(self._query(noise.write_message()))\n assert not noise.handshake_finished\n send_msg = noise.write_message()\n assert noise.handshake_finished\n pairing_code = base64.b32encode(noise.get_handshake_hash()).decode(\"ascii\")\n show_pairing_callback(\n \"{} {}\\n{} {}\".format(\n pairing_code[:5], pairing_code[5:10], pairing_code[10:15], pairing_code[15:20]\n )\n )\n response = self._query(send_msg)\n\n # Can be set to False if the remote static pubkey was previously confirmed.\n pairing_verification_required_by_host = True\n\n pairing_verification_required_by_device = response == b\"\\x01\"\n if pairing_verification_required_by_host or pairing_verification_required_by_device:\n pairing_response = self._query(OP_I_CAN_HAS_PAIRIN_VERIFICASHUN)\n if pairing_response == RESPONSE_SUCCESS:\n pass\n elif pairing_response == RESPONSE_FAILURE:\n raise Exception(\"pairing rejected by the user\")\n else:\n raise Exception(\"unexpected response\")\n self.noise = noise\n\n def close(self):\n self.device.close()\n\n def _query(self, msg):\n \"\"\"\n Sends msg bytes and retrieves response bytes.\n \"\"\"\n hid_send_frames(self.device, msg, cmd=HWW_CMD)\n return bytes(hid_read_frames(self.device, cmd=HWW_CMD))\n\n def _encrypted_query(self, msg):\n \"\"\"\n Sends msg bytes and reads response bytes over an encrypted channel.\n \"\"\"\n return self.noise.decrypt(self._query(self.noise.encrypt(msg)))\n\n def _msg_query(self, request, expected_response=None):\n \"\"\"\n Sends protobuf msg and retrieves protobuf response over an encrypted\n channel.\n \"\"\"\n # pylint: disable=no-member\n if self.debug:\n print(request)\n response_bytes = self._encrypted_query(request.SerializeToString())\n response = hww.Response()\n response.ParseFromString(response_bytes)\n if response.WhichOneof(\"response\") == \"error\":\n raise Bitbox02Exception(response.error.code, response.error.message)\n if expected_response is not None and response.WhichOneof(\"response\") != expected_response:\n raise Exception(\n \"Unexpected response: {}, expected: {}\".format(\n response.WhichOneof(\"response\"), expected_response\n )\n )\n if self.debug:\n print(response)\n return response\n\n def random_number(self):\n # pylint: disable=no-member\n request = hww.Request()\n request.random_number.CopyFrom(hww.RandomNumberRequest())\n response = self._msg_query(request, expected_response=\"random_number\")\n return response.random_number.number\n\n def device_info(self):\n # pylint: disable=no-member\n request = hww.Request()\n device_info_request = hww.DeviceInfoRequest()\n request.device_info.CopyFrom(device_info_request)\n response = self._msg_query(request, expected_response=\"device_info\")\n return {\n \"name\": response.device_info.name,\n \"version\": response.device_info.version,\n \"initialized\": response.device_info.initialized,\n \"mnemonic_passphrase_enabled\": response.device_info.mnemonic_passphrase_enabled,\n \"monotonic_increments_remaining\": response.device_info.monotonic_increments_remaining,\n }\n\n def set_device_name(self, device_name):\n # pylint: disable=no-member\n request = hww.Request()\n request.device_name.name = device_name\n self._msg_query(request, expected_response=\"success\")\n\n def set_password(self):\n \"\"\"\n Returns True if the user entered the password correctly (passwords match).\n Returns False otherwise.\n \"\"\"\n # pylint: disable=no-member\n request = hww.Request()\n request.set_password.entropy = os.urandom(32)\n try:\n self._msg_query(request, expected_response=\"success\")\n except Bitbox02Exception as err:\n if err.code == ERR_GENERIC:\n return False\n raise\n return True\n\n def create_backup(self):\n \"\"\"\n Returns True if the backup was created successfully.\n Returns False otherwise.\n \"\"\"\n # pylint: disable=no-member\n request = hww.Request()\n request.create_backup.timestamp = int(time.time())\n request.create_backup.timezone_offset = time.localtime().tm_gmtoff\n try:\n self._msg_query(request, expected_response=\"success\")\n except Bitbox02Exception as err:\n if err.code == ERR_GENERIC:\n return False\n raise\n return True\n\n def list_backups(self):\n \"\"\"\n Returns a pair of id and timestamp's strings that identify the backups.\n \"\"\"\n # pylint: disable=no-member\n self.insert_or_remove_sdcard(insert=True)\n request = hww.Request()\n request.list_backups.CopyFrom(hww.ListBackupsRequest())\n response = self._msg_query(request, expected_response=\"list_backups\")\n for info in response.list_backups.info:\n utcdate = datetime.utcfromtimestamp(info.timestamp)\n yield (info.id, info.name, utcdate)\n\n def restore_backup(self, backup_id):\n \"\"\"\n Sends a restore API call to the BitBox.\n \"\"\"\n # pylint: disable=no-member\n request = hww.Request()\n request.restore_backup.id = backup_id\n try:\n self._msg_query(request, expected_response=\"success\")\n except Bitbox02Exception as err:\n if err.code == ERR_GENERIC:\n return False\n raise\n return True\n\n def check_backup(self, silent=False):\n \"\"\"\n Sends a check backup API call to the BitBox.\n Returns the backup ID if the backup was found and can be restored.\n Otherwise, returns None. If silent is True, the result won't be shown on the device screen.\n \"\"\"\n # pylint: disable=no-member\n self.insert_or_remove_sdcard(insert=True)\n request = hww.Request()\n request.check_backup.CopyFrom(hww.CheckBackupRequest(silent=silent))\n try:\n response = self._msg_query(request, expected_response=\"check_backup\")\n except Bitbox02Exception as err:\n if err.code == ERR_GENERIC:\n return None\n raise\n return response.check_backup.id\n\n def show_mnemonic(self):\n \"\"\"\n Returns True if mnemonic was successfully shown and confirmed.\n Returns False otherwise.\n \"\"\"\n # pylint: disable=no-member\n request = hww.Request()\n request.show_mnemonic.CopyFrom(hww.ShowMnemonicRequest())\n try:\n self._msg_query(request, expected_response=\"success\")\n except Bitbox02Exception as err:\n if err.code == ERR_GENERIC:\n return False\n raise\n return True\n\n def btc_pub(\n self,\n keypath=None,\n coin=hww.BTC,\n output_type=hww.BTCPubRequest.XPUB,\n script_type=hww.SCRIPT_UNKNOWN,\n display=True,\n ):\n \"\"\"\n keypath is a list of child derivation numbers.\n e.g. m/44'/0'/1'/5 corresponds to [44+HARDENED, 0+HARDENED, 1+HARDENED, 5].\n \"\"\"\n # pylint: disable=no-member,too-many-arguments\n keypath = [] if keypath is None else keypath\n request = hww.Request()\n request.btc_pub.CopyFrom(\n hww.BTCPubRequest(\n coin=coin,\n keypath=keypath,\n output_type=output_type,\n script_type=script_type,\n display=display,\n )\n )\n return self._msg_query(request).pub.pub\n\n def check_sdcard(self):\n # pylint: disable=no-member\n request = hww.Request()\n request.check_sdcard.CopyFrom(hww.CheckSDCardRequest())\n response = self._msg_query(request, expected_response=\"check_sdcard\")\n return response.check_sdcard.inserted\n\n def insert_or_remove_sdcard(self, insert=False, remove=False):\n \"\"\"TODO: document\"\"\"\n # pylint: disable=no-member\n request = hww.Request()\n if insert:\n request.insert_remove_sdcard.CopyFrom(\n hww.InsertRemoveSDCardRequest(action=hww.InsertRemoveSDCardRequest.INSERT_CARD)\n )\n elif remove:\n request.insert_remove_sdcard.CopyFrom(\n hww.InsertRemoveSDCardRequest(action=hww.InsertRemoveSDCardRequest.REMOVE_CARD)\n )\n else:\n raise Exception(\"Invalid action\")\n self._msg_query(request, expected_response=\"success\")\n\n def set_mnemonic_passphrase_enabled(self, enabled):\n \"\"\"\n Enable or disable the bip39 passphrase.\n \"\"\"\n # pylint: disable=no-member\n request = hww.Request()\n request.set_mnemonic_passphrase_enabled.enabled = enabled\n self._msg_query(request, expected_response=\"success\")\n\n def _perform_attestation(self):\n \"\"\"Sends a random challenge and verifies that the response can be verified with\n Shift's root attestation pubkeys. Returns True if the verification is successful.\"\"\"\n\n challenge = os.urandom(32)\n response = self._query(OP_ATTESTATION + challenge)\n if response[:1] != RESPONSE_SUCCESS:\n return False\n\n # parse data\n response = response[1:]\n bootloader_hash, response = response[:32], response[32:]\n device_pubkey_bytes, response = response[:64], response[64:]\n certificate, response = response[:64], response[64:]\n root_pubkey_identifier, response = response[:32], response[32:]\n challenge_signature, response = response[:64], response[64:]\n\n # check attestation\n if root_pubkey_identifier not in ATTESTATION_PUBKEYS_MAP:\n # root pubkey could not be identified.\n return False\n\n root_pubkey_bytes_uncompressed = ATTESTATION_PUBKEYS_MAP[root_pubkey_identifier]\n root_pubkey = ecdsa.VerifyingKey.from_string(\n root_pubkey_bytes_uncompressed[1:], ecdsa.curves.SECP256k1\n )\n\n device_pubkey = ecdsa.VerifyingKey.from_string(device_pubkey_bytes, ecdsa.curves.NIST256p)\n\n try:\n # Verify certificate\n if not root_pubkey.verify(\n certificate, bootloader_hash + device_pubkey_bytes, hashfunc=hashlib.sha256\n ):\n return False\n\n # Verify challenge\n if not device_pubkey.verify(challenge_signature, challenge, hashfunc=hashlib.sha256):\n return False\n except ecdsa.BadSignatureError:\n return False\n return True\n\n def reboot(self):\n \"\"\"TODO: Document\"\"\"\n # pylint: disable=no-member\n request = hww.Request()\n request.reboot.CopyFrom(hww.RebootRequest())\n try:\n self._msg_query(request)\n except OSError:\n # In case of reboot we can't read the response.\n return True\n except Bitbox02Exception:\n return False\n return True\n\n def _eth_msg_query(self, eth_request, expected_response=None):\n \"\"\"\n Same as _msg_query, but one nesting deeper for ethereum messages.\n \"\"\"\n # pylint: disable=no-member\n request = hww.Request()\n request.eth.CopyFrom(eth_request)\n eth_response = self._msg_query(request, expected_response=\"eth\").eth\n if (\n expected_response is not None\n and eth_response.WhichOneof(\"response\") != expected_response\n ):\n raise Exception(\n \"Unexpected response: {}, expected: {}\".format(\n eth_response.WhichOneof(\"response\"), expected_response\n )\n )\n return eth_response\n\n def eth_pub(\n self, keypath=None, coin=hww.ETH, output_type=hww.ETHPubRequest.ADDRESS, display=True\n ):\n \"\"\"\n keypath is a list of child derivation numbers.\n e.g. m/44'/60'/0'/0/5 corresponds to [44+HARDENED, 60+HARDENED, 0+HARDENED, 0, 5].\n \"\"\"\n # pylint: disable=no-member\n keypath = [] if keypath is None else keypath\n request = hww.ETHRequest()\n request.pub.CopyFrom(\n hww.ETHPubRequest(coin=coin, keypath=keypath, output_type=output_type, display=display)\n )\n return self._eth_msg_query(request, expected_response=\"pub\").pub.pub\n\n def eth_sign(self, transaction, keypath=None, coin=hww.ETH):\n \"\"\"\n transaction should be given as a full rlp encoded eth transaction.\n \"\"\"\n nonce, gas_price, gas_limit, recipient, value, data, _, _, _ = rlp.decode(transaction)\n keypath = [] if keypath is None else keypath\n request = hww.ETHRequest()\n # pylint: disable=no-member\n request.sign.CopyFrom(\n hww.ETHSignRequest(\n coin=coin,\n keypath=keypath,\n nonce=nonce,\n gas_price=gas_price,\n gas_limit=gas_limit,\n recipient=recipient,\n value=value,\n data=data,\n )\n )\n return self._eth_msg_query(request, expected_response=\"sign\").sign\n", "sub_path": "py/bitbox02/bitbox02.py", "file_name": "bitbox02.py", "file_ext": "py", "file_size_in_byte": 16975, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "sys.exit", "line_number": 37, "usage_type": "call"}, {"api_name": "binascii.unhexlify", "line_number": 55, "usage_type": "call"}, {"api_name": "hashlib.sha256", "line_number": 61, "usage_type": "call"}, {"api_name": "devices.parse_device_version", "line_number": 92, "usage_type": "call"}, {"api_name": "hid.device", "line_number": 95, "usage_type": "call"}, {"api_name": "semver.VersionInfo", "line_number": 98, "usage_type": "call"}, {"api_name": "noise.connection", "line_number": 111, "usage_type": "name"}, {"api_name": "noise.connection.NoiseConnection.from_name", "line_number": 111, "usage_type": "call"}, {"api_name": "noise.connection.NoiseConnection", "line_number": 111, "usage_type": "name"}, {"api_name": "noise.connection.set_as_initiator", "line_number": 112, "usage_type": "call"}, {"api_name": "noise.connection", "line_number": 112, "usage_type": "name"}, {"api_name": "os.urandom", "line_number": 113, "usage_type": "call"}, {"api_name": "noise.connection.set_keypair_from_private_bytes", "line_number": 114, "usage_type": "call"}, {"api_name": "noise.connection", "line_number": 114, "usage_type": "name"}, {"api_name": "noise.connection.Keypair.STATIC", "line_number": 114, "usage_type": "attribute"}, {"api_name": "noise.connection.Keypair", "line_number": 114, "usage_type": "name"}, {"api_name": "noise.connection.set_prologue", "line_number": 115, "usage_type": "call"}, {"api_name": "noise.connection", "line_number": 115, "usage_type": "name"}, {"api_name": "noise.connection.start_handshake", "line_number": 116, "usage_type": "call"}, {"api_name": "noise.connection", "line_number": 116, "usage_type": "name"}, {"api_name": "noise.connection.read_message", "line_number": 117, "usage_type": "call"}, {"api_name": "noise.connection", "line_number": 117, "usage_type": "name"}, {"api_name": "noise.connection.write_message", "line_number": 117, "usage_type": "call"}, {"api_name": "noise.connection.handshake_finished", "line_number": 118, "usage_type": "attribute"}, {"api_name": "noise.connection", "line_number": 118, "usage_type": "name"}, {"api_name": "noise.connection.write_message", "line_number": 119, "usage_type": "call"}, {"api_name": "noise.connection", "line_number": 119, "usage_type": "name"}, {"api_name": "noise.connection.handshake_finished", "line_number": 120, "usage_type": "attribute"}, {"api_name": "noise.connection", "line_number": 120, "usage_type": "name"}, {"api_name": "base64.b32encode", "line_number": 121, "usage_type": "call"}, {"api_name": "noise.connection.get_handshake_hash", "line_number": 121, "usage_type": "call"}, {"api_name": "noise.connection", "line_number": 121, "usage_type": "name"}, {"api_name": "noise.connection", "line_number": 141, "usage_type": "name"}, {"api_name": "usb.hid_send_frames", "line_number": 150, "usage_type": "call"}, {"api_name": "usb.hid_read_frames", "line_number": 151, "usage_type": "call"}, {"api_name": "generated.hww_pb2.Response", "line_number": 168, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 168, "usage_type": "name"}, {"api_name": "generated.hww_pb2.Request", "line_number": 184, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 184, "usage_type": "name"}, {"api_name": "generated.hww_pb2.RandomNumberRequest", "line_number": 185, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 185, "usage_type": "name"}, {"api_name": "generated.hww_pb2.Request", "line_number": 191, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 191, "usage_type": "name"}, {"api_name": "generated.hww_pb2.DeviceInfoRequest", "line_number": 192, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 192, "usage_type": "name"}, {"api_name": "generated.hww_pb2.Request", "line_number": 205, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 205, "usage_type": "name"}, {"api_name": "generated.hww_pb2.Request", "line_number": 215, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 215, "usage_type": "name"}, {"api_name": "os.urandom", "line_number": 216, "usage_type": "call"}, {"api_name": "generated.hww_pb2.Request", "line_number": 231, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 231, "usage_type": "name"}, {"api_name": "time.time", "line_number": 232, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 233, "usage_type": "call"}, {"api_name": "generated.hww_pb2.Request", "line_number": 248, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 248, "usage_type": "name"}, {"api_name": "generated.hww_pb2.ListBackupsRequest", "line_number": 249, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 249, "usage_type": "name"}, {"api_name": "datetime.datetime.utcfromtimestamp", "line_number": 252, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 252, "usage_type": "name"}, {"api_name": "generated.hww_pb2.Request", "line_number": 260, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 260, "usage_type": "name"}, {"api_name": "generated.hww_pb2.Request", "line_number": 278, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 278, "usage_type": "name"}, {"api_name": "generated.hww_pb2.CheckBackupRequest", "line_number": 279, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 279, "usage_type": "name"}, {"api_name": "generated.hww_pb2.Request", "line_number": 294, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 294, "usage_type": "name"}, {"api_name": "generated.hww_pb2.ShowMnemonicRequest", "line_number": 295, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 295, "usage_type": "name"}, {"api_name": "generated.hww_pb2.BTC", "line_number": 307, "usage_type": "attribute"}, {"api_name": "generated.hww_pb2", "line_number": 307, "usage_type": "name"}, {"api_name": "generated.hww_pb2.BTCPubRequest", "line_number": 308, "usage_type": "attribute"}, {"api_name": "generated.hww_pb2", "line_number": 308, "usage_type": "name"}, {"api_name": "generated.hww_pb2.SCRIPT_UNKNOWN", "line_number": 309, "usage_type": "attribute"}, {"api_name": "generated.hww_pb2", "line_number": 309, "usage_type": "name"}, {"api_name": "generated.hww_pb2.Request", "line_number": 318, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 318, "usage_type": "name"}, {"api_name": "generated.hww_pb2.BTCPubRequest", "line_number": 320, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 320, "usage_type": "name"}, {"api_name": "generated.hww_pb2.Request", "line_number": 332, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 332, "usage_type": "name"}, {"api_name": "generated.hww_pb2.CheckSDCardRequest", "line_number": 333, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 333, "usage_type": "name"}, {"api_name": "generated.hww_pb2.Request", "line_number": 340, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 340, "usage_type": "name"}, {"api_name": "generated.hww_pb2.InsertRemoveSDCardRequest", "line_number": 343, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 343, "usage_type": "name"}, {"api_name": "generated.hww_pb2.InsertRemoveSDCardRequest", "line_number": 347, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 347, "usage_type": "name"}, {"api_name": "generated.hww_pb2.Request", "line_number": 358, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 358, "usage_type": "name"}, {"api_name": "os.urandom", "line_number": 366, "usage_type": "call"}, {"api_name": "ecdsa.VerifyingKey.from_string", "line_number": 385, "usage_type": "call"}, {"api_name": "ecdsa.VerifyingKey", "line_number": 385, "usage_type": "attribute"}, {"api_name": "ecdsa.curves", "line_number": 386, "usage_type": "attribute"}, {"api_name": "ecdsa.VerifyingKey.from_string", "line_number": 389, "usage_type": "call"}, {"api_name": "ecdsa.VerifyingKey", "line_number": 389, "usage_type": "attribute"}, {"api_name": "ecdsa.curves", "line_number": 389, "usage_type": "attribute"}, {"api_name": "hashlib.sha256", "line_number": 394, "usage_type": "attribute"}, {"api_name": "hashlib.sha256", "line_number": 399, "usage_type": "attribute"}, {"api_name": "ecdsa.BadSignatureError", "line_number": 401, "usage_type": "attribute"}, {"api_name": "generated.hww_pb2.Request", "line_number": 408, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 408, "usage_type": "name"}, {"api_name": "generated.hww_pb2.RebootRequest", "line_number": 409, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 409, "usage_type": "name"}, {"api_name": "generated.hww_pb2.Request", "line_number": 424, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 424, "usage_type": "name"}, {"api_name": "generated.hww_pb2.ETH", "line_number": 439, "usage_type": "attribute"}, {"api_name": "generated.hww_pb2", "line_number": 439, "usage_type": "name"}, {"api_name": "generated.hww_pb2.ETHPubRequest", "line_number": 439, "usage_type": "attribute"}, {"api_name": "generated.hww_pb2.ETHRequest", "line_number": 447, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 447, "usage_type": "name"}, {"api_name": "generated.hww_pb2.ETHPubRequest", "line_number": 449, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 449, "usage_type": "name"}, {"api_name": "generated.hww_pb2.ETH", "line_number": 453, "usage_type": "attribute"}, {"api_name": "generated.hww_pb2", "line_number": 453, "usage_type": "name"}, {"api_name": "rlp.decode", "line_number": 457, "usage_type": "call"}, {"api_name": "generated.hww_pb2.ETHRequest", "line_number": 459, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 459, "usage_type": "name"}, {"api_name": "generated.hww_pb2.ETHSignRequest", "line_number": 462, "usage_type": "call"}, {"api_name": "generated.hww_pb2", "line_number": 462, "usage_type": "name"}]} +{"seq_id": "234975468", "text": "from tkinter import *\nimport numpy as np\nimport pandas as pd\nfrom subprocess import call\nfrom PIL import ImageTk,Image\nimport pyttsx3\nimport datetime\nimport speech_recognition as sr\nimport sqlite3 as s\n\n\n# from gui_stuff import *\ndf=pd.read_csv(\"Training.csv\")\ntr=pd.read_csv(\"Testing.csv\")\n\n\nl1=['itching','skin rash','nodal skin eruptions','continuous sneezing','shivering','chills','joint pain','stomach pain',\n'acidity','ulcers on tongue','muscle wasting','vomiting','burning micturition','spotting urination','fatigue',\n'weight gain','anxiety','cold hands and feets','mood swings','weight loss','restlessness','lethargy','patches in throat',\n'irregular sugar level','cough','high fever','sunken eyes','breathlessness','sweating','dehydration','indigestion',\n'headache','yellowish skin','dark urine','nausea','loss of appetite','pain behind the eyes','back pain','constipation',\n'abdominal pain','diarrhoea','mild fever','yellow urine','yellowing of eyes','acute liver failure','fluid overload',\n'swelling of stomach','swelled lymph nodes','malaise','blurred and distorted vision','phlegm','throat irritation',\n'redness of eyes','sinus pressure','runny nose','congestion','chest pain','weakness in limbs','fast heart rate',\n'pain during bowel movements','pain in anal region','bloody stool',\n'irritation in anus','neck pain','dizziness','cramps','bruising','obesity','swollen legs','swollen blood vessels',\n'puffy face and eyes','enlarged thyroid','brittle nails','swollen extremeties','excessive hunger',\n'extra marital contacts','drying and tingling lips','slurred speech','knee pain','hip joint pain','muscle weakness',\n'stiff neck','swelling joints','movement stiffness','spinning movements','loss of balance','unsteadiness',\n'weakness of one body side','loss of smell','bladder discomfort','foul smell of urine','continuous feel of urine',\n'passage of gases','internal itching','toxic look (typhos)','depression','irritability','muscle pain',\n'altered sensorium','red spots over body','belly pain','abnormal menstruation','dischromic patches',\n'watering from eyes','increased appetite','polyuria','family history','mucoid sputum','rusty sputum',\n'lack of concentration','visual disturbances','receiving blood transfusion','receiving unsterile injections',\n'coma','stomach bleeding','distention of abdomen','history of alcohol consumption','fluid overload','blood in sputum',\n'prominent veins on calf','palpitations','painful walking','pus filled pimples','blackheads','scurring','skin peeling',\n'silver like dusting','small dents in nails','inflammatory nails','blister','red sore around nose','yellow crust ooze']\n\n\n\n\ndisease=['Fungal infection','Allergy','GERD','Chronic cholestasis','Drug Reaction',\n'Peptic ulcer diseae','AIDS','Diabetes','Gastroenteritis','Bronchial Asthma','Hypertension',\n' Migraine','Cervical spondylosis',\n'Paralysis (brain hemorrhage)','Jaundice','Malaria','Chicken pox','Dengue','Typhoid','hepatitis A',\n'Hepatitis B','Hepatitis C','Hepatitis D','Hepatitis E','Alcoholic hepatitis','Tuberculosis',\n'Common Cold','Pneumonia','Dimorphic hemmorhoids(piles)',\n'Heartattack','Varicoseveins','Hypothyroidism','Hyperthyroidism','Hypoglycemia','Osteoarthristis',\n'Arthritis','(vertigo) Paroymsal Positional Vertigo','Acne','Urinary tract infection','Psoriasis',\n'Impetigo']\n\nl2=[]\nfor x in range(0,len(l1)):\n l2.append(0)\n\n# TESTING DATA df -------------------------------------------------------------------------------------\ndf=pd.read_csv(\"Training.csv\")\ndf.columns = df.columns.str.replace('dischromic _patches', 'dischromic patches')\ndf.columns = df.columns.str.replace('spotting_ urination','spotting urination' )\ndf.columns = df.columns.str.replace('_', ' ')\n\n\ndf.replace({'prognosis':{'Fungal infection':0,'Allergy':1,'GERD':2,'Chronic cholestasis':3,'Drug Reaction':4,\n'Peptic ulcer diseae':5,'AIDS':6,'Diabetes ':7,'Gastroenteritis':8,'Bronchial Asthma':9,'Hypertension ':10,\n'Migraine':11,'Cervical spondylosis':12,\n'Paralysis (brain hemorrhage)':13,'Jaundice':14,'Malaria':15,'Chicken pox':16,'Dengue':17,'Typhoid':18,'hepatitis A':19,\n'Hepatitis B':20,'Hepatitis C':21,'Hepatitis D':22,'Hepatitis E':23,'Alcoholic hepatitis':24,'Tuberculosis':25,\n'Common Cold':26,'Pneumonia':27,'Dimorphic hemmorhoids(piles)':28,'Heart attack':29,'Varicose veins':30,'Hypothyroidism':31,\n'Hyperthyroidism':32,'Hypoglycemia':33,'Osteoarthristis':34,'Arthritis':35,\n'(vertigo) Paroymsal Positional Vertigo':36,'Acne':37,'Urinary tract infection':38,'Psoriasis':39,\n'Impetigo':40}},inplace=True)\n\n# print(df.head())\n\nX= df[l1]\ny = df[[\"prognosis\"]]\nnp.ravel(y)\n# print(y)\n\n# TRAINING DATA tr --------------------------------------------------------------------------------\ntr=pd.read_csv(\"Testing.csv\")\ntr.columns = tr.columns.str.replace('dischromic _patches', 'dischromic patches')\ntr.columns = tr.columns.str.replace('spotting_ urination','spotting urination' )\ntr.columns = tr.columns.str.replace('_', ' ')\n\ntr.replace({'prognosis':{'Fungal infection':0,'Allergy':1,'GERD':2,'Chronic cholestasis':3,'Drug Reaction':4,\n'Peptic ulcer diseae':5,'AIDS':6,'Diabetes ':7,'Gastroenteritis':8,'Bronchial Asthma':9,'Hypertension ':10,\n'Migraine':11,'Cervical spondylosis':12,\n'Paralysis (brain hemorrhage)':13,'Jaundice':14,'Malaria':15,'Chicken pox':16,'Dengue':17,'Typhoid':18,'hepatitis A':19,\n'Hepatitis B':20,'Hepatitis C':21,'Hepatitis D':22,'Hepatitis E':23,'Alcoholic hepatitis':24,'Tuberculosis':25,\n'Common Cold':26,'Pneumonia':27,'Dimorphic hemmorhoids(piles)':28,'Heart attack':29,'Varicose veins':30,'Hypothyroidism':31,\n'Hyperthyroidism':32,'Hypoglycemia':33,'Osteoarthristis':34,'Arthritis':35,\n'(vertigo) Paroymsal Positional Vertigo':36,'Acne':37,'Urinary tract infection':38,'Psoriasis':39,\n'Impetigo':40}},inplace=True)\n\nX_test= tr[l1]\ny_test = tr[[\"prognosis\"]]\nnp.ravel(y_test)\n# ------------------------------------------------------------------------------------------------------\nlst=[]\ndiseaselist=[]\ndef DecisionTree():\n\n from sklearn import tree\n\n clf3 = tree.DecisionTreeClassifier() # empty model of the decision tree\n clf3 = clf3.fit(X,y)\n\n # calculating accuracy-------------------------------------------------------------------\n from sklearn.metrics import accuracy_score\n y_pred=clf3.predict(X_test)\n pre=(accuracy_score(y_test, y_pred,normalize=False))\n # -----------------------------------------------------\n\n psymptoms = [Symptom1.get(),Symptom2.get(),Symptom3.get(),Symptom4.get(),Symptom5.get()]\n\n for k in range(0,len(l1)):\n # print (k,)\n for z in psymptoms:\n if(z==l1[k]):\n l2[k]=1\n\n inputtest = [l2]\n predict = clf3.predict(inputtest)\n predicted=predict[0]\n\n h='no'\n for a in range(0,len(disease)):\n if(predicted == a):\n h='yes'\n break\n\n\n if (h=='yes'):\n t1.delete(\"1.0\", END)\n t1.insert(END, disease[a])\n else:\n t1.delete(\"1.0\", END)\n t1.insert(END, \"Not Found\")\n lst.append(int(pre))\n diseaselist.append(disease[a])\n \n \n\ndef randomforest():\n from sklearn.ensemble import RandomForestClassifier\n clf4 = RandomForestClassifier()\n clf4 = clf4.fit(X,np.ravel(y))\n\n # calculating accuracy-------------------------------------------------------------------\n from sklearn.metrics import accuracy_score\n y_pred=clf4.predict(X_test)\n print(accuracy_score(y_test, y_pred))\n pre1=(accuracy_score(y_test, y_pred,normalize=False))\n # -----------------------------------------------------\n\n psymptoms = [Symptom1.get(),Symptom2.get(),Symptom3.get(),Symptom4.get(),Symptom5.get()]\n\n for k in range(0,len(l1)):\n for z in psymptoms:\n if(z==l1[k]):\n l2[k]=1\n\n inputtest = [l2]\n predict = clf4.predict(inputtest)\n predicted=predict[0]\n\n h='no'\n for b in range(0,len(disease)):\n if(predicted == b):\n h='yes'\n break\n\n if (h=='yes'):\n t2.delete(\"1.0\", END)\n t2.insert(END, disease[b])\n else:\n t2.delete(\"1.0\", END)\n t2.insert(END, \"Not Found\")\n lst.append(int(pre1))\n diseaselist.append(disease[b])\n\ndef NaiveBayes():\n try:\n client=s.connect(\"C://Users//Hackers world//Desktop//projects//healthprediction//register.db\")\n cu=client.cursor()\n cu.execute(\"create table patient(name varchar(50),disease varchar(80))\")\n except:\n pass\n\n from sklearn.naive_bayes import GaussianNB\n gnb = GaussianNB()\n gnb=gnb.fit(X,np.ravel(y))\n\n # calculating accuracy-------------------------------------------------------------------\n from sklearn.metrics import accuracy_score\n y_pred=gnb.predict(X_test)\n print(accuracy_score(y_test, y_pred))\n pre2=(accuracy_score(y_test, y_pred,normalize=False))\n # -----------------------------------------------------\n\n psymptoms = [Symptom1.get(),Symptom2.get(),Symptom3.get(),Symptom4.get(),Symptom5.get()]\n for k in range(0,len(l1)):\n for z in psymptoms:\n if(z==l1[k]):\n l2[k]=1\n\n inputtest = [l2]\n predict = gnb.predict(inputtest)\n predicted=predict[0]\n\n h='no'\n for c in range(0,len(disease)):\n if(predicted == c):\n h='yes'\n break\n\n if (h=='yes'):\n t3.delete(\"1.0\", END)\n t3.insert(END, disease[c])\n else:\n t3.delete(\"1.0\", END)\n t3.insert(END, \"Not Found\")\n lst.append(int(pre2))\n diseaselist.append(disease[c])\n print(diseaselist)\n # IMPORTING COLLECTION\n import collections\n final=([item for item, count in collections.Counter(diseaselist).items() if count > 1])\n t4.insert(END,final)\n cu.execute(\"insert into patient values(%r,%r)\"%(NameEn.get(),diseaselist[2]))\n client.commit() \n \n# gui_stuff------------------------------------------------------------------------------------\n\nroot = Tk(className=' Disease Predictor')\nroot.geometry('450x740')\nroot.maxsize(740,450)\nroot.minsize(740,450)\n\n\n\n# FOR BACKGROUND IMAGE\ni=ImageTk.PhotoImage(Image.open('C:\\\\Users\\\\Hackers world\\\\Desktop\\\\projects\\\\healthprediction\\\\images\\\\png.png'))\nl=Label(root,image=i)\nl.grid(row=0,column=0)\n# entry variables\nSymptom1 = StringVar()\nSymptom1.set(None)\nSymptom2 = StringVar()\nSymptom2.set(None)\nSymptom3 = StringVar()\nSymptom3.set(None)\nSymptom4 = StringVar()\nSymptom4.set(None)\nSymptom5 = StringVar()\nSymptom5.set(None)\nName = StringVar()\n\ndef sym1():\n Symptom1.set(None)\ndef sym2():\n Symptom2.set(None)\ndef sym3():\n Symptom3.set(None)\ndef sym4():\n Symptom4.set(None)\ndef sym5():\n Symptom5.set(None)\n\n#logout function\ndef logoutfun():\n root.destroy()\n call([\"python\",\"home.py\"])\n\n\n# Heading\nw2 = Label(root,width=62,height=1,justify='center', text=\" Disease Predictor \",font=('times',15,'bold'), fg=\"white\",bg='darkcyan')\n\nw2.place(x=0,y=0)\n\n# labels\nNameLb = Label(root, text=\"Patient Name\", font=('times',15,'bold'),fg=\"white\",bg='darkcyan')\nNameLb.place(x=10,y=72)\n\n\nS1Lb = Label(root, text=\"Symptom 1 \", font=('times',15,'bold'),fg=\"white\",bg='darkcyan')\nS1Lb.place(x=10,y=118)\n\nS2Lb = Label(root, text=\"Symptom 2 \", font=('times',15,'bold'),fg=\"white\",bg='darkcyan')\nS2Lb.place(x=10,y=158)\n\nS3Lb = Label(root, text=\"Symptom 3 \", font=('times',15,'bold'),fg=\"white\",bg='darkcyan')\nS3Lb.place(x=10,y=198)\n\nS4Lb = Label(root, text=\"Symptom 4 \",font=('times',15,'bold'), fg=\"white\",bg='darkcyan')\nS4Lb.place(x=10,y=238)\n\nS5Lb = Label(root, text=\"Symptom 5 \",font=('times',15,'bold'), fg=\"white\",bg='darkcyan')\nS5Lb.place(x=10,y=278)\n\n\nlrLb = Label(root, text=\"DecisionTree\", font=('times',15,'bold'),fg=\"white\",bg='green')\nlrLb.place(x=10,y=322)\n\ndestreeLb = Label(root, text=\"Rand_Forest \", font=('times',15,'bold'),fg=\"white\",bg='green')\ndestreeLb.place(x=10,y=362)\n\nranfLb = Label(root, text=\"NaiveBayes \",font=('times',15,'bold'), fg=\"white\",bg='green')\nranfLb.place(x=10,y=402)\n# entries\nOPTIONS = sorted(l1)\n\nNameEn = Entry(root, textvariable=Name,width=25)\nNameEn.place(x=140,y=73)\n\nS1En = OptionMenu(root, Symptom1,*OPTIONS)\nS1En.place(x=140,y=113)\n\nS2En = OptionMenu(root, Symptom2,*OPTIONS)\nS2En.place(x=140,y=153)\n\n\nS3En = OptionMenu(root, Symptom3,*OPTIONS)\nS3En.place(x=140,y=193)\n\n\nS4En = OptionMenu(root, Symptom4,*OPTIONS)\nS4En.place(x=140,y=233)\n\n\nS5En = OptionMenu(root, Symptom5,*OPTIONS)\nS5En.place(x=140,y=273)\n\n# clear button for option menu\nbtn1=Button(root,text='CLEAR',command=sym1)\nbtn1.place(x=420,y=116)\nbtn2=Button(root,text='CLEAR',command=sym2)\nbtn2.place(x=420,y=156)\nbtn3=Button(root,text='CLEAR',command=sym3)\nbtn3.place(x=420,y=196)\nbtn4=Button(root,text='CLEAR',command=sym4)\nbtn4.place(x=420,y=236)\nbtn5=Button(root,text='CLEAR',command=sym5)\nbtn5.place(x=420,y=276)\n\n\n\ndst = Button(root, text=\"DecisionTree \", font=('times',15,'bold'),command=DecisionTree,bg=\"green\",fg=\"white\")\ndst.place(x=600,y=113)\n\nrnf = Button(root, text=\"Randomforest\", font=('times',15,'bold'),command=randomforest,bg=\"green\",fg=\"white\")\nrnf.place(x=600,y=173)\n\nlr = Button(root, text=\"NaiveBayes \",font=('times',15,'bold'), command=NaiveBayes,bg=\"green\",fg=\"white\")\nlr.place(x=600,y=233)\n\n#textfileds\nt1 = Text(root, height=1, width=40,bg=\"orange\",fg=\"black\")\nt1.place(x=150,y=323)\n\nt2 = Text(root, height=1, width=40,bg=\"orange\",fg=\"black\")\nt2.place(x=150,y=363)\n\nt3 = Text(root, height=1, width=40,bg=\"orange\",fg=\"black\")\nt3.place(x=150,y=403)\nfl=Label(root,text=' FINAL PREDICTION ',font=('times',10,'bold'),bg='darkcyan',fg='black')\nfl.place(x=600,y=283)\nt4=Text(root,width=18,height=40)\nt4.place(x=600,y=303)\n\n\n# Speech recognintion\n\ndef SpeechRecog():\n \n #voices \n engine=pyttsx3.init('sapi5')\n voices=engine.getProperty('voices')\n engine.setProperty('voice',voices[0].id)\n\n #for speak\n def speak (audio):\n engine.say(audio)\n engine.runAndWait()\n\n #wishing\n\n def wishme():\n speak(\"hello, this is disease predictor\")\n\n\n # it takes microphone input from user and return string output\n def takeCommand():\n r=sr.Recognizer()\n with sr.Microphone() as source:\n speak(\"please tell your name.\")\n print('listening...')\n r.pause_threshold=1\n audio=r.listen(source)\n try:\n print('recogninzing...')\n query=r.recognize_google(audio,language='en-in')\n print(query)\n Name.set(query)\n except Exception as e:\n print('say that again please...')\n return takeCommand()\n \n nolist=['first','second','third','fourth','fifth']\n symoptions=list(set(OPTIONS))\n \n def symptomsrec():\n r=sr.Recognizer()\n with sr.Microphone() as source:\n for i in range(len(nolist)):\n speak('your'+ nolist[i]+ 'symptom please')\n print('listening..')\n r.pause_threshold=1\n audio=r.listen(source)\n try:\n print('recongnizing...')\n query=r.recognize_google(audio,language='en-in')\n print(query)\n if query in symoptions:\n if i==0:\n Symptom1.set(query)\n if i==1:\n Symptom2.set(query)\n if i==2:\n Symptom3.set(query)\n if i==3:\n Symptom4.set(query)\n if i==4:\n Symptom5.set(query) \n except:\n print('say that again please...')\n \n speak('do you want to fill your symptoms by using me.')\n r=sr.Recognizer()\n with sr.Microphone() as source:\n print('listening..')\n r.pause_threshold=1\n audio=r.listen(source)\n print('recognizing..')\n \n query1=r.recognize_google(audio,language='en-in')\n \n mainquery=[]\n mainquery.append(query1)\n if 'yes' in mainquery:\n print(mainquery)\n symptomsrec()\n else:\n pass\n \n\n\n\n #main function\n if __name__==\"__main__\":\n wishme()\n takeCommand()\n DecisionTree()\n randomforest()\n NaiveBayes()\n\n \n\n\n\n\n#mic\nmic=ImageTk.PhotoImage(Image.open('C:\\\\Users\\\\Hackers world\\\\Desktop\\\\projects\\\\healthprediction\\\\images\\\\mic.jpg'))\nmicrophone=Button(master=root,width=13,height=13,image=mic,relief='flat',command=SpeechRecog)\nmicrophone.place(x=300,y=74)\n#logout\nlogout=ImageTk.PhotoImage(Image.open('C:\\\\Users\\\\Hackers world\\\\Desktop\\\\projects\\\\healthprediction\\\\images\\\\logoutbtn.jpg'))\nlog_out=Button(master=root,width=41,height=41,image=logout,relief='flat',command=logoutfun)\nlog_out.place(x=600,y=30)\n\nbtn=Button(master=root,width=9,height=2,bg='green',fg='white',relief='flat',text='LOGOUT',font=('times',11,'bold'),command=logoutfun)\nbtn.place(x=645,y=30)\n\n\ndef reload():\n root.destroy()\n call(['python','clean_code.py'])\n\n\nrepeatbtn=Button(root,text='PREDICT AGAIN',bg='red',fg='white',width=18,height=2,font=('times',11,'bold'),command=reload)\nrepeatbtn.place(x=400,y=30)\n\nroot.mainloop()\n", "sub_path": "healthprediction/clean_code.py", "file_name": "clean_code.py", "file_ext": "py", "file_size_in_byte": 17372, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "pandas.read_csv", "line_number": 13, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 14, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 77, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 98, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 106, "usage_type": "call"}, {"api_name": "sklearn.tree", "line_number": 106, "usage_type": "name"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 112, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 148, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 153, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 154, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 185, "usage_type": "call"}, {"api_name": "sklearn.naive_bayes.GaussianNB", "line_number": 192, "usage_type": "call"}, {"api_name": "numpy.ravel", "line_number": 193, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 198, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 199, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 229, "usage_type": "call"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 244, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 244, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 244, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 244, "usage_type": "name"}, {"api_name": "subprocess.call", "line_number": 274, "usage_type": "call"}, {"api_name": "pyttsx3.init", "line_number": 378, "usage_type": "call"}, {"api_name": "speech_recognition.Recognizer", "line_number": 395, "usage_type": "call"}, {"api_name": "speech_recognition.Microphone", "line_number": 396, "usage_type": "call"}, {"api_name": "speech_recognition.Recognizer", "line_number": 414, "usage_type": "call"}, {"api_name": "speech_recognition.Microphone", "line_number": 415, "usage_type": "call"}, {"api_name": "speech_recognition.Recognizer", "line_number": 440, "usage_type": "call"}, {"api_name": "speech_recognition.Microphone", "line_number": 441, "usage_type": "call"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 474, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 474, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 474, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 474, "usage_type": "name"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 478, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 478, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 478, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 478, "usage_type": "name"}, {"api_name": "subprocess.call", "line_number": 488, "usage_type": "call"}]} +{"seq_id": "446967454", "text": "from django.contrib.auth import login\nfrom django.http import JsonResponse, HttpResponseRedirect\nfrom django.shortcuts import render, redirect\nfrom .forms import *\nfrom django.contrib import messages\nimport json\nfrom django.template.loader import render_to_string\nfrom index.views import get_user_profile_photo, get_org_photo\n\n\n# with open('data.json', 'r', encoding='utf-8') as f:\n# text = json.load(f)\n# for i in text:\n# org = Organization.objects.create(title=str(text[i]['title']).replace('\\n', ''),\n# description=str(text[i]['description']),\n# rating=str(text[i]['rating']).replace('\\n', ''),\n# type_of_kitchen=str(text[i]['type_of_kitchen']).replace('\\n', ''),\n# type_of_establishment=str(text[i]['type_of_establishment']).replace('\\n', ''),\n# services=str(text[i]['services']).replace('\\n', ''),\n# phone=str(text[i]['phone']),\n# address=str(text[i]['address']).replace('\\n', ''),\n# working_hours=str(text[i]['working_hours']).replace('\\n', ''),\n# website=str(text[i]['website']),\n# average_check=str(text[i]['average_check']).replace('\\n', ''))\n# org.save()\n# user = org.owner_user\n# all_review = text[i]['reviews_link']\n# if all_review is not None:\n# for j in all_review:\n# review = Review.objects.create(\n# owner_user=user,\n# organization=org,\n# owner_name=str(all_review[j]['name']).replace('\\n', ''),\n# rating=str(all_review[j]['rating']).replace('\\n', ''),\n# content=str(all_review[j]['content'])\n# )\n# review.save()\n\n\ndef create_org(request):\n if request.user.is_authenticated:\n if request.method == 'POST':\n form = CreateOrganizationForm(request.POST)\n form_Photo = CreateOrganizationPhotoForm(request.POST, request.FILES)\n if form.is_valid():\n organization = form.save(commit=False)\n organization.phone = split_toList(organization.phone)\n organization.website = split_toList(organization.website)\n organization.owner_user = request.user\n organization.save()\n messages.success(request, 'Заведение добавлено')\n if form_Photo.is_valid():\n photo = form_Photo.save(commit=False)\n photo.organization = organization\n photo.save()\n return redirect('index')\n else:\n error = 'Ошибка проверьте ввод данных'\n return render(request, 'organization/add_organization.html', {'form': form,\n 'formPhoto': form_Photo,\n 'errors': error})\n else:\n form = CreateOrganizationForm()\n form_Photo = CreateOrganizationPhotoForm()\n else:\n return redirect('login')\n return render(request, 'organization/add_organization.html', {'form': form,\n 'formPhoto': form_Photo})\n\n\ndef top_ten(request):\n content = []\n org = Organization.objects.all().filter(rating=5, verified=True)\n for item in org:\n org_review = [item, len(Review.objects.all().filter(organization=item))]\n content.append(org_review)\n content.sort(key=lambda i: i[1], reverse=True)\n content = content[:10]\n result = []\n for item in content:\n result.append(item[0])\n org_n_photo = get_org_photo(result)\n data = {\n 'org': org_n_photo,\n 'user_profile_photo': get_user_profile_photo(request)\n }\n return render(request, 'organization/top_ten.html', data)\n\n\ndef org_details(request, pk):\n org = Organization.objects.get(id=pk, verified=True)\n reviews = Review.objects.all().order_by('-id').filter(organization=org)[:20]\n photo = OrganizationPhoto.objects.all().filter(organization=org)\n if len(photo) != 0:\n photo = photo[0]\n else:\n photo = 'img_from_user/foto.png'\n phone = str_to_list(org.phone)\n rev_n_photo = review_user_photo(reviews)\n website = org.website.split(',')[0].replace('[', '').replace(\"'\", '')\n context = {\n 'org': org,\n 'phones': phone,\n 'website': website,\n 'reviews': rev_n_photo,\n 'photo': photo,\n 'user_profile_photo': get_user_profile_photo(request),\n }\n if request.method == \"POST\":\n form = CreateReviewForm(request.POST)\n if form.is_valid():\n review = form.save(commit=False)\n review.owner_user = request.user\n review.owner_name = request.user.first_name\n review.organization = Organization.objects.get(id=pk)\n review.save()\n return redirect('org_details', pk)\n return render(request, 'organization/org_details.html', context)\n\n\ndef load_more_reviews(request):\n if request.method == 'GET':\n last_review_id = request.GET.get('lastReviewId')\n org_id = request.GET.get('orgId')\n reviews = Review.objects.all().order_by('-id').filter(id__lt=last_review_id,\n organization_id=org_id)[:20]\n more_review = review_user_photo(reviews)\n rendered = render_to_string('organization/load_more_ajax.html', {'reviews': more_review})\n if not more_review:\n return JsonResponse({'data': False})\n data = {\n 'success': True,\n 'rendered': rendered\n }\n print(data)\n return JsonResponse(data)\n return JsonResponse({'data': False})\n\n\ndef add_review_ajax(request):\n if request.method == 'GET':\n form = CreateReviewForm()\n token = request.GET.get('csrfmiddlewaretoken')\n rendered = render_to_string('organization/add_review_ajax.html',\n {'csrf_token': token,\n 'form': form})\n data = {\n 'success': True,\n 'rendered': rendered\n }\n return JsonResponse(data)\n return JsonResponse({'data': False})\n\n\ndef review_user_photo(values):\n reviews = []\n for r in values:\n rev_n_photo = [r]\n photo = UserPhoto.objects.all().filter(user=r.owner_user)\n if len(photo) != 0:\n for p in photo:\n if p.user.id == r.owner_user.id:\n rev_n_photo.append(p.photo)\n else:\n rev_n_photo.append('img_from_user/foto.png')\n reviews.append(rev_n_photo)\n return reviews\n\n\ndef split_toList(text):\n text.replace(' ', '')\n list_text = text.split(',')\n return list_text\n\n\ndef str_to_list(text):\n a = text.split(',')\n res = []\n for i in a:\n k = i.split(':')\n if len(k) >= 2:\n res.append(str(k[1]).replace(\"'\", '').replace(']', ''))\n else:\n return ''\n return res\n", "sub_path": "organization/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 7465, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "django.contrib.messages.success", "line_number": 51, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 51, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 56, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 59, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 66, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 67, "usage_type": "call"}, {"api_name": "index.views.get_org_photo", "line_number": 82, "usage_type": "call"}, {"api_name": "index.views.get_user_profile_photo", "line_number": 85, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 87, "usage_type": "call"}, {"api_name": "index.views.get_user_profile_photo", "line_number": 107, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 117, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 118, "usage_type": "call"}, {"api_name": "django.template.loader.render_to_string", "line_number": 128, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 130, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 136, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 137, "usage_type": "call"}, {"api_name": "django.template.loader.render_to_string", "line_number": 144, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 151, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 152, "usage_type": "call"}]} +{"seq_id": "360055987", "text": "import requests\r\n# json解析库,对应到lxml\r\nimport json\r\n# json的解析语法,对应到xpath\r\nimport jsonpath\r\n\r\nproxies = {\r\n 'http': '120.83.106.8:9999',\r\n # 'https': '175.148.71.202:9999'\r\n}\r\nheaders={\r\n 'User-Agent':'Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/34.0.1847.137 Safari/537.36 LBBROWSER',\r\n }\r\nresponse = requests.get(url = \"请输入json网站\", headers=headers,proxies=proxies)\r\n\r\nlottery_message=json.dumps(response.json(), ensure_ascii=False)\r\n\r\n# # 读取reponse\r\nhtml = response.text\r\n\r\n# 把json格式字符串转换成python对象\r\nhtml = json.loads(html)\r\n# 获取score节点下的数据\r\nqq = jsonpath.jsonpath(html, '$..content')\r\nfor i in qq:\r\n print(i)\r\n", "sub_path": "huoqujson.py", "file_name": "huoqujson.py", "file_ext": "py", "file_size_in_byte": 747, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "requests.get", "line_number": 14, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 16, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 22, "usage_type": "call"}, {"api_name": "jsonpath.jsonpath", "line_number": 24, "usage_type": "call"}]} +{"seq_id": "654201029", "text": "from discord.ext import commands\nfrom discord import Embed, Color\nfrom aiohttp import ClientSession\nfrom random import choice\nfrom ast import literal_eval\n\n\nasync def get(session: object, url: object) -> object:\n async with session.get(url) as response:\n return await response.text()\n\n\nclass Photo(commands.Cog):\n \"\"\"Search for photos.\n \"\"\"\n\n def __init__(self, bot):\n self.bot = bot\n\n @commands.command(name='img')\n async def _image(self, ctx, *, query):\n image = Embed(color=Color.dark_teal())\n image.title = query.title()\n if len(query.split()) > 1:\n query = '+'.join(query.split())\n key = '17191614-063633dedf733f61470d1198b'\n base = f'https://pixabay.com/api/?key={key}&q={query}&lang=en&per_page=3'\n async with ClientSession() as session:\n url = await get(session, base)\n url = literal_eval(url)\n url = url['hits']\n url = choice(url)\n image.set_footer(text=url['tags'], icon_url=\"https://i.ibb.co/Vqgtj2z/pix.png\")\n image.set_image(url=url['largeImageURL'])\n image.set_author(name=url['user'], icon_url=url['userImageURL'], url=url['pageURL'])\n await ctx.send(embed=image)\n\n\ndef setup(bot):\n bot.add_cog(Photo(bot))\n", "sub_path": "cogs/photo.py", "file_name": "photo.py", "file_ext": "py", "file_size_in_byte": 1270, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "discord.ext.commands.Cog", "line_number": 13, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 13, "usage_type": "name"}, {"api_name": "discord.Embed", "line_number": 22, "usage_type": "call"}, {"api_name": "discord.Color.dark_teal", "line_number": 22, "usage_type": "call"}, {"api_name": "discord.Color", "line_number": 22, "usage_type": "name"}, {"api_name": "aiohttp.ClientSession", "line_number": 28, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 30, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 32, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 20, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 20, "usage_type": "name"}]} +{"seq_id": "467201406", "text": "import time\n\nimport sys\nimport cv2\nimport torch\n\nfrom tqdm import tqdm\n\nfrom yolo_human_counter import YoloHumanCounter\nfrom yolo_human_counter import plot_bboxes, plot_count\n\n\ndef main():\n print('Connecting to camera')\n cap = cv2.VideoCapture(0)\n assert cap.isOpened(), 'Unable to connect to camera!'\n device = 'cuda:0' if torch.cuda.is_available() else 'cpu'\n\n print('Loading models')\n human_counter = YoloHumanCounter('weights/yolov5s.pt', img_size=(640, 640),\n conf_thresh=0.4, iou_thresh=0.5, agnostic_nms=False,\n device=device)\n\n width, height = int(cap.get(cv2.CAP_PROP_FRAME_WIDTH)), int(cap.get(cv2.CAP_PROP_FRAME_HEIGHT))\n print(f'Starting capture, camera_fps={int(cap.get(cv2.CAP_PROP_FPS))}')\n\n # Start of demo\n win_name = 'Camera Pi Demo'\n cv2.namedWindow(win_name, cv2.WINDOW_GUI_NORMAL | cv2.WINDOW_FREERATIO)\n cv2.resizeWindow(win_name, width, height)\n pbar = tqdm(desc=f'[{win_name}]', file=sys.stdout)\n\n while True:\n start_it = time.time()\n ret, img = cap.read()\n if not ret:\n print('Unable to read camera')\n break\n num_people, det = human_counter(img, return_detection=True)\n\n # visualize\n img = plot_bboxes(img, det,\n label='Person',\n thickness=5)\n img = plot_count(img, num_people,\n label='Person',\n thickness=5)\n\n # show\n cv2.imshow(win_name, img)\n elapsed_time = time.time() - start_it\n pbar.set_description(f'[{win_name}] num_detections={num_people} elapsed_time={elapsed_time:.03f}')\n pbar.update(1)\n\n # check key pressed\n key = cv2.waitKey(1)\n if key == ord('q') or key == 27: # q or esc to quit\n break\n elif key == 32: # space to pause\n key = cv2.waitKey(0)\n if key == ord('q') or key == 27:\n break\n cv2.destroyAllWindows()\n cap.release()\n\n\nif __name__ == '__main__':\n main()\n", "sub_path": "example_camera.py", "file_name": "example_camera.py", "file_ext": "py", "file_size_in_byte": 2108, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "cv2.VideoCapture", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 17, "usage_type": "attribute"}, {"api_name": "yolo_human_counter.YoloHumanCounter", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_FRAME_WIDTH", "line_number": 24, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_HEIGHT", "line_number": 24, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FPS", "line_number": 25, "usage_type": "attribute"}, {"api_name": "cv2.namedWindow", "line_number": 29, "usage_type": "call"}, {"api_name": "cv2.WINDOW_GUI_NORMAL", "line_number": 29, "usage_type": "attribute"}, {"api_name": "cv2.WINDOW_FREERATIO", "line_number": 29, "usage_type": "attribute"}, {"api_name": "cv2.resizeWindow", "line_number": 30, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 31, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 31, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 34, "usage_type": "call"}, {"api_name": "yolo_human_counter.plot_bboxes", "line_number": 42, "usage_type": "call"}, {"api_name": "yolo_human_counter.plot_count", "line_number": 45, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 50, "usage_type": "call"}, {"api_name": "time.time", "line_number": 51, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 56, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 60, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 63, "usage_type": "call"}]} +{"seq_id": "151620429", "text": "from django.shortcuts import render, redirect\nfrom django.core.urlresolvers import reverse\nfrom django.views.generic import View\nfrom django.http import JsonResponse\nfrom django.db import transaction\nfrom django.conf import settings\nfrom django_redis import get_redis_connection\nfrom datetime import datetime\nimport os\n\nfrom utils.mixin import LoginRequiredMixin\nfrom goods.models import GoodsSKU\nfrom user.models import Address\nfrom order.models import OrderInfo, OrderGoods\n\nfrom alipay import AliPay, ISVAliPay\n\n# Create your views here.\n\n\nclass OrderPlaceView(LoginRequiredMixin, View):\n '''提交订单'''\n def post(self, request):\n ''''''\n user = request.user\n sku_ids = request.POST.getlist('sku_ids')\n if not sku_ids:\n return redirect(reverse('cart:show'))\n conn = get_redis_connection('default')\n cart_key = 'cart_%d' % user.id\n # todo:遍历信息\n skus = []\n total_count = 0\n total_amount = 0\n for sku_id in sku_ids:\n sku = GoodsSKU.objects.get(id=sku_id)\n count = conn.hget(cart_key, sku_id)\n amount = sku.price * int(count)\n sku.count = count\n sku.amount = amount\n skus.append(sku)\n total_count += int(count)\n total_amount += int(amount)\n # todo:运费有子系统\n transit_price = 10\n total_pay = transit_price + total_amount\n addrs = Address.objects.filter(user=user)\n # todo: 生成上下文\n sku_ids = ','.join(sku_ids)\n context = {\n 'skus': skus,\n 'total_count': total_count,\n 'total_amount': total_amount,\n 'transit_price': transit_price,\n 'total_pay': total_pay,\n 'addrs': addrs,\n 'sku_ids': sku_ids,\n }\n # todo: 使用模板\n return render(request, 'place_order.html', context)\n\n\nclass OrderCommitView(View):\n '''创建订单'''\n @transaction.atomic\n def post(self, request):\n user = request.user\n if not user.is_authenticated():\n return JsonResponse({'res': 0, 'errmsg': 'user not login'})\n # todo: receiver params\n addr_id = request.POST.get('addr_id')\n pay_method = request.POST.get('pay_method')\n sku_ids = request.POST.get('sku_ids')\n # todo:检验参数\n if not all([addr_id, pay_method, sku_ids]):\n return JsonResponse({'res': 1, 'errmsg': 'incomplete parameters'})\n if pay_method not in OrderInfo.PAY_METHOD.keys():\n return JsonResponse({'res': 2, 'errmsg': 'error pay'})\n try:\n addr = Address.objects.get(id=addr_id)\n except Address.DoesNotExist:\n return JsonResponse({'res': 3, 'errmsg': 'address error'})\n\n # todo: add data to OrderInfo\n order_id = datetime.now().strftime('%Y%m%d%H%M%S') + str(user.id)\n transit_price = 10\n total_count = 0\n total_amount = 0\n # set save point\n save_id = transaction.savepoint()\n try:\n order = OrderInfo.objects.create(order_id=order_id,\n user=user,\n addr=addr,\n pay_method=pay_method,\n total_count=total_count,\n total_price=total_amount,\n transit_price=transit_price)\n # todo:\n print('ok')\n sku_ids = sku_ids.split(',')\n conn = get_redis_connection('default')\n cart_key = 'cart_%d' % user.id\n for sku_id in sku_ids:\n try:\n sku = GoodsSKU.objects.select_for_update().get(id=sku_id)\n except GoodsSKU.DoesNotExist:\n transaction.savepoint_rollback(save_id)\n return JsonResponse({'res': 4, 'errmsg': 'sku error'})\n count = conn.hget(cart_key, sku_id)\n # todo: judge product reserve\n if int(count) > sku.stock:\n transaction.savepoint_rollback(save_id)\n return JsonResponse({\n 'res': 6,\n 'errmsg': 'product shortage'\n })\n # todo: add data to OrderGoods\n OrderGoods.objects.create(order=order,\n sku=sku,\n count=count,\n price=sku.price)\n # todo: update product reserve\n sku.stock -= int(count)\n sku.sales += int(count)\n sku.save()\n # todo: compute count and amount\n total_count += int(count)\n amount = sku.price * int(count)\n total_amount += amount\n # todo: update total_count and total_amount of OrderInfo\n order.total_count = total_count\n order.total_price = total_amount\n order.save()\n except Exception as e:\n transaction.savepoint_rollback(save_id)\n return JsonResponse({'res': 7, 'message': 'failure'})\n # todo: del data from shopping cart\n transaction.savepoint_commit(save_id)\n conn.hdel(cart_key, *sku_ids)\n return JsonResponse({'res': 5, 'message': 'successful'})\n\n\nclass OrderPay(View):\n '''订单支付'''\n def post(self, request):\n user = request.user\n if not user.is_authenticated():\n return JsonResponse({'res': 0, 'errmsg': 'user not login'})\n order_id = request.POST.get('order_id')\n if not order_id:\n return JsonResponse({'res': 1, 'errmsg': 'error order'})\n way = OrderInfo.objects\n try:\n order = OrderInfo.objects.get(order_id=order_id,\n user=user,\n pay_method=3,\n order_status=1)\n except OrderInfo.DoesNotExist:\n return JsonResponse({'res': 2, 'errmsg': 'error order 2'})\n print('111111')\n # 使用alipay SDK\n # 配置地址\n private_path = os.path.join(settings.BASE_DIR,\n 'apps/order/app_private_key.pem')\n public_path = os.path.join(settings.BASE_DIR,\n 'apps/order/alipay_public_key.pem')\n # 获取公私钥字符串\n app_private_key_string = open(private_path).read()\n alipay_public_key_string = open(public_path).read()\n alipay = AliPay(appid='2016101700711419',\n app_notify_url=None,\n app_private_key_string=app_private_key_string,\n alipay_public_key_string=alipay_public_key_string,\n sign_type='RSA2',\n debug=True)\n total_pay = order.total_price + order.transit_price\n order_string = alipay.api_alipay_trade_page_pay(\n out_trade_no=order_id,\n total_amount=str(total_pay),\n subject='ttsx%s' % order_id,\n return_url=None,\n notify_url=None,\n )\n pay_url = 'https://openapi.alipaydev.com/gateway.do?' + order_string\n return JsonResponse({'res': 3, 'pay_url': pay_url})\n\n\nclass CheckPayView(View):\n '''获取交易结果'''\n def post(self, request):\n user = request.user\n if not user.is_authenticated():\n return JsonResponse({'res': 0, 'errmsg': 'user not login'})\n order_id = request.POST.get('order_id')\n if not order_id:\n return JsonResponse({'res': 1, 'errmsg': 'error order'})\n\n try:\n\n order = OrderInfo.objects.get(order_id=order_id,\n user=user,\n pay_method=3,\n order_status=1)\n except OrderInfo.DoesNotExist:\n return JsonResponse({'res': 2, 'errmsg': 'error order 2'})\n print('111111')\n # 使用alipay SDK\n # 配置地址\n private_path = os.path.join(settings.BASE_DIR,\n 'apps/order/app_private_key.pem')\n public_path = os.path.join(settings.BASE_DIR,\n 'apps/order/alipay_public_key.pem')\n # 获取公私钥字符串\n app_private_key_string = open(private_path).read()\n alipay_public_key_string = open(public_path).read()\n alipay = AliPay(appid='2016101700711419',\n app_notify_url=None,\n app_private_key_string=app_private_key_string,\n alipay_public_key_string=alipay_public_key_string,\n sign_type='RSA2',\n debug=True)\n while True:\n response = alipay.api_alipay_trade_query(out_trade_no=order_id)\n code = response.get('code')\n if code == '10000' and response.get(\n 'trade_status') == 'TRADE_SUCCESS':\n trade_no = response.get('trade_no')\n order.trade_no = trade_no\n order.order_status = 4\n order.save()\n return JsonResponse({'res': 3, 'message': 'success'})\n elif code == '40004' or (code == '10000'\n and response.get('trade_status')\n == 'WAIT_BUYER_PAY'):\n # wait\n import time\n time.sleep(5)\n continue\n else:\n return JsonResponse({'res': 4, 'errmsg': 'pay fail'})\n\n\nclass CommentView(LoginRequiredMixin, View):\n \"\"\"订单评论\"\"\"\n def get(self, request, order_id):\n \"\"\"提供评论页面\"\"\"\n user = request.user\n # 校验数据\n if not order_id:\n return redirect(reverse('user:order'))\n\n try:\n order = OrderInfo.objects.get(order_id=order_id, user=user)\n except OrderInfo.DoesNotExist:\n return redirect(reverse(\"user:order\"))\n\n # 根据订单的状态获取订单的状态标题\n order.status_name = OrderInfo.ORDER_STATUS[order.order_status]\n\n # 获取订单商品信息\n order_skus = OrderGoods.objects.filter(order_id=order_id)\n for order_sku in order_skus:\n # 计算商品的小计\n amount = order_sku.count * order_sku.price\n # 动态给order_sku增加属性amount,保存商品小计\n order_sku.amount = amount\n # 动态给order增加属性order_skus, 保存订单商品信息\n order.order_skus = order_skus\n\n # 使用模板\n return render(request, \"order_comment.html\", {\"order\": order})\n\n def post(self, request, order_id):\n \"\"\"处理评论内容\"\"\"\n user = request.user\n # 校验数据\n if not order_id:\n return redirect(reverse('user:order'))\n\n try:\n order = OrderInfo.objects.get(order_id=order_id, user=user)\n except OrderInfo.DoesNotExist:\n return redirect(reverse(\"user:order\"))\n\n # 获取评论条数\n total_count = request.POST.get(\"total_count\")\n total_count = int(total_count)\n\n # 循环获取订单中商品的评论内容\n for i in range(1, total_count + 1):\n # 获取评论的商品的id\n sku_id = request.POST.get(\"sku_%d\" % i) # sku_1 sku_2\n # 获取评论的商品的内容\n content = request.POST.get('content_%d' % i,\n '') # cotent_1 content_2 content_3\n try:\n order_goods = OrderGoods.objects.get(order=order,\n sku_id=sku_id)\n except OrderGoods.DoesNotExist:\n continue\n\n order_goods.comment = content\n order_goods.save()\n\n order.order_status = 5 # 已完成\n order.save()\n\n return redirect(reverse(\"user:order\", kwargs={\"page\": 1}))\n", "sub_path": "dailyfresh/apps/order/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 12277, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "utils.mixin.LoginRequiredMixin", "line_number": 21, "usage_type": "name"}, {"api_name": "django.views.generic.View", "line_number": 21, "usage_type": "name"}, {"api_name": "user.models", "line_number": 25, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 28, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 28, "usage_type": "call"}, {"api_name": "django_redis.get_redis_connection", "line_number": 29, "usage_type": "call"}, {"api_name": "user.models.id", "line_number": 30, "usage_type": "attribute"}, {"api_name": "user.models", "line_number": 30, "usage_type": "name"}, {"api_name": "goods.models.GoodsSKU.objects.get", "line_number": 36, "usage_type": "call"}, {"api_name": "goods.models.GoodsSKU.objects", "line_number": 36, "usage_type": "attribute"}, {"api_name": "goods.models.GoodsSKU", "line_number": 36, "usage_type": "name"}, {"api_name": "user.models.Address.objects.filter", "line_number": 47, "usage_type": "call"}, {"api_name": "user.models.Address.objects", "line_number": 47, "usage_type": "attribute"}, {"api_name": "user.models.Address", "line_number": 47, "usage_type": "name"}, {"api_name": "user.models", "line_number": 47, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 60, "usage_type": "call"}, {"api_name": "django.views.generic.View", "line_number": 63, "usage_type": "name"}, {"api_name": "user.models", "line_number": 67, "usage_type": "name"}, {"api_name": "user.models.is_authenticated", "line_number": 68, "usage_type": "call"}, {"api_name": "user.models", "line_number": 68, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 69, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 76, "usage_type": "call"}, {"api_name": "order.models.OrderInfo.PAY_METHOD.keys", "line_number": 77, "usage_type": "call"}, {"api_name": "order.models.OrderInfo.PAY_METHOD", "line_number": 77, "usage_type": "attribute"}, {"api_name": "order.models.OrderInfo", "line_number": 77, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 78, "usage_type": "call"}, {"api_name": "user.models.Address.objects.get", "line_number": 80, "usage_type": "call"}, {"api_name": "user.models.Address.objects", "line_number": 80, "usage_type": "attribute"}, {"api_name": "user.models.Address", "line_number": 80, "usage_type": "name"}, {"api_name": "user.models.Address.DoesNotExist", "line_number": 81, "usage_type": "attribute"}, {"api_name": "user.models.Address", "line_number": 81, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 82, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 85, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 85, "usage_type": "name"}, {"api_name": "user.models.id", "line_number": 85, "usage_type": "attribute"}, {"api_name": "user.models", "line_number": 85, "usage_type": "name"}, {"api_name": "django.db.transaction.savepoint", "line_number": 90, "usage_type": "call"}, {"api_name": "django.db.transaction", "line_number": 90, "usage_type": "name"}, {"api_name": "order.models", "line_number": 92, "usage_type": "name"}, {"api_name": "order.models.OrderInfo.objects.create", "line_number": 92, "usage_type": "call"}, {"api_name": "order.models.OrderInfo.objects", "line_number": 92, "usage_type": "attribute"}, {"api_name": "order.models.OrderInfo", "line_number": 92, "usage_type": "name"}, {"api_name": "user.models", "line_number": 93, "usage_type": "name"}, {"api_name": "django_redis.get_redis_connection", "line_number": 102, "usage_type": "call"}, {"api_name": "user.models.id", "line_number": 103, "usage_type": "attribute"}, {"api_name": "user.models", "line_number": 103, "usage_type": "name"}, {"api_name": "goods.models.GoodsSKU.objects.select_for_update", "line_number": 106, "usage_type": "call"}, {"api_name": "goods.models.GoodsSKU.objects", "line_number": 106, "usage_type": "attribute"}, {"api_name": "goods.models.GoodsSKU", "line_number": 106, "usage_type": "name"}, {"api_name": "goods.models.GoodsSKU.DoesNotExist", "line_number": 107, "usage_type": "attribute"}, {"api_name": "goods.models.GoodsSKU", "line_number": 107, "usage_type": "name"}, {"api_name": "django.db.transaction.savepoint_rollback", "line_number": 108, "usage_type": "call"}, {"api_name": "django.db.transaction", "line_number": 108, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 109, "usage_type": "call"}, {"api_name": "django.db.transaction.savepoint_rollback", "line_number": 113, "usage_type": "call"}, {"api_name": "django.db.transaction", "line_number": 113, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 114, "usage_type": "call"}, {"api_name": "order.models.OrderGoods.objects.create", "line_number": 119, "usage_type": "call"}, {"api_name": "order.models.OrderGoods.objects", "line_number": 119, "usage_type": "attribute"}, {"api_name": "order.models.OrderGoods", "line_number": 119, "usage_type": "name"}, {"api_name": "order.models", "line_number": 119, "usage_type": "name"}, {"api_name": "order.models.total_count", "line_number": 132, "usage_type": "attribute"}, {"api_name": "order.models", "line_number": 132, "usage_type": "name"}, {"api_name": "order.models.total_price", "line_number": 133, "usage_type": "attribute"}, {"api_name": "order.models", "line_number": 133, "usage_type": "name"}, {"api_name": "order.models.save", "line_number": 134, "usage_type": "call"}, {"api_name": "order.models", "line_number": 134, "usage_type": "name"}, {"api_name": "django.db.transaction.savepoint_rollback", "line_number": 136, "usage_type": "call"}, {"api_name": "django.db.transaction", "line_number": 136, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 137, "usage_type": "call"}, {"api_name": "django.db.transaction.savepoint_commit", "line_number": 139, "usage_type": "call"}, {"api_name": "django.db.transaction", "line_number": 139, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 141, "usage_type": "call"}, {"api_name": "django.db.transaction.atomic", "line_number": 65, "usage_type": "attribute"}, {"api_name": "django.db.transaction", "line_number": 65, "usage_type": "name"}, {"api_name": "django.views.generic.View", "line_number": 144, "usage_type": "name"}, {"api_name": "user.models", "line_number": 147, "usage_type": "name"}, {"api_name": "user.models.is_authenticated", "line_number": 148, "usage_type": "call"}, {"api_name": "user.models", "line_number": 148, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 149, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 152, "usage_type": "call"}, {"api_name": "order.models.OrderInfo.objects", "line_number": 153, "usage_type": "attribute"}, {"api_name": "order.models.OrderInfo", "line_number": 153, "usage_type": "name"}, {"api_name": "order.models", "line_number": 155, "usage_type": "name"}, {"api_name": "order.models.OrderInfo.objects.get", "line_number": 155, "usage_type": "call"}, {"api_name": "order.models.OrderInfo.objects", "line_number": 155, "usage_type": "attribute"}, {"api_name": "order.models.OrderInfo", "line_number": 155, "usage_type": "name"}, {"api_name": "user.models", "line_number": 156, "usage_type": "name"}, {"api_name": "order.models.OrderInfo.DoesNotExist", "line_number": 159, "usage_type": "attribute"}, {"api_name": "order.models.OrderInfo", "line_number": 159, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 160, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 164, "usage_type": "call"}, {"api_name": "os.path", "line_number": 164, "usage_type": "attribute"}, {"api_name": "django.conf.settings.BASE_DIR", "line_number": 164, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 164, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 166, "usage_type": "call"}, {"api_name": "os.path", "line_number": 166, "usage_type": "attribute"}, {"api_name": "django.conf.settings.BASE_DIR", "line_number": 166, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 166, "usage_type": "name"}, {"api_name": "alipay.AliPay", "line_number": 171, "usage_type": "call"}, {"api_name": "order.models.total_price", "line_number": 177, "usage_type": "attribute"}, {"api_name": "order.models", "line_number": 177, "usage_type": "name"}, {"api_name": "order.models.transit_price", "line_number": 177, "usage_type": "attribute"}, {"api_name": "alipay.api_alipay_trade_page_pay", "line_number": 178, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 186, "usage_type": "call"}, {"api_name": "django.views.generic.View", "line_number": 189, "usage_type": "name"}, {"api_name": "user.models", "line_number": 192, "usage_type": "name"}, {"api_name": "user.models.is_authenticated", "line_number": 193, "usage_type": "call"}, {"api_name": "user.models", "line_number": 193, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 194, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 197, "usage_type": "call"}, {"api_name": "order.models", "line_number": 201, "usage_type": "name"}, {"api_name": "order.models.OrderInfo.objects.get", "line_number": 201, "usage_type": "call"}, {"api_name": "order.models.OrderInfo.objects", "line_number": 201, "usage_type": "attribute"}, {"api_name": "order.models.OrderInfo", "line_number": 201, "usage_type": "name"}, {"api_name": "user.models", "line_number": 202, "usage_type": "name"}, {"api_name": "order.models.OrderInfo.DoesNotExist", "line_number": 205, "usage_type": "attribute"}, {"api_name": "order.models.OrderInfo", "line_number": 205, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 206, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 210, "usage_type": "call"}, {"api_name": "os.path", "line_number": 210, "usage_type": "attribute"}, {"api_name": "django.conf.settings.BASE_DIR", "line_number": 210, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 210, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 212, "usage_type": "call"}, {"api_name": "os.path", "line_number": 212, "usage_type": "attribute"}, {"api_name": "django.conf.settings.BASE_DIR", "line_number": 212, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 212, "usage_type": "name"}, {"api_name": "alipay.AliPay", "line_number": 217, "usage_type": "call"}, {"api_name": "alipay.api_alipay_trade_query", "line_number": 224, "usage_type": "call"}, {"api_name": "order.models.trade_no", "line_number": 229, "usage_type": "attribute"}, {"api_name": "order.models", "line_number": 229, "usage_type": "name"}, {"api_name": "order.models.order_status", "line_number": 230, "usage_type": "attribute"}, {"api_name": "order.models", "line_number": 230, "usage_type": "name"}, {"api_name": "order.models.save", "line_number": 231, "usage_type": "call"}, {"api_name": "order.models", "line_number": 231, "usage_type": "name"}, {"api_name": "django.http.JsonResponse", "line_number": 232, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 238, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 241, "usage_type": "call"}, {"api_name": "utils.mixin.LoginRequiredMixin", "line_number": 244, "usage_type": "name"}, {"api_name": "django.views.generic.View", "line_number": 244, "usage_type": "name"}, {"api_name": "user.models", "line_number": 248, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 251, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 251, "usage_type": "call"}, {"api_name": "order.models", "line_number": 254, "usage_type": "name"}, {"api_name": "order.models.OrderInfo.objects.get", "line_number": 254, "usage_type": "call"}, {"api_name": "order.models.OrderInfo.objects", "line_number": 254, "usage_type": "attribute"}, {"api_name": "order.models.OrderInfo", "line_number": 254, "usage_type": "name"}, {"api_name": "user.models", "line_number": 254, "usage_type": "name"}, {"api_name": "order.models.OrderInfo.DoesNotExist", "line_number": 255, "usage_type": "attribute"}, {"api_name": "order.models.OrderInfo", "line_number": 255, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 256, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 256, "usage_type": "call"}, {"api_name": "order.models.status_name", "line_number": 259, "usage_type": "attribute"}, {"api_name": "order.models", "line_number": 259, "usage_type": "name"}, {"api_name": "order.models.OrderInfo.ORDER_STATUS", "line_number": 259, "usage_type": "attribute"}, {"api_name": "order.models.OrderInfo", "line_number": 259, "usage_type": "name"}, {"api_name": "order.models.order_status", "line_number": 259, "usage_type": "attribute"}, {"api_name": "order.models.OrderGoods.objects.filter", "line_number": 262, "usage_type": "call"}, {"api_name": "order.models.OrderGoods.objects", "line_number": 262, "usage_type": "attribute"}, {"api_name": "order.models.OrderGoods", "line_number": 262, "usage_type": "name"}, {"api_name": "order.models.order_skus", "line_number": 269, "usage_type": "attribute"}, {"api_name": "order.models", "line_number": 269, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 272, "usage_type": "call"}, {"api_name": "order.models", "line_number": 272, "usage_type": "name"}, {"api_name": "user.models", "line_number": 276, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 279, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 279, "usage_type": "call"}, {"api_name": "order.models", "line_number": 282, "usage_type": "name"}, {"api_name": "order.models.OrderInfo.objects.get", "line_number": 282, "usage_type": "call"}, {"api_name": "order.models.OrderInfo.objects", "line_number": 282, "usage_type": "attribute"}, {"api_name": "order.models.OrderInfo", "line_number": 282, "usage_type": "name"}, {"api_name": "user.models", "line_number": 282, "usage_type": "name"}, {"api_name": "order.models.OrderInfo.DoesNotExist", "line_number": 283, "usage_type": "attribute"}, {"api_name": "order.models.OrderInfo", "line_number": 283, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 284, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 284, "usage_type": "call"}, {"api_name": "order.models.OrderGoods.objects.get", "line_number": 298, "usage_type": "call"}, {"api_name": "order.models.OrderGoods.objects", "line_number": 298, "usage_type": "attribute"}, {"api_name": "order.models.OrderGoods", "line_number": 298, "usage_type": "name"}, {"api_name": "order.models", "line_number": 298, "usage_type": "name"}, {"api_name": "order.models.OrderGoods.DoesNotExist", "line_number": 300, "usage_type": "attribute"}, {"api_name": "order.models.OrderGoods", "line_number": 300, "usage_type": "name"}, {"api_name": "order.models.order_status", "line_number": 306, "usage_type": "attribute"}, {"api_name": "order.models", "line_number": 306, "usage_type": "name"}, {"api_name": "order.models.save", "line_number": 307, "usage_type": "call"}, {"api_name": "order.models", "line_number": 307, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 309, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 309, "usage_type": "call"}]} +{"seq_id": "283556602", "text": "import re\nimport random\nfrom pyleri import Choice\nfrom pyleri import Grammar\nfrom pyleri import Keyword\nfrom pyleri import Repeat\nfrom pyleri import Sequence\nfrom pyleri import end_of_statement\n\n\n# Create a Grammar Class to define your language.\nclass MyGrammar(Grammar):\n RE_KEYWORDS = re.compile('\\S+')\n r_name = Keyword('\"pyleri\"')\n k_hi = Keyword('hi')\n k_bye = Keyword('bye')\n START = Repeat(Sequence(Choice(k_hi, k_bye), r_name), mi=2)\n\n\n# Print the expected elements as a indented and numbered list.\ndef print_expecting(node_expecting, string_expecting):\n for loop, e in enumerate(node_expecting):\n string_expecting = '{}\\n\\t({}) {}'.format(string_expecting, loop, e)\n print(string_expecting)\n\n\n# Complete a string until it is valid according to the grammar.\ndef auto_correction(string, my_grammar):\n node = my_grammar.parse(string)\n print('\\nParsed string: {}'.format(node.tree.string))\n\n if node.is_valid:\n string_expecting = 'String is valid. \\nExpected: '\n print_expecting(node.expecting, string_expecting)\n\n else:\n string_expecting = 'String is NOT valid.\\nExpected: ' \\\n if not node.pos \\\n else 'String is NOT valid. \\nAfter \"{}\" expected: '.format(\n node.tree.string[:node.pos])\n print_expecting(node.expecting, string_expecting)\n\n selected = random.choice(list(node.expecting))\n string = '{} {}'.format(node.tree.string[:node.pos],\n selected\n if selected\n is not end_of_statement else '')\n\n auto_correction(string, my_grammar)\n\n\nif __name__ == '__main__':\n # Compile your grammar by creating an instance of the Grammar Class.\n my_grammar = MyGrammar()\n string = 'hello \"pyleri\"'\n auto_correction(string, my_grammar)\n", "sub_path": "test_expecting.py", "file_name": "test_expecting.py", "file_ext": "py", "file_size_in_byte": 1896, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "pyleri.Grammar", "line_number": 12, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 13, "usage_type": "call"}, {"api_name": "pyleri.Keyword", "line_number": 14, "usage_type": "call"}, {"api_name": "pyleri.Keyword", "line_number": 15, "usage_type": "call"}, {"api_name": "pyleri.Keyword", "line_number": 16, "usage_type": "call"}, {"api_name": "pyleri.Repeat", "line_number": 17, "usage_type": "call"}, {"api_name": "pyleri.Sequence", "line_number": 17, "usage_type": "call"}, {"api_name": "pyleri.Choice", "line_number": 17, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 43, "usage_type": "call"}, {"api_name": "pyleri.end_of_statement", "line_number": 47, "usage_type": "name"}]} +{"seq_id": "16450695", "text": "from django.contrib import admin\n\nfrom .models import Match\n\n\nclass MatchAdmin(admin.ModelAdmin):\n list_display = (\n 'id',\n 'original_home_team_name', 'original_away_team_name',\n 'original_home_team_odds', 'original_away_team_odds',\n 'date_of_match', 'is_active',\n 'is_live', 'is_started',\n 'is_finished', 'is_cancelled',\n )\n list_display_links = (\n 'id', 'original_home_team_name', 'original_away_team_name',\n )\n list_filter = ['date_of_match']\n\n\nadmin.site.register(Match, MatchAdmin)\n", "sub_path": "gamble/apps/bets/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 551, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "django.contrib.admin.ModelAdmin", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 6, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 21, "usage_type": "call"}, {"api_name": "models.Match", "line_number": 21, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 21, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 21, "usage_type": "name"}]} +{"seq_id": "67009261", "text": "from .models import Profile, Projects, Rating\nfrom django import forms\nfrom django.contrib.auth.models import User\n\nclass ProjectsForm(forms.ModelForm):\n class Meta:\n model = Projects\n fields = ['title', 'image','description', 'url']\n\n\nclass ProfileForm(forms.ModelForm):\n class Meta:\n model = Profile\n fields = ['name', 'profile_picture', 'bio']\n\n\nclass UserForm(forms.ModelForm):\n email = forms.EmailField(max_length=300, help_text='Required. Inform a valid email address.')\n\n class Meta:\n model = User\n fields = ('username', 'email')\n\n\nclass RatingsForm(forms.ModelForm):\n class Meta:\n model = Rating\n fields = ['design', 'usability', 'content']", "sub_path": "awwwards/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 719, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "django.forms.ModelForm", "line_number": 5, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 5, "usage_type": "name"}, {"api_name": "models.Projects", "line_number": 7, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 11, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 11, "usage_type": "name"}, {"api_name": "models.Profile", "line_number": 13, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 17, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 17, "usage_type": "name"}, {"api_name": "django.forms.EmailField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 18, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User", "line_number": 21, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 25, "usage_type": "name"}, {"api_name": "models.Rating", "line_number": 27, "usage_type": "name"}]} +{"seq_id": "491821417", "text": "# Copyright (c) 2019,20-21 NVIDIA CORPORATION & AFFILIATES.\n# All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport torch\n\n__all__ = [\n 'index_vertices_by_faces',\n 'adjacency_matrix',\n 'uniform_laplacian',\n]\n\ndef index_vertices_by_faces(vertices_features, faces):\n r\"\"\"Index vertex features to convert per vertex tensor to per vertex per face tensor.\n\n Args:\n vertices_features (torch.FloatTensor):\n vertices features, of shape\n :math:`(\\text{batch_size}, \\text{num_points}, \\text{knum})`,\n ``knum`` is feature dimension, the features could be xyz position,\n rgb color, or even neural network features.\n faces (torch.LongTensor):\n face index, of shape :math:`(\\text{num_faces}, \\text{num_vertices})`.\n Returns:\n (torch.FloatTensor):\n the face features, of shape\n :math:`(\\text{batch_size}, \\text{num_faces}, \\text{num_vertices}, \\text{knum})`.\n \"\"\"\n assert vertices_features.ndim == 3, \\\n \"vertices_features must have 3 dimensions of shape (batch_size, num_points, knum)\"\n assert faces.ndim == 2, \"faces must have 2 dimensions of shape (num_faces, num_vertices)\"\n input = vertices_features.unsqueeze(2).expand(-1, -1, faces.shape[-1], -1)\n indices = faces[None, ..., None].expand(vertices_features.shape[0], -1, -1, vertices_features.shape[-1])\n return torch.gather(input=input, index=indices, dim=1)\n\n\ndef adjacency_matrix(num_vertices, faces, sparse=True):\n r\"\"\"Calculates a adjacency matrix of a mesh.\n\n Args:\n num_vertices (int): Number of vertices of the mesh.\n faces (torch.LongTensor):\n Faces of shape :math:`(\\text{num_faces}, \\text{face_size})` of the mesh.\n sparse (bool): Whether to return a sparse tensor or not. Default: True.\n\n Returns:\n (torch.FloatTensor or torch.sparse.FloatTensor): adjacency matrix\n\n Example:\n >>> faces = torch.tensor([[0, 1, 2]])\n >>> adjacency_matrix(3, faces)\n tensor(indices=tensor([[0, 0, 1, 1, 2, 2],\n [1, 2, 0, 2, 0, 1]]),\n values=tensor([1., 1., 1., 1., 1., 1.]),\n size=(3, 3), nnz=6, layout=torch.sparse_coo)\n \"\"\"\n device = faces.device\n\n forward_i = torch.stack([faces, torch.roll(faces, 1, dims=-1)], dim=-1)\n backward_i = torch.stack([torch.roll(faces, 1, dims=-1), faces], dim=-1)\n indices = torch.cat([forward_i, backward_i], dim=1).reshape(-1, 2)\n indices = indices.unique(dim=0)\n\n if sparse:\n indices = indices.t()\n # If vertex i and j have an edge connect to it, A[i, j] = 1\n values = torch.ones(indices.shape[1], device=device)\n adjacency = torch.sparse.FloatTensor(indices, values, (num_vertices, num_vertices))\n else:\n adjacency = torch.zeros((num_vertices, num_vertices), device=device, dtype=torch.float)\n adjacency[indices[:, 0], indices[:, 1]] = 1\n\n return adjacency\n\ndef uniform_laplacian(num_vertices, faces):\n r\"\"\"Calculates the uniform laplacian of a mesh.\n :math:`L[i, j] = \\frac{1}{num\\_neighbours(i)}` if i, j are neighbours.\n :math:`L[i, j] = -1` if i == j. \n :math:`L[i, j] = 0` otherwise.\n\n Args:\n num_vertices (int): Number of vertices for the mesh.\n faces (torch.LongTensor):\n Faces of shape :math:`(\\text{num_faces}, \\text{face_size})` of the mesh.\n\n Returns:\n (torch.Tensor):\n Uniform laplacian of the mesh of size :math:`(\\text{num_vertices}, \\text{num_vertices})`\n Example:\n >>> faces = torch.tensor([[0, 1, 2]])\n >>> uniform_laplacian(3, faces)\n tensor([[-1.0000, 0.5000, 0.5000],\n [ 0.5000, -1.0000, 0.5000],\n [ 0.5000, 0.5000, -1.0000]])\n \"\"\"\n batch_size = faces.shape[0]\n\n dense_adjacency = adjacency_matrix(num_vertices, faces).to_dense()\n\n # Compute the number of neighbours of each vertex\n num_neighbour = torch.sum(dense_adjacency, dim=1).view(-1, 1)\n\n L = torch.div(dense_adjacency, num_neighbour)\n\n mask = torch.eye(num_vertices, num_vertices, device=faces.device, dtype=torch.bool)\n L = L.masked_fill_(mask, -1)\n\n # Fill NaN value with 0\n L[torch.isnan(L)] = 0\n\n return L\n", "sub_path": "kaolin/ops/mesh/mesh.py", "file_name": "mesh.py", "file_ext": "py", "file_size_in_byte": 4773, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "torch.gather", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.roll", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.roll", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.sparse.FloatTensor", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.sparse", "line_number": 79, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 81, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 81, "usage_type": "attribute"}, {"api_name": "torch.sum", "line_number": 112, "usage_type": "call"}, {"api_name": "torch.div", "line_number": 114, "usage_type": "call"}, {"api_name": "torch.eye", "line_number": 116, "usage_type": "call"}, {"api_name": "torch.bool", "line_number": 116, "usage_type": "attribute"}, {"api_name": "torch.isnan", "line_number": 120, "usage_type": "call"}]} +{"seq_id": "67654481", "text": "\nimport pdb\nimport matplotlib.pyplot as plt\nimport matplotlib.dates as mdates\nimport seaborn as sns\nfrom matplotlib.backends.backend_pdf import PdfPages\n\n\ndef plot(all_data, type='buoy'):\n\n if type=='wide':\n fig, ax = plt.subplots()\n plt.style.use('seaborn')\n ax.plot(all_data.index, all_data[col],label=col)\n ax.xaxis.set_major_locator(mdates.YearLocator())\n ax.xaxis.set_major_formatter(mdates.DateFormatter('\\n%Y'))\n ax.xaxis.set_minor_locator(mdates.MonthLocator())\n ax.grid(which='major',linewidth=0.5)\n ax.set_ylabel('Temp.')\n ax.set_xlabel('Date')\n ax.set_title('Station: 51101h; Air Temperature')\n\n pdb.set_trace()\n\n if type=='buoy':\n plt.style.use('seaborn')\n pdf = PdfPages('./buoy_1.pdf')\n for key,data in all_data.items():\n print(key)\n fig, axs = plt.subplots(4, sharex=True, tight_layout=True)\n\n for i in range(4):\n axs[i].xaxis.set_major_locator(mdates.YearLocator())\n axs[i].xaxis.set_major_formatter(mdates.DateFormatter('\\n%Y'))\n axs[i].xaxis.set_minor_locator(mdates.MonthLocator())\n axs[i].tick_params(axis=\"y\", labelsize=5)\n# axs[i].xaxis.set_minor_formatter(mdates.DateFormatter('%B'))\n axs[i].grid(which='major',color='white',linewidth=0.5)\n\n axs[0].plot(data.index, data['air_temp'], label='air_temp')\n axs[0].set_ylabel('Temp.', fontsize=6)\n axs[0].set_title('Air temperature', fontsize=8)\n\n axs[1].plot(data.index, data['average_wave_period'], label='av_wav_prd')\n axs[1].set_ylabel('Wave prd.', fontsize=6)\n axs[1].set_title('Av. wave prd.', fontsize=8)\n\n axs[2].plot(data.index, data['dominant_wave_period'], label='dom_wav_prd')\n axs[2].set_ylabel('Wave prd.', fontsize=6)\n axs[2].set_title('Dom. wave prd.', fontsize=8)\n\n axs[3].plot(data.index, data['wave_height'], label='wav_hgt')\n axs[3].set_xlabel('Date',fontsize=7)\n axs[3].set_ylabel('Wave hgt.', fontsize=6)\n axs[3].set_title('Wave height', fontsize=8)\n axs[3].tick_params(axis=\"x\", rotation=45,labelsize=6)\n plt.rcParams['font.size'] = '5'\n fig.suptitle('Station: ' + key, fontsize=10)\n \n pdf.savefig(fig)\n# break;\n pdf.close()\n\n pdb.set_trace()\n return \n", "sub_path": "assignment/strong/plot.py", "file_name": "plot.py", "file_ext": "py", "file_size_in_byte": 2492, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "matplotlib.pyplot.subplots", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 12, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 13, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 13, "usage_type": "name"}, {"api_name": "matplotlib.dates.YearLocator", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 15, "usage_type": "name"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.dates.MonthLocator", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 17, "usage_type": "name"}, {"api_name": "pdb.set_trace", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 26, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.backends.backend_pdf.PdfPages", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.dates.YearLocator", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.dates.MonthLocator", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 57, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "pdb.set_trace", "line_number": 64, "usage_type": "call"}]} +{"seq_id": "182034393", "text": "from abc import abstractmethod\nfrom typing import Any, Callable, Sequence, Tuple, Type, Union\n\nimport click\nimport numpy as np\nimport xarray as xr\n\nfrom starfish.imagestack.imagestack import ImageStack\nfrom starfish.intensity_table.intensity_table import IntensityTable\nfrom starfish.pipeline.algorithmbase import AlgorithmBase\nfrom starfish.pipeline.pipelinecomponent import PipelineComponent\nfrom starfish.types import Axes, Number, SpotAttributes\n\nCOMPONENT_NAME = \"detect_spots\"\n\n\nclass SpotFinder(PipelineComponent):\n @classmethod\n def pipeline_component_type_name(cls) -> str:\n return COMPONENT_NAME\n\n @classmethod\n def _cli_run(cls, ctx, instance):\n output = ctx.obj[\"output\"]\n blobs_stack = ctx.obj[\"blobs_stack\"]\n image_stack = ctx.obj[\"image_stack\"]\n ref_image = ctx.obj[\"reference_image_from_max_projection\"]\n if blobs_stack is not None:\n blobs_stack = ImageStack.from_path_or_url(blobs_stack) # type: ignore\n mp = blobs_stack.max_proj(Axes.ROUND, Axes.CH)\n mp_numpy = mp._squeezed_numpy(Axes.ROUND, Axes.CH)\n intensities = instance.run(\n image_stack,\n blobs_image=mp_numpy,\n reference_image_from_max_projection=ref_image,\n )\n else:\n intensities = instance.run(image_stack)\n\n # When run() returns a tuple, we only save the intensities for now\n # TODO ambrosejcarr find a way to save arbitrary detector results\n if isinstance(intensities, tuple):\n intensities = intensities[0]\n intensities.save(output)\n\n @staticmethod\n @click.group(COMPONENT_NAME)\n @click.option(\"-i\", \"--input\", required=True, type=click.Path(exists=True))\n @click.option(\"-o\", \"--output\", required=True)\n @click.option(\n '--blobs-stack', default=None, required=False, help=(\n 'ImageStack that contains the blobs. Will be max-projected across imaging round '\n 'and channel to produce the blobs_image'\n )\n )\n @click.option(\n '--reference-image-from-max-projection', default=False, is_flag=True, help=(\n 'Construct a reference image by max projecting imaging rounds and channels. Spots '\n 'are found in this image and then measured across all images in the input stack.'\n )\n )\n @click.pass_context\n def _cli(ctx, input, output, blobs_stack, reference_image_from_max_projection):\n \"\"\"detect spots\"\"\"\n print('Detecting Spots ...')\n ctx.obj = dict(\n component=SpotFinder,\n image_stack=ImageStack.from_path_or_url(input),\n output=output,\n blobs_stack=blobs_stack,\n reference_image_from_max_projection=reference_image_from_max_projection,\n )\n\n\nclass SpotFinderAlgorithmBase(AlgorithmBase):\n @classmethod\n def get_pipeline_component_class(cls) -> Type[PipelineComponent]:\n return SpotFinder\n\n @abstractmethod\n def run(\n self,\n primary_image: ImageStack,\n *args,\n ) -> Union[IntensityTable, Tuple[IntensityTable, Any]]:\n \"\"\"Finds spots in an ImageStack\"\"\"\n raise NotImplementedError()\n\n @abstractmethod\n def image_to_spots(self, data_image: Union[np.ndarray, xr.DataArray]) -> SpotAttributes:\n \"\"\"Finds spots in a 3d volume\"\"\"\n raise NotImplementedError()\n\n @staticmethod\n def _get_measurement_function(measurement_type: str) -> Callable[[Sequence], Number]:\n try:\n measurement_function = getattr(np, measurement_type)\n except AttributeError:\n raise ValueError(\n f'measurement_type must be a numpy reduce function such as \"max\" or \"mean\". '\n f'{measurement_type} not found.')\n return measurement_function\n", "sub_path": "starfish/spots/_detector/_base.py", "file_name": "_base.py", "file_ext": "py", "file_size_in_byte": 3840, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "starfish.pipeline.pipelinecomponent.PipelineComponent", "line_number": 17, "usage_type": "name"}, {"api_name": "starfish.imagestack.imagestack.ImageStack.from_path_or_url", "line_number": 29, "usage_type": "call"}, {"api_name": "starfish.imagestack.imagestack.ImageStack", "line_number": 29, "usage_type": "name"}, {"api_name": "starfish.types.Axes.ROUND", "line_number": 30, "usage_type": "attribute"}, {"api_name": "starfish.types.Axes", "line_number": 30, "usage_type": "name"}, {"api_name": "starfish.types.Axes.CH", "line_number": 30, "usage_type": "attribute"}, {"api_name": "starfish.types.Axes.ROUND", "line_number": 31, "usage_type": "attribute"}, {"api_name": "starfish.types.Axes", "line_number": 31, "usage_type": "name"}, {"api_name": "starfish.types.Axes.CH", "line_number": 31, "usage_type": "attribute"}, {"api_name": "starfish.imagestack.imagestack.ImageStack.from_path_or_url", "line_number": 68, "usage_type": "call"}, {"api_name": "starfish.imagestack.imagestack.ImageStack", "line_number": 68, "usage_type": "name"}, {"api_name": "click.group", "line_number": 47, "usage_type": "call"}, {"api_name": "click.option", "line_number": 48, "usage_type": "call"}, {"api_name": "click.Path", "line_number": 48, "usage_type": "call"}, {"api_name": "click.option", "line_number": 49, "usage_type": "call"}, {"api_name": "click.option", "line_number": 50, "usage_type": "call"}, {"api_name": "click.option", "line_number": 56, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 62, "usage_type": "attribute"}, {"api_name": "starfish.pipeline.algorithmbase.AlgorithmBase", "line_number": 75, "usage_type": "name"}, {"api_name": "typing.Type", "line_number": 77, "usage_type": "name"}, {"api_name": "starfish.pipeline.pipelinecomponent.PipelineComponent", "line_number": 77, "usage_type": "name"}, {"api_name": "starfish.imagestack.imagestack.ImageStack", "line_number": 83, "usage_type": "name"}, {"api_name": "abc.abstractmethod", "line_number": 80, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 85, "usage_type": "name"}, {"api_name": "starfish.intensity_table.intensity_table.IntensityTable", "line_number": 85, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 85, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 85, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 90, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 90, "usage_type": "attribute"}, {"api_name": "xarray.DataArray", "line_number": 90, "usage_type": "attribute"}, {"api_name": "abc.abstractmethod", "line_number": 89, "usage_type": "name"}, {"api_name": "starfish.types.SpotAttributes", "line_number": 90, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 95, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 95, "usage_type": "name"}, {"api_name": "starfish.types.Number", "line_number": 95, "usage_type": "name"}]} +{"seq_id": "132219837", "text": "from collections import OrderedDict\nimport utils\nimport torch\nimport torch.nn as nn\n\n\nclass EncoderConv1d(nn.Module):\n def __init__(self, tw, n_vars, kernels, strides, paddings, n_ch,\n actfn_name='relu', use_fc=False, fc_size=20):\n super(EncoderConv1d, self).__init__()\n self.tw = tw\n self.n_vars = n_vars\n self.kernels = kernels\n self.strides = strides\n self.paddings = paddings\n self.use_fc = use_fc\n self.n_ch =n_ch\n\n k, s, p = kernels, strides, paddings\n length = tw\n act_fn = utils.get_actfn(actfn_name)\n\n layers = OrderedDict()\n in_ch, out_ch = n_vars, self.n_ch\n for i in range(len(k)-1):\n layers[f'conv{i+1}'] = nn.Conv1d(in_ch, out_ch, k[i], s[i], p[i], bias=False)\n layers[f'bn{i+1}'] = nn.BatchNorm1d(num_features=out_ch)\n layers[f'act{i+1}'] = act_fn\n length = utils.conv1d_output_size(length, k[i], s[i], p[i])\n in_ch, out_ch = out_ch, out_ch*2\n i += 1\n layers[f'conv{i+1}'] = nn.Conv1d(in_ch, out_ch, k[i], s[i], p[i], bias=True) # No batchnorm -> Bias!\n length = utils.conv1d_output_size(length, k[i], s[i], p[i])\n self.layers = nn.Sequential(layers)\n if self.use_fc:\n self.fc = nn.Linear(length * out_ch, fc_size)\n\n def forward(self, x):\n out = self.layers(x)\n if self.use_fc:\n out = torch.flatten(out, start_dim=1)\n out = self.fc(out)\n return out\n\n\nclass DecoderConv1d(nn.Module):\n def __init__(self, tw, n_vars, kernels, strides, paddings, n_ch,\n embed_length, actfn_name='relu', outactfn_name='sigmoid', use_fc=False, fc_size=20):\n super(DecoderConv1d, self).__init__()\n self.tw = tw\n self.n_vars = n_vars\n self.kernels = [x for x in reversed(kernels)]\n self.strides = [x for x in reversed(strides)]\n self.paddings = [x for x in reversed(paddings)]\n self.n_ch = n_ch\n self.embed_length = embed_length\n self.use_fc = use_fc\n self.fc_size = fc_size\n\n k, s, p = self.kernels, self.strides, self.paddings\n\n in_ch = self.n_ch * (2 ** (len(k) - 1))\n out_ch = int(in_ch / 2)\n if self.use_fc:\n self.fc = nn.Linear(self.fc_size, embed_length * in_ch)\n length = embed_length\n act_fn = utils.get_actfn(actfn_name)\n outact_fn = utils.get_actfn(outactfn_name)\n layers = OrderedDict()\n for i in range(len(k) - 1):\n layers[f'convtr{i + 1}'] = nn.ConvTranspose1d(in_ch, out_ch, k[i], s[i], p[i], bias=False)\n layers[f'bn{i + 1}'] = nn.BatchNorm1d(num_features=out_ch)\n layers[f'act{i + 1}'] = act_fn\n length = utils.convtr1d_output_size(length, k[i], s[i], p[i])\n in_ch, out_ch = out_ch, int(out_ch / 2)\n i += 1\n layers[f'convtr{i + 1}'] = nn.ConvTranspose1d(in_ch, n_vars, k[i], s[i], p[i], bias=True)\n # length = utils.convtr1d_output_size(length, k[i], s[i], p[i])\n self.layers = nn.Sequential(layers)\n self.outact_fn = outact_fn\n\n def forward(self, z):\n if self.use_fc:\n z = self.fc(z)\n out_ch = self.n_ch * (2 ** (len(self.kernels) - 1))\n z = z.view(-1, out_ch, self.embed_length)\n out = self.layers(z)\n if self.outact_fn is not None:\n out = self.outact_fn(out)\n return out\n\n", "sub_path": "networks.py", "file_name": "networks.py", "file_ext": "py", "file_size_in_byte": 3475, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "torch.nn.Module", "line_number": 7, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 7, "usage_type": "name"}, {"api_name": "utils.get_actfn", "line_number": 21, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn.Conv1d", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}, {"api_name": "utils.conv1d_output_size", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.nn.Conv1d", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 32, "usage_type": "name"}, {"api_name": "utils.conv1d_output_size", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.nn.Sequential", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 34, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 36, "usage_type": "name"}, {"api_name": "torch.flatten", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 46, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 46, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 65, "usage_type": "name"}, {"api_name": "utils.get_actfn", "line_number": 67, "usage_type": "call"}, {"api_name": "utils.get_actfn", "line_number": 68, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.nn.ConvTranspose1d", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 71, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 72, "usage_type": "name"}, {"api_name": "utils.convtr1d_output_size", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.nn.ConvTranspose1d", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 77, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 79, "usage_type": "name"}]} +{"seq_id": "9973165", "text": "from flask import request, jsonify\nfrom flask.views import MethodView\n\nfrom app.model.genre import Genre\nfrom app.model.language import Language\nfrom app.model.video import Video\nfrom app.model.year import Year\nfrom app.utils.database import get_session\nfrom app.utils.filters import year_filters\n\n\nclass SearchApi(MethodView):\n def get(self):\n year = request.args.get('year')\n language = request.args.get('language')\n genre = request.args.get('genre')\n limit = 10 if request.args.get('limit') is None else int(request.args.get('limit'))\n page = 1 if request.args.get('page') is None else int(request.args.get('page'))\n sort = request.args.get('sort')\n search_title = request.args.get('title')\n duration = request.args.get('duration')\n\n session = get_session()\n session_query = session.query(Video)\n if year is not None:\n session_query = session_query.join(Video.year).filter(Year.value.\\\n between(year_filters[str(year)]['start'],\n year_filters[str(year)]['end']))\n if language is not None:\n session_query = session_query.join(Video.language).filter(Language.id == language)\n if genre is not None:\n genres = genre.split(\",\")\n session_query = session_query.filter(Video.genres.any(Genre.value.in_(genres)))\n if search_title is not None:\n session_query = session_query.filter(Video.title.contains(search_title))\n if duration is not None:\n if duration == 'short':\n session_query = session_query.filter(Video.duration < 300)\n else:\n session_query = session_query.filter(Video.duration > 300)\n\n videos = session_query.limit(limit).offset((page * limit) - limit).all()\n\n if sort is not None:\n sort = sort.lower()\n if sort == 'new':\n videos.sort(key=lambda vd: vd.year.value, reverse=True)\n elif sort == 'popular':\n videos.sort(key=lambda vd: vd.view_count, reverse=True)\n\n data = dict()\n data['videos'] = []\n for v in videos:\n data['videos'].append(v.serialize)\n session.close()\n return jsonify(data), 200\n", "sub_path": "app/controller/search.py", "file_name": "search.py", "file_ext": "py", "file_size_in_byte": 2377, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "flask.views.MethodView", "line_number": 12, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 14, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 14, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 14, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 15, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 15, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 16, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 16, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 16, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 17, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 17, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 18, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 18, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 19, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 19, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 19, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 20, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 20, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 20, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 21, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 21, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 21, "usage_type": "name"}, {"api_name": "app.utils.database.get_session", "line_number": 23, "usage_type": "call"}, {"api_name": "app.model.video.Video", "line_number": 24, "usage_type": "argument"}, {"api_name": "app.model.video.Video.year", "line_number": 26, "usage_type": "attribute"}, {"api_name": "app.model.video.Video", "line_number": 26, "usage_type": "name"}, {"api_name": "app.model.year.Year.value.between", "line_number": 26, "usage_type": "call"}, {"api_name": "app.model.year.Year.value", "line_number": 26, "usage_type": "attribute"}, {"api_name": "app.model.year.Year", "line_number": 26, "usage_type": "name"}, {"api_name": "app.utils.filters.year_filters", "line_number": 27, "usage_type": "name"}, {"api_name": "app.utils.filters.year_filters", "line_number": 28, "usage_type": "name"}, {"api_name": "app.model.video.Video.language", "line_number": 30, "usage_type": "attribute"}, {"api_name": "app.model.video.Video", "line_number": 30, "usage_type": "name"}, {"api_name": "app.model.language.Language.id", "line_number": 30, "usage_type": "attribute"}, {"api_name": "app.model.language.Language", "line_number": 30, "usage_type": "name"}, {"api_name": "app.model.video.Video.genres.any", "line_number": 33, "usage_type": "call"}, {"api_name": "app.model.video.Video.genres", "line_number": 33, "usage_type": "attribute"}, {"api_name": "app.model.video.Video", "line_number": 33, "usage_type": "name"}, {"api_name": "app.model.genre.Genre.value.in_", "line_number": 33, "usage_type": "call"}, {"api_name": "app.model.genre.Genre.value", "line_number": 33, "usage_type": "attribute"}, {"api_name": "app.model.genre.Genre", "line_number": 33, "usage_type": "name"}, {"api_name": "app.model.video.Video.title.contains", "line_number": 35, "usage_type": "call"}, {"api_name": "app.model.video.Video.title", "line_number": 35, "usage_type": "attribute"}, {"api_name": "app.model.video.Video", "line_number": 35, "usage_type": "name"}, {"api_name": "app.model.video.Video.duration", "line_number": 38, "usage_type": "attribute"}, {"api_name": "app.model.video.Video", "line_number": 38, "usage_type": "name"}, {"api_name": "app.model.video.Video.duration", "line_number": 40, "usage_type": "attribute"}, {"api_name": "app.model.video.Video", "line_number": 40, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 56, "usage_type": "call"}]} +{"seq_id": "221875796", "text": "import sys\nimport logging\nimport traceback\nfrom epee.errors import NoParamsError\n\ndef get_logger(name, settings={}):\n if not bool(settings):\n settings = {\n \"logfile\": \"/tmp/epee.log\",\n \"loglevel\": \"DEBUG\"\n }\n logger = logging.Logger(name)\n LOG_FORMAT = '%(asctime)s [%(levelname)s] [%(filename)s:%(lineno)d]: %(message)s'\n formatter = logging.Formatter(LOG_FORMAT)\n # set console log\n console = logging.StreamHandler(sys.stdout)\n console.setFormatter(formatter)\n logger.addHandler(console)\n # try to configure logger from settings or raise Error\n try:\n filename = settings[\"logfile\"]\n logger.setLevel(settings[\"loglevel\"])\n filehandler = logging.FileHandler(filename)\n filehandler.setFormatter(formatter)\n logger.addHandler(filehandler)\n except Exception as e:\n logger.error(e)\n logger.debug(traceback.format_exc())\n raise SystemExit\n return logger\n\n\ndef getid(dd={}):\n \"\"\"get value of key 'id' from any level of inputted dictionary\"\"\"\n if \"id\" in dd:\n return dd[\"id\"]\n for k in dd:\n if isinstance(dd[k], dict):\n result = getid(dd[k])\n if result is not None:\n return result\n return\n\n\ndef validate_task(task):\n \"\"\"validate input params\"\"\"\n keys = (\n \"id\",\n \"first\",\n \"second\",\n )\n errors = []\n for k in keys:\n if task.get(k) is None:\n errors.append(\"%s is not set\" % k)\n if errors:\n print(errors)\n raise NoParamsError(errors)\n", "sub_path": "epee/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 1587, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "logging.Logger", "line_number": 12, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 14, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 16, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 16, "usage_type": "attribute"}, {"api_name": "logging.FileHandler", "line_number": 23, "usage_type": "call"}, {"api_name": "traceback.format_exc", "line_number": 28, "usage_type": "call"}, {"api_name": "epee.errors.NoParamsError", "line_number": 58, "usage_type": "call"}]} +{"seq_id": "293367484", "text": "from datetime import datetime\n\nimport discord\n\nfrom Core import esi\nimport random\nimport requests\nfrom Miscellaneous.config import DOOSTER_PHRASES, JUMP_MASS_CATEGORIES, DESTINATION_CATEGORIES, COLOR\nfrom Miscellaneous.wormhole_data import WORMHOLE_IDS\nfrom Core.esi import get_type\n\n\ndef get_xkcd_url(arg):\n \"\"\"\n Get the XKCD comic URL specified by arg\n :param arg: an integer representing the comic to be retreived\n :return: string containing the xkcd url\n \"\"\"\n response = requests.get('https://xkcd.com/info.0.json')\n xkcd_json = response.json()\n max_url = xkcd_json['num']\n if arg.isdigit() and 0 < int(arg) <= max_url:\n return 'https://xkcd.com/{comic_num}'.format(comic_num=arg)\n elif arg == 'random':\n return 'https://xkcd.com/{comic_num}'.format(comic_num=random.randint(1, max_url))\n else:\n return 'Invalid webcomic. Try again with an integer between 1 and ' + str(max_url)\n\n\ndef get_dustey_phrase():\n return random.choice(DOOSTER_PHRASES)\n\n\ndef _get_dogma_value(wh_data, key):\n \"\"\"\n Find a dictionary identified by a key, in a list of dictionaries\n :param wh_data: the dictionary to search\n :param key: the key to identify the correct dictionary\n :return: the value in the correct dictionary\n \"\"\"\n dogma_attr = wh_data.get('dogma_attributes')\n return next(element for element in dogma_attr if element.get('attribute_id') == key).get('value')\n\n\ndef get_wormhole_stats(id):\n \"\"\"\n Get attributes of a wormhole\n :param id: 4 character wormhole id\n :return: dict containing the relevant wormhole attributes\n \"\"\"\n for wormhole in WORMHOLE_IDS:\n if id in wormhole.get('name'):\n wh_data = get_type(wormhole.get('id'))\n\n leads_to = _get_dogma_value(wh_data, 1381)\n lifetime = _get_dogma_value(wh_data, 1382)\n total_mass = _get_dogma_value(wh_data, 1383)\n regen_mass = _get_dogma_value(wh_data, 1384)\n jump_mass = _get_dogma_value(wh_data, 1385)\n\n return {'leads_to': DESTINATION_CATEGORIES.get(leads_to),\n 'lifetime': int(lifetime / 60),\n 'total_mass': int(total_mass),\n 'regen_mass': int(regen_mass),\n 'jump_mass': JUMP_MASS_CATEGORIES.get(jump_mass)}\n\n\ndef _running_for(start_time):\n running_for = int((datetime.utcnow() - start_time).total_seconds())\n if running_for < 60:\n return \"less than a minute\"\n\n running_for_list = []\n units = [\n (running_for // (60 * 60), \"hour\"),\n ((running_for // 60) % 60, \"minute\"),\n ]\n for number, unit in units:\n running_for_list.append(\"{} {}{}\".format(\n number,\n unit,\n \"s\" * (number != 1)\n ))\n return \", \".join(running_for_list)\n\n\ndef get_server_status(datasource='tranquility'):\n \"\"\"Generate a discord embed describing the status of an EVE server/datasource.\"\"\"\n\n response = esi.get_status(datasource)\n server_name = datasource.capitalize()\n\n if response == \"offline\" or response == \"indeterminate\":\n\n attachment = discord.Embed(\n title=\"{}: {}\".format(server_name, response.capitalize()),\n color=COLOR.RED,\n ).add_field(\n name=\"Server time\",\n value=datetime.strftime(datetime.utcnow(), \"%Y-%m-%d %H:%M:%S\")\n )\n\n else:\n vip = response.get(\"vip\")\n started = datetime.strptime(response[\"start_time\"], \"%Y-%m-%dT%H:%M:%SZ\")\n\n attachment = discord.Embed(\n title=\"{}: Online\".format(server_name),\n color=COLOR.ORANGE if vip else COLOR.GREEN,\n ).add_field(\n name=\"Server time\",\n value=datetime.strftime(datetime.utcnow(), \"%Y-%m-%d %H:%M:%S\"),\n inline=False\n ).add_field(\n name=\"Players online\",\n value=\"{:,}\".format(response[\"players\"]),\n inline=False\n ).add_field(\n name=\"Started at\",\n value=datetime.strftime(started, \"%Y-%m-%d %H:%M:%S\"),\n inline=False\n ).add_field(\n name=\"Running for\",\n value=_running_for(started),\n inline=False\n )\n\n if vip:\n attachment.title = \"{}: In VIP mode\".format(server_name)\n\n return attachment", "sub_path": "Miscellaneous/controller.py", "file_name": "controller.py", "file_ext": "py", "file_size_in_byte": 4319, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "requests.get", "line_number": 19, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 25, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 31, "usage_type": "call"}, {"api_name": "Miscellaneous.config.DOOSTER_PHRASES", "line_number": 31, "usage_type": "argument"}, {"api_name": "Miscellaneous.wormhole_data.WORMHOLE_IDS", "line_number": 51, "usage_type": "name"}, {"api_name": "Core.esi.get_type", "line_number": 53, "usage_type": "call"}, {"api_name": "Miscellaneous.config.DESTINATION_CATEGORIES.get", "line_number": 61, "usage_type": "call"}, {"api_name": "Miscellaneous.config.DESTINATION_CATEGORIES", "line_number": 61, "usage_type": "name"}, {"api_name": "Miscellaneous.config.JUMP_MASS_CATEGORIES.get", "line_number": 65, "usage_type": "call"}, {"api_name": "Miscellaneous.config.JUMP_MASS_CATEGORIES", "line_number": 65, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 69, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 69, "usage_type": "name"}, {"api_name": "Core.esi.get_status", "line_number": 90, "usage_type": "call"}, {"api_name": "Core.esi", "line_number": 90, "usage_type": "name"}, {"api_name": "discord.Embed", "line_number": 95, "usage_type": "call"}, {"api_name": "Miscellaneous.config.COLOR.RED", "line_number": 97, "usage_type": "attribute"}, {"api_name": "Miscellaneous.config.COLOR", "line_number": 97, "usage_type": "name"}, {"api_name": "datetime.datetime.strftime", "line_number": 100, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 100, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 100, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 105, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 105, "usage_type": "name"}, {"api_name": "discord.Embed", "line_number": 107, "usage_type": "call"}, {"api_name": "Miscellaneous.config.COLOR.ORANGE", "line_number": 109, "usage_type": "attribute"}, {"api_name": "Miscellaneous.config.COLOR", "line_number": 109, "usage_type": "name"}, {"api_name": "Miscellaneous.config.COLOR.GREEN", "line_number": 109, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strftime", "line_number": 112, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 112, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 112, "usage_type": "call"}, {"api_name": "datetime.datetime.strftime", "line_number": 120, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 120, "usage_type": "name"}]} +{"seq_id": "462181502", "text": "from django.contrib.auth.models import User\nfrom django.test.client import Client\nfrom django.core.urlresolvers import reverse\n\nfrom django.test import TestCase\n\nfrom entities.models import Domain, Division, Entity\nfrom .models import *\n\nclass UserTest(TestCase):\n user = {\n 'first_name': 'Django',\n 'last_name': 'Reinhardt',\n 'verified': True,\n 'name': 'Django Reinhardt',\n 'locale': 'en_US',\n 'hometown': {\n 'id': '12345678',\n 'name': 'Any Town, Any State'\n },\n 'expires': '4812',\n 'updated_time': '2012-01-29T19:27:32+0000',\n 'access_token': 'dummyToken',\n 'link': 'http://www.facebook.com/profile.php?id=1234',\n 'location': {\n 'id': '108659242498155',\n 'name': 'Chicago, Illinois'\n },\n 'gender': 'male',\n 'timezone': -6,\n 'id': '1234',\n 'email': 'user@domain.com'\n }\n def setUp(self):\n self.user = User.objects.create_user(\"user\",\n \"user@example.com\", \"pass\")\n self.user.profile.save()\n self.candidate = User.objects.create_user(\"candidate\",\n \"candidate@example.com\", \"pass\")\n self.candidate.profile.save()\n\n def test_edit_profile(self):\n c = Client()\n clist_url = reverse('edit_profile')\n response = c.get(clist_url)\n self.assertEquals(response.status_code, 302)\n self.assertTrue(c.login(username=\"user\", password=\"pass\"))\n response = c.get(clist_url)\n self.assertEquals(response.status_code, 200)\n\n def test_public_profile(self):\n c = Client()\n # assert {% url %} and get_absolute_url are one and the same\n public_profile = reverse('public-profile',\n kwargs={'username': self.user.username})\n self.assertEquals(public_profile,\n Profile.objects.get(user=self.user).get_absolute_url())\n response = c.get(public_profile)\n self.assertEquals(response.status_code, 200)\n self.assertTemplateUsed(response, \"user/public_profile.html\")\n\n def test_entities(self):\n domain = Domain.objects.create(name=\"test\")\n division = Division.objects.create(name=\"localities\", domain=domain,\n index=3)\n locality = Entity.objects.create(name=\"the moon\", division=division)\n self.user.profile.add_entity(locality)\n self.assertTrue(self.user.profile.is_member_of(locality))\n self.assertTrue(locality.id in self.user.profile.get_entity_ids())\n", "sub_path": "user/tests.py", "file_name": "tests.py", "file_ext": "py", "file_size_in_byte": 2578, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "django.test.TestCase", "line_number": 10, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.create_user", "line_number": 35, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 35, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 35, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.create_user", "line_number": 38, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 38, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 38, "usage_type": "name"}, {"api_name": "django.test.client.Client", "line_number": 43, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 44, "usage_type": "call"}, {"api_name": "django.test.client.Client", "line_number": 52, "usage_type": "call"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 54, "usage_type": "call"}, {"api_name": "entities.models.Domain.objects.create", "line_number": 63, "usage_type": "call"}, {"api_name": "entities.models.Domain.objects", "line_number": 63, "usage_type": "attribute"}, {"api_name": "entities.models.Domain", "line_number": 63, "usage_type": "name"}, {"api_name": "entities.models.Division.objects.create", "line_number": 64, "usage_type": "call"}, {"api_name": "entities.models.Division.objects", "line_number": 64, "usage_type": "attribute"}, {"api_name": "entities.models.Division", "line_number": 64, "usage_type": "name"}, {"api_name": "entities.models.Entity.objects.create", "line_number": 66, "usage_type": "call"}, {"api_name": "entities.models.Entity.objects", "line_number": 66, "usage_type": "attribute"}, {"api_name": "entities.models.Entity", "line_number": 66, "usage_type": "name"}]} +{"seq_id": "476396285", "text": "# Notes:-\n# This code uses CART instead of c4.5. They both are pretty close to each other. Refer documentation for difference.\n# This code uses SAMME.R for Adaboost instea of ADaboost.MI used for boosting in the paper.\n#\n# Links to documentations\n# http://scikit-learn.org/stable/modules/generated/sklearn.ensemble.BaggingClassifier.html#sklearn.ensemble.BaggingClassifier\n# http://scikit-learn.org/stable/modules/generated/sklearn.tree.DecisionTreeClassifier.html\n# http://scikit-learn.org/stable/modules/generated/sklearn.ensemble.ExtraTreesClassifier.html#sklearn.ensemble.ExtraTreesClassifier\n# http://scikit-learn.org/stable/modules/classes.html#module-sklearn.ensemble\n# http://scikit-learn.org/stable/modules/generated/sklearn.ensemble.AdaBoostClassifier.html#sklearn.ensemble.AdaBoostClassifier.fit\n\n# TODO\n# Re-verify all parameters to ensembles methods and make sure we are not diverging from the paper.\n# Do fancy charts with the accuracy data.\n# Find out if there is a way in extratreeclassifier to make it use only positive information gain splits.\n# Are we going to use pruning?\n# calculate Error rate using t-test\n# Support other data source format besides csv\n\n\n\n# Import stuff\nimport numpy as np\nimport pandas as pd\nimport scipy.stats as stats\nimport matplotlib.pyplot as plt\nimport statsmodels.api as sm\nimport sklearn\nfrom sklearn.ensemble import BaggingClassifier\nfrom sklearn.tree import DecisionTreeClassifier\nfrom sklearn.ensemble import AdaBoostClassifier\nfrom sklearn.ensemble import ExtraTreesClassifier\nfrom sklearn.cross_validation import cross_val_score\nfrom sklearn import cross_validation\nfrom sklearn.preprocessing import Imputer\n\nCSV_SEP = ','\n\n# get list of datasets of UCI rep to use\ndef get_dataset_list():\n #datasets = ['https://archive.ics.uci.edu/ml/machine-learning-databases/wine-quality/winequality-red.csv',\n # 'https://archive.ics.uci.edu/ml/machine-learning-databases/wine-quality/winequality-white.csv']\n datasets = ['data/anneal.data.txt','data/glass.data.txt','data/heart.dat.txt','data/house-votes-84.data.txt']\n #,'data/krkopt.data.txt',\n #'data/letter-recognition.data.txt','data/sat.trn.txt','data/segment.dat.txt','data/sonar.all-data.txt',\n #'data/soybean-large.data.txt','splice.data.txt']\n\n return datasets\n\n# Do bagging.\n# X : {array-like, sparse matrix} of shape = [n_samples, n_features]\n# Y : array-like, shape = [n_samples]\ndef do_bagging(X, y):\n bagging = BaggingClassifier(DecisionTreeClassifier(), 200, 0.67, 1.0, True, True)\n return cross_val_score(bagging, X, y, cv=10)\n\n# Do boosting\n# X : {array-like, sparse matrix} of shape = [n_samples, n_features]\n# Y : array-like, shape = [n_samples]\ndef do_boosting(X, y):\n boosting = AdaBoostClassifier(DecisionTreeClassifier(), n_estimators=100)\n return cross_val_score(boosting, X, y, cv=10)\n\n# Do Randomization\n# X : {array-like, sparse matrix} of shape = [n_samples, n_features]\n# Y : array-like, shape = [n_samples]\ndef do_randomization(X, y):\n random = ExtraTreesClassifier(200)\n return cross_val_score(random, X, y, cv=10)\n\n# Do plain vanilla CART\n# X : {array-like, sparse matrix} of shape = [n_samples, n_features]\n# Y : array-like, shape = [n_samples]\ndef do_cart(X, y):\n cart = DecisionTreeClassifier()\n return cross_val_score(cart, X, y, cv=10)\n\ndef convert_to_error_rate(score):\n error = []\n for foldscore in score:\n errorscore = 1 - foldscore\n error.append(errorscore)\n return error\n \n\n# Main\nscores = []\nerror_rates = []\nmean_error_rates = []\nfor dataset_url in get_dataset_list():\n print(\"Testing: \"+dataset_url)\n datascore = []\n dataerrorrate = []\n df = pd.read_csv(dataset_url, CSV_SEP)\n Y = df['class'].values\n df = df.drop('class', 1)\n df.fillna(df.mode().iloc[0])\n X = df.as_matrix()\n print(X)\n #imp = Imputer(missing_values='?', strategy='mode',axis=0)\n #X = imp.fit_transform(df)\n Y = np.array([1 if y >= 7 else 0 for y in Y])\n \n cart_score = do_cart(X, Y)\n bagging_score = do_bagging(X, Y)\n boosting_score = do_boosting(X, Y)\n random_score = do_randomization(X, Y)\n \n datascore.append(cart_score)\n datascore.append(bagging_score)\n datascore.append(boosting_score)\n datascore.append(random_score)\n scores.append(datascore)\n \n cart_error = convert_to_error_rate(cart_score)\n bagging_error = convert_to_error_rate(bagging_score)\n boosting_error = convert_to_error_rate(boosting_score)\n random_error = convert_to_error_rate(random_score)\n \n dataerrorrate.append(cart_error)\n dataerrorrate.append(bagging_error)\n dataerrorrate.append(boosting_error)\n dataerrorrate.append(random_error)\n error_rates.append(dataerrorrate)\n \n mean_error_rates.append([np.mean(cart_error), np.mean(bagging_error), np.mean(boosting_error), np.mean(random_error)])\n\n\n# Plot box plot.\nplt.boxplot(scores)\nplt.ylabel('Score')\nplt.xlabel('Classification data set')\nplt.title('Box plot of classification score for various datasets')\nplt.show()\n\n# Plot box plot.\nplt.boxplot(error_rates)\nplt.ylabel('Error rates')\nplt.xlabel('Classification data set')\nplt.title('Box plot of error rates for various datasets')\nplt.show()\n\n# Plot error rates of various algorithms\nmean_error_rates = np.array(mean_error_rates)\nplt.plot(mean_error_rates.T[0], label='CART')\nplt.plot(mean_error_rates.T[1], label='Bagging')\nplt.plot(mean_error_rates.T[2], label='Boosting')\nplt.plot(mean_error_rates.T[3], label='Randomization')\nplt.legend(loc='best')\nplt.ylabel('Error rates')\nplt.xlabel('Classification data set')\nplt.title('Line plot of error rates for various datasets using various algorithms')\nplt.show()\n\nprint(scores)", "sub_path": "run.py", "file_name": "run.py", "file_ext": "py", "file_size_in_byte": 5696, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "sklearn.ensemble.BaggingClassifier", "line_number": 54, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 54, "usage_type": "call"}, {"api_name": "sklearn.cross_validation.cross_val_score", "line_number": 55, "usage_type": "call"}, {"api_name": "sklearn.ensemble.AdaBoostClassifier", "line_number": 61, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 61, "usage_type": "call"}, {"api_name": "sklearn.cross_validation.cross_val_score", "line_number": 62, "usage_type": "call"}, {"api_name": "sklearn.ensemble.ExtraTreesClassifier", "line_number": 68, "usage_type": "call"}, {"api_name": "sklearn.cross_validation.cross_val_score", "line_number": 69, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 75, "usage_type": "call"}, {"api_name": "sklearn.cross_validation.cross_val_score", "line_number": 76, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.boxplot", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.boxplot", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 137, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 139, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 139, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 141, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 141, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 145, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 146, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 147, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 148, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 148, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 149, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 151, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 152, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 153, "usage_type": "name"}]} +{"seq_id": "476125793", "text": "#Created and updated on 2018.10.22\n#call ms tts api and get multipil audio files\n#multiple languages and voices, features kept in pandas dataframe\n#this version is for English (US, UK, India), Japanese and Chinese (mainland China, HK and Taiwan)\n\nimport http.client, urllib.parse, json\nfrom xml.etree import ElementTree\n\n#global variables\napiKey = \"e32b3ea6ff3a4e27baa64bec7f54de57\"\nvoiceDIR = '/Users/maryzhu/ggdrive/language/tts/msazure/'\n\ndef getaccesstoken(apiKey):\n \n params = \"\"\n headers = {\"Ocp-Apim-Subscription-Key\": apiKey}\n\n #AccessTokenUri = \"https://westus.api.cognitive.microsoft.com/sts/v1.0/issueToken\";\n AccessTokenHost = \"westus.api.cognitive.microsoft.com\"\n path = \"/sts/v1.0/issueToken\"\n \n # Connect to server to get the Access Token\n print (\"Connect to server to get the Access Token\")\n conn = http.client.HTTPSConnection(AccessTokenHost)\n conn.request(\"POST\", path, params, headers)\n response = conn.getresponse()\n print(response.status, response.reason)\n \n data = response.read()\n conn.close()\n \n accesstoken = data.decode(\"UTF-8\")\n print (\"Access Token: \" + accesstoken)\n return accesstoken\n\n#global variables\nmyaccesstoken = getaccesstoken(apiKey)\n\ndef getdata2file(lang, gender, voicetype, text, filename):\n \n print('===parameters====')\n print(lang, gender, voicetype, text, filename)\n \n body = ElementTree.Element('speak', version='1.0')\n body.set('{http://www.w3.org/XML/1998/namespace}lang', lang)\n voice = ElementTree.SubElement(body, 'voice')\n voice.set('{http://www.w3.org/XML/1998/namespace}lang', lang)\n voice.set('{http://www.w3.org/XML/1998/namespace}gender', gender)\n voice.set('name', 'Microsoft Server Speech Text to Speech Voice (' + lang + ', ' + voicetype + ')')\n voice.text = text\n\n headers = {\"Content-type\": \"application/ssml+xml\", \n #\"X-Microsoft-OutputFormat\": \"riff-24khz-16bit-mono-pcm\",\n \"X-Microsoft-OutputFormat\": \"audio-16khz-128kbitrate-mono-mp3\",\n \"Authorization\": \"Bearer \" + myaccesstoken, \n \"X-Search-AppId\": \"07D3234E49CE426DAA29772419F436CA\", \n \"X-Search-ClientID\": \"1ECFAE91408841A480F00935DC390960\", \n \"User-Agent\": \"TTSForPython\"}\n \n print('===body===')\n print(ElementTree.tostring(body))\n \n print('===headers===')\n print(headers)\n \n \n #Connect to server to synthesize the wave\n print (\"\\nConnect to server to synthesize the wave\")\n conn = http.client.HTTPSConnection(\"westus.tts.speech.microsoft.com\")\n conn.request(\"POST\", \"/cognitiveservices/v1\", ElementTree.tostring(body), headers)\n response = conn.getresponse()\n print(response.status, response.reason)\n \n data = response.read()\n conn.close()\n print(\"The synthesized wave length: %d\" %(len(data)))\n \n #write audio data to file\n with open(filename, 'wb') as f:\n f.write(data)\n f.close()\n\ndef getallvoices(frame, lang, text):\n i = 0\n while i < len(frame):\n gender = frame.iloc[i].gender\n voicetype = frame.iloc[i].voicetype\n print(gender, voicetype)\n i += 1\n print(i)\n filename = lang + '_' + gender + '_' + voicetype + '.mp3'\n getdata2file(lang, gender, voicetype, text, voiceDIR + filename) \n \n#main program starts here\nimport pandas as pd\nfrom pandas import DataFrame\n\ntexten = 'No. 1: this technology is wonderful. we would like to use it in different areas. No.2: We will go out together this weekend.'\ntextjp = 'この技術は世界最先端のものです。研究者たちが長い年月をかけて開発しました。いちばん:いらっしゃいませ。こちらへどうぞ。にばん:はい。さんばん:決まりましたら、お呼びください。よんばん:はい、わかりました。'\ntextcn = '亲爱的佳佳:祝贺你自然考试取得好成绩!我们大家都为你高兴。今后还要加油哦!今天天气很好。我想和妈妈一起出去吃饭。我最喜欢吃日本菜,其中又最喜欢吃生鱼片。我爱爸爸妈妈,也爱我的外婆和奶奶。希望他们每天都开心。'\n\nlangus = 'en-US'\nlanguk = 'en-GB'\nlangin = 'en-IN'\nlangjp = 'ja-JP'\nlangcn = 'zh-CN'\nlanghk = 'zh-HK'\nlangtw = 'zh-TW'\n\n#US English voices\ndataus = {'gender': ['female', 'female', 'male', 'female', 'male'],\n 'voicetype': ['ZiraRUS', 'JessaRUS', 'BenjaminRUS', 'Jessa24kRUS', 'Guy24kRUS']}\n\nframeus = pd.DataFrame(dataus)\n\ngetallvoices(frameus, langus, texten)\n\n#UK English voices\ndatauk = {'gender': ['female', 'female', 'male'],\n 'voicetype': ['Susan, Apollo', 'HazelRUS', 'George, Apollo']}\n\nframeuk = pd.DataFrame(datauk)\n\ngetallvoices(frameuk, languk, texten)\n\n#India English voices\ndatain = {'gender': ['female', 'female', 'male'],\n 'voicetype': ['Heera, Apollo', 'PriyaRUS', 'Ravi, Apollo']}\n\nframein = pd.DataFrame(datain)\n\ngetallvoices(framein, langin, texten)\n\n#Japanese voices\ndatajp = {'gender': ['female', 'male', 'female'],\n 'voicetype': ['Ayumi, Apollo', 'Ichiro, Apollo', 'HarukaRUS']}\n\nframejp = pd.DataFrame(datajp)\n\ngetallvoices(framejp, langjp, textjp)\n\n#CN Chinese voices\ndatacn = {'gender': ['female', 'female', 'male'],\n 'voicetype': ['HuihuiRUS', 'Yaoyao, Apollo', 'Kangkang, Apollo']}\n\nframecn = pd.DataFrame(datacn)\n\ngetallvoices(framecn, langcn, textcn)\n\n#HK Chinese voices\ndatahk = {'gender': ['female', 'female', 'male'],\n 'voicetype': ['Tracy, Apollo', 'TracyRUS', 'Danny, Apollo']}\n\nframehk = pd.DataFrame(datahk)\n\ngetallvoices(framehk, langhk, textcn)\n\n#TW Chinese voices\ndatatw = {'gender': ['female', 'female', 'male'],\n 'voicetype': ['Yating, Apollo', 'HanHanRUS', 'Zhiwei, Apollo']}\n\nframetw = pd.DataFrame(datatw)\n\ngetallvoices(frametw, langtw, textcn)\n\n", "sub_path": "msttswebapi.py", "file_name": "msttswebapi.py", "file_ext": "py", "file_size_in_byte": 5827, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "http.client.client.HTTPSConnection", "line_number": 24, "usage_type": "call"}, {"api_name": "http.client.client", "line_number": 24, "usage_type": "attribute"}, {"api_name": "http.client", "line_number": 24, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 44, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 44, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 46, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 46, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.tostring", "line_number": 61, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 61, "usage_type": "name"}, {"api_name": "http.client.client.HTTPSConnection", "line_number": 69, "usage_type": "call"}, {"api_name": "http.client.client", "line_number": 69, "usage_type": "attribute"}, {"api_name": "http.client", "line_number": 69, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.tostring", "line_number": 70, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 70, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 114, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 122, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 130, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 138, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 146, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 154, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 162, "usage_type": "call"}]} +{"seq_id": "226549445", "text": "# allgemeine imports\nimport csv\nimport itertools\nfrom datetime import datetime\nfrom math import pi\n\nimport pandas as pd\n\n# bokeh imports\nfrom bokeh.embed import components\nfrom bokeh.layouts import column\nfrom bokeh.models import Panel, Tabs\nfrom bokeh.palettes import Dark2_5 as palette\nfrom bokeh.plotting import figure\n\n# django imports\nfrom django.conf import settings\nfrom django.http import HttpResponse\nfrom django.shortcuts import redirect, render, reverse\nfrom django.views import View\nfrom django.views.generic import CreateView, TemplateView\n\nfrom trading.mixins import (\n fehlerViewMixin,\n plotKonfigViewMixin,\n\n allgemeineFehlerPruefung,\n datenAnBackendSenden,\n)\n\n# Strategie app imports\nfrom strategie.models import Strategie\n\n# Simulation app imports\nfrom .forms import SimulationModelForm\nfrom .models import Simulation\n\nglobalerHauptPfad = \"simulation\"\nappName = \"simulation\"\n\nclass SimulationFehlerView(fehlerViewMixin):\n \"\"\"\n Klasse für das Anzeigen von jeglichen Fehlermeldungen.\n \"\"\" \n\n appName = appName\n\nclass SimulationConfigView(plotKonfigViewMixin): \n \"\"\"\n Klasse der Ansicht für das Konfigurieren einer Simulation.\n \"\"\" \n\n form_class = SimulationModelForm \n appName = appName\n model = Simulation\n \nclass SimulationErgebnisView(View):\n \"\"\"\n Klasse der Ansicht für das Darstellen des Simulations-Ergebnis.\n\n Methoden\n --------\n get\n Zuständig für das Holen und Darstellen der Ergebnis-Daten.\n \"\"\"\n\n template_name = 'modulViews/generisch_ansicht_plotErgebnis.html'\n\n def get(self, request):\n \"\"\"\n Diese Funktion holt sich alle Daten die benötigt werden für das Darstellen der Simulations-Ergebnisse.\n Dazu gehören die Aktienkurs Daten der zum simulieren verwendeten ISIN und die Simulationsergebnisse.\n\n \"\"\"\n if(\"konfigDaten\" not in request.session):\n request.session[\"fehler\"] = \"Keine Simulationsdaten vorhanden. Versuchen Sie es nochmal...\" \n return redirect(reverse('simulation:simulation-fehler')) \n\n\n # Zuletzt erstelltes Objekt holen\n # Prüfen ob es existiert, wenn nicht, auf FehlerSeite weiterleiten und entsprechende Fehlermeldung in session speichern.\n \n konfigDaten = request.session[\"konfigDaten\"]\n\n strategie_id = konfigDaten[\"strategie_id\"]\n isin = konfigDaten[\"isin\"]\n start_kapital = konfigDaten[\"start_kapital\"]\n start_datum = konfigDaten[\"start_datum\"]\n end_datum = konfigDaten[\"end_datum\"]\n # Hier werden die Simulation-Konfigurations-Daten an das Backend geschickt.\n simulationsDaten = datenAnBackendSenden(\n hauptPfad = globalerHauptPfad, \n unterPfad = \"\", \n daten = {\n \"benutzer_id\" : self.request.user.username,\n \"strategie_id\" : strategie_id,\n \"isin\" : isin,\n \"start_kapital\" : start_kapital,\n \"start_datum\" : start_datum,\n \"end_datum\" : end_datum,\n }\n )\n # Bei Fehlern auf Fehlerseite weiterleiten\n if(allgemeineFehlerPruefung(simulationsDaten,request)):\n return redirect(reverse('simulation:simulation-fehler')) \n #------------------\n\n self.request.session[\"simulationsDaten\"] = simulationsDaten # Simulationsdaten in Session speichern um für die Funktion downloadCSV zugänglich zu machen\n \n #Von- und Bis-Datum in passendes Format TT.MM.JJJJ umwandeln\n vonDatumAnzeigeFormat = datetime.strptime(start_datum, '%Y-%m-%d').strftime('%d.%m.%Y') \n bisDatumAnzeigeFormat = datetime.strptime(end_datum, '%Y-%m-%d').strftime('%d.%m.%Y')\n\n # AnzeigeDaten Dict für Template zusammenbauen\n anzeigeDaten = {\n \"strategie_id\" : strategie_id,\n \"strategie_name\" : simulationsDaten[\"strategie\"][\"name\"],\n \"isin\" : simulationsDaten[\"wertpapier\"][\"isin\"],\n \"name\" : simulationsDaten[\"wertpapier\"][\"name\"],\n \"start_kapital\" : start_kapital,\n \"start_datum\" : start_datum,\n \"end_datum\" : end_datum,\n \"statistik\" : simulationsDaten[\"strategie_statistik\"],\n }\n\n # Statistischen Werte auf 3 Nachkommestellen kürzen\n anzeigeDaten[\"statistik\"][\"performance_gesamt\"] = '%.3f' % (float(anzeigeDaten[\"statistik\"][\"performance_gesamt\"]))\n anzeigeDaten[\"statistik\"][\"performance_pro_jahr\"] = '%.3f' % (float(anzeigeDaten[\"statistik\"][\"performance_pro_jahr\"]))\n anzeigeDaten[\"statistik\"][\"hoch_gesamt\"] = '%.3f' % (float(anzeigeDaten[\"statistik\"][\"hoch_gesamt\"]))\n anzeigeDaten[\"statistik\"][\"tief_gesamt\"] = '%.3f' % (float(anzeigeDaten[\"statistik\"][\"tief_gesamt\"]))\n anzeigeDaten[\"statistik\"][\"maximum_drawdown\"] = '%.3f' % (float(anzeigeDaten[\"statistik\"][\"maximum_drawdown\"]))\n\n # Alle Koordinatensysteme erstellen\n \n \n \n p1 = figure( # p1 ist Koordinatensystem mit Aktienkurs als Candlesticks\n plot_width=1000, \n plot_height=700, \n x_axis_type=\"datetime\", \n tools=\"pan, wheel_zoom, box_zoom, reset, save\" # verfügbaren Tools\n )\n p2 = figure( # p2 ist Koordinatensystem mit Aktienkurs als Linie\n plot_width=1000, \n plot_height=700, \n x_axis_type=\"datetime\", \n tools=\"pan, wheel_zoom, box_zoom, reset, save\", \n x_range=p1.x_range, # Verschiebungen entlang der X-Achse von p1 übernehmen\n y_range=p1.y_range # Verschiebungen entlang der Y-Achse von p1 übernehmen\n )\n p3 = figure( # p3 ist Koordinatensystem für Performance-Entwicklung\n plot_width=1000, \n plot_height=700, \n x_axis_type=\"datetime\", \n tools=\"pan, wheel_zoom, box_zoom, reset, save\",\n x_range=p1.x_range # Verschiebungen entlang der X-Achse von p1 übernehmen\n # Verschiebung entlang der Y-Achse wird nicht übernommen, da es keinen Sinn ergibt.\n )\n plotListe = [p1,p2,p3] # Alle Koordinatensyteme in eine Liste\n\n indikatoren = simulationsDaten[\"indikator_zeitreihe\"] # Alle indikatordaten werden in einem eigenen Objekt gespeichert\n keyList = []\n listenDict = {}\n\n # Jeder einzelne Key wird in die Key-Liste gespeichert und für jeden Key wird im ListenDict eine leere Liste erstellt.\n # Der erste key ist immer der Zeitstempel, die darauf folgenden sind die keys der Indikatoren\n for key in indikatoren[0]:\n keyList.append(key)\n listenDict[key] = []\n\n # daten im indikatoren-Objekt werden aufgeteilt und abhängig vom key an die entsprechende Liste im listenDict angefügt.\n for daten in indikatoren:\n for key in keyList:\n listenDict[key].append(daten[key])\n\n # Farb-Palette für das Darstellen der Graphen in verschiedenen Farben\n colors = itertools.cycle(palette)\n\n # Alle benötigten DataFrames generieren und deren Zeitstempel entsprechend formatieren\n zeitstempelDataFrame = pd.DataFrame(listenDict[\"zeitstempel\"], columns = ['zeitstempel'])\n zeitstempelDataFrame[\"zeitstempel\"] = pd.to_datetime(zeitstempelDataFrame[\"zeitstempel\"])\n\n kursZeitreiheDF = pd.DataFrame(simulationsDaten[\"kurs_zeitreihe\"])\n kursZeitreiheDF[\"zeitstempel\"] = pd.to_datetime(kursZeitreiheDF[\"zeitstempel\"])\n\n stratZeitreiheDF = pd.DataFrame(simulationsDaten[\"strategie_kurs_zeitreihe\"])\n stratZeitreiheDF[\"zeitstempel\"] = pd.to_datetime(stratZeitreiheDF[\"zeitstempel\"])\n\n kaufVerkaufDF = pd.DataFrame(simulationsDaten[\"strategie_kaeufe_verkaeufe_zeitreihe\"])\n\n\n inc = kursZeitreiheDF.close > kursZeitreiheDF.open # inc ist ein boolean, welcher True zurückgibt wenn Close > Open\n dec = kursZeitreiheDF.open > kursZeitreiheDF.close # dec ist ein boolean, welcher True zurückgibt wenn Open > Close\n w = 12*60*60*1000 # halber Tag in ms, für die Breite der Kerzen\n\n # Hier werden die einzelnen Elemente der Kerze zum Koordinatensystem p1 hinzugefügt\n \n p1.segment( # Hier werden die High und Lows für jeden Tag erstellt, dargestellt durch zwei verbunde Punkte \n kursZeitreiheDF.zeitstempel, # X Werte für Punkt 1\n kursZeitreiheDF.high, # Y Werte für Punkt 1\n kursZeitreiheDF.zeitstempel, # X Werte für Punkt 2 \n kursZeitreiheDF.low, # Y Werte für Punkt 2\n color = \"black\",\n legend_label = simulationsDaten[\"wertpapier\"][\"name\"] + \" - High-Low\"\n )\n \n p1.vbar( # Hier werden die grünen Kerzen \"Torsos\" erstellt, dargestellt durch Balken je einem X-Wert, zwei Y-Werten und einer Breite\n kursZeitreiheDF.zeitstempel[inc], # X Werte der Balken \n w, # Breite \n kursZeitreiheDF.open[inc], # Y1 Werte der Balken\n kursZeitreiheDF.close[inc], # Y2 Werte der Balken\n fill_color = \"green\", line_color = \"black\",\n legend_label = simulationsDaten[\"wertpapier\"][\"name\"] + \" - Grüne Kerzen\"\n )\n \n p1.vbar(# Hier werden die roten Kerzen \"Torsos\" erstellt, dargestellt durch Balken je einem X-Wert, zwei Y-Werten und einer Breite\n kursZeitreiheDF.zeitstempel[dec], # X Werte der Balken \n w, # Breite \n kursZeitreiheDF.open[dec], # Y1 Werte der Balken\n kursZeitreiheDF.close[dec], # Y2 Werte der Balken\n fill_color = \"red\", line_color = \"black\",\n legend_label = simulationsDaten[\"wertpapier\"][\"name\"] + \" - Rote Kerzen\"\n )\n # ---------------------------------\n \n p2.line( # Hier wird der Aktienkurs als Linie zu dem Koordinatensystem p2 hinzugefügt\n kursZeitreiheDF.zeitstempel, # X Werte der Linie\n kursZeitreiheDF.close, # Y Werte der Linie\n line_width = 3, color = \"red\", alpha = 0.5, \n legend_label = simulationsDaten[\"wertpapier\"][\"name\"] \n ) \n p3.line( # Hier wird die Performance als Linie zu dem Koordinatensystem p3 hinzugefügt\n stratZeitreiheDF.zeitstempel, \n stratZeitreiheDF.kurs_prozentual, \n line_width = 3, color = \"black\", alpha = 0.5, \n legend_label = \"Kapital in Prozent\"\n )\n\n # Für jeden Key in der KeyListe\n for key in keyList:\n # Der nicht gleich zeitstempel, also für jeden Indikator\n if(key != \"zeitstempel\"):\n # wird ein eigenes DataFrame angelegt mit den entsprechenden Daten aus listenDict\n elementDataFrame = pd.DataFrame(listenDict[key])\n\n # Wenn Indikator eine eigene Skala braucht, wird ein neues Koordinatensystem erstellt\n inEigenesKoord = False\n if(pruefeObEigeneSkala(key,self)):\n newPlot = figure(\n plot_width = 1000, \n plot_height = 700, \n x_axis_type = \"datetime\", \n tools = \"pan, wheel_zoom, box_zoom, reset, save, crosshair\",\n x_range = p1.x_range # Verschiebungen entlang der X-Achse von p1 übernehmen\n )\n inEigenesKoord = True\n \n # für jeden key in diesem DataFrame, also für jeden Graph des Indikators\n for graph in listenDict[key][0].keys(): \n # wird eine eigene Linie erstellt mit den entsprechenden Y-Werten. Die X-Werte werden dabei vom ZeitstempelDataFrame übernommen,\n # um es einheitlich zu halten\n line = pd.concat(\n [zeitstempelDataFrame, elementDataFrame[graph]], axis = 1)\n\n # Wenn eigenes Koordianten System erstellt wurde, an dieses Koordinatensystem fügen \n if(inEigenesKoord):\n newPlot.line(\n zeitstempelDataFrame[\"zeitstempel\"], \n line[graph], \n line_width = 3, color = next(colors), alpha = 1, \n legend_label = \"ID \" + key + \" - \" + graph\n )\n \n # Ansonsten in beide Koordinatensystemen p1 und p2 \n else:\n for plot in [p1,p2]:\n plot.line(\n zeitstempelDataFrame[\"zeitstempel\"], \n line[graph],\n line_width = 3, color = next(colors), alpha = 1, \n legend_label = \"ID \" + key + \" - \" + graph,\n )\n\n # Wenn eigenes Koordinatensystem erstellt wurde, dann dieses zur KoordinatenSystem Liste hinzufügen\n if(inEigenesKoord):\n plotListe.append(newPlot)\n\n # Wenn das KaufVerkaufDataFrame nicht leer ist\n if(not kaufVerkaufDF.empty):\n # boolischen Ausdrücke generieren\n kauf = kaufVerkaufDF.typ == \"Kauf\"\n verkauf = kaufVerkaufDF.typ == \"Verkauf\"\n\n # In beide Koordinatensystemen p1 und p2 werden die Markierungen für Käufe und Verkäufe eingefügt\n hauptPlotListe = [p1,p2]\n for plot in hauptPlotListe:\n plot.circle( # Punkte für Käufe\n pd.to_datetime(kaufVerkaufDF.zeitstempel[kauf]), # X-Werte, Zeitstempel der Daten wenn TransaktionsTyp = Kauf\n # Für die Y-Werte wird der Kaufpreis der Aktie berechnet. Also die Menge an ausgegebenem Kapital durch die Anzahl der gekauften Aktien\n (-1)*(kaufVerkaufDF.kapital_bestand_aenderung[kauf] / kaufVerkaufDF.stueck_bestand_aenderung[kauf]), \n size=20, color=\"blue\", alpha=0.5, legend_label=\"Käufe\"\n )\n plot.circle( # Punkte für Verkäufe\n pd.to_datetime(kaufVerkaufDF.zeitstempel[verkauf]), # X-Werte Zeitstempel der Daten wenn TransaktionsTyp = Verkauf\n # Für die Y-Werte wird der Verkaufspreis der Aktie berechnet. Also die Menge an erhaltenem Kapital durch die Anzahl der verkauften Aktien\n (-1)*(kaufVerkaufDF.kapital_bestand_aenderung[verkauf] /kaufVerkaufDF.stueck_bestand_aenderung[verkauf]), \n size=20, color=\"yellow\", alpha=0.5, legend_label=\"Verkäufe\"\n )\n\n # Hier werden alle Legenden Einstellungen und allg. Koordinaten System Einstellungen gemacht\n for plot in plotListe:\n plot.legend.location = \"top_left\"\n plot.legend.title = 'Graphen'\n plot.legend.title_text_font_style = \"bold\"\n plot.legend.title_text_font_size = \"20px\"\n plot.legend.click_policy = \"hide\"\n\n plot.output_backend = \"svg\"\n plot.background_fill_color = \"#f5f5f5\"\n plot.grid.grid_line_color = \"white\"\n plot.axis.axis_line_color = None\n plot.xaxis.major_label_orientation = pi/4\n \n\n # script, div = components(Tabs(tabs=[tab1, tab2]))\n script, div = components(column(plotListe))\n\n return render(request, 'modulViews/generisch_ansicht_plotErgebnis.html', {'script': script, 'div': div, 'daten': anzeigeDaten, \"appName\" : \"simulation\"})\n\ndef pruefeObEigeneSkala(ID,callerSelf):\n \"\"\"\n Funktion zum Prüfen ob ein Indikator mit der angegebenen ID eine eigene Skala hat. \n Hierbei werden die Indikatordaten geholt und, wenn kein Fehler aufgetreten ist, der Wert für eigene_skala zurückgegeben\n \"\"\"\n serverAntwort = datenAnBackendSenden(\n hauptPfad = \"indikator/\", \n unterPfad = \"get\", \n daten = {\n \"id\" : int(ID),\n \"benutzer_id\" : callerSelf.request.user.username,\n }\n )\n\n if(allgemeineFehlerPruefung(serverAntwort, callerSelf.request)):\n return redirect(reverse('simulation:simulation-fehler')) \n\n return serverAntwort[\"indikator\"][\"eigene_skala\"]\n \ndef downloadCSV(request):\n \"\"\"\n Funktion für das Generieren eines CSV mit allen Simulationsergebnis-Daten.\n Dabei werden alle Daten für eine deutsche Ansicht umgewandelt, \n d.h. Punkte bei Zahlenwerten werden zu Kommas und Zeitstempel werden in ein akzeptiertes Format umgewandelt\n \"\"\"\n\n daten = request.session[\"simulationsDaten\"] # Daten aus der Session holen\n datumUhrzeit = datetime.now().strftime('--%d-%m-%Y--%H-%M-%S') # aktueller Zeitstempel für Datei Namensgebung\n response = HttpResponse(content_type='text/csv')\n response['Content-Disposition'] = 'attachment; filename=\"'+daten[\"strategie\"][\"name\"]+datumUhrzeit+'.csv\"' # Festlegung des Datei-Namens\n writer = csv.writer(response, delimiter=';')\n \n \n # Die Gewählte ISIN, die Bezeichnung des wertpapiers und der eingestellte Testzeitraum in einzelne Zeilen schreiben \n writer.writerow([\"ISIN\", daten[\"wertpapier\"][\"isin\"]])\n writer.writerow([\"Bezeichnung\", daten[\"wertpapier\"][\"name\"]])\n writer.writerow([\"Start Datum\", daten[\"start_datum\"]])\n writer.writerow([\"End Datum\", daten[\"end_datum\"]])\n\n # Hier werden alle einzelnen Strategie Attribute in einzelne Zeilen mit entsprechender Bezeichnung geschrieben.\n writer.writerow([\"Strategie Daten\"])\n for strategieAttribut in daten[\"strategie\"]: # \n writer.writerow(\n [\"\", strategieAttribut, daten[\"strategie\"][strategieAttribut]])\n\n # Hier werden alle einzelnen Strategie Statistik Daten in einzelne Zeilen mit entsprechender Bezeichnung geschrieben.\n writer.writerow([\"Strategie Statistik Daten\"])\n for strategieStatistikWert in daten[\"strategie_statistik\"]:\n writer.writerow([\"\", strategieStatistikWert,\n daten[\"strategie_statistik\"][strategieStatistikWert]])\n\n leerzeichenListe = [1, 5, 2, 4] # Wie viele Leere Spalten zwischen den einzelnen Spalten sein sollen, dabei wird angefangen mit den leeren Spalten\n spaltenNamenListe = [\"Kurs Zeitreihe\", \"Strategie Kurs Zeitreihe\",\n \"Kaeufe/Verkaeufe Zeitreihe\", \"Indikator Zeitreihe(n)\"]\n spaltenNamenZeile = []\n\n # Für jeden Spaltennamen\n for idx, spaltenName in enumerate(spaltenNamenListe):\n \n for i in range(leerzeichenListe[idx]): # wird die Anzahl der vorlaufenden Leerzeichen entsprechend der leerzeichenListe der spaltenNamenZeile angefügt\n spaltenNamenZeile.append(\"\")\n # dann der Spaltennamen wert angehängt\n spaltenNamenZeile.append(spaltenName)\n # spaltenNamenZeile wird als eine Zeile in die CSV geschrieben\n writer.writerow(spaltenNamenZeile)\n\n \n\n \n # Danach kommen der Bereich mit den Kursdaten also close, high, low, open und volumen. Zusätzlich wird hier der zeitstempel mit aufgelistet\n # Dann der Bereich mit den Strategiekursdaten also kurs_kapital und kurs_prozentual\n # Dann kommt der Bereich mit den Kauf/Verkauf daten also regel, typ, stueck_bestand_aenderung und kapital_bestand_aenderung\n kursDatenBereich = [\"zeitstempel\", \"close\", \"high\", \"low\", \"open\", \"volumen\"]\n strategiekursdatenBereich = [\"kurs_prozentual\", \"kurs_kapital\"]\n kaufVerkaufDatenBereich = [\"regel\", \"typ\", \"stueck_bestand_aenderung\", \"kapital_bestand_aenderung\"]\n\n alleBereiche = [kursDatenBereich,strategiekursdatenBereich,kaufVerkaufDatenBereich]\n\n spaltenNamenListe = []\n spaltenNamenListe.extend(kursDatenBereich) \n spaltenNamenListe.extend(strategiekursdatenBereich)\n spaltenNamenListe.extend(kaufVerkaufDatenBereich)\n\n bereichBreiten = [] # Liste mit den Breiten der einzelnen Bereiche\n for bereich in alleBereiche: # Die Breite von jedem Bereich in die bereichBreiten Liste hinzufügen\n bereichBreiten.append(len(bereich))\n\n indikatorBereichBreite = 0\n # Die Bereich mit den einzelnen Indikatoren wird dynamisch wie folgt generiert:\n # für jeden key != zeitstempel, also jeden Indikator\n for key in daten[\"indikator_zeitreihe\"][0]:\n if(key != \"zeitstempel\"):\n # wird jeder graph des Indikators mit entsprechendem Namen zur spaltenNamenListe angefügt\n for graph in daten[\"indikator_zeitreihe\"][0][key]:\n spaltenNamenListe.append(\"Indikator \" + key + \" - \" + graph)\n # und die breite des Indikatorbereichs um eins erhöht.\n indikatorBereichBreite += 1\n bereichBreiten.append(indikatorBereichBreite) # Der entgültige Wert wird zur bereichBreiten-liste hinzugefügt\n \n spaltenNamenZeile = [] \n counter = 0\n # Hier werden die einzelnen Spaltennamen zur spaltenNamenZeile Liste angefügt \n # Abhängig von der bereichBreite, werden entsprechend viele Bezeichnungen zur spaltenNamenZeile hinzugefügt. Nach jedem Bereich kommt eine leere Spalte\n for breite in bereichBreiten:\n for i in range(breite):\n spaltenNamenZeile.append(spaltenNamenListe[counter])\n counter += 1\n spaltenNamenZeile.append(\"\") # Leere Spalte zwischen Bereichen\n writer.writerow(spaltenNamenZeile)\n\n kaufVerkaufCounter = 0 # Für das Zählen von bereits aufgeschriebenen Käufen/Verkäufen\n for idx, datensatz in enumerate(daten[\"indikator_zeitreihe\"]): # für jeden Datensatz in den indikator-zeitreihe\n einzelneZeile = [] # Neue Zeile initialisieren\n einzelneZeile.append(datetime.strptime(datensatz[\"zeitstempel\"],'%Y-%m-%dT%H:%M:%S%z').strftime('%d.%m.%Y')) # Zeitstempel anfügen\n\n # für jeden einzelne Wert in der kurs_zeitreihe außer zeitstempel\n for wert in daten[\"kurs_zeitreihe\"][0]: # wert ist hierbei der key \n if(wert != \"zeitstempel\"):\n # Werte an Zeile anfügen, Punkt mit Komma ersetzen für deutsche Darstellung\n einzelneZeile.append(str(daten[\"kurs_zeitreihe\"][idx][wert]).replace(\".\",\",\")) \n\n einzelneZeile.append(\"\") #Spalte zwischen Bereichen \n\n # für jeden einzelne Wert in der strategie_kurs_zeitreihe außer zeitstempel\n for element in daten[\"strategie_kurs_zeitreihe\"][0]:\n if(element != \"zeitstempel\"):\n # Werte an Zeile anfügen, Punkt mit Komma ersetzen für deutsche Darstellung\n einzelneZeile.append(str(daten[\"strategie_kurs_zeitreihe\"][idx][element]).replace(\".\",\",\")) \n \n einzelneZeile.append(\"\") #Spalte zwischen Bereichen\n\n # Prüfen ob strategie_kaeufe_verkaeufe_zeitreihe Werte enthält\n if(len(daten[\"strategie_kaeufe_verkaeufe_zeitreihe\"]) != 0):\n # Prüfen ob der kaufVerkaufCounter kleiner als die Anzahl der kaufVerkauf-Daten in strategie_kaeufe_verkaeufe_zeitreihe\n if(kaufVerkaufCounter < len(daten[\"strategie_kaeufe_verkaeufe_zeitreihe\"])): \n # prüfen ob für den aktuellen zeitstempel ein Kauf/verkauf getätig wurde\n if(datensatz[\"zeitstempel\"] == daten[\"strategie_kaeufe_verkaeufe_zeitreihe\"][kaufVerkaufCounter][\"zeitstempel\"]):\n # wenn Kauf/Verkauf getätigt wurde dann die einzelnen Werte an die Zeile hängen\n for wert in daten[\"strategie_kaeufe_verkaeufe_zeitreihe\"][0]:\n if(wert != \"zeitstempel\"):\n einzelneZeile.append(\n str(daten[\"strategie_kaeufe_verkaeufe_zeitreihe\"][kaufVerkaufCounter][wert]).replace(\".\",\",\"))\n\n kaufVerkaufCounter += 1 # Anzahl der bereits aufgeschriebenen Käufen/Verkäufen um eins erhöhen\n else:\n # Wenn kein Kauf/Verkauf getätigt wurde, jeden Wert eine leere Spalte anhängen um Format beizubehalten\n for element in daten[\"strategie_kaeufe_verkaeufe_zeitreihe\"][0]:\n if(element != \"zeitstempel\"):\n einzelneZeile.append(\"\")\n else:\n # Wenn kaufVerkaufCounter >= als die Anzahl der kaufVerkauf-Daten in strategie_kaeufe_verkaeufe_zeitreihe\n # Alle Käufe/Verkäufe wurden bereits aufgeschrieben, also leere Spalte anhängen um Format beizubehalten\n for element in daten[\"strategie_kaeufe_verkaeufe_zeitreihe\"][0]:\n if(element != \"zeitstempel\"):\n einzelneZeile.append(\"\")\n else:\n # Wenn keine Kauf/Verkauf Daten vorhanden, leere Spalte anhängen um Format beizubehalten\n einzelneZeile.append(\"\")\n einzelneZeile.append(\"\")\n einzelneZeile.append(\"\")\n einzelneZeile.append(\"\")\n\n einzelneZeile.append(\"\") #Spalte zwischen Bereichen\n\n # Für jedes Element in der indikator_zeitreihe ungleich dem zeitstempel, also für jeden Indikator\n for element in daten[\"indikator_zeitreihe\"][0]:\n if(element != \"zeitstempel\"):\n # Für jeden einzelnen Graphen des Indikators\n for graph in daten[\"indikator_zeitreihe\"][idx][element]:\n # Den Wert des jeweiligen Graphen an die Zeile hängen, Punkt mit Komma ersetzen für deutsche Darstellung\n einzelneZeile.append(\n str(daten[\"indikator_zeitreihe\"][idx][element][graph]).replace(\".\",\",\"))\n \n \n # die Zeile in CSV schreiben\n writer.writerow(einzelneZeile) \n\n return response", "sub_path": "src/simulation/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 25852, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "trading.mixins.fehlerViewMixin", "line_number": 41, "usage_type": "name"}, {"api_name": "trading.mixins.plotKonfigViewMixin", "line_number": 48, "usage_type": "name"}, {"api_name": "forms.SimulationModelForm", "line_number": 53, "usage_type": "name"}, {"api_name": "models.Simulation", "line_number": 55, "usage_type": "name"}, {"api_name": "django.views.View", "line_number": 57, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 77, "usage_type": "call"}, {"api_name": "django.shortcuts.reverse", "line_number": 77, "usage_type": "call"}, {"api_name": "trading.mixins.datenAnBackendSenden", "line_number": 91, "usage_type": "call"}, {"api_name": "trading.mixins.allgemeineFehlerPruefung", "line_number": 104, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 105, "usage_type": "call"}, {"api_name": "django.shortcuts.reverse", "line_number": 105, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 111, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 111, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 112, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 112, "usage_type": "name"}, {"api_name": "bokeh.plotting.figure", "line_number": 137, "usage_type": "call"}, {"api_name": "bokeh.plotting.figure", "line_number": 143, "usage_type": "call"}, {"api_name": "bokeh.plotting.figure", "line_number": 151, "usage_type": "call"}, {"api_name": "itertools.cycle", "line_number": 177, "usage_type": "call"}, {"api_name": "bokeh.palettes.Dark2_5", "line_number": 177, "usage_type": "argument"}, {"api_name": "pandas.DataFrame", "line_number": 180, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 181, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 183, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 184, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 186, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 187, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 189, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 244, "usage_type": "call"}, {"api_name": "bokeh.plotting.figure", "line_number": 249, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 262, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 298, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 304, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 322, "usage_type": "name"}, {"api_name": "bokeh.embed.components", "line_number": 326, "usage_type": "call"}, {"api_name": "bokeh.layouts.column", "line_number": 326, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 328, "usage_type": "call"}, {"api_name": "trading.mixins.datenAnBackendSenden", "line_number": 335, "usage_type": "call"}, {"api_name": "trading.mixins.allgemeineFehlerPruefung", "line_number": 344, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 345, "usage_type": "call"}, {"api_name": "django.shortcuts.reverse", "line_number": 345, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 357, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 357, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 358, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 360, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 443, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 443, "usage_type": "name"}]} +{"seq_id": "631952616", "text": "import torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport random\n\nfrom .neural_blocks import (\n SkipConnMLP, UpdateOperator, FourierEncoder, PositionalEncoder, NNEncoder,\n EncodedGRU,\n)\nfrom .utils import (\n dir_to_elev_azim, autograd, sample_random_hemisphere, laplace_cdf, load_sigmoid,\n)\nimport src.refl as refl\nfrom .renderers import ( load_occlusion_kind, direct )\nimport src.march as march\n\n@torch.jit.script\ndef cumuprod_exclusive(t):\n cp = torch.cumprod(t, dim=0)\n cp = torch.roll(cp, 1, dims=0)\n cp[0, ...] = 1.0\n return cp\n\n#@torch.jit.script # cannot jit script cause of tensordot :)\ndef compute_pts_ts(\n rays, near, far, steps, lindisp=False,\n perturb: float = 0,\n):\n r_o, r_d = rays.split([3,3], dim=-1)\n device = r_o.device\n if lindisp:\n t_vals = torch.linspace(0, 1, steps, device=device, dtype=r_o.dtype)\n ts = 1/(1/max(near, 1e-10) * (1-t_vals) + 1/far * (t_vals))\n else:\n ts = torch.linspace(near, far, steps=steps, device=device, dtype=r_o.dtype)\n\n if perturb > 0:\n mids = 0.5 * (ts[:-1] + ts[1:])\n lower = torch.cat([mids, ts[-1:]])\n upper = torch.cat([ts[:1], mids])\n rand = torch.rand_like(lower) * perturb\n ts = lower + (upper - lower) * rand\n pts = r_o.unsqueeze(0) + torch.tensordot(ts, r_d, dims = 0)\n return pts, ts, r_o, r_d\n\n# given a set of densities, and distances between the densities,\n# compute alphas from them.\n#@torch.jit.script\ndef alpha_from_density(\n density, ts, r_d,\n softplus: bool = True,\n):\n device=density.device\n\n if softplus: sigma_a = F.softplus(density-1)\n else: sigma_a = F.relu(density)\n\n end_val = torch.full_like(ts[..., :1], 1e10)\n dists = torch.cat([ts[..., 1:] - ts[..., :-1], end_val], dim=-1)\n while len(dists.shape) < 4: dists = dists[..., None]\n dists = dists * torch.linalg.norm(r_d, dim=-1)\n alpha = 1 - torch.exp(-sigma_a * dists)\n weights = alpha * cumuprod_exclusive(1.0 - alpha + 1e-10)\n return alpha, weights\n\n# TODO delete these for utils\n\n# sigmoids which shrink or expand the total range to prevent gradient vanishing,\n# or prevent it from representing full density items.\n# fat sigmoid has no vanishing gradient, but thin sigmoid leads to better outlines.\ndef fat_sigmoid(v, eps: float = 1e-3): return v.sigmoid() * (1+2*eps) - eps\ndef thin_sigmoid(v, eps: float = 1e-2): return fat_sigmoid(v, -eps)\ndef cyclic_sigmoid(v, eps:float=-1e-2,period:int=5):\n return ((v/period).sin()+1)/2 * (1+2*eps) - eps\n\n# perform volumetric integration of density with some other quantity\n# returns the integrated 2nd value over density at timesteps.\n@torch.jit.script\ndef volumetric_integrate(weights, other):\n return torch.sum(weights[..., None] * other, dim=0)\n\n# perform volumetric integration but only using some of other's values where the weights\n# are big enough.\n#\n# TODO the computation of `other` itself should be sparse, so that it doesn't need to be\n# computed in the first place.\n@torch.jit.script\ndef sparse_volumetric_integrate(weights, other, eps:float=1e-3):\n vals = torch.full_like(other, 1e-3)\n mask = weights > 1e-3\n vals[mask] = other[mask]\n return torch.sum(weights[..., None] * vals, dim=0)\n\n\n# bg functions, need to be here for pickling\ndef black(_elaz_r_d, _weights): return 0\ndef white(_, weights): 1-weights.sum(dim=0).unsqueeze(-1)\n# having a random color will probably help prevent any background\ndef random_color(_elaz_r_d, weights):\n # TODO need to think this through more\n # This will make it so that there never is a background.\n summed = (1-weights.sum(dim=0).unsqueeze(-1))\n return torch.rand_like(summed) * summed\n\nclass CommonNeRF(nn.Module):\n def __init__(\n self,\n\n steps: int = 64,\n\n #out_features: int = 3, # 3 is for RGB\n t_near: float = 0,\n t_far: float = 1,\n density_std: float = 0.01,\n noise_std: int = 1e-2,\n mip = None,\n instance_latent_size: int = 0,\n per_pixel_latent_size: int = 0,\n per_point_latent_size: int = 0,\n\n sigmoid_kind: str = \"thin\",\n bg: str = \"black\",\n\n record_depth: bool = False,\n\n device=\"cuda\",\n ):\n super().__init__()\n self.empty_latent = torch.zeros(1,1,1,1,0, device=device, dtype=torch.float)\n\n self.t_near = t_near\n self.t_far = t_far\n self.steps = steps\n self.mip = mip\n\n self.per_pixel_latent_size = per_pixel_latent_size\n self.per_pixel_latent = None\n\n self.instance_latent_size = instance_latent_size\n self.instance_latent = None\n\n self.per_pt_latent_size = per_point_latent_size\n self.per_pt_latent = None\n\n self.alpha = None\n self.noise_std = 0.2\n # TODO add activation for using sigmoid or fat sigmoid\n\n self.set_bg(bg)\n self.set_sigmoid(sigmoid_kind)\n\n self.record_depth = record_depth\n self.depth = None\n\n def forward(self, _x): raise NotImplementedError()\n def set_bg(self, bg=\"black\"):\n if bg == \"black\":\n self.sky_color = black\n elif bg == \"white\":\n self.sky_color = white\n elif bg == \"mlp\":\n self.sky_mlp = SkipConnMLP(\n in_size=2, out=3, enc=NNEncoder(in_size=2,out=3),\n num_layers=3, hidden_size=32, device=device, xavier_init=True,\n )\n self.sky_color = self.sky_from_mlp\n elif bg == \"random\":\n self.sky_color = random_color\n else:\n raise NotImplementedError(f\"Unexpected bg: {bg}\")\n\n def set_sigmoid(self, kind=\"thin\"):\n act = load_sigmoid(kind)\n self.feat_act = act\n if isinstance(self.refl, refl.LightAndRefl): self.refl.refl.act = act\n else: self.refl.act = act\n def sky_from_mlp(self, elaz_r_d, weights):\n return (1-weights.sum(dim=0)).unsqueeze(-1) * fat_sigmoid(self.sky_mlp(elaz_r_d))\n def total_latent_size(self) -> int:\n return self.mip_size() + \\\n self.per_pixel_latent_size + \\\n self.instance_latent_size + \\\n self.per_pt_latent_size\n def set_per_pt_latent(self, latent):\n assert(latent.shape[-1] == self.per_pt_latent_size), \\\n f\"expected latent in [T, B, H, W, L={self.per_pixel_latent_size}], got {latent.shape}\"\n assert(len(latent.shape) == 5), \\\n f\"expected latent in [T, B, H, W, L], got {latent.shape}\"\n self.per_pt_latent = latent\n def set_per_pixel_latent(self, latent):\n assert(latent.shape[-1] == self.per_pixel_latent_size), \\\n f\"expected latent in [B, H, W, L={self.per_pixel_latent_size}], got {latent.shape}\"\n assert(len(latent.shape) == 4), \\\n f\"expected latent in [B, H, W, L], got {latent.shape}\"\n self.per_pixel_latent = latent\n def set_instance_latent(self, latent):\n assert(latent.shape[-1] == self.instance_latent_size), \"expected latent in [B, L]\"\n assert(len(latent.shape) == 2), \"expected latent in [B, L]\"\n self.instance_latent = latent\n\n # produces a segmentation mask of sorts, using the alpha for occupancy.\n def acc(self): return self.alpha.max(dim=0)[0]\n def acc_smooth(self): return self.weights.sum(dim=0).unsqueeze(-1)\n def set_refl(self, refl):\n if hasattr(self, \"refl\"): self.refl = refl\n\n def depths(self, depths):\n with torch.no_grad():\n return volumetric_integrate(self.alpha, depths[..., None, None, None])\n\n @property\n def nerf(self): return self\n\n def mip_size(self): return 0 if self.mip is None else self.mip.size() * 6\n def mip_encoding(self, r_o, r_d, ts):\n if self.mip is None: return None\n end_val = torch.tensor([1e10], device=ts.device, dtype=ts.dtype)\n ts = torch.cat([ts, end_val], dim=-1)\n return self.mip(r_o, r_d, ts[..., :-1], ts[..., 1:])\n\n # gets the current latent vector for this NeRF instance\n def curr_latent(self, pts_shape) -> [\"T\", \"B\", \"H\", \"W\", \"L_pp + L_inst\"]:\n curr = self.empty_latent.expand(pts_shape[:-1] + (0,)) if self.per_pt_latent is None \\\n else self.per_pt_latent\n\n if self.per_pixel_latent is not None:\n ppl = self.per_pixel_latent[None, ...].expand(pts_shape[:-1] + (-1,))\n curr = torch.cat([curr, ppl], dim=-1)\n\n if self.instance_latent is not None:\n il = self.instance_latent[None, :, None, None, :].expand(pts_shape[:-1] + (-1,))\n curr = torch.cat([curr, il], dim=-1)\n\n return curr\n\nclass TinyNeRF(CommonNeRF):\n # No frills, single MLP NeRF\n def __init__(\n self,\n out_features: int = 3,\n device=\"cuda\",\n **kwargs,\n ):\n super().__init__(**kwargs, device=device)\n self.estim = SkipConnMLP(\n in_size=3, out=1 + out_features,\n latent_size = self.total_latent_size(),\n num_layers=6, hidden_size=128,\n\n xavier_init=True,\n )\n\n def forward(self, rays):\n pts, ts, r_o, r_d = compute_pts_ts(\n rays, self.t_near, self.t_far, self.steps,\n perturb = 1 if self.training else 0,\n )\n self.ts = ts\n return self.from_pts(pts, ts, r_o, r_d)\n\n def from_pts(self, pts, ts, r_o, r_d):\n latent = self.curr_latent(pts.shape)\n mip_enc = self.mip_encoding(r_o, r_d, ts)\n if mip_enc is not None: latent = torch.cat([latent, mip_enc], dim=-1)\n\n density, feats = self.estim(pts, latent).split([1, 3], dim=-1)\n\n self.alpha, self.weights = alpha_from_density(density, ts, r_d)\n return volumetric_integrate(self.weights, self.feat_act(feats)) + \\\n self.sky_color(None, self.weights)\n\n# A plain old nerf\nclass PlainNeRF(CommonNeRF):\n def __init__(\n self,\n intermediate_size: int = 32,\n out_features: int = 3,\n\n device: torch.device = \"cuda\",\n\n **kwargs,\n ):\n super().__init__(**kwargs, device=device)\n self.latent_size = self.total_latent_size()\n\n self.first = SkipConnMLP(\n in_size=3, out=1 + intermediate_size, latent_size=self.latent_size,\n enc=FourierEncoder(input_dims=3, device=device),\n\n num_layers = 6, hidden_size = 128, xavier_init=True,\n )\n\n self.refl = refl.View(\n out_features=out_features,\n latent_size=self.latent_size+intermediate_size,\n )\n\n def forward(self, rays):\n pts, ts, r_o, r_d = compute_pts_ts(\n rays, self.t_near, self.t_far, self.steps, perturb = 1 if self.training else 0,\n )\n self.ts = ts\n return self.from_pts(pts, ts, r_o, r_d)\n\n def from_pts(self, pts, ts, r_o, r_d):\n latent = self.curr_latent(pts.shape)\n\n mip_enc = self.mip_encoding(r_o, r_d, ts)\n\n # If there is a mip encoding, stack it with the latent encoding.\n if mip_enc is not None: latent = torch.cat([latent, mip_enc], dim=-1)\n\n first_out = self.first(pts, latent if latent.shape[-1] != 0 else None)\n\n density = first_out[..., 0]\n if self.training and self.noise_std > 0:\n density = density + torch.randn_like(density) * self.noise_std\n\n intermediate = first_out[..., 1:]\n\n #n = None\n #if self.refl.can_use_normal: n = autograd(pts, density)\n\n view = r_d[None, ...].expand_as(pts)\n rgb = self.refl(\n x=pts, view=view,\n latent=torch.cat([latent, intermediate], dim=-1),\n )\n\n self.alpha, self.weights = alpha_from_density(density, ts, r_d)\n return volumetric_integrate(self.weights, rgb) + self.sky_color(view, self.weights)\n\n# NeRF with a thin middle layer, for encoding information\nclass NeRFAE(CommonNeRF):\n def __init__(\n self,\n intermediate_size: int = 32,\n out_features: int = 3,\n\n encoding_size: int = 32,\n normalize_latent: bool = True,\n\n device=\"cuda\",\n **kwargs,\n ):\n super().__init__(**kwargs, device=device)\n\n self.latent_size = self.total_latent_size()\n\n self.encode = SkipConnMLP(\n in_size=3, out=encoding_size,\n latent_size=self.latent_size,\n num_layers=5, hidden_size=128,\n enc=FourierEncoder(input_dims=3, device=device),\n xavier_init=True,\n )\n\n self.density_tform = SkipConnMLP(\n in_size=encoding_size, out=1+intermediate_size, latent_size=0,\n num_layers=5, hidden_size=64, xavier_init=True,\n )\n\n self.refl = refl.View(\n out_features=out_features,\n latent_size=encoding_size+intermediate_size,\n )\n self.encoding_size = encoding_size\n self.regularize_latent = False\n self.normalize_latent = normalize_latent\n\n def set_regularize_latent(self):\n self.regularize_latent = True\n self.latent_l2_loss = 0\n def forward(self, rays):\n pts, ts, r_o, r_d = compute_pts_ts(\n rays, self.t_near, self.t_far, self.steps,\n perturb = 1 if self.training else 0,\n )\n self.ts = ts\n return self.from_pts(pts, ts, r_o, r_d)\n\n def from_pts(self, pts, ts, r_o, r_d):\n encoded = self.compute_encoded(pts, ts, r_o, r_d)\n if self.regularize_latent:\n self.latent_l2_loss = torch.linalg.norm(encoded, dim=-1).square().mean()\n return self.from_encoded(encoded, ts, r_d, pts)\n\n def compute_encoded(self, pts, ts, r_o, r_d):\n latent = self.curr_latent(pts.shape)\n\n mip_enc = self.mip_encoding(r_o, r_d, ts)\n\n # If there is a mip encoding, stack it with the latent encoding.\n if mip_enc is not None: latent = torch.cat([latent, mip_enc], dim=-1)\n\n return self.encode(pts, latent if latent.shape[-1] != 0 else None)\n def from_encoded(self, encoded, ts, r_d, pts):\n if self.normalize_latent: encoded = F.normalize(encoded, dim=-1)\n\n first_out = self.density_tform(encoded)\n density = first_out[..., 0]\n intermediate = first_out[..., 1:]\n\n if self.training and self.noise_std > 0:\n density = density + torch.randn_like(density) * self.noise_std\n\n rgb = self.refl(\n x=pts, view=r_d[None,...].expand_as(pts),\n latent=torch.cat([encoded,intermediate],dim=-1),\n )\n\n self.alpha, self.weights = alpha_from_density(density, ts, r_d)\n\n color = volumetric_integrate(self.weights, rgb)\n sky = self.sky_color(None, self.weights)\n return color + sky\n\ndef identity(x): return x\n\n# https://arxiv.org/pdf/2106.12052.pdf\nclass VolSDF(CommonNeRF):\n def __init__(\n self, sdf,\n # how many features to pass from density to RGB\n intermediate_size: int = 32, out_features: int = 3,\n device: torch.device = \"cuda\",\n\n occ_kind=None,\n w_missing:bool = False,\n integrator_kind=\"direct\",\n scale_softplus=False,\n **kwargs,\n ):\n super().__init__(**kwargs, device=device)\n self.sdf = sdf\n # the reflectance model is in the SDF, so don't encode it here.\n self.scale = nn.Parameter(torch.tensor(0.1, requires_grad=True, device=device))\n self.secondary = None\n self.out_features = out_features\n self.scale_act = identity if not scale_softplus else nn.Softplus()\n if occ_kind is not None:\n assert(isinstance(self.sdf.refl, refl.LightAndRefl)), \\\n f\"Must have light w/ volsdf integration {type(self.sdf.refl)}\"\n self.occ = load_occlusion_kind(occ_kind, self.sdf.latent_size)\n if integrator_kind == \"direct\": self.secondary = self.direct\n elif integrator_kind == \"path\": self.convert_to_path(w_missing)\n else: raise NotImplementedError(f\"unknown integrator kind {integrator_kind}\")\n def convert_to_path(self, w_missing: bool):\n if self.secondary == self.path: return False\n self.secondary = self.path\n self.path_n = N = 3\n missing_cmpts = 3 * (N + 1) + 6\n\n # this is a function of the seen pts and the sampled lighting dir\n self.missing = None\n if w_missing:\n self.missing = SkipConnMLP(\n in_size=missing_cmpts, out=self.out_features,\n enc=FourierEncoder(input_dims=missing_cmpts),\n # here we care about the aggregate set of all point, so bundle them all up.\n latent_size = self.sdf.latent_size * (N + 1), hidden_size=512,\n )\n\n # transfer_fn := G(x1, x2) -> [0,1]\n self.transfer_fn = SkipConnMLP(\n in_size=6, out=1, enc=FourierEncoder(input_dims=6),\n # multiply by two here ince it's the pair of latent values at sets of point\n latent_size = self.sdf.latent_size * 2,\n hidden_size=512,\n )\n return True\n def direct(self, r_o, weights, pts, view, n, latent):\n out = torch.zeros_like(pts)\n for light in self.sdf.refl.light.iter():\n light_dir, light_val = self.occ(pts, light, self.sdf.intersect_mask, latent=latent)\n bsdf_val = self.sdf.refl(x=pts, view=view, normal=n, light=light_dir, latent=latent)\n out = out + bsdf_val * light_val\n return out\n def path(self, r_o, weights, pts, view, n, latent):\n out = torch.zeros_like(pts)\n\n N = self.path_n # number of samples for 1st order bounces\n\n # for each point sample some number of directions\n dirs = sample_random_hemisphere(n, num_samples=N)\n # compute intersection of random directions with surface\n ext_pts, ext_hits, dists, _ = march.bisect(\n self.sdf.underlying, pts[None,...].expand_as(dirs), dirs, iters=64, near=5e-3, far=6,\n )\n # TODO does not decay with the square of distance, need to add in a flag for this\n # if the model assumes that it does.\n # decays = 1/dists.square().clamp(min=1e-8)\n\n ext_sdf_vals, ext_latent = self.sdf.from_pts(ext_pts)\n\n ext_view = F.normalize(ext_pts - r_o[None,None,...], dim=-1)\n # detach secondary normals\n ext_n = F.normalize(self.sdf.normals(ext_pts), dim=-1).detach()\n\n fit = lambda x: x.unsqueeze(0).expand(N,-1,-1,-1,-1,-1)\n # reflection at the intersection points from light incoming from the random directions\n first_step_bsdf = self.sdf.refl(\n x=fit(pts), view=ext_view, normal=fit(n), light=-dirs, latent=fit(latent),\n )\n # compute transfer function (G) between ext_pts and pts (which is a proxy for the density).\n tf = self.transfer_fn(\n torch.cat([ext_pts, pts.unsqueeze(0).expand_as(ext_pts)],dim=-1),\n torch.cat([ext_latent, latent.unsqueeze(0).expand_as(ext_latent)], dim=-1),\n ).sigmoid()\n first_step_bsdf = first_step_bsdf * tf # * decays\n\n for light in self.sdf.refl.light.iter():\n # compute direct lighting at each point (identical to direct)\n light_dir, light_val = self.occ(pts, light, self.sdf.intersect_mask, latent=latent)\n bsdf_val = self.sdf.refl(x=pts, view=view, normal=n, light=light_dir, latent=latent)\n out = out + bsdf_val * light_val\n # compute light contribution and bsdf at 2ndary points from this light\n ext_light_dir, ext_light_val = \\\n self.occ(ext_pts, light, self.sdf.intersect_mask, latent=ext_latent)\n path_bsdf = self.sdf.refl(\n x=ext_pts, view=dirs, normal=ext_n, light=ext_light_dir, latent=ext_latent,\n )\n second_step = ext_light_val * path_bsdf\n # sum over the contributions at each point adding with each secondary contribution\n secondary = (first_step_bsdf * second_step).sum(dim=0)\n out = out + secondary\n # because we have high sampling variance, add in a secondary component which accounts for\n # unsampled values by taking the points sampled and the current set of points.\n # This makes it possible to learn outside of the scope of what is possible, but should\n # converge faster?\n # we explicitly allow it to be negative in case the points we pick are all sampled with\n # super high value.\n if self.missing is None: continue\n missing = self.missing(\n torch.cat([\n ext_pts.reshape((*ext_pts.shape[1:-1], 3 * N)), pts, light_dir, view,\n ], dim=-1),\n torch.cat([\n ext_latent.reshape((*ext_latent.shape[1:-1], self.sdf.latent_size * N)), latent,\n ], dim=-1),\n )\n missing = self.feat_act(missing)\n out = out + missing\n return out\n def forward(self, rays):\n pts, ts, r_o, r_d = compute_pts_ts(\n rays, self.t_near, self.t_far, self.steps, perturb = 1 if self.training else 0,\n )\n self.ts = ts\n return self.from_pts(pts, ts, r_o, r_d)\n def total_latent_size(self): return self.sdf.latent_size\n def set_refl(self, refl): self.sdf.refl = refl\n\n @property\n def refl(self): return self.sdf.refl\n\n def from_pts(self, pts, ts, r_o, r_d):\n latent = self.curr_latent(pts.shape)\n mip_enc = self.mip_encoding(r_o, r_d, ts)\n if mip_enc is not None: latent = torch.cat([latent, mip_enc], dim=-1)\n\n sdf_vals, latent = self.sdf.from_pts(pts)\n scale = self.scale_act(self.scale)\n self.scale_post_act = scale\n density = 1/scale * laplace_cdf(-sdf_vals, scale)\n self.alpha, self.weights = alpha_from_density(density, ts, r_d, softplus=False)\n\n n = None\n if self.sdf.refl.can_use_normal or self.secondary is not None:\n self.n = n = F.normalize(self.sdf.normals(pts), dim=-1)\n\n view = r_d.unsqueeze(0).expand_as(pts)\n if self.secondary is None: rgb = self.sdf.refl(x=pts, view=view, normal=n, latent=latent)\n else: rgb = self.secondary(r_o, self.weights, pts, view, n, latent)\n\n return volumetric_integrate(self.weights, rgb)\n def set_sigmoid(self, kind=\"thin\"):\n if not hasattr(self, \"sdf\"): return\n act = load_sigmoid(kind)\n if isinstance(self.refl, refl.LightAndRefl): self.refl.refl.act = act\n else: self.refl.act = act\n\nclass RecurrentNeRF(CommonNeRF):\n def __init__(\n self,\n intermediate_size: int = 64,\n out_features: int = 3,\n\n device: torch.device = \"cuda\",\n\n **kwargs,\n ):\n super().__init__(**kwargs, device=device)\n self.latent_size = self.total_latent_size()\n\n self.first = EncodedGRU(\n encs=[\n FourierEncoder(input_dims=3, sigma=1<<1, device=device),\n FourierEncoder(input_dims=3, sigma=1<<2, device=device),\n FourierEncoder(input_dims=3, sigma=1<<3, device=device),\n FourierEncoder(input_dims=3, sigma=1<<3, device=device),\n FourierEncoder(input_dims=3, sigma=1<<4, device=device),\n FourierEncoder(input_dims=3, sigma=1<<4, device=device),\n FourierEncoder(input_dims=3, sigma=1<<5, device=device),\n ],\n state_size=256,\n in_size=3, out=1,\n latent_out=intermediate_size,\n )\n\n self.refl = refl.View(\n out_features=out_features,\n latent_size=self.latent_size+intermediate_size,\n )\n\n def forward(self, rays):\n pts, ts, r_o, r_d = compute_pts_ts(\n rays, self.t_near, self.t_far, self.steps, perturb = 1 if self.training else 0,\n )\n self.ts = ts\n return self.from_pts(pts, ts, r_o, r_d)\n\n def from_pts(self, pts, ts, r_o, r_d):\n latent = self.curr_latent(pts.shape)\n\n densities, intermediate = self.first(pts, latent if latent.shape[-1] != 0 else None)\n acc_density = (torch.cumsum(densities, dim=-1) - densities).detach() + densities\n if self.training and self.noise_std > 0:\n acc_density = acc_density + torch.randn_like(acc_density) * self.noise_std\n\n view = r_d[None, ...].expand_as(pts)\n rgb = self.refl(x=pts, view=view, latent=torch.cat([latent, intermediate], dim=-1))\n images = []\n for i in range(acc_density.shape[-1]):\n density = acc_density[..., i]\n alpha, weights = alpha_from_density(density, ts, r_d)\n img = volumetric_integrate(weights, rgb)\n images.append(img)\n return images\n\ndef alternating_volsdf_loss(model, nerf_loss, sdf_loss):\n def aux(x, ref): return nerf_loss(x, ref[..., :3]) if model.vol_render else sdf_loss(x, ref)\n return aux\n\n# An odd module which alternates between volume rendering and SDF rendering\nclass AlternatingVolSDF(nn.Module):\n def __init__(\n self,\n volsdf: VolSDF,\n # run_len is how many iterations of volume/SDF rendering it will perform.\n # it performs run_len/2 volume, and run_len/2 SDF\n run_len:int = 4096,\n ):\n super().__init__()\n assert(isinstance(volsdf, VolSDF))\n self.volsdf = volsdf\n self.i = 0\n self.force_volume = False\n self.force_sdf = False\n self.run_len = run_len\n # TODO add some count for doing only sdfs first?\n\n # forward a few properties to sdf\n @property\n def sdf(self): return self.volsdf.sdf\n @property\n def nerf(self): return self.volsdf\n @property\n def n(self): return self.volsdf.n\n @property\n def total_latent_size(self): return self.volsdf.total_latent_size\n @property\n def refl(self): return self.volsdf.refl\n def set_refl(self, refl): return self.volsdf.set_refl(refl)\n\n def forward(self, rays):\n if not self.training: return self.volsdf(rays)\n self.i = (self.i + 1) % self.run_len\n self.vol_render = (self.i < self.run_len//2 or self.force_volume) and not self.force_sdf\n if self.vol_render:\n return self.volsdf(rays)\n else:\n return direct(self.volsdf.sdf, self.volsdf.refl, self.volsdf.occ, rays, self.training)\n\n# Dynamic NeRF for multiple frams\nclass DynamicNeRF(nn.Module):\n def __init__(self, canonical: CommonNeRF, gru_flow:bool=False, device=\"cuda\"):\n super().__init__()\n self.canonical = canonical\n\n if gru_flow:\n self.delta_estim = UpdateOperator(in_size=4, out_size=3, hidden_size=32)\n else:\n self.delta_estim = SkipConnMLP(\n # x,y,z,t -> dx, dy, dz\n in_size=4, out=3,\n\n num_layers = 5, hidden_size = 128,\n enc=NNEncoder(input_dims=4),\n activation=nn.Softplus(),\n zero_init=True,\n )\n self.time_noise_std = 3e-3\n self.smooth_delta = False\n self.delta_smoothness = 0\n\n @property\n def nerf(self): return self.canonical\n\n @property\n def sdf(self): return getattr(self.canonical, \"sdf\", None)\n\n def set_smooth_delta(self): setattr(self, \"smooth_delta\", True)\n def forward(self, rays_t):\n rays, t = rays_t\n device=rays.device\n pts, ts, r_o, r_d = compute_pts_ts(\n rays, self.canonical.t_near, self.canonical.t_far, self.canonical.steps,\n perturb = 1 if self.training else 0,\n )\n self.ts = ts\n # small deviation for regularization\n if self.training and self.time_noise_std > 0:\n t = t + self.time_noise_std * torch.randn_like(t)\n\n t = t[None, :, None, None, None].expand(pts.shape[:-1] + (1,))\n\n pts_t = torch.cat([pts, t], dim=-1)\n dp = self.delta_estim(pts_t)\n dp = torch.where(t.abs() < 1e-6, torch.zeros_like(pts), dp)\n #if self.training and self.smooth_delta:\n # self.delta_smoothness = self.delta_estim.l2_smoothness(pts_t, dp)\n pts = pts + dp\n return self.canonical.from_pts(pts, ts, r_o, r_d)\n\n# Dynamic NeRFAE for multiple framss with changing materials\nclass DynamicNeRFAE(nn.Module):\n def __init__(self, canonical: NeRFAE, gru_flow: bool=False, device=\"cuda\"):\n super().__init__()\n assert(isinstance(canonical, NeRFAE)), \"Must use NeRFAE for DynamicNeRFAE\"\n self.canon = canonical.to(device)\n\n self.delta_estim = SkipConnMLP(\n # x,y,z,t -> dx, dy, dz\n in_size=4, out=3 + canonical.encoding_size,\n num_layers = 6, hidden_size = 128,\n enc=NNEncoder(input_dims=4, device=device),\n\n activation=nn.Softplus(), zero_init=True,\n )\n\n self.smooth_delta = False\n self.tf_smoothness = 0\n self.time_noise_std = 1e-3\n\n @property\n def nerf(self): return self.canon\n def set_smooth_delta(self): setattr(self, \"smooth_delta\", True)\n def forward(self, rays_t):\n rays, t = rays_t\n device=rays.device\n\n pts, ts, r_o, r_d = compute_pts_ts(\n rays, self.canon.t_near, self.canon.t_far, self.canon.steps,\n )\n self.ts = ts\n # small deviation for regularization\n if self.training and self.time_noise_std > 0: t = t + self.time_noise_std * torch.randn_like(t)\n t = t[None, :, None, None, None].expand(pts.shape[:-1] + (1,))\n # compute encoding using delta positions at a given time\n pts_t = torch.cat([pts, t], dim=-1)\n delta = self.delta_estim(pts_t)\n #delta = torch.where(t.abs() < 1e-6, torch.zeros_like(delta), delta)\n dp, d_enc = delta.split([3, self.canon.encoding_size], dim=-1)\n encoded = self.canon.compute_encoded(pts + dp, ts, r_o, r_d)\n\n # TODO is this best as a sum, or is some other kind of tform better?\n return self.canon.from_encoded(encoded + d_enc, ts, r_d, pts)\n\nclass SinglePixelNeRF(nn.Module):\n def __init__(\n self,\n canon: CommonNeRF,\n encoder,\n img,\n\n device: torch.device = \"cuda\",\n ):\n super().__init__()\n self.canon = canon\n self.encoder = encoder\n # encode image\n self.encoder(img)\n\n self.device = device\n\n @property\n def nerf(self): return self.canon\n def forward(self, rays_uvs):\n rays, uvs = rays_uvs\n latent = self.encoder.sample(uvs)\n self.canon.set_per_pixel_latent(latent)\n return self.canon(rays)\n\nclass MPI(nn.Module):\n # Multi Plane Imaging.\n def __init__(\n self,\n canonical: CommonNeRF,\n\n position = [0,0,0],\n normal = [0,0,-1],\n delta=0.1,\n\n n_planes: int = 6,\n\n device=\"cuda\",\n ):\n super().__init__()\n\n self.n_planes = torch.linspace(canon.t_near, canon.t_far, steps=n_planes, device=device)\n self.position = torch.tensor(position, device=device, dtype=torch.float)\n self.normal = torch.tensor(normal, device=device, dtype=torch.float)\n self.delta = delta\n\n self.canonical = canonical.to(device)\n def forward(self, rays):\n r_o, r_d = rays.split([3,3], dim=-1)\n device = r_o.device\n\n n = self.normal.expand_as(r_d)\n denom = (n * r_d).sum(dim=-1, keepdim=True)\n centers = self.position.unsqueeze(0) + torch.tensordot(\n self.delta * torch.arange(self.n_planes, device=device, dtype=torch.float),\n -self.normal, dims=0,\n )\n ts = ((centers - r_o) * n).sum(dim=-1, keepdim=True)/denom\n # if denom is too small it will have numerical instability because it's near parallel.\n hits = torch.where(denom.abs() > 1e-3, ts, torch.zeros_like(denom))\n pts = r_o.unsqueeze(0) + r_d.unsqueeze(0) * hits\n\n return self.canonical.from_pts(pts, ts, r_o, r_d)\n\n @property\n def nerf(self): return self.canon\n def from_pts(self, pts, ts, r_o, r_d):\n density, feats = self.estim(pts).split([1, 3], dim=-1)\n\n alpha, weights = alpha_from_density(density, ts, r_d)\n return volumetric_integrate(weights, self.feat_act(feats))\n\n\n# TODO test this as well\ndef metropolis_sampling(\n density_estimator,\n ts_init, r_o, r_d,\n iters: int = 6,\n):\n # need to make this the shape of r_d exit with last dim 1\n curr = ts_init\n print(r_o.shape)\n exit()\n with torch.no_grad():\n candidates = torch.rand_like(curr) + curr\n curr_density = density_estimator(candidates)\n for i in range(iters):\n candidates = torch.randn_like(curr) + curr\n density = density_estimator(candidates)\n acceptance = density/curr_density\n alphas = torch.rand_like(density)\n mask = acceptance <= alphas\n curr = torch.where(mask, candidates, curr)\n curr_density = torch.where(mask, density, curr_density)\n return curr, r_o + curr * r_d\n\n# TODO need to test this more, doesn't seem to work that well\ndef inverse_sample(\n density_estimator,\n pts, ts, r_o, r_d,\n):\n with torch.no_grad():\n _, weights = alpha_from_density(density_estimator(pts.squeeze(-1)), ts, r_d)\n weights = weights.clamp(min=1e-10)\n pdf = weights/weights.sum(dim=0,keepdim=True)\n cdf = torch.cumsum(pdf, dim=0)\n N = ts.shape[0]\n # XXX this only works because we assume that the number of samples (N) is the same.\n #u = torch.rand_like(cdf)\n u = torch.linspace(0, 1, N, device=cdf.device)\\\n [..., None, None, None].expand_as(cdf)\n # XXX this operates on innermost dimension, so need to do this transpose\n inds = torch.searchsorted(\n cdf.transpose(0, -1).contiguous(), u.transpose(0, -1).contiguous(), right=True\n ).transpose(0, -1)\n below = (inds - 1).clamp(min=0)\n above = inds.clamp(max=N-1)\n inds_g = torch.stack([below, above], dim=-1)\n\n # TODO what is the right dimension to add here?\n cdf_g = torch.gather(cdf.unsqueeze(1).expand_as(inds_g), 0, inds_g)\n bins_g = torch.gather(ts[:, None, None, None, None].expand_as(inds_g), 0, inds_g)\n\n denom = cdf_g[..., 1] - cdf_g[..., 0]\n denom = torch.where(denom < 1e-5, torch.ones_like(denom), denom)\n t = (u - cdf_g[..., 0]) / denom\n samples = bins_g[..., 0] + t * (bins_g[..., 1] - bins_g[..., 0])\n new_pts = r_o + samples.unsqueeze(-1) * r_d\n return samples, new_pts\n", "sub_path": "src/nerf.py", "file_name": "nerf.py", "file_ext": "py", "file_size_in_byte": 31485, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "torch.cumprod", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.roll", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.jit", "line_number": 17, "usage_type": "attribute"}, {"api_name": "torch.linspace", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.linspace", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.rand_like", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.tensordot", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.nn.functional.softplus", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 55, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 56, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 56, "usage_type": "name"}, {"api_name": "torch.full_like", "line_number": 58, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.linalg.norm", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.linalg", "line_number": 61, "usage_type": "attribute"}, {"api_name": "torch.exp", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 80, "usage_type": "call"}, {"api_name": "torch.jit", "line_number": 78, "usage_type": "attribute"}, {"api_name": "torch.full_like", "line_number": 89, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.jit", "line_number": 87, "usage_type": "attribute"}, {"api_name": "torch.rand_like", "line_number": 103, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 105, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 105, "usage_type": "name"}, {"api_name": "torch.zeros", "line_number": 129, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 129, "usage_type": "attribute"}, {"api_name": "neural_blocks.SkipConnMLP", "line_number": 162, "usage_type": "call"}, {"api_name": "neural_blocks.NNEncoder", "line_number": 163, "usage_type": "call"}, {"api_name": "utils.load_sigmoid", "line_number": 173, "usage_type": "call"}, {"api_name": "src.refl.LightAndRefl", "line_number": 175, "usage_type": "attribute"}, {"api_name": "src.refl", "line_number": 175, "usage_type": "name"}, {"api_name": "src.refl", "line_number": 205, "usage_type": "name"}, {"api_name": "torch.no_grad", "line_number": 208, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 217, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 218, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 228, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 232, "usage_type": "call"}, {"api_name": "neural_blocks.SkipConnMLP", "line_number": 245, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 264, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 279, "usage_type": "attribute"}, {"api_name": "neural_blocks.SkipConnMLP", "line_number": 286, "usage_type": "call"}, {"api_name": "neural_blocks.FourierEncoder", "line_number": 288, "usage_type": "call"}, {"api_name": "src.refl.View", "line_number": 293, "usage_type": "call"}, {"api_name": "src.refl", "line_number": 293, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 311, "usage_type": "call"}, {"api_name": "torch.randn_like", "line_number": 317, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 327, "usage_type": "call"}, {"api_name": "neural_blocks.SkipConnMLP", "line_number": 350, "usage_type": "call"}, {"api_name": "neural_blocks.FourierEncoder", "line_number": 354, "usage_type": "call"}, {"api_name": "neural_blocks.SkipConnMLP", "line_number": 358, "usage_type": "call"}, {"api_name": "src.refl.View", "line_number": 363, "usage_type": "call"}, {"api_name": "src.refl", "line_number": 363, "usage_type": "name"}, {"api_name": "torch.linalg.norm", "line_number": 385, "usage_type": "call"}, {"api_name": "torch.linalg", "line_number": 385, "usage_type": "attribute"}, {"api_name": "torch.cat", "line_number": 394, "usage_type": "call"}, {"api_name": "torch.nn.functional.normalize", "line_number": 398, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 398, "usage_type": "name"}, {"api_name": "torch.randn_like", "line_number": 405, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 409, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 426, "usage_type": "attribute"}, {"api_name": "torch.nn.Parameter", "line_number": 437, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 437, "usage_type": "name"}, {"api_name": "torch.tensor", "line_number": 437, "usage_type": "call"}, {"api_name": "torch.nn.Softplus", "line_number": 440, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 440, "usage_type": "name"}, {"api_name": "src.refl.LightAndRefl", "line_number": 442, "usage_type": "attribute"}, {"api_name": "src.refl", "line_number": 442, "usage_type": "name"}, {"api_name": "renderers.load_occlusion_kind", "line_number": 444, "usage_type": "call"}, {"api_name": "neural_blocks.SkipConnMLP", "line_number": 457, "usage_type": "call"}, {"api_name": "neural_blocks.FourierEncoder", "line_number": 459, "usage_type": "call"}, {"api_name": "neural_blocks.SkipConnMLP", "line_number": 465, "usage_type": "call"}, {"api_name": "neural_blocks.FourierEncoder", "line_number": 466, "usage_type": "call"}, {"api_name": "torch.zeros_like", "line_number": 473, "usage_type": "call"}, {"api_name": "torch.zeros_like", "line_number": 480, "usage_type": "call"}, {"api_name": "utils.sample_random_hemisphere", "line_number": 485, "usage_type": "call"}, {"api_name": "src.march.bisect", "line_number": 487, "usage_type": "call"}, {"api_name": "src.march", "line_number": 487, "usage_type": "name"}, {"api_name": "torch.nn.functional.normalize", "line_number": 496, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 496, "usage_type": "name"}, {"api_name": "torch.nn.functional.normalize", "line_number": 498, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 498, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 507, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 508, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 535, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 538, "usage_type": "call"}, {"api_name": "src.refl", "line_number": 552, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 560, "usage_type": "call"}, {"api_name": "utils.laplace_cdf", "line_number": 565, "usage_type": "call"}, {"api_name": "torch.nn.functional.normalize", "line_number": 570, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 570, "usage_type": "name"}, {"api_name": "utils.load_sigmoid", "line_number": 579, "usage_type": "call"}, {"api_name": "src.refl.LightAndRefl", "line_number": 580, "usage_type": "attribute"}, {"api_name": "src.refl", "line_number": 580, "usage_type": "name"}, {"api_name": "torch.device", "line_number": 589, "usage_type": "attribute"}, {"api_name": "neural_blocks.EncodedGRU", "line_number": 596, "usage_type": "call"}, {"api_name": "neural_blocks.FourierEncoder", "line_number": 598, "usage_type": "call"}, {"api_name": "neural_blocks.FourierEncoder", "line_number": 599, "usage_type": "call"}, {"api_name": "neural_blocks.FourierEncoder", "line_number": 600, "usage_type": "call"}, {"api_name": "neural_blocks.FourierEncoder", "line_number": 601, "usage_type": "call"}, {"api_name": "neural_blocks.FourierEncoder", "line_number": 602, "usage_type": "call"}, {"api_name": "neural_blocks.FourierEncoder", "line_number": 603, "usage_type": "call"}, {"api_name": "neural_blocks.FourierEncoder", "line_number": 604, "usage_type": "call"}, {"api_name": "src.refl.View", "line_number": 611, "usage_type": "call"}, {"api_name": "src.refl", "line_number": 611, "usage_type": "name"}, {"api_name": "torch.cumsum", "line_number": 627, "usage_type": "call"}, {"api_name": "torch.randn_like", "line_number": 629, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 632, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 646, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 646, "usage_type": "name"}, {"api_name": "src.refl", "line_number": 674, "usage_type": "argument"}, {"api_name": "renderers.direct", "line_number": 683, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 686, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 686, "usage_type": "name"}, {"api_name": "neural_blocks.UpdateOperator", "line_number": 692, "usage_type": "call"}, {"api_name": "neural_blocks.SkipConnMLP", "line_number": 694, "usage_type": "call"}, {"api_name": "neural_blocks.NNEncoder", "line_number": 699, "usage_type": "call"}, {"api_name": "torch.nn.Softplus", "line_number": 700, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 700, "usage_type": "name"}, {"api_name": "torch.randn_like", "line_number": 724, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 728, "usage_type": "call"}, {"api_name": "torch.where", "line_number": 730, "usage_type": "call"}, {"api_name": "torch.zeros_like", "line_number": 730, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 737, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 737, "usage_type": "name"}, {"api_name": "neural_blocks.SkipConnMLP", "line_number": 743, "usage_type": "call"}, {"api_name": "neural_blocks.NNEncoder", "line_number": 747, "usage_type": "call"}, {"api_name": "torch.nn.Softplus", "line_number": 749, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 749, "usage_type": "name"}, {"api_name": "torch.randn_like", "line_number": 768, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 771, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 780, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 780, "usage_type": "name"}, {"api_name": "torch.device", "line_number": 787, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 805, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 805, "usage_type": "name"}, {"api_name": "torch.linspace", "line_number": 821, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 822, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 822, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 823, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 823, "usage_type": "attribute"}, {"api_name": "torch.tensordot", "line_number": 833, "usage_type": "call"}, {"api_name": "torch.arange", "line_number": 834, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 834, "usage_type": "attribute"}, {"api_name": "torch.where", "line_number": 839, "usage_type": "call"}, {"api_name": "torch.zeros_like", "line_number": 839, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 863, "usage_type": "call"}, {"api_name": "torch.rand_like", "line_number": 864, "usage_type": "call"}, {"api_name": "torch.randn_like", "line_number": 867, "usage_type": "call"}, {"api_name": "torch.rand_like", "line_number": 870, "usage_type": "call"}, {"api_name": "torch.where", "line_number": 872, "usage_type": "call"}, {"api_name": "torch.where", "line_number": 873, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 881, "usage_type": "call"}, {"api_name": "torch.cumsum", "line_number": 885, "usage_type": "call"}, {"api_name": "torch.linspace", "line_number": 889, "usage_type": "call"}, {"api_name": "torch.searchsorted", "line_number": 892, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 897, "usage_type": "call"}, {"api_name": "torch.gather", "line_number": 900, "usage_type": "call"}, {"api_name": "torch.gather", "line_number": 901, "usage_type": "call"}, {"api_name": "torch.where", "line_number": 904, "usage_type": "call"}, {"api_name": "torch.ones_like", "line_number": 904, "usage_type": "call"}]} +{"seq_id": "272949142", "text": "import os\nimport re\nimport logging\nfrom typing import Union\n\nimport settings\nfrom impf.constructors import zulip_client, zulip_send_payload, zulip_read_payload, get_command\n\nimport requests\n\nHEADERS = {\n 'Accept': 'application/json',\n 'User-Agent': 'https://github.com/alfonsrv/impf-botpy'\n}\n\nlogger = logging.getLogger(__name__)\np = re.compile(r\"sms:\\d{3}-?\\d{3}\")\n\n\ndef sms_code(string: str) -> Union[str, None]:\n \"\"\" Checks if string contains a valid SMS code \"\"\"\n m = p.search(string.strip())\n if m: m = m.group().replace('-', '').replace('sms:', '')\n return m\n\n\ndef read_code() -> str:\n \"\"\" Reads the alert code from any given platform and returns it as string \"\"\"\n code = ''\n if settings.ZULIP_ENABLED:\n _code = zulip_read()\n if _code:\n logger.info(f'Read SMS Code from Zulip: {_code}')\n code = _code\n\n return code\n\n\ndef send_alert(message: str) -> None:\n logger.info(f'Sending alert \"{message}\"')\n if settings.COMMAND_ENABLED:\n try: os.system(get_command())\n except: pass\n if settings.ZULIP_ENABLED:\n zulip_send(message)\n if settings.TELEGRAM_ENABLED:\n telegram_send(message)\n if settings.PUSHOVER_ENABLED:\n pushover_send(message)\n\n\ndef zulip_send(message: str) -> None:\n client = zulip_client()\n if client is None: return\n request = zulip_send_payload()\n request.setdefault('content', message)\n r = client.send_message(request)\n if r.get('result') != 'success':\n logger.error(f'Error sending Zulip message - got {r}')\n\n\ndef zulip_read() -> str:\n client = zulip_client()\n if client is None: return\n request = zulip_read_payload()\n r = client.get_messages(request)\n\n for message in r.get('messages'):\n if sms_code(message.get('content')):\n return sms_code(message.get('content'))\n\n\ndef telegram_send(message: str) -> None:\n api_token = settings.TELEGRAM_BOT_TOKEN\n chat_id = settings.TELEGRAM_BOT_CHATID\n\n url = f'https://api.telegram.org/bot{api_token}/sendMessage'\n params = {\n 'chat_id': chat_id,\n 'parse_mode': 'Markdown',\n 'text': message\n }\n\n response = requests.get(url, params=params, headers=HEADERS)\n logger.debug(response)\n\n\ndef pushover_send(message: str) -> None:\n app_token = settings.PUSHOVER_APP_TOKEN\n user_key = settings.PUSHOVER_USER_KEY\n\n url = f'https://api.pushover.net/1/messages.json'\n data = {\n 'token': app_token,\n 'user': user_key,\n 'message': message\n }\n\n response = requests.post(url, data=data)\n logger.debug(response)", "sub_path": "impf/alert.py", "file_name": "alert.py", "file_ext": "py", "file_size_in_byte": 2612, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "logging.getLogger", "line_number": 16, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 17, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 20, "usage_type": "name"}, {"api_name": "settings.ZULIP_ENABLED", "line_number": 30, "usage_type": "attribute"}, {"api_name": "settings.COMMAND_ENABLED", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 42, "usage_type": "call"}, {"api_name": "impf.constructors.get_command", "line_number": 42, "usage_type": "call"}, {"api_name": "settings.ZULIP_ENABLED", "line_number": 44, "usage_type": "attribute"}, {"api_name": "settings.TELEGRAM_ENABLED", "line_number": 46, "usage_type": "attribute"}, {"api_name": "settings.PUSHOVER_ENABLED", "line_number": 48, "usage_type": "attribute"}, {"api_name": "impf.constructors.zulip_client", "line_number": 53, "usage_type": "call"}, {"api_name": "impf.constructors.zulip_send_payload", "line_number": 55, "usage_type": "call"}, {"api_name": "impf.constructors.zulip_client", "line_number": 63, "usage_type": "call"}, {"api_name": "impf.constructors.zulip_read_payload", "line_number": 65, "usage_type": "call"}, {"api_name": "settings.TELEGRAM_BOT_TOKEN", "line_number": 74, "usage_type": "attribute"}, {"api_name": "settings.TELEGRAM_BOT_CHATID", "line_number": 75, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 84, "usage_type": "call"}, {"api_name": "settings.PUSHOVER_APP_TOKEN", "line_number": 89, "usage_type": "attribute"}, {"api_name": "settings.PUSHOVER_USER_KEY", "line_number": 90, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 99, "usage_type": "call"}]} +{"seq_id": "590053400", "text": "\n# coding: utf-8\n\n# In[ ]:\n\n\n# pip install keras\n\n\n# In[2]:\n\n\n# load the data\nimport numpy as np\nimport seaborn as sns\nfrom keras.preprocessing.image import load_img, img_to_array\nimport matplotlib.pyplot as plt;\nimport os\nos.environ[\"TF_CPP_MIN_LOG_LEVEL\"]=\"3\"\n# size of the pic\npic_size = 48\n\n# base path\nbase_path = \"./images/\"\n\nplt.figure(0, figsize=(12,20))\ncpt = 0\n\nfor expression in os.listdir(base_path + \"train/\"):\n for i in range(5):\n cpt = cpt + 1\n plt.subplot(7,5,cpt)\n img = load_img(base_path + \"train/\" + expression + \"/\" + os.listdir(base_path + \"train/\" + expression)[i], target_size=(pic_size, pic_size))\n plt.imshow(img, cmap=\"gray\")\n\nplt.tight_layout()\nplt.show() \n\n\n# In[3]:\n\n\n# check the number of pictures\nfor expression in os.listdir(base_path + \"train\"):\n print(str(len(os.listdir(base_path + \"train/\" + expression))) + \" \" + expression + \" images\")\n\n\n# In[4]:\n\n\n# split data.\n# 4 convolutional layers\n# 2 fully connected layers\nfrom keras.preprocessing.image import ImageDataGenerator\n# number of images to feed into the NN for every batch\nbatch_size = 128\n\ndatagen_train = ImageDataGenerator()\ndatagen_validation = ImageDataGenerator()\n\ntrain_generator = datagen_train.flow_from_directory(base_path + \"train\",\n target_size=(pic_size,pic_size),\n color_mode=\"grayscale\",\n batch_size=batch_size,\n class_mode='categorical',\n shuffle=True)\n\nvalidation_generator = datagen_validation.flow_from_directory(base_path + \"validation\",\n target_size=(pic_size,pic_size),\n color_mode=\"grayscale\",\n batch_size=batch_size,\n class_mode='categorical',\n shuffle=False)\n\n\n# In[5]:\n\n\n# setup CNN\nfrom keras.layers import Dense, Input, Dropout, GlobalAveragePooling2D, Flatten, Conv2D, BatchNormalization, Activation, MaxPooling2D\nfrom keras.models import Model, Sequential\nfrom keras.optimizers import Adam\n\n# number of possible label values\nnb_classes = 7\n\n# Initialising the CNN\nmodel = Sequential()\n\n# 1 - Convolution\nmodel.add(Conv2D(64,(3,3), padding='same', input_shape=(48, 48,1)))\nmodel.add(BatchNormalization())\nmodel.add(Activation('relu'))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\nmodel.add(Dropout(0.25))\n\n# 2nd Convolution layer\nmodel.add(Conv2D(128,(5,5), padding='same'))\nmodel.add(BatchNormalization())\nmodel.add(Activation('relu'))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\nmodel.add(Dropout(0.25))\n\n# 3rd Convolution layer\nmodel.add(Conv2D(512,(3,3), padding='same'))\nmodel.add(BatchNormalization())\nmodel.add(Activation('relu'))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\nmodel.add(Dropout(0.25))\n\n# 4th Convolution layer\nmodel.add(Conv2D(512,(3,3), padding='same'))\nmodel.add(BatchNormalization())\nmodel.add(Activation('relu'))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\nmodel.add(Dropout(0.25))\n\n# Flattening\nmodel.add(Flatten())\n\n# Fully connected layer 1st layer\nmodel.add(Dense(256))\nmodel.add(BatchNormalization())\nmodel.add(Activation('relu'))\nmodel.add(Dropout(0.25))\n\n# Fully connected layer 2nd layer\nmodel.add(Dense(512))\nmodel.add(BatchNormalization())\nmodel.add(Activation('relu'))\nmodel.add(Dropout(0.25))\n\nmodel.add(Dense(nb_classes, activation='softmax'))\n\nopt = Adam(lr=0.0001)\nmodel.compile(optimizer=opt, loss='categorical_crossentropy', metrics=['accuracy'])\n\n\n# In[6]:\n\n\nget_ipython().run_cell_magic('time', '', '# train the model\\n\\n# number of epochs to train the NN\\nepochs = 10\\n\\nfrom keras.callbacks import ModelCheckpoint\\n\\ncheckpoint = ModelCheckpoint(\"model_weights.h5\", monitor=\\'val_acc\\', verbose=1, save_best_only=True, mode=\\'max\\')\\ncallbacks_list = [checkpoint]\\n\\nhistory = model.fit_generator(generator=train_generator,\\n steps_per_epoch=train_generator.n//train_generator.batch_size,\\n epochs=epochs,\\n validation_data = validation_generator,\\n validation_steps = validation_generator.n//validation_generator.batch_size,\\n callbacks=callbacks_list\\n )')\n\n\n# In[7]:\n\n\n# serialize model structure to JSON\nmodel_json = model.to_json()\nwith open(\"model.json\", \"w\") as json_file:\n json_file.write(model_json)\n\n\n# In[115]:\n\n\n# plot the evolution of Loss on the train and validation sets\n\nimport matplotlib.pyplot as plt\n\nxlabel = [\"\"+str(2 * x + 1) for x in range(10)]\nplt.figure(figsize=(10,10))\nplt.xticks(np.arange(0, epochs + 2, step=2) , tuple(xlabel))\nplt.title('Loss - Iteration Number (Optimizer = Adam)')\nplt.ylabel('Loss', fontsize=16)\nplt.xlabel('Iteration Number', fontsize = 16)\nplt.plot(history.history['loss'], label='Training Loss')\nplt.plot(history.history['val_loss'], label='Validation Loss')\nplt.legend(loc='upper right')\n\n\n# In[116]:\n\n\n# plot the evolution of Accuracy on the train and validation sets\nxlabel = [\"\"+str(2 * x + 1) for x in range(10)]\nplt.figure(figsize=(10,10))\nplt.xticks(np.arange(0, epochs + 2, step=2) , tuple(xlabel))\nprint(np.arange(2, epochs + 2, step=2))\nplt.title('Loss - Iteration Number (Optimizer = Adam)')\nplt.ylabel('Accuracy', fontsize=16)\nplt.xlabel('Iteration Number', fontsize = 16)\nplt.plot(history.history['accuracy'], label='Training Accuracy')\nprint(history.history['accuracy'])\nplt.plot(history.history['val_accuracy'], label='Validation Accuracy')\nplt.legend(loc='upper left')\nplt.show()\n\n\n# In[33]:\n\n\n# show the confusion matrix of our predictions\n\n# compute predictions\npredictions = model.predict_generator(generator=validation_generator)\ny_pred = [np.argmax(probas) for probas in predictions]\ny_test = validation_generator.classes\nclass_names = validation_generator.class_indices.keys()\n\nfrom sklearn.metrics import confusion_matrix\nimport itertools\n\ndef plot_confusion_matrix(cm, classes, title='Confusion matrix', cmap=plt.cm.Blues):\n cm = cm.astype('float') / cm.sum(axis=1)[:, np.newaxis]\n plt.figure(figsize=(10,10))\n plt.imshow(cm, interpolation='nearest', cmap=cmap)\n plt.title(title)\n plt.colorbar()\n tick_marks = np.arange(len(classes))\n plt.xticks(tick_marks, classes, rotation=45)\n plt.yticks(tick_marks, classes)\n\n fmt = '.2f'\n thresh = cm.max() / 2.\n for i, j in itertools.product(range(cm.shape[0]), range(cm.shape[1])):\n plt.text(j, i, format(cm[i, j], fmt),\n horizontalalignment=\"center\",\n color=\"white\" if cm[i, j] > thresh else \"black\")\n\n plt.ylabel('True label')\n plt.xlabel('Predicted label')\n plt.tight_layout()\n \n# compute confusion matrix\ncnf_matrix = confusion_matrix(y_test, y_pred)\nnp.set_printoptions(precision=2)\n\n# plot normalized confusion matrix\nplt.figure()\nplot_confusion_matrix(cnf_matrix, classes=class_names, title='Normalized confusion matrix')\nplt.show()\n\n\n# In[66]:\n\n\n# testing.\nfilenames=validation_generator.filenames\nclasses = [\"angry\", \"disgust\", \"fear\",\"happy\",\"neutral\",\"sad\",\"surprise\"]\nprint(\"file name is :\" , filenames[0])\nprint(\"number of files are :\", str(len(filenames)))\nprint(\"classes are :\", str(class_names))\nprint(\"first prediction is :\", str(predictions[0]))\nprint(\"length of y_test :\", str(len(y_test)))\nprint(\"length of y_pred :\", str(len(y_pred)))\nprint(\"first test class is :\", str(y_test[0]))\nprint(\"first pred class is :\", str(y_pred[0]))\n\n\n# In[76]:\n\n\n# get 5 true predictions of \"angry\" picture\n# get 5 false predictions of \"angry\" picture\ntrue_angry_path = []\nfalse_angry_path = []\nfalse_angry_class = []\ncount = 0\niterate = 0\nwhile count != 10:\n if y_test[iterate] == 0 and y_test[iterate] == y_pred[iterate] and len(true_angry_path) != 5:\n true_angry_path.append(filenames[iterate])\n count += 1\n else:\n if y_test[iterate] == 0 and y_test[iterate] != y_pred[iterate] and len(false_angry_path) != 5:\n false_angry_path.append(filenames[iterate])\n false_angry_class.append(y_pred[iterate])\n count += 1\n iterate += 1\n\nprint(true_angry_path)\nprint(false_angry_path)\nprint(false_angry_class)\n\n\n# In[80]:\n\n\n# plot 5 true predictions of \"angry\" picture\n# plot 5 false predictions of \"angry\" picture\n\nplt.figure(0, figsize=(15,5))\ncpt = 0\ntrue_count = 0\nfalse_count = 0\nfor i in range(2):\n for j in range(5):\n cpt = cpt + 1\n plt.subplot(2,5,cpt)\n plt.ylabel(\"True label: \" + \"anger\")\n if i == 0:\n img = load_img(base_path + \"validation/\" +true_angry_path[true_count].replace(\"\\\\\",\"/\"), target_size=(pic_size, pic_size))\n plt.xlabel(\"Pred label: \" + \"anger\")\n true_count += 1\n else:\n img = load_img(base_path + \"validation/\" +false_angry_path[false_count].replace(\"\\\\\",\"/\"), target_size=(pic_size, pic_size))\n plt.xlabel(\"Pred label:\" + classes[false_angry_class[false_count]])\n false_count += 1 \n\n plt.imshow(img, cmap=\"gray\")\n\nplt.tight_layout()\nplt.show() \n\n\n# In[88]:\n\n\n# get 5 true predictions of \"disgust\" picture\n# get 5 false predictions of \"disgust\" picture\ntrue_disgust_path = []\nfalse_disgust_path = []\nfalse_disgust_class = []\ncount = 0\niterate = 0\nwhile count != 10:\n if y_test[iterate] == 1 and y_test[iterate] == y_pred[iterate] and len(true_disgust_path) != 5:\n true_disgust_path.append(filenames[iterate])\n count += 1\n else:\n if y_test[iterate] == 1 and y_test[iterate] != y_pred[iterate] and len(false_disgust_path) != 5:\n false_disgust_path.append(filenames[iterate])\n false_disgust_class.append(y_pred[iterate])\n count += 1\n iterate += 1\n\nprint(true_disgust_path)\nprint(false_disgust_path)\nprint(false_disgust_class)\n\n\n# In[89]:\n\n\n# plot 5 true predictions of \"disgust\" picture\n# plot 5 false predictions of \"disgust\" picture\n\nplt.figure(0, figsize=(15,5))\ncpt = 0\ntrue_count = 0\nfalse_count = 0\nfor i in range(2):\n for j in range(5):\n cpt = cpt + 1\n plt.subplot(2,5,cpt)\n plt.ylabel(\"True label: \" + \"disgust\")\n if i == 0:\n img = load_img(base_path + \"validation/\" +true_disgust_path[true_count].replace(\"\\\\\",\"/\"), target_size=(pic_size, pic_size))\n plt.xlabel(\"Pred label: \" + \"disgust\")\n true_count += 1\n else:\n img = load_img(base_path + \"validation/\" +false_disgust_path[false_count].replace(\"\\\\\",\"/\"), target_size=(pic_size, pic_size))\n plt.xlabel(\"Pred label:\" + classes[false_disgust_class[false_count]])\n false_count += 1 \n\n plt.imshow(img, cmap=\"gray\")\n\nplt.tight_layout()\nplt.show() \n\n\n# In[91]:\n\n\n# get 5 true predictions of \"fear\" picture\n# get 5 false predictions of \"fear\" picture\ntrue_fear_path = []\nfalse_fear_path = []\nfalse_fear_class = []\ncount = 0\niterate = 0\nwhile count != 10:\n if y_test[iterate] == 2 and y_test[iterate] == y_pred[iterate] and len(true_fear_path) != 5:\n true_fear_path.append(filenames[iterate])\n count += 1\n else:\n if y_test[iterate] == 2 and y_test[iterate] != y_pred[iterate] and len(false_fear_path) != 5:\n false_fear_path.append(filenames[iterate])\n false_fear_class.append(y_pred[iterate])\n count += 1\n iterate += 1\n\nprint(true_fear_path)\nprint(false_fear_path)\nprint(false_fear_class)\n\n\n# In[92]:\n\n\n# plot 5 true predictions of \"fear\" picture\n# plot 5 false predictions of \"fear\" picture\n\nplt.figure(0, figsize=(15,5))\ncpt = 0\ntrue_count = 0\nfalse_count = 0\nfor i in range(2):\n for j in range(5):\n cpt = cpt + 1\n plt.subplot(2,5,cpt)\n plt.ylabel(\"True label: \" + \"fear\")\n if i == 0:\n img = load_img(base_path + \"validation/\" +true_fear_path[true_count].replace(\"\\\\\",\"/\"), target_size=(pic_size, pic_size))\n plt.xlabel(\"Pred label: \" + \"fear\")\n true_count += 1\n else:\n img = load_img(base_path + \"validation/\" +false_fear_path[false_count].replace(\"\\\\\",\"/\"), target_size=(pic_size, pic_size))\n plt.xlabel(\"Pred label:\" + classes[false_fear_class[false_count]])\n false_count += 1 \n\n plt.imshow(img, cmap=\"gray\")\n\nplt.tight_layout()\nplt.show() \n\n\n# In[94]:\n\n\n# get 5 true predictions of \"happy\" picture\n# get 5 false predictions of \"happy\" picture\ntrue_happy_path = []\nfalse_happy_path = []\nfalse_happy_class = []\ncount = 0\niterate = 0\nwhile count != 10:\n if y_test[iterate] == 3 and y_test[iterate] == y_pred[iterate] and len(true_happy_path) != 5:\n true_happy_path.append(filenames[iterate])\n count += 1\n else:\n if y_test[iterate] == 3 and y_test[iterate] != y_pred[iterate] and len(false_happy_path) != 5:\n false_happy_path.append(filenames[iterate])\n false_happy_class.append(y_pred[iterate])\n count += 1\n iterate += 1\n\nprint(true_happy_path)\nprint(false_happy_path)\nprint(false_happy_class)\n\n\n# In[95]:\n\n\n# plot 5 true predictions of \"happy\" picture\n# plot 5 false predictions of \"happy\" picture\n\nplt.figure(0, figsize=(15,5))\ncpt = 0\ntrue_count = 0\nfalse_count = 0\nfor i in range(2):\n for j in range(5):\n cpt = cpt + 1\n plt.subplot(2,5,cpt)\n plt.ylabel(\"True label: \" + \"happy\")\n if i == 0:\n img = load_img(base_path + \"validation/\" +true_happy_path[true_count].replace(\"\\\\\",\"/\"), target_size=(pic_size, pic_size))\n plt.xlabel(\"Pred label: \" + \"happy\")\n true_count += 1\n else:\n img = load_img(base_path + \"validation/\" +false_happy_path[false_count].replace(\"\\\\\",\"/\"), target_size=(pic_size, pic_size))\n plt.xlabel(\"Pred label:\" + classes[false_happy_class[false_count]])\n false_count += 1 \n\n plt.imshow(img, cmap=\"gray\")\n\nplt.tight_layout()\nplt.show() \n\n\n# In[97]:\n\n\n# get 5 true predictions of \"neutral\" picture\n# get 5 false predictions of \"neutral\" picture\ntrue_neutral_path = []\nfalse_neutral_path = []\nfalse_neutral_class = []\ncount = 0\niterate = 0\nwhile count != 10:\n if y_test[iterate] == 4 and y_test[iterate] == y_pred[iterate] and len(true_neutral_path) != 5:\n true_neutral_path.append(filenames[iterate])\n count += 1\n else:\n if y_test[iterate] == 4 and y_test[iterate] != y_pred[iterate] and len(false_neutral_path) != 5:\n false_neutral_path.append(filenames[iterate])\n false_neutral_class.append(y_pred[iterate])\n count += 1\n iterate += 1\n\nprint(true_neutral_path)\nprint(false_neutral_path)\nprint(false_neutral_class)\n\n\n# In[98]:\n\n\n# plot 5 true predictions of \"neutral\" picture\n# plot 5 false predictions of \"neutral\" picture\n\nplt.figure(0, figsize=(15,5))\ncpt = 0\ntrue_count = 0\nfalse_count = 0\nfor i in range(2):\n for j in range(5):\n cpt = cpt + 1\n plt.subplot(2,5,cpt)\n plt.ylabel(\"True label: \" + \"neutral\")\n if i == 0:\n img = load_img(base_path + \"validation/\" +true_neutral_path[true_count].replace(\"\\\\\",\"/\"), target_size=(pic_size, pic_size))\n plt.xlabel(\"Pred label: \" + \"neutral\")\n true_count += 1\n else:\n img = load_img(base_path + \"validation/\" +false_neutral_path[false_count].replace(\"\\\\\",\"/\"), target_size=(pic_size, pic_size))\n plt.xlabel(\"Pred label:\" + classes[false_neutral_class[false_count]])\n false_count += 1 \n\n plt.imshow(img, cmap=\"gray\")\n\nplt.tight_layout()\nplt.show() \n\n\n# In[100]:\n\n\n# get 5 true predictions of \"sad\" picture\n# get 5 false predictions of \"sad\" picture\ntrue_sad_path = []\nfalse_sad_path = []\nfalse_sad_class = []\ncount = 0\niterate = 0\nwhile count != 10:\n if y_test[iterate] == 5 and y_test[iterate] == y_pred[iterate] and len(true_sad_path) != 5:\n true_sad_path.append(filenames[iterate])\n count += 1\n else:\n if y_test[iterate] == 5 and y_test[iterate] != y_pred[iterate] and len(false_sad_path) != 5:\n false_sad_path.append(filenames[iterate])\n false_sad_class.append(y_pred[iterate])\n count += 1\n iterate += 1\n\nprint(true_sad_path)\nprint(false_sad_path)\nprint(false_sad_class)\n\n\n# In[101]:\n\n\n# plot 5 true predictions of \"sad\" picture\n# plot 5 false predictions of \"sad\" picture\n\nplt.figure(0, figsize=(15,5))\ncpt = 0\ntrue_count = 0\nfalse_count = 0\nfor i in range(2):\n for j in range(5):\n cpt = cpt + 1\n plt.subplot(2,5,cpt)\n plt.ylabel(\"True label: \" + \"sad\")\n if i == 0:\n img = load_img(base_path + \"validation/\" +true_sad_path[true_count].replace(\"\\\\\",\"/\"), target_size=(pic_size, pic_size))\n plt.xlabel(\"Pred label: \" + \"sad\")\n true_count += 1\n else:\n img = load_img(base_path + \"validation/\" +false_sad_path[false_count].replace(\"\\\\\",\"/\"), target_size=(pic_size, pic_size))\n plt.xlabel(\"Pred label:\" + classes[false_sad_class[false_count]])\n false_count += 1 \n\n plt.imshow(img, cmap=\"gray\")\n\nplt.tight_layout()\nplt.show() \n\n\n# In[103]:\n\n\n# get 5 true predictions of \"surprise\" picture\n# get 5 false predictions of \"surprise\" picture\ntrue_surprise_path = []\nfalse_surprise_path = []\nfalse_surprise_class = []\ncount = 0\niterate = 0\nwhile count != 10:\n if y_test[iterate] == 6 and y_test[iterate] == y_pred[iterate] and len(true_surprise_path) != 5:\n true_surprise_path.append(filenames[iterate])\n count += 1\n else:\n if y_test[iterate] == 6 and y_test[iterate] != y_pred[iterate] and len(false_surprise_path) != 5:\n false_surprise_path.append(filenames[iterate])\n false_surprise_class.append(y_pred[iterate])\n count += 1\n iterate += 1\n\nprint(true_surprise_path)\nprint(false_surprise_path)\nprint(false_surprise_class)\n\n\n# In[104]:\n\n\n# plot 5 true predictions of \"surprise\" picture\n# plot 5 false predictions of \"surprise\" picture\n\nplt.figure(0, figsize=(15,5))\ncpt = 0\ntrue_count = 0\nfalse_count = 0\nfor i in range(2):\n for j in range(5):\n cpt = cpt + 1\n plt.subplot(2,5,cpt)\n plt.ylabel(\"True label: \" + \"surprise\")\n if i == 0:\n img = load_img(base_path + \"validation/\" +true_surprise_path[true_count].replace(\"\\\\\",\"/\"), target_size=(pic_size, pic_size))\n plt.xlabel(\"Pred label: \" + \"surprise\")\n true_count += 1\n else:\n img = load_img(base_path + \"validation/\" +false_surprise_path[false_count].replace(\"\\\\\",\"/\"), target_size=(pic_size, pic_size))\n plt.xlabel(\"Pred label:\" + classes[false_surprise_class[false_count]])\n false_count += 1 \n\n plt.imshow(img, cmap=\"gray\")\n\nplt.tight_layout()\nplt.show() \n\n", "sub_path": "project.py", "file_name": "project.py", "file_ext": "py", "file_size_in_byte": 19022, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "76", "api": [{"api_name": "os.environ", "line_number": 19, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 33, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 44, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 45, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 58, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 59, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 88, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 91, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 92, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 93, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 94, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 95, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 98, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 99, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 100, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 101, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 102, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 105, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 106, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 107, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 108, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 109, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 112, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 113, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 114, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 115, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 116, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 119, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 122, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 123, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 124, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 125, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 128, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 129, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 130, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 131, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 133, "usage_type": "call"}, {"api_name": "keras.optimizers.Adam", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 162, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 162, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 163, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 163, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 163, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 164, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 165, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 166, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 166, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 167, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 167, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 168, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 168, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 169, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 169, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 177, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 177, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 178, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 179, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 180, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 181, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 181, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 182, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 185, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 185, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 186, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 186, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 187, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 187, "usage_type": "name"}, {"api_name": "numpy.argmax", "line_number": 197, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 204, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 204, "usage_type": "name"}, {"api_name": "numpy.newaxis", "line_number": 205, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 206, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 206, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 207, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 208, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 208, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 209, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 209, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 210, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 211, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 211, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 212, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 212, "usage_type": "name"}, {"api_name": "itertools.product", "line_number": 216, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.text", "line_number": 217, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 217, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 221, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 221, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 222, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 222, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 223, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 223, "usage_type": "name"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 226, "usage_type": "call"}, {"api_name": "numpy.set_printoptions", "line_number": 227, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 230, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 230, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 232, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 232, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 283, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 283, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 290, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 290, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 291, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 291, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 293, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 294, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 294, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 297, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 298, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 298, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 301, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 301, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 303, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 303, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 304, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 304, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 339, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 339, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 346, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 346, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 347, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 347, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 349, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 350, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 350, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 353, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 354, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 354, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 357, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 357, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 359, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 359, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 360, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 360, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 395, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 395, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 402, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 402, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 403, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 403, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 405, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 406, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 406, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 409, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 410, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 410, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 413, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 413, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 415, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 415, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 416, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 416, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 451, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 451, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 458, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 458, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 459, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 459, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 461, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 462, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 462, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 465, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 466, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 466, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 469, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 469, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 471, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 471, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 472, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 472, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 507, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 507, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 514, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 514, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 515, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 515, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 517, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 518, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 518, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 521, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 522, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 522, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 525, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 525, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 527, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 527, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 528, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 528, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 563, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 563, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 570, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 570, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 571, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 571, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 573, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 574, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 574, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 577, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 578, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 578, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 581, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 581, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 583, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 583, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 584, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 584, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 619, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 619, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 626, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 626, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 627, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 627, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 629, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 630, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 630, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 633, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 634, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 634, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 637, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 637, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 639, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 639, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 640, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 640, "usage_type": "name"}]} +{"seq_id": "565204414", "text": "#!/usr/bin/env python\n\nfrom __future__ import print_function\n\nimport os\nimport numpy as np\nnp.seterr(all=\"ignore\") # noqa E402\nimport functools\n\nfrom collections import defaultdict\n\nfrom lsst.daf.persistence.butler import Butler\nfrom lsst.pex.config import (Config, Field, ConfigField, ListField, DictField, ConfigDictField,\n ConfigurableField)\nfrom lsst.pipe.base import CmdLineTask, ArgumentParser, TaskRunner, TaskError\nfrom lsst.coadd.utils import TractDataIdContainer\nfrom lsst.afw.table.catalogMatches import matchesToCatalog\nfrom lsst.meas.astrom import AstrometryConfig\nfrom lsst.meas.extensions.astrometryNet import LoadAstrometryNetObjectsTask\nfrom lsst.pipe.tasks.colorterms import Colorterm, ColortermLibrary\n\nfrom lsst.meas.algorithms import LoadIndexedReferenceObjectsTask\n\nfrom .analysis import AnalysisConfig, Analysis\nfrom .utils import (Filenamer, Enforcer, MagDiff, MagDiffMatches, MagDiffCompare,\n AstrometryDiff, TraceSize, PsfTraceSizeDiff, TraceSizeCompare, PercentDiff,\n E1Resids, E2Resids, E1ResidsHsmRegauss, E2ResidsHsmRegauss, FootNpixDiffCompare,\n MagDiffErr, CentroidDiff, deconvMom,\n deconvMomStarGal, concatenateCatalogs, joinMatches, checkPatchOverlap,\n addColumnsToSchema, addApertureFluxesHSC, addFpPoint,\n addFootprintNPix, makeBadArray, addIntFloatOrStrColumn,\n calibrateCoaddSourceCatalog, backoutApCorr, matchJanskyToDn,\n fluxToPlotString, andCatalog, writeParquet, getRepoInfo, setAliasMaps)\nfrom .plotUtils import (CosmosLabeller, StarGalaxyLabeller, OverlapsStarGalaxyLabeller,\n MatchesStarGalaxyLabeller)\n\nimport lsst.afw.geom as afwGeom\nimport lsst.afw.image as afwImage\nimport lsst.afw.table as afwTable\n\n__all__ = [\"CoaddAnalysisConfig\", \"CoaddAnalysisRunner\", \"CoaddAnalysisTask\", \"CompareCoaddAnalysisConfig\",\n \"CompareCoaddAnalysisRunner\", \"CompareCoaddAnalysisTask\"]\n\n\nclass CoaddAnalysisConfig(Config):\n coaddName = Field(dtype=str, default=\"deep\", doc=\"Name for coadd\")\n matchRadius = Field(dtype=float, default=0.5, doc=\"Matching radius (arcseconds)\")\n matchOverlapRadius = Field(dtype=float, default=0.5, doc=\"Matching radius for overlaps (arcseconds)\")\n colorterms = ConfigField(dtype=ColortermLibrary,\n doc=(\"Library of color terms.\"\n \"\\nNote that the colorterms, if any, need to be loaded in a config \"\n \"override file. See obs_subaru/config/hsc/coaddAnalysis.py for an \"\n \"example. If the colorterms for the appropriate reference dataset are \"\n \"loaded, they will be applied. Otherwise, no colorterms will be applied \"\n \"to the reference catalog.\"))\n analysis = ConfigField(dtype=AnalysisConfig, doc=\"Analysis plotting options\")\n analysisMatches = ConfigField(dtype=AnalysisConfig, doc=\"Analysis plotting options for matches\")\n matchesMaxDistance = Field(dtype=float, default=0.15, doc=\"Maximum plotting distance for matches\")\n externalCatalogs = ConfigDictField(keytype=str, itemtype=AstrometryConfig, default={},\n doc=\"Additional external catalogs for matching\")\n refObjLoader = ConfigurableField(target=LoadIndexedReferenceObjectsTask, doc=\"Reference object loader\")\n doPlotMags = Field(dtype=bool, default=True, doc=\"Plot magnitudes?\")\n doPlotSizes = Field(dtype=bool, default=True, doc=\"Plot PSF sizes?\")\n doPlotCentroids = Field(dtype=bool, default=True, doc=\"Plot centroids?\")\n doApCorrs = Field(dtype=bool, default=True, doc=\"Plot aperture corrections?\")\n doBackoutApCorr = Field(dtype=bool, default=False, doc=\"Backout aperture corrections?\")\n doAddAperFluxHsc = Field(dtype=bool, default=False,\n doc=\"Add a field containing 12 pix circular aperture flux to HSC table?\")\n doPlotStarGalaxy = Field(dtype=bool, default=True, doc=\"Plot star/galaxy?\")\n doPlotOverlaps = Field(dtype=bool, default=True, doc=\"Plot overlaps?\")\n doPlotMatches = Field(dtype=bool, default=True, doc=\"Plot matches?\")\n doPlotCompareUnforced = Field(dtype=bool, default=True,\n doc=\"Plot difference between forced and unforced?\")\n doPlotQuiver = Field(dtype=bool, default=True, doc=\"Plot ellipticity residuals quiver plot?\")\n doPlotFootprintNpix = Field(dtype=bool, default=True, doc=\"Plot histogram of footprint nPix?\")\n doPlotInputCounts = Field(dtype=bool, default=True, doc=\"Make input counts plot?\")\n onlyReadStars = Field(dtype=bool, default=False, doc=\"Only read stars (to save memory)?\")\n toMilli = Field(dtype=bool, default=True, doc=\"Print stats in milli units (i.e. mas, mmag)?\")\n srcSchemaMap = DictField(keytype=str, itemtype=str, default=None, optional=True,\n doc=\"Mapping between different stack (e.g. HSC vs. LSST) schema names\")\n fluxToPlotList = ListField(dtype=str, default=[\"base_GaussianFlux\", \"ext_photometryKron_KronFlux\",\n \"modelfit_CModel\"],\n doc=\"List of fluxes to plot: mag(flux)-mag(base_PsfFlux) vs mag(fluxColumn)\")\n columnsToCopy = ListField(dtype=str,\n default=[\"calib_psf_used\", \"calib_psf_candidate\", \"detect_isPatchInner\",\n \"detect_isTractInner\", \"merge_peak_sky\", \"calib_psfUsed\",\n \"calib_psfCandidate\", ],\n doc=\"List of columns to copy from one source catalog to another.\")\n flagsToAlias = DictField(keytype=str, itemtype=str,\n default={\"calib_psf_used\": \"calib_psfUsed\",\n \"calib_psf_candidate\": \"calib_psfCandidate\",\n \"calib_astrometry_used\": \"calib_astrometryUsed\"},\n doc=(\"List of flags to alias to old, pre-RFC-498, names for backwards \"\n \"compatibility with old processings\"))\n doWriteParquetTables = Field(dtype=bool, default=True,\n doc=(\"Write out Parquet tables (for subsequent interactive analysis)?\"\n \"\\nNOTE: if True but fastparquet package is unavailable, a warning is \"\n \"issued and table writing is skipped.\"))\n writeParquetOnly = Field(dtype=bool, default=False,\n doc=\"Only write out Parquet tables (i.e. do not produce any plots)?\")\n\n def saveToStream(self, outfile, root=\"root\"):\n \"\"\"Required for loading colorterms from a Config outside the 'lsst' namespace\"\"\"\n print(\"import lsst.meas.photocal.colorterms\", file=outfile)\n return Config.saveToStream(self, outfile, root)\n\n def setDefaults(self):\n Config.setDefaults(self)\n # self.externalCatalogs = {\"sdss-dr9-fink-v5b\": astrom}\n self.analysisMatches.magThreshold = 21.0 # External catalogs like PS1 & SDSS used smaller telescopes\n self.refObjLoader.ref_dataset_name = \"ps1_pv3_3pi_20170110\"\n\n def validate(self):\n Config.validate(self)\n if self.writeParquetOnly and not self.doWriteParquetTables:\n raise ValueError(\"Cannot writeParquetOnly if doWriteParquetTables is False\")\n\n\nclass CoaddAnalysisRunner(TaskRunner):\n @staticmethod\n def getTargetList(parsedCmd, **kwargs):\n kwargs[\"cosmos\"] = parsedCmd.cosmos\n\n # Partition all inputs by tract,filter\n FilterRefsDict = functools.partial(defaultdict, list) # Dict for filter-->dataRefs\n # Make sure the actual input files requested exist (i.e. do not follow the parent chain)\n # First check for forced catalogs. Break out of datasets loop if forced catalogs were found,\n # otherwise continue search for existence of unforced catalogs\n for dataset in [\"forced_src\", \"meas\"]:\n tractFilterRefs = defaultdict(FilterRefsDict) # tract-->filter-->dataRefs\n for patchRef in sum(parsedCmd.id.refList, []):\n tract = patchRef.dataId[\"tract\"]\n filterName = patchRef.dataId[\"filter\"]\n inputDataFile = patchRef.get(\"deepCoadd_\" + dataset + \"_filename\")[0]\n if parsedCmd.input not in parsedCmd.output:\n inputDataFile = inputDataFile.replace(parsedCmd.output, parsedCmd.input)\n if os.path.exists(inputDataFile):\n tractFilterRefs[tract][filterName].append(patchRef)\n if tractFilterRefs:\n break\n\n if not tractFilterRefs:\n raise RuntimeError(\"No suitable datasets found.\")\n\n return [(tractFilterRefs[tract][filterName], kwargs) for tract in tractFilterRefs for\n filterName in tractFilterRefs[tract]]\n\n\nclass CoaddAnalysisTask(CmdLineTask):\n _DefaultName = \"coaddAnalysis\"\n ConfigClass = CoaddAnalysisConfig\n RunnerClass = CoaddAnalysisRunner\n AnalysisClass = Analysis\n outputDataset = \"plotCoadd\"\n\n @classmethod\n def _makeArgumentParser(cls):\n parser = ArgumentParser(name=cls._DefaultName)\n parser.add_argument(\"--cosmos\", default=None, help=\"Filename for Leauthaud Cosmos catalog\")\n parser.add_id_argument(\"--id\", \"deepCoadd_meas\",\n help=\"data ID, e.g. --id tract=12345 patch=1,2 filter=HSC-X\",\n ContainerClass=TractDataIdContainer)\n return parser\n\n def __init__(self, *args, **kwargs):\n CmdLineTask.__init__(self, *args, **kwargs)\n self.unitScale = 1000.0 if self.config.toMilli else 1.0\n\n def runDataRef(self, patchRefList, cosmos=None):\n haveForced = False # do forced datasets exits (may not for single band datasets)\n dataset = \"Coadd_forced_src\"\n # Explicit input file was checked in CoaddAnalysisRunner, so a check on datasetExists\n # is sufficient here (modulo the case where a forced dataset exists higher up the parent\n # tree than the specified input, but does not exist in the input directory as the former\n # will be found)\n if patchRefList[0].datasetExists(self.config.coaddName + dataset):\n haveForced = True\n if not haveForced:\n self.log.warn(\"No forced dataset exists for, e.g.,: {:} (only showing first dataId in \"\n \"patchRefList).\\nPlotting unforced results only.\".format(patchRefList[0].dataId))\n dataset = \"Coadd_meas\"\n if not patchRefList[0].datasetExists(self.config.coaddName + dataset):\n raise TaskError(\"No data exists in patRefList: %s\" %\n ([patchRef.dataId for patchRef in patchRefList]))\n patchList = [patchRef.dataId[\"patch\"] for patchRef in patchRefList]\n self.log.info(\"patchList size: {:d}\".format(len(patchList)))\n repoInfo = getRepoInfo(patchRefList[0], coaddName=self.config.coaddName, coaddDataset=dataset)\n filenamer = Filenamer(repoInfo.butler, self.outputDataset, repoInfo.dataId)\n if (self.config.doPlotMags or self.config.doPlotStarGalaxy or self.config.doPlotOverlaps or\n self.config.doPlotCompareUnforced or cosmos or self.config.externalCatalogs):\n if haveForced:\n forced = self.readCatalogs(patchRefList, self.config.coaddName + \"Coadd_forced_src\")\n forced = self.calibrateCatalogs(forced, wcs=repoInfo.wcs)\n unforced = self.readCatalogs(patchRefList, self.config.coaddName + \"Coadd_meas\")\n unforced = self.calibrateCatalogs(unforced, wcs=repoInfo.wcs)\n\n if haveForced:\n # copy over some fields from unforced to forced catalog\n forced = addColumnsToSchema(unforced, forced,\n [col for col in list(self.config.columnsToCopy) +\n list(self.config.analysis.flags) if\n col not in forced.schema and col in unforced.schema and\n not (repoInfo.hscRun and col == \"slot_Centroid_flag\")])\n # Add the reference band flags for forced photometry to forced catalog\n refBandCat = self.readCatalogs(patchRefList, self.config.coaddName + \"Coadd_ref\")\n if len(forced) != len(refBandCat):\n raise RuntimeError((\"Lengths of forced (N = {0:d}) and ref (N = {0:d}) cats don't match\").\n format(len(forced), len(refBandCat)))\n refBandList = list(s.field.getName() for s in refBandCat.schema if \"merge_measurement_\"\n in s.field.getName())\n forced = addColumnsToSchema(refBandCat, forced,\n [col for col in refBandList if col not in forced.schema and\n col in refBandCat.schema])\n\n # Set some aliases for differing schema naming conventions\n coaddList = [unforced, ]\n if haveForced:\n coaddList += [forced]\n aliasDictList = [self.config.flagsToAlias, ]\n if repoInfo.hscRun is not None and self.config.srcSchemaMap is not None:\n aliasDictList += [self.config.srcSchemaMap]\n for cat in coaddList:\n cat = setAliasMaps(cat, aliasDictList)\n\n forcedStr = \"forced\" if haveForced else \"unforced\"\n\n if self.config.doPlotFootprintNpix:\n unforced = addFootprintNPix(unforced, fromCat=unforced)\n if haveForced:\n forced = addFootprintNPix(forced, fromCat=unforced)\n\n # Must do the overlaps before purging the catalogs of non-primary sources\n if self.config.doPlotOverlaps:\n # Determine if any patches in the patchList actually overlap\n overlappingPatches = checkPatchOverlap(patchList, repoInfo.tractInfo)\n if not overlappingPatches:\n self.log.info(\"No overlapping patches...skipping overlap plots\")\n else:\n self.catLabel = \"nChild = 0\"\n if haveForced:\n forcedOverlaps = self.overlaps(forced)\n if forcedOverlaps:\n self.plotOverlaps(forcedOverlaps, filenamer, repoInfo.dataId, butler=repoInfo.butler,\n camera=repoInfo.camera, tractInfo=repoInfo.tractInfo,\n patchList=patchList, hscRun=repoInfo.hscRun,\n matchRadius=self.config.matchOverlapRadius, zpLabel=self.zpLabel,\n forcedStr=forcedStr, postFix=\"_forced\",\n fluxToPlotList=[\"modelfit_CModel\", ])\n self.log.info(\"Number of forced overlap objects matched = {:d}\".\n format(len(forcedOverlaps)))\n unforcedOverlaps = self.overlaps(unforced)\n if unforcedOverlaps:\n self.plotOverlaps(unforcedOverlaps, filenamer, repoInfo.dataId, butler=repoInfo.butler,\n camera=repoInfo.camera, tractInfo=repoInfo.tractInfo,\n patchList=patchList, hscRun=repoInfo.hscRun,\n matchRadius=self.config.matchOverlapRadius, zpLabel=self.zpLabel,\n forcedStr=\"unforced\", postFix=\"_unforced\",\n fluxToPlotList=[\"modelfit_CModel\", ])\n self.log.info(\"Number of unforced overlap objects matched = {:d}\".\n format(len(unforcedOverlaps)))\n\n # Set boolean array indicating sources deemed unsuitable for qa analyses\n self.catLabel = \"noDuplicates\"\n bad = makeBadArray(unforced, flagList=self.config.analysis.flags,\n onlyReadStars=self.config.onlyReadStars)\n if haveForced:\n bad |= makeBadArray(forced, flagList=self.config.analysis.flags,\n onlyReadStars=self.config.onlyReadStars)\n\n # Create and write parquet tables\n if self.config.doWriteParquetTables:\n tableFilenamer = Filenamer(repoInfo.butler, 'qaTableCoadd', repoInfo.dataId)\n if haveForced:\n writeParquet(forced, tableFilenamer(repoInfo.dataId, description='forced'), badArray=bad)\n writeParquet(unforced, tableFilenamer(repoInfo.dataId, description='unforced'), badArray=bad)\n if self.config.writeParquetOnly:\n self.log.info(\"Exiting after writing Parquet tables. No plots generated.\")\n return\n\n # Purge the catalogs of flagged sources\n unforced = unforced[~bad].copy(deep=True)\n if haveForced:\n forced = forced[~bad].copy(deep=True)\n else:\n forced = unforced\n self.zpLabel = self.zpLabel + \" \" + self.catLabel\n if haveForced:\n self.log.info(\"\\nNumber of sources in catalogs: unforced = {0:d} and forced = {1:d}\".format(\n len(unforced), len(forced)))\n else:\n self.log.info(\"\\nNumber of sources in catalog: unforced = {0:d}\".format(len(unforced)))\n\n flagsCat = unforced\n\n if self.config.doPlotFootprintNpix:\n self.plotFootprintHist(forced, filenamer(repoInfo.dataId, description=\"footNpix\", style=\"hist\"),\n repoInfo.dataId, butler=repoInfo.butler, camera=repoInfo.camera,\n tractInfo=repoInfo.tractInfo, patchList=patchList, hscRun=repoInfo.hscRun,\n zpLabel=self.zpLabel, flagsCat=flagsCat)\n self.plotFootprint(forced, filenamer, repoInfo.dataId, butler=repoInfo.butler,\n camera=repoInfo.camera, tractInfo=repoInfo.tractInfo, patchList=patchList,\n hscRun=repoInfo.hscRun, zpLabel=self.zpLabel, flagsCat=flagsCat)\n\n if self.config.doPlotQuiver:\n self.plotQuiver(unforced, filenamer(repoInfo.dataId, description=\"ellipResids\", style=\"quiver\"),\n dataId=repoInfo.dataId, butler=repoInfo.butler, camera=repoInfo.camera,\n tractInfo=repoInfo.tractInfo, patchList=patchList, hscRun=repoInfo.hscRun,\n zpLabel=self.zpLabel, forcedStr=\"unforced\", scale=2)\n\n if self.config.doPlotInputCounts:\n self.plotInputCounts(unforced, filenamer(repoInfo.dataId, description=\"inputCounts\",\n style=\"tract\"),\n dataId=repoInfo.dataId, butler=repoInfo.butler, tractInfo=repoInfo.tractInfo,\n patchList=patchList, camera=repoInfo.camera, hscRun=repoInfo.hscRun,\n forcedStr=\"unforced\", alpha=0.5, doPlotTractImage=True,\n doPlotPatchOutline=True, sizeFactor=5.0, maxDiamPix=1000)\n\n if self.config.doPlotMags:\n self.plotMags(unforced, filenamer, repoInfo.dataId, butler=repoInfo.butler,\n camera=repoInfo.camera, tractInfo=repoInfo.tractInfo, patchList=patchList,\n hscRun=repoInfo.hscRun, zpLabel=self.zpLabel, forcedStr=\"unforced\",\n postFix=\"_unforced\", flagsCat=flagsCat)\n if haveForced:\n self.plotMags(forced, filenamer, repoInfo.dataId, butler=repoInfo.butler,\n camera=repoInfo.camera, tractInfo=repoInfo.tractInfo, patchList=patchList,\n hscRun=repoInfo.hscRun, zpLabel=self.zpLabel, forcedStr=forcedStr,\n postFix=\"_forced\", flagsCat=flagsCat,\n highlightList=[(\"merge_measurement_\" + repoInfo.genericFilterName, 0,\n \"yellow\"), ])\n if self.config.doPlotStarGalaxy:\n if \"ext_shapeHSM_HsmSourceMoments_xx\" in unforced.schema:\n self.plotStarGal(unforced, filenamer, repoInfo.dataId, butler=repoInfo.butler,\n camera=repoInfo.camera, tractInfo=repoInfo.tractInfo, patchList=patchList,\n hscRun=repoInfo.hscRun, zpLabel=self.zpLabel, forcedStr=\"unforced\")\n else:\n self.log.warn(\"Cannot run plotStarGal: ext_shapeHSM_HsmSourceMoments_xx not in forced.schema\")\n\n if self.config.doPlotSizes:\n if all(ss in unforced.schema for ss in [\"base_SdssShape_psf_xx\", \"calib_psf_used\"]):\n self.plotSizes(unforced, filenamer, repoInfo.dataId, butler=repoInfo.butler,\n camera=repoInfo.camera, tractInfo=repoInfo.tractInfo, patchList=patchList,\n hscRun=repoInfo.hscRun, zpLabel=self.zpLabel, forcedStr=\"unforced\",\n postFix=\"_unforced\")\n else:\n self.log.warn(\"Cannot run plotSizes: base_SdssShape_psf_xx and/or calib_psf_used \"\n \"not in unforced.schema\")\n if haveForced:\n if all(ss in forced.schema for ss in [\"base_SdssShape_psf_xx\", \"calib_psf_used\"]):\n self.plotSizes(forced, filenamer, repoInfo.dataId, butler=repoInfo.butler,\n camera=repoInfo.camera, tractInfo=repoInfo.tractInfo, patchList=patchList,\n hscRun=repoInfo.hscRun, zpLabel=self.zpLabel, forcedStr=forcedStr)\n else:\n self.log.warn(\"Cannot run plotSizes: base_SdssShape_psf_xx and/or calib_psf_used \"\n \"not in forced.schema\")\n if cosmos:\n self.plotCosmos(forced, filenamer, cosmos, repoInfo.dataId)\n if self.config.doPlotCompareUnforced and haveForced:\n self.plotCompareUnforced(forced, unforced, filenamer, repoInfo.dataId, butler=repoInfo.butler,\n camera=repoInfo.camera, tractInfo=repoInfo.tractInfo,\n patchList=patchList, hscRun=repoInfo.hscRun,\n matchRadius=self.config.matchRadius, zpLabel=self.zpLabel)\n\n if self.config.doPlotMatches:\n if haveForced:\n matches = self.readSrcMatches(patchRefList, self.config.coaddName + \"Coadd_forced_src\",\n hscRun=repoInfo.hscRun, wcs=repoInfo.wcs,\n aliasDictList=aliasDictList)\n else:\n matches = self.readSrcMatches(patchRefList, self.config.coaddName + \"Coadd_meas\",\n hscRun=repoInfo.hscRun, wcs=repoInfo.wcs,\n aliasDictList=aliasDictList)\n self.plotMatches(matches, repoInfo.filterName, filenamer, repoInfo.dataId, butler=repoInfo.butler,\n camera=repoInfo.camera, tractInfo=repoInfo.tractInfo, patchList=patchList,\n hscRun=repoInfo.hscRun, zpLabel=self.zpLabel, forcedStr=forcedStr)\n\n for cat in self.config.externalCatalogs:\n with andCatalog(cat):\n matches = self.matchCatalog(forced, repoInfo.filterName, self.config.externalCatalogs[cat])\n self.plotMatches(matches, repoInfo.filterName, filenamer, repoInfo.dataId,\n butler=repoInfo.butler, camera=repoInfo.camera, tractInfo=repoInfo.tractInfo,\n patchList=patchList, hscRun=repoInfo.hscRun, zpLabel=self.zpLabel,\n forcedStr=forcedStr, matchRadius=self.config.matchRadius)\n\n def readCatalogs(self, patchRefList, dataset):\n \"\"\"Read in and concatenate catalogs of type dataset in lists of data references\n\n If self.config.doWriteParquetTables is True, before appending each catalog to a single\n list, an extra column indicating the patch is added to the catalog. This is useful for\n the subsequent interactive QA analysis.\n\n Parameters\n ----------\n patchRefList : `list` of `lsst.daf.persistence.butlerSubset.ButlerDataRef`\n A list of butler data references whose catalogs of dataset type are to be read in\n dataset : `str`\n Name of the catalog dataset to be read in\n\n Raises\n ------\n `TaskError`\n If no data is read in for the dataRefList\n\n Returns\n -------\n `list` of concatenated `lsst.afw.table.source.source.SourceCatalog`s\n \"\"\"\n catList = []\n for patchRef in patchRefList:\n if patchRef.datasetExists(dataset):\n cat = patchRef.get(dataset, immediate=True, flags=afwTable.SOURCE_IO_NO_HEAVY_FOOTPRINTS)\n if self.config.doWriteParquetTables:\n cat = addIntFloatOrStrColumn(cat, patchRef.dataId[\"patch\"], \"patchId\",\n \"Patch on which source was detected\")\n catList.append(cat)\n if not catList:\n raise TaskError(\"No catalogs read: %s\" % ([patchRef.dataId for patchRef in patchRefList]))\n return concatenateCatalogs(catList)\n\n def readSrcMatches(self, dataRefList, dataset, hscRun=None, wcs=None, aliasDictList=None):\n catList = []\n for dataRef in dataRefList:\n if not dataRef.datasetExists(dataset):\n self.log.info(\"Dataset does not exist: {0:r}, {1:s}\".format(dataRef.dataId, dataset))\n continue\n butler = dataRef.getButler()\n\n # Generate unnormalized match list (from normalized persisted one) with joinMatchListWithCatalog\n # (which requires a refObjLoader to be initialized).\n catalog = dataRef.get(dataset, immediate=True, flags=afwTable.SOURCE_IO_NO_FOOTPRINTS)\n # Set some aliases for differing schema naming conventions\n if aliasDictList is not None:\n catalog = setAliasMaps(catalog, aliasDictList)\n if dataset != \"deepCoadd_meas\" and any(ss not in catalog.schema\n for ss in self.config.columnsToCopy):\n unforced = dataRef.get(\"deepCoadd_meas\", immediate=True,\n flags=afwTable.SOURCE_IO_NO_FOOTPRINTS)\n # copy over some fields from unforced to forced catalog\n catalog = addColumnsToSchema(unforced, catalog,\n [col for col in list(self.config.columnsToCopy) +\n list(self.config.analysis.flags) if\n col not in catalog.schema and col in unforced.schema and\n not (hscRun and col == \"slot_Centroid_flag\")])\n if aliasDictList is not None:\n catalog = setAliasMaps(catalog, aliasDictList)\n\n # Set boolean array indicating sources deemed unsuitable for qa analyses\n bad = makeBadArray(catalog, flagList=self.config.analysis.flags,\n onlyReadStars=self.config.onlyReadStars)\n\n catalog = self.calibrateCatalogs(catalog, wcs=wcs)\n\n if dataset.startswith(\"deepCoadd_\"):\n packedMatches = butler.get(\"deepCoadd_measMatch\", dataRef.dataId)\n else:\n packedMatches = butler.get(dataset + \"Match\", dataRef.dataId)\n\n # Purge the match list of sources flagged in the catalog\n badIds = catalog[\"id\"][bad]\n badMatch = np.zeros(len(packedMatches), dtype=bool)\n for iMat, iMatch in enumerate(packedMatches):\n if iMatch[\"second\"] in badIds:\n badMatch[iMat] = True\n self.catLabel = \"noDuplicates\"\n self.zpLabel = self.zpLabel + \" \" + self.catLabel\n packedMatches = packedMatches[~badMatch].copy(deep=True)\n if not packedMatches:\n self.log.warn(\"No good matches for %s\" % (dataRef.dataId,))\n continue\n # The reference object loader grows the bbox by the config parameter pixelMargin. This\n # is set to 50 by default but is not reflected by the radius parameter set in the\n # metadata, so some matches may reside outside the circle searched within this radius\n # Thus, increase the radius set in the metadata fed into joinMatchListWithCatalog() to\n # accommodate.\n matchmeta = packedMatches.table.getMetadata()\n rad = matchmeta.getDouble(\"RADIUS\")\n matchmeta.setDouble(\"RADIUS\", rad*1.05, \"field radius in degrees, approximate, padded\")\n refObjLoader = self.config.refObjLoader.apply(butler=butler)\n matches = refObjLoader.joinMatchListWithCatalog(packedMatches, catalog)\n if not hasattr(matches[0].first, \"schema\"):\n raise RuntimeError(\"Unable to unpack matches. \"\n \"Do you have the correct astrometry_net_data setup?\")\n # LSST reads in a_net catalogs with flux in \"janskys\", so must convert back to DN\n matches = matchJanskyToDn(matches)\n if hscRun and self.config.doAddAperFluxHsc:\n addApertureFluxesHSC(matches, prefix=\"second_\")\n\n if not matches:\n self.log.warn(\"No matches for %s\" % (dataRef.dataId,))\n continue\n\n # Set the alias maps for the matches sources (i.e. the .second attribute schema for each match)\n if aliasDictList is not None:\n for mm in matches:\n mm.second = setAliasMaps(mm.second, aliasDictList)\n\n matchMeta = butler.get(dataset, dataRef.dataId,\n flags=afwTable.SOURCE_IO_NO_FOOTPRINTS).getTable().getMetadata()\n catalog = matchesToCatalog(matches, matchMeta)\n # Compute Focal Plane coordinates for each source if not already there\n if self.config.analysisMatches.doPlotFP:\n if \"src_base_FPPosition_x\" not in catalog.schema and \"src_focalplane_x\" not in catalog.schema:\n exp = butler.get(\"calexp\", dataRef.dataId)\n det = exp.getDetector()\n catalog = addFpPoint(det, catalog, prefix=\"src_\")\n # Optionally backout aperture corrections\n if self.config.doBackoutApCorr:\n catalog = backoutApCorr(catalog)\n # Set the alias maps for the matched catalog sources\n if aliasDictList is not None:\n catalog = setAliasMaps(catalog, aliasDictList, prefix=\"src_\")\n\n catList.append(catalog)\n\n if not catList:\n raise TaskError(\"No matches read: %s\" % ([dataRef.dataId for dataRef in dataRefList]))\n\n return concatenateCatalogs(catList)\n\n def calibrateCatalogs(self, catalog, wcs=None):\n self.zpLabel = \"common (\" + str(self.config.analysis.coaddZp) + \")\"\n # My persisted catalogs in lauren/LSST/DM-6816new all have nan for ra dec (see DM-9556)\n if np.all(np.isnan(catalog[\"coord_ra\"])):\n if wcs is None:\n self.log.warn(\"Bad ra, dec entries but can't update because wcs is None\")\n else:\n for src in catalog:\n src.updateCoord(wcs)\n # Optionally backout aperture corrections\n if self.config.doBackoutApCorr:\n catalog = backoutApCorr(catalog)\n calibrated = calibrateCoaddSourceCatalog(catalog, self.config.analysis.coaddZp)\n return calibrated\n\n def plotMags(self, catalog, filenamer, dataId, butler=None, camera=None, ccdList=None, tractInfo=None,\n patchList=None, hscRun=None, matchRadius=None, zpLabel=None, forcedStr=None,\n fluxToPlotList=None, postFix=\"\", flagsCat=None, highlightList=None):\n if fluxToPlotList is None:\n fluxToPlotList = self.config.fluxToPlotList\n unitStr = \"mmag\" if self.config.toMilli else \"mag\"\n enforcer = Enforcer(requireLess={\"star\": {\"stdev\": 0.02*self.unitScale}})\n for col in fluxToPlotList:\n if col + \"_instFlux\" in catalog.schema:\n shortName = \"mag_\" + col + postFix\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(catalog,\n MagDiff(col + \"_instFlux\", \"base_PsfFlux_instFlux\",\n unitScale=self.unitScale),\n \"Mag(%s) - PSFMag (%s)\" % (fluxToPlotString(col), unitStr),\n shortName, self.config.analysis,\n flags=[col + \"_flag\"], labeller=StarGalaxyLabeller(),\n flagsCat=flagsCat, unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr,\n highlightList=highlightList)\n\n def plotSizes(self, catalog, filenamer, dataId, butler=None, camera=None, ccdList=None, tractInfo=None,\n patchList=None, hscRun=None, matchRadius=None, zpLabel=None, forcedStr=None, postFix=\"\",\n flagsCat=None):\n enforcer = None\n unitStr = \" (milli)\" if self.config.toMilli else \"\"\n for col in [\"base_PsfFlux\", ]:\n if col + \"_instFlux\" in catalog.schema:\n compareCol = \"base_SdssShape\"\n # Set limits dynamically...can be very different visit-to-visit due to seeing differences\n # SDSS and HSM should be similar, so limits based on one should be valid for the other and\n # having the same scale eases comparisons between the two.\n traceSizeFunc = TraceSize(compareCol)\n\n # First do for calib_psf_used only.\n shortName = \"trace\" + postFix + \"_calib_psf_used\"\n psfUsed = catalog[catalog[\"calib_psf_used\"]].copy(deep=True)\n sdssTrace = traceSizeFunc(psfUsed)\n sdssTrace = sdssTrace[np.where(np.isfinite(sdssTrace))]\n traceMean = np.around(np.nanmean(sdssTrace), 2)\n traceStd = max(0.03, np.around(4.5*np.nanstd(sdssTrace), 2))\n qMin = traceMean - traceStd\n qMax = traceMean + traceStd\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(psfUsed, sdssTrace,\n (\" SdssShape Trace (calib_psf_used): \"\n \"$\\sqrt{0.5*(I_{xx}+I_{yy})}$ (pixels)\"),\n shortName, self.config.analysis, flags=[col + \"_flag\"],\n goodKeys=[\"calib_psf_used\"], qMin=qMin, qMax=qMax,\n labeller=StarGalaxyLabeller(), flagsCat=flagsCat,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n if \"ext_shapeHSM_HsmSourceMoments_xx\" in catalog.schema:\n shortName = \"hsmTrace\" + postFix + \"_calib_psf_used\"\n compareCol = \"ext_shapeHSM_HsmSourceMoments\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(psfUsed, TraceSize(compareCol),\n (\" HSM Trace (calib_psf_used): $\\sqrt{0.5*(I_{xx}+I_{yy})}$\"\n \" (pixels)\"), shortName, self.config.analysis, flags=[col + \"_flag\"],\n goodKeys=[\"calib_psf_used\"], qMin=qMin, qMax=qMax,\n labeller=StarGalaxyLabeller(), flagsCat=flagsCat,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer,\n butler=butler, camera=camera, ccdList=ccdList,\n tractInfo=tractInfo, patchList=patchList, hscRun=hscRun,\n matchRadius=matchRadius, zpLabel=zpLabel,\n forcedStr=forcedStr)\n\n # Now for all stars.\n shortName = \"trace\" + postFix\n starsOnly = catalog[catalog[\"base_ClassificationExtendedness_value\"] < 0.5].copy(deep=True)\n sdssTrace = traceSizeFunc(starsOnly)\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(starsOnly, sdssTrace,\n \" SdssShape Trace: $\\sqrt{0.5*(I_{xx}+I_{yy})}$ (pixels)\", shortName,\n self.config.analysis, flags=[col + \"_flag\"], qMin=qMin, qMax=qMax,\n labeller=StarGalaxyLabeller(), flagsCat=flagsCat,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n if \"ext_shapeHSM_HsmSourceMoments_xx\" in catalog.schema:\n shortName = \"hsmTrace\" + postFix\n compareCol = \"ext_shapeHSM_HsmSourceMoments\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(starsOnly, TraceSize(compareCol),\n \"HSM Trace: $\\sqrt{0.5*(I_{xx}+I_{yy})}$ (pixels)\", shortName,\n self.config.analysis, flags=[col + \"_flag\"], qMin=qMin, qMax=qMax,\n labeller=StarGalaxyLabeller(), flagsCat=flagsCat,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer,\n butler=butler, camera=camera, ccdList=ccdList,\n tractInfo=tractInfo, patchList=patchList, hscRun=hscRun,\n matchRadius=matchRadius, zpLabel=zpLabel,\n forcedStr=forcedStr)\n\n if col + \"_instFlux\" in catalog.schema:\n shortName = \"psfTraceDiff\" + postFix\n compareCol = \"base_SdssShape\"\n psfCompareCol = \"base_SdssShape_psf\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(catalog, PsfTraceSizeDiff(compareCol, psfCompareCol),\n \" SdssShape Trace % diff (psf_used - PSFmodel)\", shortName,\n self.config.analysis, flags=[col + \"_flag\"],\n goodKeys=[\"calib_psf_used\"], qMin=-3.0, qMax=3.0,\n labeller=StarGalaxyLabeller(), flagsCat=flagsCat,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n\n shortName = \"e1Resids\" + postFix\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(catalog, E1Resids(compareCol, psfCompareCol, unitScale=self.unitScale),\n \" SdssShape e1 resids (psf_used - PSFmodel)%s\" % unitStr, shortName,\n self.config.analysis, flags=[col + \"_flag\"], goodKeys=[\"calib_psf_used\"],\n qMin=-0.05, qMax=0.05, labeller=StarGalaxyLabeller(), flagsCat=flagsCat,\n unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n\n shortName = \"e2Resids\" + postFix\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(catalog, E2Resids(compareCol, psfCompareCol, unitScale=self.unitScale),\n \" SdssShape e2 resids (psf_used - PSFmodel)%s\" % unitStr, shortName,\n self.config.analysis, flags=[col + \"_flag\"], goodKeys=[\"calib_psf_used\"],\n qMin=-0.05, qMax=0.05, labeller=StarGalaxyLabeller(), flagsCat=flagsCat,\n unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n\n if \"ext_shapeHSM_HsmSourceMoments_xx\" in catalog.schema:\n shortName = \"psfHsmTraceDiff\" + postFix\n compareCol = \"ext_shapeHSM_HsmSourceMoments\"\n psfCompareCol = \"ext_shapeHSM_HsmPsfMoments\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(catalog, PsfTraceSizeDiff(compareCol, psfCompareCol),\n \"HSM Trace % diff (psf_used - PSFmodel)\", shortName,\n self.config.analysis, flags=[col + \"_flag\"],\n goodKeys=[\"calib_psf_used\"], qMin=-3.0, qMax=3.0,\n labeller=StarGalaxyLabeller(), flagsCat=flagsCat,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer,\n butler=butler, camera=camera, ccdList=ccdList,\n tractInfo=tractInfo, patchList=patchList, hscRun=hscRun,\n matchRadius=matchRadius, zpLabel=zpLabel,\n forcedStr=forcedStr)\n shortName = \"e1ResidsHsm\" + postFix\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(catalog, E1Resids(compareCol, psfCompareCol, unitScale=self.unitScale),\n \" HSM e1 resids (psf_used - PSFmodel)%s\" % unitStr, shortName,\n self.config.analysis, flags=[col + \"_flag\"],\n goodKeys=[\"calib_psf_used\"], qMin=-0.05, qMax=0.05,\n labeller=StarGalaxyLabeller(), flagsCat=flagsCat,\n unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer,\n butler=butler, camera=camera, ccdList=ccdList,\n tractInfo=tractInfo, patchList=patchList, hscRun=hscRun,\n matchRadius=matchRadius, zpLabel=zpLabel,\n forcedStr=forcedStr)\n shortName = \"e2ResidsHsm\" + postFix\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(catalog, E2Resids(compareCol, psfCompareCol, unitScale=self.unitScale),\n \" HSM e2 resids (psf_used - PSFmodel)%s\" % unitStr, shortName,\n self.config.analysis, flags=[col + \"_flag\"],\n goodKeys=[\"calib_psf_used\"], qMin=-0.05, qMax=0.05,\n labeller=StarGalaxyLabeller(), flagsCat=flagsCat,\n unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer,\n butler=butler, camera=camera, ccdList=ccdList,\n tractInfo=tractInfo, patchList=patchList, hscRun=hscRun,\n matchRadius=matchRadius, zpLabel=zpLabel,\n forcedStr=forcedStr)\n\n shortName = \"e1ResidsHsmRegauss\" + postFix\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(catalog, E1ResidsHsmRegauss(unitScale=self.unitScale),\n \" HsmRegauss e1 resids (psf_used - HsmPsfMoments)%s\" % unitStr,\n shortName, self.config.analysis,\n flags=[col + \"_flag\", \"ext_shapeHSM_HsmShapeRegauss_flag\"],\n goodKeys=[\"calib_psf_used\"], qMin=-0.05, qMax=0.05,\n labeller=StarGalaxyLabeller(), flagsCat=flagsCat,\n unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer,\n butler=butler, camera=camera, ccdList=ccdList,\n tractInfo=tractInfo, patchList=patchList, hscRun=hscRun,\n matchRadius=matchRadius, zpLabel=zpLabel,\n forcedStr=forcedStr)\n\n shortName = \"e2ResidsHsmRegauss\" + postFix\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(catalog, E2ResidsHsmRegauss(unitScale=self.unitScale),\n \" HsmRegauss e2 resids (psf_used - HsmPsfMoments)%s\" % unitStr,\n shortName, self.config.analysis,\n flags=[col + \"_flag\", \"ext_shapeHSM_HsmShapeRegauss_flag\"],\n goodKeys=[\"calib_psf_used\"], qMin=-0.05, qMax=0.05,\n labeller=StarGalaxyLabeller(), flagsCat=flagsCat,\n unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer,\n butler=butler, camera=camera, ccdList=ccdList,\n tractInfo=tractInfo, patchList=patchList, hscRun=hscRun,\n matchRadius=matchRadius, zpLabel=zpLabel,\n forcedStr=forcedStr)\n\n def plotCentroidXY(self, catalog, filenamer, dataId, butler=None, camera=None, ccdList=None,\n tractInfo=None, patchList=None, hscRun=None, matchRadius=None, zpLabel=None,\n forcedStr=None, flagsCat=None):\n enforcer = None # Enforcer(requireLess={\"star\": {\"stdev\": 0.02*self.unitScale}})\n for col in [\"base_SdssCentroid_x\", \"base_SdssCentroid_y\"]:\n if col in catalog.schema:\n shortName = col\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(catalog, catalog[col], \"(%s)\" % col, shortName, self.config.analysis,\n labeller=StarGalaxyLabeller(), flagsCat=flagsCat,\n ).plotFP(dataId, filenamer, self.log, enforcer=enforcer, camera=camera,\n ccdList=ccdList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n\n def plotFootprint(self, catalog, filenamer, dataId, butler=None, camera=None, ccdList=None,\n tractInfo=None, patchList=None, hscRun=None, matchRadius=None, zpLabel=None,\n forcedStr=None, postFix=\"\", flagsCat=None, plotRunStats=False, highlightList=None):\n enforcer = None\n if \"calib_psf_used\" in catalog.schema:\n shortName = \"footNpix_calib_psf_used\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(catalog, catalog[\"base_Footprint_nPix\"], \"%s\" % shortName, shortName,\n self.config.analysis, flags=[\"base_Footprint_nPix_flag\"],\n goodKeys=[\"calib_psf_used\"], qMin=-100, qMax=2000,\n labeller=StarGalaxyLabeller(), flagsCat=flagsCat,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr, plotRunStats=plotRunStats,\n highlightList=highlightList)\n shortName = \"footNpix\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(catalog, catalog[\"base_Footprint_nPix\"], \"%s\" % shortName, shortName,\n self.config.analysis, flags=[\"base_Footprint_nPix_flag\"],\n qMin=0, qMax=3000, labeller=StarGalaxyLabeller(), flagsCat=flagsCat,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo, patchList=patchList,\n hscRun=hscRun, matchRadius=matchRadius, zpLabel=zpLabel,\n forcedStr=forcedStr, plotRunStats=plotRunStats)\n\n def plotFootprintHist(self, catalog, filenamer, dataId, butler=None, camera=None, ccdList=None,\n tractInfo=None, patchList=None, hscRun=None, matchRadius=None, zpLabel=None,\n postFix=\"\", flagsCat=None):\n stats = None\n shortName = \"footNpix\"\n self.log.info(\"shortName = {:s}\".format(shortName + \"Hist\"))\n self.AnalysisClass(catalog, catalog[\"base_Footprint_nPix\"], \"%s\" % shortName, shortName,\n self.config.analysis, flags=[\"base_Footprint_nPix_flag\"], qMin=0, qMax=3000,\n labeller=StarGalaxyLabeller(), flagsCat=flagsCat,\n ).plotHistogram(filenamer, stats=stats, camera=camera, hscRun=hscRun,\n matchRadius=matchRadius, zpLabel=zpLabel,\n filterStr=dataId['filter'])\n\n def plotStarGal(self, catalog, filenamer, dataId, butler=None, camera=None, ccdList=None, tractInfo=None,\n patchList=None, hscRun=None, matchRadius=None, zpLabel=None, forcedStr=None,\n flagsCat=None):\n enforcer = None\n shortName = \"pStar\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(catalog, deconvMomStarGal, \"P(star) from deconvolved moments\",\n shortName, self.config.analysis, qMin=-0.1, qMax=1.39,\n labeller=StarGalaxyLabeller(), flagsCat=flagsCat,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo, patchList=patchList,\n hscRun=hscRun, matchRadius=matchRadius, zpLabel=zpLabel,\n forcedStr=forcedStr)\n shortName = \"deconvMom\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(catalog, deconvMom, \"Deconvolved moments\", shortName,\n self.config.analysis, qMin=-1.0, qMax=3.0, labeller=StarGalaxyLabeller(),\n flagsCat=flagsCat,\n ).plotAll(dataId, filenamer, self.log,\n enforcer=Enforcer(requireLess={\"star\": {\"stdev\": 0.2}}),\n butler=butler, camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n\n if \"ext_shapeHSM_HsmShapeRegauss_resolution\" in catalog.schema:\n shortName = \"resolution\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(catalog, catalog[\"ext_shapeHSM_HsmShapeRegauss_resolution\"],\n \"Resolution Factor from HsmRegauss\",\n shortName, self.config.analysis, qMin=-0.1, qMax=1.15,\n labeller=StarGalaxyLabeller(), flagsCat=flagsCat,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n\n def plotCompareUnforced(self, forced, unforced, filenamer, dataId, butler=None, camera=None, ccdList=None,\n tractInfo=None, patchList=None, hscRun=None, matchRadius=None, zpLabel=None,\n fluxToPlotList=None):\n if fluxToPlotList is None:\n fluxToPlotList = self.config.fluxToPlotList\n unitStr = \"mmag\" if self.config.toMilli else \"mag\"\n enforcer = None\n catalog = joinMatches(afwTable.matchRaDec(forced, unforced, matchRadius*afwGeom.arcseconds),\n \"forced_\", \"unforced_\")\n for col in fluxToPlotList:\n shortName = \"compareUnforced_\" + col\n self.log.info(\"shortName = {:s}\".format(shortName))\n if \"forced_\" + col + \"_instFlux\" in catalog.schema:\n self.AnalysisClass(catalog, MagDiff(\"forced_\" + col + \"_instFlux\",\n \"unforced_\" + col + \"_instFlux\",\n unitScale=self.unitScale),\n \" Forced - Unforced mag [%s] (%s)\" % (fluxToPlotString(col), unitStr),\n shortName, self.config.analysis, prefix=\"forced_\", flags=[col + \"_flag\"],\n labeller=OverlapsStarGalaxyLabeller(\"forced_\", \"unforced_\"),\n unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel)\n\n def isBad(self, source):\n \"\"\"Return True if any of config.badFlags are set for this source.\"\"\"\n for flag in self.config.analysis.flags:\n if source.get(flag):\n return True\n return False\n\n def overlaps(self, catalog):\n badForOverlap = makeBadArray(catalog, flagList=self.config.analysis.flags,\n onlyReadStars=self.config.onlyReadStars, patchInnerOnly=False)\n goodCat = catalog[~badForOverlap]\n matches = afwTable.matchRaDec(goodCat, self.config.matchOverlapRadius*afwGeom.arcseconds)\n if not matches:\n self.log.info(\"Did not find any overlapping matches\")\n return joinMatches(matches, \"first_\", \"second_\")\n\n def plotOverlaps(self, overlaps, filenamer, dataId, butler=None, camera=None, ccdList=None,\n tractInfo=None, patchList=None, hscRun=None, matchRadius=None, zpLabel=None,\n forcedStr=None, postFix=\"\", fluxToPlotList=None, flagsCat=None):\n if fluxToPlotList is None:\n fluxToPlotList = self.config.fluxToPlotList\n unitStr = \"mmag\" if self.config.toMilli else \"mag\"\n magEnforcer = Enforcer(requireLess={\"star\": {\"stdev\": 0.003*self.unitScale}})\n for col in fluxToPlotList:\n shortName = \"overlap_\" + col + postFix\n self.log.info(\"shortName = {:s}\".format(shortName))\n if \"first_\" + col + \"_instFlux\" in overlaps.schema:\n self.AnalysisClass(overlaps, MagDiff(\"first_\" + col + \"_instFlux\",\n \"second_\" + col + \"_instFlux\",\n unitScale=self.unitScale),\n \" Overlap mag difference (%s) (%s)\" % (fluxToPlotString(col), unitStr),\n shortName, self.config.analysis, prefix=\"first_\", flags=[col + \"_flag\"],\n labeller=OverlapsStarGalaxyLabeller(), magThreshold=23,\n unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, enforcer=magEnforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n unitStr = \"mas\" if self.config.toMilli else \"arcsec\"\n distEnforcer = Enforcer(requireLess={\"star\": {\"stdev\": 0.005*self.unitScale}})\n shortName = \"overlap_distance\" + postFix\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(overlaps,\n lambda cat: cat[\"distance\"]*(1.0*afwGeom.radians).asArcseconds()*self.unitScale,\n \"Distance (%s)\" % unitStr, shortName, self.config.analysis, prefix=\"first_\",\n qMin=-0.01, qMax=0.11, labeller=OverlapsStarGalaxyLabeller(), flagsCat=flagsCat,\n forcedMean=0.0, unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, enforcer=distEnforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo, patchList=patchList,\n hscRun=hscRun, matchRadius=matchRadius, zpLabel=zpLabel,\n forcedStr=forcedStr)\n\n def plotMatches(self, matches, filterName, filenamer, dataId, description=\"matches\", butler=None,\n camera=None, ccdList=None, tractInfo=None, patchList=None, hscRun=None, matchRadius=None,\n zpLabel=None, forcedStr=None, flagsCat=None):\n unitStr = \"mmag\" if self.config.toMilli else \"mag\"\n enforcer = None # Enforcer(requireLess={\"star\": {\"stdev\": 0.030*self.unitScale}}),\n\n try:\n ct = self.config.colorterms.getColorterm(filterName, self.config.refObjLoader.ref_dataset_name)\n except Exception:\n # Pass in a null colorterm. Note the filterName must match for the source and reference catalogs\n ct = Colorterm(primary=filterName, secondary=filterName)\n self.log.warn(\"Note: no colorterms loaded for {:s}, thus no colorterms will be applied to \"\n \"the reference catalog\".format(self.config.refObjLoader.ref_dataset_name))\n if \"src_calib_psf_used\" in matches.schema:\n shortName = description + \"_mag_calib_psf_used\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(matches, MagDiffMatches(\"base_PsfFlux_instFlux\", ct, zp=0.0,\n unitScale=self.unitScale),\n \"MagPsf - ref (calib_psf_used) (%s)\" % unitStr, shortName,\n self.config.analysisMatches, prefix=\"src_\", goodKeys=[\"calib_psf_used\"],\n qMin=-0.15, qMax=0.1, labeller=MatchesStarGalaxyLabeller(), flagsCat=flagsCat,\n unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n if \"src_calib_photometry_used\" in matches.schema:\n shortName = description + \"_mag_calib_photometry_used\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(matches, MagDiffMatches(\"base_PsfFlux_instFlux\", ct, zp=0.0,\n unitScale=self.unitScale),\n \" MagPsf - ref (calib_photom_used) (%s)\" % unitStr, shortName,\n self.config.analysisMatches, prefix=\"src_\", goodKeys=[\"calib_photometry_used\"],\n qMin=-0.15, qMax=0.15, labeller=MatchesStarGalaxyLabeller(), flagsCat=flagsCat,\n unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n shortName = description + \"_mag\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(matches, MagDiffMatches(\"base_PsfFlux_instFlux\", ct, zp=0.0,\n unitScale=self.unitScale),\n \"MagPsf - ref (%s)\" % unitStr, shortName, self.config.analysisMatches,\n prefix=\"src_\", qMin=-0.15, qMax=0.5, labeller=MatchesStarGalaxyLabeller(),\n unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo, patchList=patchList,\n hscRun=hscRun, matchRadius=matchRadius, zpLabel=zpLabel,\n forcedStr=forcedStr)\n\n unitStr = \"mas\" if self.config.toMilli else \"arcsec\"\n if \"src_calib_astrometry_used\" in matches.schema:\n shortName = description + \"_distance_calib_astrometry_used\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(matches,\n lambda cat:\n cat[\"distance\"]*(1.0*afwGeom.radians).asArcseconds()*self.unitScale,\n \"Distance (%s) (calib_astrom_used)\" % unitStr, shortName,\n self.config.analysisMatches, prefix=\"src_\", goodKeys=[\"calib_astrometry_used\"],\n qMin=-0.01*self.config.matchRadius, qMax=0.5*self.config.matchRadius,\n labeller=MatchesStarGalaxyLabeller(), flagsCat=flagsCat,\n unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n shortName = description + \"_distance\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(matches,\n lambda cat: cat[\"distance\"]*(1.0*afwGeom.radians).asArcseconds()*self.unitScale,\n \"Distance (%s)\" % unitStr, shortName, self.config.analysisMatches, prefix=\"src_\",\n qMin=-0.05*self.config.matchRadius, qMax=0.3*self.config.matchRadius,\n labeller=MatchesStarGalaxyLabeller(), flagsCat=flagsCat, forcedMean=0.0,\n unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log,\n enforcer=Enforcer(requireLess={\"star\": {\"stdev\": 0.050*self.unitScale}}),\n butler=butler, camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n if \"src_calib_astrometry_used\" in matches.schema:\n shortName = description + \"_raCosDec_calib_astrometry_used\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(matches, AstrometryDiff(\"src_coord_ra\", \"ref_coord_ra\",\n declination1=\"src_coord_dec\",\n declination2=\"ref_coord_dec\",\n unitScale=self.unitScale),\n \" $\\delta_{Ra}$ = $\\Delta$RA*cos(Dec) (%s) (calib_astrom_used)\" % unitStr,\n shortName, self.config.analysisMatches, prefix=\"src_\",\n goodKeys=[\"calib_astrometry_used\"], qMin=-0.2*self.config.matchRadius,\n qMax=0.2*self.config.matchRadius, labeller=MatchesStarGalaxyLabeller(),\n flagsCat=flagsCat, unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n shortName = description + \"_raCosDec\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(matches, AstrometryDiff(\"src_coord_ra\", \"ref_coord_ra\",\n declination1=\"src_coord_dec\", declination2=\"ref_coord_dec\",\n unitScale=self.unitScale),\n \"$\\delta_{Ra}$ = $\\Delta$RA*cos(Dec) (%s)\" % unitStr, shortName,\n self.config.analysisMatches, prefix=\"src_\", qMin=-0.2*self.config.matchRadius,\n qMax=0.2*self.config.matchRadius, labeller=MatchesStarGalaxyLabeller(),\n flagsCat=flagsCat, unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log,\n enforcer=Enforcer(requireLess={\"star\": {\"stdev\": 0.050*self.unitScale}}),\n butler=butler, camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n if \"src_calib_astrometry_used\" in matches.schema:\n shortName = description + \"_ra_calib_astrometry_used\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(matches,\n AstrometryDiff(\"src_coord_ra\", \"ref_coord_ra\", unitScale=self.unitScale),\n \"$\\Delta$RA (%s) (calib_astrom_used)\" % unitStr, shortName,\n self.config.analysisMatches, prefix=\"src_\", goodKeys=[\"calib_astrometry_used\"],\n qMin=-0.25*self.config.matchRadius, qMax=0.25*self.config.matchRadius,\n labeller=MatchesStarGalaxyLabeller(), flagsCat=flagsCat,\n unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n shortName = description + \"_ra\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(matches, AstrometryDiff(\"src_coord_ra\", \"ref_coord_ra\", unitScale=self.unitScale),\n \"$\\Delta$RA (%s)\" % unitStr, shortName, self.config.analysisMatches,\n prefix=\"src_\", qMin=-0.25*self.config.matchRadius,\n qMax=0.25*self.config.matchRadius, labeller=MatchesStarGalaxyLabeller(),\n flagsCat=flagsCat, unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log,\n enforcer=Enforcer(requireLess={\"star\": {\"stdev\": 0.050*self.unitScale}}),\n butler=butler, camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n if \"src_calib_astrometry_used\" in matches.schema:\n shortName = description + \"_dec_calib_astrometry_used\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(matches,\n AstrometryDiff(\"src_coord_dec\", \"ref_coord_dec\", unitScale=self.unitScale),\n \"$\\delta_{Dec}$ (%s) (calib_astrom_used)\" % unitStr, shortName,\n self.config.analysisMatches, prefix=\"src_\", goodKeys=[\"calib_astrometry_used\"],\n qMin=-0.25*self.config.matchRadius, qMax=0.25*self.config.matchRadius,\n labeller=MatchesStarGalaxyLabeller(), flagsCat=flagsCat,\n unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n shortName = description + \"_dec\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(matches,\n AstrometryDiff(\"src_coord_dec\", \"ref_coord_dec\", unitScale=self.unitScale),\n \"$\\delta_{Dec}$ (%s)\" % unitStr, shortName, self.config.analysisMatches,\n prefix=\"src_\", qMin=-0.3*self.config.matchRadius, qMax=0.3*self.config.matchRadius,\n labeller=MatchesStarGalaxyLabeller(), flagsCat=flagsCat, unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log,\n enforcer=Enforcer(requireLess={\"star\": {\"stdev\": 0.050*self.unitScale}}),\n butler=butler, camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n\n def plotCosmos(self, catalog, filenamer, cosmos, dataId):\n labeller = CosmosLabeller(cosmos, self.config.matchRadius*afwGeom.arcseconds)\n self.AnalysisClass(catalog, deconvMom, \"Deconvolved moments\", \"cosmos\", self.config.analysis,\n qMin=-1.0, qMax=6.0, labeller=labeller,\n ).plotAll(dataId, filenamer, self.log,\n enforcer=Enforcer(requireLess={\"star\": {\"stdev\": 0.2}}))\n\n def matchCatalog(self, catalog, filterName, astrometryConfig):\n refObjLoader = LoadAstrometryNetObjectsTask(self.config.refObjLoaderConfig)\n center = afwGeom.averageSpherePoint([src.getCoord() for src in catalog])\n radius = max(center.separation(src.getCoord()) for src in catalog)\n filterName = afwImage.Filter(afwImage.Filter(filterName).getId()).getName() # Get primary name\n refs = refObjLoader.loadSkyCircle(center, radius, filterName).refCat\n matches = afwTable.matchRaDec(refs, catalog, self.config.matchRadius*afwGeom.arcseconds)\n matches = matchJanskyToDn(matches)\n return joinMatches(matches, \"ref_\", \"src_\")\n\n def plotQuiver(self, catalog, filenamer, dataId=None, butler=None, camera=None, ccdList=None,\n tractInfo=None, patchList=None, hscRun=None, matchRadius=None, zpLabel=None,\n forcedStr=None, postFix=\"\", flagsCat=None, scale=1):\n stats = None\n shortName = \"quiver\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(catalog, None, \"%s\" % shortName, shortName,\n self.config.analysis, labeller=None,\n ).plotQuiver(catalog, filenamer, self.log, stats=stats, dataId=dataId,\n butler=butler, camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, zpLabel=zpLabel,\n forcedStr=forcedStr, scale=scale)\n\n def plotInputCounts(self, catalog, filenamer, dataId, butler, tractInfo, patchList=None, camera=None,\n hscRun=None, forcedStr=None, alpha=0.5, doPlotTractImage=True,\n doPlotPatchOutline=True, sizeFactor=5.0, maxDiamPix=1000):\n shortName = \"inputCounts\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n self.AnalysisClass(catalog, None, \"%s\" % shortName, shortName,\n self.config.analysis, labeller=None,\n ).plotInputCounts(catalog, filenamer, self.log, dataId, butler, tractInfo,\n patchList=patchList, camera=camera, forcedStr=forcedStr,\n alpha=alpha, doPlotTractImage=doPlotTractImage,\n doPlotPatchOutline=doPlotPatchOutline,\n sizeFactor=sizeFactor, maxDiamPix=maxDiamPix)\n\n def _getConfigName(self):\n return None\n\n def _getMetadataName(self):\n return None\n\n def _getEupsVersionsName(self):\n return None\n\n\nclass CompareCoaddAnalysisConfig(CoaddAnalysisConfig):\n\n def setDefaults(self):\n CoaddAnalysisConfig.setDefaults(self)\n self.matchRadius = 0.2\n if \"base_PsfFlux\" not in self.fluxToPlotList:\n self.fluxToPlotList.append(\"base_PsfFlux\") # Add PSF flux to default list for comparison scripts\n\n\nclass CompareCoaddAnalysisRunner(TaskRunner):\n @staticmethod\n def getTargetList(parsedCmd, **kwargs):\n rootDir = parsedCmd.input.split(\"rerun\")[0] if len(parsedCmd.rerun) == 2 else parsedCmd.input\n butlerArgs = dict(root=os.path.join(rootDir, \"rerun\", parsedCmd.rerun2))\n if parsedCmd.calib is not None:\n butlerArgs[\"calibRoot\"] = parsedCmd.calib\n butler2 = Butler(**butlerArgs)\n idParser = parsedCmd.id.__class__(parsedCmd.id.level)\n idParser.idList = parsedCmd.id.idList\n butler = parsedCmd.butler\n parsedCmd.butler = butler2\n idParser.makeDataRefList(parsedCmd)\n parsedCmd.butler = butler\n\n return [(refList1, dict(patchRefList2=refList2, **kwargs)) for\n refList1, refList2 in zip(parsedCmd.id.refList, idParser.refList)]\n\n\nclass CompareCoaddAnalysisTask(CmdLineTask):\n ConfigClass = CompareCoaddAnalysisConfig\n RunnerClass = CompareCoaddAnalysisRunner\n _DefaultName = \"compareCoaddAnalysis\"\n\n @classmethod\n def _makeArgumentParser(cls):\n parser = ArgumentParser(name=cls._DefaultName)\n parser.add_argument(\"--rerun2\", required=True, help=\"Second rerun, for comparison\")\n parser.add_id_argument(\"--id\", \"deepCoadd_forced_src\",\n help=\"data ID, e.g. --id tract=12345 patch=1,2 filter=HSC-X\",\n ContainerClass=TractDataIdContainer)\n return parser\n\n def __init__(self, *args, **kwargs):\n CmdLineTask.__init__(self, *args, **kwargs)\n self.unitScale = 1000.0 if self.config.toMilli else 1.0\n\n def runDataRef(self, patchRefList1, patchRefList2):\n haveForced = True # do forced datasets exits (may not for single band datasets)\n dataset = \"Coadd_forced_src\"\n patchRefExistsList1 = [patchRef1 for patchRef1 in patchRefList1 if\n patchRef1.datasetExists(self.config.coaddName + dataset)]\n if not patchRefExistsList1:\n haveForced = False\n\n if not haveForced:\n self.log.warn(\"No forced dataset exist for tract: {0:d} filter: {1:s}. \"\n \"Plotting unforced results only.\".format(patchRefList1[0].dataId[\"tract\"],\n patchRefList1[0].dataId[\"filter\"]))\n dataset = \"Coadd_meas\"\n patchRefExistsList1 = [patchRef1 for patchRef1 in patchRefList1 if\n patchRef1.datasetExists(self.config.coaddName + dataset)]\n if not patchRefExistsList1:\n raise TaskError(\"No data exists in patRefList1: %s\" %\n ([patchRef1.dataId for patchRef1 in patchRefList1]))\n patchRefList2 = [dataRef2 for dataRef2 in patchRefList2 if\n dataRef2.datasetExists(self.config.coaddName + dataset)]\n\n patchList1 = [dataRef1.dataId[\"patch\"] for dataRef1 in patchRefList1 if\n dataRef1.datasetExists(self.config.coaddName + dataset)]\n patchRefList1 = patchRefExistsList1\n\n repoInfo1 = getRepoInfo(patchRefList1[0], coaddName=self.config.coaddName, coaddDataset=dataset)\n repoInfo2 = getRepoInfo(patchRefList2[0], coaddName=self.config.coaddName, coaddDataset=dataset)\n if haveForced:\n forced1 = self.readCatalogs(patchRefList1, self.config.coaddName + \"Coadd_forced_src\")\n forced1 = self.calibrateCatalogs(forced1, wcs=repoInfo1.wcs)\n forced2 = self.readCatalogs(patchRefList2, self.config.coaddName + \"Coadd_forced_src\")\n forced2 = self.calibrateCatalogs(forced2, wcs=repoInfo2.wcs)\n unforced1 = self.readCatalogs(patchRefList1, self.config.coaddName + \"Coadd_meas\")\n unforced1 = self.calibrateCatalogs(unforced1, wcs=repoInfo1.wcs)\n unforced2 = self.readCatalogs(patchRefList2, self.config.coaddName + \"Coadd_meas\")\n unforced2 = self.calibrateCatalogs(unforced2, wcs=repoInfo2.wcs)\n\n forcedStr = \"forced\" if haveForced else \"unforced\"\n\n if haveForced:\n # copy over some fields from unforced to forced catalog\n forced1 = addColumnsToSchema(unforced1, forced1,\n [col for col in list(self.config.columnsToCopy) +\n list(self.config.analysis.flags) if\n col not in forced1.schema and col in unforced1.schema and\n not (repoInfo1.hscRun and col == \"slot_Centroid_flag\")])\n forced2 = addColumnsToSchema(unforced2, forced2,\n [col for col in list(self.config.columnsToCopy) +\n list(self.config.analysis.flags) if\n col not in forced2.schema and col in unforced2.schema and\n not (repoInfo2.hscRun and col == \"slot_Centroid_flag\")])\n\n # Set an alias map for differing schema naming conventions of different stacks (if any)\n repoList = [repoInfo1.hscRun, repoInfo2.hscRun]\n coaddList = [unforced1, unforced2]\n if haveForced:\n repoList += repoList\n coaddList += [forced1, forced2]\n aliasDictList0 = [self.config.flagsToAlias, ]\n for hscRun, catalog in zip(repoList, coaddList):\n aliasDictList = aliasDictList0\n if hscRun is not None and self.config.srcSchemaMap is not None:\n aliasDictList += [self.config.srcSchemaMap]\n if aliasDictList is not None:\n catalog = setAliasMaps(catalog, aliasDictList)\n\n # Set boolean array indicating sources deemed unsuitable for qa analyses\n self.catLabel = \"noDuplicates\"\n bad1 = makeBadArray(unforced1, flagList=self.config.analysis.flags,\n onlyReadStars=self.config.onlyReadStars)\n bad2 = makeBadArray(unforced2, flagList=self.config.analysis.flags,\n onlyReadStars=self.config.onlyReadStars)\n if haveForced:\n bad1 |= makeBadArray(forced1, flagList=self.config.analysis.flags,\n onlyReadStars=self.config.onlyReadStars)\n bad2 |= makeBadArray(forced2, flagList=self.config.analysis.flags,\n onlyReadStars=self.config.onlyReadStars)\n\n # Purge the catalogs of flagged sources\n unforced1 = unforced1[~bad1].copy(deep=True)\n unforced2 = unforced2[~bad2].copy(deep=True)\n if haveForced:\n forced1 = forced1[~bad1].copy(deep=True)\n forced2 = forced2[~bad2].copy(deep=True)\n else:\n forced1 = unforced1\n forced2 = unforced2\n unforced = self.matchCatalogs(unforced1, unforced2)\n forced = self.matchCatalogs(forced1, forced2)\n\n aliasDictList = aliasDictList0\n if hscRun is not None and self.config.srcSchemaMap is not None:\n aliasDictList += [self.config.srcSchemaMap]\n if aliasDictList is not None:\n forced = setAliasMaps(forced, aliasDictList)\n unforced = setAliasMaps(unforced, aliasDictList)\n\n self.log.info(\"\\nNumber of sources in forced catalogs: first = {0:d} and second = {1:d}\".format(\n len(forced1), len(forced2)))\n\n filenamer = Filenamer(repoInfo1.butler, \"plotCompareCoadd\", repoInfo1.dataId)\n hscRun = repoInfo1.hscRun if repoInfo1.hscRun is not None else repoInfo2.hscRun\n if self.config.doPlotMags:\n self.plotMags(forced, filenamer, repoInfo1.dataId, butler=repoInfo1.butler,\n camera=repoInfo1.camera, tractInfo=repoInfo1.tractInfo, patchList=patchList1,\n hscRun=hscRun, matchRadius=self.config.matchRadius, zpLabel=self.zpLabel,\n forcedStr=forcedStr)\n\n if self.config.doPlotSizes:\n if (\"first_base_SdssShape_psf_xx\" in forced.schema and\n \"second_base_SdssShape_psf_xx\" in forced.schema):\n self.plotSizes(forced, filenamer, repoInfo1.dataId, butler=repoInfo1.butler,\n camera=repoInfo1.camera, tractInfo=repoInfo1.tractInfo, patchList=patchList1,\n hscRun=hscRun, matchRadius=self.config.matchRadius, zpLabel=self.zpLabel,\n forcedStr=forcedStr)\n else:\n self.log.warn(\"Cannot run plotSizes: base_SdssShape_psf_xx not in catalog.schema\")\n\n if self.config.doApCorrs:\n self.plotApCorrs(unforced, filenamer, repoInfo1.dataId, butler=repoInfo1.butler,\n camera=repoInfo1.camera, tractInfo=repoInfo1.tractInfo, patchList=patchList1,\n hscRun=hscRun, matchRadius=self.config.matchRadius, zpLabel=self.zpLabel,\n forcedStr=\"unforced\")\n if self.config.doPlotCentroids:\n self.plotCentroids(forced, filenamer, repoInfo1.dataId, butler=repoInfo1.butler,\n camera=repoInfo1.camera, tractInfo=repoInfo1.tractInfo, patchList=patchList1,\n hscRun=hscRun, hscRun1=repoInfo1.hscRun, hscRun2=repoInfo2.hscRun,\n matchRadius=self.config.matchRadius, zpLabel=self.zpLabel, forcedStr=forcedStr)\n if self.config.doPlotStarGalaxy:\n self.plotStarGal(forced, filenamer, repoInfo1.dataId, butler=repoInfo1.butler,\n camera=repoInfo1.camera, tractInfo=repoInfo1.tractInfo, patchList=patchList1,\n hscRun=hscRun, matchRadius=self.config.matchRadius, zpLabel=self.zpLabel,\n forcedStr=forcedStr)\n\n def readCatalogs(self, patchRefList, dataset):\n catList = [patchRef.get(dataset, immediate=True, flags=afwTable.SOURCE_IO_NO_FOOTPRINTS) for\n patchRef in patchRefList if patchRef.datasetExists(dataset)]\n if not catList:\n raise TaskError(\"No catalogs read: %s\" % ([patchRef.dataId for patchRef in patchRefList]))\n return concatenateCatalogs(catList)\n\n def matchCatalogs(self, catalog1, catalog2):\n matches = afwTable.matchRaDec(catalog1, catalog2, self.config.matchRadius*afwGeom.arcseconds)\n if not matches:\n raise TaskError(\"No matches found\")\n return joinMatches(matches, \"first_\", \"second_\")\n\n def calibrateCatalogs(self, catalog, wcs=None):\n self.zpLabel = \"common (\" + str(self.config.analysis.coaddZp) + \")\"\n # For some reason my persisted catalogs in lauren/LSST/DM-6816new all have nan for ra dec\n if np.all(np.isnan(catalog[\"coord_ra\"])):\n if wcs is None:\n self.log.warn(\"Bad ra, dec entries but can't update because wcs is None\")\n else:\n for src in catalog:\n src.updateCoord(wcs)\n calibrated = calibrateCoaddSourceCatalog(catalog, self.config.analysis.coaddZp)\n return calibrated\n\n def plotMags(self, catalog, filenamer, dataId, butler=None, camera=None, ccdList=None, tractInfo=None,\n patchList=None, hscRun=None, matchRadius=None, zpLabel=None, forcedStr=None,\n fluxToPlotList=None, postFix=\"\", flagsCat=None, highlightList=None):\n if fluxToPlotList is None:\n fluxToPlotList = self.config.fluxToPlotList\n unitStr = \"mmag\" if self.config.toMilli else \"mag\"\n if fluxToPlotList is None:\n fluxToPlotList = self.config.fluxToPlotList\n enforcer = None # Enforcer(requireLess={\"star\": {\"stdev\": 0.02*self.unitScale}})\n for col in fluxToPlotList:\n if (\"first_\" + col + \"_instFlux\" in catalog.schema and \"second_\" + col + \"_instFlux\" in\n catalog.schema):\n shortName = \"diff_\" + col + postFix\n self.log.info(\"shortName = {:s}\".format(shortName))\n Analysis(catalog, MagDiffCompare(col + \"_instFlux\", unitScale=self.unitScale),\n \" Run Comparison: %s mag diff (%s)\" % (fluxToPlotString(col), unitStr),\n shortName, self.config.analysis, prefix=\"first_\", qMin=-0.05, qMax=0.05,\n flags=[col + \"_flag\"], errFunc=MagDiffErr(col + \"_instFlux\",\n unitScale=self.unitScale),\n labeller=OverlapsStarGalaxyLabeller(), flagsCat=flagsCat, unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo, patchList=patchList,\n hscRun=hscRun, matchRadius=matchRadius, zpLabel=zpLabel,\n forcedStr=forcedStr, highlightList=highlightList)\n\n def plotCentroids(self, catalog, filenamer, dataId, butler=None, camera=None, ccdList=None,\n tractInfo=None, patchList=None, hscRun=None, hscRun1=None, hscRun2=None,\n matchRadius=None, zpLabel=None, forcedStr=None, flagsCat=None, highlightList=None):\n unitStr = \"milliPixels\" if self.config.toMilli else \"pixels\"\n distEnforcer = None\n centroidStr1, centroidStr2 = \"base_SdssCentroid\", \"base_SdssCentroid\"\n if bool(hscRun1) ^ bool(hscRun2):\n if hscRun1 is None:\n centroidStr1 = \"base_SdssCentroid_Rot\"\n if hscRun2 is None:\n centroidStr2 = \"base_SdssCentroid_Rot\"\n\n shortName = \"diff_x\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n Analysis(catalog, CentroidDiff(\"x\", centroid1=centroidStr1, centroid2=centroidStr2,\n unitScale=self.unitScale),\n \"Run Comparison: x offset (%s)\" % unitStr, shortName, self.config.analysis, prefix=\"first_\",\n qMin=-0.08, qMax=0.08, errFunc=None, labeller=OverlapsStarGalaxyLabeller(),\n ).plotAll(dataId, filenamer, self.log, enforcer=distEnforcer, butler=butler, camera=camera,\n ccdList=ccdList, tractInfo=tractInfo, patchList=patchList, hscRun=hscRun,\n matchRadius=matchRadius, zpLabel=zpLabel, forcedStr=forcedStr)\n shortName = \"diff_y\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n Analysis(catalog, CentroidDiff(\"y\", centroid1=centroidStr1, centroid2=centroidStr2,\n unitScale=self.unitScale),\n \"Run Comparison: y offset (%s)\" % unitStr, shortName, self.config.analysis, prefix=\"first_\",\n qMin=-0.08, qMax=0.08, errFunc=None, labeller=OverlapsStarGalaxyLabeller(),\n ).plotAll(dataId, filenamer, self.log, enforcer=distEnforcer, butler=butler, camera=camera,\n ccdList=ccdList, tractInfo=tractInfo, patchList=patchList, hscRun=hscRun,\n matchRadius=matchRadius, zpLabel=zpLabel, forcedStr=forcedStr)\n\n unitStr = \"mas\" if self.config.toMilli else \"arcsec\"\n shortName = \"diff_raCosDec\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n Analysis(catalog, AstrometryDiff(\"first_coord_ra\", \"second_coord_ra\", declination1=\"first_coord_dec\",\n declination2=\"second_coord_dec\", unitScale=self.unitScale),\n \" Run Comparison: $\\delta_{Ra}$ = $\\Delta$RA*cos(Dec) (%s)\" % unitStr, shortName,\n self.config.analysisMatches, prefix=\"first_\", qMin=-0.2*self.config.matchRadius,\n qMax=0.2*self.config.matchRadius, labeller=OverlapsStarGalaxyLabeller(),\n flagsCat=flagsCat, unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, butler=butler, camera=camera, ccdList=ccdList,\n tractInfo=tractInfo, patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n shortName = \"diff_ra\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n Analysis(catalog, AstrometryDiff(\"first_coord_ra\", \"second_coord_ra\", declination1=None,\n declination2=None, unitScale=self.unitScale),\n \"Run Comparison: $\\Delta$RA (%s)\" % unitStr, shortName, self.config.analysisMatches,\n prefix=\"first_\", qMin=-0.25*self.config.matchRadius, qMax=0.25*self.config.matchRadius,\n labeller=OverlapsStarGalaxyLabeller(), flagsCat=flagsCat, unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, butler=butler, camera=camera, ccdList=ccdList,\n tractInfo=tractInfo, patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n shortName = \"diff_dec\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n Analysis(catalog, AstrometryDiff(\"first_coord_dec\", \"second_coord_dec\", unitScale=self.unitScale),\n \"$\\delta_{Dec}$ (%s)\" % unitStr, shortName, self.config.analysisMatches, prefix=\"first_\",\n qMin=-0.3*self.config.matchRadius, qMax=0.3*self.config.matchRadius,\n labeller=OverlapsStarGalaxyLabeller(), flagsCat=flagsCat, unitScale=self.unitScale,\n ).plotAll(dataId, filenamer, self.log, butler=butler, camera=camera, ccdList=ccdList,\n tractInfo=tractInfo, patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n\n def plotFootprint(self, catalog, filenamer, dataId, butler=None, camera=None, ccdList=None,\n tractInfo=None, patchList=None, hscRun=None, matchRadius=None, zpLabel=None,\n forcedStr=None, postFix=\"\", flagsCat=None, highlightList=None):\n enforcer = None\n shortName = \"diff_footNpix\"\n col = \"base_Footprint_nPix\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n Analysis(catalog, FootNpixDiffCompare(col), \" Run Comparison: Footprint nPix difference\", shortName,\n self.config.analysis, prefix=\"first_\", qMin=-250, qMax=250, flags=[col + \"_flag\"],\n labeller=OverlapsStarGalaxyLabeller(), flagsCat=flagsCat,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler, camera=camera,\n ccdList=ccdList, tractInfo=tractInfo, patchList=patchList, hscRun=hscRun,\n matchRadius=matchRadius, zpLabel=zpLabel, forcedStr=forcedStr, postFix=postFix)\n shortName = \"diff_footNpix_calib_psf_used\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n Analysis(catalog, FootNpixDiffCompare(col), \" Run Comparison: Footprint nPix diff (psf_used)\",\n shortName, self.config.analysis, prefix=\"first_\", goodKeys=[\"calib_psf_used\"],\n qMin=-150, qMax=150, flags=[col + \"_flag\"], labeller=OverlapsStarGalaxyLabeller(),\n flagsCat=flagsCat,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler, camera=camera,\n ccdList=ccdList, tractInfo=tractInfo, patchList=patchList, hscRun=hscRun,\n matchRadius=matchRadius, zpLabel=zpLabel, forcedStr=forcedStr, postFix=postFix,\n highlightList=highlightList)\n\n def plotSizes(self, catalog, filenamer, dataId, butler=None, camera=None, ccdList=None, tractInfo=None,\n patchList=None, hscRun=None, matchRadius=None, zpLabel=None, forcedStr=None):\n enforcer = None # Enforcer(requireLess={\"star\": {\"stdev\": 0.02*self.unitScale}})\n for col in [\"base_PsfFlux\"]:\n if \"first_\" + col + \"_flux\" in catalog.schema and \"second_\" + col + \"_flux\" in catalog.schema:\n # Make comparison plots for all objects and calib_psf_used only objects\n for goodFlags in [[], [\"calib_psf_used\"]]:\n subCatString = \" (calib_psf_used)\" if \"calib_psf_used\" in goodFlags else \"\"\n shortNameBase = \"trace\"\n shortName = (shortNameBase + \"_calib_psf_used\" if \"calib_psf_used\" in goodFlags else\n shortNameBase)\n compareCol = \"base_SdssShape\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n Analysis(catalog, TraceSizeCompare(compareCol),\n \" SdssShape Trace Radius Diff (%)\" + subCatString,\n shortName, self.config.analysis, flags=[col + \"_flag\"], prefix=\"first_\",\n goodKeys=goodFlags, qMin=-0.5, qMax=1.5, labeller=OverlapsStarGalaxyLabeller(),\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n\n shortNameBase = \"psfTrace\"\n shortName = (shortNameBase + \"_calib_psf_used\" if \"calib_psf_used\" in goodFlags else\n shortNameBase)\n compareCol = \"base_SdssShape_psf\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n Analysis(catalog, TraceSizeCompare(compareCol),\n \" SdssShape PSF Trace Radius Diff (%)\" + subCatString,\n shortName, self.config.analysis, flags=[col + \"_flag\"], prefix=\"first_\",\n goodKeys=goodFlags, qMin=-1.1, qMax=1.1, labeller=OverlapsStarGalaxyLabeller(),\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n\n if \"first_ext_shapeHSM_HsmSourceMoments_xx\" in catalog.schema:\n shortNameBase = \"hsmTrace\"\n shortName = (shortNameBase + \"_calib_psf_used\" if \"calib_psf_used\" in goodFlags else\n shortNameBase)\n compareCol = \"ext_shapeHSM_HsmSourceMoments\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n Analysis(catalog, TraceSizeCompare(compareCol),\n \" HSM Trace Radius Diff (%)\" + subCatString, shortName,\n self.config.analysis, flags=[col + \"_flag\"], prefix=\"first_\",\n goodKeys=goodFlags, qMin=-0.5, qMax=1.5,\n labeller=OverlapsStarGalaxyLabeller(),\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n shortNameBase = \"hsmPsfTrace\"\n shortName = (shortNameBase + \"_calib_psf_used\" if \"calib_psf_used\" in goodFlags else\n shortNameBase)\n if \"first_ext_shapeHSM_PsfMoments_xx\" in catalog.schema:\n compareCol = \"ext_shapeHSM_HsmPsfMoments\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n Analysis(catalog, TraceSizeCompare(compareCol),\n \" HSM PSF Trace Radius Diff (%)\" + subCatString,\n shortName, self.config.analysis, flags=[col + \"_flag\"], prefix=\"first_\",\n goodKeys=goodFlags, qMin=-1.1, qMax=1.1,\n labeller=OverlapsStarGalaxyLabeller(),\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n\n shortName = \"sdssXx\"\n compareCol = \"base_SdssShape_xx\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n Analysis(catalog, PercentDiff(compareCol), \"SdssShape xx Moment Diff (%)\", shortName,\n self.config.analysis, flags=[col + \"_flag\"], prefix=\"first_\",\n qMin=-0.5, qMax=1.5, labeller=OverlapsStarGalaxyLabeller(),\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n shortName = \"sdssYy\"\n compareCol = \"base_SdssShape_yy\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n Analysis(catalog, PercentDiff(compareCol), \"SdssShape yy Moment Diff (%)\", shortName,\n self.config.analysis, flags=[col + \"_flag\"], prefix=\"first_\",\n qMin=-0.5, qMax=1.5, labeller=OverlapsStarGalaxyLabeller(),\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n\n def plotStarGal(self, catalog, filenamer, dataId, butler=None, camera=None, ccdList=None, tractInfo=None,\n patchList=None, hscRun=None, matchRadius=None, zpLabel=None, forcedStr=None,\n flagsCat=None):\n enforcer = None\n col = \"ext_shapeHSM_HsmShapeRegauss_resolution\"\n if \"first_\" + col in catalog.schema:\n shortName = \"diff_resolution\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n Analysis(catalog, PercentDiff(col),\n \" Run Comparison: HsmRegauss Resolution (% diff)\",\n shortName, self.config.analysis, prefix=\"first_\", qMin=-0.2, qMax=0.2,\n labeller=OverlapsStarGalaxyLabeller(), flagsCat=flagsCat,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n col = \"ext_shapeHSM_HsmShapeRegauss_e1\"\n if \"first_\" + col in catalog.schema:\n shortName = \"diff_HsmShapeRegauss_e1\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n Analysis(catalog, PercentDiff(col),\n \" Run Comparison: HsmRegauss e1 (% diff)\",\n shortName, self.config.analysis, prefix=\"first_\", qMin=-0.2, qMax=0.2,\n labeller=OverlapsStarGalaxyLabeller(), flagsCat=flagsCat,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n col = \"ext_shapeHSM_HsmShapeRegauss_e2\"\n if \"first_\" + col in catalog.schema:\n shortName = \"diff_HsmShapeRegauss_e2\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n Analysis(catalog, PercentDiff(col),\n \" Run Comparison: HsmRegauss e2 (% diff)\",\n shortName, self.config.analysis, prefix=\"first_\", qMin=-0.2, qMax=0.2,\n labeller=OverlapsStarGalaxyLabeller(), flagsCat=flagsCat,\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=zpLabel, forcedStr=forcedStr)\n\n def plotApCorrs(self, catalog, filenamer, dataId, butler=None, camera=None, ccdList=None,\n tractInfo=None, patchList=None, hscRun=None, matchRadius=None, zpLabel=None,\n forcedStr=None, fluxToPlotList=None):\n if fluxToPlotList is None:\n fluxToPlotList = self.config.fluxToPlotList\n enforcer = None # Enforcer(requireLess={\"star\": {\"stdev\": 0.02*self.unitScale}})\n for col in fluxToPlotList:\n if \"first_\" + col + \"_apCorr\" in catalog.schema and \"second_\" + col + \"_apCorr\" in catalog.schema:\n shortName = \"diff_\" + col + \"_apCorr\"\n self.log.info(\"shortName = {:s}\".format(shortName))\n # apCorrs in coadds can be all nan if they weren't run in sfm, so add a check for valid data\n # but here so we don't encounter the fatal error in Analysis\n if (len(np.where(np.isfinite(catalog[\"first_\" + col + \"_apCorr\"]))[0]) > 0 and\n len(np.where(np.isfinite(catalog[\"second_\" + col + \"_apCorr\"]))[0]) > 0):\n Analysis(catalog, MagDiffCompare(col + \"_apCorr\"),\n \" Run Comparison: %s apCorr diff\" % fluxToPlotString(col),\n shortName, self.config.analysis,\n prefix=\"first_\", qMin=-0.025, qMax=0.025, flags=[col + \"_flag_apCorr\"],\n labeller=OverlapsStarGalaxyLabeller(),\n ).plotAll(dataId, filenamer, self.log, enforcer=enforcer, butler=butler,\n camera=camera, ccdList=ccdList, tractInfo=tractInfo,\n patchList=patchList, hscRun=hscRun, matchRadius=matchRadius,\n zpLabel=None, forcedStr=forcedStr)\n else:\n self.log.warn(\"No valid data points for shortName = {:s}. Skipping...\".format(shortName))\n\n def _getConfigName(self):\n return None\n\n def _getMetadataName(self):\n return None\n\n def _getEupsVersionsName(self):\n return None\n", "sub_path": "python/lsst/pipe/analysis/coaddAnalysis.py", "file_name": "coaddAnalysis.py", "file_ext": "py", "file_size_in_byte": 108760, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "65", "api": [{"api_name": "numpy.seterr", "line_number": 7, "usage_type": "call"}, {"api_name": "lsst.pex.config.Config", "line_number": 45, "usage_type": "name"}, {"api_name": "lsst.pex.config.Field", "line_number": 46, "usage_type": "call"}, {"api_name": "lsst.pex.config.Field", "line_number": 47, "usage_type": "call"}, {"api_name": "lsst.pex.config.Field", "line_number": 48, "usage_type": "call"}, {"api_name": "lsst.pex.config.ConfigField", "line_number": 49, "usage_type": "call"}, {"api_name": "lsst.pipe.tasks.colorterms.ColortermLibrary", "line_number": 49, "usage_type": "name"}, {"api_name": "lsst.pex.config.ConfigField", "line_number": 56, "usage_type": "call"}, {"api_name": "analysis.AnalysisConfig", "line_number": 56, "usage_type": "name"}, {"api_name": "lsst.pex.config.ConfigField", "line_number": 57, "usage_type": "call"}, {"api_name": "analysis.AnalysisConfig", "line_number": 57, "usage_type": "name"}, {"api_name": "lsst.pex.config.Field", "line_number": 58, "usage_type": "call"}, {"api_name": "lsst.pex.config.ConfigDictField", "line_number": 59, "usage_type": "call"}, {"api_name": "lsst.meas.astrom.AstrometryConfig", "line_number": 59, "usage_type": "name"}, {"api_name": "lsst.pex.config.ConfigurableField", "line_number": 61, "usage_type": "call"}, {"api_name": "lsst.meas.algorithms.LoadIndexedReferenceObjectsTask", "line_number": 61, "usage_type": "name"}, {"api_name": "lsst.pex.config.Field", "line_number": 62, "usage_type": "call"}, {"api_name": "lsst.pex.config.Field", "line_number": 63, "usage_type": "call"}, {"api_name": "lsst.pex.config.Field", "line_number": 64, "usage_type": "call"}, {"api_name": "lsst.pex.config.Field", "line_number": 65, "usage_type": "call"}, {"api_name": "lsst.pex.config.Field", "line_number": 66, "usage_type": "call"}, {"api_name": "lsst.pex.config.Field", "line_number": 67, "usage_type": "call"}, {"api_name": "lsst.pex.config.Field", "line_number": 69, "usage_type": "call"}, {"api_name": "lsst.pex.config.Field", "line_number": 70, "usage_type": "call"}, {"api_name": "lsst.pex.config.Field", "line_number": 71, "usage_type": "call"}, {"api_name": "lsst.pex.config.Field", "line_number": 72, "usage_type": "call"}, {"api_name": "lsst.pex.config.Field", "line_number": 74, "usage_type": "call"}, {"api_name": "lsst.pex.config.Field", "line_number": 75, "usage_type": "call"}, {"api_name": "lsst.pex.config.Field", "line_number": 76, "usage_type": "call"}, {"api_name": "lsst.pex.config.Field", "line_number": 77, "usage_type": "call"}, {"api_name": "lsst.pex.config.Field", "line_number": 78, "usage_type": "call"}, {"api_name": "lsst.pex.config.DictField", "line_number": 79, "usage_type": "call"}, {"api_name": "lsst.pex.config.ListField", "line_number": 81, "usage_type": "call"}, {"api_name": "lsst.pex.config.ListField", "line_number": 84, "usage_type": "call"}, {"api_name": "lsst.pex.config.DictField", "line_number": 89, "usage_type": "call"}, {"api_name": "lsst.pex.config.Field", "line_number": 95, "usage_type": "call"}, {"api_name": "lsst.pex.config.Field", "line_number": 99, "usage_type": "call"}, {"api_name": "lsst.pex.config.Config.saveToStream", "line_number": 105, "usage_type": "call"}, {"api_name": "lsst.pex.config.Config", "line_number": 105, "usage_type": "name"}, {"api_name": "lsst.pex.config.Config.setDefaults", "line_number": 108, "usage_type": "call"}, {"api_name": "lsst.pex.config.Config", "line_number": 108, "usage_type": "name"}, {"api_name": "lsst.pex.config.Config.validate", "line_number": 114, "usage_type": "call"}, {"api_name": "lsst.pex.config.Config", "line_number": 114, "usage_type": "name"}, {"api_name": "lsst.pipe.base.TaskRunner", "line_number": 119, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 125, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 125, "usage_type": "argument"}, {"api_name": "collections.defaultdict", "line_number": 130, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 137, "usage_type": "call"}, {"api_name": "os.path", "line_number": 137, "usage_type": "attribute"}, {"api_name": "lsst.pipe.base.CmdLineTask", "line_number": 149, "usage_type": "name"}, {"api_name": "analysis.Analysis", "line_number": 153, "usage_type": "name"}, {"api_name": "lsst.pipe.base.ArgumentParser", "line_number": 158, "usage_type": "call"}, {"api_name": "lsst.coadd.utils.TractDataIdContainer", "line_number": 162, "usage_type": "name"}, {"api_name": "lsst.pipe.base.CmdLineTask.__init__", "line_number": 166, "usage_type": "call"}, {"api_name": "lsst.pipe.base.CmdLineTask", "line_number": 166, "usage_type": "name"}, {"api_name": "lsst.pipe.base.TaskError", "line_number": 183, "usage_type": "call"}, {"api_name": "utils.getRepoInfo", "line_number": 187, "usage_type": "call"}, {"api_name": "utils.Filenamer", "line_number": 188, "usage_type": "call"}, {"api_name": "utils.addColumnsToSchema", "line_number": 199, "usage_type": "call"}, {"api_name": "utils.addColumnsToSchema", "line_number": 211, "usage_type": "call"}, {"api_name": "utils.setAliasMaps", "line_number": 223, "usage_type": "call"}, {"api_name": "utils.addFootprintNPix", "line_number": 228, "usage_type": "call"}, {"api_name": "utils.addFootprintNPix", "line_number": 230, "usage_type": "call"}, {"api_name": "utils.checkPatchOverlap", "line_number": 235, "usage_type": "call"}, {"api_name": "utils.makeBadArray", "line_number": 264, "usage_type": "call"}, {"api_name": "utils.makeBadArray", "line_number": 267, "usage_type": "call"}, {"api_name": "utils.Filenamer", "line_number": 272, "usage_type": "call"}, {"api_name": "utils.writeParquet", "line_number": 274, "usage_type": "call"}, {"api_name": "utils.writeParquet", "line_number": 275, "usage_type": "call"}, {"api_name": "utils.andCatalog", "line_number": 377, "usage_type": "call"}, {"api_name": "lsst.afw.table.SOURCE_IO_NO_HEAVY_FOOTPRINTS", "line_number": 410, "usage_type": "attribute"}, {"api_name": "lsst.afw.table", "line_number": 410, "usage_type": "name"}, {"api_name": "utils.addIntFloatOrStrColumn", "line_number": 412, "usage_type": "call"}, {"api_name": "lsst.pipe.base.TaskError", "line_number": 416, "usage_type": "call"}, {"api_name": "utils.concatenateCatalogs", "line_number": 417, "usage_type": "call"}, {"api_name": "lsst.afw.table.SOURCE_IO_NO_FOOTPRINTS", "line_number": 429, "usage_type": "attribute"}, {"api_name": "lsst.afw.table", "line_number": 429, "usage_type": "name"}, {"api_name": "utils.setAliasMaps", "line_number": 432, "usage_type": "call"}, {"api_name": "lsst.afw.table.SOURCE_IO_NO_FOOTPRINTS", "line_number": 436, "usage_type": "attribute"}, {"api_name": "lsst.afw.table", "line_number": 436, "usage_type": "name"}, {"api_name": "utils.addColumnsToSchema", "line_number": 438, "usage_type": "call"}, {"api_name": "utils.setAliasMaps", "line_number": 444, "usage_type": "call"}, {"api_name": "utils.makeBadArray", "line_number": 447, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 459, "usage_type": "call"}, {"api_name": "utils.matchJanskyToDn", "line_number": 483, "usage_type": "call"}, {"api_name": "utils.addApertureFluxesHSC", "line_number": 485, "usage_type": "call"}, {"api_name": "utils.setAliasMaps", "line_number": 494, "usage_type": "call"}, {"api_name": "lsst.afw.table.SOURCE_IO_NO_FOOTPRINTS", "line_number": 497, "usage_type": "attribute"}, {"api_name": "lsst.afw.table", "line_number": 497, "usage_type": "name"}, {"api_name": "lsst.afw.table.catalogMatches.matchesToCatalog", "line_number": 498, "usage_type": "call"}, {"api_name": "utils.addFpPoint", "line_number": 504, "usage_type": "call"}, {"api_name": "utils.backoutApCorr", "line_number": 507, "usage_type": "call"}, {"api_name": "utils.setAliasMaps", "line_number": 510, "usage_type": "call"}, {"api_name": "lsst.pipe.base.TaskError", "line_number": 515, "usage_type": "call"}, {"api_name": "utils.concatenateCatalogs", "line_number": 517, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 522, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 522, "usage_type": "call"}, {"api_name": "utils.backoutApCorr", "line_number": 530, "usage_type": "call"}, {"api_name": "utils.calibrateCoaddSourceCatalog", "line_number": 531, "usage_type": "call"}, {"api_name": "utils.Enforcer", "line_number": 540, "usage_type": "call"}, {"api_name": "utils.MagDiff", "line_number": 546, "usage_type": "call"}, {"api_name": "utils.fluxToPlotString", "line_number": 548, "usage_type": "call"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 550, "usage_type": "call"}, {"api_name": "utils.TraceSize", "line_number": 569, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 575, "usage_type": "call"}, {"api_name": "numpy.isfinite", "line_number": 575, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 576, "usage_type": "call"}, {"api_name": "numpy.nanmean", "line_number": 576, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 577, "usage_type": "call"}, {"api_name": "numpy.nanstd", "line_number": 577, "usage_type": "call"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 586, "usage_type": "call"}, {"api_name": "utils.TraceSize", "line_number": 595, "usage_type": "call"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 599, "usage_type": "call"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 614, "usage_type": "call"}, {"api_name": "utils.TraceSize", "line_number": 623, "usage_type": "call"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 626, "usage_type": "call"}, {"api_name": "utils.PsfTraceSizeDiff", "line_number": 638, "usage_type": "call"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 642, "usage_type": "call"}, {"api_name": "utils.E1Resids", "line_number": 650, "usage_type": "call"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 653, "usage_type": "call"}, {"api_name": "utils.E2Resids", "line_number": 662, "usage_type": "call"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 665, "usage_type": "call"}, {"api_name": "utils.PsfTraceSizeDiff", "line_number": 677, "usage_type": "call"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 681, "usage_type": "call"}, {"api_name": "utils.E1Resids", "line_number": 689, "usage_type": "call"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 693, "usage_type": "call"}, {"api_name": "utils.E2Resids", "line_number": 702, "usage_type": "call"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 706, "usage_type": "call"}, {"api_name": "utils.E1ResidsHsmRegauss", "line_number": 716, "usage_type": "call"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 721, "usage_type": "call"}, {"api_name": "utils.E2ResidsHsmRegauss", "line_number": 731, "usage_type": "call"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 736, "usage_type": "call"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 753, "usage_type": "call"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 768, "usage_type": "call"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 778, "usage_type": "call"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 792, "usage_type": "call"}, {"api_name": "utils.deconvMomStarGal", "line_number": 803, "usage_type": "argument"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 805, "usage_type": "call"}, {"api_name": "utils.deconvMom", "line_number": 812, "usage_type": "argument"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 813, "usage_type": "call"}, {"api_name": "utils.Enforcer", "line_number": 816, "usage_type": "call"}, {"api_name": "plotUtils.StarGalaxyLabeller", "line_number": 827, "usage_type": "call"}, {"api_name": "utils.joinMatches", "line_number": 840, "usage_type": "call"}, {"api_name": "lsst.afw.table.matchRaDec", "line_number": 840, "usage_type": "call"}, {"api_name": "lsst.afw.table", "line_number": 840, "usage_type": "name"}, {"api_name": "lsst.afw.geom.arcseconds", "line_number": 840, "usage_type": "attribute"}, {"api_name": "lsst.afw.geom", "line_number": 840, "usage_type": "name"}, {"api_name": "utils.MagDiff", "line_number": 846, "usage_type": "call"}, {"api_name": "utils.fluxToPlotString", "line_number": 849, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 851, "usage_type": "call"}, {"api_name": "utils.makeBadArray", "line_number": 866, "usage_type": "call"}, {"api_name": "lsst.afw.table.matchRaDec", "line_number": 869, "usage_type": "call"}, {"api_name": "lsst.afw.table", "line_number": 869, "usage_type": "name"}, {"api_name": "lsst.afw.geom.arcseconds", "line_number": 869, "usage_type": "attribute"}, {"api_name": "lsst.afw.geom", "line_number": 869, "usage_type": "name"}, {"api_name": "utils.joinMatches", "line_number": 872, "usage_type": "call"}, {"api_name": "utils.Enforcer", "line_number": 880, "usage_type": "call"}, {"api_name": "utils.MagDiff", "line_number": 885, "usage_type": "call"}, {"api_name": "utils.fluxToPlotString", "line_number": 888, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 890, "usage_type": "call"}, {"api_name": "utils.Enforcer", "line_number": 897, "usage_type": "call"}, {"api_name": "lsst.afw.geom.radians", "line_number": 901, "usage_type": "attribute"}, {"api_name": "lsst.afw.geom", "line_number": 901, "usage_type": "name"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 903, "usage_type": "call"}, {"api_name": "lsst.pipe.tasks.colorterms.Colorterm", "line_number": 920, "usage_type": "call"}, {"api_name": "utils.MagDiffMatches", "line_number": 926, "usage_type": "call"}, {"api_name": "plotUtils.MatchesStarGalaxyLabeller", "line_number": 930, "usage_type": "call"}, {"api_name": "utils.MagDiffMatches", "line_number": 939, "usage_type": "call"}, {"api_name": "plotUtils.MatchesStarGalaxyLabeller", "line_number": 943, "usage_type": "call"}, {"api_name": "utils.MagDiffMatches", "line_number": 951, "usage_type": "call"}, {"api_name": "plotUtils.MatchesStarGalaxyLabeller", "line_number": 954, "usage_type": "call"}, {"api_name": "lsst.afw.geom.radians", "line_number": 967, "usage_type": "attribute"}, {"api_name": "lsst.afw.geom", "line_number": 967, "usage_type": "name"}, {"api_name": "plotUtils.MatchesStarGalaxyLabeller", "line_number": 971, "usage_type": "call"}, {"api_name": "lsst.afw.geom.radians", "line_number": 980, "usage_type": "attribute"}, {"api_name": "lsst.afw.geom", "line_number": 980, "usage_type": "name"}, {"api_name": "plotUtils.MatchesStarGalaxyLabeller", "line_number": 983, "usage_type": "call"}, {"api_name": "utils.Enforcer", "line_number": 986, "usage_type": "call"}, {"api_name": "utils.AstrometryDiff", "line_number": 993, "usage_type": "call"}, {"api_name": "plotUtils.MatchesStarGalaxyLabeller", "line_number": 1000, "usage_type": "call"}, {"api_name": "utils.AstrometryDiff", "line_number": 1008, "usage_type": "call"}, {"api_name": "plotUtils.MatchesStarGalaxyLabeller", "line_number": 1013, "usage_type": "call"}, {"api_name": "utils.Enforcer", "line_number": 1016, "usage_type": "call"}, {"api_name": "utils.AstrometryDiff", "line_number": 1024, "usage_type": "call"}, {"api_name": "plotUtils.MatchesStarGalaxyLabeller", "line_number": 1028, "usage_type": "call"}, {"api_name": "utils.AstrometryDiff", "line_number": 1036, "usage_type": "call"}, {"api_name": "plotUtils.MatchesStarGalaxyLabeller", "line_number": 1039, "usage_type": "call"}, {"api_name": "utils.Enforcer", "line_number": 1042, "usage_type": "call"}, {"api_name": "utils.AstrometryDiff", "line_number": 1050, "usage_type": "call"}, {"api_name": "plotUtils.MatchesStarGalaxyLabeller", "line_number": 1054, "usage_type": "call"}, {"api_name": "utils.AstrometryDiff", "line_number": 1063, "usage_type": "call"}, {"api_name": "plotUtils.MatchesStarGalaxyLabeller", "line_number": 1066, "usage_type": "call"}, {"api_name": "utils.Enforcer", "line_number": 1068, "usage_type": "call"}, {"api_name": "plotUtils.CosmosLabeller", "line_number": 1074, "usage_type": "call"}, {"api_name": "lsst.afw.geom.arcseconds", "line_number": 1074, "usage_type": "attribute"}, {"api_name": "lsst.afw.geom", "line_number": 1074, "usage_type": "name"}, {"api_name": "utils.deconvMom", "line_number": 1075, "usage_type": "argument"}, {"api_name": "utils.Enforcer", "line_number": 1078, "usage_type": "call"}, {"api_name": "lsst.meas.extensions.astrometryNet.LoadAstrometryNetObjectsTask", "line_number": 1081, "usage_type": "call"}, {"api_name": "lsst.afw.geom.averageSpherePoint", "line_number": 1082, "usage_type": "call"}, {"api_name": "lsst.afw.geom", "line_number": 1082, "usage_type": "name"}, {"api_name": "lsst.afw.image.Filter", "line_number": 1084, "usage_type": "call"}, {"api_name": "lsst.afw.image", "line_number": 1084, "usage_type": "name"}, {"api_name": "lsst.afw.table.matchRaDec", "line_number": 1086, "usage_type": "call"}, {"api_name": "lsst.afw.table", "line_number": 1086, "usage_type": "name"}, {"api_name": "lsst.afw.geom.arcseconds", "line_number": 1086, "usage_type": "attribute"}, {"api_name": "lsst.afw.geom", "line_number": 1086, "usage_type": "name"}, {"api_name": "utils.matchJanskyToDn", "line_number": 1087, "usage_type": "call"}, {"api_name": "utils.joinMatches", "line_number": 1088, "usage_type": "call"}, {"api_name": "lsst.pipe.base.TaskRunner", "line_number": 1135, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 1139, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1139, "usage_type": "attribute"}, {"api_name": "lsst.daf.persistence.butler.Butler", "line_number": 1142, "usage_type": "call"}, {"api_name": "lsst.pipe.base.CmdLineTask", "line_number": 1154, "usage_type": "name"}, {"api_name": "lsst.pipe.base.ArgumentParser", "line_number": 1161, "usage_type": "call"}, {"api_name": "lsst.coadd.utils.TractDataIdContainer", "line_number": 1165, "usage_type": "name"}, {"api_name": "lsst.pipe.base.CmdLineTask.__init__", "line_number": 1169, "usage_type": "call"}, {"api_name": "lsst.pipe.base.CmdLineTask", "line_number": 1169, "usage_type": "name"}, {"api_name": "lsst.pipe.base.TaskError", "line_number": 1188, "usage_type": "call"}, {"api_name": "utils.getRepoInfo", "line_number": 1197, "usage_type": "call"}, {"api_name": "utils.getRepoInfo", "line_number": 1198, "usage_type": "call"}, {"api_name": "utils.addColumnsToSchema", "line_number": 1213, "usage_type": "call"}, {"api_name": "utils.addColumnsToSchema", "line_number": 1218, "usage_type": "call"}, {"api_name": "utils.setAliasMaps", "line_number": 1236, "usage_type": "call"}, {"api_name": "utils.makeBadArray", "line_number": 1240, "usage_type": "call"}, {"api_name": "utils.makeBadArray", "line_number": 1242, "usage_type": "call"}, {"api_name": "utils.makeBadArray", "line_number": 1245, "usage_type": "call"}, {"api_name": "utils.makeBadArray", "line_number": 1247, "usage_type": "call"}, {"api_name": "utils.setAliasMaps", "line_number": 1266, "usage_type": "call"}, {"api_name": "utils.setAliasMaps", "line_number": 1267, "usage_type": "call"}, {"api_name": "utils.Filenamer", "line_number": 1272, "usage_type": "call"}, {"api_name": "lsst.afw.table.SOURCE_IO_NO_FOOTPRINTS", "line_number": 1307, "usage_type": "attribute"}, {"api_name": "lsst.afw.table", "line_number": 1307, "usage_type": "name"}, {"api_name": "lsst.pipe.base.TaskError", "line_number": 1310, "usage_type": "call"}, {"api_name": "utils.concatenateCatalogs", "line_number": 1311, "usage_type": "call"}, {"api_name": "lsst.afw.table.matchRaDec", "line_number": 1314, "usage_type": "call"}, {"api_name": "lsst.afw.table", "line_number": 1314, "usage_type": "name"}, {"api_name": "lsst.afw.geom.arcseconds", "line_number": 1314, "usage_type": "attribute"}, {"api_name": "lsst.afw.geom", "line_number": 1314, "usage_type": "name"}, {"api_name": "lsst.pipe.base.TaskError", "line_number": 1316, "usage_type": "call"}, {"api_name": "utils.joinMatches", "line_number": 1317, "usage_type": "call"}, {"api_name": "numpy.all", "line_number": 1322, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 1322, "usage_type": "call"}, {"api_name": "utils.calibrateCoaddSourceCatalog", "line_number": 1328, "usage_type": "call"}, {"api_name": "analysis.Analysis", "line_number": 1345, "usage_type": "call"}, {"api_name": "utils.MagDiffCompare", "line_number": 1345, "usage_type": "call"}, {"api_name": "utils.fluxToPlotString", "line_number": 1346, "usage_type": "call"}, {"api_name": "utils.MagDiffErr", "line_number": 1348, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 1350, "usage_type": "call"}, {"api_name": "analysis.Analysis", "line_number": 1370, "usage_type": "call"}, {"api_name": "utils.CentroidDiff", "line_number": 1370, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 1373, "usage_type": "call"}, {"api_name": "analysis.Analysis", "line_number": 1379, "usage_type": "call"}, {"api_name": "utils.CentroidDiff", "line_number": 1379, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 1382, "usage_type": "call"}, {"api_name": "analysis.Analysis", "line_number": 1390, "usage_type": "call"}, {"api_name": "utils.AstrometryDiff", "line_number": 1390, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 1394, "usage_type": "call"}, {"api_name": "analysis.Analysis", "line_number": 1401, "usage_type": "call"}, {"api_name": "utils.AstrometryDiff", "line_number": 1401, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 1405, "usage_type": "call"}, {"api_name": "analysis.Analysis", "line_number": 1411, "usage_type": "call"}, {"api_name": "utils.AstrometryDiff", "line_number": 1411, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 1414, "usage_type": "call"}, {"api_name": "analysis.Analysis", "line_number": 1426, "usage_type": "call"}, {"api_name": "utils.FootNpixDiffCompare", "line_number": 1426, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 1428, "usage_type": "call"}, {"api_name": "analysis.Analysis", "line_number": 1434, "usage_type": "call"}, {"api_name": "utils.FootNpixDiffCompare", "line_number": 1434, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 1436, "usage_type": "call"}, {"api_name": "analysis.Analysis", "line_number": 1456, "usage_type": "call"}, {"api_name": "utils.TraceSizeCompare", "line_number": 1456, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 1459, "usage_type": "call"}, {"api_name": "analysis.Analysis", "line_number": 1470, "usage_type": "call"}, {"api_name": "utils.TraceSizeCompare", "line_number": 1470, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 1473, "usage_type": "call"}, {"api_name": "analysis.Analysis", "line_number": 1485, "usage_type": "call"}, {"api_name": "utils.TraceSizeCompare", "line_number": 1485, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 1489, "usage_type": "call"}, {"api_name": "analysis.Analysis", "line_number": 1500, "usage_type": "call"}, {"api_name": "utils.TraceSizeCompare", "line_number": 1500, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 1504, "usage_type": "call"}, {"api_name": "analysis.Analysis", "line_number": 1513, "usage_type": "call"}, {"api_name": "utils.PercentDiff", "line_number": 1513, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 1515, "usage_type": "call"}, {"api_name": "analysis.Analysis", "line_number": 1523, "usage_type": "call"}, {"api_name": "utils.PercentDiff", "line_number": 1523, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 1525, "usage_type": "call"}, {"api_name": "analysis.Analysis", "line_number": 1539, "usage_type": "call"}, {"api_name": "utils.PercentDiff", "line_number": 1539, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 1542, "usage_type": "call"}, {"api_name": "analysis.Analysis", "line_number": 1551, "usage_type": "call"}, {"api_name": "utils.PercentDiff", "line_number": 1551, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 1554, "usage_type": "call"}, {"api_name": "analysis.Analysis", "line_number": 1563, "usage_type": "call"}, {"api_name": "utils.PercentDiff", "line_number": 1563, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 1566, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1584, "usage_type": "call"}, {"api_name": "numpy.isfinite", "line_number": 1584, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 1585, "usage_type": "call"}, {"api_name": "numpy.isfinite", "line_number": 1585, "usage_type": "call"}, {"api_name": "analysis.Analysis", "line_number": 1586, "usage_type": "call"}, {"api_name": "utils.MagDiffCompare", "line_number": 1586, "usage_type": "call"}, {"api_name": "utils.fluxToPlotString", "line_number": 1587, "usage_type": "call"}, {"api_name": "plotUtils.OverlapsStarGalaxyLabeller", "line_number": 1590, "usage_type": "call"}]}