diff --git "a/1088.jsonl" "b/1088.jsonl"
new file mode 100644--- /dev/null
+++ "b/1088.jsonl"
@@ -0,0 +1,462 @@
+{"seq_id": "463020699", "text": "class Solution:\n def minWindow(self, s:str, t:str) -> str:\n from collections import Counter\n if not t or not s:\n return ''\n dict_t = Counter(t)\n required = len(dict_t)\n l,r = 0, 0\n formed = 0\n window_counts = {}\n ans = float('inf'), None, None\n while r < len(s):\n character = s[r]\n window_counts[character] = window_counts.get(character, 0) + 1\n if character in dict_t and window_counts[character] == dict_t[character]:\n formed += 1\n while l <= r and formed == required:\n character = s[l]\n if r - l +1 < ans[0]:\n ans = (r-1+1, l, r)\n window_counts[character] -= 1\n if character in dict_t and window_counts[character] < dict_t[character]:\n formed -= 1\n\n l +=1\n r+=1\n return ''if ans[0] == float('inf') else s[ans[1]:ans[2] + 1]\n\nif __name__ == '__main__':\n solution = Solution()\n s = 'ab'\n t = 'b'\n solution.minWindow(s, t)\n\n", "sub_path": "sliding window/min_strs.py", "file_name": "min_strs.py", "file_ext": "py", "file_size_in_byte": 1103, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "collections.Counter", "line_number": 6, "usage_type": "call"}, {"api_name": "{'Counter': 'collections.Counter'}", "line_number": 30, "usage_type": "call"}]}
+{"seq_id": "109707999", "text": "#!usr/bin/env python3\n#\n# Author(s): Roman Rivera (Invisible Institute)\n\n'''export script for awards_1967-2017_2017-08_p061715'''\n\nimport pandas as pd\nimport __main__\nimport yaml\n\nimport setup\n\n\ndef get_setup():\n ''' encapsulates args.\n calls setup.do_setup() which returns constants and logger\n constants contains args and a few often-useful bits in it\n including constants.write_yamlvar()\n logger is used to write logging messages\n '''\n script_path = __main__.__file__\n args = {\n 'input_file': 'input/awards_1967-2017_2017-08.csv.gz',\n 'input_profiles_file': 'input/awards_1967-2017_2017-08_profiles.csv.gz',\n 'output_file': 'output/awards_1967-2017_2017-08.csv.gz',\n 'output_profiles_file': 'output/awards_1967-2017_2017-08_profiles.csv.gz',\n 'export_cols': [\n 'pps_award_detail_id', 'award_type', 'award_start_date',\n 'current_award_status', 'award_request_date',\n 'award_end_date', 'rank', 'last_promotion_date',\n 'requester_full_name', 'ceremony_date', 'tracking_no'\n ],\n 'id': 'awards_1967-2017_2017-08_ID'\n }\n\n assert (args['input_file'].startswith('input/') and\n args['input_file'].endswith('.csv.gz')),\\\n \"input_file is malformed: {}\".format(args['input_file'])\n assert (args['output_file'].startswith('output/') and\n args['output_file'].endswith('.csv.gz')),\\\n \"output_file is malformed: {}\".format(args['output_file'])\n\n return setup.do_setup(script_path, args)\n\n\ncons, log = get_setup()\n\ndf = pd.read_csv(cons.input_file)\n\nwith open(\"hand/award_po_ranks.yaml\", \"r\") as f:\n po_ranks = yaml.load(f)\nwith open(\"hand/maybe_po_ranks.yaml\", \"r\") as f:\n maybe_po_ranks = yaml.load(f)\n\npo_ids = df.loc[(df['rank'].isin(po_ranks)) |\n ((df['rank'].isin(maybe_po_ranks)) &\n (df['appointed_date'] < \"2010-01-01\")),\n cons.id].unique()\n\ndf_rows = df.shape[0]\ndf = df[['row_id', cons.id] + cons.export_cols]\ndf.to_csv(cons.output_file, **cons.csv_opts)\n\nprofiles_df = pd.read_csv(cons.input_profiles_file)\nprofiles_df.loc[profiles_df[cons.id].isin(po_ids), 'merge'] = 1\nprofiles_df['merge'] = profiles_df['merge'].fillna(0)\nlog.info('%d IDs with PO ranks marked for merging', len(po_ids))\nprofiles_df.to_csv(cons.output_profiles_file, **cons.csv_opts)\n", "sub_path": "get_data/utils/folder_structures/awards/export/src/export.py", "file_name": "export.py", "file_ext": "py", "file_size_in_byte": 2394, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "__main__.__file__", "line_number": 21, "usage_type": "attribute"}, {"api_name": "setup.do_setup", "line_number": 43, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 48, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 51, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 53, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 64, "usage_type": "call"}]}
+{"seq_id": "316795424", "text": "from telegram import Update, InlineKeyboardButton, InlineKeyboardMarkup, ReplyKeyboardMarkup\nfrom telegram.ext import CallbackContext, DispatcherHandlerStop\n\nfrom src import dynamo_db as db\nfrom src.communication.basics import send_markup_msg, edit_message\nfrom src.support.m17n import strings\n\n\ndef delete(update: Update, context: CallbackContext):\n _del_inline_keyboard_buttons = [\n [InlineKeyboardButton(strings()['button:yes'], callback_data='delete#yes')],\n [InlineKeyboardButton(strings()['button:no'], callback_data='delete#no')],\n ]\n\n _callback_query_options = [\n 'delete#yes',\n 'delete#no'\n ]\n\n chat_id = update.effective_chat.id\n if update.callback_query is None or update.callback_query.data not in _callback_query_options:\n reply_markup = InlineKeyboardMarkup(_del_inline_keyboard_buttons)\n send_markup_msg(update, strings()[\"delete:confirm\"], reply_markup, True)\n raise DispatcherHandlerStop\n else:\n if update.callback_query.data == 'delete#yes':\n deleted_user_entries = db.delete_user_entries(chat_id)\n deleted_user_info = db.delete_user_info(chat_id)\n if not deleted_user_entries and not deleted_user_info:\n update.effective_message.delete()\n send_markup_msg(update, strings()['delete:nothing'], ReplyKeyboardMarkup([['/start']],\n resize_keyboard=True,\n one_time_keyboard=True))\n else:\n update.effective_message.delete()\n send_markup_msg(update, strings()['delete:success'], ReplyKeyboardMarkup([['/start']],\n resize_keyboard=True,\n one_time_keyboard=True))\n raise DispatcherHandlerStop\n else:\n edit_message(update, strings()['delete:cancelled'])\n raise DispatcherHandlerStop\n", "sub_path": "bot_files/src/handlers/delete_all_data.py", "file_name": "delete_all_data.py", "file_ext": "py", "file_size_in_byte": 2164, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "telegram.Update", "line_number": 9, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 9, "usage_type": "name"}, {"api_name": "telegram.InlineKeyboardButton", "line_number": 11, "usage_type": "call"}, {"api_name": "src.support.m17n.strings", "line_number": 11, "usage_type": "call"}, {"api_name": "telegram.InlineKeyboardButton", "line_number": 12, "usage_type": "call"}, {"api_name": "src.support.m17n.strings", "line_number": 12, "usage_type": "call"}, {"api_name": "telegram.InlineKeyboardMarkup", "line_number": 22, "usage_type": "call"}, {"api_name": "src.communication.basics.send_markup_msg", "line_number": 23, "usage_type": "call"}, {"api_name": "src.support.m17n.strings", "line_number": 23, "usage_type": "call"}, {"api_name": "telegram.ext.DispatcherHandlerStop", "line_number": 24, "usage_type": "name"}, {"api_name": "src.dynamo_db.delete_user_entries", "line_number": 27, "usage_type": "call"}, {"api_name": "src.dynamo_db", "line_number": 27, "usage_type": "name"}, {"api_name": "src.dynamo_db.delete_user_info", "line_number": 28, "usage_type": "call"}, {"api_name": "src.dynamo_db", "line_number": 28, "usage_type": "name"}, {"api_name": "src.communication.basics.send_markup_msg", "line_number": 31, "usage_type": "call"}, {"api_name": "src.support.m17n.strings", "line_number": 31, "usage_type": "call"}, {"api_name": "telegram.ReplyKeyboardMarkup", "line_number": 31, "usage_type": "call"}, {"api_name": "src.communication.basics.send_markup_msg", "line_number": 36, "usage_type": "call"}, {"api_name": "src.support.m17n.strings", "line_number": 36, "usage_type": "call"}, {"api_name": "telegram.ReplyKeyboardMarkup", "line_number": 36, "usage_type": "call"}, {"api_name": "telegram.ext.DispatcherHandlerStop", "line_number": 39, "usage_type": "name"}, {"api_name": "src.communication.basics.edit_message", "line_number": 41, "usage_type": "call"}, {"api_name": "src.support.m17n.strings", "line_number": 41, "usage_type": "call"}, {"api_name": "telegram.ext.DispatcherHandlerStop", "line_number": 42, "usage_type": "name"}]}
+{"seq_id": "144147674", "text": "import sys, csv, scipy.stats\nfrom collections import defaultdict\n\ndef loadGeneLen(geneLenFile):\n counts = {}\n with open(geneLenFile) as f:\n for line in f:\n gene, l = line.strip().split('\\t')\n counts[gene] = int(l)\n return counts\n\ndef countGenes(countFile):\n counts = defaultdict(dict)\n with open(countFile) as f:\n f.readline()\n for line in f:\n gene, lof, lof_dbnsfp = line.strip().split('\\t')\n counts[gene]['lof'] = int(lof)\n counts[gene]['lof_dbnsfp'] = int(lof_dbnsfp)\n return counts\n\nif __name__ == '__main__':\n fgCountFile, bgCountFile, geneLenFile, outFile = sys.argv[1:]\n totFgSamples = 380\n totBgSamples = 54346\n\n fgGenes = countGenes(fgCountFile)\n bgGenes = countGenes(bgCountFile)\n\n geneLen = loadGeneLen(geneLenFile)\n\n with open(outFile, 'w') as fout:\n print >> fout, 'gene\\tvarType\\tfgCount\\tbgCount\\tpval'\n for gene in fgGenes:\n for varType in fgGenes[gene]:\n bgCount = 0\n if varType in bgGenes[gene]:\n bgCount = bgGenes[gene][varType]\n if gene in geneLen:\n fgSize = totFgSamples * geneLen[gene] - fgGenes[gene][varType]\n bgSize = totBgSamples * geneLen[gene] - bgCount\n (oratio, pval) = scipy.stats.fisher_exact([[fgGenes[gene][varType], fgSize], [bgCount, bgSize]])\n # pval = fisher.pvalue(fgGenes[gene][varType], fgSize,\n # bgGenes[gene][varType],\n # bgSize).right_tail\n else:\n pval = 'NA'\n \n\n ls = [str(x) for x in (gene, varType, fgGenes[gene][varType],\n bgCount, pval)]\n print >> fout, '\\t'.join(ls)\n \n \n", "sub_path": "code/enrich.py", "file_name": "enrich.py", "file_ext": "py", "file_size_in_byte": 1921, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "collections.defaultdict", "line_number": 13, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 23, "usage_type": "attribute"}, {"api_name": "scipy.stats.stats.fisher_exact", "line_number": 42, "usage_type": "call"}, {"api_name": "scipy.stats.stats", "line_number": 42, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 42, "usage_type": "name"}]}
+{"seq_id": "639665475", "text": "from datetime import datetime, timedelta\nimport datetime\nimport os\nfrom airflow import conf\nfrom airflow import DAG\nfrom airflow.operators.postgres_operator import PostgresOperator\nfrom airflow.operators.dummy_operator import DummyOperator\n\nfrom airflow.operators import (StageToRedshiftOperator, LoadFactOperator, LoadDimensionOperator, DataQualityOperator)\nfrom helpers import SqlQueries\n\n# Default args \ndefault_args = {\n 'owner': 'xingya-zhou',\n 'depends_on_past': False,\n 'start_date': datetime.datetime.now(),\n 'email_on_failure': False,\n 'email_on_retry': False,\n 'retries': 3,\n 'retry_delay': timedelta(minutes=5),\n 'catchup': False,\n 'retry_delay': timedelta(minutes=5)\n}\n\ndag = DAG(\n 'sparkify_dag',\n default_args = default_args,\n start_date = datetime.datetime.now()\n)\n\nf= open(os.path.join(conf.get('core','dags_folder'),'create_tables.sql'))\ncreate_tables_sql = f.read()\n\ncreate_trips_table = PostgresOperator(\n task_id=\"create_trips_table\",\n dag=dag,\n postgres_conn_id=\"redshift\",\n sql=create_tables_sql\n)\n\nstart_operator = DummyOperator(task_id='Begin_execution', dag=dag)\n\nstage_events_to_redshift = StageToRedshiftOperator(\n task_id='Stage_events',\n dag=dag, \n redshift_conn_id=\"redshift\",\n aws_credentials_id=\"aws_credentials\", \n table = \"staging_events\",\n s3_path = \"s3://udacity-dend/log_data\",\n json_path=\"s3://udacity-dend/log_json_path.json\"\n)\n\n\nstage_songs_to_redshift = StageToRedshiftOperator(\n task_id='Stage_songs',\n dag=dag, \n redshift_conn_id=\"redshift\",\n aws_credentials_id=\"aws_credentials\", \n table = \"staging_songs\",\n s3_path = \"s3://udacity-dend/song_data\",\n json_path=\"auto\"\n)\n\n\nload_songplays_table = LoadFactOperator(\n task_id='Load_songplays_fact_table',\n dag=dag, \n redshift_conn_id=\"redshift\",\n table=\"songplays\",\n sql=SqlQueries.songplay_table_insert,\n append_only=False\n)\n\nload_songs_table = LoadDimensionOperator(\n task_id='Load_songs_table',\n dag=dag, \n redshift_conn_id=\"redshift\",\n table=\"songs\",\n sql=SqlQueries.song_table_insert,\n append_only=False\n)\n\n\nload_users_table = LoadDimensionOperator(\n task_id='Load_users_table',\n dag=dag, \n redshift_conn_id=\"redshift\",\n table=\"users\",\n sql=SqlQueries.user_table_insert,\n append_only=False\n)\n\nload_artists_table = LoadDimensionOperator(\n task_id='Load_artists_table',\n dag=dag, \n redshift_conn_id=\"redshift\",\n table=\"artists\",\n sql=SqlQueries.artist_table_insert,\n append_only=False\n)\n\nload_time_table = LoadDimensionOperator(\n task_id='Load_time_table',\n dag=dag, \n redshift_conn_id=\"redshift\",\n table=\"time\",\n sql=SqlQueries.time_table_insert,\n append_only=False\n)\n\n\nrun_quality_checks = DataQualityOperator(\n task_id='Run_data_quality_checks',\n dag=dag,\n redshift_conn_id=\"redshift\",\n tables=[ \"songplays\", \"songs\", \"artists\", \"time\", \"users\"]\n)\n\nend_operator = DummyOperator(task_id='Stop_execution', dag=dag)\n\nstart_operator \\\n >> create_trips_table \\\n >> [stage_events_to_redshift, stage_songs_to_redshift] \\\n >> load_songplays_table \\\n >> [ load_songs_table, load_artists_table, load_time_table, load_users_table] \\\n >> run_quality_checks \\\n >> end_operator\n\n \n\n \n", "sub_path": "airflow/dags/sparkify_dag.py", "file_name": "sparkify_dag.py", "file_ext": "py", "file_size_in_byte": 3303, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "datetime.datetime.now", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 16, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 20, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 22, "usage_type": "call"}, {"api_name": "airflow.DAG", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 28, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "airflow.conf.get", "line_number": 31, "usage_type": "call"}, {"api_name": "airflow.conf", "line_number": 31, "usage_type": "name"}, {"api_name": "airflow.operators.postgres_operator.PostgresOperator", "line_number": 34, "usage_type": "call"}, {"api_name": "airflow.operators.dummy_operator.DummyOperator", "line_number": 41, "usage_type": "call"}, {"api_name": "airflow.operators.StageToRedshiftOperator", "line_number": 43, "usage_type": "call"}, {"api_name": "airflow.operators.StageToRedshiftOperator", "line_number": 54, "usage_type": "call"}, {"api_name": "airflow.operators.LoadFactOperator", "line_number": 65, "usage_type": "call"}, {"api_name": "helpers.SqlQueries.songplay_table_insert", "line_number": 70, "usage_type": "attribute"}, {"api_name": "helpers.SqlQueries", "line_number": 70, "usage_type": "name"}, {"api_name": "airflow.operators.LoadDimensionOperator", "line_number": 74, "usage_type": "call"}, {"api_name": "helpers.SqlQueries.song_table_insert", "line_number": 79, "usage_type": "attribute"}, {"api_name": "helpers.SqlQueries", "line_number": 79, "usage_type": "name"}, {"api_name": "airflow.operators.LoadDimensionOperator", "line_number": 84, "usage_type": "call"}, {"api_name": "helpers.SqlQueries.user_table_insert", "line_number": 89, "usage_type": "attribute"}, {"api_name": "helpers.SqlQueries", "line_number": 89, "usage_type": "name"}, {"api_name": "airflow.operators.LoadDimensionOperator", "line_number": 93, "usage_type": "call"}, {"api_name": "helpers.SqlQueries.artist_table_insert", "line_number": 98, "usage_type": "attribute"}, {"api_name": "helpers.SqlQueries", "line_number": 98, "usage_type": "name"}, {"api_name": "airflow.operators.LoadDimensionOperator", "line_number": 102, "usage_type": "call"}, {"api_name": "helpers.SqlQueries.time_table_insert", "line_number": 107, "usage_type": "attribute"}, {"api_name": "helpers.SqlQueries", "line_number": 107, "usage_type": "name"}, {"api_name": "airflow.operators.DataQualityOperator", "line_number": 112, "usage_type": "call"}, {"api_name": "airflow.operators.dummy_operator.DummyOperator", "line_number": 119, "usage_type": "call"}]}
+{"seq_id": "136721298", "text": "import pika\nimport pymongo\nimport pymysql\nimport datetime\n# import schedule\nimport os\nimport time\nimport CommonConnection\n\n\nclass DynamicQuening:\n def __init__(self):\n\n self.mongodb = CommonConnection.MongoConnection()\n self.db = CommonConnection.MySQLConnection()\n self.RabbitCon = CommonConnection.RabbitMQConnection()\n self.IPAddr = CommonConnection.ServivesIP()\n self.maxLength = 100000000\n self.maxPriority = 9\n\n\n def QueueCreatorDatabase(self):\n\n\n cur = self.db.cursor()\n cur.execute(\"select * from tbl_Bli_GroupMaster\")\n CurData = cur.fetchall()\n\n cur.close()\n self.db.close()\n return CurData\n\n\n def RetailGroupSelector(self):\n cur = self.db.cursor()\n cur.execute(\"select * from tbl_Bli_GroupMaster where businessType = 'Retail'\")\n CurData = cur.fetchall()\n\n cur.close()\n self.db.close()\n return CurData\n\n\n def HotelGroupSelector(self):\n cur = self.db.cursor()\n cur.execute(\"select * from tbl_Bli_GroupMaster where businessType = 'Hotel'\")\n CurData = cur.fetchall()\n\n cur.close()\n self.db.close()\n return CurData\n\n\n\n\n def run(self):\n self.connection = pika.BlockingConnection(pika.ConnectionParameters(host='localhost'))\n self.channel = self.connection.channel()\n args = {}\n #args[\"x-max-length\"] = self.maxLength\n args['x-max-priority'] = self.maxPriority\n\n CurData = DynamicQuening.QueueCreatorDatabase(self)\n for rows in CurData:\n print(rows)\n Groupname = rows[1]\n\n self.channel.queue_declare(queue=str(Groupname), durable=True, arguments=args)\n self.channel.queue_declare(queue=\"Parser\" + str(Groupname), durable=True, arguments=args)\n\n print(\"Queues updated\")\n\n\n\nwhile True:\n a = DynamicQuening()\n a.run()\n time.sleep(3600)\n\n\n\n", "sub_path": "eCube_Hotel_2/HotelMessaging/Ecube2.0MessagingQueueLatest/ScrappingProducer/Queues/ScraperQueue/DynamicQueueCreator.py", "file_name": "DynamicQueueCreator.py", "file_ext": "py", "file_size_in_byte": 1934, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "CommonConnection.MongoConnection", "line_number": 14, "usage_type": "call"}, {"api_name": "CommonConnection.MySQLConnection", "line_number": 15, "usage_type": "call"}, {"api_name": "CommonConnection.RabbitMQConnection", "line_number": 16, "usage_type": "call"}, {"api_name": "CommonConnection.ServivesIP", "line_number": 17, "usage_type": "call"}, {"api_name": "pika.BlockingConnection", "line_number": 57, "usage_type": "call"}, {"api_name": "pika.ConnectionParameters", "line_number": 57, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 78, "usage_type": "call"}]}
+{"seq_id": "87507547", "text": "from bs4 import BeautifulSoup, NavigableString\nfrom urllib.request import urlopen\n\nclass Host:\n\n ALLOWED_KEYS = [\n 'name',\n 'status',\n 'uptime',\n 'users',\n 'load'\n ]\n\n def __init__(self, **kwargs: dict) -> None:\n for k in kwargs:\n if k not in self.ALLOWED_KEYS:\n raise AttributeError\n\n self.name = kwargs.get('name')\n self.status = self.__status(kwargs.get('status'))\n self.uptime = self.__uptime(kwargs.get('uptime'))\n self.users = self.__users(kwargs.get('users'))\n self.load = self.__load(kwargs.get('load'))\n\n def __status(self, status):\n if status == 'up':\n return True\n if status == 'down':\n return False\n\n def __uptime(self, uptime):\n if not '+' in uptime:\n return 0\n else:\n hour = uptime.split('+')[0]\n return int(hour)\n\n def __users(self, users):\n if len(users) == 0:\n return 0\n else:\n return int(users)\n\n def __load(self, load):\n if len(load) == 0:\n return 0.0\n else:\n return float(load)\n\nclass Scraper:\n\n PARSER = 'html.parser'\n\n def __init__(self, url):\n self.url = url\n self.hosts = []\n self.timestamp = ''\n\n def __get_soup(self):\n html = urlopen(self.url)\n soup = BeautifulSoup(html, self.PARSER)\n return soup\n\n def __parse_html(self):\n soup = self.__get_soup()\n rows = soup.find_all('tr')\n\n for _ in rows[0].children:\n time = _.get_text()\n self.timestamp = time\n\n for i in range(3, len(rows)):\n data = []\n for row in rows[i].children:\n if type(row) != NavigableString:\n text = row.get_text()\n data.append(text)\n\n host = Host(name=data[0], status=data[1], uptime=data[2],\n users=data[3], load=data[4])\n self.hosts.append(host)\n\n def get_hosts(self):\n return self.hosts\n\n def get_timestamp(self):\n return self.timestamp\n\n def update(self):\n self.__parse_html()\n", "sub_path": "lab/scraper.py", "file_name": "scraper.py", "file_ext": "py", "file_size_in_byte": 2199, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "urllib.request.urlopen", "line_number": 60, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 61, "usage_type": "call"}, {"api_name": "bs4.NavigableString", "line_number": 75, "usage_type": "name"}]}
+{"seq_id": "85931122", "text": "#!/usr/bin/env python\nimport barcode\nfrom barcode.writer import ImageWriter\nfrom docx import Document\nfrom docx.enum.text import WD_ALIGN_PARAGRAPH\nfrom docx.shared import Cm\nfrom egcg_core.config import cfg\n\nfrom EPPs.common import SendMailEPP\n\n\nclass GenerateTrackingLetter(SendMailEPP):\n '''Automatically generates the tracking letter sent to customers for tube shipments by populating a word template\n with a 128 format barcode containing the container name'''\n\n _max_nb_project = 1\n\n\n # additional argument required to obtain the file location for newly created tracking letter in the LIMS step\n def __init__(self, argv=None):\n super().__init__(argv)\n self.letter = self.cmd_args.letter\n\n @staticmethod\n def add_args(argparser):\n argparser.add_argument(\n '-t', '--letter', type=str, required=True, help='Tracking letter generated by the LIMS'\n )\n\n def _run(self):\n\n # obtain all of the inputs for the step\n all_inputs = self.artifacts\n\n # 96 well plate so don't need tracking letter\n if all_inputs[0].container.type.name == \"96 well plate\":\n return 0\n\n EAN = barcode.get_barcode_class('code128')\n\n ean = EAN(all_inputs[0].container.name, writer=ImageWriter())\n save_options = {'font_size': 20,\n 'text_distance': 2,\n 'module_height': 15,\n 'module_width': 0.3}\n ean.save('code128', options=save_options)\n\n document = Document(cfg.query('file_templates', 'tracking_letter'))\n\n for paragraph in document.paragraphs:\n if 'The barcode(s) above provides confirmation' in paragraph.text:\n p = paragraph.insert_paragraph_before('')\n p = p.insert_paragraph_before('')\n p.alignment = WD_ALIGN_PARAGRAPH.CENTER\n r = p.add_run()\n r.add_picture('code128.png', width=Cm(5))\n\n document.save(self.letter + '-Edinburgh_Genomics_Sample_Tracking_Letter_' + self.projects[0].name + '.docx')\n\n\nif __name__ == '__main__':\n GenerateTrackingLetter().run()\n", "sub_path": "scripts/create_sample_tracking_letter.py", "file_name": "create_sample_tracking_letter.py", "file_ext": "py", "file_size_in_byte": 2145, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "EPPs.common.SendMailEPP", "line_number": 12, "usage_type": "name"}, {"api_name": "barcode.get_barcode_class", "line_number": 39, "usage_type": "call"}, {"api_name": "barcode.writer.ImageWriter", "line_number": 41, "usage_type": "call"}, {"api_name": "docx.Document", "line_number": 48, "usage_type": "call"}, {"api_name": "egcg_core.config.cfg.query", "line_number": 48, "usage_type": "call"}, {"api_name": "egcg_core.config.cfg", "line_number": 48, "usage_type": "name"}, {"api_name": "docx.enum.text.WD_ALIGN_PARAGRAPH.CENTER", "line_number": 54, "usage_type": "attribute"}, {"api_name": "docx.enum.text.WD_ALIGN_PARAGRAPH", "line_number": 54, "usage_type": "name"}, {"api_name": "docx.shared.Cm", "line_number": 56, "usage_type": "call"}]}
+{"seq_id": "634505129", "text": "from burp import IBurpExtender, IProxyListener, IHttpListener, IResponseInfo, ITab, ITextEditor\nfrom java.io import PrintWriter\nfrom datetime import datetime\nfrom javax import swing\nfrom java.awt import BorderLayout\nfrom ast import literal_eval\n\nclass BurpExtender(IBurpExtender, IProxyListener, IHttpListener, IResponseInfo, ITab, ITextEditor):\n filePathBase = \"/tmp\"\n fileMimeTypes = [\"JPEG\", \"PNG\", \"GIF\"]\n\n # Used to store the config in Burp\n FILELOCATION = \"location\"\n MIMETYPES = \"mimetypes\"\n\n def saveData(self, e):\n # self._stdout.println (e.getSource().getText() + \" was clicked\")\n # self._stdout.println (self.saveLocationInput.getText())\n\n location = self.saveLocationInput.getText()\n\n # force a / on the end if not provided\n\n if (location[len(location)-1] != \"/\"):\n location = location + \"/\"\n self._stdout.println(\"Saving files to: \" + location)\n\n self.filePathBase = location\n\n text = self.mimeTypesInput.getText()\n upper = text.upper()\n mimeTypesToList = upper.split(\",\")\n self._stdout.println(mimeTypesToList)\n self.fileMimeTypes = mimeTypesToList\n\n self._stdout.println(\"Matching MIME Types: \" + repr(mimeTypesToList))\n\n # Save the location\n self._callbacks.saveExtensionSetting (self.FILELOCATION, location)\n self._callbacks.saveExtensionSetting (self.MIMETYPES, repr(self.fileMimeTypes))\n\n def initUI(self):\n self.tab = swing.JPanel()\n\n # Create the text area at the top of the tab\n textPanel = swing.JPanel()\n boxVertical = swing.Box.createVerticalBox()\n\n # Create the label for save location\n boxHorizontal = swing.Box.createHorizontalBox()\n textLabel = swing.JLabel(\"Save location: \")\n boxHorizontal.add(textLabel)\n boxVertical.add(boxHorizontal)\n\n # Create save location input\n boxHorizontal = swing.Box.createHorizontalBox()\n self.saveLocationInput = swing.JTextField(100)\n boxHorizontal.add(self.saveLocationInput)\n boxVertical.add(boxHorizontal)\n\n # Create the label for the mime type\n boxHorizontal = swing.Box.createHorizontalBox()\n textLabel = swing.JLabel(\"MIME Types - comma separated: \")\n boxHorizontal.add(textLabel)\n boxVertical.add(boxHorizontal)\n\n # Create MIME type input\n boxHorizontal = swing.Box.createHorizontalBox()\n self.mimeTypesInput = swing.JTextField(100)\n boxHorizontal.add(self.mimeTypesInput)\n boxVertical.add(boxHorizontal)\n\n # Save button\n boxHorizontal = swing.Box.createHorizontalBox()\n saveButton = swing.JButton(\"Save\")\n saveButton.addActionListener(self.saveData)\n boxHorizontal.add(saveButton)\n boxVertical.add(boxHorizontal)\n\n # Output pane label\n boxHorizontal = swing.Box.createHorizontalBox()\n textLabel = swing.JLabel(\"Output\")\n boxHorizontal.add(textLabel)\n boxVertical.add(boxHorizontal)\n\n # Output pane\n boxHorizontal = swing.Box.createHorizontalBox()\n # This is an attempt at using a Burp ITextEditor, but \n # I need to work out how to add it to the box\n self.outputBox = self._callbacks.createTextEditor()\n self.outputBox.setEditable(False)\n boxHorizontal.add(self.outputBox.getComponent())\n boxVertical.add(boxHorizontal)\n\n # Add the text label and area to the text panel\n textPanel.add(boxVertical)\n\n # Add the text panel to the top of the main tab\n self.tab.add(textPanel, BorderLayout.NORTH) \n\n def getTabCaption(self):\n return \"Save Browsing Files\"\n\n def getUiComponent(self):\n return self.tab\n\n def registerExtenderCallbacks( self, callbacks):\n extName = \"Save Files\"\n # keep a reference to our callbacks object and add helpers\n self._callbacks = callbacks\n self._helpers = self._callbacks.getHelpers()\n\n # set our extension name\n self._callbacks.setExtensionName(extName)\n\n # obtain our output streams\n self._stdout = PrintWriter(self._callbacks.getStdout(), True)\n self._stderr = PrintWriter(self._callbacks.getStderr(), True)\n\n # register ourselves as a Proxy listener\n self._callbacks.registerHttpListener(self)\n\n # print extension name\n self._stdout.println(extName)\n\n # Build list to compare against\n # Need to load this from storage as well\n self.fileMimeTypes = [\"JPEG\", \"PNG\", \"GIF\"]\n\n # Load the location from Burp storage\n self.filePathBase = self._callbacks.loadExtensionSetting(self.FILELOCATION)\n\n # Default to /tmp\n # May be better to check the OS to make this decision, but sticking\n # with this for now.\n if self.filePathBase is None:\n self.filePathBase = \"/tmp/\"\n\n self._stdout.println(\"Saving files to: \" + self.filePathBase)\n\n loadedMimeTypes = self._callbacks.loadExtensionSetting(self.MIMETYPES)\n if loadedMimeTypes is None:\n self.mimeTypesInput = [\"JPEG\", \"PNG\", \"GIF\"]\n else:\n # should probably check to see what happens if loadedMimeTypes does\n # not eval correctly.\n self.mimeTypesInput = literal_eval(loadedMimeTypes)\n self._stdout.println(\"loaded: \" + loadedMimeTypes)\n\n mimeTypesAsString = ','.join(self.mimeTypesInput)\n self._stdout.println(\"parsed: \" + mimeTypesAsString)\n\n self.initUI()\n self._callbacks.addSuiteTab(self)\n self.saveLocationInput.setText(self.filePathBase)\n self.mimeTypesInput.setText(mimeTypesAsString)\n\n return\n\n def processHttpMessage(self, toolflag, messageIsRequest, messageInfo):\n if (messageIsRequest == False):\n response = messageInfo.getResponse()\n responseInfo = self._helpers.analyzeResponse(response)\n\n # request = messageInfo.getRequest()\n # self._stdout.println(type(request))\n\n # for header in request:\n # self._stdout.println(header)\n\n # for header in request.getHeaders():\n # self._stdout.println(header)\n #\n # self._stdout.println(request)\n\n # Get MIME types\n inferredMime = responseInfo.getInferredMimeType()\n statedMime = responseInfo.getStatedMimeType()\n\n # Get response body\n bodyOffset = responseInfo.getBodyOffset()\n # self._stdout.println(bodyOffset)\n # Build image request body\n imgData = response[bodyOffset:]\n # self._stdout.println(imgData)\n\n self._stdout.println(\"Stated MIME Type: \" + statedMime)\n self._stdout.println(\"Inferred MIME Type: \" + inferredMime)\n\n # If multiple files are loaded in the same second they will all get\n # the same name and be overwritten so need to add something extra to \n # the name to ensure it is unique.\n\n if (statedMime in self.fileMimeTypes) or (inferredMime in self.fileMimeTypes):\n # Build file path\n fileName = datetime.now().strftime('%Y-%m-%d-%H-%M-%S-%f')\n fileExtension = \".\" + inferredMime.lower()\n fullFilename = self.filePathBase + fileName + fileExtension\n self.outputBox.append(\"Writing to file: \" + fullFilename + \"\\n\")\n\n # This forces the textarea to autoscroll after the update\n self.outputBox.setCaretPosition(self.outputBox.getDocument().getLength());\n # Write to file\n f = open(fullFilename, \"wb\")\n f.write(imgData)\n f.close()\n return\n", "sub_path": "saveImages.py", "file_name": "saveImages.py", "file_ext": "py", "file_size_in_byte": 7798, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "burp.IBurpExtender", "line_number": 8, "usage_type": "name"}, {"api_name": "burp.IProxyListener", "line_number": 8, "usage_type": "name"}, {"api_name": "burp.IHttpListener", "line_number": 8, "usage_type": "name"}, {"api_name": "burp.IResponseInfo", "line_number": 8, "usage_type": "name"}, {"api_name": "burp.ITab", "line_number": 8, "usage_type": "name"}, {"api_name": "burp.ITextEditor", "line_number": 8, "usage_type": "name"}, {"api_name": "javax.swing.JPanel", "line_number": 43, "usage_type": "call"}, {"api_name": "javax.swing", "line_number": 43, "usage_type": "name"}, {"api_name": "javax.swing.JPanel", "line_number": 46, "usage_type": "call"}, {"api_name": "javax.swing", "line_number": 46, "usage_type": "name"}, {"api_name": "javax.swing.Box.createVerticalBox", "line_number": 47, "usage_type": "call"}, {"api_name": "javax.swing.Box", "line_number": 47, "usage_type": "attribute"}, {"api_name": "javax.swing", "line_number": 47, "usage_type": "name"}, {"api_name": "javax.swing.Box.createHorizontalBox", "line_number": 50, "usage_type": "call"}, {"api_name": "javax.swing.Box", "line_number": 50, "usage_type": "attribute"}, {"api_name": "javax.swing", "line_number": 50, "usage_type": "name"}, {"api_name": "javax.swing.JLabel", "line_number": 51, "usage_type": "call"}, {"api_name": "javax.swing", "line_number": 51, "usage_type": "name"}, {"api_name": "javax.swing.Box.createHorizontalBox", "line_number": 56, "usage_type": "call"}, {"api_name": "javax.swing.Box", "line_number": 56, "usage_type": "attribute"}, {"api_name": "javax.swing", "line_number": 56, "usage_type": "name"}, {"api_name": "javax.swing.JTextField", "line_number": 57, "usage_type": "call"}, {"api_name": "javax.swing", "line_number": 57, "usage_type": "name"}, {"api_name": "javax.swing.Box.createHorizontalBox", "line_number": 62, "usage_type": "call"}, {"api_name": "javax.swing.Box", "line_number": 62, "usage_type": "attribute"}, {"api_name": "javax.swing", "line_number": 62, "usage_type": "name"}, {"api_name": "javax.swing.JLabel", "line_number": 63, "usage_type": "call"}, {"api_name": "javax.swing", "line_number": 63, "usage_type": "name"}, {"api_name": "javax.swing.Box.createHorizontalBox", "line_number": 68, "usage_type": "call"}, {"api_name": "javax.swing.Box", "line_number": 68, "usage_type": "attribute"}, {"api_name": "javax.swing", "line_number": 68, "usage_type": "name"}, {"api_name": "javax.swing.JTextField", "line_number": 69, "usage_type": "call"}, {"api_name": "javax.swing", "line_number": 69, "usage_type": "name"}, {"api_name": "javax.swing.Box.createHorizontalBox", "line_number": 74, "usage_type": "call"}, {"api_name": "javax.swing.Box", "line_number": 74, "usage_type": "attribute"}, {"api_name": "javax.swing", "line_number": 74, "usage_type": "name"}, {"api_name": "javax.swing.JButton", "line_number": 75, "usage_type": "call"}, {"api_name": "javax.swing", "line_number": 75, "usage_type": "name"}, {"api_name": "javax.swing.Box.createHorizontalBox", "line_number": 81, "usage_type": "call"}, {"api_name": "javax.swing.Box", "line_number": 81, "usage_type": "attribute"}, {"api_name": "javax.swing", "line_number": 81, "usage_type": "name"}, {"api_name": "javax.swing.JLabel", "line_number": 82, "usage_type": "call"}, {"api_name": "javax.swing", "line_number": 82, "usage_type": "name"}, {"api_name": "javax.swing.Box.createHorizontalBox", "line_number": 87, "usage_type": "call"}, {"api_name": "javax.swing.Box", "line_number": 87, "usage_type": "attribute"}, {"api_name": "javax.swing", "line_number": 87, "usage_type": "name"}, {"api_name": "java.awt.BorderLayout.NORTH", "line_number": 99, "usage_type": "attribute"}, {"api_name": "java.awt.BorderLayout", "line_number": 99, "usage_type": "name"}, {"api_name": "java.io.PrintWriter", "line_number": 117, "usage_type": "call"}, {"api_name": "java.io.PrintWriter", "line_number": 118, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 147, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 196, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 196, "usage_type": "name"}]}
+{"seq_id": "273992088", "text": "import pandas as pd\nfrom pandas import DataFrame\nimport numpy as np\nimport matplotlib.dates as mdates\nimport matplotlib\nmatplotlib.use('TkAgg')\nimport matplotlib.pyplot as plt\nimport re\nimport scipy.stats as stats\n\narr = []\ny_arr = []\n\n# Read CSV\ndf = pd.read_csv('./potato_rates.csv')\n# Change values to float\ndf['Modal Price'] = df['Modal Price'].astype('float')\ndf['Max Price'] = df['Max Price'].astype('float')\ndf['Min Price'] = df['Min Price'].astype('float')\ndf['Price Date'] = df['Price Date'].astype('string')\n\n# Select only 2016 data\n# df = df[df['Price Date'].str.contains('-16')]\n\n\ndef it(a, b):\n x = (b-1)*7\n x = 1 if(x==0) else x\n return str(a)+\"-\"+str(x).zfill(3)\n\n\n# Convert Rs/kg\ndf['Modal Price'] = np.divide(df['Modal Price'], 100)\ndf['Max Price'] = np.divide(df['Max Price'], 100)\ndf['Min Price'] = np.divide(df['Min Price'], 100)\ndf['Price Date'] = pd.to_datetime(df['Price Date'])\n\n\ntimes = pd.DatetimeIndex(df['Price Date'])\ndf = DataFrame({\"Modal Price\": df.groupby([times.year])['Modal Price'].mean()}).reset_index()\n# print df\n# exit()\n# df['Date'] = pd.Series(pd.to_datetime(map(it, df['index']), format='%Y-%j'))\n\n\nfig = plt.figure()\nfig.suptitle('Potato', fontsize=14)\nplt.scatter(df['index'], df['Modal Price'])\n# plt.plot_date(df['index'], df['Modal Price'])\nplt.show()\n\n# print df['Price Date']\n", "sub_path": "rate_analysis/test2.py", "file_name": "test2.py", "file_ext": "py", "file_size_in_byte": 1335, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "matplotlib.use", "line_number": 6, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 35, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 36, "usage_type": "call"}, {"api_name": "pandas.DatetimeIndex", "line_number": 39, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}]}
+{"seq_id": "346061059", "text": "import torch\nimport attr\nimport numpy as np\nfrom skultrafast.dataset import TimeResSpec\nimport math\nexp_half = math.exp(1 / 2.)\nfrom scipy.optimize import least_squares\n\ndef lstsq(b, y, alpha=0.01):\n \"\"\"\n Batched linear least-squares for pytorch with optional L1 regularization.\n\n Parameters\n ----------\n\n b : shape(L, M, N)\n y : shape(L, M)\n\n Returns\n -------\n tuple of (coefficients, model, residuals)\n\n \"\"\"\n bT = b.transpose(-1, -2)\n AA = torch.bmm(bT, b)\n if alpha != 0:\n diag = torch.diagonal(AA, dim1=1, dim2=2)\n diag += alpha\n RHS = torch.bmm(bT, y[:, :, None])\n X, LU = torch.gesv(RHS, AA)\n fit = torch.bmm(b, X)[..., 0]\n res = y - fit\n return X[..., 0], fit, res\n\n@attr.s(auto_attribs=True)\nclass FitterTorch:\n dataset : TimeResSpec = attr.ib()\n zero_func : function = attr.ib(lambda x: np.zeros_like(x))\n done_eval : bool = attr.ib(False)\n use_cuda : bool = attr.ib(True)\n disp_poly_deg : int = attr.ib(2)\n model_coh : bool = attr.ib(0)\n\n def __attrs_post_init__(self):\n ds = self.dataset\n self.dev_data = torch.from_numpy(ds.data.T)\n if self.use_cuda:\n self.dev_data = self.dev_data.cuda()\n\n\n def eval(self, tt, w, tau, model_coh=False):\n \"\"\"\n Evaluates a model for given arrays\n\n Parameters\n ----------\n tt : ndarray\n Contains the delay-times, should have the same shape as the data.\n w : float\n The IRF width.\n tau : ndarray\n Contains the decay times.\n \"\"\"\n tt = torch.from_numpy(tt)\n tau = torch.from_numpy(tau)\n if self.use_cuda:\n tt = tt.cuda()\n tau= tau.cuda()\n\n k = 1 / (tau[None, None, ...])\n t = (tt)[..., None]\n if w == 0:\n A = torch.exp(-k*tt)\n else:\n A = torch.exp(k * (w * w * k / (4.0) - t)) \\\n * 0.5 * torch.erfc(-t / w + w * k / (2.0))\n if model_coh:\n coh = torch.exp(-0.5 * (tt / w) * (tt / w))\n coh = coh[:, :, None].repeat((1, 1, 3))\n coh[..., 1] *= (-tt * exp_half / w)\n coh[..., 2] *= (tt * tt / w / w - 1)\n A = torch.cat((A, coh), dim=-1)\n\n X, fit, res = lstsq(A, self.data)\n self.done_eval = True\n self.c = X\n self.model = fit\n self.residuals = res\n return X, fit, res\n\n\n def fit_func(self, x):\n ds = self.dataset\n disp_coefs = x[:self.disp_poly_deg]\n w = float(x[self.disp_poly_deg])\n taus = x[self.disp_poly_deg+1:]\n t_zeros = np.poly1d(disp_coefs)(ds.wavenumbers)\n tt = np.subtract.outer(ds.t, t_zeros).T\n c, model, res = self.eval_torch(tt, w, taus, True)\n return res.cpu().numpy().ravel()\n\n def start_fit(self, w, taus, fix_last_tau=False, fix_width=False,\n fix_disp=False):\n ds = self.dataset\n time_zeros = self.zero_func(ds.wavenumbers)\n disp_guess = np.polyfit(ds.wavenumbers, time_zeros)\n x0 = np.hstack((disp_guess, w, taus))\n idx = np.ones_like(x0, dtype='bool')\n if fix_last_tau:\n idx[-1] = False\n if fix_width:\n idx[self.disp_poly_deg] = False\n if fix_disp:\n idx[:self.disp_poly_deg] = False\n\n start_guess = x0[idx]\n\n def fix_func(x):\n x0[idx] = x\n return self.fit_func(x0)\n\n bounds = np.array([(-np.inf, np.inf)] * len(x0))\n bounds[self.disp_poly_deg:, 0] = 0\n bounds = bounds[idx, :]\n x = least_squares(fix_func, start_guess, bounds=bounds.T)\n return x, x0\n\n", "sub_path": "skultrafast/base_funcs/pytorch_fitter.py", "file_name": "pytorch_fitter.py", "file_ext": "py", "file_size_in_byte": 3671, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "math.exp", "line_number": 6, "usage_type": "call"}, {"api_name": "torch.bmm", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.diagonal", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.bmm", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.gesv", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.bmm", "line_number": 31, "usage_type": "call"}, {"api_name": "skultrafast.dataset.TimeResSpec", "line_number": 37, "usage_type": "name"}, {"api_name": "attr.ib", "line_number": 37, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 38, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 39, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 40, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 41, "usage_type": "call"}, {"api_name": "attr.ib", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.exp", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.exp", "line_number": 75, "usage_type": "call"}, {"api_name": "torch.erfc", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.exp", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.poly1d", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.subtract.outer", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.subtract", "line_number": 98, "usage_type": "attribute"}, {"api_name": "numpy.polyfit", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.ones_like", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 122, "usage_type": "attribute"}, {"api_name": "scipy.optimize.least_squares", "line_number": 125, "usage_type": "call"}, {"api_name": "attr.s", "line_number": 35, "usage_type": "call"}]}
+{"seq_id": "451957585", "text": "from django.http import Http404\nfrom django.shortcuts import render\n\n# my import\nfrom .models import Company, Phone\n\n# -------------------------------------------------------------------\n# -------------------------------------------------------------------\ndef index(request):\n all_companies = Company.objects.all()\n\n context = {\n 'all_companies': all_companies\n }\n return render(request, 'main/index.html', context)\n\n# -------------------------------------------------------------------\n# -------------------------------------------------------------------\ndef company(request, company_id):\n try:\n company = Company.objects.get(id=company_id)\n except Company.DoesNotExist:\n raise Http404(\"Company does not exist!\")\n\n context = {\n 'company': company\n }\n return render(request, 'main/company.html', context)\n", "sub_path": "main/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 862, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "models.Company.objects.all", "line_number": 10, "usage_type": "call"}, {"api_name": "models.Company.objects", "line_number": 10, "usage_type": "attribute"}, {"api_name": "models.Company", "line_number": 10, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 15, "usage_type": "call"}, {"api_name": "models.Company.objects.get", "line_number": 21, "usage_type": "call"}, {"api_name": "models.Company.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "models.Company", "line_number": 21, "usage_type": "name"}, {"api_name": "models.Company.DoesNotExist", "line_number": 22, "usage_type": "attribute"}, {"api_name": "models.Company", "line_number": 22, "usage_type": "name"}, {"api_name": "django.http.Http404", "line_number": 23, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 28, "usage_type": "call"}]}
+{"seq_id": "597167877", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('prices', '0002_prices_gasoline_price'),\n ]\n\n operations = [\n migrations.AddField(\n model_name='prices',\n name='crack_spread',\n field=models.FloatField(default=b'11.00'),\n ),\n ]\n", "sub_path": "prices/migrations/0003_prices_crack_spread.py", "file_name": "0003_prices_crack_spread.py", "file_ext": "py", "file_size_in_byte": 412, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}]}
+{"seq_id": "40259208", "text": "from django.db import models\nfrom django.contrib.auth.models import User\n\n\n\n# Create your models here.\n\n\n\nclass Department(models.Model):\n name=models.CharField(max_length=40);\n leader=models.CharField(max_length=40);\n main_speciality=models.CharField(max_length=40);\n number_of_members=models.IntegerField(max_length=4);\n #worker=models.ForeignKey(Worker)\n\nclass Worker(models.Model):\n name=models.CharField(max_length=40);\n gift=models.IntegerField(max_length=5);\n photo=models.ImageField(null=True,blank=True)\n department=models.ManyToManyField(Department, related_name='workers',blank=True,null=True);\n #user=models.OneToOneField(User)\n def bit (self):\n if self.article_image:\n return u'
'% self.article_image.url\n else:\n return u'(none)'\n bit.short_description = 'Изображение'\n bit.allow_tags = True\n\n\n\n", "sub_path": "dz/dz/app1/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 919, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.db.models.Model", "line_number": 10, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 10, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 11, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 11, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 12, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 12, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 17, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.ImageField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}]}
+{"seq_id": "406423597", "text": "\"\"\"\nNOAA ISD Lite import routine\n\nGet hourly weather data for weather stations worldwide.\n\nThe code is licensed under the MIT license.\n\"\"\"\n\nimport os\nfrom sys import argv\nfrom datetime import datetime\nfrom io import BytesIO\nfrom ftplib import FTP\nimport gzip\nimport pandas as pd\nfrom routines import Routine\nfrom routines.convert import ms_to_kmh, temp_dwpt_to_rhum\nfrom routines.schema import hourly_global\n\n# Configuration\nMODE = argv[1]\nSTATIONS_PER_CYCLE = 1 if MODE == 'recent' else 4\nUSAF_WBAN_PATH = os.path.abspath(\n os.path.join(\n os.path.dirname(__file__),\n '../../..',\n 'resources')) + '/usaf_wban.csv'\nCURRENT_YEAR = datetime.now().year\n\n# Required columns\nusecols = [0, 1, 2, 3, 4, 5, 6, 7, 8, 10]\n\n# Column names\nNAMES = ['time', 'temp', 'dwpt', 'pres', 'wdir', 'wspd', 'prcp']\n\n# Create new task\ntask = Routine('import.noaa.hourly.global')\n\n# Get counter value\ncounter = task.get_var('station_counter_' + MODE)\nskip = 0 if counter is None else int(counter)\n\n# Get year\nif MODE == 'historical':\n year = task.get_var('year')\n year = 1901 if year is None else int(year)\n\n# Get ISD Lite stations\ntry:\n stations = pd.read_csv(\n USAF_WBAN_PATH,\n dtype='str',\n skiprows=skip,\n nrows=STATIONS_PER_CYCLE,\n names=[\n 'id',\n 'usaf',\n 'wban'])\nexcept pd.errors.EmptyDataError:\n stations = None\n pass\n\n# Update counter\nif stations is None or len(stations.index) < STATIONS_PER_CYCLE:\n # Reset counter\n task.set_var('station_counter_' + MODE, 0)\n # Reset year\n if MODE == 'historical':\n if year >= CURRENT_YEAR - 2:\n task.set_var('year', 1901)\n else:\n task.set_var('year', year + 1)\n exit()\nelse:\n task.set_var('station_counter_' + MODE, skip + STATIONS_PER_CYCLE)\n\n# Connect to NOAA FTP Server\nftp = FTP('ftp.ncdc.noaa.gov')\nftp.login()\n\n# Get list of years\nif MODE == 'recent':\n years = range(CURRENT_YEAR - 1, CURRENT_YEAR + 1)\nelse:\n years = range(year, year + 1)\n\n# Import data for each weather station\nfor station in stations.to_dict(orient='records'):\n\n for year in years:\n\n try:\n\n ftp.cwd('/pub/data/noaa/isd-lite/' + str(year))\n\n filename = station[\"usaf\"] + '-' + \\\n station[\"wban\"] + '-' + str(year) + '.gz'\n\n if filename in ftp.nlst():\n\n # Download file\n local_file = os.path.dirname(__file__) + os.sep + filename\n ftp.retrbinary(\n \"RETR \" + filename,\n open(\n local_file,\n 'wb').write)\n\n # Unzip file\n file = gzip.open(local_file, 'rb')\n raw = file.read()\n file.close()\n\n # Remove .gz file\n os.remove(local_file)\n\n df = pd.read_fwf(\n BytesIO(raw),\n parse_dates={\n 'time': [\n 0,\n 1,\n 2,\n 3]},\n na_values=-\n 9999,\n header=None,\n usecols=usecols)\n\n # Rename columns\n df.columns = NAMES\n\n # Adapt columns\n df['temp'] = df['temp'].div(10)\n df['dwpt'] = df['dwpt'].div(10)\n df['pres'] = df['pres'].div(10)\n df['wspd'] = df['wspd'].div(10).apply(ms_to_kmh)\n df['prcp'] = df['prcp'].div(10)\n\n # Calculate humidity data\n df['rhum'] = df.apply(\n lambda row: temp_dwpt_to_rhum(row), axis=1)\n\n # Drop dew point column\n df = df.drop('dwpt', axis=1)\n\n # Add station column\n df['station'] = station['id']\n\n # Set index\n df = df.set_index(['station', 'time'])\n\n # Round decimals\n df = df.round(1)\n\n # Write data into Meteostat database\n task.write(df, hourly_global)\n\n except BaseException:\n\n pass\n\n# Quit FTP connection\nftp.quit()\n", "sub_path": "import/noaa/hourly/global.py", "file_name": "global.py", "file_ext": "py", "file_size_in_byte": 4269, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sys.argv", "line_number": 21, "usage_type": "name"}, {"api_name": "os.path.abspath", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 28, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 28, "usage_type": "name"}, {"api_name": "routines.Routine", "line_number": 37, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 50, "usage_type": "call"}, {"api_name": "pandas.errors", "line_number": 59, "usage_type": "attribute"}, {"api_name": "ftplib.FTP", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path", "line_number": 102, "usage_type": "attribute"}, {"api_name": "os.sep", "line_number": 102, "usage_type": "attribute"}, {"api_name": "gzip.open", "line_number": 110, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 115, "usage_type": "call"}, {"api_name": "pandas.read_fwf", "line_number": 117, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 118, "usage_type": "call"}, {"api_name": "routines.convert.ms_to_kmh", "line_number": 137, "usage_type": "argument"}, {"api_name": "routines.convert.temp_dwpt_to_rhum", "line_number": 142, "usage_type": "call"}, {"api_name": "routines.schema.hourly_global", "line_number": 157, "usage_type": "argument"}]}
+{"seq_id": "313040177", "text": "from unittest import TestCase\n\nfrom ..lib.binarystreams import stream_compare\nfrom ..lib.contextlib import tempdir\nfrom ..lib.pyx2py import pyx_to_py\n\nfrom _pyio import DEFAULT_BUFFER_SIZE\nfrom itertools import repeat\n\nclass T(TestCase):\n def test(self):\n with tempdir() as root:\n a, b = (root.joinpath(filename) for filename in ('a', 'b'))\n with open(pyx_to_py(__file__), 'rb') as istream:\n data = istream.read()\n with a.open('wb') as ostream:\n for none in repeat(None, DEFAULT_BUFFER_SIZE):\n none # pylint: disable=pointless-statement\n ostream.write(data)\n with b.open('wb') as ostream:\n for none in repeat(None, DEFAULT_BUFFER_SIZE):\n ostream.write(data)\n ostream.write(br'.')\n expected = True, False\n with a.open('rt') as istream0:\n with a.open('rt') as istream1:\n gotten0 = stream_compare(istream0, istream1)\n with a.open('rt') as istream0:\n with b.open('rt') as istream1:\n gotten1 = stream_compare(istream0, istream1)\n actual = gotten0, gotten1\n self.assertEqual(expected, actual)\n", "sub_path": "x19290/test/t00binarystreams.py", "file_name": "t00binarystreams.py", "file_ext": "py", "file_size_in_byte": 1285, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "unittest.TestCase", "line_number": 10, "usage_type": "name"}, {"api_name": "lib.contextlib.tempdir", "line_number": 12, "usage_type": "call"}, {"api_name": "lib.pyx2py.pyx_to_py", "line_number": 14, "usage_type": "call"}, {"api_name": "itertools.repeat", "line_number": 17, "usage_type": "call"}, {"api_name": "_pyio.DEFAULT_BUFFER_SIZE", "line_number": 17, "usage_type": "argument"}, {"api_name": "itertools.repeat", "line_number": 21, "usage_type": "call"}, {"api_name": "_pyio.DEFAULT_BUFFER_SIZE", "line_number": 21, "usage_type": "argument"}, {"api_name": "lib.binarystreams.stream_compare", "line_number": 27, "usage_type": "call"}, {"api_name": "lib.binarystreams.stream_compare", "line_number": 30, "usage_type": "call"}]}
+{"seq_id": "157941863", "text": "import pandas as pd\n# from constants import *\nfrom .constants import *\nimport json\nfrom pyrebase import pyrebase\nfrom urllib.parse import parse_qsl, urljoin, urlparse\n# from database import Database\nfrom .database import Database\n\nfirebase = pyrebase.initialize_app(config)\ndb = firebase.database() # Get a reference to the database service\n\n\nclass BagOfIngredients:\n def __init__(self, username):\n self.username = \"\\'\" + username + \"\\'\" # use a session variable\n self.ingredients = []\n self.number_of_ingredients = 0\n self.boi = None\n self.db = Database()\n self.db.open()\n\n def get_boi(self):\n # Gets bag of ingredients for a certain User\n\n #print(\"Getting Bag of Ingredients from DB>>>\\n\", self.db.get(\"bagofingredients\", \"*\", where=\"user_id=\"+self.username))\n return self.db.get(\"BagOfIngredients\", \"*\", where=\"user_id=\"+self.username)\n\n def push_boi(self, ing: Ingredient):\n # Pushes an ingredient into Bag of Ingredients for the User\n\n columns = \"user_id, ingredient, ingredient_name, amount, unit\"\n data = \"{0},'{1}','{2}',{3},'{4}'\".format(self.username,\n ing.ingredient_full, ing.ingredient, ing.amount, ing.units)\n print(\"Pushing \"+ing.ingredient_full+\" into DB>>> Bag of Ingredients.\")\n push_success = self.db.write(\"BagOfIngredients\", columns, data)\n self.number_of_ingredients += 1\n self.ingredients.append(ing)\n return push_success\n\n def delete_all(self):\n # Deletes all ingredients from Bag for a User\n try:\n print(\"DELETING ALL from BOI with user_id>>>\"+self.username)\n delete_query = \"DELETE FROM bagofingredients WHERE user_id=\"+self.username+\";\"\n self.db.query(delete_query)\n except:\n print(\"ERROR OCCURED IN DELETION!\")\n return False\n return True\n\n def delete_ingredient(self, ingredient_name):\n # Deletes one ingredient\n\n try:\n print(\"DELETING ingredient \"+ingredient_name+\" from BOI with user_id>>>\"+self.username)\n delete_query = \"DELETE FROM bagofingredients WHERE user_id=\"+self.username+ \"AND ingredient_name=\"+ingredient_name+\";\"\n self.db.query(delete_query)\n except:\n print(\"ERROR OCCURED IN DELETION!\")\n return False\n return True\n\n def update_ingredient(self, ingredient_name, new_quantity):\n # Updates ingredient with new quantity\n\n try:\n print(\"UPDATING ingredient \"+ingredient_name+\" from BOI with user_id>>>\"+self.username)\n delete_query = \"UPDATE bagofingredients SET amount=\"+new_quantity+\"WHERE user_id=\"+self.username+\"AND ingredient_name=\"+ingredient_name+\";\"\n self.db.query(delete_query)\n except:\n print(\"ERROR OCCURED IN UPDATING!\")\n return False\n return True\n\n def update_new_boi(self):\n # Deletes boi and adds new one\n \n pass\n\n\n# TEST CASES FOR BOI FOR POSTGRESQL\n# boi_sample = BagOfIngredients(username)\n# boi_sample.get_boi()\n# boi_sample.push_boi(sample_ingredient)\n\n'''\nTHIS CAN BE USED FOR TESTING FIREBASE (OLD).\ndata = sample_user #check constants.py\n\n# CRUD operations example with predefined user from constants.py\nboi_sample = BagOfIngredients()\nauthenticated = boi_sample.authenticate_user(username, password)\nif authenticated:\n print(\"AUTHENTICATED!!\")\n boi_sample.get_boi()\n boi_sample.push_boi(sample_user)\n boi_sample.update_boi(\"diet\",\"non-vegetarian\")\n # boi_sample.delete_boi()\n'''\n", "sub_path": "modules/bag_of_ingredients.py", "file_name": "bag_of_ingredients.py", "file_ext": "py", "file_size_in_byte": 3612, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pyrebase.pyrebase.initialize_app", "line_number": 10, "usage_type": "call"}, {"api_name": "pyrebase.pyrebase", "line_number": 10, "usage_type": "name"}, {"api_name": "database.Database", "line_number": 20, "usage_type": "call"}]}
+{"seq_id": "484576949", "text": "# Twisted Imports\nfrom twisted.internet import reactor, defer, error\nfrom twisted.python import failure, log\n\n# System Imports\nfrom time import time as now\nfrom collections import deque\nimport functools\n\n# NumPy\nimport numpy as np\n\n\nclass Event (object):\n\tdef __init__(self):\n\t\tself.handlers = set()\n\n\tdef handle(self, handler):\n\t\tself.handlers.add(handler)\n\t\treturn self\n\n\tdef unhandle(self, handler):\n\t\tself.handlers.discard(handler)\n\t\treturn self\n\n\tdef fire(self, *args, **kargs):\n\t\tfor handler in self.handlers:\n\t\t\thandler(*args, **kargs)\n\n\tdef getHandlerCount(self):\n\t\treturn len(self.handlers)\n\n\t__iadd__ = handle\n\t__isub__ = unhandle\n\t__call__ = fire\n\t__len__ = getHandlerCount\n\n\nclass EventEmitter (object):\n\tdef on (self, name, function = None):\n\t\tdef _on (function):\n\t\t\ttry:\n\t\t\t\tself._events[name]\n\t\t\texcept (TypeError, AttributeError):\n\t\t\t\tself._events = {}\n\t\t\t\tself._events[name] = []\n\t\t\texcept KeyError:\n\t\t\t\tself._events[name] = []\n\n\t\t\t# Use is instead of in to avoid equality comparison\n\t\t\t# (this would create extra expression objects).\n\t\t\tfor f in self._events[name]:\n\t\t\t\tif function is f:\n\t\t\t\t\treturn function\n\n\t\t\tself._events[name].append(function)\n\n\t\t\treturn function\n\n\t\tif function is None:\n\t\t\treturn _on\n\t\telse:\n\t\t\treturn _on(function)\n\n\tdef once (self, name, function = None):\n\t\tdef _once (function):\n\t\t\t@functools.wraps(function)\n\t\t\tdef g (*args, **kwargs):\n\t\t\t\tfunction(*args, **kwargs)\n\t\t\t\tself.off(name, g)\n\n\t\t\treturn g\n\n\t\tif function is None:\n\t\t\treturn lambda function: self.on(name, _once(function))\n\t\telse:\n\t\t\tself.on(name, _once(function))\n\n\tdef off (self, name = None, function = None):\n\t\ttry:\n\t\t\tself._events\n\t\texcept AttributeError:\n\t\t\treturn\n\n\t\t# If no name is passed, remove all handlers\n\t\tif name is None:\n\t\t\tself._events.clear()\n\n\t\t# If no function is passed, remove all functions\n\t\telif function is None:\n\t\t\ttry:\n\t\t\t\tself._events[name] = []\n\t\t\texcept KeyError:\n\t\t\t\tpass\n\n\t\t# Remove handler [function] from [name]\n\t\telse:\n\t\t\tself._events[name].remove(function)\n\n\tdef listeners (self, event):\n\t\ttry:\n\t\t\treturn self._events[event]\n\t\texcept (AttributeError, KeyError):\n\t\t\treturn []\n\t\n\tdef emit (self, _event, **data):\n\t\thandled = False\n\n\t\ttry:\n\t\t\tevents = self._events[_event][:]\n\t\texcept AttributeError:\n\t\t\treturn False # No events defined yet\n\t\texcept KeyError:\n\t\t\tpass\n\t\telse:\n\t\t\thandled |= bool(len(events))\n\n\t\t\tfor function in events:\n\t\t\t\ttry:\n\t\t\t\t\tfunction(data)\n\t\t\t\texcept:\n\t\t\t\t\tlog.err()\n\n\t\ttry:\n\t\t\tevents = self._events[\"all\"][:]\n\t\texcept KeyError:\n\t\t\tpass\n\t\telse:\n\t\t\thandled |= bool(len(events))\n\n\t\t\tfor function in events:\n\t\t\t\ttry:\n\t\t\t\t\tfunction(_event, data)\n\t\t\t\texcept:\n\t\t\t\t\tlog.err()\n\n\t\treturn handled\n\n\ndef timerange (start, interval, step):\n\tif start < 0:\n\t\t\tstart = now() + start\n\n\treturn np.arange(start, start + interval, step, float)\n\n\n\nclass AsyncQueue (object):\n\t@property\n\tdef running (self):\n\t\treturn self._workers > 0\n\n\t@property\n\tdef current (self):\n\t\treturn self._current\n\n\tdef __init__ (self, worker, concurrency = 1, paused = False):\n\t\tself._tasks = deque()\n\t\tself._worker = worker\n\t\tself._workers = 0\n\t\tself._concurrency = concurrency\n\t\tself._paused = int(paused)\n\t\tself._current = set()\n\n\t\tself.drained = Event()\n\n\tdef pause (self):\n\t\tself._paused += 1\n\n\tdef resume (self):\n\t\tself._paused -= 1\n\t\tself._process()\n\n\tdef append (self, data):\n\t\ttask = _AsyncQueueTask(data)\n\t\tself._tasks.append(task)\n\t\treactor.callLater(0, self._process)\n\t\treturn task.d\n\n\tdef appendleft (self, data):\n\t\ttask = _AsyncQueueTask(data)\n\t\tself._tasks.appendleft(task)\n\t\treactor.callLater(0, self._process)\n\t\treturn task.d\n\n\tdef _process (self):\n\t\tif not self._paused and self._workers < self._concurrency:\n\t\t\tdef run (task):\n\t\t\t\tworker_d = defer.maybeDeferred(self._worker, task.data)\n\t\t\t\tworker_d.addCallbacks(success, error)\n\n\t\t\tdef success (result):\n\t\t\t\ttask.d.callback(result)\n\t\t\t\tnext()\n\n\t\t\tdef error (reason):\n\t\t\t\tif reason.type is AsyncQueueRetry:\n\t\t\t\t\trun(task)\n\t\t\t\telse:\n\t\t\t\t\ttask.d.errback(reason)\n\t\t\t\t\tnext()\n\n\t\t\tdef next ():\n\t\t\t\tself._workers -= 1\n\t\t\t\tself._current.discard(task)\n\t\t\t\treactor.callLater(0, self._process)\n\n\t\t\ttry:\n\t\t\t\ttask = self._tasks.popleft()\n\t\t\texcept IndexError:\n\t\t\t\tself.drained()\n\t\t\telse:\n\t\t\t\tself._workers += 1\n\t\t\t\tself._current.add(task)\n\t\t\t\trun(task)\n\n\tdef __len__ (self):\n\t\treturn len(self._tasks)\n\nclass AsyncQueueRetry (Exception):\n\tpass\n\nclass _AsyncQueueTask (object):\n\tdef __init__ (self, data, deferred = None):\n\t\tself.data = data\n\t\tself.d = deferred or defer.Deferred()\n", "sub_path": "octopus/util.py", "file_name": "util.py", "file_ext": "py", "file_size_in_byte": 4458, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "functools.wraps", "line_number": 67, "usage_type": "call"}, {"api_name": "twisted.python.log.err", "line_number": 122, "usage_type": "call"}, {"api_name": "twisted.python.log", "line_number": 122, "usage_type": "name"}, {"api_name": "twisted.python.log.err", "line_number": 135, "usage_type": "call"}, {"api_name": "twisted.python.log", "line_number": 135, "usage_type": "name"}, {"api_name": "time.time", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 144, "usage_type": "call"}, {"api_name": "collections.deque", "line_number": 158, "usage_type": "call"}, {"api_name": "twisted.internet.reactor.callLater", "line_number": 177, "usage_type": "call"}, {"api_name": "twisted.internet.reactor", "line_number": 177, "usage_type": "name"}, {"api_name": "twisted.internet.reactor.callLater", "line_number": 183, "usage_type": "call"}, {"api_name": "twisted.internet.reactor", "line_number": 183, "usage_type": "name"}, {"api_name": "twisted.internet.defer.maybeDeferred", "line_number": 189, "usage_type": "call"}, {"api_name": "twisted.internet.defer", "line_number": 189, "usage_type": "name"}, {"api_name": "twisted.internet.error", "line_number": 190, "usage_type": "argument"}, {"api_name": "twisted.internet.reactor.callLater", "line_number": 206, "usage_type": "call"}, {"api_name": "twisted.internet.reactor", "line_number": 206, "usage_type": "name"}, {"api_name": "twisted.internet.defer.Deferred", "line_number": 226, "usage_type": "call"}, {"api_name": "twisted.internet.defer", "line_number": 226, "usage_type": "name"}]}
+{"seq_id": "410136424", "text": "#--*coding: utf8*--\nimport pymongo\nimport hashlib\nimport uuid\nimport time\nimport random\nimport yaml\nfrom bson.objectid import ObjectId\n\n\n# local database\n# client = pymongo.MongoClient('mongodb://localhost:27017/')\n# conn = client.iubcoder\n\n# remote database\nwith open(\"config.yml\", \"r\") as f:\n doc = yaml.load(f)\n MONGODB_URI = doc[\"database\"][\"path\"]\n\nclient = pymongo.MongoClient(MONGODB_URI) # database connection\nconn = client.get_default_database()\n\n\ndef check_username(username):\n '''\n Check if the username has been used\n '''\n user_db = conn[\"users\"]\n user = user_db.find_one({\"username\": username})\n if user:\n return True\n else:\n return False\n\n\ndef check_email(email):\n '''\n Check if the email has been used\n '''\n user_db = conn[\"users\"]\n user = user_db.find_one({\"email\": email})\n if user:\n return True\n else:\n return False\n\n\ndef insert_user(username, password, salt, email):\n '''\n Inser a new user into database.\n '''\n user_db = conn[\"users\"] # user database\n # register time\n reg_time = time.time()\n # generate verify code\n code = username + email + str(reg_time)\n m2 = hashlib.md5(code.encode('utf8'))\n md5code = m2.hexdigest() + '.' + str(random.randint(11111111, 99999999))\n # generate cookie, _id is a ObjectId type, which cannot be used as cookie.\n gen_cookie = username + email\n cookie = hashlib.md5(gen_cookie.encode('utf8')).hexdigest()\n\n new_user = {\n \"reg_time\": reg_time,\n \"username\": username,\n \"password\": password,\n \"email\": email,\n \"salt\": salt,\n \"verified\": False,\n \"verify_code\": md5code,\n \"cookie\": cookie,\n \"user_info\": {},\n \"posts\": [],\n \"comments\": []\n }\n # insert new user\n user_db.insert(new_user)\n return new_user\n\n\ndef is_verified(cookie):\n '''\n Check the given user is verified.\n '''\n user_db = conn[\"users\"]\n user = user_db.find_one( {\"cookie\": cookie} )\n return user[\"verified\"]\n\n\ndef get_name_by_cookie(cookie):\n '''\n Return user name of the given user_id.\n '''\n if cookie:\n user_db = conn[\"users\"]\n user = user_db.find_one( {\"cookie\": cookie} )\n if user:\n return user[\"username\"]\n else:\n return None\n else:\n return None\n\n\ndef get_user_by_cookie(cookie):\n '''\n Return user by given cookie\n '''\n if cookie:\n user_db = conn[\"users\"]\n user = user_db.find_one( {\"cookie\": cookie} )\n return user\n else:\n return None\n\n\ndef get_user_by_username(username):\n '''\n Check if user exists\n '''\n user_db = conn[\"users\"]\n user = user_db.find_one( {\"username\": username} )\n if user:\n return user\n else:\n return None\n\n\ndef get_user_by_email(email):\n '''\n Return user according to the given email\n '''\n user_db = conn[\"users\"]\n user = user_db.find_one( {\"email\": email} )\n return user\n\ndef verify_account(email, verify_code):\n '''\n Verify user account\n '''\n user_db = conn[\"users\"]\n user = user_db.find_one( {\"email\": email} )\n if user:\n if user[\"verify_code\"] == verify_code:\n user[\"verified\"] = True\n user_db.save(user)\n return True, \"\"\n else:\n return False, \"验证码不正确\"\n else:\n return False, \"用户不存在\"\n\n\ndef insert_post(title, content, author, post_time):\n '''\n Insert a new post, and update user's 'posts'.\n '''\n post_db = conn[\"posts\"]\n user_db = conn[\"users\"] # also need to update user\n post_num = gen_new_post_num() # generate a post num(id)\n post_id = post_db.insert({\n \"post_num\": post_num,\n \"title\": title,\n \"content\": content,\n \"author\": author,\n \"post_time\": post_time,\n \"last_modified\": post_time,\n \"comments\": []\n })\n # update user's post\n user = user_db.find_one( {\"username\": author} )\n user[\"posts\"].append(ObjectId(post_id))\n user_db.save(user)\n\n\ndef fetch_all_posts():\n '''\n Get all posts from database, in reversed order of last_modified\n '''\n post_db = conn[\"posts\"]\n all_posts = post_db.find().sort( [(\"last_modified\", -1)] )\n return all_posts\n\n\ndef gen_new_post_num():\n '''\n Check the number of latest post in database, then generate a new number \n by adding one\n '''\n post_db = conn[\"posts\"]\n all_posts = post_db.find()\n if all_posts.count():\n latest_post = max(all_posts, key=lambda item: item[\"post_time\"])\n new_num = latest_post[\"post_num\"] + 1\n else:\n new_num = 0\n return new_num\n\n\ndef fetch_post_by_num(post_num):\n '''\n Return post according to given number\n '''\n post_db = conn[\"posts\"]\n post = post_db.find_one({\"post_num\": int(post_num)})\n return post\n\n\ndef insert_comment(post, username, content):\n '''\n Insert a comment into a post\n '''\n # update on posts collection\n post_db = conn[\"posts\"]\n comment_time = time.time()\n post[\"comments\"].append({\n \"content\": content,\n \"post_time\": comment_time,\n \"author\": username\n })\n post[\"last_modified\"] = comment_time\n post_db.save(post)\n # update on user collection\n user_db = conn[\"users\"]\n user = get_user_by_username(username)\n user[\"comments\"].append({\n \"post_num\": post[\"post_num\"],\n \"content\": content,\n \"last_modified\": comment_time\n })\n user_db.save(user)\n\n\ndef update_post(post):\n '''\n Save updated post into database\n '''\n post_db = conn[\"posts\"]\n post_db.save(post)\n\n\ndef update_user(user):\n '''\n Save updated user info into database\n '''\n user_db = conn[\"users\"]\n user_db.save(user)\n", "sub_path": "db_operations.py", "file_name": "db_operations.py", "file_ext": "py", "file_size_in_byte": 5799, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "yaml.load", "line_number": 17, "usage_type": "call"}, {"api_name": "pymongo.MongoClient", "line_number": 20, "usage_type": "call"}, {"api_name": "time.time", "line_number": 54, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 57, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 58, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 61, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 172, "usage_type": "call"}, {"api_name": "time.time", "line_number": 215, "usage_type": "call"}]}
+{"seq_id": "500194789", "text": "import sqlite3\n\n# connect with the myTable database\nconnection = sqlite3.connect(\"XE\")\n\n# cursor object\ncrsr = connection.cursor()\nsql_command = \"\"\"INSERT INTO employee VALUES (10, \"Abeer\", \"Gates\", \"M\", \"1980-10-28\");\"\"\"\n# execute the command to fetch all the data from the table emp\ncrsr.execute(sql_command)\ncrsr.execute(\"SELECT * FROM employee\")\n\nans = crsr.fetchall()\n\n# loop to print all the data\nfor i in ans:\n print(i)", "sub_path": "pandas_numpy/database_2.py", "file_name": "database_2.py", "file_ext": "py", "file_size_in_byte": 429, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sqlite3.connect", "line_number": 4, "usage_type": "call"}]}
+{"seq_id": "206797544", "text": "import json\nfrom bs4 import BeautifulSoup\nfrom task5 import get_movie_list_details\n\nwith open('task5.json','r') as f:\n a=json.load(f)\nmovies_data=a\nprint(movies_data)\ndef analyse_movies_directors(movies_data):\n l1=[]\n for i in movies_data:\n if i['Director'] not in l1:\n l1.append(i[\"Director\"])\n i=0\n L=[] \n while i= 200 and response.status < 300:\n return True\n else:\n try:\n _ = response.dict\n raise virtmedia_exception.VirtmediaOperationError(\"Response status is %d, %s\"% (response.status, response.dict[\"error\"][\"@Message.ExtendedInfo\"][0][\"MessageId\"].split(\".\")))\n except Exception:\n raise virtmedia_exception.VirtmediaOperationError(\"Response status is not 200, %s\"% response)\n\n def _check_supported_idrac_version(self, connection):\n response = connection.get('%s/VirtualMedia/CD'%self.idrac_location)\n self._check_success(response)\n data = response.dict\n for i in data.get('Actions', []):\n if i == \"#VirtualMedia.InsertMedia\" or i == \"#VirtualMedia.EjectMedia\":\n return True\n raise virtmedia_exception.VirtmediaOperationError(\"Unsupported version of iDRAC, please update before continuing\")\n\n def _get_virtual_media_devices(self, connection):\n idr = connection.get(\"%s\" % self.idrac_location)\n self._check_success(idr)\n try:\n virtual_media = connection.get(idr.dict[\"VirtualMedia\"][\"@odata.id\"])\n self._check_success(virtual_media)\n except KeyError:\n self.log.error(\"Cannot find a single virtual media device\")\n raise virtmedia_exception.VirtmediaOperationError(\"Cannot find any virtual media device on the server\")\n return virtual_media.dict[\"Members\"]\n\n def _umount_virtual_device(self, connection, media_uri):\n self.log.debug(\"Unmount\")\n unmount_location = media_uri + \"/Actions/VirtualMedia.EjectMedia\"\n resp = connection.post(unmount_location, body={})\n self._check_success(resp)\n\n def _mount_virtual_device(self, connection, media_uri, image_location):\n self.log.debug(\"Mount\")\n mount_location = media_uri + \"/Actions/VirtualMedia.InsertMedia\"\n payload = {'Image': image_location, 'Inserted':True, 'WriteProtected':True}\n resp = connection.post(mount_location, body=payload)\n self._check_success(resp)\n\n def _unmount_all(self, connection):\n medias = self._get_virtual_media_devices(connection)\n for media in medias:\n uri = media.get(\"@odata.id\", None)\n if not uri or connection.get(uri).dict[\"ConnectedVia\"] == \"NotConnected\":\n continue\n self._umount_virtual_device(connection, uri)\n\n def _find_first_media(self, connection, typeinfo):\n medias = self._get_virtual_media_devices(connection)\n for media in medias:\n response = connection.get(media[\"@odata.id\"])\n if typeinfo in response.dict[\"MediaTypes\"]:\n return media[\"@odata.id\"]\n return None\n\n def _mount_virtual_cd(self, connection, image_location):\n self._unmount_all(connection)\n self.log.debug(\"Mount\")\n media_uri = self._find_first_media(connection, \"DVD\")\n self._mount_virtual_device(connection, media_uri, image_location)\n\n def attach_virtual_cd(self, image_filename, driver_info, task):\n connection = None\n try:\n self.log.debug(\"attach_virtual_cd\")\n connection = self._init_connection(driver_info)\n self._check_supported_idrac_version(connection)\n image_location = 'http://' + str(driver_info['provisioning_server']) + ':' + str(driver_info['provisioning_server_http_port']) + self.remote_share + image_filename\n self._mount_virtual_cd(connection, image_location)\n\n connection.logout()\n return True\n except Exception:\n if connection:\n connection.logout()\n raise\n\n def detach_virtual_cd(self, driver_info, task):\n connection = None\n try:\n self.log.debug(\"detach_virtual_cd\")\n connection = self._init_connection(driver_info)\n self._check_supported_idrac_version(connection)\n self._unmount_all(connection)\n connection.logout()\n return True\n except Exception:\n if connection:\n connection.logout()\n raise\n\n def set_boot_device(self, task):\n try:\n #BMC boot flag valid bit clearing 1f -> all bit set\n #P 420 of ipmi spec\n # https://www.intel.com/content/www/us/en/servers/ipmi/ipmi-second-gen-interface-spec-v2-rev1-1.html\n cmd = '0x00 0x08 0x03 0x1f'\n ipmitool.send_raw(task, cmd)\n self.log.info('Disable timeout for booting')\n except Exception as err:\n self.log.warning('Failed to disable booting options: %s', str(err))\n #For time being lets do the boot order with ipmitool since, well dell doesn't provide open support\n #for this.\n try:\n # 0x00 0x08 0x05 0x80 0x20: chassis|set|bootdev|for next boot only|remote CD\n # other options for device (per ipmitool's \"ipmi_chassis.c\"):\n # 04: PXE\n # 08: HDD\n # 0c: Safe\n # 10: Diag\n # 14: CDROM\n # 18: Setup\n # 1c: Remote FDD\n # 24: Remote primary media\n # 2c: Remote HDD\n # 3c: FDD\n ipmitool.send_raw(task, '0x00 0x08 0x05 0x80 0x20 0x00 0x00 0x00')\n self.log.info('Set next boot to remote media')\n except Exception as err:\n self.log.warning('Failed to set next boot to remote media: %s', str(err))\n", "sub_path": "src/ironic_virtmedia_driver/vendors/dell/dell.py", "file_name": "dell.py", "file_ext": "py", "file_size_in_byte": 7746, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "ironic_virtmedia_driver.vendors.ironic_virtmedia_hw.IronicVirtMediaHW", "line_number": 26, "usage_type": "name"}, {"api_name": "redfish.redfish_client", "line_number": 40, "usage_type": "call"}, {"api_name": "redfish.AuthMethod.SESSION", "line_number": 42, "usage_type": "attribute"}, {"api_name": "redfish.AuthMethod", "line_number": 42, "usage_type": "name"}, {"api_name": "redfish.rest.v1.ServerDownOrUnreachableError", "line_number": 43, "usage_type": "name"}, {"api_name": "ironic.common.i18n._", "line_number": 44, "usage_type": "call"}, {"api_name": "ironic_virtmedia_driver.virtmedia_exception.VirtmediaOperationError", "line_number": 45, "usage_type": "call"}, {"api_name": "ironic_virtmedia_driver.virtmedia_exception", "line_number": 45, "usage_type": "name"}, {"api_name": "ironic.common.i18n._", "line_number": 48, "usage_type": "call"}, {"api_name": "ironic_virtmedia_driver.virtmedia_exception.VirtmediaOperationError", "line_number": 49, "usage_type": "call"}, {"api_name": "ironic_virtmedia_driver.virtmedia_exception", "line_number": 49, "usage_type": "name"}, {"api_name": "ironic.common.i18n._", "line_number": 59, "usage_type": "name"}, {"api_name": "ironic_virtmedia_driver.virtmedia_exception.VirtmediaOperationError", "line_number": 60, "usage_type": "call"}, {"api_name": "ironic_virtmedia_driver.virtmedia_exception", "line_number": 60, "usage_type": "name"}, {"api_name": "ironic_virtmedia_driver.virtmedia_exception.VirtmediaOperationError", "line_number": 62, "usage_type": "call"}, {"api_name": "ironic_virtmedia_driver.virtmedia_exception", "line_number": 62, "usage_type": "name"}, {"api_name": "ironic_virtmedia_driver.virtmedia_exception.VirtmediaOperationError", "line_number": 71, "usage_type": "call"}, {"api_name": "ironic_virtmedia_driver.virtmedia_exception", "line_number": 71, "usage_type": "name"}, {"api_name": "ironic_virtmedia_driver.virtmedia_exception.VirtmediaOperationError", "line_number": 81, "usage_type": "call"}, {"api_name": "ironic_virtmedia_driver.virtmedia_exception", "line_number": 81, "usage_type": "name"}, {"api_name": "ironic.drivers.modules.ipmitool.send_raw", "line_number": 155, "usage_type": "call"}, {"api_name": "ironic.drivers.modules.ipmitool", "line_number": 155, "usage_type": "name"}, {"api_name": "ironic.drivers.modules.ipmitool.send_raw", "line_number": 174, "usage_type": "call"}, {"api_name": "ironic.drivers.modules.ipmitool", "line_number": 174, "usage_type": "name"}]}
+{"seq_id": "136888270", "text": "\"\"\"\n FileParser.py - implementation of input file parser\n Using XML version of input data would be probably easier, but you need\n to find it prior to writing vast and complicated text parser\n\"\"\"\nfrom typing import Any, Dict, List, Tuple\n\n\ndef parse(file_name: str) -> (List[Dict[str, Any]], List[Dict[str, Any]], List[Dict[str, Any]], List[Dict[str, Any]]):\n \"\"\"\n Parse text file with network description\n\n @param file_name - path to input text file\n returns dicts - nodes, links, demands, paths\n \"\"\"\n with open(file_name, 'r') as f:\n text = f.read().split()\n\n nodes = []\n links = []\n demands = []\n paths = []\n\n i = 0\n while i < len(text):\n if text[i] == 'LINKS':\n i += 2 # skip open bracket\n while text[i] != ')':\n name = text[i]\n source = text[i + 2]\n target = text[i + 3]\n i += 10\n\n module_capacities = []\n module_costs = []\n while text[i] != ')':\n module_capacities.append(float(text[i]))\n module_costs.append(float(text[i + 1]))\n i += 2\n\n links.append({\n 'name': name,\n 'source': source,\n 'target': target,\n 'moduleCap': module_capacities,\n 'moduleCost': module_costs\n })\n i += 1 # skip link-end close bracket\n elif text[i] == 'DEMANDS':\n i += 2 # skip open bracket\n while text[i] != ')':\n name = text[i]\n source = text[i + 2]\n target = text[i + 3]\n demand_value = float(text[i + 6])\n\n max_path_length = float('inf')\n if text[i + 7] != 'UNLIMITED':\n max_path_length = float(text[i + 7])\n\n demands.append({\n 'name': name,\n 'source': source,\n 'target': target,\n 'value': demand_value,\n 'maxLen': max_path_length\n })\n i += 8\n elif text[i] == 'ADMISSIBLE_PATHS':\n i += 2 # skip open bracket\n while text[i] != ')':\n name = text[i]\n i += 2 # skip open bracket\n\n part = []\n while text[i] != ')':\n i += 2 # skip path name and open bracket\n\n path = []\n while text[i] != ')':\n path.append(text[i])\n i += 1\n\n part.append(path)\n i += 1 # skip path-end close bracket\n i += 1 # skip demand-end close bracket\n paths.append({\n 'name': name,\n 'paths': part\n })\n elif text[i] == 'NODES':\n i += 2\n while text[i] != ')':\n name = text[i]\n lon = float(text[i + 2])\n lat = float(text[i + 3])\n\n nodes.append({\n 'name': name,\n 'lon': lon,\n 'lat': lat,\n })\n i += 5\n i += 1\n\n return nodes, links, demands, paths\n\n\ndef loadSolution(file_name: str) -> Tuple[Dict[str, Any], Dict[str, Any]]:\n \"\"\"\n Load network solution provided in form of XML file. The specification of format\n can be found on SNDlib home page.\n Return the number of modules added to each link and routing for each demand\n \"\"\"\n try:\n from lxml import etree as et\n except ImportError:\n print(\"[-] Failed to load solution - lxml module is not installed !!!\")\n raise # Hard to say what should be returned in such case - raise exception anyway\n\n linksModules = {}\n demandsFlows = {}\n\n tree = et.parse(file_name)\n root = tree.getroot()\n for topGroup in root:\n if topGroup.tag == '{http://sndlib.zib.de/solution}linkConfigurations':\n for linkConfiguration in topGroup:\n linkID = linkConfiguration.attrib['linkId']\n\n if len(linkConfiguration) > 0:\n instModules = linkConfiguration[0]\n capacity = float(instModules.find('{http://sndlib.zib.de/solution}capacity').text)\n count = float(instModules.find('{http://sndlib.zib.de/solution}installCount').text)\n else:\n capacity = 0\n count = 0\n\n linksModules[linkID] = {\n 'capacity': capacity,\n 'count': count\n }\n elif topGroup.tag == '{http://sndlib.zib.de/solution}demandRoutings':\n for demandRouting in topGroup:\n demandID = demandRouting.get('demandId')\n\n flows = []\n for flowPath in demandRouting:\n value = float(flowPath.find('{http://sndlib.zib.de/solution}flowPathValue').text)\n path = []\n for link in flowPath.find('{http://sndlib.zib.de/solution}routingPath'):\n path.append(link.text)\n flows.append((value, path))\n demandsFlows[demandID] = flows\n else:\n raise KeyError(f'Unknown XML tag found in solution file - \"{topGroup.tag}\"')\n\n return linksModules, demandsFlows\n\n\ndef saveSolution(fileName: str, linksModules: Dict[str, Any], demandsFlows: Dict[str, Any]):\n \"\"\"\n Save computed solution to XML file compatible with SNDlib platform\n linksModules must be of form:\n {'LINK_0_1': {'capacity': 4.0, 'count': 2.0}, ...}\n demandsFlows must be of form:\n {'Demand_0_1': [(127.0, ['Link_1', 'Link_2', ...])], ...}\n \"\"\"\n try:\n from lxml import etree as et\n except ImportError:\n print(\"[-] Failed to save solution - lxml module is not installed !!!\")\n return\n\n root = et.Element(\"solution\")\n root.attrib['xmlns'] = 'http://sndlib.zib.de/solution'\n root.attrib['version'] = '1.0'\n\n linkConfigs = et.Element('linkConfigurations')\n for linkName in linksModules:\n linkModules = linksModules[linkName]\n\n linkConfig = et.Element('linkConfiguration')\n linkConfig.attrib['linkId'] = linkName\n\n if linkModules['count'] > 0:\n instModules = et.Element('installedModule')\n capacity = et.Element('capacity')\n count = et.Element('installCount')\n\n capacity.text = str(linkModules['capacity'])\n count.text = str(linkModules['count'])\n instModules.append(capacity)\n instModules.append(count)\n linkConfig.append(instModules)\n\n linkConfigs.append(linkConfig)\n root.append(linkConfigs)\n\n demandRoutings = et.Element('demandRoutings')\n demandRoutings.attrib['state'] = 'NOS'\n for demandName in demandsFlows:\n flows = demandsFlows[demandName]\n\n demandRouting = et.Element('demandRouting', demandId=demandName)\n for flow in flows:\n flowPath = et.Element('flowPath')\n flowPathValue = et.Element('flowPathValue')\n flowPathValue.text = str(flow[0])\n\n routingPath = et.Element('routingPath')\n for linkName in flow[1]:\n link = et.Element('linkId')\n link.text = linkName\n routingPath.append(link)\n flowPath.append(flowPathValue)\n flowPath.append(routingPath)\n\n demandRouting.append(flowPath)\n demandRoutings.append(demandRouting)\n root.append(demandRoutings)\n\n et = et.ElementTree(root)\n et.write(fileName, pretty_print=True, xml_declaration=True, encoding='UTF-8')\n", "sub_path": "src/FileParser.py", "file_name": "FileParser.py", "file_ext": "py", "file_size_in_byte": 7843, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "typing.List", "line_number": 9, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 9, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 9, "usage_type": "name"}, {"api_name": "lxml.etree.parse", "line_number": 124, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 124, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 109, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 109, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 109, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 161, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 161, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 175, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 175, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 179, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 179, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 183, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 183, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 187, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 187, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 188, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 188, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 189, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 189, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 200, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 200, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 205, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 205, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 207, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 207, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 208, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 208, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 211, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 211, "usage_type": "name"}, {"api_name": "lxml.etree.Element", "line_number": 213, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 213, "usage_type": "name"}, {"api_name": "lxml.etree", "line_number": 223, "usage_type": "name"}, {"api_name": "lxml.etree.ElementTree", "line_number": 223, "usage_type": "call"}, {"api_name": "lxml.etree.write", "line_number": 224, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 224, "usage_type": "name"}]}
+{"seq_id": "360014297", "text": "import matplotlib.pyplot as plt\nfrom matplotlib import style\nimport numpy as np\nstyle.use(\"ggplot\")\n\nclass Support_Vector_Machine:\n def __init__(self, visualization=True):\n self.visualization = visualization\n self.color = {1: 'r', -1: 'g'}\n if self.visualization:\n self.fig = plt.figure()\n self.ax = self.fig.add_subplot(1,1,1)\n\n #for training\n def fit(self, data):\n self.data = data\n\n # { |w|: [w,b]}\n opt_dict = {}\n transforms = [[1, 1], [1, -1], [-1, 1], [-1, -1]]\n all_data = []\n\n for yi in self.data:\n for featureset in self.data[yi]:\n for feature in featureset:\n all_data.append(feature)\n\n self.max_feature_value = max(all_data)\n self.min_feature_value = min(all_data)\n\n # find w, b\n step_size = [self.max_feature_value*.1,\n self.max_feature_value*.01,\n # point of expense:\n self.max_feature_value*.001]\n # extremly expensive\n b_range_multiple = 5\n b_multiple = 5\n latest_optimum = self.max_feature_value*10\n\n for step in step_size:\n w = np.array([latest_optimum, latest_optimum])\n\n # convex\n optimized = False\n while not optimized:\n for b in np.arange(-1*(self.max_feature_value*b_range_multiple), self.max_feature_value*b_range_multiple, step*b_multiple):\n for transformation in transforms:\n w_t = w*transformation\n found_option = True\n # week link in the SVM fundamentally\n # SMO attemps to fix this a bit\n # yi(xi.w+b) >= 1\n for i in self.data:\n for xi in self.data[i]:\n yi = i\n if not yi*(np.dot(w_t, xi)+b) >= 1:\n found_option = False\n if found_option:\n opt_dict[np.linalg.norm(w_t)] = [w_t, b]\n if w[0] < 0:\n optimized = True\n print(\"Optimized a step\")\n else:\n # w = [5,5]\n # step = 1\n # w-step = [4,4]\n w = w-step\n norms = sorted([n for n in opt_dict])\n opt_choice = opt_dict[norms[0]]\n self.w = opt_choice[0]\n self.b = opt_choice[1]\n\n latest_optimum = opt_choice[0][0]+step*2\n\n #for predict\n def predict(self, features):\n # sign(x.w+b)\n classification = np.sign(np.dot(np.array(features), self.w) + self.b)\n return classification\n\n\ndata_dict = {-1: np.array([[1,7], [3,4], [7,9]]), 1: np.array([[2,4],[5,1],[6,9]])}", "sub_path": "list/svm-manual.py", "file_name": "svm-manual.py", "file_ext": "py", "file_size_in_byte": 2908, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "matplotlib.style.use", "line_number": 4, "usage_type": "call"}, {"api_name": "matplotlib.style", "line_number": 4, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 11, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 60, "usage_type": "attribute"}, {"api_name": "numpy.sign", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 83, "usage_type": "call"}]}
+{"seq_id": "539035076", "text": "# https://stackoverflow.com/questions/57506101/qlabel-is-not-updated-unless-the-mainwindow-is-unfocused\n# 포커스 안맞을때 label이 변경 안되는 현상// 버그라고 하고 수정되었다고함\n# pip list 해서 버전 확인 후 업데이트 : \n\nfrom PyQt5 import QtCore, QtGui, QtWidgets\n\nclass Ui_HelloWorld(object):\n def setupUi(self, Dialog):\n Dialog.setObjectName(\"Dialog\")\n Dialog.resize(400, 300)\n self.label = QtWidgets.QLabel(Dialog)\n self.label.setGeometry(QtCore.QRect(70, 40, 201, 21))\n self.label.setObjectName(\"label\")\n self.pushButton = QtWidgets.QPushButton(Dialog)\n self.pushButton.setGeometry(QtCore.QRect(130, 90, 113, 32))\n self.pushButton.setObjectName(\"pushButton\")\n\n self.retranslateUi(Dialog)\n QtCore.QMetaObject.connectSlotsByName(Dialog)\n\n def retranslateUi(self, Dialog):\n _translate = QtCore.QCoreApplication.translate\n Dialog.setWindowTitle(_translate(\"Dialog\", \"Dialog\"))\n self.label.setText(_translate(\"Dialog\", \"foobar\"))\n self.pushButton.setText(_translate(\"Dialog\", \"Click\"))\n\n\n\nimport sys\n\nfrom PyQt5.QtCore import pyqtSlot\nfrom PyQt5.QtWidgets import QApplication\nfrom PyQt5.QtWidgets import QMainWindow \n\nclass HelloWorldGui(QMainWindow, Ui_HelloWorld):\n def __init__(self, parent=None):\n super(HelloWorldGui, self).__init__(parent)\n self.setupUi(self)\n self.pushButton.clicked.connect(self.setTextHelloWorld)\n\n def setTextHelloWorld(self):\n self.label.setText(\"Hello World\")\n\n\nif __name__ == '__main__':\n argvs = sys.argv\n app = QApplication(argvs)\n hello_world_gui = HelloWorldGui()\n hello_world_gui.show()\n sys.exit(app.exec_())", "sub_path": "200514-01-qlabel_repaint_bug.py", "file_name": "200514-01-qlabel_repaint_bug.py", "file_ext": "py", "file_size_in_byte": 1734, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 11, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 11, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 12, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 12, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 14, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 14, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRect", "line_number": 15, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 15, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QMetaObject.connectSlotsByName", "line_number": 19, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QMetaObject", "line_number": 19, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 19, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QCoreApplication", "line_number": 22, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 22, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 35, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 46, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 47, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 50, "usage_type": "call"}]}
+{"seq_id": "499760372", "text": "from collections import OrderedDict\nimport six\n\nfrom fitsblender import blendheaders\n\nfrom .. import datamodels\nfrom ..datamodels import schema\nfrom ..datamodels import fits_support\n\nimport logging\nlog = logging.getLogger(__name__)\nlog.setLevel(logging.DEBUG)\n\n\ndef blendfitsdata(input_list, output_model):\n \"\"\"\n Primary interface for JWST pipeline use of fitsblender.blendheaders\n\n This function will update the output_model datamodel with the blended metadata\n from the list of FITS objects generated from the input_list of filenames.\n \"\"\"\n new_hdrs, new_table = blendheaders.get_blended_headers(input_list)\n \n # Now merge the keyword values from new_hdrs into the metatdata for the\n # output datamodel\n # \n # start by building dict which maps all FITS keywords in schema to their \n # attribute in the schema\n fits_dict = schema.build_fits_dict(output_model.schema)\n # Now assign values from new_hdrs to output_model.meta using fits_dict map\n for hdr in new_hdrs:\n for kw in hdr:\n if kw in fits_dict:\n output_model[fits_dict[kw]] = hdr[kw]\n\n # Now, append HDRTAB as new element in datamodel\n new_schema = build_tab_schema(new_table)\n output_model.add_schema_entry('hdrtab', new_schema)\n output_model.hdrtab = fits_support.from_fits_hdu(new_table, new_schema)\n \n\ndef blendmetadata(input_models, output_model):\n final_rules = build_meta_rules(input_models)\n\n # Apply rules to each set of input headers\n new_headers = []\n i=0\n # apply rules to PRIMARY headers separately, since there is only\n # 1 PRIMARY header per image, yet many extension headers\n newphdr,newtab = final_rules.apply(phdrlist)\n final_rules.add_rules_kws(newphdr) # Adds HISTORY comments on rules used\n new_headers.append(newphdr)\n for hdrs in hdrlist[1:]:\n newhdr, newtab = final_rules.apply(hdrs)\n new_headers.append(newhdr)\n\n # create list of combined PRIMARY/SCI headers for use in creating\n # the new table extensions\n tabhdrs = []\n for phdr, scihdr in zip(hdrlist[0], hdrlist[1]):\n tabhdrs.append(cat_headers(phdr, scihdr))\n # Create extension table from list of all combined PRI/SCI headers\n tabhdr, newtab = final_rules.apply(tabhdrs)\n\n # Now merge the keyword values from new_hdrs into the metatdata for the\n # output datamodel\n #\n # start by building dict which maps all FITS keywords in schema to their\n # attribute in the schema\n fits_dict = schema.build_fits_dict(output_model.schema)\n # Now assign values from new_hdrs to output_model.meta using fits_dict map\n for hdr in new_hdrs:\n for kw in hdr:\n if kw in fits_dict:\n output_model[fits_dict[kw]] = hdr[kw]\n\n # Now, append HDRTAB as new element in datamodel\n new_schema = build_tab_schema(new_table)\n output_model.add_schema_entry('hdrtab', new_schema)\n output_model.hdrtab = fits_support.from_fits_hdu(new_table, new_schema)\n\n\ndef build_meta_rules(input_models, rules_file=None):\n # Determine what blending rules need to be merged to create the final\n # blended headers. There will be a separate set of rules for each\n # instrument, and all rules get merged into a composite set of rules that\n # get applied to all input headers regardless of instrument.\n #\n # Instrument identification will be extracted from the INSTRUME keyword\n # from the PRIMARY header of each input\n\n icache = {}\n for model in input_models:\n inst = model.meta.instrument.name.lower()\n tel = model.meta.telescope.lower()\n if inst not in icache:\n # initialize the appropriate class for this data's instrument\n inst_class = blendheaders.KeywordRules(inst, telescope=tel,\n rules_file=rules_file)\n log.debug(\"Found blendheaders RULEFILE for {}/{} of: {}\".format(\n tel, inst, inst_class.rules_file))\n # Interpret rules for this class based on image that\n # initialized this instrument's rules\n inst_class.interpret_rules(model.meta)\n # Now add this interpreted class to the cache\n icache[inst] = inst_class\n\n # Create final merged set of rules\n final_rules = None\n for inst in icache:\n if final_rules is None:\n final_rules = icache[inst]\n else:\n final_rules.merge(icache[inst])\n\n return final_rules\n\n\ndef build_tab_schema(new_table):\n \"\"\"\n Return new schema definition that describes the input table.\n \n \"\"\"\n hdrtab = OrderedDict()\n hdrtab['title']='Combined header table'\n hdrtab['fits_hdu'] = 'HDRTAB'\n datatype = []\n for col in new_table.columns:\n cname = col.name\n ctype = convert_dtype(str(col.dtype))\n c = OrderedDict()\n c['name'] = cname\n c['datatype'] = ctype\n datatype.append(c)\n hdrtab['datatype']=datatype\n \n return hdrtab\n\n\ndef convert_dtype(value):\n \"\"\"\n Convert numarray column dtype into YAML-compatible format description\n \"\"\"\n if 'S' in value:\n # working with a string description\n str_len = int(value[value.find('S')+1:])\n new_dtype = [u'ascii', str_len] ## CHANGED\n else:\n new_dtype = unicode(str(value))\n\n return new_dtype\n", "sub_path": "jwst/resample/blend.py", "file_name": "blend.py", "file_ext": "py", "file_size_in_byte": 5316, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 12, "usage_type": "attribute"}, {"api_name": "fitsblender.blendheaders.get_blended_headers", "line_number": 22, "usage_type": "call"}, {"api_name": "fitsblender.blendheaders", "line_number": 22, "usage_type": "name"}, {"api_name": "datamodels.schema.build_fits_dict", "line_number": 29, "usage_type": "call"}, {"api_name": "datamodels.schema", "line_number": 29, "usage_type": "name"}, {"api_name": "datamodels.fits_support.from_fits_hdu", "line_number": 39, "usage_type": "call"}, {"api_name": "datamodels.fits_support", "line_number": 39, "usage_type": "name"}, {"api_name": "datamodels.schema.build_fits_dict", "line_number": 70, "usage_type": "call"}, {"api_name": "datamodels.schema", "line_number": 70, "usage_type": "name"}, {"api_name": "datamodels.fits_support.from_fits_hdu", "line_number": 80, "usage_type": "call"}, {"api_name": "datamodels.fits_support", "line_number": 80, "usage_type": "name"}, {"api_name": "fitsblender.blendheaders.KeywordRules", "line_number": 98, "usage_type": "call"}, {"api_name": "fitsblender.blendheaders", "line_number": 98, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 124, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 131, "usage_type": "call"}]}
+{"seq_id": "302974618", "text": "#rdkit imports\nimport rdkit\nfrom rdkit import Chem\nfrom rdkit.Chem import Draw\nfrom rdkit.Chem.EState import Fingerprinter\nfrom rdkit.Chem import Descriptors\nfrom rdkit.Chem import rdFMCS\nfrom rdkit.Chem.rdmolops import RDKFingerprint\nfrom rdkit.Chem.Fingerprints import FingerprintMols\nfrom rdkit import DataStructs\nfrom rdkit.Avalon.pyAvalonTools import GetAvalonFP\n\n#housekeeping imports\nimport pandas as pd\nimport matplotlib\nimport numpy as np\nimport scipy as sp\n\n\ndef input_data(input_df): #cleans input df and returns neccessary elements\n '''From the input dataframe, removes rows that do not contain product\n SMILES strings. Returns the cleaned dataframe'''\n for index, row in input_df.iterrows():\n\n if row['SMILES'] == 'none':\n\n input_df.drop(index, inplace=True)\n\n return input_df\n\ndef fingerprint_products(input_df): #fingerprints all products in a given df\n '''From the input dataframe, makes a list of rdkit Mol objects and makes a\n list of rdkit fingerprints generated from those Mol objects. Inserts both\n lists as new columns and returns the expanded dataframe.'''\n mol_list = []\n fp_list = []\n\n for index, row in input_df.iterrows():\n mol_list.append(Chem.rdmolfiles.MolFromSmiles(row['SMILES'])) #get mols from SMILES and add mols to list\n fp_list.append(FingerprintMols.FingerprintMol(Chem.rdmolfiles.MolFromSmiles(row['SMILES']))) #get fingerprints from mols and and fingerprints to list\n\n input_df['Mol'] = mol_list\n input_df['Fingerprint'] = fp_list\n\n return input_df\n\n# def split_by_enzyme(input_df):\n# '''From the input dataframe, makes a set of unique enzmyes from the KEGG\n# entry column. For each unique enzyme, makes an enzyme dataframe and fills it\n# with all products in the input dataframe that are made by the unique enzyme.\n# After filling the enzyme dataframe, adds it to a list of enzyme dataframes.\n# Returns the list of unique enzyme dataframes.'''\n# unique_enzymes = set(input_df['entry'].unique())\n#\n# enzyme_df_list = []\n#\n# for entry in unique_enzymes: #for each unique enzyme in the input dataframe...\n#\n# enzyme_df = pd.DataFrame(columns=input_df.columns) #...initialize a new dataframe with the same columns as the input dataframe...\n#\n# for index, row in input_df.iterrows(): #...iterate through the input dataframe...\n#\n# if row['entry'] == entry: #... and add product rows that correspond to the unique enzyme entry...\n# enzyme_df.loc[index] = row\n#\n# enzyme_df_list.append(enzyme_df) #...then add the completed dataframe of unique enzyme products to a list\n#\n# return enzyme_df_list #return list of dataframes\n\ndef sim_i_j(row_i, row_j):\n \"\"\"For two given rows of a dataframe, use the rdkit fingerprints to compute\n TanimotoSimilarity and return the resulting float\"\"\"\n return DataStructs.FingerprintSimilarity(row_i['Fingerprint'], row_j['Fingerprint'], metric=DataStructs.TanimotoSimilarity)\n\ndef sim_i_all(input_df, index_i, row_i, metric):\n \"\"\"From the input dataframe, check the passed indexes against the DataFrame,\n and construct a new dataframe which is the similarity matrix of all of the\n products contained in the dataframe.\"\"\"\n for index_j, row_j in input_df.iterrows():\n if index_j < index_i: #skip redundant rows\n continue\n elif index_i == index_j: #autocorrelate rows\n metric.loc[index_i, index_j] = 1\n else:\n metric.loc[index_i, index_j] = sim_i_j(row_i, row_j) #fill matrix with calculated similarity at two positions at once\n metric.loc[index_j, index_i] = metric.loc[index_i, index_j]\n return\n\ndef sim_metric(input_df):\n \"\"\"From an input_df, use sim_i_j and sim_i_all to build and return a\n similarity matrix dataframe.\"\"\"\n metric = pd.DataFrame()\n for index_i, row_i in input_df.iterrows():\n sim_i_all(input_df, index_i, row_i, metric)\n return metric\n\ndef calculate_dist(input_df):\n '''Main method, takes an input dataframe and builds and returns a master\n dataframe which is the original dataframe, with three additional columns,\n an rdkit Mol column, an rdkit Fingerprint column, and a column which\n describes the average distance of a product row to all the products of the\n associated enzyme entry. Requires the KEGG enzyme entry column to be named 'entry'\n\tand the SMILES string column to be named 'SMILES' '''\n\n master_df = fingerprint_products(input_data(input_df)) #expand input df: generate mols from SMILES then generate fingerprints from mols, adding columns for each\n\n # enzyme_df_list = split_by_enzyme(input_df) #split expanded df by rows, grouped by enzyme entry (1.1.1.110 etc), into a list of dataframes\n unique_enzymes = set(master_df['entry'].unique()) # create set of unique enzymes\n\n dist_lookup = {} # initialize master dist list\n\n for enzyme in unique_enzymes: #loop through list of enzyme dataframes\n\n # enzyme_df['Dist'] = '' #initialize distance column\n enzyme_df = master_df[master_df['entry'] == enzyme]\n\n metric = sim_metric(enzyme_df) #get similarity matrix dataframe\n\n vals = metric.values #use np array of similarity matrix\n\n start_at = 1 #skip autocorrelation\n\n dist_list =[] #initialize list\n\n for i in range(len(vals)-1): #row of matrix except for last row\n\n for j in range(start_at, len(vals)): #col of matrix skipping first column\n\n dist_list.append(vals[i][j]) #add distance value to list\n\n start_at += 1 #start at higher index to skip redundancy\n\n avg_dist = sum(dist_list)/len(dist_list) #compute average distance\n dist_lookup[enzyme] = avg_dist\n # for _, row in enzyme_df.iterrows(): #loop through enzyme dataframe\n # # enzyme_df['Dist'].loc[index] = avg_dist #add averaged distance to each product row of enzyme dataframe\n\n master_df['dist'] = [dist_lookup[row['entry']] for _, row in master_df.iterrows()]\n\n return master_df\n", "sub_path": "deprecated/code/mol_sim_copy.py", "file_name": "mol_sim_copy.py", "file_ext": "py", "file_size_in_byte": 6071, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "rdkit.Chem.rdmolfiles.MolFromSmiles", "line_number": 39, "usage_type": "call"}, {"api_name": "rdkit.Chem.rdmolfiles", "line_number": 39, "usage_type": "attribute"}, {"api_name": "rdkit.Chem", "line_number": 39, "usage_type": "name"}, {"api_name": "rdkit.Chem.Fingerprints.FingerprintMols.FingerprintMol", "line_number": 40, "usage_type": "call"}, {"api_name": "rdkit.Chem.Fingerprints.FingerprintMols", "line_number": 40, "usage_type": "name"}, {"api_name": "rdkit.Chem.rdmolfiles.MolFromSmiles", "line_number": 40, "usage_type": "call"}, {"api_name": "rdkit.Chem.rdmolfiles", "line_number": 40, "usage_type": "attribute"}, {"api_name": "rdkit.Chem", "line_number": 40, "usage_type": "name"}, {"api_name": "rdkit.DataStructs.FingerprintSimilarity", "line_number": 73, "usage_type": "call"}, {"api_name": "rdkit.DataStructs", "line_number": 73, "usage_type": "name"}, {"api_name": "rdkit.DataStructs.TanimotoSimilarity", "line_number": 73, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 92, "usage_type": "call"}]}
+{"seq_id": "146517901", "text": "import torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nfrom torch.nn import init\n\nclass conv_block(nn.Module):\n def __init__(self, ch_in, ch_out, k_size=3):\n super(conv_block,self).__init__()\n self.conv = nn.Sequential(\n nn.Conv2d(ch_in, ch_out, kernel_size=k_size,padding=k_size // 2,bias=False),\n nn.GroupNorm(ch_out//16, ch_out),\n nn.ReLU(inplace=True),\n nn.Conv2d(ch_out, ch_out, kernel_size=3,stride=1,padding=1,bias=False),\n nn.GroupNorm(ch_out//16, ch_out)\n )\n\n self.ident = nn.Sequential(\n nn.Conv2d(ch_in, ch_out, kernel_size=1,stride=1,padding=0,bias=False)\n )\n self.out = nn.Sequential(\n nn.ReLU(inplace=True)\n )\n\n\n def forward(self,x):\n res = self.conv(x)\n ident = self.ident(x)\n return self.out(res + ident)\n\nclass up_conv(nn.Module):\n def __init__(self, ch_in, ch_out):\n super(up_conv,self).__init__()\n self.up = nn.Sequential(\n nn.Upsample(scale_factor=2),\n nn.Conv2d(ch_in,ch_out,kernel_size=3,stride=1,padding=1,bias=False),\n nn.GroupNorm(ch_out//16, ch_out),\n nn.ReLU(inplace=True),\n )\n\n def forward(self,x):\n x = self.up(x)\n return x\n\nclass UNet(nn.Module):\n def __init__(self, img_ch=3, output_ch=1):\n super(UNet, self).__init__()\n \n self.Maxpool = nn.MaxPool2d(kernel_size=2,stride=2)\n \n ch_num = [128, 128, 128, 128, 128, 128, 256]\n \n self.Conv1 = conv_block(ch_in=img_ch,ch_out=ch_num[0], k_size=7)\n self.Conv2 = conv_block(ch_in=ch_num[0],ch_out=ch_num[1], k_size=5)\n self.Conv3 = conv_block(ch_in=ch_num[1],ch_out=ch_num[2])\n self.Conv4 = conv_block(ch_in=ch_num[2],ch_out=ch_num[3])\n self.Conv5 = conv_block(ch_in=ch_num[3],ch_out=ch_num[4])\n self.Conv6 = conv_block(ch_in=ch_num[4],ch_out=ch_num[5])\n self.Conv7 = conv_block(ch_in=ch_num[5],ch_out=ch_num[6])\n\n self.Up7 = up_conv(ch_in=ch_num[6],ch_out=ch_num[5])\n self.Up_conv7 = conv_block(ch_in=ch_num[5] + ch_num[5], ch_out=ch_num[5])\n\n self.Up6 = up_conv(ch_in=ch_num[5],ch_out=ch_num[4])\n self.Up_conv6 = conv_block(ch_in=ch_num[4] + ch_num[4], ch_out=ch_num[4])\n\n self.Up5 = up_conv(ch_in=ch_num[4],ch_out=ch_num[3])\n self.Up_conv5 = conv_block(ch_in=ch_num[3] + ch_num[3], ch_out=ch_num[3])\n\n self.Up4 = up_conv(ch_in=ch_num[3],ch_out=ch_num[2])\n self.Up_conv4 = conv_block(ch_in=ch_num[2] + ch_num[2], ch_out=ch_num[2])\n \n self.Up3 = up_conv(ch_in=ch_num[2],ch_out=ch_num[1])\n self.Up_conv3 = conv_block(ch_in=ch_num[1] + ch_num[1], ch_out=ch_num[1])\n \n self.Up2 = up_conv(ch_in=ch_num[1],ch_out=ch_num[0])\n self.Up_conv2 = conv_block(ch_in=ch_num[0] + ch_num[0], ch_out=ch_num[0])\n\n self.out = nn.Conv2d(ch_num[0],output_ch,kernel_size=1,stride=1,padding=0, bias=False)\n\n\n def forward(self,x):\n # encoding path\n x1 = self.Conv1(x)\n\n x2 = self.Maxpool(x1)\n x2 = self.Conv2(x2)\n \n x3 = self.Maxpool(x2)\n x3 = self.Conv3(x3)\n\n x4 = self.Maxpool(x3)\n x4 = self.Conv4(x4)\n\n x5 = self.Maxpool(x4)\n x5 = self.Conv5(x5)\n\n x6 = self.Maxpool(x5)\n x6 = self.Conv6(x6)\n\n x7 = self.Maxpool(x6)\n x7 = self.Conv7(x7)\n\n # decoding + concat path\n\n d7 = self.Up7(x7)\n d7 = torch.cat((x6,d7),dim=1)\n d7 = self.Up_conv7(d7)\n\n d6 = self.Up6(d7)\n d6 = torch.cat((x5,d6),dim=1)\n d6 = self.Up_conv6(d6)\n\n\n d5 = self.Up5(d6)\n d5 = torch.cat((x4,d5),dim=1)\n d5 = self.Up_conv5(d5)\n \n d4 = self.Up4(d5)\n d4 = torch.cat((x3,d4),dim=1)\n d4 = self.Up_conv4(d4)\n\n d3 = self.Up3(d4)\n d3 = torch.cat((x2,d3),dim=1)\n d3 = self.Up_conv3(d3)\n\n d2 = self.Up2(d3)\n d2 = torch.cat((x1,d2),dim=1)\n d2 = self.Up_conv2(d2)\n\n d1 = self.out(d2)\n\n return d1\n\n\n\n\n\nif __name__ == \"__main__\":\n import numpy as np\n \n model = UNet(img_ch=36, output_ch=12)\n \n model_parameters = filter(lambda p: p.requires_grad, model.parameters())\n params = sum([np.prod(p.size()) for p in model_parameters])\n\n print(\"# of parameters: \", params)\n \n input_x = torch.rand((2, 36, 496, 448))\n out = model(input_x)\n \n print(out.shape)", "sub_path": "runs/Moscow/UNet.py", "file_name": "UNet.py", "file_ext": "py", "file_size_in_byte": 4502, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "torch.nn.Module", "line_number": 6, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 6, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 9, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 9, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 10, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 10, "usage_type": "name"}, {"api_name": "torch.nn.GroupNorm", "line_number": 11, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 11, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 12, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 12, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 13, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 13, "usage_type": "name"}, {"api_name": "torch.nn.GroupNorm", "line_number": 14, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 14, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 17, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 18, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 21, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 30, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 30, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 33, "usage_type": "name"}, {"api_name": "torch.nn.Upsample", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 34, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.nn.GroupNorm", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 36, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 37, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 44, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 44, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 48, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 78, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 115, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 119, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 123, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.prod", "line_number": 144, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 148, "usage_type": "call"}]}
+{"seq_id": "198413750", "text": "from skimage import io\r\nfrom skimage import transform\r\nfrom skimage.color import rgb2gray\r\nimport numpy as np\r\nimport math\r\nimport os\r\n\r\n\r\ndef load_data(data_directory):\r\n \"\"\" Загружает готовые датасеты \"\"\"\r\n\r\n directories = [d for d in os.listdir(data_directory)\r\n if os.path.isdir(os.path.join(data_directory, d))]\r\n labels = []\r\n images = []\r\n for d in directories:\r\n label_directory = os.path.join(data_directory, d)\r\n file_names = [os.path.join(label_directory, f)\r\n for f in os.listdir(label_directory)\r\n if f.endswith(\".ppm\")]\r\n for f in file_names:\r\n images.append(io.imread(f))\r\n labels.append(int(d))\r\n return images, labels\r\n\r\n\r\ndef color_to_gray(images):\r\n \"\"\" Преобразует изображения в оттенки серого \"\"\"\r\n\r\n images_arr = np.array(images)\r\n return rgb2gray(images_arr)\r\n\r\n\r\ndef get_train_data():\r\n \"\"\" подготавливает данные для обучения и передачи их в НС \"\"\"\r\n\r\n ROOT_PATH = os.getcwd()\r\n train_data_directory = os.path.join(ROOT_PATH, \"Training\")\r\n\r\n train_images, train_labels = load_data(train_data_directory)\r\n\r\n # трансформация изображений до размера 28х28 пикселей и их конвертация в серый цвет\r\n train_images_transformed = [transform.resize(image, (28, 28)) for image in train_images]\r\n train_images_transformed = color_to_gray(train_images_transformed)\r\n return train_images_transformed, train_labels\r\n\r\n\r\ndef get_test_data():\r\n \"\"\" подготавливает данные для тестирования НС \"\"\"\r\n\r\n ROOT_PATH = os.getcwd()\r\n test_data_directory = os.path.join(ROOT_PATH, \"Testing\")\r\n\r\n test_images, test_labels = load_data(test_data_directory)\r\n\r\n # трансформация изображений до размера 28х28 пикселей и их конвертация в серый цвет\r\n test_images_transformed = [transform.resize(image, (28, 28)) for image in test_images]\r\n test_images_transformed = color_to_gray(test_images_transformed)\r\n return test_images_transformed, test_labels\r\n\r\n\r\n\r\n\r\n\r\n\r\n", "sub_path": "load_datasets.py", "file_name": "load_datasets.py", "file_ext": "py", "file_size_in_byte": 2305, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.listdir", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 19, "usage_type": "call"}, {"api_name": "skimage.io.imread", "line_number": 22, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 22, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 30, "usage_type": "call"}, {"api_name": "skimage.color.rgb2gray", "line_number": 31, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "skimage.transform.resize", "line_number": 43, "usage_type": "call"}, {"api_name": "skimage.transform", "line_number": 43, "usage_type": "name"}, {"api_name": "os.getcwd", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "skimage.transform.resize", "line_number": 57, "usage_type": "call"}, {"api_name": "skimage.transform", "line_number": 57, "usage_type": "name"}]}
+{"seq_id": "527588282", "text": "#! /usr/bin/env python\n# -*- coding: utf-8 -*-\n# vim:fenc=utf-8\n#\n# Copyright © Her Majesty the Queen in Right of Canada, as represented\n# by the Minister of Statistics Canada, 2019.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n#\n# Example for building a complete Artemis job\n\n# Tools\nfrom artemis.tools.csvtool import CsvTool\nfrom artemis.tools.filtercoltool import FilterColTool\nfrom artemis.tools.tdigesttool import TDigestTool\nfrom artemis.tools.xlstool import XlsTool\n\n# Algorithms\nfrom artemis.algorithms.dummyalgo import DummyAlgo1\nfrom artemis.algorithms.csvparseralgo import CsvParserAlgo\nfrom artemis.algorithms.filteralgo import FilterAlgo\nfrom artemis.algorithms.profileralgo import ProfilerAlgo\n\n# Other requirements\nimport dask.delayed\nimport tempfile\nimport uuid\nimport urllib.parse\nimport logging\nimport click\nimport os\n\nfrom artemis.configurables.configurable import MenuBuilder\nfrom artemis.distributed.job_builder import runjob\nfrom artemis.generators.simutablegen import SimuTableGen\nfrom artemis.io.protobuf.configuration_pb2 import Configuration\nfrom artemis.io.protobuf.cronus_pb2 import (MenuObjectInfo, ConfigObjectInfo,\n TableObjectInfo, DatasetObjectInfo)\nfrom artemis.io.protobuf.table_pb2 import Table\nfrom artemis.io.filehandler import FileHandlerTool\nfrom artemis.io.writer import BufferOutputWriter\nfrom artemis.meta.cronus import BaseObjectStore\nfrom artemis.meta.Directed_Graph import Directed_Graph, Node\nfrom artemis.core.book import TDigestBook\n\nfrom artemis.dq.plotlytool import PlotlyTool\n# Validation/graphing code requirements\nimport numpy as np\nimport sys\nimport time\nimport matplotlib.pyplot as plt\n\nfrom artemis.externals.tdigest.tdigest import TDigest\nfrom scipy import interpolate\nfrom scipy.stats import norm\nfrom scipy import interpolate\nfrom plotly.subplots import make_subplots\nimport plotly.graph_objects as go\n# ------------------------------------------\n\nlogging.getLogger().setLevel(logging.INFO)\ndef example_configuration(table_id, seed=42):\n # First define a data generator using SimuTable\n\n max_malloc = 2147483648 # Maximum memory allowed in Arrow memory pool\n max_buffer_size = 2147483648 # Maximum size serialized ipc message\n write_csv = True # Output csv files for each arrow output file\n sample_ndatums = 1 # Preprocess job to sample files from dataset\n sample_nchunks = 10 # Preprocess job to sample chunks from a file\n linesep = '\\r\\n' # Line delimiter to scan for on csv input\n delimiter = \",\" # Field delimiter\n blocksize = 2**16 # Size of chunked data in-memory\n header = '' # Predefined header\n footer = '' # Predefined footer\n header_offset = 0 # N bytes to scan past header\n footer_size = 0 # N bytes size of footer\n schema = [] # Predefined list of field names on input\n encoding = 'utf8' # encoding\n gen_nbatches = 5 # Number of batches to generator\n gen_nrows = 1000 # Number of rows per batch\n\n config = Configuration() # Cronus Configuration message\n config.uuid = str(uuid.uuid4())\n config.name = f\"{config.uuid}.config.pb\"\n config.max_malloc_size_bytes = max_malloc\n\n generator = SimuTableGen('generator',\n nbatches=gen_nbatches,\n num_rows=gen_nrows,\n file_type=1, # Output type cronus.proto filetype\n table_id=table_id,\n seed=seed)\n\n # Set the generator configuration\n config.input.generator.config.CopyFrom(generator.to_msg())\n\n filehandler = FileHandlerTool('filehandler',\n filetype='csv', # TBD use filetype metadata\n blocksize=blocksize,\n delimiter=delimiter,\n linesep=linesep,\n header=header,\n footer=footer,\n header_offset=header_offset,\n footer_size=footer_size,\n schema=schema,\n encoding=encoding,\n seed=seed)\n # Add to the tools\n config.tools[filehandler.name].CopyFrom(filehandler.to_msg())\n\n csvtool = CsvTool('csvtool', block_size=(2 * blocksize))\n config.tools[csvtool.name].CopyFrom(csvtool.to_msg())\n\n filtercoltool = FilterColTool('filtercoltool',\n columns=['record-id', 'SIN', 'DOB'])\n config.tools[filtercoltool.name].CopyFrom(filtercoltool.to_msg())\n \n writer = BufferOutputWriter('bufferwriter',\n BUFFER_MAX_SIZE=max_buffer_size,\n write_csv=write_csv)\n config.tools[writer.name].CopyFrom(writer.to_msg())\n \n tdigesttool = TDigestTool('tdigesttool')\n config.tools[tdigesttool.name].CopyFrom(tdigesttool.to_msg())\n\n sampler = config.sampler\n sampler.ndatums = sample_ndatums\n sampler.nchunks = sample_nchunks\n\n return config\n\n\nclass ExampleMenu(MenuBuilder):\n def __init__(self, name='test'):\n super().__init__(name)\n\n def _algo_builder(self):\n '''\n define all algorithms required\n '''\n self._algos['testalgo'] = DummyAlgo1('dummy',\n myproperty='ptest',\n loglevel='WARNING')\n self._algos['csvalgo'] = CsvParserAlgo('csvparser', loglevel='WARNING')\n self._algos['filteralgo'] = FilterAlgo('filter',\n loglevel='WARNING')\n self._algos['profileralgo'] = ProfilerAlgo('profiler',\n loglevel='WARNING')\n\n def _seq_builder(self):\n # Define the sequences and node names\n self._seqs['seqX'] = Node([\"initial\"],\n ('csvparser',),\n \"seqX\")\n self._seqs['seqY'] = Node([\"seqX\"],\n ('filter',),\n \"seqY\")\n self._seqs['seqA'] = Node(['seqX'],\n ('profiler'),\n 'seqA')\n self._seqs['seqB'] = Node(['seqY'],\n ('dummy'),\n 'seqB')\n\n def _chain_builder(self):\n # Add the sequences to a chain\n self._chains['csvchain'] = Directed_Graph(\"csvchain\")\n self._chains['csvchain'].add(self._seqs['seqX'])\n self._chains['csvchain'].add(self._seqs['seqY'])\n self._chains['csvchain'].add(self._seqs['seqA'])\n self._chains['csvchain'].add(self._seqs['seqB'])\n\n@click.command()\n@click.option('--location', required = True, prompt = True, help = 'Path to .xlsx')\ndef example_job(location):\n # Artemis Job requirements\n # BaseObjectStore - name, path and id\n # Menu\n # Configuration\n # Input Dataset\n # Dataset partitions\n # Table schemas for each dataset partition\n\n # Build the Menu\n mb = ExampleMenu()\n msgmenu = mb.build()\n menuinfo = MenuObjectInfo()\n menuinfo.created.GetCurrentTime()\n\n # Read schema and generator names\n xlstool = XlsTool('xlstool', location=location)\n ds_schema = xlstool.execute(location)\n # Example job only have one table\n table = ds_schema.tables[0]\n \n # Build the Configuration\n\n # Build the partition Table schemas\n\n # Register all inputs in the Cronus object store\n\n # Build the job\n # To use the local directory:\n # dirpath = os.getcwd()\n with tempfile.TemporaryDirectory() as dirpath:\n # All jobs now require an object store\n # All outputs are pesisted in the object store path\n # See github.com/mbr/simplekv\n # Factory class for simplekv provided by\n # blueyonder/storefact\n store = BaseObjectStore(dirpath, 'artemis')\n\n # Requires registering an parent dataset\n # Generator data is written to disk with\n # The parent dataset uuid\n # Register the 'generator' partition -- required\n\n g_dataset = store.register_dataset()\n store.new_partition(g_dataset.uuid, 'generator')\n job_id = store.new_job(g_dataset.uuid)\n\n # The table schema which defines the model for the generator\n # Persisted first to the object store\n # protobuf file\n tinfo = TableObjectInfo()\n table_id = store.register_content(table,\n tinfo,\n dataset_id=g_dataset.uuid,\n job_id=job_id,\n partition_key='generator').uuid\n\n store.save_store()\n\n # Now configure all tools and algorithms\n # Includes IO tools\n config = example_configuration(table_id)\n\n # Algorithms need to added from the menu to the configuration\n for key in mb._algos:\n msg = config.algos.add()\n msg.CopyFrom(mb._algos[key].to_msg())\n\n configinfo = ConfigObjectInfo()\n configinfo.created.GetCurrentTime()\n\n # Store the menu and configuration protobufs\n menu_uuid = store.register_content(msgmenu, menuinfo).uuid\n config_uuid = store.register_content(config, configinfo).uuid\n\n # Register an output dataset\n dataset = store.register_dataset(menu_id=menu_uuid,\n config_id=config_uuid)\n #Copy metadata from xlstool\n store[dataset.uuid].dataset.aux.CopyFrom(ds_schema.dataset.aux)\n store.save_store()\n\n # Now define the actual Artemis job\n # Again the input is a protobuf\n # All other information read in from the\n # object store\n inputs = store.list(prefix=g_dataset.uuid)\n\n ds_results = []\n for _ in range(2):\n job_id = store.new_job(dataset.uuid)\n config = Configuration()\n store.get(config_uuid, config)\n for p in config.input.generator.config.properties.property:\n if p.name == 'glob':\n p.value = dirpath.split('.')[-2]+'csv'\n store._put_message(config_uuid, config)\n store.get(config_uuid, config)\n\n ds_results.append(runjob(dirpath,\n store.store_name,\n store.store_uuid,\n menu_uuid,\n config_uuid,\n dataset.uuid,\n g_dataset.uuid,\n str(job_id)))\n\n results = dask.compute(*ds_results, scheduler='single-threaded')\n store.new_partition(dataset.uuid, 'seqA')\n store.new_partition(dataset.uuid, 'seqB')\n store.save_store()\n for buf in results:\n ds = DatasetObjectInfo()\n ds.ParseFromString(buf)\n store.update_dataset(dataset.uuid, buf)\n\n store.save_store()\n \n dqtool = PlotlyTool(store=store, uuid=dataset.uuid)\n dqtool.visualize(output=\"{}/test\".format(os.getcwd()),show=True,check=False)\n\nif __name__ == '__main__':\n example_job()\n", "sub_path": "examples/distributed_dq_example.py", "file_name": "distributed_dq_example.py", "file_ext": "py", "file_size_in_byte": 11898, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.getLogger", "line_number": 72, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 72, "usage_type": "attribute"}, {"api_name": "artemis.io.protobuf.configuration_pb2.Configuration", "line_number": 93, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 94, "usage_type": "call"}, {"api_name": "artemis.generators.simutablegen.SimuTableGen", "line_number": 98, "usage_type": "call"}, {"api_name": "artemis.io.filehandler.FileHandlerTool", "line_number": 108, "usage_type": "call"}, {"api_name": "artemis.tools.csvtool.CsvTool", "line_number": 123, "usage_type": "call"}, {"api_name": "artemis.tools.filtercoltool.FilterColTool", "line_number": 126, "usage_type": "call"}, {"api_name": "artemis.io.writer.BufferOutputWriter", "line_number": 130, "usage_type": "call"}, {"api_name": "artemis.tools.tdigesttool.TDigestTool", "line_number": 135, "usage_type": "call"}, {"api_name": "artemis.configurables.configurable.MenuBuilder", "line_number": 145, "usage_type": "name"}, {"api_name": "artemis.algorithms.dummyalgo.DummyAlgo1", "line_number": 153, "usage_type": "call"}, {"api_name": "artemis.algorithms.csvparseralgo.CsvParserAlgo", "line_number": 156, "usage_type": "call"}, {"api_name": "artemis.algorithms.filteralgo.FilterAlgo", "line_number": 157, "usage_type": "call"}, {"api_name": "artemis.algorithms.profileralgo.ProfilerAlgo", "line_number": 159, "usage_type": "call"}, {"api_name": "artemis.meta.Directed_Graph.Node", "line_number": 164, "usage_type": "call"}, {"api_name": "artemis.meta.Directed_Graph.Node", "line_number": 167, "usage_type": "call"}, {"api_name": "artemis.meta.Directed_Graph.Node", "line_number": 170, "usage_type": "call"}, {"api_name": "artemis.meta.Directed_Graph.Node", "line_number": 173, "usage_type": "call"}, {"api_name": "artemis.meta.Directed_Graph.Directed_Graph", "line_number": 179, "usage_type": "call"}, {"api_name": "artemis.io.protobuf.cronus_pb2.MenuObjectInfo", "line_number": 199, "usage_type": "call"}, {"api_name": "artemis.tools.xlstool.XlsTool", "line_number": 203, "usage_type": "call"}, {"api_name": "tempfile.TemporaryDirectory", "line_number": 217, "usage_type": "call"}, {"api_name": "artemis.meta.cronus.BaseObjectStore", "line_number": 223, "usage_type": "call"}, {"api_name": "artemis.io.protobuf.cronus_pb2.TableObjectInfo", "line_number": 237, "usage_type": "call"}, {"api_name": "artemis.io.protobuf.cronus_pb2.ConfigObjectInfo", "line_number": 255, "usage_type": "call"}, {"api_name": "artemis.io.protobuf.configuration_pb2.Configuration", "line_number": 278, "usage_type": "call"}, {"api_name": "artemis.distributed.job_builder.runjob", "line_number": 286, "usage_type": "call"}, {"api_name": "dask.delayed.compute", "line_number": 295, "usage_type": "call"}, {"api_name": "dask.delayed", "line_number": 295, "usage_type": "name"}, {"api_name": "artemis.io.protobuf.cronus_pb2.DatasetObjectInfo", "line_number": 300, "usage_type": "call"}, {"api_name": "artemis.dq.plotlytool.PlotlyTool", "line_number": 306, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 307, "usage_type": "call"}, {"api_name": "click.command", "line_number": 185, "usage_type": "call"}, {"api_name": "click.option", "line_number": 186, "usage_type": "call"}]}
+{"seq_id": "372560445", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n#\n# rst2db.py\n# =========\n#\n# A reStructuredText to DocBook conversion tool, using Python's docutils\n# library.\n#\n# by Eron Hennessey\n\nfrom argparse import ArgumentParser\nfrom argparse import RawDescriptionHelpFormatter\nimport os\nimport sys\n\nfrom abstrys.docutils_ext.docbook_writer import DocBookWriter\nfrom docutils.core import publish_string\n\n\nDESCRIPTION = 'rst2db - convert reStructuredText to DocBook'\n\n\ndef printerr(error_text):\n \"\"\"Prints an error message to stderr.\"\"\"\n sys.stderr.write(\"ERROR -- %s\\n\" % error_text)\n\n\ndef process_cmd_args():\n \"\"\"Parse command-line options.\"\"\"\n parser = ArgumentParser(description=DESCRIPTION,\n formatter_class=RawDescriptionHelpFormatter)\n parser.add_argument('input_filename', metavar='INPUT',\n help='Path to input ReST file.')\n parser.add_argument('-o', '--output',\n dest='output_filename', metavar='OUTPUT',\n help='Path to output DocBook file.')\n parser.add_argument('-t', '--template',\n dest='template_filename', metavar='TEMPLATE',\n help='Path to template DocBook file.')\n parser.add_argument('-e', '--element', dest='root_element',\n default='section', metavar='ROOT',\n help='Root element of the resulting DocBook file.')\n parser.add_argument('-l', '--lang', dest='lang',\n help='Language code of the resulting DocBook file.')\n return parser.parse_args()\n\n\ndef run():\n \"\"\"The main procedure.\"\"\"\n program_name = os.path.basename(sys.argv[0])\n try:\n params = process_cmd_args()\n if not os.path.exists(params.input_filename):\n printerr(\"File doesn't exist: %s\" % params.input_filename)\n sys.exit(1)\n # get the file contents first\n input_file_contents = open(params.input_filename, 'rb').read()\n docutils_writer = None\n # set up the writer\n if params.output_filename is not None:\n # If there's an output filename, use its basename as the root\n # element's ID.\n (_, filename) = os.path.split(params.output_filename)\n (doc_id, _) = os.path.splitext(filename)\n docutils_writer = DocBookWriter(params.root_element,\n doc_id,\n lang=params.lang)\n else:\n docutils_writer = DocBookWriter(params.root_element,\n lang=params.lang)\n # get the docbook output.\n overrides = {'input_encoding': 'utf-8',\n 'output_encoding': 'utf-8'}\n docbook_contents = publish_string(input_file_contents,\n writer=docutils_writer,\n settings_overrides=overrides)\n # if there's an output file, write to that. Otherwise, write to stdout.\n if params.output_filename is None:\n output_file = sys.stdout\n else:\n output_file = open(params.output_filename, 'w+')\n \n output_file.write(docbook_contents)\n # that's it, we're done here!\n return 0\n except KeyboardInterrupt:\n ### handle keyboard interrupt ###\n return 0\n except Exception as e:\n indent = len(program_name) * ' '\n sys.stderr.write(program_name + ': ' + repr(e) + '\\n')\n sys.stderr.write(indent + ' for help use --help\\n')\n return 1\n\n\nif __name__ == \"__main__\":\n sys.exit(run())\n", "sub_path": "abstrys/cmd_rst2db.py", "file_name": "cmd_rst2db.py", "file_ext": "py", "file_size_in_byte": 3650, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sys.stderr.write", "line_number": 26, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 26, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 31, "usage_type": "call"}, {"api_name": "argparse.RawDescriptionHelpFormatter", "line_number": 32, "usage_type": "name"}, {"api_name": "os.path.basename", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 51, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path", "line_number": 65, "usage_type": "attribute"}, {"api_name": "abstrys.docutils_ext.docbook_writer.DocBookWriter", "line_number": 66, "usage_type": "call"}, {"api_name": "abstrys.docutils_ext.docbook_writer.DocBookWriter", "line_number": 70, "usage_type": "call"}, {"api_name": "docutils.core.publish_string", "line_number": 75, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 80, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 92, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 92, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 93, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 93, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 98, "usage_type": "call"}]}
+{"seq_id": "557514127", "text": "\"\"\"\n Function that implements SVM using optimization functions with the primal vs dual formulations as choices.\n\"\"\"\n\nfrom Classifier import Classifier\nimport numpy as np\nimport scipy.optimize\n\n# adding a random comment\n\nclass SVM(Classifier):\n # the constructor equivalent\n def __init__(self, solve_method=\"primal\", c=1):\n \"\"\"\n :param solve_method: either the primal or the dual formulation\n \"\"\"\n super().__init__()\n self.solve_method = solve_method\n self.c = c\n self.beta = None\n self.X = None\n self.Y = None\n\n # returns the name of this classifier\n def get_name(self):\n return \"SVM\"\n\n # a fit method is instead used to make it adhere to the\n def fit(self, x, y):\n \"\"\"\n :param x: N X p numpy.ndarray\n :param y: N X 1 numpy data array\n \"\"\"\n self.X = x\n self.Y = np.array(y)\n self.Y[y == 0] = -1\n\n if self.solve_method == \"primal\":\n self.beta = self.__primal_svm()\n elif self.solve_method == \"dual\":\n self.beta = self.__dual_svm()\n else:\n raise Exception(\"solve method not found\")\n\n # method to predict\n def predict(self, x):\n \"\"\"\n :param x: N X p data numpy.ndarray\n :return: (N, ) y predicted labels\n \"\"\"\n y = np.sign(np.dot(x, self.beta.T))\n y[y == -1] = 0\n return np.round(y)\n\n # the objective function\n @staticmethod\n def func(alpha, k):\n \"\"\" Objective function\n :param alpha: the support vectors\n :param k: the kernel\n \"\"\"\n a1 = np.ones(alpha.shape)\n return -(np.dot(a1.T, alpha) - .5 * np.dot(alpha.T, np.dot(k, alpha)))\n\n # define derivative of function\n @staticmethod\n def func_deriv(alpha, k):\n \"\"\" Derivative\n :param alpha: the support vectors\n :param k: the kernel\n \"\"\"\n return -(np.ones(alpha.shape) - 1 * np.dot(k, alpha))\n\n # the dual method\n def __dual_svm(self):\n \"\"\" Support vector machine - Dual problem\n\n SVM classification for a numeric test matrix. The\n returned result is the vector of coefficients from\n the support vector machine (beta, *not* alpha!).\n\n Returns:\n a 1d numpy array of length p giving the coefficients of beta in\n the SVM model\n \"\"\"\n c = self.c\n x = self.X\n y = self.Y\n\n # calculate the linear kernel matrix\n k = np.zeros((x.shape[0], x.shape[0]))\n for i in range(x.shape[0]):\n for j in range(x.shape[0]):\n x_1 = x[i, :]\n x_2 = x[j, :]\n k[i, j] = y[i] * y[j] * np.inner(x_1, x_2)\n\n alpha = .5 * np.ones((x.shape[0], 1))\n bnds = np.concatenate((np.zeros(np.shape(alpha)), np.ones(np.shape(alpha)) * c), axis=1)\n optim = scipy.optimize.minimize(SVM.func, alpha, args=k, jac=SVM.func_deriv,\n bounds=bnds, options={'disp': False})\n alpha_updated = optim.x\n beta = np.zeros((x.shape[1]))\n for i in range(x.shape[0]):\n beta = beta + alpha_updated[i] * y[i] * x[i, :]\n\n return beta # correct dimension\n\n # the primal method\n def __primal_svm(self):\n \"\"\" Support vector machine - Dual problem\n\n SVM classification for a numeric test matrix. The\n returned result is the vector of coefficients from\n the support vector machine (beta, *not* alpha!).\n\n Args:\n X: an n by p numpy array; the data matrix of predictors\n y: a length n numpy array; the observed response\n lam: positive numeric value giving the tuning parameter\n in the (primal, penalized format) of the support vector machine\n k: positive integer giving the number of samples selected in\n each iteration of the algorithm\n T: positive integer giving the total number of iteration to run\n\n Returns:\n a 1d numpy array of length p giving the coefficients of beta in\n the SVM model\n \"\"\"\n X = self.X\n y = self.Y\n lam = 1\n k = 5\n T = 100\n\n w = np.zeros(X.shape[1]) / (X.shape[0])\n # print(\"initial norm is: \" + str(np.linalg.norm(w)))\n # print(\"1/sqrt(lambda) is: \" + str(1/np.sqrt(lam)))\n\n for t in range(T):\n k_inds = np.random.randint(0, X.shape[0], (k, 1))\n y_k = np.reshape(y[k_inds], (k, 1))\n x_k = np.reshape(X[k_inds, :], (k, X.shape[1]))\n temp1 = np.multiply(y_k, np.reshape(np.dot(x_k, w), (k, 1)))\n\n eta_t = 1 / ((t + 1) * lam)\n w_half = (1 - eta_t * lam) * w\n # print(w_half)\n\n for i in range(temp1.shape[0]):\n if temp1[i] < 1:\n w_half = w_half + eta_t * y_k[i] * x_k[i, :] / k\n\n scale_val = np.min([1, 1 / (np.sqrt(lam) * np.linalg.norm(w_half))])\n # if np.abs(scale_val - 1) > .000001:\n # print(\"scale value is: \" + str(scale_val))\n w = w_half * scale_val\n\n return w # correct dimension\n", "sub_path": "SVM.py", "file_name": "SVM.py", "file_ext": "py", "file_size_in_byte": 5172, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "Classifier.Classifier", "line_number": 11, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.sign", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.inner", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 99, "usage_type": "call"}, {"api_name": "scipy.optimize.optimize.minimize", "line_number": 100, "usage_type": "call"}, {"api_name": "scipy.optimize.optimize", "line_number": 100, "usage_type": "attribute"}, {"api_name": "scipy.optimize", "line_number": 100, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 141, "usage_type": "attribute"}, {"api_name": "numpy.reshape", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 154, "usage_type": "attribute"}]}
+{"seq_id": "239969453", "text": "#!/usr/bin/env pyhton\n\nimport numpy as np\nfrom mpl_toolkits.mplot3d import Axes3D\nimport pylab as pl\nimport matplotlib.image as mpimg\n\ndef leeImagen(filename, factor):\n img=mpimg.imread(filename)\n k = img[:,:,0]\n for j in range(k.shape[1]):\n for i in range(k.shape[0]):\n if k[i,j] == 0.0:\n k[i,j] = 0.1\n if k[i,j] < 1.0:\n k[i,j] *= factor \n\n return k.transpose()\n\n\ndef u_face(u1, u2):\n return 0.5 * (u1 + u2)\n\ndef Laplaciano2D(Nx, Ny, r, k):\n \"\"\" Esta funcion calcula los coeficientes del \n sistema lineal producido por el operador de \n Laplace en 2D. Estos coeficientes son almacenados \n en la matriz pentadiagonal correspondiente.\"\"\"\n N = Nx * Ny\n A = np.zeros((N,N))\n\n# Primero llena los bloques tridiagonales\n for j in range(1,Ny+1):\n ofs = Nx * (j-1) \n # Primer renglón del bloque, considera BC en la pared izq.\n k1 = u_face(k[0,j ], k[1,j]) # k_(i-1/2, j)\n k2 = u_face(k[2,j ], k[1,j]) # k_(i+1/2, j)\n k3 = u_face(k[1,j+1], k[1,j]) # k_(i, j-1/2)\n k4 = u_face(k[1,j-1], k[1,j]) # k_(i, j+1/2)\n A[ofs , ofs] = r * (k1 + k2 + k3 + k4) \n A[ofs + 1, ofs] = -r * k2\n\n # Renglones intermedios del bloque \n for i in range(2,Nx):\n k1 = u_face(k[i-1,j], k[i,j]) # k_(i-1/2, j)\n k2 = u_face(k[i+1,j], k[i,j]) # k_(i+1/2, j)\n k3 = u_face(k[i,j-1], k[i,j]) # k_(i, j-1/2)\n k4 = u_face(k[i,j+1], k[i,j]) # k_(i, j+1/2)\n I = ofs + i - 1\n A[I , I] = r * (k1 + k2 + k3 + k4)\n A[I-1, I] = -r * k1\n A[I+1, I] = -r * k2\n\n # Último renglón del bloque, considera BC en la pared der.\n k1 = u_face(k[Nx-1,j ], k[Nx,j]) # k_(i-1/2, j)\n k2 = u_face(k[Nx+1,j ], k[Nx,j]) # k_(i+1/2, j)\n k3 = u_face(k[Nx ,j-1], k[Nx,j]) # k_(i, j-1/2)\n k4 = u_face(k[Nx ,j+1], k[Nx,j]) # k_(i, j+1/2)\n I = ofs + Nx - 1\n A[I-1,I] = -r * k1 \n A[I ,I] = r * (k1 + k2 + k3 + k4) \n\n \n# Despues llena las dos diagonales externas\n I = 0\n for j in range(1, Ny):\n for i in range(1,Nx+1):\n k3 = u_face(k[i,j-1+1], k[i,j+1]) # k_(i, j-1/2)\n k4 = u_face(k[i,j+1], k[i,j]) # k_(i, j+1/2)\n A[I , I+Nx] = -r * k3 # South, 3, down\n A[I+Nx, I ] = -r * k4 # North, 4, up\n I += 1\n \n return A\n\n\ndef Laplaciano2D_T(Nx, Ny, r, k):\n \"\"\" Esta funcion calcula los coeficientes del \n sistema lineal producido por el operador de \n Laplace en 2D. Estos coeficientes son almacenados \n en la matriz pentadiagonal correspondiente.\"\"\"\n N = Nx * Ny\n A = np.zeros((N,N))\n\n# Primero llena los bloques tridiagonales\n for j in range(1,Ny+1):\n ofs = Nx * (j-1) \n # Primer renglón del bloque, considera BC en la pared izq.\n k1 = u_face(k[0,j ], k[1,j]) # k_(i-1/2, j)\n k2 = u_face(k[2,j ], k[1,j]) # k_(i+1/2, j)\n k3 = u_face(k[1,j+1], k[1,j]) # k_(i, j-1/2)\n k4 = u_face(k[1,j-1], k[1,j]) # k_(i, j+1/2)\n A[ofs , ofs] = 1 + r * (k1 + k2 + k3 + k4) \n A[ofs + 1, ofs] = -r * k2\n\n # Renglones intermedios del bloque \n for i in range(2,Nx):\n k1 = u_face(k[i-1,j], k[i,j]) # k_(i-1/2, j)\n k2 = u_face(k[i+1,j], k[i,j]) # k_(i+1/2, j)\n k3 = u_face(k[i,j-1], k[i,j]) # k_(i, j-1/2)\n k4 = u_face(k[i,j+1], k[i,j]) # k_(i, j+1/2)\n I = ofs + i - 1\n A[I , I] = 1 + r * (k1 + k2 + k3 + k4)\n A[I-1, I] = -r * k1\n A[I+1, I] = -r * k2\n\n # Último renglón del bloque, considera BC en la pared der.\n k1 = u_face(k[Nx-1,j ], k[Nx,j]) # k_(i-1/2, j)\n k2 = u_face(k[Nx+1,j ], k[Nx,j]) # k_(i+1/2, j)\n k3 = u_face(k[Nx ,j-1], k[Nx,j]) # k_(i, j-1/2)\n k4 = u_face(k[Nx ,j+1], k[Nx,j]) # k_(i, j+1/2)\n I = ofs + Nx - 1\n A[I-1,I] = -r * k1 \n A[I ,I] = 1 + r * (k1 + k2 + k3 + k4) \n\n \n# Despues llena las dos diagonales externas\n I = 0\n for j in range(1, Ny):\n for i in range(1,Nx+1):\n k3 = u_face(k[i,j-1+1], k[i,j+1]) # k_(i, j-1/2)\n k4 = u_face(k[i,j+1], k[i,j]) # k_(i, j+1/2)\n A[I , I+Nx] = -r * k3 # South, 3, down\n A[I+Nx, I ] = -r * k4 # North, 4, up\n I += 1\n \n return A\n\n\ndef LeeDatos(filename):\n \"\"\" Esta funcion lee los datos de un archivo. a? y b? son \n las coordenadas inicial y final del dominio respectivamente\n las direcciones x y y;\n Nx y Ny son el numero de incognitas en las direcciones \n correspondientes; A, B, C y D son las condiciones de\n frontera. Se regresa la tupla (ax,bx,ay,by,Nx,Ny,A,B,C,D).\"\"\"\n ifile = open(filename, 'r') # abre el archivo de entrada\n file_lines = ifile.readlines() # lee las lineas del archivo\n ifile.close(); # cierra el archivo de entrada\n ax, bx, ay, by, Nx, Ny, A, B, C, D = file_lines[0].split() # separa las columnas de la primera linea\n ax = float(ax); bx = float(bx); ay = float(ay); by = float(by); \n Nx = int(Nx); Ny = int(Ny); \n A = float(A); B = float(B); C = float(C); D = float(D); \n return ax, bx, ay, by, Nx, Ny, A, B, C, D\n\ndef ImprimeDatos(ax,bx,ay,by,Nx,Ny,hx,hy,A,cd1,B,cd2,C,cd3,D,cd4):\n \"\"\" Esta funcion imprime los datos del problema a resolver.\"\"\"\n print()\n print(\"+----------------------------------------------------+\")\n print(\"| Solucion de la ecuacion de Laplace en 2D |\")\n print(\"+----------------------------------------------------+\")\n print(\"| Autor: Luis M. de la Cruz S. |\")\n print(\"+----------------------------------------------------+\")\n print(\"| Datos de entrada |\")\n print(\"+----------------------------------------------------+\")\n print(\"| Punto inicial del dominio en x : ax = %g\" % ax)\n print(\"| Punto final del dominio en x : bx = %g\" % bx)\n print(\"| Punto inicial del dominio en y : ax = %g\" % ay)\n print(\"| Punto final del dominio en y : bx = %g\" % by)\n print(\"| Numero total de incognitas en x : Nx = %d\" % Nx)\n print(\"| Numero total de incognitas en y : Ny = %d\" % Ny)\n print(\"| Numero total de incognitas : N = %d\" % (Nx*Ny))\n print(\"| El tamanio de la malla en x es : hx = %g \" % hx)\n print(\"| El tamanio de la malla en y es : hy = %g \" % hy)\n print(\"| Cond. de front. \", cd1, \"en ax : A = %g\" % A)\n print(\"| Cond. de front. \", cd2, \"en bx : B = %g\" % B)\n print(\"| Cond. de front. \", cd3, \"en ay : C = %g\" % C)\n print(\"| Cond. de front. \", cd4, \"en by : D = %g\" % D)\n print(\"+----------------------------------------------------+\")\n\ndef ImprimeSistema(A,u,f):\n \"\"\" Esta funcion imprime el sistema lineal asi como la solucion\n del mismo, siempre y cuando su longitud sea menor o igual a 10.\"\"\"\n print(\"\\n Lado derecho del sistema : size = %d \\n\" % f.size, f)\n print(\"\\n Matriz del sistema : \\n\", A)\n print(\"\\n Solucion del sistema : size = %d \\n\" % u.size, u)\n\ndef GraficaSuperficieC(xg,yg,u,colormap):\n pl.contourf(xg, yg, u, 100, alpha=.95, cmap=colormap)\n C = pl.contour(xg, yg, u, 100, colors='black', alpha=0.01, linewidth=.5)\n pl.clabel(C, inline=1, fontsize=10)\n \n fig = pl.figure()\n ax = Axes3D(fig)\n ax.plot_surface(xg, yg, u, rstride=2, cstride=2, alpha=.95, cmap=colormap)\n\n pl.show()\n\ndef GuardaSolucion(filename, x, y, u):\n \"\"\" Esta funcion guarda la solucion en un archivo para su\n posterior analisis, en un archivo de nombre filename.\"\"\" \n ofile = open(filename, 'w')\n for i in range(0,x.size):\n for j in range(0,y.size):\n ofile.write('%12.10g \\t %12.10g \\t %12.10g\\n' % (x[i], y[j],u[j,i]))\n ofile.close()\n\n\n\n#if __name__ == \"__main__\":\n\n\n\n\n\n", "sub_path": "TEST/poisson2D.py", "file_name": "poisson2D.py", "file_ext": "py", "file_size_in_byte": 7915, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "matplotlib.image.imread", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.image", "line_number": 9, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 83, "usage_type": "call"}, {"api_name": "pylab.contourf", "line_number": 179, "usage_type": "call"}, {"api_name": "pylab.contour", "line_number": 180, "usage_type": "call"}, {"api_name": "pylab.clabel", "line_number": 181, "usage_type": "call"}, {"api_name": "pylab.figure", "line_number": 183, "usage_type": "call"}, {"api_name": "mpl_toolkits.mplot3d.Axes3D", "line_number": 184, "usage_type": "call"}, {"api_name": "pylab.show", "line_number": 187, "usage_type": "call"}]}
+{"seq_id": "246859062", "text": "import rospy\nimport json\n\nfrom planner_msgs.msg import PlanRequest, Plan, AgentTasksRequest, Task\n\nfrom .hatpehda import Goal\n\nclass RosNode:\n def __init__(self, name, on_new_request_cb):\n self.name = name\n self.user_callback = on_new_request_cb\n rospy.init_node(name)\n self.request_sub = rospy.Subscriber(\"~request_new_plan\", PlanRequest, self.on_new_request)\n self.plan_pub = rospy.Publisher(\"~plan_answer\", Plan, queue_size=10)\n @staticmethod\n def start_ros_node(node_name=\"planner\", on_new_request=None):\n return RosNode(node_name, on_new_request)\n\n def retrieve_agents_task(self, agents_task_msg, agents_task):\n for ag in agents_task_msg:\n agents_task[ag.agent_name] = []\n for task in ag.tasks:\n arguments = []\n for ar in task.parameters:\n try:\n print(\"goal\", ar)\n j = json.loads(ar)\n print(j)\n goal = Goal(\"goal\")\n for p, indivs in j.items():\n if not hasattr(goal, p):\n goal.__setattr__(p, {})\n for s, objs in indivs.items():\n goal.__getattribute__(p)[s] = objs\n arguments.append(goal)\n except json.JSONDecodeError as e:\n print(e)\n arguments.append(ar) # We assume that if it is not JSON, it is a simple string\n agents_task[ag.agent_name].append((task.name, arguments))\n\n def on_new_request(self, msg: PlanRequest):\n ctrl_agents_task = {}\n unctrl_agents_task = {}\n self.retrieve_agents_task(msg.controllable_agent_tasks, ctrl_agents_task)\n self.retrieve_agents_task(msg.uncontrollable_agent_tasks, unctrl_agents_task)\n\n if self.user_callback is not None:\n self.user_callback(ctrl_agents_task, unctrl_agents_task)\n\n def wait_for_request(self):\n rospy.spin()\n\n def create_primitive_task(self, action):\n print(action)\n task = Task()\n task.id = action.id\n task.type = task.PRIMITIVE_TASK\n task.name = action.name\n task.parameters = [*action.parameters]\n task.agent = action.agent\n task.successors = []\n if action.why is None:\n task.decomposition_of = -1\n return task\n\n def send_plan(self, actions, ctrlable_name, unctrlable_name):\n existing_edges = set()\n existing_tasks = {}\n msg = Plan()\n msg.tasks = []\n for action in actions:\n while action is not None:\n if action.id not in existing_tasks:\n task = self.create_primitive_task(action)\n # print(task)\n msg.tasks.append(task)\n existing_tasks[action.id] = task\n task = existing_tasks[action.id]\n if action.previous is not None and action.previous.id not in task.predecessors:\n task.predecessors.append(action.previous.id)\n if action.next is not None:\n for n in action.next:\n if n.id not in task.successors:\n task.successors.append(n.id)\n why = action.why\n how = action\n while why is not None:\n if (why.id, how.id) not in existing_edges:\n if why.id not in existing_tasks:\n print(\"adding\", why.id, how.id)\n task = Task()\n task.id = why.id\n task.type = task.ABSTRACT_TASK\n task.name = why.name\n task.parameters = []\n for param in why.parameters:\n #print(\"Parameter\", param)\n if isinstance(param, Goal):\n task.parameters.append(\"goal_{}\".format(param.__name__))\n else:\n task.parameters.append(param)\n #print(task.parameters)\n task.agent = why.agent\n if why.why is None:\n task.decomposition_of = -1\n task.successors = []\n existing_tasks[why.id] = task\n msg.tasks.append(task)\n why_task = existing_tasks[why.id]\n how_task = existing_tasks[how.id] # this one should exist\n why_task.decomposed_into.append(how.id)\n how_task.decomposition_of = why.id\n how_task.decomposition_number = how.decompo_number\n existing_edges.add((why.id, how.id))\n how = why\n why = why.why\n action = action.previous\n self.plan_pub.publish(msg)\n", "sub_path": "hatpehda/ros.py", "file_name": "ros.py", "file_ext": "py", "file_size_in_byte": 5176, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "rospy.init_node", "line_number": 12, "usage_type": "call"}, {"api_name": "rospy.Subscriber", "line_number": 13, "usage_type": "call"}, {"api_name": "planner_msgs.msg.PlanRequest", "line_number": 13, "usage_type": "argument"}, {"api_name": "rospy.Publisher", "line_number": 14, "usage_type": "call"}, {"api_name": "planner_msgs.msg.Plan", "line_number": 14, "usage_type": "argument"}, {"api_name": "json.loads", "line_number": 27, "usage_type": "call"}, {"api_name": "hatpehda.Goal", "line_number": 29, "usage_type": "call"}, {"api_name": "json.JSONDecodeError", "line_number": 36, "usage_type": "attribute"}, {"api_name": "planner_msgs.msg.PlanRequest", "line_number": 41, "usage_type": "name"}, {"api_name": "rospy.spin", "line_number": 51, "usage_type": "call"}, {"api_name": "planner_msgs.msg.Task", "line_number": 55, "usage_type": "call"}, {"api_name": "planner_msgs.msg.Plan", "line_number": 69, "usage_type": "call"}, {"api_name": "planner_msgs.msg.Task", "line_number": 91, "usage_type": "call"}, {"api_name": "hatpehda.Goal", "line_number": 98, "usage_type": "argument"}]}
+{"seq_id": "235442369", "text": "import torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport gluonnlp as nlp\nfrom model.modules import Encoder, Decoder, Attention\nimport random\n\nclass Seq2Seq(nn.Module):\n \"\"\" model class for seq2seq with attention \"\"\"\n def __init__(self, vocab_src:nlp.Vocab, vocab_tgt:nlp.Vocab, embedding_dim:int, hidden_dim:int, dev,\n num_layers:int=1, bos_idx=2, eos_idx=3, use_attention=True):\n \"\"\" initialization of the class \"\"\"\n super(Seq2Seq, self).__init__()\n\n self._encoder = Encoder(vocab_src, embedding_dim, hidden_dim)\n self._decoder = Decoder(vocab_tgt, embedding_dim, hidden_dim)\n\n self._dev = dev\n self._hidden_dim = hidden_dim\n self._bos_idx = bos_idx\n self._eos_idx = eos_idx\n self._pad_idx = self._encoder.pad_idx\n self._mask = None\n\n # global attention related\n self._use_attention = use_attention\n self._attn = Attention(self._hidden_dim) if self._use_attention else None\n\n # teacher forcing related\n self._use_teacher_forcing = None\n self._teacher_forcing_ratio = None\n\n def forward(self, inputs, use_teacher_forcing=True, teacher_forcing_ratio=0.5):\n src, tgt_in, tgt_out = inputs\n batch_size = src.size()[0]\n max_len = tgt_in.size()[1]\n mask = (tgt_out != self._pad_idx)\n\n # teacher forcing\n self._use_teacher_forcing = use_teacher_forcing\n self._teacher_forcing_ratio = teacher_forcing_ratio\n\n encoder_output, encoder_hidden = self._encoder(src) # encoder_out : (batch, max_len, hidden_dim * 2) (BiLSTM)\n decoder_input = torch.full((batch_size, 1), self._bos_idx).long().to(self._dev) # float32 -> int64\n decoder_hidden = encoder_hidden # initialize decoder's hidden state with encoder's last hidden state\n\n loss = 0\n nTotals = 0\n for di in range(max_len):\n if self._use_attention:\n decoder_hidden_top = decoder_hidden[0][0].unsqueeze(1) # top layer's hidden state (batch_size, 1, hidden)\n context_vector = self._attn(decoder_input, decoder_hidden_top, encoder_output)[0] # (batch, 1, hidden)\n decoder_output, next_decoder_hidden = self._decoder(decoder_input, decoder_hidden, context_vector)\n else:\n decoder_output, next_decoder_hidden = self._decoder(decoder_input, decoder_hidden)\n\n decoded_label = decoder_output.topk(1)[1]\n\n # calculate and accumulate loss\n mask_loss, nTotal = self.maskNLLLoss(decoder_output, tgt_out[:,di], mask[:,di], self._dev)\n loss += mask_loss\n nTotals += nTotal\n\n # Teacher forcing: Feed the target as the next input\n if self._use_teacher_forcing:\n if random.random() < self._teacher_forcing_ratio:\n decoder_input = tgt_out[:,di].unsqueeze(-1)\n else:\n decoder_input = decoded_label.squeeze(2)\n else:\n decoder_input = decoded_label.squeeze(2)\n decoder_hidden = next_decoder_hidden\n\n return loss/max_len, nTotals\n\n def maskNLLLoss(self, decoder_output, target, mask, dev):\n \"\"\"\n Calculate average Negative Log Likelihood Loss for mini batch in one time step\n Args:\n decoder_output: (batch, 1, tgt_vocab_size)\n target: (batch, )\n mask: (batch, )\n dev: current device (cpu or gpu)\n return:\n loss:\n\n \"\"\"\n mask = mask.unsqueeze(-1)\n nTotal = mask.sum()\n crossEntropy = -torch.log(torch.gather(decoder_output.squeeze(1), 1, target.unsqueeze(-1)).squeeze(1))\n loss = crossEntropy.masked_select(mask).mean()\n loss = loss.to(dev)\n return loss, nTotal.item()\n\n def to(self, dev):\n super(Seq2Seq, self).to(dev)\n self._dev = dev\n return\n", "sub_path": "wk9_NMT/model/network.py", "file_name": "network.py", "file_ext": "py", "file_size_in_byte": 3938, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "torch.nn.Module", "line_number": 8, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 8, "usage_type": "name"}, {"api_name": "gluonnlp.Vocab", "line_number": 10, "usage_type": "attribute"}, {"api_name": "model.modules.Encoder", "line_number": 15, "usage_type": "call"}, {"api_name": "model.modules.Decoder", "line_number": 16, "usage_type": "call"}, {"api_name": "model.modules.Attention", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.full", "line_number": 44, "usage_type": "call"}, {"api_name": "random.random", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.log", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.gather", "line_number": 90, "usage_type": "call"}]}
+{"seq_id": "402537865", "text": "import json\nimport sys\nimport urllib2\n\nfrom twisted.python import log\n\nimport plugin\n\nURL_ACTIVITY = \"http://api.trakt.tv/activity/user.json/{0}/{1}/all/all/{2}\"\nURL_TIME = \"http://api.trakt.tv/server/time.json/{0}\"\n\nclass Trakt(plugin.Plugin):\n\n def __init__(self):\n log.msg(\"Trakt.__init__\")\n plugin.Plugin.__init__(self, \"Trakt\")\n\n self.settings = {}\n self.users = []\n self.ticks = 0\n\n def update_time(self, users):\n log.msg(\"Trakt.update_time\", users)\n url = URL_TIME.format(self.settings[\"key\"])\n response = urllib2.urlopen(url)\n data = json.load(response)\n for user in users:\n self.users[user][\"last_sync\"] = data[\"timestamp\"]\n\n def started(self, settings):\n log.msg(\"Trakt.started\", settings)\n self.settings = json.loads(settings)\n\n self.users = dict(map(lambda user: (user, {\"last_sync\": 0}), self.settings[\"users\"]))\n self.update_time(self.users)\n\n self.join(0, str(self.settings[\"channel\"]))\n\n def joined(self, server_id, channel):\n log.msg(\"Trakt.joined\", server_id, channel)\n\n def echo(self, message):\n log.msg(\"Trakt.echo\", message)\n self.say(0, str(self.settings[\"channel\"]), \"Trakt: \" + message.encode(\"utf-8\"))\n\n def update(self):\n self.ticks += 1\n if self.ticks % self.settings[\"interval\"] == 0:\n for user in self.users:\n try:\n url = URL_ACTIVITY.format(self.settings[\"key\"], user, self.users[user][\"last_sync\"])\n response = urllib2.urlopen(url)\n data = json.load(response)\n self.users[user][\"last_sync\"] = data[\"timestamps\"][\"current\"]\n for activity in data[\"activity\"]:\n message = Trakt.format_activity(activity, user)\n if message is not None:\n self.echo(message)\n except urllib2.HTTPError as e:\n log.msg(\"HTTP error when fetching\", url, e.code)\n except (urllib2.URLError, ) as e:\n log.msg(\"URL error when fetching\", url, e.args)\n except Exception as e:\n log.msg(\"Unhandled exception when fetching\", url)\n log.msg(\"Data:\", data, \"User:\", user)\n log.err()\n\n @staticmethod\n def format_activity(activity, user):\n if activity[\"type\"] == \"list\":\n if activity[\"action\"] == \"created\":\n return \"{0} create a list '{1}'\".format(user, activity[\"list\"][\"name\"])\n elif activity[\"action\"] == \"item_added\":\n return \"{0} added {1} to the list '{2}'\".format(user, Trakt.format_item(activity[\"list_item\"]), activity[\"list\"][\"name\"])\n else:\n message = user\n\n #if activity[\"action\"] == \"watching\":\n # message += \" is watching (\" + activity[\"elapsed\"][\"short\"] + \") \"\n if activity[\"action\"] == \"scrobble\":\n message += \" scrobbled \"\n elif activity[\"action\"] == \"checkin\":\n message += \" checked in \"\n elif activity[\"action\"] == \"rating\":\n message += \" rated (as \" + Trakt.format_rating(activity) + \") \"\n elif activity[\"action\"] == \"watchlist\":\n message += \" added to watchlist, \"\n else:\n # TODO: seen, collection, shout, review\n return\n\n return message + Trakt.format_item(activity)\n\n @staticmethod\n def format_item(item):\n if item[\"type\"] == \"movie\":\n return Trakt.format_movie(item[\"movie\"])\n elif item[\"type\"] == \"episode\":\n return Trakt.format_episode(item[\"show\"], item[\"episode\"])\n elif item[\"type\"] == \"show\":\n return Trakt.format_show(item[\"show\"])\n\n @staticmethod\n def format_movie(movie):\n return \"'{0[title]} ({0[year]})' {0[url]}\".format(movie)\n\n @staticmethod\n def format_show(show):\n return \"'{0[title]}' {0[url]}\".format(show)\n\n @staticmethod\n def format_episode(show, episode):\n return \"'{0[title]}' 'S{1[season]:02d}E{1[episode]:02d} {1[title]}' {1[url]}\".format(show, episode)\n\n @staticmethod\n def format_rating(activity):\n if activity[\"use_rating_advanced\"]:\n return str(activity[\"rating_advanced\"])\n else:\n return activity[\"rating\"]\n\nif __name__ == \"__main__\":\n sys.exit(Trakt.run())\n\n", "sub_path": "plugins/trakt/trakt.py", "file_name": "trakt.py", "file_ext": "py", "file_size_in_byte": 4513, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "plugin.Plugin", "line_number": 12, "usage_type": "attribute"}, {"api_name": "twisted.python.log.msg", "line_number": 15, "usage_type": "call"}, {"api_name": "twisted.python.log", "line_number": 15, "usage_type": "name"}, {"api_name": "plugin.Plugin.__init__", "line_number": 16, "usage_type": "call"}, {"api_name": "plugin.Plugin", "line_number": 16, "usage_type": "attribute"}, {"api_name": "twisted.python.log.msg", "line_number": 23, "usage_type": "call"}, {"api_name": "twisted.python.log", "line_number": 23, "usage_type": "name"}, {"api_name": "urllib2.urlopen", "line_number": 25, "usage_type": "call"}, {"api_name": "json.load", "line_number": 26, "usage_type": "call"}, {"api_name": "twisted.python.log.msg", "line_number": 31, "usage_type": "call"}, {"api_name": "twisted.python.log", "line_number": 31, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 32, "usage_type": "call"}, {"api_name": "twisted.python.log.msg", "line_number": 40, "usage_type": "call"}, {"api_name": "twisted.python.log", "line_number": 40, "usage_type": "name"}, {"api_name": "twisted.python.log.msg", "line_number": 43, "usage_type": "call"}, {"api_name": "twisted.python.log", "line_number": 43, "usage_type": "name"}, {"api_name": "urllib2.urlopen", "line_number": 52, "usage_type": "call"}, {"api_name": "json.load", "line_number": 53, "usage_type": "call"}, {"api_name": "urllib2.HTTPError", "line_number": 59, "usage_type": "attribute"}, {"api_name": "twisted.python.log.msg", "line_number": 60, "usage_type": "call"}, {"api_name": "twisted.python.log", "line_number": 60, "usage_type": "name"}, {"api_name": "urllib2.URLError", "line_number": 61, "usage_type": "attribute"}, {"api_name": "twisted.python.log.msg", "line_number": 62, "usage_type": "call"}, {"api_name": "twisted.python.log", "line_number": 62, "usage_type": "name"}, {"api_name": "twisted.python.log.msg", "line_number": 64, "usage_type": "call"}, {"api_name": "twisted.python.log", "line_number": 64, "usage_type": "name"}, {"api_name": "twisted.python.log.msg", "line_number": 65, "usage_type": "call"}, {"api_name": "twisted.python.log", "line_number": 65, "usage_type": "name"}, {"api_name": "twisted.python.log.err", "line_number": 66, "usage_type": "call"}, {"api_name": "twisted.python.log", "line_number": 66, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 123, "usage_type": "call"}]}
+{"seq_id": "342520813", "text": "import pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nimport tensorflow as tf\nfrom sklearn.model_selection import StratifiedShuffleSplit\nfrom sklearn.model_selection import train_test_split\n\nfrom .config import CONFIG\n# Types of devices in the model\n_deviceList = ['ac', 'tv', 'fan', 'light', 'geyser']\n\ndef device_exists(command: str, ft_model) -> bool:\n '''Checks whether the given string contains one of the many devices\n supported by the model. \n '''\n # Get the 4 nearest words to each device and create a dictionary\n top_nearest_words_to_devices = {}\n for device in _deviceList:\n top_nearest_words_to_devices[device] = []\n nearestWords = ft_model.get_nearest_neighbors(device, k=4)\n top_nearest_words_to_devices[device].extend([word for _, word in nearestWords])\n top_nearest_words_to_devices[device].extend([device])\n\n for word in command.split(' '):\n for device in _deviceList:\n if word in top_nearest_words_to_devices[device]:\n return True\n \n return False\n\ndef add_class_ovr_cols(dataset: pd.DataFrame) -> pd.DataFrame:\n '''Add a column for each class representing whether that class\n is present for that instance or not (OVR Technique)\n '''\n classList = dataset['label'].unique()\n for label in classList:\n dataset[label] = np.where(dataset['label'] == label, 1, 0)\n return dataset\n\ndef shuffle_split(dataset: pd.DataFrame, label: str):\n '''A generator function to split the dataset using \n StratifiedShuffleSplit and return each split\n '''\n sss = StratifiedShuffleSplit(n_splits=10, test_size=0.2, random_state=20)\n\n # Code taken from https://scikit-learn.org/stable/modules/generated/sklearn.model_selection.StratifiedShuffleSplit.html\n X, y = dataset['sent_vec'], dataset[label]\n for train_index, test_index in sss.split(X, y):\n X_train, X_test = np.stack(X.iloc[train_index]), np.stack(X.iloc[test_index])\n y_train, y_test = y.iloc[train_index], y.iloc[test_index]\n yield np.asarray(X_train), np.asarray(X_test), \\\n np.asarray(y_train), np.asarray(y_test)\n\ndef data_split_classwise(dataset: pd.DataFrame):\n '''Split the data according to the OVR mechanism for \n per class training.\n '''\n classList = dataset['label'].unique()\n for label in classList:\n X, y = dataset['sent_vec'], dataset[label]\n X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, \\\n random_state=40, stratify=y)\n X_train, X_test = np.stack(X_train), np.stack(X_test)\n yield np.asarray(X_train), np.asarray(X_test), \\\n np.asarray(y_train), np.asarray(y_test), label\n\ndef data_split(dataset: pd.DataFrame, test_size: float = 0.25):\n '''Split the dataset into train and test sets.\n '''\n X, y = dataset['sent_vec'], dataset['label']\n X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=test_size, \\\n random_state=40, stratify=y)\n train_df, test_df = pd.DataFrame(X_train), pd.DataFrame(X_test)\n train_df['y'], test_df['y'] = y_train, y_test\n return train_df, test_df\n\ndef plot(models):\n fig = plt.figure(figsize=(20, 60))\n plot_count = 1\n for m_name in models.keys():\n history = models[m_name]['history'].history\n plt.subplot(len(models.keys()), 3, plot_count)\n plt.xlabel('epochs')\n plt.grid()\n plt.ylabel('loss')\n plt.xticks(range(0, len(history['loss']) + 1, 5))\n plt.plot(history['loss'])\n plt.plot(history['val_loss'])\n plt.title(map_label(m_name))\n plt.legend(['Train Set', 'Validation Set'], loc='upper right')\n plot_count += 1\n \n plt.subplot(len(models.keys()), 3, plot_count)\n plt.xlabel('epochs')\n plt.grid()\n plt.ylabel('F1 Score')\n plt.xticks(range(0, len(history['_f1_score']) + 1, 5))\n plt.plot(history['_f1_score'])\n plt.plot(history['val__f1_score'])\n plt.title(map_label(m_name))\n plt.legend(['Train Set', 'Validation Set'], loc='lower right')\n plot_count += 1\n\n plt.subplot(len(models.keys()), 3, plot_count)\n plt.xlabel('epochs')\n plt.grid()\n plt.ylabel('accuracy')\n plt.xticks(range(0, len(history['accuracy']) + 1, 5))\n plt.plot(history['accuracy'])\n plt.plot(history['val_accuracy'])\n plt.title(map_label(m_name))\n plt.legend(['Train Set', 'Validation Set'], loc='lower right')\n plot_count += 1\n\n return fig\n\ndef map_label(label:str)-> str:\n label_map = {\n '__label__light_off': 'light off',\n '__label__light_on': 'light on',\n '__label__geyser_on': 'geyser on',\n '__label__geyser_off': 'geyser off',\n '__label__fan_on': 'fan on',\n '__label__fan_off': 'fan off',\n '__label__tv_on': 'tv on',\n '__label__tv_off': 'tv off',\n '__label__ac_on': 'ac on',\n '__label__ac_off': 'ac off',\n 'Other': 'other'\n }\n\n return label_map[label]\n", "sub_path": "app/hats/utility.py", "file_name": "utility.py", "file_ext": "py", "file_size_in_byte": 5071, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pandas.DataFrame", "line_number": 32, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 38, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 41, "usage_type": "attribute"}, {"api_name": "sklearn.model_selection.StratifiedShuffleSplit", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 53, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 55, "usage_type": "attribute"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 66, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 68, "usage_type": "attribute"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 72, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 97, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 106, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 110, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}]}
+{"seq_id": "642584166", "text": "import argparse\n\nimport numpy as np \nimport pandas as pd \nimport csv\nimport os\nimport time\n\ndef parse_args():\n parser = argparse.ArgumentParser()\n parser.add_argument('--src', type = str, default = 'Office_Products')\n parser.add_argument('--dst', type = str, default = 'Movies_and_TV')\n parser.add_argument('--savepath', type = str, default = '../../data/dataset_1')\n \n return parser.parse_args()\n\n# 提取共享用户的评分信息\ndef draw_shareuser(src, dst):\n if(os.path.exists('src_temp1.csv')):\n os.remove('src_temp1.csv')\n if(os.path.exists('dst_temp1.csv')):\n os.remove('dst_temp1.csv')\n \n src_set = set()\n dst_set = set()\n\n# 首先获得共享用户集合\n # 源数据集的用户集合\n src_read = csv.reader(open(src, 'r'))\n for row in src_read:\n src_set.add(row[0])\n\n # 目的数据集的用户集合\n dst_read = csv.reader(open(dst, 'r'))\n for row in dst_read:\n dst_set.add(row[0])\n\n # 源数据和目的数据集的用户集合的交集,即共享用户集合\n union_set = src_set & dst_set\n print(\"Src_data:%s\\nDst_data:%s\\nShared Users#%d\" %(src, dst, len(union_set)))\n# 然后根据共享用户集合,筛选出所有共享用户的评分信息,即只保留共享用户的评分信息,写入src_temp1.csv和dst_temp1.csv文件中\n src_read = csv.reader(open(src, 'r'))\n with open('src_temp1.csv', 'a', newline = '') as src_out:\n src_write = csv.writer(src_out, dialect = 'excel')\n i = 0\n for row in src_read:\n if row[0] in union_set:\n src_write.writerow(row)\n i += 1\n print('Src Ratings#%d' %i)\n\n dst_read = csv.reader(open(dst, 'r'))\n with open('dst_temp1.csv', 'a', newline = '') as dst_out:\n dst_write = csv.writer(dst_out, dialect = 'excel')\n i = 0\n for row in dst_read:\n if row[0] in union_set:\n dst_write.writerow(row)\n i += 1\n print('Dst Ratings#%d' %i)\n\n# 重新编码userid和itemid\ndef recode_userid_itemid():\n if(os.path.exists('src_temp2.csv')):\n os.remove('src_temp2.csv')\n if(os.path.exists('dst_temp2.csv')):\n os.remove('dst_temp2.csv')\n \n userid = 0\n itemid = 0\n\n userdict = {}\n itemdict = {}\n\n src_read = csv.reader(open('src_temp1.csv', 'r'))\n with open('src_temp2.csv', 'a', newline = '') as src_out:\n src_write = csv.writer(src_out, dialect = 'excel')\n i = 0\n for row in src_read:\n ori_userid = row[0]\n ori_itemid = row[1]\n if ori_userid not in userdict:\n userdict[ori_userid] = userid\n userid += 1\n if ori_itemid not in itemdict:\n itemdict[ori_itemid] = itemid\n itemid += 1\n i += 1\n row[0] = userdict[ori_userid]\n row[1] = itemdict[ori_itemid]\n src_write.writerow(row)\n print('Src_save Ratings#%d Src_Users#%d Src_Items#%d' %(i, len(userdict), len(itemdict)))\n\n itemid = 0\n itemdict.clear()\n\n dst_read = csv.reader(open('dst_temp1.csv', 'r'))\n with open('dst_temp2.csv', 'a', newline = '') as dst_out:\n dst_write = csv.writer(dst_out, dialect = 'excel')\n i = 0\n for row in dst_read:\n ori_userid = row[0]\n ori_itemid = row[1]\n if ori_userid in userdict:\n row[0] = userdict[ori_userid]\n if ori_itemid not in itemdict:\n itemdict[ori_itemid] = itemid\n itemid += 1\n row[1] = itemdict[ori_itemid]\n dst_write.writerow(row)\n i += 1\n print('Dst_save Ratings#%d Dst_Users#%d Dst_Items%d' %(i, len(userdict), len(itemdict)))\n\n# 按userid的大小重新排序\ndef sort_data(src_save, dst_save):\n df = pd.read_csv('src_temp2.csv', names = ['userid', 'itemid', 'ratings'])\n df.sort_values('userid').to_csv(src_save, index = False, header = False)\n df = pd.read_csv('dst_temp2.csv', names = ['userid', 'itemid', 'ratings'])\n df.sort_values('userid').to_csv(dst_save, index = False, header = False)\n\n\n\n\nif __name__ == '__main__':\n ori_datapath = '../../data/ori_data/'\n\n args = parse_args()\n src = ori_datapath + 'ratings_' + args.src + '.csv'\n dst = ori_datapath + 'ratings_' + args.dst + '.csv'\n\n src_save = args.savepath + '/' + args.src + '.csv'\n dst_save = args.savepath + '/' + args.dst + '.csv'\n\n\n draw_shareuser(src, dst)\n recode_userid_itemid()\n sort_data(src_save, dst_save)\n\n os.remove('src_temp1.csv')\n os.remove('dst_temp1.csv')\n os.remove('src_temp2.csv')\n os.remove('dst_temp2.csv')", "sub_path": "utils/data_handling/datahandle.py", "file_name": "datahandle.py", "file_ext": "py", "file_size_in_byte": 4714, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 22, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 29, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 34, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 42, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 44, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 52, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 67, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 75, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 77, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 97, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 99, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 116, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 118, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 139, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 140, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 141, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 142, "usage_type": "call"}]}
+{"seq_id": "379221601", "text": "from collections import OrderedDict\nfrom pathlib import Path\nfrom typing import Union\nfrom io import IOBase\n\n\nDEFAULT_WPA_SUPPLICANT_FILEPATH = Path('/etc/wpa_supplicant/wpa_supplicant.conf')\n\n\nclass ParseError(ValueError):\n pass\n\n\nclass WpaSupplicantConf:\n \"\"\"This class parses a wpa_supplicant configuration file, allows\n manipulation of the configured networks and then writing out of\n the updated file.\n\n WARNING: Although care has been taken to preserve ordering,\n comments will be lost for any wpa_supplicant.conf which is\n round-tripped through this class.\n \"\"\"\n\n def __init__(self, lines=None, filepath=None):\n self._fields = OrderedDict()\n self._networks = OrderedDict()\n self._comments = list()\n if filepath is not None:\n self.filepath = Path(filepath)\n else:\n self.filepath = None\n self._lines = lines\n self.reload()\n\n def reload(self):\n self._fields = OrderedDict()\n self._networks = OrderedDict()\n self._comments = list()\n\n if self.filepath is not None:\n with open(self.filepath, 'r') as rfid:\n self._lines = rfid.readlines()\n\n network = None\n for linenumber, line in enumerate(self._lines):\n line = line.strip()\n if not line or line.startswith('#'):\n self._comments.append((linenumber, line))\n continue\n\n if line == \"}\":\n if network is None:\n raise ParseError(\"unxpected '}'\")\n\n ssid = network.pop('ssid', None)\n if ssid is None:\n raise ParseError('missing \"ssid\" for network')\n self._networks[dequote(ssid)] = network\n network = None\n continue\n\n parts = [x.strip() for x in line.split('=', 1)]\n if len(parts) != 2:\n raise ParseError(\"invalid line: %{!r}\".format(line))\n\n left, right = parts\n\n if right == '{':\n if left != 'network':\n raise ParseError('unsupported section: \"{}\"'.format(left))\n if network is not None:\n raise ParseError(\"can't nest networks\")\n\n network = OrderedDict()\n else:\n if network is None:\n self._fields[left] = right\n else:\n network[left] = right\n\n def fields(self):\n return self._fields\n\n def networks(self):\n return self._networks\n\n def add_network(self, ssid, **attrs):\n self._networks[ssid] = attrs\n\n def remove_network(self, ssid):\n self._networks.pop(ssid, None)\n\n def write(self, fid: Union[IOBase, Path, str] = None):\n print(f'fid={fid}')\n if fid is None and self.filepath is not None:\n fid = open(self.filepath, 'w+')\n elif fid is None:\n raise TypeError(f'write() missing 1 required positional argument: fid')\n elif isinstance(fid, str):\n fid = open(Path(fid), 'w+')\n\n for name, value in self._fields.items():\n fid.write(\"{}={}\\n\".format(name, value))\n\n for ssid, info in self._networks.items():\n fid.write(\"\\nnetwork={\\n\")\n fid.write(' ssid=\"{}\"\\n'.format(ssid))\n for name, value in info.items():\n fid.write(\" {}={}\\n\".format(name, value))\n fid.write(\"}\\n\")\n\n try:\n fid.close()\n except Exception as e:\n print(f'Couldnt close the output file {fid}: {e}')\n pass\n\n @classmethod\n def default(cls):\n return WpaSupplicantConf(filepath=DEFAULT_WPA_SUPPLICANT_FILEPATH)\n\n @classmethod\n def from_file(cls, rfilepath: Path):\n return WpaSupplicantConf(filepath=rfilepath)\n\n @classmethod\n def from_lines(cls, lines):\n return WpaSupplicantConf(lines=lines)\n\n\ndef dequote(v):\n if len(v) < 2:\n return v\n if v.startswith('\"') and v.endswith('\"'):\n return v[1:-1]\n return v\n", "sub_path": "wpasupplicantconf.py", "file_name": "wpasupplicantconf.py", "file_ext": "py", "file_size_in_byte": 4094, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pathlib.Path", "line_number": 7, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 25, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 26, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 29, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 36, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 37, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 74, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 93, "usage_type": "name"}, {"api_name": "io.IOBase", "line_number": 93, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 93, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 100, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 123, "usage_type": "name"}]}
+{"seq_id": "175762704", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nfrom keras.layers import Dense\nfrom keras.models import Sequential\n\nx_data = np.random.rand(100)\nnoise = np.random.normal(0, 0.01, x_data.shape)\ny_data = x_data * 0.1 + 0.2 + noise\nplt.scatter(x_data, y_data)\nplt.show()\n\nmodel = Sequential()\nmodel.add(Dense(units=1,input_dim=1))\nmodel.compile(optimizer='sgd', loss='mse')\n\nfor step in range(30000):\n # 每次都训练一个批次,这个地方我们使用的是全部放入\n cost = model.train_on_batch(x_data, y_data)\n if step % 500 == 0:\n print('cost:', cost)\n\nw, b = model.layers[0].get_weights()\nprint('w:', w, 'b:', b)\n\n# x 输入到网络中 得到预测的值\ny_pred = model.predict(x_data)\n\n# 显示随机点\nplt.scatter(x_data, y_data)\n# 显示预测结果\nplt.plot(x_data, y_pred, 'r-', lw=3)\nplt.show()\n", "sub_path": "test/info.py", "file_name": "info.py", "file_ext": "py", "file_size_in_byte": 833, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "numpy.random.rand", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 6, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 7, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 9, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 10, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 10, "usage_type": "name"}, {"api_name": "keras.models.Sequential", "line_number": 12, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}]}
+{"seq_id": "287253937", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Tue May 28 23:18:19 2019\r\n\r\n@author: tusha\r\n\"\"\"\r\n\r\nimport pandas as pd\r\nimport matplotlib.pyplot as plt\r\nfrom decimal import Decimal\r\nimport numpy as np\r\nimport seaborn as sb\r\nh1b_data = pd.read_csv('C://Users//tusha//Desktop//Tushar School Documents//Masters Project//ian-h-1-b-disclosure-data-fy-17//h1b_kaggle_55.csv')\r\n\r\nlen(h1b_data)\r\nh1b_data.EMPLOYER_NAME.value_counts().head(15)\r\n\r\nh1b_data['EMPLOYER_NAME'].value_counts().head(15).plot(kind = \"bar\" , title =\"Top 15 Hiring Company\")\r\n\r\n\r\nh1b_data.PREVAILING_WAGE.value_counts().sort_values(ascending = False).head(15)\r\n\r\n\r\nh1b_data.PREVAILING_WAGE.mean()\r\n\r\ndenied = h1b_data[h1b_data.CASE_STATUS=='DENIED']\r\n\r\nnooooo = h1b_data[h1b_data.YEAR == 'nan']\r\nh1b_data.dropna()\r\nDAta = h1b_data[h1b_data.JOB_TITLE == 'DATA ANALYST']\r\n\r\nDAta['EMPLOYER_NAME'].value_counts().head(50)\r\n\r\n#wages given by employee\r\nwages_employee = h1b_data.groupby(['EMPLOYER_NAME']).mean()['PREVAILING_WAGE'].nlargest(15).plot(kind = 'bar')\r\n\r\nh1b_data.WORKSITE.value_counts().head(20)\r\n\r\nh1b_data.WORKSITE.value_counts().head(20).plot(kind = 'bar', title =\"Cities with Highest Job opportunity \")\r\n\r\nh1b_data.loc[:,'WORKSITE'] = h1b_data.loc[:,'WORKSITE'].apply(lambda rec:rec.split(',')[1][1:])\r\n\r\ndef change_NA(rec):\r\n if (rec=='NA'):\r\n return 'MARINA ISLANDS'\r\n return rec\r\nh1b_data.loc[:,'WORKSITE'] = h1b_data.loc[:,'WORKSITE'].apply(lambda rec: change_NA(rec))\r\nprint(len(h1b_data['WORKSITE'].unique()))\r\n\r\n\r\nh1b_data['CASE_STATUS'].unique()\r\n\r\n\r\nstatus_freq = [0]*7\r\n\r\nstatues = ['CERTIFIED-WITHDRAWN', 'WITHDRAWN', 'CERTIFIED', 'DENIED',\r\n 'REJECTED', 'INVALIDATED',\r\n 'PENDING QUALITY AND COMPLIANCE REVIEW - UNASSIGNED']\r\n\r\nfor i in range(0,7):\r\n status_freq[i] = h1b_data[h1b_data.CASE_STATUS==statues[i]]['CASE_STATUS'].count()\r\nstatus_freq\r\n#status_freq.unique()\r\nfrom matplotlib.pyplot import pie,axis,show\r\nimport matplotlib as mpl\r\n\r\nplt.figure(figsize = (5,5))\r\nplt.title('PETITIONS BY CASE STATUS')\r\naxis('equal');\r\npie(status_freq[:4], labels = statues[:4]);\r\nshow()\r\n\r\n#h1b_data.EMPLOYMENT_START_DATE = pd.tslib.Timestamp.now()\r\nh1b_data['YEAR'] = h1b_data['YEAR'].apply(lambda year:'%g' % (Decimal(str(year))))\r\n\r\nh1b_data['PREVAILING_WAGE'] = h1b_data['PREVAILING_WAGE'].apply(lambda year:'%g' % (Decimal(str(year))))\r\n\r\nyear = ['2011','2012','2013','2014','2015','2016']\r\nyear_count = [0]*6\r\nfor i in range(0,6):\r\n year_count[i] = h1b_data[h1b_data.YEAR==year[i]]['YEAR'].count()\r\nyear_count\r\n\r\nsb.set_context(\"notebook\",font_scale=1.0)\r\nplt.figure(figsize=(13,3))\r\nplt.title('PETITIONS DISTRIBUTION BY YEAR')\r\nsb.countplot(h1b_data['YEAR'])\r\n\r\ndenied = h1b_data[h1b_data.CASE_STATUS=='DENIED']\r\nlen(denied)\r\n\r\n\r\ndel denied['CASE_STATUS']\r\ndenied = denied.reset_index()\r\ndenied.head()\r\n\r\ndenied_year_count = [0]*6\r\nfor i in range(0,6):\r\n denied_year_count[i] = denied[denied.YEAR==year[i]]['YEAR'].count()\r\ndenied_year_count\r\n\r\n\r\nsb.set_context(\"notebook\",font_scale=1.0)\r\nplt.figure(figsize=(13,3))\r\nplt.title('DENIED PETITIONS BY YEAR')\r\nsb.countplot(denied['YEAR'])\r\n\r\ndenied_rate = [0]*6\r\nfor i in range(0,6):\r\n denied_rate[i] = float(\"%.2f\" % ((denied_year_count[i] / year_count[i])*100))\r\n\r\nratio = pd.DataFrame()\r\nratio['year'] = year\r\nratio['denied rate %'] = denied_rate\r\nratio = ratio.set_index(['year'])\r\nratio.T\r\n\r\nratio = ratio.reset_index()\r\nsb.set_context(\"notebook\",font_scale=1.0)\r\nplt.figure(figsize=(13,3))\r\nplt.title('DENIED PETITIONS RATE BY YEAR')\r\ng= sb.barplot(x='year' , y = 'denied rate %', data = ratio)\r\n\r\nUS_states = ['Alabama','Alaska','Arizona','Arkansas','California','Colorado ','Connecticut','Delaware',\r\n 'District of Columbia','Florida','Georgia','Hawaii','Idaho','Illinois','Indiana','Iowa','Kansas','Kentucky','Louisiana',\r\n 'Maine','Marina Islands','Maryland','Massachusetts','Michigan','Minnesota','Mississippi','Missouri','Montana','Nebraska',\r\n 'Nevada','New Hampshire','New Jersey','New Mexico','New York','North Carolina','North Dakota',\r\n 'Ohio','Oklahoma','Oregon','Pennsylvania','Puerto Rico','Rhode Island','South Carolina','South Dakota','Tennessee',\r\n 'Texas ','Utah','Vermont','Virginia','Washington','West Virginia','Wisconsin','Wyoming']\r\n\r\nUS_states = [x.upper() for x in US_states] \r\n \r\n# printing output \r\nprint(US_states) \r\nlen(US_states)\r\npetition_by_state = [0]*53\r\nfor i in range(0,53):\r\n petition_by_state[i] = h1b_data[h1b_data.WORKSITE == US_states[i]]['WORKSITE'].count()\r\npet_state = pd.DataFrame()\r\npet_state['STATE'] = US_states\r\npet_state['FILED PETITIONS'] = petition_by_state\r\nprint(sum(petition_by_state))\r\n\r\n\r\nsb.set_context(\"notebook\",font_scale=1.0)\r\nplt.figure(figsize=(13,5))\r\nplt.title('FILED PETITIONS BY STATE')\r\nv= sb.barplot(x='STATE' , y = 'FILED PETITIONS', data = pet_state)\r\nrotg = v.set_xticklabels(v.get_xticklabels(), rotation = 90)\r\n\r\n########\r\nlen(denied)\r\ndenied_by_state = [0]*53\r\nfor i in range(0,53):\r\n denied_by_state[i] = denied[denied.WORKSITE == US_states[i]]['WORKSITE'].count()\r\nden_state = pd.DataFrame()\r\nden_state['STATE'] = US_states\r\nden_state['DENIED PETITIONS'] = denied_by_state\r\nprint(sum(denied_by_state))\r\n\r\n\r\nsb.set_context(\"notebook\",font_scale=1.0)\r\nplt.figure(figsize=(13,5))\r\nplt.title('DENIED PETITIONS BY STATE')\r\nv= sb.barplot(x='STATE' , y = 'DENIED PETITIONS', data = den_state)\r\nrotg = v.set_xticklabels(v.get_xticklabels(), rotation = 90)\r\n\r\n\r\n\r\n#######\r\n\r\ndenied_state_rate = [0]*53\r\nfor i in range(0,53):\r\n denied_state_rate[i] = float(\"%.2f\" % ((denied_by_state[i] / petition_by_state[i])*100))\r\nratios = pd.DataFrame()\r\nratios['STATE'] = US_states\r\nratios['DENIED PETITIONS %'] = denied_state_rate\r\nprint(sum(denied_state_rate))\r\n\r\n\r\nsb.set_context(\"notebook\",font_scale=1.0)\r\nplt.figure(figsize=(13,5))\r\nplt.title('DENIED PETITIONS BY STATE')\r\nv= sb.barplot(x='STATE' , y = 'DENIED PETITIONS %', data = ratios)\r\nrotg = v.set_xticklabels(v.get_xticklabels(), rotation = 90)\r\n\r\n\r\npet_state['DENIED PETITIONS'] = denied_by_state\r\npet_state['DENIED PETITIONS %'] = denied_state_rate\r\npet_state = pet_state.sort_values(by='DENIED PETITIONS %',ascending = False)\r\npet_state\r\n\r\n\r\nh1b_data.JOB_TITLE.value_counts().head(15)\r\n\r\nh1b_data['JOB_TITLE'].value_counts().head(15).plot(kind = \"bar\" , title =\"Top 15 Jobs\")", "sub_path": "h1b.py", "file_name": "h1b.py", "file_ext": "py", "file_size_in_byte": 6341, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pandas.read_csv", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.pie", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 71, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 74, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 76, "usage_type": "call"}, {"api_name": "seaborn.set_context", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "seaborn.countplot", "line_number": 87, "usage_type": "call"}, {"api_name": "seaborn.set_context", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "seaborn.countplot", "line_number": 106, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 112, "usage_type": "call"}, {"api_name": "seaborn.set_context", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 120, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 121, "usage_type": "name"}, {"api_name": "seaborn.barplot", "line_number": 122, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 139, "usage_type": "call"}, {"api_name": "seaborn.set_context", "line_number": 145, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 146, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 147, "usage_type": "name"}, {"api_name": "seaborn.barplot", "line_number": 148, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 156, "usage_type": "call"}, {"api_name": "seaborn.set_context", "line_number": 162, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 163, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 163, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 164, "usage_type": "name"}, {"api_name": "seaborn.barplot", "line_number": 165, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 175, "usage_type": "call"}, {"api_name": "seaborn.set_context", "line_number": 181, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 182, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}, {"api_name": "seaborn.barplot", "line_number": 184, "usage_type": "call"}]}
+{"seq_id": "576168236", "text": "'''\nCreated on Oct 9, 2012\n\n@author: davidgrogan\n'''\nimport db\nimport datetime\nimport named_object\nimport sub_object\nimport unittest\n\nclass GenericNamedObject(named_object.NamedObject):\n '''\n A generic named object for testing purposes.\n '''\n\n class Factory(named_object.NamedObject.Factory):\n '''\n The factory for creating generic named objects.\n '''\n \n def __init__(self):\n super(GenericNamedObject.Factory, self).__init__(\"test_named_object\")\n \n def _new_instance(self, data):\n return GenericNamedObject(data, self)\n\n def __init__(self, data, factory):\n if not data:\n data = list([-1, \"\", 45.0, \"20010101\"])\n super(GenericNamedObject, self).__init__(data, factory)\n self.__sub_object_collection = None\n \n def get_float_value(self):\n return self._get_data()[2]\n \n def set_float_value(self, float_value):\n self._get_data()[2] = float_value\n \n def get_date_value(self):\n dt = self._get_data()[3]\n return datetime.date(int(dt[:4]), int(dt[4:6]), int(dt[6:8]))\n \n def set_date_value(self, date_value):\n self._get_data()[3] = date_value.strftime(\"%Y%m%d\")\n \n def get_sub_objects(self):\n if not self.__sub_object_collection:\n self.__sub_object_collection = SubObject.Collection(self)\n return self.__sub_object_collection\n\n def save(self):\n super(GenericNamedObject, self).save()\n self.get_sub_objects().save()\n \n def _prepare_for_removal(self):\n self.get_sub_objects()._remove_all()\n\nclass SubObject(sub_object.SubObject):\n '''\n A sub object for testing purposes.\n '''\n \n class Collection(sub_object.SubObject.Collection):\n '''\n The collection class for sub-objects.\n '''\n\n def __init__(self, parent):\n super(SubObject.Collection, self).__init__(parent, \"test_sub_object\")\n\n def _new_instance(self, data, parent):\n return SubObject(data, parent)\n \n\n def __init__(self, data, owner):\n if not data:\n data = list([-1, -1, \"\", 45.0, \"20010101\"])\n super(SubObject, self).__init__(data, owner)\n\n\ndef get_factory():\n '''\n Returns the factory for creating generic named objects.\n '''\n return GenericNamedObject.Factory()\n\n\nclass Test(unittest.TestCase):\n\n\n def setUp(self):\n db.reset_connection()\n db.set_database_type(\"postgres\")\n self.assertNotEqual(db.connect_default(), None)\n self.assertNotEqual(db.get_connection(), None)\n\n\n def tearDown(self):\n db.reset_connection()\n\n\n def testNamedObject(self):\n name = \"test_object\"\n float_value = 55.5\n date_value = datetime.date(2009, 12, 17)\n factory = get_factory()\n factory.remove(name)\n self.assertFalse(factory.is_valid_name(name))\n # Test create\n object_instance = factory.create()\n object_instance.set_name(name)\n object_instance.set_float_value(float_value)\n object_instance.set_date_value(date_value)\n object_instance.save()\n self.assertIsNotNone(object_instance)\n self.assertGreaterEqual(object_instance.get_id(), 0)\n self.assertEqual(object_instance.get_name(), name)\n self.assertEqual(object_instance.get_float_value(), float_value)\n self.assertEqual(object_instance.get_date_value(), date_value)\n self.assertTrue(factory.is_valid_name(name))\n self.assertTrue(factory.is_valid_id(object_instance.get_id()))\n # Test retrieve by id\n object_instance = factory.retrieve_by_id(object_instance.get_id())\n self.assertIsNotNone(object_instance)\n self.assertGreaterEqual(object_instance.get_id(), 0)\n self.assertEqual(object_instance.get_name(), name)\n self.assertEqual(object_instance.get_float_value(), float_value)\n self.assertEqual(object_instance.get_date_value(), date_value)\n # Test retrieve\n object_instance = factory.retrieve_by_name(name)\n self.assertIsNotNone(object_instance)\n self.assertGreaterEqual(object_instance.get_id(), 0)\n self.assertEqual(object_instance.get_name(), name)\n self.assertEqual(object_instance.get_float_value(), float_value)\n self.assertEqual(object_instance.get_date_value(), date_value)\n # Test retrieve_all\n all_object_instances = factory.retrieve_all()\n self.assertGreaterEqual(len(all_object_instances), 0)\n found = False\n for object_instance in all_object_instances:\n if object_instance.get_name() == name:\n found = True\n self.assertGreaterEqual(object_instance.get_id(), 0)\n self.assertEqual(object_instance.get_name(), name)\n self.assertTrue(found)\n # This should throw an exception because a object_instance with the specified name already exists.\n try:\n factory.create(name)\n self.fail(\"factory.create didn't throw an exception.\")\n except:\n pass\n factory.remove(name)\n self.assertFalse(factory.is_valid_name(name))\n self.assertFalse(factory.is_valid_id(object_instance.get_id()))\n # This should throw an exception because the object_instance no longer exists.\n try:\n factory.retrieve(name)\n self.fail(\"factory.retrieve didn't throw an exception.\")\n except:\n pass\n\n def testSubObject(self):\n name = \"test_object\"\n factory = get_factory()\n factory.remove(name)\n # Test create\n object_instance = factory.create()\n object_instance.set_name(name)\n sub_objects = object_instance.get_sub_objects()\n self.assertTrue(len(sub_objects) == 0)\n new_item = object_instance.get_sub_objects().add_item()\n new_item = object_instance.get_sub_objects().add_item()\n new_item = object_instance.get_sub_objects().add_item()\n \n object_instance.save()\n\nif __name__ == \"__main__\":\n #import sys;sys.argv = ['', 'Test.testName']\n unittest.main()\n", "sub_path": "tests/test_named_object.py", "file_name": "test_named_object.py", "file_ext": "py", "file_size_in_byte": 6175, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "named_object.NamedObject", "line_number": 12, "usage_type": "attribute"}, {"api_name": "named_object.NamedObject", "line_number": 17, "usage_type": "attribute"}, {"api_name": "datetime.date", "line_number": 42, "usage_type": "call"}, {"api_name": "sub_object.SubObject", "line_number": 59, "usage_type": "attribute"}, {"api_name": "sub_object.SubObject", "line_number": 64, "usage_type": "attribute"}, {"api_name": "unittest.TestCase", "line_number": 89, "usage_type": "attribute"}, {"api_name": "db.reset_connection", "line_number": 93, "usage_type": "call"}, {"api_name": "db.set_database_type", "line_number": 94, "usage_type": "call"}, {"api_name": "db.connect_default", "line_number": 95, "usage_type": "call"}, {"api_name": "db.get_connection", "line_number": 96, "usage_type": "call"}, {"api_name": "db.reset_connection", "line_number": 100, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 106, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 180, "usage_type": "call"}]}
+{"seq_id": "169638277", "text": "import numpy as np\n\n\nfrom sl1m.constants_and_tools import *\nfrom sl1m import planner_l1 as pl1\nfrom sl1m import planner as pl\n\nfrom . import qp\n\n\n# try to import mixed integer solver\nMIP_OK = False \ntry:\n import gurobipy\n import cvxpy as cp\n MIP_OK = True\n\nexcept ImportError:\n pass\n\n\n\nnp.set_printoptions(formatter={'float': lambda x: \"{0:0.1f}\".format(x)})\n\n\n\n### This solver is called when the sparsity is fixed. It assumes the first contact surface for each phase\n### is the one used for contact creation.\ndef solve(pb,surfaces, draw_scene = None, plot = True ): \n \n t1 = clock()\n A, b, E, e = pl.convertProblemToLp(pb) \n C = identity(A.shape[1])\n c = zeros(A.shape[1])\n t2 = clock()\n res = qp.quadprog_solve_qp(C, c,A,b,E,e)\n t3 = clock()\n \n print(\"time to set up problem\" , timMs(t1,t2))\n print(\"time to solve problem\" , timMs(t2,t3))\n print(\"total time\" , timMs(t1,t3))\n \n coms, footpos, allfeetpos = pl.retrieve_points_from_res(pb, res)\n \n plot = plot and draw_scene is not None \n if plot:\n ax = draw_scene(surfaces)\n pl.plotQPRes(pb, res, ax=ax)\n \n return pb, coms, footpos, allfeetpos, res\n\n\n### Calls the sl1m solver. Brute-forcedly tries to solve non fixed sparsity by handling the combinatorial.\n### Ultimately calls solve which provides the approriate cost function\ndef solveL1(pb, surfaces, draw_scene = None, plot = True): \n A, b, E, e = pl1.convertProblemToLp(pb) \n C = identity(A.shape[1]) * 0.00001\n c = pl1.slackSelectionMatrix(pb)\n \n res = qp.quadprog_solve_qp(C, c,A,b,E,e)\n \n ok = pl1.isSparsityFixed(pb, res)\n solutionIndices = None\n solutionComb = None\n if not ok:\n pbs = pl1.generateAllFixedScenariosWithFixedSparsity(pb, res)\n \n t3 = clock()\n \n for (pbComb, comb, indices) in pbs:\n A, b, E, e = pl1.convertProblemToLp(pbComb, convertSurfaces = False)\n C = identity(A.shape[1]) * 0.00001\n c = pl1.slackSelectionMatrix(pbComb)\n try:\n res = qp.quadprog_solve_qp(C, c,A,b,E,e)\n if pl1.isSparsityFixed(pbComb, res): \n coms, footpos, allfeetpos = pl1.retrieve_points_from_res(pbComb, res)\n pb = pbComb\n ok = True\n solutionIndices = indices[:]\n solutionComb = comb\n if plot:\n ax = draw_scene(surfaces)\n pl1.plotQPRes(pb, res, ax=ax)\n break\n except:\n print(\"unfeasible problem\")\n pass\n \n t4 = clock() \n \n print(\"time to solve combinatorial \", timMs(t3,t4))\n \n if ok:\n surfacesret, indices = pl1.bestSelectedSurfaces(pb, res) \n for i, phase in enumerate(pb[\"phaseData\"]): \n phase[\"S\"] = [surfaces[i][indices[i]]]\n if solutionIndices is not None:\n for i, idx in enumerate(solutionIndices):\n pb[\"phaseData\"][idx][\"S\"] = [surfaces[idx][solutionComb[i]]]\n \n return solve(pb,surfaces, draw_scene = draw_scene, plot = True ) \n\n\n############### MIXED-INTEGER SOLVER ###############\n\ndef tovals(variables):\n return array([el.value for el in variables])\n\ndef solveMIP(pb, surfaces, MIP = True, draw_scene = None, plot = True): \n if not MIP_OK:\n print(\"Mixed integer formulation requires gurobi packaged in cvxpy\")\n raise ImportError\n \n gurobipy.setParam('LogFile', '')\n gurobipy.setParam('OutputFlag', 0)\n \n A, b, E, e = pl1.convertProblemToLp(pb) \n slackMatrix = pl1.slackSelectionMatrix(pb)\n \n rdim = A.shape[1]\n varReal = cp.Variable(rdim)\n constraints = []\n constraintNormalIneq = A * varReal <= b\n constraintNormalEq = E * varReal == e\n \n constraints = [constraintNormalIneq, constraintNormalEq]\n #creating boolean vars\n \n slackIndices = [i for i,el in enumerate (slackMatrix) if el > 0]\n numSlackVariables = len([el for el in slackMatrix if el > 0])\n boolvars = cp.Variable(numSlackVariables, boolean=True) \n obj = cp.Minimize(slackMatrix * varReal)\n \n if MIP: \n constraints = constraints + [varReal[el] <= 100. * boolvars[i] for i, el in enumerate(slackIndices)] \n \n currentSum = []\n previousL = 0\n for i, el in enumerate(slackIndices):\n if i!= 0 and el - previousL > 2.:\n assert len(currentSum) > 0\n constraints = constraints + [sum(currentSum) == len(currentSum) -1 ]\n currentSum = [boolvars[i]]\n elif el !=0:\n currentSum = currentSum + [boolvars[i]]\n previousL = el\n if len(currentSum) > 1:\n constraints = constraints + [sum(currentSum) == len(currentSum) -1 ]\n obj = cp.Minimize(ones(numSlackVariables) * boolvars)\n prob = cp.Problem(obj, constraints)\n t1 = clock()\n res = prob.solve(solver=cp.GUROBI, verbose=False )\n t2 = clock()\n res = tovals(varReal)\n print(\"time to solve MIP \", timMs(t1,t2))\n\n \n plot = plot and draw_scene is not None \n if plot:\n ax = draw_scene(surfaces)\n pl1.plotQPRes(pb, res, ax=ax)\n \n return timMs(t1,t2)\n \n", "sub_path": "sl1m/fix_sparsity.py", "file_name": "fix_sparsity.py", "file_ext": "py", "file_size_in_byte": 5372, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "numpy.set_printoptions", "line_number": 23, "usage_type": "call"}, {"api_name": "sl1m.planner.convertProblemToLp", "line_number": 32, "usage_type": "call"}, {"api_name": "sl1m.planner", "line_number": 32, "usage_type": "name"}, {"api_name": "sl1m.planner.retrieve_points_from_res", "line_number": 43, "usage_type": "call"}, {"api_name": "sl1m.planner", "line_number": 43, "usage_type": "name"}, {"api_name": "sl1m.planner.plotQPRes", "line_number": 48, "usage_type": "call"}, {"api_name": "sl1m.planner", "line_number": 48, "usage_type": "name"}, {"api_name": "sl1m.planner_l1.convertProblemToLp", "line_number": 56, "usage_type": "call"}, {"api_name": "sl1m.planner_l1", "line_number": 56, "usage_type": "name"}, {"api_name": "sl1m.planner_l1.slackSelectionMatrix", "line_number": 58, "usage_type": "call"}, {"api_name": "sl1m.planner_l1", "line_number": 58, "usage_type": "name"}, {"api_name": "sl1m.planner_l1.isSparsityFixed", "line_number": 62, "usage_type": "call"}, {"api_name": "sl1m.planner_l1", "line_number": 62, "usage_type": "name"}, {"api_name": "sl1m.planner_l1.generateAllFixedScenariosWithFixedSparsity", "line_number": 66, "usage_type": "call"}, {"api_name": "sl1m.planner_l1", "line_number": 66, "usage_type": "name"}, {"api_name": "sl1m.planner_l1.convertProblemToLp", "line_number": 71, "usage_type": "call"}, {"api_name": "sl1m.planner_l1", "line_number": 71, "usage_type": "name"}, {"api_name": "sl1m.planner_l1.slackSelectionMatrix", "line_number": 73, "usage_type": "call"}, {"api_name": "sl1m.planner_l1", "line_number": 73, "usage_type": "name"}, {"api_name": "sl1m.planner_l1.isSparsityFixed", "line_number": 76, "usage_type": "call"}, {"api_name": "sl1m.planner_l1", "line_number": 76, "usage_type": "name"}, {"api_name": "sl1m.planner_l1.retrieve_points_from_res", "line_number": 77, "usage_type": "call"}, {"api_name": "sl1m.planner_l1", "line_number": 77, "usage_type": "name"}, {"api_name": "sl1m.planner_l1.plotQPRes", "line_number": 84, "usage_type": "call"}, {"api_name": "sl1m.planner_l1", "line_number": 84, "usage_type": "name"}, {"api_name": "sl1m.planner_l1.bestSelectedSurfaces", "line_number": 95, "usage_type": "call"}, {"api_name": "sl1m.planner_l1", "line_number": 95, "usage_type": "name"}, {"api_name": "gurobipy.setParam", "line_number": 115, "usage_type": "call"}, {"api_name": "gurobipy.setParam", "line_number": 116, "usage_type": "call"}, {"api_name": "sl1m.planner_l1.convertProblemToLp", "line_number": 118, "usage_type": "call"}, {"api_name": "sl1m.planner_l1", "line_number": 118, "usage_type": "name"}, {"api_name": "sl1m.planner_l1.slackSelectionMatrix", "line_number": 119, "usage_type": "call"}, {"api_name": "sl1m.planner_l1", "line_number": 119, "usage_type": "name"}, {"api_name": "cvxpy.Variable", "line_number": 122, "usage_type": "call"}, {"api_name": "cvxpy.Variable", "line_number": 132, "usage_type": "call"}, {"api_name": "cvxpy.Minimize", "line_number": 133, "usage_type": "call"}, {"api_name": "cvxpy.Minimize", "line_number": 150, "usage_type": "call"}, {"api_name": "cvxpy.Problem", "line_number": 151, "usage_type": "call"}, {"api_name": "cvxpy.GUROBI", "line_number": 153, "usage_type": "attribute"}, {"api_name": "sl1m.planner_l1.plotQPRes", "line_number": 162, "usage_type": "call"}, {"api_name": "sl1m.planner_l1", "line_number": 162, "usage_type": "name"}]}
+{"seq_id": "480163346", "text": "# coding: utf-8\r\nimport shutil, tqdm\r\nimport pandas as pd\r\n\r\ndf = pd.read_csv(\"./temp/df.csv\")\r\npaths = df.file.tolist()\r\nbars = tqdm.tqdm(paths)\r\nfor i, bar in enumerate(bars):\r\n shutil.copyfile(paths[i], \"./temp/{}\".format(paths[i][13:]))\r\n bars.set_description(\"已复制 {}\".format(bar))\r\n", "sub_path": "0202/move0K.py", "file_name": "move0K.py", "file_ext": "py", "file_size_in_byte": 299, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pandas.read_csv", "line_number": 5, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 7, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 9, "usage_type": "call"}]}
+{"seq_id": "591330890", "text": "import cv2\nimport os\nimport glob\nfrom tqdm import tqdm\nimport torch\nfrom torchvision import models\nimport matplotlib.pyplot as plt\n#画出余弦学习率的变化规律\ndef visulize_cosine_lr(net,max_epoch,optimizer,lr_scheduler,iters=100):\n\n plt.figure()\n cur_lr_list = []\n cur_lr = optimizer.param_groups[-1]['lr']\n cur_lr_list.append(cur_lr)\n for epoch in range(max_epoch):\n #print('epoch_{}'.format(epoch))\n # cur_lr = optimizer.param_groups[-1]['lr']\n # cur_lr_list.append(cur_lr)\n for batch in range(iters):\n optimizer.step()\n scheduler.step(epoch + batch / iters)\n cur_lr = optimizer.param_groups[-1]['lr']\n cur_lr_list.append(cur_lr)\n #scheduler.step(epoch + batch / iters)\n #scheduler.step()\n #print('cur_lr:',cur_lr)\n #print('epoch_{}_end'.format(epoch))\n lr_scheduler.step()\n x_list = list(range(len(cur_lr_list)))\n plt.title('Cosine lr T_0:{} T_mult:{}'.format(T_0,T_mult))\n plt.xlabel('epoch')\n plt.ylabel('lr')\n plt.plot(x_list, cur_lr_list)\n plt.savefig('./lr.png')\nif __name__=='__main__':\n model=models.resnet18(pretrained=False)\n T_0=3\n T_mult=2\n optimizer = torch.optim.SGD(model.parameters(), lr=1e-2, momentum=0.9, weight_decay=5e-4)\n #scheduler = StepLR(optimizer, step_size=step_size, gamma=gamma)\n scheduler = torch.optim.lr_scheduler.CosineAnnealingWarmRestarts(optimizer, T_0=T_0, T_mult=T_mult, eta_min=1e-5, last_epoch=-1)\n visulize_cosine_lr(model,100,optimizer,scheduler,901)", "sub_path": "天池/CVPR2021-PIC-Challenge/plot_lr.py", "file_name": "plot_lr.py", "file_ext": "py", "file_size_in_byte": 1568, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 11, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "torchvision.models.resnet18", "line_number": 36, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 36, "usage_type": "name"}, {"api_name": "torch.optim.SGD", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 39, "usage_type": "attribute"}, {"api_name": "torch.optim.lr_scheduler.CosineAnnealingWarmRestarts", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 41, "usage_type": "attribute"}]}
+{"seq_id": "321273041", "text": "from __future__ import division, print_function\nimport time\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom numpy import asarray\nfrom numpy import expand_dims\nfrom numpy import log\nfrom numpy import mean\nfrom numpy import exp\nfrom numpy import std\nfrom math import floor\nimport os\nfrom keras.models import Model, Sequential\nfrom keras.layers import Activation, Dense, Flatten, BatchNormalization, Dropout, Input, Reshape, multiply\nfrom keras.layers import Embedding, ZeroPadding2D\nfrom keras.layers.advanced_activations import LeakyReLU\nfrom keras.optimizers import Nadam, Adam, SGD\nfrom keras.datasets import mnist\nimport tensorflow as tf\n\n(X_train, y_train), (X_test, y_test) = mnist.load_data()\n\n# Image shape information\n\nimg_rows = X_train.shape[1]\nimg_cols = X_train.shape[2]\nif len(X_train.shape) == 4:\n channels = X_train.shape[3]\nelse:\n channels = 1\n\nimg_shape = (img_rows, img_cols, channels)\nnum_classes = 10\nlatent_dim = 100\noptimizer = Adam(0.0002, 0.5)\n\ndef generator():\n model = Sequential()\n model.add(Dense(256, input_dim=latent_dim))\n model.add(LeakyReLU(alpha=0.2))\n model.add(BatchNormalization(momentum=0.8))\n model.add(Dense(512))\n model.add(LeakyReLU(alpha=0.2))\n model.add(BatchNormalization(momentum=0.8))\n model.add(Dense(1024))\n model.add(LeakyReLU(alpha=0.2))\n model.add(BatchNormalization(momentum=0.8))\n model.add(Dense(np.prod(img_shape), activation='tanh'))\n model.add(Reshape(img_shape))\n #model.summary()\n\n noise = Input(shape=(latent_dim,))\n label = Input(shape=(1,), dtype='int32')\n label_embedding = Flatten()(Embedding(num_classes, latent_dim)(label))\n model_input = multiply([noise, label_embedding])\n img = model(model_input)\n return Model([noise, label], img)\n\ndef discriminator():\n model = Sequential()\n model.add(Dense(512, input_dim=np.prod(img_shape)))\n model.add(LeakyReLU(alpha=0.2))\n model.add(Dense(512))\n model.add(LeakyReLU(alpha=0.2))\n model.add(Dropout(0.4))\n model.add(Dense(512))\n model.add(LeakyReLU(alpha=0.2))\n model.add(Dropout(0.4))\n model.add(Dense(1, activation='sigmoid'))\n #model.summary()\n\n img = Input(shape=img_shape)\n label = Input(shape=(1,), dtype='int32')\n label_embedding = Flatten()(Embedding(num_classes, np.prod(img_shape))(label))\n flat_img = Flatten()(img)\n\n model_input = multiply([flat_img, label_embedding])\n validity = model(model_input)\n return Model([img, label], validity)\n\n# Build the generator\n\ngenerator = generator()\n# The generator takes noise and the target label as input\n# and generates the corresponding digit of that label\ngenerator.load_weights('../Q2/saved_model_weights/version1/generator_weights_99000.h5')\n\n# the classifier\npath_save_model = 'save_weight_classifier/version_1.h5'\nmodel = tf.keras.models.load_model(path_save_model)", "sub_path": "Lab 2 - Deep Learning (GAN)/Q3/load_different_cgan_model/cgan.py", "file_name": "cgan.py", "file_ext": "py", "file_size_in_byte": 2856, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "keras.datasets.mnist.load_data", "line_number": 21, "usage_type": "call"}, {"api_name": "keras.datasets.mnist", "line_number": 21, "usage_type": "name"}, {"api_name": "keras.optimizers.Adam", "line_number": 35, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 38, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 39, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 40, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 41, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 42, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 43, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 44, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 45, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 46, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 47, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.prod", "line_number": 48, "usage_type": "call"}, {"api_name": "keras.layers.Reshape", "line_number": 49, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 52, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 53, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 54, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 54, "usage_type": "call"}, {"api_name": "keras.layers.multiply", "line_number": 55, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 57, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 60, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.prod", "line_number": 61, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 62, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 63, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 64, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 65, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 66, "usage_type": "call"}, {"api_name": "keras.layers.advanced_activations.LeakyReLU", "line_number": 67, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 68, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 69, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 72, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 73, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 74, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.prod", "line_number": 74, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 75, "usage_type": "call"}, {"api_name": "keras.layers.multiply", "line_number": 77, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 79, "usage_type": "call"}, {"api_name": "tensorflow.keras.models.load_model", "line_number": 90, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 90, "usage_type": "attribute"}]}
+{"seq_id": "543791089", "text": "import torch\nfrom torch import optim\nimport time\n\nfrom models import *\nfrom datasets import *\nfrom loss import *\n\nbatch_size = 500\nv_batch_size = 100\nepoch = 22\n\ndevice = torch.device(\"cuda:0\" if torch.cuda.is_available() else \"cpu\")\ntorch.backends.cudnn.benchmark = True\n\ntrain_imgs, train_lbls, val_imgs, val_lbls = build_dataset(device=device)\nn_train = len(train_lbls)\nn_val = len(val_lbls)\n\n\nnet = build_network()\nnet.to(device).half()\nfor layer in net.modules():\n if isinstance(layer, nn.BatchNorm2d):\n layer.float()\n if hasattr(layer, 'weight') and layer.weight is not None:\n layer.weight.data.fill_(1.0)\n layer.eps = 0.00001\n layer.momentum = 0.1\n\ncriterion = nn.CrossEntropyLoss()\ncriterion2 = CrossEntropyLabelSmooth(num_classes=10, epsilon=0.2)\noptimizer = optim.SGD(net.parameters(), lr=0.2, momentum=0.9, nesterov=True, weight_decay=0.001)\n\ndef lr(e):\n if e < 4:\n return 0.5*e/3. + 0.01\n return 0.5*(22-e)/19. + 0.01\nsched = optim.lr_scheduler.LambdaLR(optimizer, lr)\n\naugment = Augment()\naugment.to(device).half()\n\nt_start = time.time()\nfor e in range(epoch): # loop over the dataset multiple times\n start = time.time()\n\n # process training set\n a_train = []\n for i in range(n_train//batch_size):\n # get the inputs; data is a list of [inputs, labels]\n inputs = train_imgs[i*batch_size:(i+1)*batch_size, ...]\n a_train.append(augment(inputs.to(device).half()))\n a_train_imgs = torch.cat(a_train)\n perm = torch.randperm(n_train)\n a_train_imgs = a_train_imgs[perm, ...].contiguous()\n a_train_lbls = train_lbls[perm].contiguous()\n\n # a_stop = time.time()\n\n net.train()\n running_loss = []\n perm = torch.randperm(n_train)\n # t1 = 0\n # t2 = 0\n # t3 = 0\n for i in range(n_train//batch_size):\n # s = time.time()\n # get the inputs; data is a list of [inputs, labels]\n inputs = a_train_imgs[i*batch_size: (i+1)*batch_size, ...]\n labels = a_train_lbls[i*batch_size: (i+1)*batch_size]\n\n # zero the parameter gradients\n optimizer.zero_grad()\n\n # forward + backward + optimize\n outputs = net(inputs)\n loss = criterion(outputs, labels)\n loss2 = criterion2(outputs, labels)\n loss = loss + 2*loss2\n # torch.cuda.synchronize()\n # t1 += time.time() - s\n loss.backward()\n # torch.cuda.synchronize()\n # t2 += time.time() - s\n optimizer.step()\n # torch.cuda.synchronize()\n # t3 += time.time() - s\n\n # print statistics\n running_loss.append(loss)\n running_loss = torch.stack(running_loss).mean().item()\n # t_stop = time.time()\n # t1 /= n_train//batch_size\n # t2 /= n_train//batch_size\n # t3 /= n_train//batch_size\n\n if e == 0 or e%5 == 1:\n net.eval()\n val_loss = []\n val_acc = []\n for i in range(n_val//v_batch_size):\n # get the inputs; data is a list of [inputs, labels]\n inputs = val_imgs[i*v_batch_size: (i+1)*v_batch_size, ...]\n labels = val_lbls[i*v_batch_size: (i+1)*v_batch_size]\n outputs = net(inputs)\n val_loss.append(criterion(outputs, labels))\n val_acc.append((outputs.argmax(dim=1) == labels).sum()/labels.shape[0])\n\n v_stop = time.time()\n # print('{} train loss {:5.02f} val loss {:5.02f} val acc {:5.02f} time a:{:5.03f} t:{:5.03f}, v:{:5.03f}, t1:{:5.03f}, t2:{:5.03f}, t3:{:5.03f} '.format(\n # e, running_loss, torch.stack(val_loss).mean(), 100.*torch.stack(val_acc).mean(), (a_stop-start), (t_stop-start), (v_stop - start), t1, t2, t3))\n print('{} train loss {:5.02f} val loss {:5.02f} val acc {:5.02f} time v:{:5.03f}'.format(\n e, running_loss, torch.stack(val_loss).mean(), 100.*torch.stack(val_acc).mean(), (v_stop - start)))\n sched.step()\nprint('Finished Training in {:5.03f}'.format(time.time()-t_start))\n\n\n", "sub_path": "cifar_singlerun.py", "file_name": "cifar_singlerun.py", "file_ext": "py", "file_size_in_byte": 3697, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "torch.device", "line_number": 13, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 13, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 13, "usage_type": "attribute"}, {"api_name": "torch.backends", "line_number": 14, "usage_type": "attribute"}, {"api_name": "torch.optim.SGD", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 33, "usage_type": "name"}, {"api_name": "torch.optim.lr_scheduler.LambdaLR", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 39, "usage_type": "attribute"}, {"api_name": "torch.optim", "line_number": 39, "usage_type": "name"}, {"api_name": "time.time", "line_number": 44, "usage_type": "call"}, {"api_name": "time.time", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 54, "usage_type": "call"}, {"api_name": "torch.randperm", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.randperm", "line_number": 63, "usage_type": "call"}, {"api_name": "loss.backward", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 92, "usage_type": "call"}, {"api_name": "time.time", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 114, "usage_type": "call"}, {"api_name": "time.time", "line_number": 116, "usage_type": "call"}]}
+{"seq_id": "25241299", "text": "from pathlib import Path\n\nimport fire\nimport yaml\n\nMODEL_CONFIG_PATH = Path(__file__).parent / \"../model_config\"\n\n\ndef main(name: str = \"\", ls: bool = False):\n if ls:\n for fpath in MODEL_CONFIG_PATH.glob(\"**/*.yml\"):\n with fpath.open() as f:\n cfg = yaml.safe_load(f)\n print(cfg[\"name\"])\n return\n fpath = MODEL_CONFIG_PATH / f\"{name}.yml\"\n print(fpath.read_text())\n\n\nif __name__ == \"__main__\":\n fire.Fire(main)\n", "sub_path": "camphr/cli/model_config.py", "file_name": "model_config.py", "file_ext": "py", "file_size_in_byte": 477, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pathlib.Path", "line_number": 6, "usage_type": "call"}, {"api_name": "yaml.safe_load", "line_number": 13, "usage_type": "call"}, {"api_name": "fire.Fire", "line_number": 21, "usage_type": "call"}]}
+{"seq_id": "344911225", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Mon Sep 28 13:38:33 2020\r\n\r\n@author: User\r\n\"\"\"\r\nimport matplotlib \r\nimport matplotlib.font_manager as fm\r\nimport matplotlib.pyplot as plt\r\nfrom matplotlib import font_manager, rc\r\nimport pandas as pd\r\nimport utils\r\nimport numpy as np\r\nimport pickle\r\nimport os\r\nfrom load_data import load_data\r\nimport seaborn as sns\r\n\r\n\r\ndef make_product_key(): \r\n product_list = ['가구','가전','건강기능','농수축','생활용품',\r\n '속옷','의류','이미용','잡화','주방','침구']\r\n product_dict = {'상품군_%s'%k : k for k in product_list}\r\n return product_dict\r\n\r\n\r\ndef make_season_key(): \r\n season_dict = {'계절_0':0,'계절_1':1,'계절_2':2,'계절_3':3}\r\n return season_dict\r\n\r\ndef make_month_key():\r\n month_dict = {'��_%s'%k : k for k in range(1,13)}\r\n return month_dict\r\n\r\ndef make_week_key():\r\n\r\n week_dict = {'주차_%s'%k : k for k in range(1,54)}\r\n return week_dict\r\n\r\ndef make_hour_key():\r\n\r\n hour_dict = {'시간대_%s'%k : k for k in [0,1,2,6,7,8,9,10,11,12,13,14,15,16,\r\n 17,18,19,20,21,22,23]}\r\n return hour_dict\r\n\r\n\r\ndef make_not_enough_dummy(dummy_df, col):\r\n result_df = dummy_df.copy()\r\n element_list = np.unique(dummy_df['name'])\r\n for element in element_list:\r\n temp_col_name = '%s_%s'%(col,element)\r\n result_df[temp_col_name] = 0\r\n \r\n ind = dummy_df[dummy_df['name'] == element].index\r\n result_df.loc[ind,temp_col_name] = 1\r\n \r\n if col == '계절':\r\n target_dict = make_season_key()\r\n elif col == '월':\r\n target_dict = make_month_key()\r\n elif col == '주차':\r\n target_dict = make_week_key()\r\n elif col == '상품군':\r\n target_dict = make_product_key()\r\n elif col == '시간대':\r\n target_dict = make_hour_key()\r\n for key in target_dict.keys():\r\n if key not in result_df.columns:\r\n result_df[key] = 0\r\n \r\n result_df = result_df.loc[:,['name']+[k for k in target_dict.keys()]]\r\n \r\n return result_df, target_dict\r\n \r\n\r\n\r\ndef make_dummy(df, col_name):\r\n #col_name = 'new_월'\r\n not_enough_col_list = ['월','주차','계절','상품군','시간대']\r\n dummy = pd.get_dummies(df[col_name], prefix=col_name)\r\n \r\n dummy_df = pd.DataFrame({'name':df[col_name]})\r\n dummy_df = pd.concat([dummy_df, dummy], axis=1)\r\n \r\n name_dict = {}\r\n for col in dummy_df.columns[1:]:\r\n target = dummy_df[dummy_df[col]!=0].reset_index(drop=True).loc[0,['name',col]]\r\n name_dict[col] = target['name']\r\n \r\n # 보충\r\n if col_name in not_enough_col_list:\r\n dummy_df, name_dict = make_not_enough_dummy(dummy_df, col_name) \r\n \r\n col_stand_name = '_'.join(dummy_df.columns[1].split('_')[:-1])\r\n\r\n col_list = ['%s_%s'%(col_stand_name,v) for v in name_dict.values()]\r\n \r\n dummy_df = dummy_df.iloc[:,1:]\r\n dummy_df.columns = col_list\r\n\r\n df = df.drop(col_name, axis=1)\r\n df = pd.concat([df, dummy_df], axis=1)\r\n\r\n return df\r\n\r\ndef extract_dummy(X_data, save_path):\r\n \r\n # 더미변수 추출\r\n with open(save_path+'model_setting.txt', 'r') as f1:\r\n while True:\r\n line = f1.readline()\r\n if 'dummy' in line:\r\n dummy_var = line.split('=')[1]\r\n \r\n if not line:\r\n break\r\n \r\n dummy_list = dummy_var.split(',')\r\n dummy_list[-1] = dummy_list[-1][:-1]\r\n \r\n for col in dummy_list:\r\n X_data = make_dummy(X_data, col)\r\n \r\n return X_data\r\n\r\n\r\n\r\ndef load_x_col(save_path):\r\n\r\n with open(save_path+'test_result.pkl', 'rb') as f:\r\n obj_dict = pickle.load(f)\r\n \r\n x_col = obj_dict['result']['data']['changed_x_col']\r\n \r\n return x_col\r\n\r\nfont = {'family' : 'normal',\r\n 'weight' : 'bold',\r\n 'size' : 22}\r\n\r\nmatplotlib.rc('font', **font)\r\n\r\nfm.get_fontconfig_fonts()\r\nplt.rcParams['axes.unicode_minus'] = False\r\nmatplotlib.rc('font', family='Malgun Gothic')\r\n\r\nmodel_path = utils.model_path\r\n\r\ntrain_path = utils.train_path\r\nfig_path = utils.fig_path\r\n#%%\r\ndef load_model(save_path):\r\n # 0. 최적 모델 가져오기\r\n \r\n with open(save_path+'test_result.pkl', 'rb') as f:\r\n result = pickle.load(f)\r\n \r\n return result\r\n\r\n\r\ndef predict_value(X_data, save_path,fig_save_path ):\r\n # 모델\r\n result = load_model(save_path)\r\n\r\n # 예측값 구하기\r\n pred_list = []\r\n X_data = extract_dummy(X_data, save_path)\r\n x_col = load_x_col(save_path)\r\n X_data = X_data.loc[:,x_col]\r\n X_data = X_data.fillna(0)\r\n X_data = X_data.values\r\n \r\n for i in range(5):\r\n valid_ind = result['result']['data']['dataset']['valid'][i]\r\n\r\n with open(save_path+'model_%s.pkl'%(i), 'rb') as f:\r\n model = pickle.load(f)\r\n pred = model.predict(X_data).reshape(-1,)\r\n pred_list.append(pred)\r\n \r\n pred_list = np.mean(np.stack(pred_list),axis=0)\r\n \r\n plt.figure(figsize=(24,16))\r\n sns.distplot(pred_list)\r\n plt.title('test 예측값', size=30)\r\n plt.xlabel('취급액')\r\n plt.ylabel('scaled_빈도')\r\n plt.savefig(fig_save_path+'test 데이터 분포.png')\r\n plt.show()\r\n \r\n return pred_list\r\n\r\n\r\ndef extract_feature_importance(save_path, fig_save_path, fig_save_option=False):\r\n\r\n result = load_model(save_path) \r\n\r\n x_col = result['result']['data']['changed_x_col']\r\n \r\n feature_df = pd.DataFrame({'x_col':x_col})\r\n for i in range(5):\r\n feature = result['result']['feature'][i]\r\n feature_df[i] = feature\r\n \r\n feature_df['important'] = np.mean(feature_df.iloc[:,1:],axis=1)\r\n feature_df = feature_df[['x_col','important']]\r\n feature_df = feature_df.sort_values('important',ascending=False).reset_index(drop=True)\r\n \r\n plt.figure(figsize=(24,16))\r\n plt.bar(feature_df.loc[:29,'x_col'], feature_df.loc[:29,'important'])\r\n plt.title('변수 중요도')\r\n plt.ylabel('중요도')\r\n plt.xlabel('변수이름')\r\n plt.xticks(rotation=40, size=15)\r\n if fig_save_option == True:\r\n plt.savefig(fig_save_path+'변수 중요도.png')\r\n plt.show()\r\n \r\n return 0\r\n\r\ndef main(save_path = model_path+'xgb_bds100/'):\r\n # 1. 파일 불러오기\r\n train = load_data('train.csv', 'train_WordVec.pkl', True)\r\n test = load_data('test.csv', 'test_WordVec.pkl',True)\r\n \r\n # 예측값 추출\r\n y_pred = predict_value(test, save_path, fig_path+'모델/')\r\n \r\n # 저장\r\n np.save(train_path+'test_pred.npy',y_pred)\r\n \r\n # feature_importance\r\n extract_feature_importance(save_path, fig_path+'모델/', fig_save_option=True)\r\n\r\nif __name__ == '__main__':\r\n main()\r\n \r\n", "sub_path": "code/predict.py", "file_name": "predict.py", "file_ext": "py", "file_size_in_byte": 6784, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "numpy.unique", "line_number": 49, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 80, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 82, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 83, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 102, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 141, "usage_type": "call"}, {"api_name": "matplotlib.font_manager.get_fontconfig_fonts", "line_number": 143, "usage_type": "call"}, {"api_name": "matplotlib.font_manager", "line_number": 143, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 144, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 144, "usage_type": "name"}, {"api_name": "matplotlib.rc", "line_number": 145, "usage_type": "call"}, {"api_name": "utils.model_path", "line_number": 147, "usage_type": "attribute"}, {"api_name": "utils.train_path", "line_number": 149, "usage_type": "attribute"}, {"api_name": "utils.fig_path", "line_number": 150, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 156, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 181, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}, {"api_name": "seaborn.distplot", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 185, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 185, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 186, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 186, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 187, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 187, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 188, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 188, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 189, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 205, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 209, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 209, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 210, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 210, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 211, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 211, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 212, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 212, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 213, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 213, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 214, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 214, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 216, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 216, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 217, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 217, "usage_type": "name"}, {"api_name": "load_data.load_data", "line_number": 223, "usage_type": "call"}, {"api_name": "load_data.load_data", "line_number": 224, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 230, "usage_type": "call"}]}
+{"seq_id": "507696612", "text": "import torch\nimport matplotlib.pyplot as plt\nfrom torchvision import transforms\nfrom PIL import Image as I\nfrom torch.autograd import Variable\n\nimport evalMethod\n\n# GPU\ndevice = 'cuda' if torch.cuda.is_available() else 'cpu'\n\ntorch.manual_seed(777)\nif device == 'cuda':\n torch.cuda.manual_seed_all(777)\n print(\"GPU is available!!!\")\n\nimgA = I.open('data/A.jpeg')\nimgB = I.open('data/B.jpeg')\n\ntf = transforms.Compose([\n transforms.Resize((255,255)),\n transforms.ToTensor()\n])\n\nimgA_tensor = 255*tf(imgA)\nimgB_tensor = 255*tf(imgB)\n\nimg1 = Variable(torch.unsqueeze(imgA_tensor,0))\nimg2 = Variable(torch.unsqueeze(imgB_tensor,0))\n\nprint(evalMethod.ssim(img1, img2))\n\nplt.imshow(imgA)\nplt.title('A')\nplt.axis('off')\nplt.show(block=True)\n\nplt.imshow(imgB)\nplt.title('B')\nplt.axis('off')\nplt.show(block=True)", "sub_path": "imageEvaluation/.ipynb_checkpoints/testEval-checkpoint.py", "file_name": "testEval-checkpoint.py", "file_ext": "py", "file_size_in_byte": 815, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "torch.cuda.is_available", "line_number": 10, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 10, "usage_type": "attribute"}, {"api_name": "torch.manual_seed", "line_number": 12, "usage_type": "call"}, {"api_name": "torch.cuda.manual_seed_all", "line_number": 14, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 14, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 17, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 17, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 18, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 18, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 20, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 20, "usage_type": "name"}, {"api_name": "torchvision.transforms.Resize", "line_number": 21, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 21, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 22, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 22, "usage_type": "name"}, {"api_name": "torch.autograd.Variable", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 29, "usage_type": "call"}, {"api_name": "evalMethod.ssim", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}]}
+{"seq_id": "578453624", "text": "#!/usr/bin/python3\n\nimport sys\nimport logging\nfrom mowcounterbot import MowCounterTelegramBot\n\nlogging.basicConfig(level=logging.DEBUG,\n format='%(asctime)s - %(name)s - %(levelname)s - %(message)s')\n\n\ndef main():\n try:\n bot = MowCounterTelegramBot(MowCounterTelegramBot.parse_cli_arguments())\n except RuntimeError as e:\n print(e)\n sys.exit(1)\n print(\"Starting up bot\")\n bot.setup_commands()\n bot.start_loop()\n bot.shutdown()\n print(\"Shutting down bot\")\n\nif __name__ == \"__main__\":\n main()\n", "sub_path": "mowcounter_bot.py", "file_name": "mowcounter_bot.py", "file_ext": "py", "file_size_in_byte": 554, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.basicConfig", "line_number": 7, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 7, "usage_type": "attribute"}, {"api_name": "mowcounterbot.MowCounterTelegramBot", "line_number": 13, "usage_type": "call"}, {"api_name": "mowcounterbot.MowCounterTelegramBot.parse_cli_arguments", "line_number": 13, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 16, "usage_type": "call"}]}
+{"seq_id": "279903549", "text": "from django.conf.urls import patterns, url\n\nfrom selector import views\n\nurlpatterns = patterns('',\n url(r'^$', views.index, name='index'),\n url(r'^show/', views.show, name='show'),\n url(r'^select/', views.select, name='select'),\n url(r'^doselect/', views.doselect, name='doselect'),\n )\n", "sub_path": "AnotherCoursePro/selector/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 395, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.conf.urls.patterns", "line_number": 5, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 6, "usage_type": "call"}, {"api_name": "selector.views.index", "line_number": 6, "usage_type": "attribute"}, {"api_name": "selector.views", "line_number": 6, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "selector.views.show", "line_number": 7, "usage_type": "attribute"}, {"api_name": "selector.views", "line_number": 7, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "selector.views.select", "line_number": 8, "usage_type": "attribute"}, {"api_name": "selector.views", "line_number": 8, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "selector.views.doselect", "line_number": 9, "usage_type": "attribute"}, {"api_name": "selector.views", "line_number": 9, "usage_type": "name"}]}
+{"seq_id": "11745840", "text": "# -*- coding: utf-8 -*-\r\nfrom datetime import datetime\r\n\r\nfrom django.shortcuts import redirect, render\r\n\r\nfrom ..forms import PaginaForm\r\nfrom ..models import Pagina\r\nfrom ..comum.contents import reescrever_url, save_in_portal_catalog, get_site_url\r\nfrom ..comum.contents import get_site_url_id, get_url_id_content, save_indice_url, format_visao_by_delete\r\nfrom security.anotation import permission_content\r\n\r\n\r\nTEMPLATE = '%s/documents.html' % 'comum'\r\n\r\n@permission_content(tipo='ATPagina', permissao='create', login_url='/security/login/')\r\ndef create(request):\r\n path_url = reescrever_url(request)\r\n form = PaginaForm(request.POST or None,)\r\n site = get_site_url(request)\r\n if form.is_valid():\r\n model = form.save(commit=False)\r\n _url = save_indice_url(request, model.titulo)\r\n model.url = _url\r\n model.tipo = 'ATPagina'\r\n model.site = site\r\n model.dono = request.user\r\n model.save()\r\n path_url += _url + '/'\r\n save_in_portal_catalog(model, path_url)\r\n return redirect(path_url)\r\n\r\n context = {\r\n 'form' : form,\r\n 'editor' : True,\r\n }\r\n \r\n _site_url = get_site_url_id(request)\r\n template = '%s/documents.html' % _site_url\r\n try:\r\n return render(request, template, context)\r\n except:\r\n return render(request, TEMPLATE, context)\r\n\r\n@permission_content(tipo='ATPagina', permissao='update', login_url='/security/login/')\r\ndef edit(request):\r\n _url = reescrever_url(request)\r\n _site_url = get_site_url_id(request)\r\n _content_url_id = get_url_id_content(request)\r\n _object = Pagina.objects.filter(site__url=_site_url).get(url=_content_url_id)\r\n form = PaginaForm(request.POST or None, instance=_object)\r\n if form.is_valid():\r\n model = form.save(commit=False)\r\n model.update_at = datetime.now()\r\n model.save()\r\n save_in_portal_catalog(model)\r\n return redirect(_url)\r\n context = {\r\n 'form' : form,\r\n 'editor' : True,\r\n }\r\n \r\n template = '%s/documents.html' % _site_url\r\n try:\r\n return render(request, template, context)\r\n except:\r\n return render(request, TEMPLATE, context)\r\n\r\n@permission_content(tipo='ATPagina', permissao='delete', login_url='/security/login/')\r\ndef delete(request, portal_catalog):\r\n content_url = get_url_id_content(request)\r\n content = portal_catalog.get_content_object()\r\n content.delete()\r\n portal_catalog.delete()\r\n _new_url = reescrever_url(request)\r\n _new_url = _new_url.replace('/'+content_url, '')\r\n _site_url = get_site_url_id(request)\r\n #verica ocorrencia de visão do content na pasta e formata para visão sumaria\r\n format_visao_by_delete(_site_url, _new_url)\r\n return redirect(_new_url)\r\n\r\n@permission_content(tipo='ATPagina', permissao='workflow', login_url='/security/login/')\r\ndef workflow(request, portal_catalog, _workflow):\r\n _site_url = get_site_url_id(request)\r\n _o = Pagina.objects.filter(site__url=_site_url).get(url=portal_catalog.url)\r\n _o.workflow = _workflow\r\n if _o.workflow == 'Publicado' and _o.public_at==None:\r\n _o.public_at = datetime.now()\r\n _o.save()\r\n save_in_portal_catalog(_o)\r\n ", "sub_path": "portalufopa/comum/paginas.py", "file_name": "paginas.py", "file_ext": "py", "file_size_in_byte": 3231, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "comum.contents.reescrever_url", "line_number": 17, "usage_type": "call"}, {"api_name": "forms.PaginaForm", "line_number": 18, "usage_type": "call"}, {"api_name": "comum.contents.get_site_url", "line_number": 19, "usage_type": "call"}, {"api_name": "comum.contents.save_indice_url", "line_number": 22, "usage_type": "call"}, {"api_name": "comum.contents.save_in_portal_catalog", "line_number": 29, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 30, "usage_type": "call"}, {"api_name": "comum.contents.get_site_url_id", "line_number": 37, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 40, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 42, "usage_type": "call"}, {"api_name": "security.anotation.permission_content", "line_number": 15, "usage_type": "call"}, {"api_name": "comum.contents.reescrever_url", "line_number": 46, "usage_type": "call"}, {"api_name": "comum.contents.get_site_url_id", "line_number": 47, "usage_type": "call"}, {"api_name": "comum.contents.get_url_id_content", "line_number": 48, "usage_type": "call"}, {"api_name": "models.Pagina.objects.filter", "line_number": 49, "usage_type": "call"}, {"api_name": "models.Pagina.objects", "line_number": 49, "usage_type": "attribute"}, {"api_name": "models.Pagina", "line_number": 49, "usage_type": "name"}, {"api_name": "forms.PaginaForm", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 53, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 53, "usage_type": "name"}, {"api_name": "comum.contents.save_in_portal_catalog", "line_number": 55, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 56, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 64, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 66, "usage_type": "call"}, {"api_name": "security.anotation.permission_content", "line_number": 44, "usage_type": "call"}, {"api_name": "comum.contents.get_url_id_content", "line_number": 70, "usage_type": "call"}, {"api_name": "comum.contents.reescrever_url", "line_number": 74, "usage_type": "call"}, {"api_name": "comum.contents.get_site_url_id", "line_number": 76, "usage_type": "call"}, {"api_name": "comum.contents.format_visao_by_delete", "line_number": 78, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 79, "usage_type": "call"}, {"api_name": "security.anotation.permission_content", "line_number": 68, "usage_type": "call"}, {"api_name": "comum.contents.get_site_url_id", "line_number": 83, "usage_type": "call"}, {"api_name": "models.Pagina.objects.filter", "line_number": 84, "usage_type": "call"}, {"api_name": "models.Pagina.objects", "line_number": 84, "usage_type": "attribute"}, {"api_name": "models.Pagina", "line_number": 84, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 87, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 87, "usage_type": "name"}, {"api_name": "comum.contents.save_in_portal_catalog", "line_number": 89, "usage_type": "call"}, {"api_name": "security.anotation.permission_content", "line_number": 81, "usage_type": "call"}]}
+{"seq_id": "591212337", "text": "from math import ceil\n\nfrom scipy.fftpack import diff\nfrom scipy.signal import medfilt, lfilter\nfrom scipy.signal import resample, filtfilt\nfrom scipy.stats import mode\n\nfrom utils import matlab, common\nfrom utils.matlab import *\n\n\ndef frange(x, y, jump):\n while x < y:\n yield x\n x += jump\n\n\ndef qrs_detect2(ecg, thres=0.6, ref_period=0.25, fs=300):\n \"\"\"\n QRS detector based on the P&T method\n See: https://github.com/alistairewj/peak-detector/blob/master/sources/qrs_detect2.m\n :param ecg: one ecg channel on which to run the detector\n :param thres: energy threshold of the detector [arbitrary units]\n :param ref_period: refractory period in sec between two R-peaks [ms]\n :param fs: sampling frequency [Hz]\n :return: list, positions of R picks\n \"\"\"\n\n WIN_SAMP_SZ = 7\n MED_SMOOTH_NB_COEFF = int(round(fs / 100))\n INT_NB_COEFF = int(round(WIN_SAMP_SZ * fs / 256))\n SEARCH_BACK = True\n MAX_FORCE = []\n MIN_AMP = 0.1\n\n NB_SAMP = len(ecg)\n\n tm = list(frange(1 / fs, ceil(NB_SAMP / fs), 1 / fs))\n\n # == Bandpass filtering for ECG signal\n # this sombrero hat has shown to give slightly better results than a\n # standard band-pass filter. Plot the frequency response to convince\n # yourself of what it does\n b1 = [-7.757327341237223e-05, -2.357742589814283e-04, -6.689305101192819e-04, -0.001770119249103,\n -0.004364327211358, -0.010013251577232, -0.021344241245400, -0.042182820580118, -0.077080889653194,\n -0.129740392318591, -0.200064921294891, -0.280328573340852, -0.352139052257134, - 0.386867664739069,\n -0.351974030208595, -0.223363323458050, 0, 0.286427448595213, 0.574058766243311,\n 0.788100265785590, 0.867325070584078, 0.788100265785590, 0.574058766243311, 0.286427448595213, 0,\n -0.223363323458050, -0.351974030208595, -0.386867664739069, -0.352139052257134,\n -0.280328573340852, -0.200064921294891, -0.129740392318591, -0.077080889653194, -0.042182820580118,\n -0.021344241245400, -0.010013251577232, -0.004364327211358, -0.001770119249103, -6.689305101192819e-04,\n -2.357742589814283e-04, -7.757327341237223e-05]\n\n # NOTE: resample works differently than in matlab\n b1 = resample(b1, int(ceil(len(b1) * fs / 250)))\n bpfecg = np.transpose(filtfilt(b1, 1, ecg))\n\n if (sum(abs(ecg - common.mode(ecg)) > MIN_AMP) / NB_SAMP) > 0.2:\n \"\"\"\n if 20% of the samples have an absolute amplitude which is higher\n than MIN_AMP then we are good to go\n \"\"\"\n\n # == P&T operations\n dffecg = matlab.diff(np.transpose(bpfecg))\n sqrecg = [x * x for x in dffecg]\n intecg = lfilter(np.ones(INT_NB_COEFF), 1, sqrecg)\n mdfint = medfilt(intecg, [MED_SMOOTH_NB_COEFF])\n delay = int(ceil(INT_NB_COEFF / 2))\n mdfint = np.roll(mdfint, -delay)\n\n mdfintFidel = mdfint\n\n if NB_SAMP / fs > 90:\n xs = np.sort(mdfintFidel[fs:fs * 90])\n else:\n xs = np.sort(mdfintFidel[fs:])\n\n if len(MAX_FORCE) == 0:\n if NB_SAMP / fs > 10:\n ind_xs = ceil(98 / 100 * len(xs))\n en_thres = xs[ind_xs]\n else:\n ind_xs = ceil(99 / 100 * len(xs))\n en_thres = xs[ind_xs]\n else:\n en_thres = MAX_FORCE\n\n poss_reg = apply(mdfint, lambda x: x > (thres * en_thres))\n\n if len(poss_reg) == 0:\n poss_reg[10] = 1\n\n if SEARCH_BACK:\n # ind of samples above threshold\n indAboveThreshold = find(poss_reg, lambda x: x > 0)\n # compute RRv\n RRv = np.diff([tm[i] for i in indAboveThreshold])\n medRRv = mode([RRv[i] for i in find(RRv, lambda x: x > 0.01)]).mode[0]\n # missed a peak?\n indMissedBeat = find(RRv, lambda x: x > 1.5 * medRRv)\n # find interval onto which a beat might have been missed\n indStart = [indAboveThreshold[i] for i in indMissedBeat]\n indEnd = [indAboveThreshold[i + 1] for i in indMissedBeat]\n\n for i in range(len(indStart)):\n # look for a peak on this interval by lowering the energy threshold\n poss_reg[indStart[i]:indEnd[i]] = apply(mdfint[indStart[i]:indEnd[i]],\n lambda x: x > 0.3 * thres * en_thres)\n\n left = find(diff(np.append([0], np.transpose(poss_reg))), lambda x: x == 1)\n right = find(diff(np.append(np.transpose(poss_reg), [0])), lambda x: x == -1)\n\n all = [(left, right) for left, right in zip(left, right) if left != right]\n\n left = [x[0] for x in all]\n right = [x[1] for x in all]\n\n nb_s = len(apply(left, lambda x: x < 30 * fs))\n loc = np.zeros(nb_s, dtype=np.int32)\n for j in range(nb_s):\n a, loc[j] = np_max(abs(bpfecg[left[j]:right[j]]))\n loc[j] = loc[j] + left[j]\n sign = np.mean([ecg[i] for i in loc])\n\n compt = 0\n NB_PEAKS = len(left)\n maxval = np.zeros(NB_PEAKS)\n maxloc = np.zeros(NB_PEAKS, dtype=np.int32)\n\n for i in range(NB_PEAKS):\n if sign > 0:\n v, l = np_max(ecg[left[i]:right[i]])\n else:\n v, l = np_min(ecg[left[i]:right[i]])\n\n maxval[compt] = v\n maxloc[compt] = l + left[i]\n\n if compt > 0:\n if maxloc[compt] - maxloc[compt - 1] < fs * ref_period and abs(maxval[compt]) < abs(maxval[compt - 1]):\n continue\n elif maxloc[compt] - maxloc[compt - 1] < fs * ref_period and abs(maxval[compt]) >= abs(\n maxval[compt - 1]):\n maxloc[compt - 1] = maxloc[compt]\n maxval[compt - 1] = maxval[compt]\n else:\n compt += 1\n else:\n # if first peak then increment\n compt += 1\n\n # datapoints QRS positions\n qrs_pos = maxloc[:compt]\n else:\n qrs_pos = []\n sign = None\n en_thres = None\n\n return qrs_pos\n", "sub_path": "extract_features/biosppyex/signals/qrs_detect2.py", "file_name": "qrs_detect2.py", "file_ext": "py", "file_size_in_byte": 6121, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "math.ceil", "line_number": 38, "usage_type": "call"}, {"api_name": "scipy.signal.resample", "line_number": 55, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 55, "usage_type": "call"}, {"api_name": "scipy.signal.filtfilt", "line_number": 56, "usage_type": "call"}, {"api_name": "utils.common.mode", "line_number": 58, "usage_type": "call"}, {"api_name": "utils.common", "line_number": 58, "usage_type": "name"}, {"api_name": "utils.matlab.diff", "line_number": 65, "usage_type": "call"}, {"api_name": "utils.matlab", "line_number": 65, "usage_type": "name"}, {"api_name": "scipy.signal.lfilter", "line_number": 67, "usage_type": "call"}, {"api_name": "scipy.signal.medfilt", "line_number": 68, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 69, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 81, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 84, "usage_type": "call"}, {"api_name": "scipy.stats.mode", "line_number": 99, "usage_type": "call"}, {"api_name": "scipy.fftpack.diff", "line_number": 111, "usage_type": "call"}, {"api_name": "scipy.fftpack.diff", "line_number": 112, "usage_type": "call"}]}
+{"seq_id": "586186951", "text": "import config\nimport requests\nimport re\nimport json\nimport traceback\n\n\ndef spider(city, job):\n # ################得到页数\n url_page = 'https://www.lagou.com/jobs/list_' + job + '?px=default&city=' + city + '#filterBox'\n html = requests.get(url_page, headers=config.headers).text\n try:\n pages = re.findall('totalNum\">(.*?)<', html, re.S)[0]\n except Exception as e:\n traceback.print_exc()\n pages = 0\n print(pages)\n\n # 最终的完整数据\n final_data = []\n\n # 对每一页进行解析\n for i in range(int(pages)):\n res = get_one_page(city, job, i + 1)\n if res is None:\n continue\n final_data.extend(res)\n\n return final_data\n\n\n# 得到一页的数据\ndef get_one_page(city, job, page):\n # ################得到json数据\n # url模板 https://www.lagou.com/jobs/positionAjax.json?px=default&city=北京&needAddtionalResult=false\n # post数据\n # {\n # first: false,\n # pn: 1, 页码\n # kd: Java 工作\n # }\n\n # 构造post数据\n post_data = {\n 'first': 'false',\n 'pn': str(page),\n 'kd': job\n }\n # 打开网页,得到json数据\n url = 'https://www.lagou.com/jobs/positionAjax.json?px=default&city=' + city + '&needAddtionalResult=false'\n html = requests.post(url, data=post_data, headers=config.headers, timeout=config.time_out).text\n if len(html) < 100:\n return None\n\n # 解析json\n json_data = json.loads(html)\n\n # 最终数据的列表\n res_data = json_data['content']['positionResult']['result']\n\n # 清洗后的数据\n _data = []\n for each in res_data:\n job_url = 'https://www.lagou.com/jobs/' + str(each['positionId']) + '.html'\n item = {'posname': each['positionName'], 'company': each['companyShortName'], 'salary': each['salary'],\n 'workYear': each['workYear'], 'education': each['education'], 'financeStage': each['financeStage'],\n 'positionAdvantage': each['positionAdvantage'], 'industryField': each['industryField'],\n 'job_url': job_url}\n # 尝试解析dict,如果失败,略过\n try:\n item = str(item).replace(\"'\", '\"')\n item = json.loads(item)\n except Exception as e:\n print(each['companyShortName'] + \"parse error\")\n traceback.print_exc()\n continue\n _data.append(item)\n\n return _data", "sub_path": "spider.py", "file_name": "spider.py", "file_ext": "py", "file_size_in_byte": 2450, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "requests.get", "line_number": 11, "usage_type": "call"}, {"api_name": "config.headers", "line_number": 11, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 13, "usage_type": "call"}, {"api_name": "re.S", "line_number": 13, "usage_type": "attribute"}, {"api_name": "traceback.print_exc", "line_number": 15, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 51, "usage_type": "call"}, {"api_name": "config.headers", "line_number": 51, "usage_type": "attribute"}, {"api_name": "config.time_out", "line_number": 51, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 56, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 72, "usage_type": "call"}, {"api_name": "traceback.print_exc", "line_number": 75, "usage_type": "call"}]}
+{"seq_id": "242599782", "text": "import gym\nimport numpy as np\n\nenv = gym.make(\"Taxi-v2\")\n\n#Algorithm memory\n#This is a table which stores all combos of states and actions\n#The entered values will be associated rewards with these combos\nQ = np.zeros([env.observation_space.n, env.action_space.n])\n\n#Total accumulated reward for each session\nG = 0\n\n#Learning rate\nalpha = 0.618\n\nfor episode in range(1, 1001):\n #initialize the session\n done = False\n G, reward = 0, 0\n state = env.reset()\n #main loop\n counter = 0\n while done != True:\n counter += 1\n #Do the most valuable action available\n action = np.argmax(Q[state])\n #Collect the info of that action\n state2, reward, done, info = env.step(action)\n #Update the reward for this particular action\n #Add the actual reward plus the new potential reward (in our new state)\n #minus the previous total reward?\n Q[state, action] += alpha*(reward+np.max(Q[state2])-Q[state, action])\n G += reward\n state = state2\n if episode % 50 == 0:\n print('Episode {} Total Reward: {}'.format(episode, G))\n print('Moves taken to finish: {}'.format(counter))\n", "sub_path": "GymNet/gymTestNet/learningAlgTaxi.py", "file_name": "learningAlgTaxi.py", "file_ext": "py", "file_size_in_byte": 1167, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "gym.make", "line_number": 4, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 33, "usage_type": "call"}]}
+{"seq_id": "69017610", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Nov 15 09:37:25 2018\n문장 입력 다중클래스분류 모델\n참조: https://tykimos.github.io/2017/08/17/Text_Input_Multiclass_Classification_Model_Recipe/\n@author: SDEDU\n\"\"\"\n\nfrom keras.datasets import reuters\nfrom keras.utils import np_utils\nfrom keras.preprocessing import sequence\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Embedding, LSTM\n\n#1. 데이터셋 생성\nmax_features=15000\ntext_max_words=120\n\n#훈련셋과 시험셋 불러오기\n(x_train,y_train),(x_test,y_test)=reuters.load_data(num_words=max_features)\n\n#훈련셋과 검증셋 분리\nx_val=x_train[7000:]\ny_val=y_train[7000:]\nx_train=x_train[:7000]\ny_train=y_train[:7000]\n\n#데이터셋 전처리 :문장길이 맞추기\n#문장길이를 maxlen인자로 맞춤(=120보다 짧은 문장은 0을 채워서 120단어로 맞춰주고 120보다 긴 문장은 120단어까지 잘라냄)\nx_train=sequence.pad_sequences(x_train,maxlen=text_max_words)\nx_val=sequence.pad_sequences(x_val,maxlen=text_max_words)\nx_test=sequence.pad_sequences(x_test,maxlen=text_max_words)\n\n#one-hot 인코딩 onverts a class vecoter (integers) to binary class matrix\ny_train=np_utils.to_categorical(y_train)\ny_val=np_utils.to_categorical(y_val)\ny_test=np_utils.to_categorical(y_test)\n\n\n#2. 모델 구성하기\n#순환 신경망 모델\nmodel=Sequential()\nmodel.add(Embedding(max_features,128))\nmodel.add(LSTM(128))\nmodel.add(Dense(46,activation='softmax'))\n\n#3. 모델 학습과정 설정하기\nmodel.compile(loss='categorical_crossentropy',optimizer='adam',metrics=['accuracy'])\n\n#4. 모델 학습시키기\nhist=model.fit(x_train,y_train,epochs=10, batch_size=64,validation_data=(x_val,y_val))\n\n#5. 학습과정 살펴보기\nimport matplotlib.pyplot as plt\nfig,loss_ax=plt.subplots()\n\nacc_ax=loss_ax.twinx()\n\nloss_ax.plot(hist.history['loss'],'y',label='train loss')\nloss_ax.plot(hist.history['val_loss'],'r',label='val loss')\nloss_ax.set_ylim([0.0,3.0])\n\nacc_ax.plot(hist.history['acc'],'b',label='train acc')\nacc_ax.plot(hist.history['val_acc'],'g',label='val acc')\nacc_ax.set_ylim([0.0,1.0])\n\nloss_ax.set_xlabel('epoch')\nloss_ax.set_ylabel('loss')\nacc_ax.set_ylabel('accuracy')\n\nloss_ax.legend(loc='upper left')\nacc_ax.legend(loc='lower left')\n\n#6. 모델평가하기\nloss_and_metrics=model.evaluate(x_test,y_test,batch_size=64)\nprint(loss_and_metrics)\n\n\n", "sub_path": "Python/Lab30/Lab30.py", "file_name": "Lab30.py", "file_ext": "py", "file_size_in_byte": 2371, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "keras.datasets.reuters.load_data", "line_number": 20, "usage_type": "call"}, {"api_name": "keras.datasets.reuters", "line_number": 20, "usage_type": "name"}, {"api_name": "keras.preprocessing.sequence.pad_sequences", "line_number": 30, "usage_type": "call"}, {"api_name": "keras.preprocessing.sequence", "line_number": 30, "usage_type": "name"}, {"api_name": "keras.preprocessing.sequence.pad_sequences", "line_number": 31, "usage_type": "call"}, {"api_name": "keras.preprocessing.sequence", "line_number": 31, "usage_type": "name"}, {"api_name": "keras.preprocessing.sequence.pad_sequences", "line_number": 32, "usage_type": "call"}, {"api_name": "keras.preprocessing.sequence", "line_number": 32, "usage_type": "name"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 35, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 35, "usage_type": "name"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 36, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 36, "usage_type": "name"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 37, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 37, "usage_type": "name"}, {"api_name": "keras.models.Sequential", "line_number": 42, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 43, "usage_type": "call"}, {"api_name": "keras.layers.LSTM", "line_number": 44, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}]}
+{"seq_id": "257673265", "text": "from django.conf.urls import include, url\nfrom django.contrib import admin\nfrom .views import (\n RestaurantListAPIView, RestaurantDetailAPIView, \n RestaurantCreateAPIView, RestaurantUpdateAPIView,\n OperatingTimeCreateAPIView, OperatingTimeUpdateAPIView,\n MenuListAPIView, MenuDetailAPIView, \n MenuCreateAPIView, MenuUpdateAPIView, MenuCategoryDetailAPIView\n )\n\nurlpatterns = [\n url(r'^$', RestaurantListAPIView.as_view(), name=\"restaurantlistapiview\"),\n url(r'^menu/$', MenuListAPIView.as_view(), name=\"menu_list_api\"),\n url(r'^create/$', RestaurantCreateAPIView.as_view(), name=\"restaurantcreateapiview\"),\n url(r'^update/(?P[\\w-]+)$', RestaurantUpdateAPIView.as_view(), name=\"restaurantupdateapiview\"),\n url(r'^(?P[\\w-]+)/$', RestaurantDetailAPIView.as_view(), name=\"restaurantdetailapiview\"),\n url(r'^menu/(?P[\\w-]+)/$', MenuDetailAPIView.as_view(), name=\"menu_detail_api\"),\n url(r'^menu_category/(?P[\\w-]+)/$', MenuCategoryDetailAPIView.as_view(), name=\"menu_category_detail_api\"),\n\n # menu post put\n url(r'^create/menu/$', MenuCreateAPIView.as_view(), name=\"menu_create_api\"),\n url(r'^update/menu/(?P[\\w-]+)$', MenuUpdateAPIView.as_view(), name=\"menu_update_api\"),\n\n url(r'^create/operating_time/$', OperatingTimeCreateAPIView.as_view(), name=\"operatingtime_create_api\"),\n url(r'^update/operating_time/(?P\\d+)$', OperatingTimeUpdateAPIView.as_view(), name=\"operatingtime_update_api\"),\n \t \n]\n\nfrom django.conf import settings\nfrom django.conf.urls.static import static\n\nurlpatterns = urlpatterns + \\\n static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)", "sub_path": "restaurants/api/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1672, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "views.RestaurantListAPIView.as_view", "line_number": 12, "usage_type": "call"}, {"api_name": "views.RestaurantListAPIView", "line_number": 12, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "views.MenuListAPIView.as_view", "line_number": 13, "usage_type": "call"}, {"api_name": "views.MenuListAPIView", "line_number": 13, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 14, "usage_type": "call"}, {"api_name": "views.RestaurantCreateAPIView.as_view", "line_number": 14, "usage_type": "call"}, {"api_name": "views.RestaurantCreateAPIView", "line_number": 14, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 15, "usage_type": "call"}, {"api_name": "views.RestaurantUpdateAPIView.as_view", "line_number": 15, "usage_type": "call"}, {"api_name": "views.RestaurantUpdateAPIView", "line_number": 15, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 16, "usage_type": "call"}, {"api_name": "views.RestaurantDetailAPIView.as_view", "line_number": 16, "usage_type": "call"}, {"api_name": "views.RestaurantDetailAPIView", "line_number": 16, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 17, "usage_type": "call"}, {"api_name": "views.MenuDetailAPIView.as_view", "line_number": 17, "usage_type": "call"}, {"api_name": "views.MenuDetailAPIView", "line_number": 17, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 18, "usage_type": "call"}, {"api_name": "views.MenuCategoryDetailAPIView.as_view", "line_number": 18, "usage_type": "call"}, {"api_name": "views.MenuCategoryDetailAPIView", "line_number": 18, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 21, "usage_type": "call"}, {"api_name": "views.MenuCreateAPIView.as_view", "line_number": 21, "usage_type": "call"}, {"api_name": "views.MenuCreateAPIView", "line_number": 21, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 22, "usage_type": "call"}, {"api_name": "views.MenuUpdateAPIView.as_view", "line_number": 22, "usage_type": "call"}, {"api_name": "views.MenuUpdateAPIView", "line_number": 22, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 24, "usage_type": "call"}, {"api_name": "views.OperatingTimeCreateAPIView.as_view", "line_number": 24, "usage_type": "call"}, {"api_name": "views.OperatingTimeCreateAPIView", "line_number": 24, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 25, "usage_type": "call"}, {"api_name": "views.OperatingTimeUpdateAPIView.as_view", "line_number": 25, "usage_type": "call"}, {"api_name": "views.OperatingTimeUpdateAPIView", "line_number": 25, "usage_type": "name"}, {"api_name": "django.conf.urls.static.static", "line_number": 33, "usage_type": "call"}, {"api_name": "django.conf.settings.MEDIA_URL", "line_number": 33, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 33, "usage_type": "name"}, {"api_name": "django.conf.settings.MEDIA_ROOT", "line_number": 33, "usage_type": "attribute"}]}
+{"seq_id": "386365577", "text": "from django.shortcuts import render\nfrom django.http import HttpResponse\nfrom django.utils.decorators import method_decorator\nimport requests\n\ndef home(request):\n return render(request, 'home.html', {})\n\ndef submit_url(request):\n if(request.POST):\n login_data = request.POST.dict()\n post_url = login_data.get(\"url\")\n crawler_url = 'http://flask-image-crawler:5000/crawl'\n data = {'url':post_url}\n response=requests.post(crawler_url, data)\n response_status = response.content\n status_code=response.status_code\n return render(request, \"done.html\",{'post_url':post_url, 'response_status':response_status, 'status_code':status_code})\n else:\n return render(request, \"crawl.html\")\n", "sub_path": "django-frontend/src/my_crawler/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 747, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.shortcuts.render", "line_number": 7, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 15, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 18, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 20, "usage_type": "call"}]}
+{"seq_id": "533003565", "text": "from django.db import models\nfrom django.contrib import admin\nfrom django.contrib.auth.models import User\n\n# Create your models here.\nclass LeaveNote(models.Model):\n user = models.ForeignKey(User)#user ID\n dateSumbit = models.DateTimeField(auto_now=True)\n dateLeave = models.DateField()\n leaveReason = models.CharField(max_length=200)\n leaveDuration = models.IntegerField()\n isApproved = models.BooleanField()\n\nclass Level(models.Model):\n user = models.OneToOneField(User)\n", "sub_path": "web/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 494, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.db.models.Model", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 6, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 7, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 7, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 8, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.DateField", "line_number": 9, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 9, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 10, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 10, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 11, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 11, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 12, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 12, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 14, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.OneToOneField", "line_number": 15, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 15, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 15, "usage_type": "name"}]}
+{"seq_id": "530670222", "text": "#!/usr/bin/python\n\nimport http.server\nimport socketserver\nimport os\nimport datetime\nimport threading\nimport time\nimport numpy as np\nimport pandas as pd\nimport matplotlib\nmatplotlib.use('Agg')\nimport matplotlib.pyplot as plt\nfrom tqdm import tqdm\n\nPORT = int(os.environ.get(\"PORT\", 5000))\ndef server_files():\n os.chdir('data')\n Handler = http.server.SimpleHTTPRequestHandler\n Handler.extensions_map.update({\n '.webapp': 'application/x-web-app-manifest+json',\n })\n\n httpd = socketserver.TCPServer((\"\", PORT), Handler)\n print(\"Serving at port\", PORT)\n httpd.serve_forever()\n\ndef cleandata(df_raw):\n df_cleaned=df_raw.melt(id_vars=['Province/State','Country/Region','Lat','Long'],value_name='Cases',var_name='Date')\n df_cleaned=df_cleaned.set_index(['Country/Region','Province/State','Date'])\n return df_cleaned\n\ndef countrydata(df_cleaned,oldname,newname):\n df_country=df_cleaned.groupby(['Country/Region','Date'])['Cases'].sum().reset_index()\n df_country=df_country.set_index(['Country/Region','Date'])\n df_country.index=df_country.index.set_levels([df_country.index.levels[0], pd.to_datetime(df_country.index.levels[1])])\n df_country=df_country.sort_values(['Country/Region','Date'],ascending=True)\n df_country=df_country.rename(columns={oldname:newname})\n return df_country\n\ndef plotcountry(Country, CountryConsolidated):\n fig, axs = plt.subplots(3, 2)\n\n CountryConsolidated.loc[Country].reset_index().plot(ax=axs[0,0], style='.-', x='Date', y='Total Confirmed Cases')\n CountryConsolidated.loc[Country].reset_index().plot(ax=axs[0,1], style='.-', x='Date', y='Active Cases')\n CountryConsolidated.loc[Country].reset_index().plot(ax=axs[1,0], style='.-', x='Date', y='Total Deaths')\n CountryConsolidated.loc[Country].reset_index().plot(ax=axs[1,1], style='.-', x='Date', y='Total Recoveries')\n CountryConsolidated.loc[Country].reset_index().plot(ax=axs[2,0], style='.-', x='Date', y='Death to Cases Ratio')\n # CountryConsolidated.loc[Country].reset_index().plot(ax=axs[2,1], style='.-', x='Date', y='Total Confirmed Cases')\n # CountryConsolidated.plot()\n return fig\n\ndef dailydata(dfcountry,oldname,newname):\n dfcountrydaily=dfcountry.groupby(level=0).diff().fillna(0)\n dfcountrydaily=dfcountrydaily.rename(columns={oldname:newname})\n return dfcountrydaily\n\ndef update():\n ConfirmedCases=cleandata(pd.read_csv('https://raw.githubusercontent.com/CSSEGISandData/COVID-19/master/csse_covid_19_data/csse_covid_19_time_series/time_series_19-covid-Confirmed.csv'))\n Deaths=cleandata(pd.read_csv('https://raw.githubusercontent.com/CSSEGISandData/COVID-19/master/csse_covid_19_data/csse_covid_19_time_series/time_series_19-covid-Deaths.csv'))\n Recoveries=cleandata(pd.read_csv('https://raw.githubusercontent.com/CSSEGISandData/COVID-19/master/csse_covid_19_data/csse_covid_19_time_series/time_series_19-covid-Recovered.csv'))\n\n ConfirmedCasesCountry=countrydata(ConfirmedCases,'Cases','Total Confirmed Cases')\n DeathsCountry=countrydata(Deaths,'Cases','Total Deaths')\n RecoveriesCountry=countrydata(Recoveries,'Cases','Total Recoveries')\n\n NewCasesCountry=dailydata(ConfirmedCasesCountry,'Total Confirmed Cases','Daily New Cases')\n NewDeathsCountry=dailydata(DeathsCountry,'Total Deaths','Daily New Deaths')\n NewRecoveriesCountry=dailydata(RecoveriesCountry,'Total Recoveries','Daily New Recoveries')\n\n CountryConsolidated=pd.merge(ConfirmedCasesCountry,NewCasesCountry,how='left',left_index=True,right_index=True)\n CountryConsolidated=pd.merge(CountryConsolidated,NewDeathsCountry,how='left',left_index=True,right_index=True)\n CountryConsolidated=pd.merge(CountryConsolidated,DeathsCountry,how='left',left_index=True,right_index=True)\n CountryConsolidated=pd.merge(CountryConsolidated,RecoveriesCountry,how='left',left_index=True,right_index=True)\n CountryConsolidated=pd.merge(CountryConsolidated,NewRecoveriesCountry,how='left',left_index=True,right_index=True)\n CountryConsolidated['Active Cases']=CountryConsolidated['Total Confirmed Cases']-CountryConsolidated['Total Deaths']-CountryConsolidated['Total Recoveries']\n CountryConsolidated['Share of Recoveries - Closed Cases']=np.round(CountryConsolidated['Total Recoveries']/(CountryConsolidated['Total Recoveries']+CountryConsolidated['Total Deaths']),2)\n CountryConsolidated['Death to Cases Ratio']=np.round(CountryConsolidated['Total Deaths']/CountryConsolidated['Total Confirmed Cases'],3)\n\n Countries = []\n for ind in list(ConfirmedCasesCountry.index):\n Countries += [ind[0]]\n\n Countries = np.unique(Countries)\n # print(CountryConsolidated)\n cs = (CountryConsolidated['Total Confirmed Cases'] != 0).groupby(level=0).cumsum()\n countryConsolidated = CountryConsolidated.drop(cs[cs == 0].index)\n for country in tqdm(Countries):\n fig = plotcountry(country, countryConsolidated)\n fig.savefig(country + \".png\", dpi=300)\n plt.close()\n\ndef main():\n threading.Thread(target=server_files).start()\n update()\n\nmain()", "sub_path": "run.py", "file_name": "run.py", "file_ext": "py", "file_size_in_byte": 5052, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "matplotlib.use", "line_number": 12, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 16, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 18, "usage_type": "call"}, {"api_name": "http.server.server", "line_number": 19, "usage_type": "attribute"}, {"api_name": "http.server", "line_number": 19, "usage_type": "name"}, {"api_name": "socketserver.TCPServer", "line_number": 24, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 59, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 60, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 61, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 71, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 72, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 73, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 74, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 84, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.close", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "threading.Thread", "line_number": 94, "usage_type": "call"}]}
+{"seq_id": "335916784", "text": "from RLAgent import *\nfrom Common import *\nfrom Model import VelocityModel\nimport pygame\n\n\ndef verifyModel(agent):\n keymap = defaultdict(lambda: 'hover')\n keymap.update([('Key.up', 'moveForward'), ('Key.left', 'yawCCW'), ('Key.right', 'yawCW'), ('Key.down', 'hover')])\n\n while True:\n initialState, a = agent.getState(), agent.keyPressed.value\n yield keymap[a]\n\n r, nextState, isTerminal = (yield)\n draw_multirotor(nextState)\n\n yield\n\n\ndef draw_multirotor(state):\n roll, pitch, yaw = toEulerianAngle(state.orientation)\n x, y, z = state.position\n\n screen.fill((58, 58, 58))\n\n img_r = pygame.transform.rotate(img, -np.rad2deg(yaw))\n screen.blit(img_r, (x*10, y*10))\n\n pygame.display.update()\n\n\ndef main():\n model = VelocityModel(regressionModel=joblib.load('models/gradient-m.model'), frequency=10.0)\n agent = RLAgent('agent', decisionFrequency=20.0, defaultSpeed=4, defaultAltitude=6, yawRate=60,\n alternativeModel=model, maxDepth=math.inf, initialState=None)\n\n agent.setRl(verifyModel)\n agent.start()\n agent.join()\n\n\npygame.init()\nscreen = pygame.display.set_mode((300, 600))\npygame.display.set_caption('Model Verification')\nimg = pygame.image.load('quadcopter.jpg')\nmain()", "sub_path": "multirotor/src/ModelVerification.py", "file_name": "ModelVerification.py", "file_ext": "py", "file_size_in_byte": 1268, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pygame.transform.rotate", "line_number": 27, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 30, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 30, "usage_type": "attribute"}, {"api_name": "Model.VelocityModel", "line_number": 34, "usage_type": "call"}, {"api_name": "pygame.init", "line_number": 43, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 44, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 44, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 45, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 45, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 46, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 46, "usage_type": "attribute"}]}
+{"seq_id": "502293104", "text": "import gridfs\nimport json\nimport zlib\nimport io\nimport os\nimport traceback\nfrom shutil import which\nimport numpy as np\nfrom datetime import datetime\nfrom monty.tempfile import ScratchDir\nfrom maggma.builder import Builder\nimport prettyplotlib as ppl\nimport matplotlib\nimport scipy.interpolate as scint\nfrom prettyplotlib import brewer2mpl\nfrom pymatgen.core.structure import Structure\nfrom pymatgen.symmetry.bandstructure import HighSymmKpath\nfrom pymatgen.electronic_structure.core import Spin, Orbital\nfrom pymatgen.electronic_structure.bandstructure import BandStructureSymmLine, BandStructure\nfrom pymatgen.electronic_structure.dos import CompleteDos\nfrom pymatgen.electronic_structure.plotter import BSDOSPlotter, DosPlotter, BSPlotter\nfrom pymatgen.electronic_structure.boltztrap import BoltztrapRunner, BoltztrapAnalyzer\nfrom pymatgen.util.plotting import pretty_plot\n__author__ = \"Shyam Dwaraknath \"\n\nmatplotlib.use('agg')\n\nclass ElectronicStructureBuilder(Builder):\n\n def __init__(self, materials, electronic_structure, bandstructure_fs=\"bandstructure_fs\",\n dos_fs=\"dos_fs\", query=None, interpolate_dos=True, small_plot=True,\n static_images=True, **kwargs):\n \"\"\"\n Creates an electronic structure from a tasks collection, the associated band structures and density of states, and the materials structure\n\n :param tasks:\n :param materials:\n :param electronic_structure:\n \"\"\"\n\n self.materials = materials\n self.electronic_structure = electronic_structure\n self.query = query if query else {}\n self.bandstructure_fs = bandstructure_fs\n self.dos_fs = dos_fs\n self.interpolate_dos = interpolate_dos and bool(which(\"x_trans\"))\n self.small_plot = small_plot\n self.static_images = static_images\n\n super().__init__(sources=[materials],\n targets=[electronic_structure],\n **kwargs)\n\n def get_items(self):\n \"\"\"\n Gets all items to process into materials documents\n\n Returns:\n generator or list relevant tasks and materials to process into materials documents\n \"\"\"\n\n self.logger.info(\"Electronic Structure Builder Started\")\n\n # only consider materials that were updated since the electronic structure was last updated\n # and there is either a dos or bandstructure\n q = dict(self.query)\n q.update(self.materials.lu_filter(self.electronic_structure))\n q[\"$or\"] = [{\"bandstructure.bs_oid\": {\"$exists\": 1}},\n {\"bandstructure.dos_oid\": {\"$exists\": 1}}]\n\n # initialize the gridfs\n self.bfs = gridfs.GridFS(\n self.materials.collection.database, self.bandstructure_fs)\n self.dfs = gridfs.GridFS(\n self.materials.collection.database, self.dos_fs)\n\n mats = list(self.materials.distinct(self.materials.key, criteria=q))\n\n for m in mats:\n\n mat = self.materials.query([self.materials.key, \"structure\", \"bandstructure\", \"calc_settings\"],\n {self.materials.key: m}).limit(1)[0]\n self.get_bandstructure(mat)\n self.get_dos(mat)\n if self.interpolate_dos:\n self.get_uniform_bandstructure(mat)\n\n yield mat\n\n def process_item(self, mat):\n \"\"\"\n Process the tasks and materials into just a list of materials\n\n Args:\n mat (dict): material document\n\n Returns:\n (dict): electronic_structure document\n \"\"\"\n\n self.logger.info(\"Processing: {}\".format(mat[self.materials.key]))\n\n d = {self.electronic_structure.key: mat[\n self.materials.key], \"bandstructure\": {}}\n bs = None\n dos = None\n interpolated_dos = None\n\n # Process the bandstructure for information\n if \"bs\" in mat[\"bandstructure\"]:\n if \"structure\" not in mat[\"bandstructure\"][\"bs\"]:\n mat[\"bandstructure\"][\"bs\"][\"structure\"] = mat[\"structure\"]\n if len(mat[\"bandstructure\"][\"bs\"].get(\"labels_dict\", {})) == 0:\n struc = Structure.from_dict(mat[\"structure\"])\n kpath = HighSymmKpath(struc)._kpath[\"kpoints\"]\n mat[\"bandstructure\"][\"bs\"][\"labels_dict\"] = kpath\n # Somethign is wrong with the as_dict / from_dict encoding in the two band structure objects so have to use this hodge podge serialization\n # TODO: Fix bandstructure objects in pymatgen\n bs = BandStructureSymmLine.from_dict(\n BandStructure.from_dict(mat[\"bandstructure\"][\"bs\"]).as_dict())\n d[\"bandstructure\"][\"band_gap\"] = {\"band_gap\": bs.get_band_gap()[\"energy\"],\n \"direct_gap\": bs.get_direct_band_gap(),\n \"is_direct\": bs.get_band_gap()[\"direct\"],\n \"transition\": bs.get_band_gap()[\"transition\"]}\n\n if self.small_plot:\n d[\"bandstructure\"][\"plot_small\"] = get_small_plot(bs)\n\n if \"dos\" in mat[\"bandstructure\"]:\n dos = CompleteDos.from_dict(mat[\"bandstructure\"][\"dos\"])\n\n if self.interpolate_dos and \"uniform_bs\" in mat[\"bandstructure\"]:\n try:\n interpolated_dos = self.get_interpolated_dos(mat)\n except Exception:\n self.logger.warning(\"Boltztrap interpolation failed for {}. Continuing with regular DOS\".format(mat[self.materials.key]))\n\n # Generate static images\n if self.static_images:\n try:\n ylim = None\n if bs:\n plotter = WebBSPlotter(bs)\n fig = plotter.get_plot()\n ylim = fig.ylim() # Used by DOS plot\n fig.close()\n\n d[\"bandstructure\"][\"bs_plot\"] = image_from_plotter(plotter)\n\n if dos:\n plotter = WebDosVertPlotter()\n plotter.add_dos_dict(dos.get_element_dos())\n\n if interpolated_dos:\n plotter.add_dos(\"Total DOS\", interpolated_dos)\n d[\"bandstructure\"][\"dos_plot\"] = image_from_plotter(plotter, ylim=ylim)\n\n d[\"bandstructure\"][\"dos_plot\"] = image_from_plotter(plotter, ylim=ylim)\n\n except Exception:\n self.logger.warning(\n \"Caught error in electronic structure plotting for {}: {}\".format(mat[self.materials.key], traceback.format_exc()))\n return None\n\n return d\n\n def update_targets(self, items):\n \"\"\"\n Inserts the new task_types into the task_types collection\n\n Args:\n items ([([dict],[int])]): A list of tuples of materials to update and the corresponding processed task_ids\n \"\"\"\n items = list(filter(None, items))\n\n if len(items) > 0:\n self.logger.info(\"Updating {} band structures\".format(len(items)))\n self.electronic_structure.update(items)\n else:\n self.logger.info(\"No items to update\")\n\n def get_bandstructure(self, mat):\n\n # If a bandstructure oid exists\n if \"bs_oid\" in mat.get(\"bandstructure\", {}):\n bs_json = self.bfs.get(mat[\"bandstructure\"][\"bs_oid\"]).read()\n\n if \"zlib\" in mat[\"bandstructure\"].get(\"bs_compression\", \"\"):\n bs_json = zlib.decompress(bs_json)\n\n bs_dict = json.loads(bs_json.decode())\n mat[\"bandstructure\"][\"bs\"] = bs_dict\n\n def get_uniform_bandstructure(self, mat):\n\n # If a bandstructure oid exists\n if \"uniform_bs_oid\" in mat.get(\"bandstructure\", {}):\n bs_json = self.bfs.get(mat[\"bandstructure\"][\n \"uniform_bs_oid\"]).read()\n\n if \"zlib\" in mat[\"bandstructure\"].get(\"uniform_bs_compression\", \"\"):\n bs_json = zlib.decompress(bs_json)\n\n bs_dict = json.loads(bs_json.decode())\n mat[\"bandstructure\"][\"uniform_bs\"] = bs_dict\n\n def get_dos(self, mat):\n\n # if a dos oid exists\n if \"dos_oid\" in mat.get(\"bandstructure\", {}):\n dos_json = self.dfs.get(mat[\"bandstructure\"][\"dos_oid\"]).read()\n\n if \"zlib\" in mat[\"bandstructure\"].get(\"dos_compression\", \"\"):\n dos_json = zlib.decompress(dos_json)\n\n dos_dict = json.loads(dos_json.decode())\n mat[\"bandstructure\"][\"dos\"] = dos_dict\n\n def get_interpolated_dos(self, mat):\n\n nelect = mat[\"calc_settings\"][\"nelect\"]\n\n bs_dict = mat[\"bandstructure\"][\"uniform_bs\"]\n bs_dict[\"structure\"] = mat['structure']\n bs = BandStructure.from_dict(bs_dict)\n\n if bs.is_spin_polarized:\n with ScratchDir(\".\"):\n BoltztrapRunner(bs=bs,\n nelec=nelect,\n run_type=\"DOS\",\n dos_type=\"TETRA\",\n spin=1,\n timeout=60).run(path_dir=os.getcwd())\n an_up = BoltztrapAnalyzer.from_files(\"boltztrap/\", dos_spin=1)\n\n with ScratchDir(\".\"):\n BoltztrapRunner(bs=bs,\n nelec=nelect,\n run_type=\"DOS\",\n dos_type=\"TETRA\",\n spin=-1,\n timeout=60).run(path_dir=os.getcwd())\n an_dw = BoltztrapAnalyzer.from_files(\"boltztrap/\", dos_spin=-1)\n\n cdos = an_up.get_complete_dos(bs.structure, an_dw)\n\n else:\n with ScratchDir(\".\"):\n BoltztrapRunner(bs=bs,\n nelec=nelect,\n run_type=\"DOS\",\n dos_type=\"TETRA\",\n timeout=60).run(path_dir=os.getcwd())\n an = BoltztrapAnalyzer.from_files(\"boltztrap/\")\n cdos = an.get_complete_dos(bs.structure)\n\n return cdos\n\n\ndef image_from_plotter(plotter, ylim=None):\n plot = plotter.get_plot()\n imgdata = io.BytesIO()\n plot.savefig(imgdata, format=\"png\", dpi=100)\n plot_img = imgdata.getvalue()\n plot.close()\n return plot_img\n\n\ndef get_small_plot(bs):\n\n plot_small = BSPlotter(bs).bs_plot_data()\n\n gap = bs.get_band_gap()[\"energy\"]\n for branch in plot_small['energy']:\n for spin, v in branch.items():\n new_bands = []\n for band in v:\n if min(band) < gap + 3 and max(band) > -3:\n new_bands.append(band)\n branch[spin] = new_bands\n return plot_small\n\n\n#\n# Obtain web-friendly images by subclassing pymatgen plotters.\n#\n\nclass WebBSPlotter(BSPlotter):\n\n def get_plot(self, zero_to_efermi=True, ylim=None, smooth=False):\n \"\"\"\n get a matplotlib object for the bandstructure plot.\n Blue lines are up spin, red lines are down\n spin.\n Args:\n zero_to_efermi: Automatically subtract off the Fermi energy from\n the eigenvalues and plot (E-Ef).\n ylim: Specify the y-axis (energy) limits; by default None let\n the code choose. It is vbm-4 and cbm+4 if insulator\n efermi-10 and efermi+10 if metal\n smooth: interpolates the bands by a spline cubic\n \"\"\"\n\n plt = pretty_plot(6, 5.5) # Was 12, 8\n\n matplotlib.rc('text', usetex=True)\n\n width = 4\n ticksize = int(width * 2.5)\n axes = plt.gca()\n axes.set_title(axes.get_title(), size=width * 4)\n labelsize = int(width * 3)\n axes.set_xlabel(axes.get_xlabel(), size=labelsize)\n axes.set_ylabel(axes.get_ylabel(), size=labelsize)\n\n plt.xticks(fontsize=ticksize)\n plt.yticks(fontsize=ticksize)\n\n for axis in ['top', 'bottom', 'left', 'right']:\n axes.spines[axis].set_linewidth(0.5)\n\n # main internal config options\n e_min = -4\n e_max = 4\n if self._bs.is_metal():\n e_min = -10\n e_max = 10\n band_linewidth = 1 # Was 3 in pymatgen\n\n data = self.bs_plot_data(zero_to_efermi)\n if not smooth:\n for d in range(len(data['distances'])):\n for i in range(self._nb_bands):\n plt.plot(data['distances'][d],\n [data['energy'][d][str(Spin.up)][i][j]\n for j in range(len(data['distances'][d]))], 'b-',\n linewidth=band_linewidth)\n if self._bs.is_spin_polarized:\n plt.plot(data['distances'][d],\n [data['energy'][d][str(Spin.down)][i][j]\n for j in range(len(data['distances'][d]))],\n 'r--', linewidth=band_linewidth)\n else:\n for d in range(len(data['distances'])):\n for i in range(self._nb_bands):\n tck = scint.splrep(\n data['distances'][d],\n [data['energy'][d][str(Spin.up)][i][j]\n for j in range(len(data['distances'][d]))])\n step = (data['distances'][d][-1]\n - data['distances'][d][0]) / 1000\n\n plt.plot([x * step + data['distances'][d][0]\n for x in range(1000)],\n [scint.splev(x * step + data['distances'][d][0],\n tck, der=0)\n for x in range(1000)], 'b-',\n linewidth=band_linewidth)\n\n if self._bs.is_spin_polarized:\n\n tck = scint.splrep(\n data['distances'][d],\n [data['energy'][d][str(Spin.down)][i][j]\n for j in range(len(data['distances'][d]))])\n step = (data['distances'][d][-1]\n - data['distances'][d][0]) / 1000\n\n plt.plot([x * step + data['distances'][d][0]\n for x in range(1000)],\n [scint.splev(x * step + data['distances'][d][0],\n tck, der=0)\n for x in range(1000)], 'r--',\n linewidth=band_linewidth)\n self._maketicks(plt)\n\n # Main X and Y Labels\n plt.xlabel(r'$\\mathrm{Wave\\ Vector}$')\n ylabel = r'$\\mathrm{E\\ -\\ E_f\\ (eV)}$' if zero_to_efermi \\\n else r'$\\mathrm{Energy\\ (eV)}$'\n plt.ylabel(ylabel)\n\n # Draw Fermi energy, only if not the zero\n if not zero_to_efermi:\n ef = self._bs.efermi\n plt.axhline(ef, linewidth=2, color='k')\n\n # X range (K)\n # last distance point\n x_max = data['distances'][-1][-1]\n plt.xlim(0, x_max)\n\n if ylim is None:\n if self._bs.is_metal():\n # Plot A Metal\n if zero_to_efermi:\n plt.ylim(e_min, e_max)\n else:\n plt.ylim(self._bs.efermi + e_min, self._bs._efermi + e_max)\n else:\n for cbm in data['cbm']:\n plt.scatter(cbm[0], cbm[1], color='r', marker='o', s=100)\n\n for vbm in data['vbm']:\n plt.scatter(vbm[0], vbm[1], color='g', marker='o', s=100)\n plt.ylim(data['vbm'][0][1] + e_min, data['cbm'][0][1] + e_max)\n else:\n plt.ylim(ylim)\n\n plt.tight_layout()\n\n return plt\n\n\nclass WebDosVertPlotter(DosPlotter):\n\n def get_plot(self, xlim=None, ylim=None,\n plt=None, handle_only=False):\n \"\"\"\n Get a matplotlib plot showing the DOS.\n Args:\n xlim: Specifies the x-axis limits. Set to None for automatic\n determination.\n ylim: Specifies the y-axis limits.\n plt: Handle on existing plot.\n handle_only: Quickly return just a handle. Useful if this method\n raises an exception so that one can close() the figure.\n \"\"\"\n\n plt = plt or pretty_plot(2, 5.5)\n if handle_only:\n return plt\n\n ncolors = max(3, len(self._doses))\n ncolors = min(9, ncolors)\n colors = brewer2mpl.get_map('Set1', 'qualitative', ncolors).mpl_colors\n\n y = None\n alldensities = []\n allenergies = []\n\n width = 4\n ticksize = int(width * 2.5)\n axes = plt.gca()\n axes.set_title(axes.get_title(), size=width * 4)\n labelsize = int(width * 3)\n axes.set_xlabel(axes.get_xlabel(), size=labelsize)\n axes.set_ylabel(axes.get_ylabel(), size=labelsize)\n axes.xaxis.labelpad = 6\n\n # Note that this complicated processing of energies is to allow for\n # stacked plots in matplotlib.\n for key, dos in self._doses.items():\n energies = dos['energies']\n densities = dos['densities']\n if not y:\n y = {Spin.up: np.zeros(energies.shape),\n Spin.down: np.zeros(energies.shape)}\n newdens = {}\n for spin in [Spin.up, Spin.down]:\n if spin in densities:\n if self.stack:\n y[spin] += densities[spin]\n newdens[spin] = y[spin].copy()\n else:\n newdens[spin] = densities[spin]\n allenergies.append(energies)\n alldensities.append(newdens)\n\n keys = list(self._doses.keys())\n keys.reverse()\n alldensities.reverse()\n allenergies.reverse()\n allpts = []\n for i, key in enumerate(keys):\n x = []\n y = []\n for spin in [Spin.up, Spin.down]:\n if spin in alldensities[i]:\n densities = list(int(spin) * alldensities[i][spin])\n energies = list(allenergies[i])\n if spin == Spin.down:\n energies.reverse()\n densities.reverse()\n y.extend(energies)\n x.extend(densities)\n allpts.extend(list(zip(x, y)))\n if self.stack:\n plt.fill(x, y, color=colors[i % ncolors],\n label=str(key))\n else:\n ppl.plot(x, y, color=colors[i % ncolors],\n label=str(key), linewidth=1)\n if not self.zero_at_efermi:\n xlim = plt.xlim()\n ppl.plot(xlim, [self._doses[key]['efermi'],\n self._doses[key]['efermi']],\n color=colors[i % ncolors],\n linestyle='--', linewidth=1)\n\n if ylim:\n plt.ylim(ylim)\n if xlim:\n plt.xlim(xlim)\n else:\n ylim = plt.ylim()\n relevantx = [p[0] for p in allpts\n if ylim[0] < p[1] < ylim[1]]\n plt.xlim(min(relevantx), max(relevantx))\n if self.zero_at_efermi:\n xlim = plt.xlim()\n plt.plot(xlim, [0, 0], 'k--', linewidth=1)\n\n plt.ylabel(r'$\\mathrm{E\\ -\\ E_f\\ (eV)}$')\n plt.xlabel(r'$\\mathrm{Density\\ of\\ states}$')\n\n locs, _ = plt.xticks()\n plt.xticks([0], fontsize=ticksize)\n plt.yticks(fontsize=ticksize)\n plt.grid(which='major', axis='y')\n\n plt.legend(fontsize='x-small',\n loc='upper right', bbox_to_anchor=(1.15, 1))\n leg = plt.gca().get_legend()\n leg.get_frame().set_alpha(0.25)\n plt.tight_layout()\n return plt\n", "sub_path": "emmet/vasp/builders/electronic_structure.py", "file_name": "electronic_structure.py", "file_ext": "py", "file_size_in_byte": 19983, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "matplotlib.use", "line_number": 26, "usage_type": "call"}, {"api_name": "maggma.builder.Builder", "line_number": 28, "usage_type": "name"}, {"api_name": "shutil.which", "line_number": 46, "usage_type": "call"}, {"api_name": "gridfs.GridFS", "line_number": 72, "usage_type": "call"}, {"api_name": "gridfs.GridFS", "line_number": 74, "usage_type": "call"}, {"api_name": "pymatgen.core.structure.Structure.from_dict", "line_number": 114, "usage_type": "call"}, {"api_name": "pymatgen.core.structure.Structure", "line_number": 114, "usage_type": "name"}, {"api_name": "pymatgen.symmetry.bandstructure.HighSymmKpath", "line_number": 115, "usage_type": "call"}, {"api_name": "pymatgen.electronic_structure.bandstructure.BandStructureSymmLine.from_dict", "line_number": 119, "usage_type": "call"}, {"api_name": "pymatgen.electronic_structure.bandstructure.BandStructureSymmLine", "line_number": 119, "usage_type": "name"}, {"api_name": "pymatgen.electronic_structure.bandstructure.BandStructure.from_dict", "line_number": 120, "usage_type": "call"}, {"api_name": "pymatgen.electronic_structure.bandstructure.BandStructure", "line_number": 120, "usage_type": "name"}, {"api_name": "pymatgen.electronic_structure.dos.CompleteDos.from_dict", "line_number": 130, "usage_type": "call"}, {"api_name": "pymatgen.electronic_structure.dos.CompleteDos", "line_number": 130, "usage_type": "name"}, {"api_name": "traceback.format_exc", "line_number": 162, "usage_type": "call"}, {"api_name": "zlib.decompress", "line_number": 189, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 191, "usage_type": "call"}, {"api_name": "zlib.decompress", "line_number": 202, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 204, "usage_type": "call"}, {"api_name": "zlib.decompress", "line_number": 214, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 216, "usage_type": "call"}, {"api_name": "pymatgen.electronic_structure.bandstructure.BandStructure.from_dict", "line_number": 225, "usage_type": "call"}, {"api_name": "pymatgen.electronic_structure.bandstructure.BandStructure", "line_number": 225, "usage_type": "name"}, {"api_name": "monty.tempfile.ScratchDir", "line_number": 228, "usage_type": "call"}, {"api_name": "pymatgen.electronic_structure.boltztrap.BoltztrapRunner", "line_number": 229, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 234, "usage_type": "call"}, {"api_name": "pymatgen.electronic_structure.boltztrap.BoltztrapAnalyzer.from_files", "line_number": 235, "usage_type": "call"}, {"api_name": "pymatgen.electronic_structure.boltztrap.BoltztrapAnalyzer", "line_number": 235, "usage_type": "name"}, {"api_name": "monty.tempfile.ScratchDir", "line_number": 237, "usage_type": "call"}, {"api_name": "pymatgen.electronic_structure.boltztrap.BoltztrapRunner", "line_number": 238, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 243, "usage_type": "call"}, {"api_name": "pymatgen.electronic_structure.boltztrap.BoltztrapAnalyzer.from_files", "line_number": 244, "usage_type": "call"}, {"api_name": "pymatgen.electronic_structure.boltztrap.BoltztrapAnalyzer", "line_number": 244, "usage_type": "name"}, {"api_name": "monty.tempfile.ScratchDir", "line_number": 249, "usage_type": "call"}, {"api_name": "pymatgen.electronic_structure.boltztrap.BoltztrapRunner", "line_number": 250, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 254, "usage_type": "call"}, {"api_name": "pymatgen.electronic_structure.boltztrap.BoltztrapAnalyzer.from_files", "line_number": 255, "usage_type": "call"}, {"api_name": "pymatgen.electronic_structure.boltztrap.BoltztrapAnalyzer", "line_number": 255, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 263, "usage_type": "call"}, {"api_name": "pymatgen.electronic_structure.plotter.BSPlotter", "line_number": 272, "usage_type": "call"}, {"api_name": "pymatgen.electronic_structure.plotter.BSPlotter", "line_number": 289, "usage_type": "name"}, {"api_name": "pymatgen.util.plotting.pretty_plot", "line_number": 305, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 307, "usage_type": "call"}, {"api_name": "pymatgen.electronic_structure.core.Spin.up", "line_number": 336, "usage_type": "attribute"}, {"api_name": "pymatgen.electronic_structure.core.Spin", "line_number": 336, "usage_type": "name"}, {"api_name": "pymatgen.electronic_structure.core.Spin.down", "line_number": 341, "usage_type": "attribute"}, {"api_name": "pymatgen.electronic_structure.core.Spin", "line_number": 341, "usage_type": "name"}, {"api_name": "scipy.interpolate.splrep", "line_number": 347, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 347, "usage_type": "name"}, {"api_name": "pymatgen.electronic_structure.core.Spin.up", "line_number": 349, "usage_type": "attribute"}, {"api_name": "pymatgen.electronic_structure.core.Spin", "line_number": 349, "usage_type": "name"}, {"api_name": "scipy.interpolate.splev", "line_number": 356, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 356, "usage_type": "name"}, {"api_name": "scipy.interpolate.splrep", "line_number": 363, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 363, "usage_type": "name"}, {"api_name": "pymatgen.electronic_structure.core.Spin.down", "line_number": 365, "usage_type": "attribute"}, {"api_name": "pymatgen.electronic_structure.core.Spin", "line_number": 365, "usage_type": "name"}, {"api_name": "scipy.interpolate.splev", "line_number": 372, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 372, "usage_type": "name"}, {"api_name": "pymatgen.electronic_structure.plotter.DosPlotter", "line_number": 416, "usage_type": "name"}, {"api_name": "pymatgen.util.plotting.pretty_plot", "line_number": 431, "usage_type": "call"}, {"api_name": "prettyplotlib.brewer2mpl.get_map", "line_number": 437, "usage_type": "call"}, {"api_name": "prettyplotlib.brewer2mpl", "line_number": 437, "usage_type": "name"}, {"api_name": "pymatgen.electronic_structure.core.Spin.up", "line_number": 458, "usage_type": "attribute"}, {"api_name": "pymatgen.electronic_structure.core.Spin", "line_number": 458, "usage_type": "name"}, {"api_name": "pymatgen.electronic_structure.core.Spin.down", "line_number": 459, "usage_type": "attribute"}, {"api_name": "pymatgen.electronic_structure.core.Spin", "line_number": 459, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 458, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 459, "usage_type": "call"}, {"api_name": "pymatgen.electronic_structure.core.Spin.up", "line_number": 461, "usage_type": "attribute"}, {"api_name": "pymatgen.electronic_structure.core.Spin", "line_number": 461, "usage_type": "name"}, {"api_name": "pymatgen.electronic_structure.core.Spin.down", "line_number": 461, "usage_type": "attribute"}, {"api_name": "pymatgen.electronic_structure.core.Spin.up", "line_number": 479, "usage_type": "attribute"}, {"api_name": "pymatgen.electronic_structure.core.Spin", "line_number": 479, "usage_type": "name"}, {"api_name": "pymatgen.electronic_structure.core.Spin.down", "line_number": 479, "usage_type": "attribute"}, {"api_name": "pymatgen.electronic_structure.core.Spin.down", "line_number": 483, "usage_type": "attribute"}, {"api_name": "pymatgen.electronic_structure.core.Spin", "line_number": 483, "usage_type": "name"}, {"api_name": "prettyplotlib.plot", "line_number": 493, "usage_type": "call"}, {"api_name": "prettyplotlib.plot", "line_number": 497, "usage_type": "call"}]}
+{"seq_id": "333570831", "text": "\"\"\"\ngit tasks\n\"\"\"\nimport logging\nfrom invoke import task\nimport click\nfrom tasks.utils import get_compose_env, is_venv\n\n# from tasks.core import clean, execute_sql\n\nlogger = logging.getLogger(__name__)\nlogger.setLevel(\"DEBUG\")\n\n\n# git rev-parse HEAD\n\n\n# @task(incrementable=[\"verbose\"])\n# def verbosity(ctx, loc=\"local\", quiet=False, verbose=0):\n# \"\"\"\n# Return `git rev-parse HEAD` for project.\n# Usage: inv docker.lint-test or inv local.lint-test\n# \"\"\"\n# env = get_compose_env(ctx, loc=loc)\n\n# # Only display result\n# ctx.config[\"run\"][\"echo\"] = False\n\n# # Override run commands env variables one key at a time\n# for k, v in env.items():\n# ctx.config[\"run\"][\"env\"][k] = v\n\n\n@task(incrementable=[\"verbose\"])\ndef pr_sha(ctx, loc=\"local\", quiet=False, verbose=0):\n \"\"\"\n Return `git rev-parse HEAD` for project.\n Usage: inv docker.lint-test or inv local.lint-test\n \"\"\"\n env = get_compose_env(ctx, loc=loc)\n\n # Only display result\n ctx.config[\"run\"][\"echo\"] = False\n\n # Override run commands env variables one key at a time\n for k, v in env.items():\n ctx.config[\"run\"][\"env\"][k] = v\n\n res = ctx.run(\"git rev-parse HEAD\")\n\n # override CI_IMAGE value\n ctx.config[\"run\"][\"env\"][\"PR_SHA\"] = \"{}\".format(res.stdout)\n ctx.config[\"run\"][\"env\"][\"REPO_NAME\"] = \"bossjones/fake-medium-fastapi-ci\"\n ctx.config[\"run\"][\"env\"][\"IMAGE_TAG\"] = \"{}:{}\".format(\n ctx.config[\"run\"][\"env\"][\"REPO_NAME\"], ctx.config[\"run\"][\"env\"][\"PR_SHA\"]\n )\n ctx.config[\"run\"][\"env\"][\"TAG\"] = ctx.config[\"run\"][\"env\"][\"IMAGE_TAG\"]\n\n if verbose >= 1:\n msg = \"[PR_SHA] {}\".format(ctx.config[\"run\"][\"env\"][\"PR_SHA\"])\n click.secho(msg, fg=\"green\")\n", "sub_path": "tasks/git.py", "file_name": "git.py", "file_ext": "py", "file_size_in_byte": 1727, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "tasks.utils.get_compose_env", "line_number": 40, "usage_type": "call"}, {"api_name": "click.secho", "line_number": 61, "usage_type": "call"}, {"api_name": "invoke.task", "line_number": 34, "usage_type": "call"}]}
+{"seq_id": "133422965", "text": "import os\nimport requests\nimport time\nimport subprocess\nimport json\nimport re\nimport sys\nfrom moviepy.editor import VideoFileClip\nimport shutil\nfrom requests import exceptions\n\nmain_api = \"api.mgghot.com\"\npath = os.getcwd()\n\n\ndef get_data_file(path_file):\n fo = open(path_file, \"r\")\n lines = fo.readlines()\n fo.close()\n stt_video = ''\n\n if len(lines) == 0:\n return ''\n\n return lines[0]\n\n\nkey = get_data_file(\"config/key.txt\")\n\nkey_apis = get_data_file(\"config/key_api.txt\").split(\",\")\n\n\ndef get_list_video_by_api(channel_id, data_channel):\n results = []\n max_result = 30\n page_token = ''\n stt = 0\n key_api = key_apis[0]\n len_key_api = len(key_apis)\n\n while len(results) < 31:\n url = \"https://www.googleapis.com/youtube/v3/search?part=id&key=\" \\\n + str(key_api) + \"&channelId=\" + str(channel_id) + \"&maxResults=\" + str(max_result) \\\n + \"&order=date&pageToken=\" + str(page_token)\n\n req = requests.get(url)\n\n list_item = json.loads(req.content)\n\n if 'items' not in list_item:\n stt = stt + 1\n\n if stt >= len_key_api:\n return []\n\n key_api = key_apis[stt]\n continue\n\n items = list_item['items']\n\n try:\n page_token = list_item['nextPageToken']\n except KeyError:\n page_token = ''\n\n for item in items:\n try:\n id_video = item['id']['videoId']\n except KeyError:\n id_video = ''\n\n if id_video != '' and id_video not in data_channel:\n results.append(id_video)\n\n if page_token == '':\n break\n\n return results\n\n\ndef get_thumbnail(url, path_thumb):\n print(url)\n try:\n stdout = subprocess.check_output(['youtube-dl', '--list-thumbnails', url])\n\n arr = str(stdout).split('\\\\n')\n url = ''\n\n for i in arr:\n temp = re.findall(r'http(.*?).jpg', str(i))\n\n if len(temp) > 0:\n url = 'http' + temp[0] + '.jpg'\n\n if url == '':\n return ''\n\n r = requests.get(url)\n\n if r.status_code == 200:\n with open(path_thumb + '/thumbnail.jpg', 'wb') as file:\n for chunk in r.iter_content(1024):\n file.write(chunk)\n except:\n return ''\n\n return path_thumb + '/thumbnail.jpg'\n\n\ndef get_number_video(url):\n result = []\n\n try:\n stdout = subprocess.check_output(['youtube-dl', '-F', url])\n arr = str(stdout).split('\\\\n')\n\n audio = ''\n\n for item in arr:\n if 'm4a' in item:\n audio = item.split(' ')[0]\n\n for item in arr:\n if '1080' in item and 'mp4' in item:\n result.append(str(item.split(' ')[0]) + '+' + str(audio))\n\n if '720' in item and 'mp4' in item:\n result.append(str(item.split(' ')[0]) + '+' + str(audio))\n\n for item in arr:\n if '480' in item and 'mp4' in item:\n result.append(str(item.split(' ')[0]) + '+' + str(audio))\n\n for item in arr:\n if '360' in item and 'mp4' in item:\n result.append(str(item.split(' ')[0]) + '+' + str(audio))\n\n for item in arr:\n if '240' in item and 'mp4' in item:\n result.append(str(item.split(' ')[0]) + '+' + str(audio))\n except:\n return False\n\n return result\n\n\ndef download_video_from_youtube(id_video, path_page):\n numbers = get_number_video(\"https://www.youtube.com/watch?v=\" + str(id_video))\n\n platform = get_platform()\n\n if numbers is False:\n return False\n\n print(\"Downloading...\")\n for number in numbers:\n url = \"youtube-dl -f \" + str(number) + \" -o \" + path_page + '/' \\\n + \"input/input.%\\(ext\\)s https://www.youtube.com/watch?v=\" + str(id_video)\n\n if platform == 'Windows':\n url = \"youtube-dl -f \" + str(number) + \" -o \" + path_page + '/' \\\n + \"input/input.%(ext)s https://www.youtube.com/watch?v=\" + str(id_video)\n\n os.system(url)\n\n check = get_file_upload(path_page)\n\n if check:\n return True\n\n empty_folder(path_page + '/input')\n\n return True\n\n\ndef get_tags(id_video):\n for key_api in key_apis:\n url = \"https://www.googleapis.com/youtube/v3/videos?part=snippet,contentDetails&key=\" + key_api + \"&id=\" + str(id_video)\n\n req = requests.get(url)\n items = json.loads(req.content)\n tags = ''\n title = ''\n result = []\n\n try:\n if 'items' not in items:\n continue\n\n title = items['items'][0]['snippet']['title']\n except KeyError as e:\n print('I got a KeyError - reason \"%s\"' % str(e))\n\n try:\n tags = items['items'][0]['snippet']['tags']\n except KeyError as e:\n print('I got a KeyError - reason \"%s\"' % str(e))\n\n list_tag = ','.join(tags)\n result.append(title)\n result.append(list_tag)\n\n return result\n\n\ndef get_file_upload(path_page):\n filelist = os.listdir(path_page + '/input')\n\n for fichier in filelist:\n if \"input.mp4\" in fichier:\n return fichier\n\n return False\n\n\ndef getLength():\n filename = \"input/input.mp4\"\n clip = VideoFileClip(filename)\n\n return clip.duration\n\n\ndef get_ffmpeg(file_video, file_ffmpeg):\n path_file = 'ffmpeg-files/' + file_ffmpeg\n fo = open(path_file, \"r\")\n lines = fo.readlines()\n\n if len(lines) > 0:\n string_process = lines[0]\n string_process = string_process.replace(\"input.mp4\", 'input/input.ts')\n string_process = string_process.replace(\"output.mp4\", \"output/\" + str(file_video))\n\n return string_process\n\n return False\n\n\ndef process_video(file_name, length_cut):\n total_lentgh = getLength()\n\n string1 = \"ffmpeg -ss \" + str(length_cut) + \" -i input/input.mp4 -t \" \\\n + str(total_lentgh) + \" -c copy output/output.mp4\"\n os.system(string1)\n\n # string = \"ffmpeg -i /input/temp_input.mp4 -c copy -bsf:v h264_mp4toannexb -f mpegts /input/input.ts\"\n # os.system(string)\n\n # string_ffmpeg = get_ffmpeg(file_name, 'text.txt')\n # os.system(string_ffmpeg)\n\n return 'output/output.mp4'\n\n\ndef uploadVideoToFacebook(file_name, access_token, cookie, title, des, thumb, account_id):\n file_size = os.path.getsize(file_name)\n\n headers = {\n 'user-agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/78.0.3904.97 Safari/537.36',\n 'referer': 'https://www.facebook.com/'\n }\n\n try:\n url1 = \"https://graph-video.facebook.com/v2.3/me/videos?\"\n data1 = {\n 'access_token': access_token,\n 'upload_phase': \"start\",\n \"file_size\": file_size\n }\n\n req1 = requests.post(url1, data=data1, headers=headers, cookies=cookie)\n\n content1 = json.loads(req1.content)\n\n if 'upload_session_id' not in content1:\n update_check_point(account_id)\n return False\n\n upload_session_id = content1['upload_session_id']\n\n data2 = {\n 'access_token': access_token,\n 'upload_phase': 'transfer',\n 'start_offset': 0,\n 'upload_session_id': upload_session_id\n }\n\n up = {'video_file_chunk': (file_name, open(file_name, 'rb'), \"multipart/form-data\")}\n req2 = requests.post(url1, files=up, data=data2, headers=headers, cookies=cookie)\n\n data3 = {\n 'access_token': access_token,\n 'upload_phase': 'finish',\n 'upload_session_id': upload_session_id,\n 'title': title,\n 'description': des\n }\n\n thumb = {'thumb': (thumb, open(thumb, 'rb'), \"multipart/form-data\")}\n\n req3 = requests.post(url1, files=thumb, data=data3, headers=headers, cookies=cookie)\n\n print(req3.content)\n result = json.loads(req3.content)['success'] is True\n except KeyError as e:\n print(e)\n return False\n\n return result\n\n\ndef getLengthVideo(input_video):\n platform = get_platform()\n\n if platform == 'Windows':\n string = 'ffprobe -i ' + input_video + ' -show_entries format=duration -v quiet -of csv=\"p=0\"'\n result = subprocess.Popen(string, stdout=subprocess.PIPE,stderr=subprocess.STDOUT)\n output = result.communicate()[0].strip()\n\n index = str(output).find('.')\n output = output[:index - 2]\n\n return int(output)\n\n string = 'ffprobe -i ' + input_video + ' -show_entries format=duration -v quiet -of csv=\"p=0\"'\n\n result = subprocess.getoutput(string)\n output = result.split()[0].strip()\n output = float(output)\n output = int(output)\n # index = str(output).find('.')\n # output = output[:index - 2]\n\n return int(output)\n\n\ndef hanlde(access_token, cookie, name_title, description, genres, thumbnail, path_page, path_thumb, account_id):\n check = False\n file = get_file_upload(path_page)\n\n if file is False:\n empty_folder(path_page + \"/input\")\n\n return False\n\n file_name = path_thumb + '/input/' + file\n\n title = name_title\n des = description\n id_page = \"me\"\n link_video = file_name\n length_video = getLengthVideo(file_name)\n print(length_video)\n print(\"Uploading...\")\n\n if length_video < 1200:\n check = uploadVideoToFacebook(link_video, access_token, cookie, title, des, thumbnail, account_id)\n else:\n if length_video > 3000:\n pass\n else:\n print(\"Upload by nodejs\")\n\n if get_platform() == 'Windows':\n string_upload = \"node upload-video-to-facebook/main.js --id=\\\"\" + id_page + \"\\\" --token=\\\"\" + access_token \\\n + \"\\\" --title=\\\"\" + title + \"\\\" --des=\\\"\" \\\n + des + \" \\\" --video=\\\"\" + link_video + \"\\\" --tags=\\\"\" + genres + \"\\\" --thumb=\\\"\" + thumbnail + \"\\\"\"\n else:\n string_upload = \"sudo node upload-video-to-facebook/main.js --id=\\\"\" + id_page + \"\\\" --token=\\\"\" + access_token \\\n + \"\\\" --title=\\\"\" + title + \"\\\" --des=\\\"\" \\\n + des + \" \\\" --video=\\\"\" + link_video + \"\\\" --tags=\\\"\" + genres + \"\\\" --thumb=\\\"\" + thumbnail + \"\\\"\"\n\n os.system(string_upload)\n check = True\n\n empty_folder(path_page + \"/input\")\n\n if thumbnail != '':\n os.remove(path_page + '/thumbnail.jpg')\n\n return check\n\n\ndef get_list_video(info_api, path_page, path_thumb, account_id):\n print(\"Get list video..\")\n source = info_api['source']\n data_channel = info_api['data_channel']\n\n channel_id = info_api['channel_id']\n access_token = info_api['access_token']\n cookie = info_api['cookie']\n\n items = get_list_video_by_api(source, data_channel)\n\n for id_video in items:\n info = get_tags(id_video)\n\n title = info[0]\n tags = info[1]\n\n description = title\n\n check = False\n\n thumbnail = get_thumbnail(\"https://www.youtube.com/watch?v=\" + str(id_video), path_thumb)\n\n has_video = download_video_from_youtube(id_video, path_page)\n\n if has_video:\n check = hanlde(access_token, cookie, title, description, tags, thumbnail, path_page, path_thumb, account_id)\n else:\n save_data_by_api(channel_id, id_video)\n\n if check:\n save_data_by_api(channel_id, id_video)\n\n print(\"Done\")\n print(\"Channel id:\" + str(channel_id))\n # time.sleep(7200)\n else:\n # update_check_point(account_id)\n pass\n break\n\n\ndef save_data_by_api(channel_id, video_id):\n url = \"http://\" + main_api + \"/data/set.php\"\n\n data = {\n 'video_id': video_id,\n 'channel_id': channel_id,\n 'key': key\n }\n\n req = requests.post(url, data=data)\n\n if req.status_code != 200:\n return False\n\n return True\n\n\ndef get_info_by_api(page_number, account_id):\n url = \"http://\" + main_api + \"/accesstoken/get.php\"\n\n results = {\n 'status_code': 200,\n 'data': []\n }\n data = {\n 'page_number': page_number,\n 'account_id': account_id,\n 'key': key\n }\n\n req = requests.get(url, params=data)\n\n results['status_code'] = req.status_code\n\n if req.status_code != 200:\n return results\n\n datas = req.json()\n\n if int(datas['status']) != 0:\n access_token = datas['accesstoken']\n cookie = generate_cookie(datas['cookie'])\n channel_id = datas['channel_id']\n\n source = get_source(channel_id)\n data_channel = get_data_channel(channel_id)\n\n result = {\n 'access_token': access_token,\n 'cookie': cookie,\n 'channel_id': channel_id,\n 'source': source,\n 'data_channel': data_channel\n }\n\n results['data'] = result\n\n return results\n\n\ndef get_source(channel_id):\n url = \"http://\" + main_api + \"/channel/get.php\"\n\n data = {\n 'channel_id': channel_id,\n 'key': key\n }\n\n req = requests.get(url, params=data)\n\n if req.status_code != 200:\n return False\n data = req.json()\n\n return data['records'][0]['source']\n\n\ndef get_data_channel(channel_id):\n url = \"http://\" + main_api + \"/data/get.php\"\n\n data = {\n 'channel_id': channel_id,\n 'key': key\n }\n\n req = requests.get(url, params=data)\n\n if req.status_code != 200:\n return False\n\n data = req.json()\n\n return data['records']\n\n\ndef check_and_create_dir(account_id, page_number):\n path_account = path + '/' + account_id\n path_page = path + '/' + account_id + '/' + page_number\n\n if os.path.isdir(account_id) is False:\n os.mkdir(path_account)\n\n if os.path.isdir(path_page) is False:\n os.mkdir(path_page)\n\n if os.path.isdir(path_page + \"/input\") is False:\n os.mkdir(path_page + \"/input\")\n\n if os.path.isdir(path_page + \"/output\") is False:\n os.mkdir(path_page + \"/output\")\n\n\ndef empty_folder(path_folder):\n shutil.rmtree(path_folder)\n os.makedirs(path_folder)\n\n\ndef update_data():\n channel_id = str(input(\"Channel id: \"))\n list_id = str(input(\"List id: \"))\n\n url = \"http://\" + main_api + \"/data/update_data.php\"\n\n data = {\n 'channel_id': channel_id,\n 'list_id': list_id\n }\n\n req = requests.post(url, data=data)\n print(req.status_code)\n\n\ndef setupDataToServer(account_id, page_number):\n source = str(input(\"Source: \"))\n access_token = str(input(\"Access token: \"))\n\n url = \"http://\" + main_api + \"/accesstoken/set.php\"\n\n data = {\n 'account_id': account_id,\n 'page_number': page_number,\n 'source': source,\n 'accesstoken': access_token,\n 'key': key\n }\n\n req = requests.post(url, data=data)\n\n if req.status_code == 200:\n return True\n\n return False\n\n\ndef auto(arr):\n count_reset = 0\n stt = 0\n\n while True:\n try:\n for i in range(len(arr)):\n account_id = str(i + 1)\n\n if stt > len(arr[i]) - 1:\n count_reset = count_reset + 1\n continue\n\n try:\n page_number = str(arr[i][stt])\n except IndexError:\n count_reset = count_reset + 1\n continue\n\n path_page = path + '/' + account_id + '/' + page_number\n path_thumb = account_id + '/' + page_number\n check_and_create_dir(account_id, page_number)\n\n result = get_info_by_api(page_number, account_id)\n\n if result['status_code'] != 200:\n print(\"Setup new data!\")\n setupDataToServer(account_id, page_number)\n result = get_info_by_api(page_number, account_id)\n\n info = result['data']\n\n if len(info) == 0:\n count_reset = count_reset + 1\n continue\n\n get_list_video(info, path_page, path_thumb, account_id)\n\n stt = stt + 1\n\n if count_reset >= len(arr):\n count_reset = 0\n stt = 0\n\n time.sleep(600)\n except exceptions.ConnectionError:\n print(\"Error Connect!\")\n time.sleep(300)\n\n\ndef default():\n account_id = str(input(\"Account id: \"))\n page_number = str(input(\"Page number: \"))\n\n path_page = path + '/' + account_id + '/' + page_number\n path_thumb = account_id + '/' + page_number\n check_and_create_dir(account_id, page_number)\n\n while True:\n try:\n result = get_info_by_api(page_number, account_id)\n\n if result['status_code'] != 200:\n print(\"Setup new data!\")\n setupDataToServer(account_id, page_number)\n result = get_info_by_api(page_number, account_id)\n\n info = result['data']\n\n if len(info) == 0:\n print(\"Account have been checkpoint!\")\n continue\n\n get_list_video(info, path_page, path_thumb, account_id)\n\n time.sleep(600)\n except exceptions.ConnectionError:\n print(\"Error connect!\")\n time.sleep(100)\n\n\ndef get_platform():\n platforms = {\n 'linux1': 'Linux',\n 'linux2': 'Linux',\n 'darwin': 'OS X',\n 'win32': 'Windows'\n }\n if sys.platform not in platforms:\n return sys.platform\n\n return platforms[sys.platform]\n\n\ndef update_check_point(account_id):\n print(\"Update Check point! Account id: \" + str(account_id))\n url = \"http://\" + main_api + \"/account/updateCheckpoint.php\"\n\n data = {\n 'account_id': account_id,\n 'key': key\n }\n\n req = requests.post(url, data=data)\n\n if req.status_code != 200:\n return False\n\n return True\n\n\ndef generate_cookie(string_cookie):\n if string_cookie == '':\n return {}\n\n string_cookie = string_cookie.replace(\" \", \"\")\n arr = string_cookie.split(\";\")\n result = {}\n\n for i in range(len(arr)):\n key, value = arr[i].split(\"=\")\n\n result[key] = value\n\n return result\n\n\nif __name__ == '__main__':\n arr_page = [[1, 2, 3], [3, 4], [1, 2], [1, 2, 3], [], []]\n\n option = str(input(\"One page (0) OR All page (1) ? \"))\n\n if option == \"0\":\n default()\n else:\n auto(arr_page)\n", "sub_path": "main-blk.py", "file_name": "main-blk.py", "file_ext": "py", "file_size_in_byte": 18464, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.getcwd", "line_number": 13, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 46, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 48, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 84, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 90, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 98, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 114, "usage_type": "call"}, {"api_name": "os.system", "line_number": 164, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 180, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 181, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 207, "usage_type": "call"}, {"api_name": "moviepy.editor.VideoFileClip", "line_number": 218, "usage_type": "call"}, {"api_name": "os.system", "line_number": 243, "usage_type": "call"}, {"api_name": "os.path.getsize", "line_number": 255, "usage_type": "call"}, {"api_name": "os.path", "line_number": 255, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 270, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 272, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 288, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 300, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 303, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 316, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 316, "usage_type": "attribute"}, {"api_name": "subprocess.STDOUT", "line_number": 316, "usage_type": "attribute"}, {"api_name": "subprocess.getoutput", "line_number": 326, "usage_type": "call"}, {"api_name": "os.system", "line_number": 372, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 378, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 434, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 455, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 493, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 510, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 524, "usage_type": "call"}, {"api_name": "os.path", "line_number": 524, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 525, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 527, "usage_type": "call"}, {"api_name": "os.path", "line_number": 527, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 528, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 530, "usage_type": "call"}, {"api_name": "os.path", "line_number": 530, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 531, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 533, "usage_type": "call"}, {"api_name": "os.path", "line_number": 533, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 534, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 538, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 539, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 553, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 571, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 623, "usage_type": "call"}, {"api_name": "requests.exceptions.ConnectionError", "line_number": 624, "usage_type": "attribute"}, {"api_name": "requests.exceptions", "line_number": 624, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 626, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 654, "usage_type": "call"}, {"api_name": "requests.exceptions.ConnectionError", "line_number": 655, "usage_type": "attribute"}, {"api_name": "requests.exceptions", "line_number": 655, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 657, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 667, "usage_type": "attribute"}, {"api_name": "sys.platform", "line_number": 668, "usage_type": "attribute"}, {"api_name": "sys.platform", "line_number": 670, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 682, "usage_type": "call"}]}
+{"seq_id": "314674165", "text": "\"\"\" Commands for downloading from s3 \"\"\"\nimport os\nimport sys\n\nimport click\n\nfrom bwdt.lib.aws.s3 import S3\nfrom bwdt.constants import (APT_TARGZ_KEY, BWDT_TARGZ_KEY, CLOUDYML_KEY,\n S3_BUCKET)\n\n\ndef _save(path, key, force):\n \"\"\" Download the specified file \"\"\"\n files_dir = '{}/{}'.format(path, 'files')\n if os.path.isdir(path):\n full_path = '{}/{}'.format(files_dir, key)\n else:\n err = 'ERROR: path {} must be a directory and exist\\n'.format(path)\n sys.stderr.write(err)\n return False\n if not os.path.exists(files_dir):\n os.mkdir(files_dir)\n if os.path.exists(full_path) and not force:\n err = 'WARN: File {} exists. Use --force to replace'.format(full_path)\n click.echo(err)\n return False\n click.echo('Saving {}'.format(full_path))\n full_path = '{}/{}'.format(files_dir, key)\n S3().download(full_path, S3_BUCKET, key)\n return True\n\n\ndef offline_apt(path, force):\n \"\"\" Download a subset of breqwatr/apt for offline installs \"\"\"\n _save(path, APT_TARGZ_KEY, force)\n\n\ndef offline_bwdt(path, force):\n \"\"\" Download an offline export of this bwdt tool \"\"\"\n _save(path, BWDT_TARGZ_KEY, force)\n\n\ndef cloud_yml(path, force):\n \"\"\" Download a commented cloud.yml template \"\"\"\n _save(path, CLOUDYML_KEY, force)\n", "sub_path": "bwdt/lib/download.py", "file_name": "download.py", "file_ext": "py", "file_size_in_byte": 1330, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.path.isdir", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 19, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "click.echo", "line_number": 25, "usage_type": "call"}, {"api_name": "click.echo", "line_number": 27, "usage_type": "call"}, {"api_name": "bwdt.constants.S3_BUCKET", "line_number": 29, "usage_type": "argument"}, {"api_name": "bwdt.lib.aws.s3.S3", "line_number": 29, "usage_type": "call"}, {"api_name": "bwdt.constants.APT_TARGZ_KEY", "line_number": 35, "usage_type": "argument"}, {"api_name": "bwdt.constants.BWDT_TARGZ_KEY", "line_number": 40, "usage_type": "argument"}, {"api_name": "bwdt.constants.CLOUDYML_KEY", "line_number": 45, "usage_type": "argument"}]}
+{"seq_id": "494584323", "text": "from itertools import islice\n\nfrom rest_framework.response import Response\nfrom rest_framework import status\n\nfrom query_base.query_base import QueryBaseView\n\n\nclass DefaultGeneModelsId(QueryBaseView):\n def get(self, request):\n default_gene_models_id = \\\n self.gpf_instance.dae_config.gene_models.resource_id\n return Response(default_gene_models_id, status=status.HTTP_200_OK)\n\n\nclass GeneModels(QueryBaseView):\n def get(self, request, gene_symbol):\n gene_symbol = gene_symbol.lower()\n gene_models = self.gpf_instance.gene_models.gene_models\n for k, v in gene_models.items():\n if gene_symbol == k.lower():\n transcripts = v\n response_data = {\n \"gene\": k,\n \"transcripts\": [],\n }\n for tr in transcripts:\n response_data[\"transcripts\"].append(\n self.transcript_to_dict(tr)\n )\n\n return Response(\n response_data,\n status=status.HTTP_200_OK,\n )\n return Response(None, status=status.HTTP_404_NOT_FOUND)\n\n def transcript_to_dict(self, transcript):\n output = dict()\n output[\"transcript_id\"] = transcript.tr_id\n output[\"strand\"] = transcript.strand\n output[\"chrom\"] = transcript.chrom\n output[\"cds\"] = self.cds_to_dictlist(transcript.cds)\n output[\"utr3\"] = list()\n for region in transcript.UTR3_regions():\n output[\"utr3\"].append(self.region_to_dict(region))\n output[\"utr5\"] = list()\n for region in transcript.UTR5_regions():\n output[\"utr5\"].append(self.region_to_dict(region))\n output[\"exons\"] = list()\n for exon in transcript.exons:\n output[\"exons\"].append(self.exon_to_dict(exon))\n return output\n\n def cds_to_dictlist(self, cds):\n return [\n {\"start\": a, \"stop\": b}\n for (a, b) in zip(cds[::2], cds[1::2])\n ]\n\n def region_to_dict(self, region):\n return {\n \"start\": region.start,\n \"stop\": region.stop\n }\n\n def exon_to_dict(self, exon):\n return {\n \"start\": exon.start,\n \"stop\": exon.stop\n }\n\n\nclass GeneSymbolsSearch(QueryBaseView):\n\n RESPONSE_LIMIT = 20\n\n def get(self, request, search_term):\n search_term = search_term.lower()\n gene_models = self.gpf_instance.gene_models.gene_models\n\n matching_gene_symbols = filter(\n lambda gs: gs.lower().startswith(search_term),\n gene_models.keys()\n )\n\n matching_gene_symbols = islice(\n matching_gene_symbols, None, self.RESPONSE_LIMIT\n )\n\n return Response(\n {\"gene_symbols\": list(matching_gene_symbols)},\n status=status.HTTP_200_OK,\n )\n", "sub_path": "wdae/wdae/genomes_api/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2914, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "query_base.query_base.QueryBaseView", "line_number": 9, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 13, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 13, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 13, "usage_type": "name"}, {"api_name": "query_base.query_base.QueryBaseView", "line_number": 16, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 32, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 34, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 34, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 36, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 36, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 36, "usage_type": "name"}, {"api_name": "query_base.query_base.QueryBaseView", "line_number": 74, "usage_type": "name"}, {"api_name": "itertools.islice", "line_number": 87, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 91, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 93, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 93, "usage_type": "name"}]}
+{"seq_id": "540447348", "text": "#!/usr/bin/env python3\n\nimport argparse\nfrom collections import namedtuple\nfrom shard_prometheus_utils import handle_install, install_parser, \\\n handle_upgrade, upgrade_parser, handle_delete, delete_parser\n\nCommandHandler = namedtuple('CommandHandler', ['handler', 'parser'])\n\nops = {\n 'install': CommandHandler(handle_install, install_parser),\n 'upgrade': CommandHandler(handle_upgrade, upgrade_parser),\n 'delete': CommandHandler(handle_delete, delete_parser)\n}\n\n\n \ndef parse_args():\n parser = argparse.ArgumentParser(\n formatter_class=argparse.ArgumentDefaultsHelpFormatter)\n subparsers = parser.add_subparsers(help='install prometheus and thanos',\n dest=\"subcommand\")\n for key in ops:\n ops[key].parser(subparsers, key)\n\n return parser.parse_args() \n\ndef main():\n args = parse_args()\n ops[args.subcommand].handler(args)\n\nif __name__ == \"__main__\":\n main()\n", "sub_path": "charts/prometheus/shard_prometheus.py", "file_name": "shard_prometheus.py", "file_ext": "py", "file_size_in_byte": 938, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "collections.namedtuple", "line_number": 8, "usage_type": "call"}, {"api_name": "shard_prometheus_utils.handle_install", "line_number": 11, "usage_type": "argument"}, {"api_name": "shard_prometheus_utils.install_parser", "line_number": 11, "usage_type": "argument"}, {"api_name": "shard_prometheus_utils.handle_upgrade", "line_number": 12, "usage_type": "argument"}, {"api_name": "shard_prometheus_utils.upgrade_parser", "line_number": 12, "usage_type": "argument"}, {"api_name": "shard_prometheus_utils.handle_delete", "line_number": 13, "usage_type": "argument"}, {"api_name": "shard_prometheus_utils.delete_parser", "line_number": 13, "usage_type": "argument"}, {"api_name": "argparse.ArgumentParser", "line_number": 19, "usage_type": "call"}, {"api_name": "argparse.ArgumentDefaultsHelpFormatter", "line_number": 20, "usage_type": "attribute"}]}
+{"seq_id": "302774469", "text": "import requests\nfrom bs4 import BeautifulSoup\n\nurl = \"http://python123.io/ws/demo.html\"\nr = requests.get(url)\ndemo = r.text\nsoup = BeautifulSoup(demo, \"html.parser\")\n# print(soup.prettify())\nfor link in soup.find_all('a'):\n print(link.get('href'))\n\n", "sub_path": "Requests/week2/extracthtml.py", "file_name": "extracthtml.py", "file_ext": "py", "file_size_in_byte": 252, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "requests.get", "line_number": 5, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 7, "usage_type": "call"}]}
+{"seq_id": "611661175", "text": "# -*- coding:utf-8 -*-\nimport logging\ndef init_log(logfile):\n logger = logging.getLogger()\n hdlr = logging.FileHandler(logfile)\n formatter = logging.Formatter(\"%(asctime)s%(message)s\")\n hdlr.setFormatter(formatter)\n logger.addHandler(hdlr)\n logger.setLevel(logging.NOTSET)\n return logger\nif __name__==\"__main__\":\n file = r\"C:\\Users\\Wuxiaoshen\\Desktop\\test(1).py\"\n A=init_log(file)\n print(A.info(file))\n\n", "sub_path": "001/009TestLog.py", "file_name": "009TestLog.py", "file_ext": "py", "file_size_in_byte": 433, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.getLogger", "line_number": 4, "usage_type": "call"}, {"api_name": "logging.FileHandler", "line_number": 5, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 6, "usage_type": "call"}, {"api_name": "logging.NOTSET", "line_number": 9, "usage_type": "attribute"}]}
+{"seq_id": "375616141", "text": "import os\nimport torch\nimport numpy as np\nimport torch.nn as nn\nfrom torchvision import transforms\nimport datetime\nimport sys\nfrom pycrayon import CrayonClient\nfrom torch.utils.data import DataLoader\n\nimport network\nimport utils\nfrom model import MultiColumnNet\nfrom data_loader import DownSampleGT, ToTensor, CnrParkExtDataset\nimport evaluate_model\n\n# General parameters\nROOT_DIR = os.getcwd()\nTRAIN_DATASET_DIR = os.path.join(ROOT_DIR, \"../datasets/cnr_park_ext/train/patches\")\nTRAIN_DENSITY_MAPS_DIR = os.path.join(TRAIN_DATASET_DIR, \"density_maps\")\nVAL_DATASET_DIR = os.path.join(ROOT_DIR, \"../datasets/cnr_park_ext/val/patches\")\nVAL_DENSITY_MAPS_DIR = os.path.join(VAL_DATASET_DIR, \"density_maps\")\nIS_CUDA = True\nOUTPUT_DIR = os.path.join(ROOT_DIR, \"../output_images\")\nTIMESTAMP = datetime.datetime.now().strftime(\"%y%m%d%H%M\")\nSAVED_MODELS_DIR = os.path.join(ROOT_DIR, \"../saved_models\", TIMESTAMP)\n\n# Training parameters\nLEARNING_RATE = 0.0001\nMOMENTUM = 0.9\nSTART_STEP = 0\nEND_STEP = 2000\ncriterion = nn.MSELoss()\nLOSS = nn.MSELoss()\nTRAIN_BATCH_SIZE = 1\nDOWN_SAMPLE_FACTOR = 4\nSHUFFLE = True\nSAVE_INTERVAL = 1\n\n# Tensorboard\ncc = CrayonClient(hostname='127.0.0.1')\nexp_name = \"exp_{}\".format(TIMESTAMP)\n# cc.remove_all_experiments()\nexp = cc.create_experiment(exp_name)\n\n# Loading train and validation dataset\ntrain_dataset = CnrParkExtDataset(TRAIN_DATASET_DIR, TRAIN_DENSITY_MAPS_DIR, transform=transforms.Compose([\n DownSampleGT(DOWN_SAMPLE_FACTOR),\n ToTensor(is_cuda=IS_CUDA)\n ]))\nval_dataset = CnrParkExtDataset(VAL_DATASET_DIR, VAL_DENSITY_MAPS_DIR, transform=transforms.Compose([\n DownSampleGT(DOWN_SAMPLE_FACTOR),\n ToTensor(is_training=False, is_cuda=IS_CUDA)\n ]))\n# Creating PyTorch data loader\n#dataloader = DataLoader(train_dataset, batch_size=TRAIN_BATCH_SIZE, shuffle=SHUFFLE)\n\n# Loading network\nnet = MultiColumnNet()\n# network.weightsInit(net)\nif IS_CUDA:\n net.cuda()\nnet.train()\n\nparameters = list(net.parameters())\noptimizer = torch.optim.Adam(filter(lambda p: p.requires_grad, net.parameters()), lr=LEARNING_RATE)\n\n# Training\nutils.log_print(\"Start training\", color='red', attrs=['bold'])\nbest_mae, best_mse = sys.maxsize, sys.maxsize\n\nfor epoch in range(START_STEP, END_STEP + 1):\n train_loss = 0\n if epoch == 0:\n os.mkdir(os.path.join(OUTPUT_DIR, TIMESTAMP))\n os.mkdir(SAVED_MODELS_DIR)\n if epoch % SAVE_INTERVAL == 0:\n output_dir = os.path.join(OUTPUT_DIR, TIMESTAMP, \"epoch{}\".format(epoch))\n os.mkdir(output_dir)\n\n for i in range(len(train_dataset)):\n sample = train_dataset[i]\n utils.log_print(\"Processing image: {} with name: {}\".format(i, sample['image_name']))\n image, gt_density_map, image_name = sample['image'], sample['density_map'], sample['image_name']\n binary_gt_density_map = gt_density_map.clone()\n binary_gt_density_map[binary_gt_density_map > 0] = 1\n density_map = net(image, binary_gt_density_map)\n # Debug\n #unique, counts = np.unique(gt_density_map.data.cpu().numpy(), return_counts=True)\n #print(dict(zip(unique, counts)))\n utils.log_print(\"Input image shape: {}, Density Map GT Shape {}, Density Map Shape {}\".format(image.shape, gt_density_map.shape, density_map.shape))\n loss = LOSS(density_map, gt_density_map)\n train_loss += loss.item()\n optimizer.zero_grad()\n loss.backward()\n optimizer.step()\n\n if epoch % SAVE_INTERVAL == 0:\n gt_count = np.sum(gt_density_map.data.cpu().numpy())\n estimated_count = np.sum(density_map.data.cpu().numpy())\n utils.log_print(\"Saving results: epoch: {}, gt_count: {}, estimated_count: {}\".format(epoch, gt_count, estimated_count), color='green')\n utils.saveResults(image, gt_density_map, density_map, output_dir, image_name, DOWN_SAMPLE_FACTOR)\n\n if epoch % (SAVE_INTERVAL * 2) == 0:\n file_name = os.path.join(SAVED_MODELS_DIR, \"network_epoch{}.h5\".format(epoch))\n network.saveSnapshot(file_name, net)\n mae, mse = evaluate_model.evaluateModel(file_name, val_dataset, IS_CUDA)\n if mae < best_mae:\n best_mae = mae\n best_model = \"network_epoch{}.h5\".format(epoch)\n if mse < best_mse:\n best_mse = mse\n utils.log_print(\"Saving snapshot: epoch: {}, mae: {}, mse: {}, best_mae: {}, best_mse: {}, best_model: {}\".format(epoch, mae, mse, best_mae, best_mse, best_model), color='green', attrs=['bold'])\n exp.add_scalar_value('MAE', mae, step=epoch)\n exp.add_scalar_value('MSE', mse, step=epoch)\n exp.add_scalar_value('train_loss', train_loss / len(train_dataset), step=epoch)\n\n\n", "sub_path": "src/train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 5187, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.getcwd", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "torch.nn.MSELoss", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 33, "usage_type": "name"}, {"api_name": "torch.nn.MSELoss", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 34, "usage_type": "name"}, {"api_name": "pycrayon.CrayonClient", "line_number": 41, "usage_type": "call"}, {"api_name": "data_loader.CnrParkExtDataset", "line_number": 47, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 47, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 47, "usage_type": "name"}, {"api_name": "data_loader.DownSampleGT", "line_number": 48, "usage_type": "call"}, {"api_name": "data_loader.ToTensor", "line_number": 49, "usage_type": "call"}, {"api_name": "data_loader.CnrParkExtDataset", "line_number": 51, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 51, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 51, "usage_type": "name"}, {"api_name": "data_loader.DownSampleGT", "line_number": 52, "usage_type": "call"}, {"api_name": "data_loader.ToTensor", "line_number": 53, "usage_type": "call"}, {"api_name": "model.MultiColumnNet", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 66, "usage_type": "attribute"}, {"api_name": "utils.log_print", "line_number": 69, "usage_type": "call"}, {"api_name": "sys.maxsize", "line_number": 70, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path", "line_number": 75, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path", "line_number": 78, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 79, "usage_type": "call"}, {"api_name": "utils.log_print", "line_number": 83, "usage_type": "call"}, {"api_name": "utils.log_print", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 100, "usage_type": "call"}, {"api_name": "utils.log_print", "line_number": 101, "usage_type": "call"}, {"api_name": "utils.saveResults", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 105, "usage_type": "call"}, {"api_name": "os.path", "line_number": 105, "usage_type": "attribute"}, {"api_name": "network.saveSnapshot", "line_number": 106, "usage_type": "call"}, {"api_name": "evaluate_model.evaluateModel", "line_number": 107, "usage_type": "call"}, {"api_name": "utils.log_print", "line_number": 113, "usage_type": "call"}]}
+{"seq_id": "407270580", "text": "import copy\nimport os\nfrom collections import OrderedDict, defaultdict\nfrom contextlib import ExitStack\nfrom functools import wraps\nfrom typing import Union, Tuple, List, Optional, Iterator\n\nfrom ..cli.parser import set_router_parser, set_indexer_parser, \\\n set_frontend_parser, set_preprocessor_parser, \\\n set_encoder_parser, set_client_cli_parser\nfrom ..client.cli import CLIClient\nfrom ..helper import set_logger\nfrom ..service.base import SocketType, BaseService, BetterEnum, ServiceManager\nfrom ..service.encoder import EncoderService\nfrom ..service.frontend import FrontendService\nfrom ..service.indexer import IndexerService\nfrom ..service.preprocessor import PreprocessorService\nfrom ..service.router import RouterService\n\n\nclass Service(BetterEnum):\n Frontend = 0\n Encoder = 1\n Router = 2\n Indexer = 3\n Preprocessor = 4\n\n\nclass FlowImcompleteError(ValueError):\n \"\"\"Exception when the flow missing some important component to run\"\"\"\n\n\nclass FlowTopologyError(ValueError):\n \"\"\"Exception when the topology is ambiguous\"\"\"\n\n\nclass FlowBuildLevelMismatch(ValueError):\n \"\"\"Exception when required level is higher than the current build level\"\"\"\n\n\ndef _build_level(required_level: 'Flow.BuildLevel'):\n def __build_level(func):\n @wraps(func)\n def arg_wrapper(self, *args, **kwargs):\n if hasattr(self, '_build_level'):\n if self._build_level.value >= required_level.value:\n return func(self, *args, **kwargs)\n else:\n raise FlowBuildLevelMismatch(\n 'build_level check failed for %r, required level: %s, actual level: %s' % (\n func, required_level, self._build_level))\n else:\n raise AttributeError('%r has no attribute \"_build_level\"' % self)\n\n return arg_wrapper\n\n return __build_level\n\n\nclass Flow:\n \"\"\"\n GNES Flow: an intuitive way to build workflow for GNES.\n\n You can use :py:meth:`.add()` then :py:meth:`.build()` to customize your own workflow.\n For example:\n\n .. highlight:: python\n .. code-block:: python\n\n from gnes.flow import Flow, Service as gfs\n\n f = (Flow(check_version=False, route_table=True)\n .add(gfs.Preprocessor, yaml_path='BasePreprocessor')\n .add(gfs.Encoder, yaml_path='BaseEncoder')\n .add(gfs.Router, yaml_path='BaseRouter'))\n\n with f.build(backend='thread') as flow:\n flow.index()\n ...\n\n You can also use the shortcuts, e.g. :py:meth:`add_encoder`, :py:meth:`add_preprocessor`.\n\n It is recommend to use flow in the context manner as showed above.\n\n Note the different default copy behaviors in :py:meth:`.add()` and :py:meth:`.build()`:\n :py:meth:`.add()` always copy the flow by default, whereas :py:meth:`.build()` modify the flow in place.\n You can change this behavior by giving an argument `copy_flow=False`.\n\n \"\"\"\n _supported_orch = {'swarm', 'k8s'}\n _service2parser = {\n Service.Encoder: set_encoder_parser,\n Service.Router: set_router_parser,\n Service.Indexer: set_indexer_parser,\n Service.Frontend: set_frontend_parser,\n Service.Preprocessor: set_preprocessor_parser,\n }\n _service2builder = {\n Service.Encoder: lambda x: ServiceManager(EncoderService, x),\n Service.Router: lambda x: ServiceManager(RouterService, x),\n Service.Indexer: lambda x: ServiceManager(IndexerService, x),\n Service.Preprocessor: lambda x: ServiceManager(PreprocessorService, x),\n Service.Frontend: FrontendService,\n }\n\n class BuildLevel(BetterEnum):\n EMPTY = 0\n GRAPH = 1\n RUNTIME = 2\n\n def __init__(self, with_frontend: bool = True, **kwargs):\n self.logger = set_logger(self.__class__.__name__)\n self._service_nodes = OrderedDict()\n self._service_edges = {}\n self._service_name_counter = {k: 0 for k in Flow._service2parser.keys()}\n self._service_contexts = []\n self._last_add_service = None\n self._common_kwargs = kwargs\n self._frontend = None\n self._client = None\n self._build_level = Flow.BuildLevel.EMPTY\n self._backend = None\n if with_frontend:\n self.add_frontend(copy_flow=False)\n else:\n self.logger.warning('with_frontend is set to False, you need to add_frontend() by yourself')\n\n @_build_level(BuildLevel.GRAPH)\n def to_yaml(self, orchestration: str) -> str:\n if orchestration not in Flow._supported_orch:\n raise TypeError(\n '%s is not valid type of orchestration, should be one of %s' % (orchestration, Flow._supported_orch))\n\n @staticmethod\n def from_yaml(orchestration: str) -> 'Flow':\n if orchestration not in Flow._supported_orch:\n raise TypeError(\n '%s is not valid type of orchestration, should be one of %s' % (orchestration, Flow._supported_orch))\n\n @_build_level(BuildLevel.GRAPH)\n def to_mermaid(self, left_right: bool = True):\n \"\"\"\n Output the mermaid graph for visualization\n\n :param left_right: render the flow in left-to-right manner, otherwise top-down manner.\n :return:\n \"\"\"\n mermaid_graph = OrderedDict()\n for k in self._service_nodes.keys():\n mermaid_graph[k] = []\n cls_dict = defaultdict(set)\n\n for k, ed_type in self._service_edges.items():\n start_node, end_node = k.split('-')\n s_service = self._service_nodes[start_node]['service']\n e_service = self._service_nodes[end_node]['service']\n cls_dict[s_service].add(start_node)\n cls_dict[e_service].add(end_node)\n p_s = '((%s))' if s_service == Service.Router else '(%s)'\n p_e = '((%s))' if e_service == Service.Router else '(%s)'\n mermaid_graph[start_node].append('\\t%s%s-- %s -->%s%s' % (\n start_node, p_s % start_node, ed_type,\n end_node, p_e % end_node))\n\n style = ['classDef FrontendCLS fill:#FFE0E0,stroke:#FFE0E0,stroke-width:1px;',\n 'classDef EncoderCLS fill:#FFDAAF,stroke:#FFDAAF,stroke-width:1px;',\n 'classDef IndexerCLS fill:#FFFBC1,stroke:#FFFBC1,stroke-width:1px;',\n 'classDef RouterCLS fill:#C9E8D2,stroke:#C9E8D2,stroke-width:1px;',\n 'classDef PreprocessorCLS fill:#CEEEEF,stroke:#CEEEEF,stroke-width:1px;']\n class_def = ['class %s %sCLS;' % (','.join(v), k) for k, v in cls_dict.items()]\n mermaid_str = '\\n'.join(\n ['graph %s' % ('LR' if left_right else 'TD')] + [ss for s in mermaid_graph.values() for ss in\n s] + style + class_def)\n\n return mermaid_str\n\n @_build_level(BuildLevel.GRAPH)\n def to_jpg(self, path: str = 'flow.jpg', left_right: bool = True):\n \"\"\"\n Rendering the current flow as a jpg image, this will call :py:meth:`to_mermaid` and it needs internet connection\n\n :param path: the file path of the image\n :param left_right: render the flow in left-to-right manner, otherwise top-down manner.\n :return:\n \"\"\"\n import base64\n from urllib.request import Request, urlopen\n mermaid_str = self.to_mermaid(left_right)\n encoded_str = base64.b64encode(bytes(mermaid_str, 'utf-8')).decode('utf-8')\n print('https://mermaidjs.github.io/mermaid-live-editor/#/view/%s' % encoded_str)\n self.logger.info('saving jpg...')\n req = Request('https://mermaid.ink/img/%s' % encoded_str, headers={'User-Agent': 'Mozilla/5.0'})\n with open(path, 'wb') as fp:\n fp.write(urlopen(req).read())\n self.logger.info('done')\n\n def train(self, bytes_gen: Iterator[bytes] = None, **kwargs):\n \"\"\"Do training on the current flow\n\n It will start a :py:class:`CLIClient` and call :py:func:`train`.\n\n :param bytes_gen: An iterator of bytes. If not given, then you have to specify it in `kwargs`.\n :param kwargs: accepts all keyword arguments of `gnes client` CLI\n \"\"\"\n self._call_client(bytes_gen, mode='train', **kwargs)\n\n def index(self, bytes_gen: Iterator[bytes] = None, **kwargs):\n \"\"\"Do indexing on the current flow\n\n It will start a :py:class:`CLIClient` and call :py:func:`index`.\n\n :param bytes_gen: An iterator of bytes. If not given, then you have to specify it in `kwargs`.\n :param kwargs: accepts all keyword arguments of `gnes client` CLI\n \"\"\"\n self._call_client(bytes_gen, mode='index', **kwargs)\n\n def query(self, bytes_gen: Iterator[bytes] = None, **kwargs):\n \"\"\"Do indexing on the current flow\n\n It will start a :py:class:`CLIClient` and call :py:func:`query`.\n\n :param bytes_gen: An iterator of bytes. If not given, then you have to specify it in `kwargs`.\n :param kwargs: accepts all keyword arguments of `gnes client` CLI\n \"\"\"\n self._call_client(bytes_gen, mode='query', **kwargs)\n\n @_build_level(BuildLevel.RUNTIME)\n def _call_client(self, bytes_gen: Iterator[bytes] = None, **kwargs):\n os.unsetenv('http_proxy')\n os.unsetenv('https_proxy')\n args, p_args = self._get_parsed_args(self, set_client_cli_parser, kwargs)\n p_args.grpc_port = self._service_nodes[self._frontend]['parsed_args'].grpc_port\n p_args.grpc_host = self._service_nodes[self._frontend]['parsed_args'].grpc_host\n c = CLIClient(p_args, start_at_init=False)\n if bytes_gen:\n c.bytes_generator = bytes_gen\n c.start()\n\n def add_frontend(self, *args, **kwargs) -> 'Flow':\n \"\"\"Add a frontend to the current flow, a shortcut of :py:meth:`add(Service.Frontend)`.\n Usually you dont need to call this function explicitly, a flow object contains a frontend service by default.\n This function is useful when you build a flow without the frontend and want to customize the frontend later.\n \"\"\"\n return self.add(Service.Frontend, *args, **kwargs)\n\n def add_encoder(self, *args, **kwargs) -> 'Flow':\n \"\"\"Add an encoder to the current flow, a shortcut of :py:meth:`add(Service.Encoder)`\"\"\"\n return self.add(Service.Encoder, *args, **kwargs)\n\n def add_indexer(self, *args, **kwargs) -> 'Flow':\n \"\"\"Add an indexer to the current flow, a shortcut of :py:meth:`add(Service.Indexer)`\"\"\"\n return self.add(Service.Indexer, *args, **kwargs)\n\n def add_preprocessor(self, *args, **kwargs) -> 'Flow':\n \"\"\"Add a preprocessor to the current flow, a shortcut of :py:meth:`add(Service.Preprocessor)`\"\"\"\n return self.add(Service.Preprocessor, *args, **kwargs)\n\n def add_router(self, *args, **kwargs) -> 'Flow':\n \"\"\"Add a router to the current flow, a shortcut of :py:meth:`add(Service.Router)`\"\"\"\n return self.add(Service.Router, *args, **kwargs)\n\n def add(self, service: 'Service',\n name: str = None,\n service_in: Union[str, Tuple[str], List[str], 'Service'] = None,\n service_out: Union[str, Tuple[str], List[str], 'Service'] = None,\n copy_flow: bool = True,\n **kwargs) -> 'Flow':\n \"\"\"\n Add a service to the current flow object and return the new modified flow object\n\n :param service: a 'Service' enum, possible choices: Encoder, Router, Preprocessor, Indexer, Frontend\n :param name: the name indentifier of the service, useful in 'service_in' and 'service_out'\n :param service_in: the name of the service(s) that this service receives data from.\n One can also use 'Service.Frontend' to indicate the connection with the frontend.\n :param service_out: the name of the service(s) that this service sends data to.\n One can also use 'Service.Frontend' to indicate the connection with the frontend.\n :param copy_flow: when set to true, then always copy the current flow and do the modification on top of it then return, otherwise, do in-line modification\n :param kwargs: other keyword-value arguments that the service CLI supports\n :return: a (new) flow object with modification\n \"\"\"\n\n op_flow = copy.deepcopy(self) if copy_flow else self\n\n if service not in Flow._service2parser:\n raise ValueError('service: %s is not supported, should be one of %s' % (service, Flow._service2parser))\n\n if name in op_flow._service_nodes:\n raise FlowTopologyError('name: %s is used in this Flow already!' % name)\n if not name:\n name = '%s%d' % (service, op_flow._service_name_counter[service])\n op_flow._service_name_counter[service] += 1\n if not name.isidentifier():\n raise ValueError('name: %s is invalid, please follow the python variable name conventions' % name)\n\n if service == Service.Frontend:\n if op_flow._frontend:\n raise FlowTopologyError('frontend is already in this Flow')\n op_flow._frontend = name\n\n service_in = op_flow._parse_service_endpoints(op_flow, name, service_in, connect_to_last_service=True)\n service_out = op_flow._parse_service_endpoints(op_flow, name, service_out, connect_to_last_service=False)\n\n args, p_args = op_flow._get_parsed_args(op_flow, Flow._service2parser[service], kwargs)\n\n op_flow._service_nodes[name] = {\n 'service': service,\n 'parsed_args': p_args,\n 'args': args,\n 'incomes': service_in,\n 'outgoings': service_out}\n\n # direct all income services' output to the current service\n for s in service_in:\n op_flow._service_nodes[s]['outgoings'].add(name)\n for s in service_out:\n op_flow._service_nodes[s]['incomes'].add(name)\n\n op_flow._last_add_service = name\n\n # graph is now changed so we need to\n # reset the build level to the lowest\n op_flow._build_level = Flow.BuildLevel.EMPTY\n\n return op_flow\n\n @staticmethod\n def _parse_service_endpoints(op_flow, cur_service_name, service_endpoint, connect_to_last_service=False):\n # parsing service_in\n if isinstance(service_endpoint, str):\n service_endpoint = [service_endpoint]\n elif service_endpoint == Service.Frontend:\n service_endpoint = [op_flow._frontend]\n elif not service_endpoint:\n if op_flow._last_add_service and connect_to_last_service:\n service_endpoint = [op_flow._last_add_service]\n else:\n service_endpoint = []\n if isinstance(service_endpoint, list) or isinstance(service_endpoint, tuple):\n for s in service_endpoint:\n if s == cur_service_name:\n raise FlowTopologyError('the income of a service can not be itself')\n if s not in op_flow._service_nodes:\n raise FlowTopologyError('service_in: %s can not be found in this Flow' % s)\n else:\n raise ValueError('service_in=%s is not parsable' % service_endpoint)\n return set(service_endpoint)\n\n @staticmethod\n def _get_parsed_args(op_flow, service_arg_parser, kwargs):\n kwargs.update(op_flow._common_kwargs)\n args = []\n for k, v in kwargs.items():\n if isinstance(v, bool):\n if v:\n if not k.startswith('no_') and not k.startswith('no-'):\n args.append('--%s' % k)\n else:\n args.append('--%s' % k[3:])\n else:\n if k.startswith('no_') or k.startswith('no-'):\n args.append('--%s' % k)\n else:\n args.append('--no_%s' % k)\n else:\n args.extend(['--%s' % k, str(v)])\n try:\n p_args, unknown_args = service_arg_parser().parse_known_args(args)\n if unknown_args:\n op_flow.logger.warning('not sure what these arguments are: %s' % unknown_args)\n except SystemExit:\n raise ValueError('bad arguments for service \"%s\", '\n 'you may want to double check your args \"%s\"' % (service_arg_parser, args))\n return args, p_args\n\n def _build_graph(self, copy_flow: bool) -> 'Flow':\n op_flow = copy.deepcopy(self) if copy_flow else self\n\n op_flow._service_edges.clear()\n\n if not op_flow._frontend:\n raise FlowImcompleteError('frontend does not exist, you may need to add_frontend()')\n\n if not op_flow._last_add_service or not op_flow._service_nodes:\n raise FlowTopologyError('flow is empty?')\n\n # close the loop\n op_flow._service_nodes[op_flow._frontend]['incomes'].add(op_flow._last_add_service)\n\n # build all edges\n for k, v in op_flow._service_nodes.items():\n for s in v['incomes']:\n op_flow._service_edges['%s-%s' % (s, k)] = ''\n for t in v['outgoings']:\n op_flow._service_edges['%s-%s' % (k, t)] = ''\n\n for k in op_flow._service_edges.keys():\n start_node, end_node = k.split('-')\n edges_with_same_start = [ed for ed in op_flow._service_edges.keys() if ed.startswith(start_node)]\n edges_with_same_end = [ed for ed in op_flow._service_edges.keys() if ed.endswith(end_node)]\n\n s_pargs = op_flow._service_nodes[start_node]['parsed_args']\n e_pargs = op_flow._service_nodes[end_node]['parsed_args']\n\n # Rule\n # if a node has multiple income/outgoing services,\n # then its socket_in/out must be PULL_BIND or PUB_BIND\n # otherwise it should be different than its income\n # i.e. income=BIND => this=CONNECT, income=CONNECT => this = BIND\n #\n # when a socket is BIND, then host must NOT be set, aka default host 0.0.0.0\n # host_in and host_out is only set when corresponding socket is CONNECT\n\n if len(edges_with_same_start) > 1 and len(edges_with_same_end) == 1:\n s_pargs.socket_out = SocketType.PUB_BIND\n s_pargs.host_out = BaseService.default_host\n e_pargs.socket_in = SocketType.SUB_CONNECT\n e_pargs.host_in = start_node\n e_pargs.port_in = s_pargs.port_out\n op_flow._service_edges[k] = 'PUB-sub'\n elif len(edges_with_same_end) > 1 and len(edges_with_same_start) == 1:\n s_pargs.socket_out = SocketType.PUSH_CONNECT\n s_pargs.host_out = end_node\n e_pargs.socket_in = SocketType.PULL_BIND\n e_pargs.host_in = BaseService.default_host\n s_pargs.port_out = e_pargs.port_in\n op_flow._service_edges[k] = 'push-PULL'\n elif len(edges_with_same_start) == 1 and len(edges_with_same_end) == 1:\n # in this case, either side can be BIND\n # we prefer frontend to be always BIND\n # check if either node is frontend\n if start_node == op_flow._frontend:\n s_pargs.socket_out = SocketType.PUSH_BIND\n e_pargs.socket_in = SocketType.PULL_CONNECT\n elif end_node == op_flow._frontend:\n s_pargs.socket_out = SocketType.PUSH_CONNECT\n e_pargs.socket_in = SocketType.PULL_BIND\n else:\n e_pargs.socket_in = s_pargs.socket_out.paired\n\n if s_pargs.socket_out.is_bind:\n s_pargs.host_out = BaseService.default_host\n e_pargs.host_in = start_node\n e_pargs.port_in = s_pargs.port_out\n op_flow._service_edges[k] = 'PUSH-pull'\n elif e_pargs.socket_in.is_bind:\n s_pargs.host_out = end_node\n e_pargs.host_in = BaseService.default_host\n s_pargs.port_out = e_pargs.port_in\n op_flow._service_edges[k] = 'push-PULL'\n else:\n raise FlowTopologyError('edge %s -> %s is ambiguous, at least one socket should be BIND')\n else:\n raise FlowTopologyError('found %d edges start with %s and %d edges end with %s, '\n 'this type of topology is ambiguous and should not exist, '\n 'i can not determine the socket type' % (\n len(edges_with_same_start), start_node, len(edges_with_same_end), end_node))\n\n op_flow._build_level = Flow.BuildLevel.GRAPH\n return op_flow\n\n def build(self, backend: Optional[str] = 'thread', copy_flow: bool = False, *args, **kwargs) -> 'Flow':\n \"\"\"\n Build the current flow and make it ready to use\n\n :param backend: supported 'thread', 'process', 'swarm', 'k8s', 'shell', if None then only build graph only\n :param copy_flow: return the copy of the current flow\n :return: the current flow (by default)\n \"\"\"\n\n op_flow = self._build_graph(copy_flow)\n\n if not backend:\n op_flow.logger.warning('no specified backend, build_level stays at %s, '\n 'and you can not run this flow.' % op_flow._build_level)\n elif backend in {'thread', 'process'}:\n op_flow._service_contexts.clear()\n for v in op_flow._service_nodes.values():\n p_args = v['parsed_args']\n p_args.parallel_backend = backend\n # for thread and process backend which runs locally, host_in and host_out should not be set\n p_args.host_in = BaseService.default_host\n p_args.host_out = BaseService.default_host\n op_flow._service_contexts.append((Flow._service2builder[v['service']], p_args))\n op_flow._build_level = Flow.BuildLevel.RUNTIME\n else:\n raise NotImplementedError('backend=%s is not supported yet' % backend)\n\n return op_flow\n\n def __call__(self, *args, **kwargs):\n return self.build(*args, **kwargs)\n\n def __enter__(self):\n if self._build_level.value < Flow.BuildLevel.RUNTIME.value:\n self.logger.warning(\n 'current build_level=%s, lower than required. '\n 'build the flow now via build() with default parameters' % self._build_level)\n self.build(copy_flow=False)\n self._service_stack = ExitStack()\n for k, v in self._service_contexts:\n self._service_stack.enter_context(k(v))\n\n self.logger.critical('flow is built and ready, current build level is %s' % self._build_level)\n return self\n\n def __exit__(self, exc_type, exc_val, exc_tb):\n self.close()\n\n def close(self):\n if hasattr(self, '_service_stack'):\n self._service_stack.close()\n self._build_level = Flow.BuildLevel.EMPTY\n self.logger.critical(\n 'flow is closed and all resources should be released already, current build level is %s' % self._build_level)\n\n def __getstate__(self):\n d = dict(self.__dict__)\n del d['logger']\n return d\n\n def __setstate__(self, d):\n self.__dict__.update(d)\n self.logger = set_logger(self.__class__.__name__)\n", "sub_path": "gnes/flow/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 23541, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "service.base.BetterEnum", "line_number": 21, "usage_type": "name"}, {"api_name": "functools.wraps", "line_number": 43, "usage_type": "call"}, {"api_name": "cli.parser.set_encoder_parser", "line_number": 92, "usage_type": "name"}, {"api_name": "cli.parser.set_router_parser", "line_number": 93, "usage_type": "name"}, {"api_name": "cli.parser.set_indexer_parser", "line_number": 94, "usage_type": "name"}, {"api_name": "cli.parser.set_frontend_parser", "line_number": 95, "usage_type": "name"}, {"api_name": "cli.parser.set_preprocessor_parser", "line_number": 96, "usage_type": "name"}, {"api_name": "service.base.ServiceManager", "line_number": 99, "usage_type": "call"}, {"api_name": "service.encoder.EncoderService", "line_number": 99, "usage_type": "argument"}, {"api_name": "service.base.ServiceManager", "line_number": 100, "usage_type": "call"}, {"api_name": "service.router.RouterService", "line_number": 100, "usage_type": "argument"}, {"api_name": "service.base.ServiceManager", "line_number": 101, "usage_type": "call"}, {"api_name": "service.indexer.IndexerService", "line_number": 101, "usage_type": "argument"}, {"api_name": "service.base.ServiceManager", "line_number": 102, "usage_type": "call"}, {"api_name": "service.preprocessor.PreprocessorService", "line_number": 102, "usage_type": "argument"}, {"api_name": "service.frontend.FrontendService", "line_number": 103, "usage_type": "name"}, {"api_name": "service.base.BetterEnum", "line_number": 106, "usage_type": "name"}, {"api_name": "helper.set_logger", "line_number": 112, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 113, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 148, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 151, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 189, "usage_type": "call"}, {"api_name": "urllib.request.Request", "line_number": 192, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 194, "usage_type": "call"}, {"api_name": "typing.Iterator", "line_number": 197, "usage_type": "name"}, {"api_name": "typing.Iterator", "line_number": 207, "usage_type": "name"}, {"api_name": "typing.Iterator", "line_number": 217, "usage_type": "name"}, {"api_name": "typing.Iterator", "line_number": 228, "usage_type": "name"}, {"api_name": "os.unsetenv", "line_number": 229, "usage_type": "call"}, {"api_name": "os.unsetenv", "line_number": 230, "usage_type": "call"}, {"api_name": "cli.parser.set_client_cli_parser", "line_number": 231, "usage_type": "argument"}, {"api_name": "client.cli.CLIClient", "line_number": 234, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 264, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 264, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 264, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 265, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 265, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 265, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 282, "usage_type": "call"}, {"api_name": "service.base", "line_number": 284, "usage_type": "name"}, {"api_name": "{'base64': 'base64', 'Request': 'urllib.request.Request', 'urlopen': 'urllib.request.urlopen'}._service2parser", "line_number": 284, "usage_type": "attribute"}, {"api_name": "service.base", "line_number": 285, "usage_type": "name"}, {"api_name": "{'base64': 'base64', 'Request': 'urllib.request.Request', 'urlopen': 'urllib.request.urlopen'}._service2parser", "line_number": 285, "usage_type": "attribute"}, {"api_name": "service.base", "line_number": 290, "usage_type": "name"}, {"api_name": "service.base", "line_number": 291, "usage_type": "name"}, {"api_name": "service.base", "line_number": 295, "usage_type": "name"}, {"api_name": "{'base64': 'base64', 'Request': 'urllib.request.Request', 'urlopen': 'urllib.request.urlopen'}._service2parser", "line_number": 303, "usage_type": "attribute"}, {"api_name": "service.base", "line_number": 303, "usage_type": "name"}, {"api_name": "service.base", "line_number": 306, "usage_type": "name"}, {"api_name": "{'base64': 'base64', 'Request': 'urllib.request.Request', 'urlopen': 'urllib.request.urlopen'}.BuildLevel", "line_number": 322, "usage_type": "attribute"}, {"api_name": "copy.deepcopy", "line_number": 376, "usage_type": "call"}, {"api_name": "service.base.SocketType.PUB_BIND", "line_number": 414, "usage_type": "attribute"}, {"api_name": "service.base.SocketType", "line_number": 414, "usage_type": "name"}, {"api_name": "service.base.BaseService.default_host", "line_number": 415, "usage_type": "attribute"}, {"api_name": "service.base.BaseService", "line_number": 415, "usage_type": "name"}, {"api_name": "service.base.SocketType.SUB_CONNECT", "line_number": 416, "usage_type": "attribute"}, {"api_name": "service.base.SocketType", "line_number": 416, "usage_type": "name"}, {"api_name": "service.base.SocketType.PUSH_CONNECT", "line_number": 421, "usage_type": "attribute"}, {"api_name": "service.base.SocketType", "line_number": 421, "usage_type": "name"}, {"api_name": "service.base.SocketType.PULL_BIND", "line_number": 423, "usage_type": "attribute"}, {"api_name": "service.base.SocketType", "line_number": 423, "usage_type": "name"}, {"api_name": "service.base.BaseService.default_host", "line_number": 424, "usage_type": "attribute"}, {"api_name": "service.base.BaseService", "line_number": 424, "usage_type": "name"}, {"api_name": "service.base.SocketType.PUSH_BIND", "line_number": 432, "usage_type": "attribute"}, {"api_name": "service.base.SocketType", "line_number": 432, "usage_type": "name"}, {"api_name": "service.base.SocketType.PULL_CONNECT", "line_number": 433, "usage_type": "attribute"}, {"api_name": "service.base.SocketType", "line_number": 433, "usage_type": "name"}, {"api_name": "service.base.SocketType.PUSH_CONNECT", "line_number": 435, "usage_type": "attribute"}, {"api_name": "service.base.SocketType", "line_number": 435, "usage_type": "name"}, {"api_name": "service.base.SocketType.PULL_BIND", "line_number": 436, "usage_type": "attribute"}, {"api_name": "service.base.SocketType", "line_number": 436, "usage_type": "name"}, {"api_name": "service.base.BaseService.default_host", "line_number": 441, "usage_type": "attribute"}, {"api_name": "service.base.BaseService", "line_number": 441, "usage_type": "name"}, {"api_name": "service.base.BaseService.default_host", "line_number": 447, "usage_type": "attribute"}, {"api_name": "service.base.BaseService", "line_number": 447, "usage_type": "name"}, {"api_name": "{'base64': 'base64', 'Request': 'urllib.request.Request', 'urlopen': 'urllib.request.urlopen'}.BuildLevel", "line_number": 458, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 461, "usage_type": "name"}, {"api_name": "service.base.BaseService.default_host", "line_number": 481, "usage_type": "attribute"}, {"api_name": "service.base.BaseService", "line_number": 481, "usage_type": "name"}, {"api_name": "service.base.BaseService.default_host", "line_number": 482, "usage_type": "attribute"}, {"api_name": "service.base.BaseService", "line_number": 482, "usage_type": "name"}, {"api_name": "{'base64': 'base64', 'Request': 'urllib.request.Request', 'urlopen': 'urllib.request.urlopen'}._service2builder", "line_number": 483, "usage_type": "attribute"}, {"api_name": "{'base64': 'base64', 'Request': 'urllib.request.Request', 'urlopen': 'urllib.request.urlopen'}.BuildLevel", "line_number": 484, "usage_type": "attribute"}, {"api_name": "{'base64': 'base64', 'Request': 'urllib.request.Request', 'urlopen': 'urllib.request.urlopen'}.BuildLevel", "line_number": 494, "usage_type": "attribute"}, {"api_name": "contextlib.ExitStack", "line_number": 499, "usage_type": "call"}, {"api_name": "{'base64': 'base64', 'Request': 'urllib.request.Request', 'urlopen': 'urllib.request.urlopen'}.BuildLevel", "line_number": 512, "usage_type": "attribute"}, {"api_name": "helper.set_logger", "line_number": 523, "usage_type": "call"}]}
+{"seq_id": "534325607", "text": "import os\nfrom keras.models import Sequential\nfrom keras.layers import Conv2D, MaxPooling2D\nfrom keras.layers import Activation, Dropout, Flatten, Dense\nfrom keras.preprocessing.image import ImageDataGenerator, array_to_img, img_to_array, load_img\nimport keras_metrics as km\nfrom keras import backend as K\n\nK.set_image_dim_ordering('th')\nimport numpy as np\nfrom keras.preprocessing import image\nfrom sklearn.metrics import roc_auc_score, roc_curve, auc, average_precision_score, confusion_matrix\nfrom scipy import interp\nimport matplotlib.pyplot as plt\nfrom itertools import cycle\n\n\ndef get_model():\n model = Sequential()\n model.add(Conv2D(32, (3, 3), input_shape=(3, 100, 300)))\n model.add(Activation('relu'))\n model.add(MaxPooling2D(pool_size=(2, 2)))\n\n model.add(Conv2D(32, (3, 3)))\n model.add(Activation('relu'))\n model.add(MaxPooling2D(pool_size=(2, 2)))\n\n model.add(Conv2D(64, (3, 3)))\n model.add(Activation('relu'))\n model.add(MaxPooling2D(pool_size=(2, 2)))\n\n # the model so far outputs 3D feature maps (height, width, features)\n\n\n model.add(Flatten()) # this converts our 3D feature maps to 1D feature vectors\n model.add(Dense(64))\n model.add(Activation('relu'))\n model.add(Dropout(0.5))\n model.add(Dense(9))\n model.add(Activation('sigmoid'))\n return model\n\n\ndef build_models(weights_file):\n model = get_model()\n model.compile(loss='categorical_crossentropy',\n optimizer='adam',\n metrics=['mean_squared_error', 'accuracy'])\n\n batch_size = 16\n\n # this is the augmentation configuration we will use for training\n train_datagen = ImageDataGenerator(\n rescale=1. / 255,\n shear_range=0.2,\n zoom_range=0.2,\n horizontal_flip=True)\n\n # this is the augmentation configuration we will use for testing:\n # only rescaling\n test_datagen = ImageDataGenerator(rescale=1. / 255)\n\n # this is a generator that will read pictures found in\n # subfolers of 'data/train', and indefinitely generate\n # batches of augmented image data\n train_generator = train_datagen.flow_from_directory(\n '/Users/preethi/Allclass/297/cnn_data/s_pattern_images/', # this is the target directory\n target_size=(100, 300), # all images will be resized to 150x150\n batch_size=batch_size,\n class_mode='categorical') # since we use binary_crossentropy loss, we need binary labels #categorical\n\n model.fit_generator(\n train_generator,\n steps_per_epoch=80,\n epochs=5)\n\n model.save_weights(weights_file) # always save your weights after training or during training\n\n\ndef rebuild_model(weights_file):\n rebuilt_model = get_model()\n rebuilt_model.load_weights(weights_file) # 'first_try.h5')\n return rebuilt_model\n\n\ndef try_roc_sklearn(y_pred, y_true):\n # Compute ROC curve and ROC area for each class\n fpr = dict()\n tpr = dict()\n roc_auc = dict()\n n_classes = 9\n lw = 2\n # Compute macro-average ROC curve and ROC area\n\n\n for i in range(n_classes):\n fpr[i], tpr[i], thresholdd = roc_curve(y_pred[:, i], y_true[:, i])\n # fpr[i], tpr[i], thresholdd = roc_curve(y_pred, y_true)\n '''\n fnr = 1 - tpr[i]\n\n err_threshold = thresholdd[np.nanargmin(np.absolute((fnr[1] - fpr[i])))]\n EER = fpr[i][np.nanargmin(np.absolute((fnr - fpr[i])))]\n print(\"ERR = \")\n print(EER)\n print(err_threshold)\n '''\n roc_auc[i] = auc(fpr[i], tpr[i])\n\n # First aggregate all false positive rates\n all_fpr = np.unique(np.concatenate([fpr[i] for i in range(n_classes)]))\n\n # Then interpolate all ROC curves at this points\n mean_tpr = np.zeros_like(all_fpr)\n for i in range(n_classes):\n mean_tpr += interp(all_fpr, fpr[i], tpr[i])\n\n # Finally average it and compute AUC\n mean_tpr /= n_classes\n\n roc_auc[i] = auc(fpr[i], tpr[i])\n\n # Compute micro-average ROC curve and ROC area\n # Plot all ROC curves\n plt.figure()\n\n colors = ['darkblue', 'darkorange', 'cornflowerblue', 'r', 'b', 'g', 'c', 'y', 'm']\n # users = ['intruder', 'user']\n for i, color in zip(range(n_classes), colors):\n plt.plot(fpr[i], tpr[i], color=color, lw=lw,\n label='ROC curve of user {0} (area = {1:0.2f})'\n ''.format(str(i + 1), roc_auc[i]))\n\n plt.plot([1, 0], [0, 1], 'k--', lw=lw)\n plt.xlim([0.0, 1.0])\n plt.ylim([0.0, 1.05])\n plt.xlabel('False Positive Rate')\n plt.ylabel('True Positive Rate')\n plt.title('ROC curve for Common User Patterns(Acceleration)')\n plt.legend(loc=\"lower right\")\n plt.show()\n\n\ndef try_predict(weights):\n # build_models()\n model = rebuild_model(weights_file=weights)\n # test_image = image.load_img('/Users/preethi/Allclass/297/data_validation/ansu/ansu2.jpg',\n # target_size=(100, 300))\n # test_image = image.img_to_array(test_image)\n # test_image = np.expand_dims(test_image, axis=0)\n # result = model.predict_classes(test_image)\n # print(result)\n\n test_datagen = ImageDataGenerator(rescale=1. / 255)\n\n # this is a generator that will read pictures found in\n # subfolers of 'data/train', and indefinitely generate\n # batches of augmented image data\n test_datagen = test_datagen.flow_from_directory(\n '/Users/preethi/Allclass/297/data_validation_unique/', # '/Users/preethi/Allclass/297/cnn_data/validation/',#\n target_size=(100, 300), # all images will be resized to 150x150\n batch_size=16,\n shuffle=False,\n class_mode='categorical') # since we use binary_crossentropy loss, we need binary labels\n\n predictions = model.predict_generator(test_datagen)\n print(predictions)\n\n # multiclass\n\n y_pred = np.zeros((39, 9), dtype=np.int8)\n\n r_num = 0\n for row in predictions:\n max_idx = np.argmax(row)\n y_pred[r_num][max_idx] = 1\n r_num += 1\n # y_pred = np.amax(predictions, axis=1)\n\n y_true_h = test_datagen.classes\n\n y_true = np.zeros((y_true_h.size, int(y_true_h.max()) + 1), dtype=np.int8)\n y_true[np.arange(y_true_h.size), y_true_h] = 1\n\n '''\n\n # binary\n y_pred = np.zeros(50, dtype=np.int8)\n\n r_num = 0\n for row in predictions:\n max_idx = np.argmax(row)\n if max_idx == 1:\n y_pred[r_num] = 1\n r_num += 1\n # y_pred = np.amax(predictions, axis=1)\n #y_pred = np.amax(predictions, axis=1)\n y_true_h = test_datagen.classes\n print(y_pred)\n print(y_true_h)\n try_roc_sklearn(y_pred, y_true_h)\n '''\n try_roc_sklearn(y_pred, y_true)\n\n\nif __name__ == \"__main__\":\n # build_models('cnn_common_2class3.h5')\n # s_pattern_model = rebuild_model('cnn_common_pattern.h5')\n # build_models('cnn_balanced_2class.h5')\n # try_predict('cnn_balanced_2class.h5')\n try_predict('first_try.h5')", "sub_path": "cnn_unique.py", "file_name": "cnn_unique.py", "file_ext": "py", "file_size_in_byte": 6863, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "keras.backend.set_image_dim_ordering", "line_number": 9, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 9, "usage_type": "name"}, {"api_name": "keras.models.Sequential", "line_number": 19, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 20, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 21, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 22, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 24, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 25, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 26, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 28, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 29, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 30, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 35, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 36, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 37, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 38, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 39, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 40, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 53, "usage_type": "call"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 61, "usage_type": "call"}, {"api_name": "sklearn.metrics.roc_curve", "line_number": 97, "usage_type": "call"}, {"api_name": "sklearn.metrics.auc", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 114, "usage_type": "call"}, {"api_name": "scipy.interp", "line_number": 116, "usage_type": "call"}, {"api_name": "sklearn.metrics.auc", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 125, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 137, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 137, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 139, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 139, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 141, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 141, "usage_type": "name"}, {"api_name": "keras.preprocessing.image.ImageDataGenerator", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 171, "usage_type": "attribute"}, {"api_name": "numpy.argmax", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 182, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 182, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 183, "usage_type": "call"}]}
+{"seq_id": "645898994", "text": "\"\"\"Build vocabulary for SYN corpus\"\"\"\n\nimport argparse\nimport os\n\nfrom tqdm import tqdm\n\nfrom syn.helpers.argparser import common_parser, vocabulary_parser\nfrom syn.helpers.environment import load_environment_variables\nfrom syn.helpers.logging import set_logger\nfrom syn.helpers.nlp.vocabulay import load_tokens, get_vocabulary_from_mongodb, save_vocabulary\nfrom syn.helpers.system import check_same_python_module_already_running\n\nload_environment_variables()\nlog = set_logger()\n\n\ndef get_input_params():\n parser = argparse.ArgumentParser(\n parents=[common_parser, vocabulary_parser],\n description='Filter word embeddings.'\n )\n\n args = parser.parse_args()\n\n return {\n 'corpus': args.corpus,\n 'query_limit': args.query_limit,\n 'vocabulary_name': args.vocabulary_name,\n 'tokens_collection': args.tokens_collection,\n }\n\n\nif __name__ == \"__main__\":\n # Check if there is a running process that contains the name of this module.\n check_same_python_module_already_running(os.path.split(__file__))\n\n # Load the parameters.\n input_params = get_input_params()\n\n # Loads dataset.\n log.info(f\"Loading Dataframe ...\")\n df = load_tokens(\n database_name=input_params['corpus'],\n collection_name=input_params['tokens_collection'],\n query_limit=input_params['query_limit']\n )\n log.info(f\"Dataframe columns: {df.columns}\")\n log.info(f\"Dataframe loaded.\")\n\n # Build vocabulary.\n log.info(f\"Building vocabulary ...\")\n vocab = set()\n for row in tqdm(df['tokens'], total=len(df['tokens']), desc='rows'):\n tmp_set = get_vocabulary_from_mongodb(row)\n vocab.update(tmp_set)\n\n log.info(f\"Vocabulary size: {len(vocab)}\")\n\n # Save vocabulary.\n inserted_documents = save_vocabulary(\n database_name=input_params['corpus'],\n collection_name=input_params['vocabulary_name'],\n vocabulary=vocab\n )\n assert inserted_documents == len(vocab)\n log.info(f\"MODULE EXECUTED.\")\n", "sub_path": "syn/model/build/common/build_vocab.py", "file_name": "build_vocab.py", "file_ext": "py", "file_size_in_byte": 2009, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "syn.helpers.environment.load_environment_variables", "line_number": 14, "usage_type": "call"}, {"api_name": "syn.helpers.logging.set_logger", "line_number": 15, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 19, "usage_type": "call"}, {"api_name": "syn.helpers.argparser.common_parser", "line_number": 20, "usage_type": "name"}, {"api_name": "syn.helpers.argparser.vocabulary_parser", "line_number": 20, "usage_type": "name"}, {"api_name": "syn.helpers.system.check_same_python_module_already_running", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path", "line_number": 36, "usage_type": "attribute"}, {"api_name": "syn.helpers.nlp.vocabulay.load_tokens", "line_number": 43, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 54, "usage_type": "call"}, {"api_name": "syn.helpers.nlp.vocabulay.get_vocabulary_from_mongodb", "line_number": 55, "usage_type": "call"}, {"api_name": "syn.helpers.nlp.vocabulay.save_vocabulary", "line_number": 61, "usage_type": "call"}]}
+{"seq_id": "537975479", "text": "import subprocess\nimport psutil\nimport re\nfrom Helper import Config, TapConfig, Level\nfrom typing import Dict, Optional, Callable\nfrom time import sleep\n\n\nclass Tap:\n initialized = False\n tapConfig: TapConfig = None\n closingRegex = None\n\n @classmethod\n def Initialize(cls):\n cls.tapConfig = Config().Tap\n cls.closingRegex = re.compile(r'.*Resource \".*\" closed.*')\n cls.initialized = True\n\n def __init__(self, tapPlan: str, externals: Dict[str, str], logger: Callable):\n if not self.initialized:\n self.Initialize()\n\n self.tapPlan = tapPlan\n self.externals = externals\n self.args = self.getArgs(tapPlan, externals)\n self.logger = logger\n self.closedInstruments = 0\n self.closeStarted = False\n self.process: Optional[psutil.Process] = None\n\n @staticmethod\n def getArgs(tapPlan: str, externals: Dict[str, str]):\n if Tap.tapConfig.OpenTap:\n args = [Tap.tapConfig.Path, 'run', '-v']\n else:\n args = [Tap.tapConfig.Path, '-v']\n\n for key, value in externals.items():\n args.extend(['-e', f'{key}={value}'])\n\n args.append(tapPlan)\n\n return args\n\n def notify(self):\n self.logger(Level.INFO, f'Executing TapPlan: {self.tapPlan}')\n if len(self.externals) != 0:\n for key, value in self.externals.items():\n self.logger(Level.INFO, f' {key}={value}')\n\n def Execute(self) -> int:\n self.closedInstruments = 0\n self.closeStarted = False\n\n self.notify()\n process = subprocess.Popen(self.args, stdout=subprocess.PIPE,\n stderr=subprocess.STDOUT, cwd=Tap.tapConfig.Folder)\n sleep(0.5) # Give some time to ensure that psutil finds the process\n self.process = psutil.Process(process.pid)\n self.tap_stdout(process)\n\n exitCode = process.wait()\n\n return exitCode\n\n def tap_stdout(self, process: subprocess.Popen):\n _levels = [('Debug', Level.DEBUG), ('Information', Level.INFO),\n ('Warning', Level.WARNING), ('Error', Level.ERROR)]\n\n def _inferLevel(l: str) -> Level:\n for level in _levels:\n string, res = level\n if re.match(f'.*:\\s*{string}\\s*:.*', l): return res\n return Level.INFO\n\n pipe = process.stdout\n\n for line in iter(pipe.readline, b''):\n try: line = line.decode('utf-8').rstrip()\n except Exception as e: line = f\"DECODING EXCEPTION: {e}\"\n\n level = _inferLevel(line)\n self.logger(level, f\"[TAP]{line}\")\n\n if self.tapConfig.EnsureClosed:\n if 'Unable to continue. Now exiting TAP CLI' in line:\n self.closeStarted = True\n Tap.ensureTapClosed(self.process, self.logger, self.tapConfig.EnsureAdbClosed)\n\n if Tap.closingRegex.match(line):\n self.closedInstruments += 1\n if self.closedInstruments >= 3 and not self.closeStarted:\n self.closeStarted = True\n Tap.ensureTapClosed(self.process, self.logger, self.tapConfig.EnsureAdbClosed)\n\n @staticmethod\n def ensureTapClosed(tapProcess: psutil.Process, logger, closeAdb):\n logger(Level.INFO, 'Ensuring that TAP is correctly closed (in 15 seconds).')\n sleep(15)\n\n if tapProcess.is_running():\n logger(Level.WARNING, f'TAP still running, stopping child processes '\n f'({len(tapProcess.children(recursive=True))})...')\n Tap.endProcessTree(tapProcess)\n logger(Level.INFO, 'Process tree closed')\n else:\n logger(Level.INFO, 'TAP closed correctly')\n\n if closeAdb:\n for p in psutil.process_iter():\n if p.name() == 'adb.exe':\n logger(Level.WARNING, f\"Closing rogue adb process with PID: {p.pid}\")\n p.kill()\n\n @classmethod\n def endProcessTree(cls, process: psutil.Process):\n def safeTerminate(p: psutil.Process):\n try: p.terminate()\n except psutil.NoSuchProcess: pass\n\n for child in process.children(recursive=True): # type: psutil.Process\n safeTerminate(child)\n safeTerminate(process)\n", "sub_path": "Helper/tap_executor.py", "file_name": "tap_executor.py", "file_ext": "py", "file_size_in_byte": 4357, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "Helper.TapConfig", "line_number": 11, "usage_type": "name"}, {"api_name": "Helper.Config", "line_number": 16, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 17, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 20, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 20, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 30, "usage_type": "name"}, {"api_name": "psutil.Process", "line_number": 30, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 33, "usage_type": "name"}, {"api_name": "Helper.Level.INFO", "line_number": 47, "usage_type": "attribute"}, {"api_name": "Helper.Level", "line_number": 47, "usage_type": "name"}, {"api_name": "Helper.Level.INFO", "line_number": 50, "usage_type": "attribute"}, {"api_name": "Helper.Level", "line_number": 50, "usage_type": "name"}, {"api_name": "subprocess.Popen", "line_number": 57, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 57, "usage_type": "attribute"}, {"api_name": "subprocess.STDOUT", "line_number": 58, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 59, "usage_type": "call"}, {"api_name": "psutil.Process", "line_number": 60, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 67, "usage_type": "attribute"}, {"api_name": "Helper.Level.DEBUG", "line_number": 68, "usage_type": "attribute"}, {"api_name": "Helper.Level", "line_number": 68, "usage_type": "name"}, {"api_name": "Helper.Level.INFO", "line_number": 68, "usage_type": "attribute"}, {"api_name": "Helper.Level.WARNING", "line_number": 69, "usage_type": "attribute"}, {"api_name": "Helper.Level", "line_number": 69, "usage_type": "name"}, {"api_name": "Helper.Level.ERROR", "line_number": 69, "usage_type": "attribute"}, {"api_name": "re.match", "line_number": 74, "usage_type": "call"}, {"api_name": "Helper.Level.INFO", "line_number": 75, "usage_type": "attribute"}, {"api_name": "Helper.Level", "line_number": 75, "usage_type": "name"}, {"api_name": "Helper.Level", "line_number": 71, "usage_type": "name"}, {"api_name": "psutil.Process", "line_number": 98, "usage_type": "attribute"}, {"api_name": "Helper.Level.INFO", "line_number": 99, "usage_type": "attribute"}, {"api_name": "Helper.Level", "line_number": 99, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 100, "usage_type": "call"}, {"api_name": "Helper.Level.WARNING", "line_number": 103, "usage_type": "attribute"}, {"api_name": "Helper.Level", "line_number": 103, "usage_type": "name"}, {"api_name": "Helper.Level.INFO", "line_number": 106, "usage_type": "attribute"}, {"api_name": "Helper.Level", "line_number": 106, "usage_type": "name"}, {"api_name": "Helper.Level.INFO", "line_number": 108, "usage_type": "attribute"}, {"api_name": "Helper.Level", "line_number": 108, "usage_type": "name"}, {"api_name": "psutil.process_iter", "line_number": 111, "usage_type": "call"}, {"api_name": "Helper.Level.WARNING", "line_number": 113, "usage_type": "attribute"}, {"api_name": "Helper.Level", "line_number": 113, "usage_type": "name"}, {"api_name": "psutil.Process", "line_number": 117, "usage_type": "attribute"}, {"api_name": "psutil.Process", "line_number": 118, "usage_type": "attribute"}, {"api_name": "psutil.NoSuchProcess", "line_number": 120, "usage_type": "attribute"}]}
+{"seq_id": "554021529", "text": "import os\r\nimport sys\r\nfrom pathlib import Path\r\n\r\n\r\n# Function to rename multiple files\r\ndef main():\r\n direc = Path(sys.argv[1])\r\n file_extsn = sys.argv[2]\r\n for filename in os.listdir(direc):\r\n dst = Path(filename).stem + file_extsn\r\n src = os.path.join(direc, filename)\r\n dst = os.path.join(direc, dst)\r\n if not os.path.exists(dst):\r\n # rename() function will\r\n # rename all the files\r\n os.rename(src, dst)\r\n else:\r\n continue\r\n\r\n\r\n# Driver Code\r\nif __name__ == '__main__':\r\n # Calling main() function\r\n main()\r\n", "sub_path": "file-extension-changer.py", "file_name": "file-extension-changer.py", "file_ext": "py", "file_size_in_byte": 608, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pathlib.Path", "line_number": 8, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 8, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 10, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.rename", "line_number": 17, "usage_type": "call"}]}
+{"seq_id": "8368323", "text": "# Curso em vídeo Python 3 - Exercicio 101\n# Funçoes para votação\n\n\ndef voto(nasc):\n from datetime import date\n now = date.today()\n\n idade = now.year - nasc\n if 0 <= idade < 16:\n return 'Idade de {} anos. NÃO VOTA.'.format(idade)\n elif 16 <= idade < 18 or idade >=60:\n return 'Idade de {} anos. VOTO OPCIONAL.'.format(idade)\n elif 18 <= idade < 60:\n return 'Idade de {} anos. VOTO OBRIGATÓRIO.'.format(idade)\n else:\n return 'Ano de nascimento inserio não válido.'\n\n\nnasc = int(input('Ano de nascimento: '))\nprint(voto(nasc))\n", "sub_path": "Mundo 3/Exercicios/ex101.py", "file_name": "ex101.py", "file_ext": "py", "file_size_in_byte": 581, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "datetime.date.today", "line_number": 7, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 7, "usage_type": "name"}]}
+{"seq_id": "312272549", "text": "# encoding:utf-8\nimport cv2\nimport os\n\n\ndef isfile(src_path):\n if os.path.isfile(src_path):\n file_path = os.path.dirname(src_path) # 去掉文件名,返回所在目录\n\n cat_video(file_path)\n return True\n else:\n infile(src_path)\n return False\n\n\ndef infile(src_path):\n for item in os.listdir(src_path):\n file_path = os.path.join(src_path, item)\n paduan = isfile(file_path)\n if paduan:\n break\n\n\ndef cat_video(file_path):\n for videofile in os.listdir(file_path):\n print(videofile)\n imagefile_name = videofile.split(\".\")[0]\n imagefile_path = os.path.join(video_dst_path, imagefile_name)\n if not os.path.exists(imagefile_path):\n os.mkdir(imagefile_path)\n videofile_path = os.path.join(file_path, videofile)\n videoName = videofile_path.split(\"/\")[-1].split(\".\")[0]\n image_name = \"\"\n if nameLength != 0 or image_name != \"\":\n image_name = str(videoName)[0:nameLength]\n else:\n image_name = str(videoName)\n cap = cv2.VideoCapture(videofile_path)\n if cap.isOpened():\n rval, frame = cap.read()\n else:\n rval = False\n count = 1\n while rval:\n rval, frame = cap.read()\n imageName=f'img_{count:05d}.jpg'\n if count % stop == 0:\n try:\n cv2.imwrite(os.path.join(imagefile_path, imageName), frame)\n except:\n continue\n count += 1\n print(count)\n\n cap.release\n\n\nif __name__ == \"__main__\":\n video_path = \"/home/lishuang/Disk/dukto/异常行为采集\" # 视频路径\n video_dst_path = \"/home/lishuang/Disk/dukto/异常行为采集图片\" # 图片保存路径\n if not os.path.exists(video_dst_path):\n os.mkdir(video_dst_path)\n nameLength = 0 # 图片名长度 如果长度为0则使用视频原名\n stop = 1 # 视频截取帧数间隔\n isfile(video_path)\n", "sub_path": "tools/customdata/catVideoToImage.py", "file_name": "catVideoToImage.py", "file_ext": "py", "file_size_in_byte": 2023, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.path.isfile", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 39, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path", "line_number": 62, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 63, "usage_type": "call"}]}
+{"seq_id": "378995266", "text": "import wx\nimport numpy as np\nfrom .matrix import MainPanel\nfrom .control import ControlPanel\n\n\nclass Panel(wx.Panel):\n\n\n\tdef __init__(self, parent, appConf, widgetConf):\n\t\tsuper().__init__(parent, size=(475, 300))\n\n\t\tself.Bind(wx.EVT_PAINT, self._OnPaint)\n\t\tself.Bind(wx.EVT_ERASE_BACKGROUND, self._OnEraseBackground)\n\t\tself.Bind(wx.EVT_BUTTON, self.OnButton)\n\n\n\t\tself.config = appConf\n\t\tself.widgetConf = widgetConf\n\n\t\tself.Display()\n\n\n\tdef Display(self):\n\n\t\tsizer = wx.BoxSizer(wx.VERTICAL)\n\t\tself.matrixPanel = MainPanel(self, self.config, self.widgetConf)\n\t\tself.controlPanel = ControlPanel(self, self.config, self.widgetConf, \n\t\t\t\t\t\t\t\t\t\t self.matrixPanel.matrix_a, \n\t\t\t\t\t\t\t\t\t\t self.matrixPanel.matrix_b, \n\t\t\t\t\t\t\t\t\t\t self.matrixPanel.matrix_c)\n\n\t\tsizer.Add(self.matrixPanel, 1, flag=wx.EXPAND)\n\t\tsizer.Add(self.controlPanel, 0, flag=wx.EXPAND)\n\n\t\tself.SetSizer(sizer)\n\t\tself.Layout()\n\n\tdef OnButton(self, e):\n\t\tname = e.GetEventObject().label\n\t\tif name == 'Calculate':\n\t\t\tMultiplyMatrices(self.matrixPanel.matrix_a, \n\t\t\t\t\t\t\t self.matrixPanel.matrix_b, \n\t\t\t\t\t\t\t self.matrixPanel.matrix_c)\n\n\n\n\tdef Draw(self, dc):\n\t\tdc.SetBackground(wx.Brush(self.config.get_color('color1', 'widget')))\n\t\tdc.Clear()\n\n\n\tdef _OnEraseBackground(self, e):\n\t\tpass\n\n\n\tdef _OnPaint(self, e):\n\t\tdc = wx.BufferedPaintDC(self)\n\t\tself.Draw(dc)\n\n\ndef MultiplyMatrices(matrix_a, matrix_b, out):\n\tA = np.array(matrix_a.ReadValues(), dtype=float)\n\tB = np.array(matrix_b.ReadValues(), dtype=float)\n\tC = A.dot(B)\n\tc_rows, c_cols = C.shape\n\tout.DisplayGrid(c_rows, c_cols)\n\tout.CalculateSize()\n\tout.WriteValues(C)", "sub_path": "scienv/widgets/matrices/panel.py", "file_name": "panel.py", "file_ext": "py", "file_size_in_byte": 1580, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "wx.Panel", "line_number": 7, "usage_type": "attribute"}, {"api_name": "wx.EVT_PAINT", "line_number": 13, "usage_type": "attribute"}, {"api_name": "wx.EVT_ERASE_BACKGROUND", "line_number": 14, "usage_type": "attribute"}, {"api_name": "wx.EVT_BUTTON", "line_number": 15, "usage_type": "attribute"}, {"api_name": "wx.BoxSizer", "line_number": 26, "usage_type": "call"}, {"api_name": "wx.VERTICAL", "line_number": 26, "usage_type": "attribute"}, {"api_name": "matrix.MainPanel", "line_number": 27, "usage_type": "call"}, {"api_name": "control.ControlPanel", "line_number": 28, "usage_type": "call"}, {"api_name": "wx.EXPAND", "line_number": 33, "usage_type": "attribute"}, {"api_name": "wx.EXPAND", "line_number": 34, "usage_type": "attribute"}, {"api_name": "wx.Brush", "line_number": 49, "usage_type": "call"}, {"api_name": "wx.BufferedPaintDC", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 64, "usage_type": "call"}]}
+{"seq_id": "110992852", "text": "#!/usr/bin/python3\n# coding=utf-8\n#\n# bme280.py\n# Read data from a digital pressure sensor.\n#\n# Official datasheet available from :\n# https://www.bosch-sensortec.com/bst/products/all_products/bme280\n#---------------------------------------\nfrom __future__ import print_function\nfrom mq import *\nfrom ky import *\nfrom mics import *\nimport os\nimport math\nimport smbus\nimport time\nimport config\nimport serial, struct, sys, json\nfrom ctypes import c_short\nfrom ctypes import c_byte\nfrom ctypes import c_ubyte\n\nDEBUG = 0\nCMD_MODE = 2\nCMD_QUERY_DATA = 4\nCMD_DEVICE_ID = 5\nCMD_SLEEP = 6\nCMD_FIRMWARE = 7\nCMD_WORKING_PERIOD = 8\nMODE_ACTIVE = 0\nMODE_QUERY = 1\nDEVICE = 0x76 # Default device I2C address\nlight_channel = 1\nNH3_channel = 2\nNO2_channel = 3\n\nbus = smbus.SMBus(1) # Rev 2 Pi, Pi 2 & Pi 3 uses bus 1\n # Rev 1 Pi uses bus 0\n\nPORT = '/dev/ttyUSB0'\n\nUNPACK_PAT = ' 127:\n result -= 256\n return result\n\ndef getUChar(data,index):\n # return one byte from data as an unsigned char\n result = data[index] & 0xFF\n return result\n\ndef readBME280ID(addr=DEVICE):\n # Chip ID Register Address\n REG_ID = 0xD0\n (chip_id, chip_version) = bus.read_i2c_block_data(addr, REG_ID, 2)\n return (chip_id, chip_version)\n\ndef readBME280All(addr=DEVICE):\n # Register Addresses\n REG_DATA = 0xF7\n REG_CONTROL = 0xF4\n REG_CONFIG = 0xF5\n\n REG_CONTROL_HUM = 0xF2\n REG_HUM_MSB = 0xFD\n REG_HUM_LSB = 0xFE\n\n # Oversample setting - page 27\n OVERSAMPLE_TEMP = 2\n OVERSAMPLE_PRES = 2\n MODE = 1\n\n # Oversample setting for humidity register - page 26\n OVERSAMPLE_HUM = 2\n bus.write_byte_data(addr, REG_CONTROL_HUM, OVERSAMPLE_HUM)\n\n control = OVERSAMPLE_TEMP<<5 | OVERSAMPLE_PRES<<2 | MODE\n bus.write_byte_data(addr, REG_CONTROL, control)\n\n # Read blocks of calibration data from EEPROM\n # See Page 22 data sheet\n cal1 = bus.read_i2c_block_data(addr, 0x88, 24)\n cal2 = bus.read_i2c_block_data(addr, 0xA1, 1)\n cal3 = bus.read_i2c_block_data(addr, 0xE1, 7)\n\n # Convert byte data to word values\n dig_T1 = getUShort(cal1, 0)\n dig_T2 = getShort(cal1, 2)\n dig_T3 = getShort(cal1, 4)\n\n dig_P1 = getUShort(cal1, 6)\n dig_P2 = getShort(cal1, 8)\n dig_P3 = getShort(cal1, 10)\n dig_P4 = getShort(cal1, 12)\n dig_P5 = getShort(cal1, 14)\n dig_P6 = getShort(cal1, 16)\n dig_P7 = getShort(cal1, 18)\n dig_P8 = getShort(cal1, 20)\n dig_P9 = getShort(cal1, 22)\n\n dig_H1 = getUChar(cal2, 0)\n dig_H2 = getShort(cal3, 0)\n dig_H3 = getUChar(cal3, 2)\n\n dig_H4 = getChar(cal3, 3)\n dig_H4 = (dig_H4 << 24) >> 20\n dig_H4 = dig_H4 | (getChar(cal3, 4) & 0x0F)\n\n dig_H5 = getChar(cal3, 5)\n dig_H5 = (dig_H5 << 24) >> 20\n dig_H5 = dig_H5 | (getUChar(cal3, 4) >> 4 & 0x0F)\n\n dig_H6 = getChar(cal3, 6)\n\n # Wait in ms (Datasheet Appendix B: Measurement time and current calculation)\n wait_time = 1.25 + (2.3 * OVERSAMPLE_TEMP) + ((2.3 * OVERSAMPLE_PRES) + 0.575) + ((2.3 * OVERSAMPLE_HUM)+0.575)\n time.sleep(wait_time/1000) # Wait the required time \n\n # Read temperature/pressure/humidity\n data = bus.read_i2c_block_data(addr, REG_DATA, 8)\n pres_raw = (data[0] << 12) | (data[1] << 4) | (data[2] >> 4)\n temp_raw = (data[3] << 12) | (data[4] << 4) | (data[5] >> 4)\n hum_raw = (data[6] << 8) | data[7]\n\n #Refine temperature\n var1 = ((((temp_raw>>3)-(dig_T1<<1)))*(dig_T2)) >> 11\n var2 = (((((temp_raw>>4) - (dig_T1)) * ((temp_raw>>4) - (dig_T1))) >> 12) * (dig_T3)) >> 14\n t_fine = var1+var2\n temperature = float(((t_fine * 5) + 128) >> 8);\n\n # Refine pressure and adjust for temperature\n var1 = t_fine / 2.0 - 64000.0\n var2 = var1 * var1 * dig_P6 / 32768.0\n var2 = var2 + var1 * dig_P5 * 2.0\n var2 = var2 / 4.0 + dig_P4 * 65536.0\n var1 = (dig_P3 * var1 * var1 / 524288.0 + dig_P2 * var1) / 524288.0\n var1 = (1.0 + var1 / 32768.0) * dig_P1\n if var1 == 0:\n pressure=0\n else:\n pressure = 1048576.0 - pres_raw\n pressure = ((pressure - var2 / 4096.0) * 6250.0) / var1\n var1 = dig_P9 * pressure * pressure / 2147483648.0\n var2 = pressure * dig_P8 / 32768.0\n pressure = pressure + (var1 + var2 + dig_P7) / 16.0\n\n # Refine humidity\n humidity = t_fine - 76800.0\n humidity = (hum_raw - (dig_H4 * 64.0 + dig_H5 / 16384.0 * humidity)) * (dig_H2 / 65536.0 * (1.0 + dig_H6 / 67108864.0 * humidity * (1.0 + dig_H3 / 67108864.0 * humidity)))\n humidity = humidity * (1.0 - dig_H1 * humidity / 524288.0)\n if humidity > 100:\n humidity = 100\n elif humidity < 0:\n humidity = 0\n\n return temperature/100.0,pressure/100.0,humidity\n\ndef dump(d, prefix=''):\n print(prefix + ' '.join(x.encode('hex') for x in d))\n\ndef construct_command(cmd, data=[]):\n assert len(data) <= 12\n data += [0,]*(12-len(data))\n checksum = (sum(data)+cmd-2)%256\n ret = \"\\xaa\\xb4\" + chr(cmd)\n ret += ''.join(chr(x) for x in data)\n ret += \"\\xff\\xff\" + chr(checksum) + \"\\xab\"\n\n if DEBUG:\n dump(ret, '> ')\n return ret\n\ndef process_data(d):\n r = struct.unpack(' 0:\n p = mp.Process(target=run_by_gpu_2, args=(sub_task_df,gpu))\n jobs.append(p)\n p.start()\n\n for proc in jobs:\n proc.join()\n\n t_all_2 = time.time()\n\n print(\"Everything Done:\", (t_all_2 - t_all)/3600, \"hours\")\n print(\"Setup(s) completed:\", task_df[\"Code\"])\n print(\"Number of json:\", len(task_df))\n\nif __name__ == \"__main__\":\n os.environ['TF_CPP_MIN_LOG_LEVEL'] = '3'\n\n optparser = optparse.OptionParser()\n optparser.add_option(\"-s\", \"--sheetname\", default=\"Z_5_DLTask\", help=\"sheet\")\n optparser.add_option(\"-g\", \"--maxgpu\", default=\"2\", help=\"max_gpu\")\n optparser.add_option(\"-p\", \"--pcname\", default=\"0\", help=\"pcname\")\n opts = optparser.parse_args()[0]\n\n sheet_name = str(opts.sheetname)\n max_gpu = int(opts.maxgpu)\n pc_name = int(opts.pcname)\n\n scope = ['https://spreadsheets.google.com/feeds',\n 'https://www.googleapis.com/auth/drive']\n\n credentials = ServiceAccountCredentials.from_json_keyfile_name('./Credential/DeepLearningAlphaC-666170c72205.json', scope)\n gc = gspread.authorize(credentials)\n spreadsheet = gc.open(\"TASK\")\n\n # worksheet_list = spreadsheet.worksheets()\n worksheet = spreadsheet.worksheet(sheet_name)\n\n column_list = worksheet.range('A1:R1')\n task_df = pd.DataFrame(worksheet.get_all_records(), columns=[cell.value for cell in column_list])\n task_df[\"row\"] = np.arange(len(task_df)) + 2\n task_df = task_df[task_df[\"Training\"] == \"\"]\n\n if pc_name == 3:\n task_df = task_df[task_df[\"PC\"] == \"THR-WS\"]\n elif pc_name == 4:\n task_df = task_df[task_df[\"PC\"] == \"FOU-WS\"]\n elif pc_name == 5:\n task_df = task_df[task_df[\"PC\"] == \"FIV-WS\"]\n\n if len(task_df) == 0:\n print(sheetname, \": Nothing to run!\")\n else:\n main(task_df, max_gpu)\n", "sub_path": "run_pool_gpu_gspread.py", "file_name": "run_pool_gpu_gspread.py", "file_ext": "py", "file_size_in_byte": 2371, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "time.time", "line_number": 18, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 25, "usage_type": "call"}, {"api_name": "time.time", "line_number": 32, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 39, "usage_type": "attribute"}, {"api_name": "oauth2client.service_account.ServiceAccountCredentials.from_json_keyfile_name", "line_number": 54, "usage_type": "call"}, {"api_name": "oauth2client.service_account.ServiceAccountCredentials", "line_number": 54, "usage_type": "name"}, {"api_name": "gspread.authorize", "line_number": 55, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 63, "usage_type": "call"}]}
+{"seq_id": "161959090", "text": "# coding=utf-8\n# Copyright 2020 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# https://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom functools import update_wrapper, reduce, partial\nimport inspect\nimport operator as op\n\nfrom . import core\nfrom . import linear_util as lu\nfrom .tree_util import tree_flatten, tree_unflatten, tree_map, tree_multimap\nfrom .util import safe_zip, safe_map, unzip2, split_list\nfrom .api_util import flatten_fun_nokwargs, argnums_partial, wrap_hashably\nfrom .abstract_arrays import raise_to_shaped\nfrom .ad_util import Zero, stop_gradient_p\nfrom .interpreters import partial_eval as pe\nfrom .interpreters import ad\nfrom .interpreters import batching\nfrom .interpreters import xla\nfrom .interpreters.batching import not_mapped\nfrom .config import config\n\nmap = safe_map\nzip = safe_zip\n\n\n### util\n\ndef _resolve_kwargs(fun, args, kwargs):\n ba = inspect.signature(fun).bind(*args, **kwargs)\n ba.apply_defaults()\n if ba.kwargs:\n raise TypeError(\"keyword arguments could not be resolved to positions\")\n else:\n return ba.args\n\ndef _add_args(f, extra_args, left):\n return _add_args_(f, tuple(map(wrap_hashably, extra_args)), left)\n\n@lu.transformation\ndef _add_args_(extra_args, left, *args, **kwargs):\n extra_args = tuple([arg.val for arg in extra_args])\n args = (extra_args + args) if left else (args + extra_args)\n yield (yield args, kwargs)\n\ndef _memoize(thunk):\n cell = []\n saved_state = core.thread_local_state.trace_state.copy()\n def memoized():\n if not cell:\n prev_state = core.thread_local_state.trace_state\n core.thread_local_state.trace_state = saved_state\n try:\n cell.append(thunk())\n finally:\n core.thread_local_state.trace_state = prev_state\n return cell[0]\n return memoized\n\ndef _initial_style_jaxpr(fun, in_avals):\n in_pvals = [pe.PartialVal.unknown(aval) for aval in in_avals]\n jaxpr, out_pvals, consts = pe.trace_to_jaxpr(fun, in_pvals, instantiate=True,\n bottom=True, stage_out=False)\n assert not any(isinstance(c, core.Tracer) for c in consts)\n out_avals = map(raise_to_shaped, unzip2(out_pvals)[0])\n typed_jaxpr = core.TypedJaxpr(jaxpr, consts, in_avals, out_avals)\n return typed_jaxpr\n\ndef _initial_style_staging() -> bool:\n if config.omnistaging_enabled:\n return core.thread_local_state.trace_state.trace_stack.dynamic.level != 0 # type: ignore\n else:\n return core.thread_local_state.trace_state.initial_style\n\ndef _sum_tangents(_, x, *xs):\n return reduce(ad.add_tangents, xs, x)\n\ndef _zeros_like_pytree(x):\n return tree_map(Zero.from_value, x)\n\n@partial(partial, tree_map)\ndef _stop_gradient(x):\n if isinstance(x, core.Tracer):\n return stop_gradient_p.bind(x)\n else:\n return x\n\n\n### JVPs\n\nclass custom_jvp:\n \"\"\"Set up a JAX-transformable function for a custom JVP rule definition.\n\n This class is meant to be used as a function decorator. Instances are\n callables that behave similarly to the underlying function to which the\n decorator was applied, except when a differentiation transformation (like\n :py:func:`jax.jvp` or :py:func:`jax.grad`) is applied, in which case a custom user-supplied\n JVP rule function is used instead of tracing into and performing automatic\n differentiation of the underlying function's implementation. There is a single\n instance method, ``defjvp``, which defines the custom JVP rule.\n\n For example::\n\n import jax.numpy as jnp\n\n @jax.custom_jvp\n def f(x, y):\n return jnp.sin(x) * y\n\n @f.defjvp\n def f_jvp(primals, tangents):\n x, y = primals\n x_dot, y_dot = tangents\n primal_out = f(x, y)\n tangent_out = jnp.cos(x) * x_dot * y + jnp.sin(x) * y_dot\n return primal_out, tangent_out\n\n For a more detailed introduction, see the tutorial_.\n\n .. _tutorial: https://jax.readthedocs.io/en/latest/notebooks/Custom_derivative_rules_for_Python_code.html\n \"\"\"\n\n def __init__(self, fun, nondiff_argnums=()):\n self.fun = fun\n self.nondiff_argnums = nondiff_argnums\n self.jvp = None\n update_wrapper(self, fun)\n\n def defjvp(self, jvp):\n \"\"\"Define a custom JVP rule for the function represented by this instance.\n\n Args:\n jvp: a Python callable representing the custom JVP rule. When there are no\n ``nondiff_argnums``, the ``jvp`` function should accept two arguments,\n where the first is a tuple of primal inputs and the second is a tuple of\n tangent inputs. The lengths of both tuples is equal to the number of\n parameters of the ``custom_jvp`` function. The ``jvp`` function should\n produce as output a pair where the first element is the primal output\n and the second element is the tangent output. Elements of the input and\n output tuples may be arrays or any nested tuples/lists/dicts thereof.\n\n Returns:\n None.\n\n Example::\n\n import jax.numpy as jnp\n\n @jax.custom_jvp\n def f(x, y):\n return jnp.sin(x) * y\n\n @f.defjvp\n def f_jvp(primals, tangents):\n x, y = primals\n x_dot, y_dot = tangents\n primal_out = f(x, y)\n tangent_out = jnp.cos(x) * x_dot * y + jnp.sin(x) * y_dot\n return primal_out, tangent_out\n \"\"\"\n self.jvp = jvp\n\n def defjvps(self, *jvps):\n \"\"\"Convenience wrapper for defining JVPs for each argument separately.\n\n This convenience wrapper cannot be used together with ``nondiff_argnums``.\n\n Args:\n *jvps: a sequence of functions, one for each positional argument of the\n ``custom_jvp`` function. Each function takes as arguments the tangent\n value for the corresponding primal input, the primal output, and the\n primal inputs. See the example below.\n\n Returns:\n None.\n\n Example::\n\n @jax.custom_jvp\n def f(x, y):\n return jnp.sin(x) * y\n\n f.defjvps(lambda x_dot, primal_out, x, y: jnp.cos(x) * x_dot * y,\n lambda y_dot, primal_out, x, y: jnp.sin(x) * y_dot)\n \"\"\"\n if self.nondiff_argnums:\n raise TypeError(\"Can't use ``defjvps`` with ``nondiff_argnums``.\")\n\n def jvp(primals, tangents):\n primal_out = self(*primals)\n zeros = _zeros_like_pytree(primal_out)\n all_tangents_out = [jvp(t, primal_out, *primals) if jvp else zeros\n for t, jvp in zip(tangents, jvps)]\n tangent_out = tree_multimap(_sum_tangents, primal_out, *all_tangents_out)\n return primal_out, tangent_out\n\n self.defjvp(jvp)\n\n def __call__(self, *args, **kwargs):\n if not self.jvp:\n msg = \"No JVP defined for custom_jvp function {} using defjvp.\"\n raise AttributeError(msg.format(self.__name__))\n args = _resolve_kwargs(self.fun, args, kwargs)\n if self.nondiff_argnums:\n is_nondiff = [False] * len(args)\n for i in self.nondiff_argnums: is_nondiff[i] = True\n args = [_stop_gradient(x) if b else x for b, x in zip(is_nondiff, args)]\n dyn_argnums = [i for i, b in enumerate(is_nondiff) if not b]\n f_, dyn_args = argnums_partial(lu.wrap_init(self.fun), dyn_argnums, args)\n static_args = [args[i] for i in self.nondiff_argnums]\n jvp = _add_args(lu.wrap_init(self.jvp), static_args, left=True)\n else:\n f_, dyn_args = lu.wrap_init(self.fun), args\n jvp = lu.wrap_init(self.jvp)\n args_flat, in_tree = tree_flatten(dyn_args)\n flat_fun, out_tree1 = flatten_fun_nokwargs(f_, in_tree)\n flat_jvp, out_tree2 = _flatten_jvp(jvp, in_tree)\n if _initial_style_staging():\n out_flat = custom_jvp_call_jaxpr(flat_fun, flat_jvp, *args_flat)\n out_tree = out_tree1()\n else:\n out_flat = custom_jvp_call(flat_fun, flat_jvp, *args_flat)\n _, out_tree = lu.merge_linear_aux(out_tree1, out_tree2)\n return tree_unflatten(out_tree, out_flat)\n\n@lu.transformation_with_aux\ndef _flatten_jvp(in_tree, *args):\n primals_in, tangents_in = split_list(args, [len(args) // 2])\n py_primals = tree_unflatten(in_tree, primals_in)\n py_tangents = tree_unflatten(in_tree, tangents_in)\n pair_out = yield (py_primals, py_tangents), {}\n if not isinstance(pair_out, (list, tuple)) or len(pair_out) != 2:\n msg = (\"Custom JVP rule must produce a pair (list or tuple of length two) \"\n \"representing primal and tangent outputs, got {}.\")\n raise TypeError(msg.format(pair_out))\n py_primals_out, py_tangents_out = pair_out\n primals_out, out_tree = tree_flatten(py_primals_out)\n tangents_out, out_tree2 = tree_flatten(py_tangents_out)\n if out_tree != out_tree2:\n msg = (\"Custom JVP rule must produce primal and tangent outputs with equal \"\n \"container (pytree) structures, but got {} and {} respectively.\")\n raise TypeError(msg.format(out_tree, out_tree2))\n primal_avals_out = [raise_to_shaped(core.get_aval(x)) for x in primals_out]\n tangent_avals_out = [raise_to_shaped(core.get_aval(t)) for t in tangents_out]\n if primal_avals_out != tangent_avals_out:\n if len(primal_avals_out) == 1:\n (av1,), (av2,) = primal_avals_out, tangent_avals_out\n msg = (\"Custom JVP rule must produce primal and tangent outputs with \"\n \"equal shapes and dtypes, but got {} and {} respectively.\")\n raise TypeError(msg.format(av1.str_short(), av2.str_short()))\n else:\n msg = (\"Custom JVP rule must produce primal and tangent outputs with \"\n \"equal shapes and dtypes, but got:\\n{}\")\n disagreements = (\n \" primal {} for tangent {}\".format(av1.str_short(), av2.str_short())\n for av1, av2 in zip(primal_avals_out, tangent_avals_out) if av1 != av2)\n raise TypeError(msg.format('\\n'.join(disagreements)))\n yield primals_out + tangents_out, out_tree\n\nclass CustomJVPCallPrimitive(core.CallPrimitive):\n def bind(self, fun, jvp, *args):\n args = map(core.full_lower, args)\n top_trace = core.find_top_trace(args)\n fun, env_trace_todo1 = core.process_env_traces(\n fun, self, top_trace and top_trace.level, ())\n jvp, env_trace_todo2 = core.process_env_traces(\n jvp, self, top_trace and top_trace.level, ())\n if top_trace is None:\n with core.new_sublevel():\n outs = self.impl(fun, jvp, *args)\n else:\n tracers = map(top_trace.full_raise, args)\n outs = top_trace.process_custom_jvp_call(self, fun, jvp, tracers)\n _, env_trace_todo = lu.merge_linear_aux(env_trace_todo1, env_trace_todo2)\n if env_trace_todo:\n raise core.UnexpectedTracerError\n return map(core.full_lower, outs)\n\n def impl(self, fun, _, *args):\n return fun.call_wrapped(*args)\n\ncustom_jvp_call_p = CustomJVPCallPrimitive('custom_jvp_call')\ncustom_jvp_call = custom_jvp_call_p.bind\n\n\ndef custom_jvp_call_jaxpr(fun, jvp, *args):\n in_avals = [raise_to_shaped(core.get_aval(x)) for x in args]\n fun_jaxpr = _initial_style_jaxpr(fun, in_avals)\n jvp_jaxpr_thunk = _memoize(lambda: _initial_style_jaxpr(jvp, in_avals * 2))\n return custom_jvp_call_jaxpr_p.bind(*args, fun_jaxpr=fun_jaxpr,\n jvp_jaxpr_thunk=jvp_jaxpr_thunk)\n\ndef _custom_jvp_call_jaxpr_impl(*args, fun_jaxpr, **_):\n return core.jaxpr_as_fun(fun_jaxpr)(*args)\n\ndef _custom_jvp_call_jaxpr_abstract_eval(*_, fun_jaxpr, **__):\n return fun_jaxpr.out_avals\n\ncustom_jvp_call_jaxpr_p = core.Primitive('custom_jvp_call_jaxpr')\ncustom_jvp_call_jaxpr_p.multiple_results = True\ncustom_jvp_call_jaxpr_p.def_impl(_custom_jvp_call_jaxpr_impl)\ncustom_jvp_call_jaxpr_p.def_abstract_eval(_custom_jvp_call_jaxpr_abstract_eval)\n\ndef _custom_jvp_call_jaxpr_jvp(primals, tangents, *, fun_jaxpr, jvp_jaxpr_thunk):\n jvp_jaxpr = jvp_jaxpr_thunk()\n tangents = map(ad.instantiate_zeros, tangents)\n outs = core.jaxpr_as_fun(jvp_jaxpr)(*primals, *tangents)\n return split_list(outs, [len(outs) // 2])\nad.primitive_jvps[custom_jvp_call_jaxpr_p] = _custom_jvp_call_jaxpr_jvp\n\ndef _custom_jvp_call_jaxpr_vmap(args, in_dims, *, fun_jaxpr, jvp_jaxpr_thunk):\n size, = {x.shape[d] for x, d in zip(args, in_dims) if d is not not_mapped}\n args = [batching.moveaxis(x, d, 0) if d is not not_mapped and d != 0\n else x for x, d in zip(args, in_dims)]\n num_out = len(fun_jaxpr.out_avals)\n\n in_batched = [d is not not_mapped for d in in_dims]\n batched_fun_jaxpr, out_batched = batching.batch_jaxpr(fun_jaxpr, size, in_batched, False)\n out_dims1 = [0 if b else not_mapped for b in out_batched]\n out_dims2 = []\n\n @_memoize\n def batched_jvp_jaxpr_thunk():\n jvp_jaxpr = jvp_jaxpr_thunk()\n _, all_batched = batching.batch_jaxpr(jvp_jaxpr, size, in_batched * 2, False)\n primals_batched, tangents_batched = split_list(all_batched, [num_out])\n out_batched = map(op.or_, primals_batched, tangents_batched)\n out_dims2.append([0 if b else not_mapped for b in out_batched])\n batched_jvp_jaxpr, _ = batching.batch_jaxpr(jvp_jaxpr, size, in_batched * 2,\n out_batched * 2)\n return batched_jvp_jaxpr\n\n batched_outs = custom_jvp_call_jaxpr_p.bind(\n *args, fun_jaxpr=batched_fun_jaxpr, jvp_jaxpr_thunk=batched_jvp_jaxpr_thunk)\n out_dims = out_dims2[0] if out_dims2 else out_dims1\n return batched_outs, out_dims\nbatching.primitive_batchers[custom_jvp_call_jaxpr_p] = _custom_jvp_call_jaxpr_vmap\n\nxla.initial_style_translations[custom_jvp_call_jaxpr_p] = \\\n xla.lower_fun_initial_style(_custom_jvp_call_jaxpr_impl)\n\n# If a (multi)linear function is defined with a custom jvp, then\n# custom_jvp_call_jaxpr can appear in jaxprs to be transposed. Since it's\n# already been linearized, we can drop the jvp rule.\ndef _custom_jvp_call_jaxpr_transpose(cts, *args, fun_jaxpr, jvp_jaxpr_thunk):\n del jvp_jaxpr_thunk\n return ad.backward_pass(fun_jaxpr.jaxpr, fun_jaxpr.literals, args, cts)\nad.primitive_transposes[custom_jvp_call_jaxpr_p] = _custom_jvp_call_jaxpr_transpose\n\n\n### VJPs\n\nclass custom_vjp:\n \"\"\"Set up a JAX-transformable function for a custom VJP rule definition.\n\n This class is meant to be used as a function decorator. Instances are\n callables that behave similarly to the underlying function to which the\n decorator was applied, except when a reverse-mode differentiation\n transformation (like :py:func:`jax.grad`) is applied, in which case a custom\n user-supplied VJP rule function is used instead of tracing into and performing\n automatic differentiation of the underlying function's implementation. There\n is a single instance method, ``defvjp``, which defines the custom VJP rule.\n\n This decorator precludes the use of forward-mode automatic differentiation.\n\n For example::\n\n import jax.numpy as jnp\n\n @jax.custom_vjp\n def f(x, y):\n return jnp.sin(x) * y\n\n def f_fwd(x, y):\n return f(x, y), (jnp.cos(x), jnp.sin(x), y)\n\n def f_bwd(res, g):\n cos_x, sin_x, y = res\n return (cos_x * g * y, sin_x * g)\n\n f.defvjp(f_fwd, f_bwd)\n\n For a more detailed introduction, see the tutorial_.\n\n .. _tutorial: https://jax.readthedocs.io/en/latest/notebooks/Custom_derivative_rules_for_Python_code.html\n \"\"\"\n\n def __init__(self, fun, nondiff_argnums=()):\n self.fun = fun\n self.nondiff_argnums = nondiff_argnums\n self.fwd = None\n self.bwd = None\n update_wrapper(self, fun)\n\n def defvjp(self, fwd, bwd):\n \"\"\"Define a custom VJP rule for the function represented by this instance.\n\n Args:\n fwd: a Python callable representing the forward pass of the custom VJP\n rule. When there are no ``nondiff_argnums``, the ``fwd`` function has\n the same input signature as the underlying primal function. It should\n return as output a pair, where the first element represents the primal\n output and the second element represents any \"residual\" values to store\n from the forward pass for use on the backward pass by the function\n ``bwd``. Input arguments and elements of the output pair may be arrays\n or nested tuples/lists/dicts thereof.\n bwd: a Python callable representing the backward pass of the custom VJP\n rule. When there are no ``nondiff_argnums``, the ``bwd`` function takes\n two arguments, where the first is the \"residual\" values produced on the\n forward pass by ``fwd``, and the second is the output cotangent with the\n same structure as the primal function output. The output of ``bwd`` must\n be a tuple of length equal to the number of arguments of the primal\n function, and the tuple elements may be arrays or nested\n tuples/lists/dicts thereof so as to match the structure of the primal\n input arguments.\n\n Returns:\n None.\n\n Example::\n\n import jax.numpy as jnp\n\n @jax.custom_vjp\n def f(x, y):\n return jnp.sin(x) * y\n\n def f_fwd(x, y):\n return f(x, y), (jnp.cos(x), jnp.sin(x), y)\n\n def f_bwd(res, g):\n cos_x, sin_x, y = res\n return (cos_x * g * y, sin_x * g)\n\n f.defvjp(f_fwd, f_bwd)\n \"\"\"\n self.fwd = fwd\n self.bwd = bwd\n\n def __call__(self, *args, **kwargs):\n if not self.fwd or not self.bwd:\n msg = \"No VJP defined for custom_vjp function {} using defvjp.\"\n raise AttributeError(msg.format(self.__name__))\n args = _resolve_kwargs(self.fun, args, kwargs)\n if self.nondiff_argnums:\n is_nondiff = [False] * len(args)\n for i in self.nondiff_argnums: is_nondiff[i] = True\n args = [_stop_gradient(x) if b else x for b, x in zip(is_nondiff, args)]\n dyn_argnums = [i for i, b in enumerate(is_nondiff) if not b]\n f_, dyn_args = argnums_partial(lu.wrap_init(self.fun), dyn_argnums, args)\n static_args = [args[i] for i in self.nondiff_argnums]\n fwd, _ = argnums_partial(lu.wrap_init(self.fwd), dyn_argnums, args)\n bwd = _add_args(lu.wrap_init(self.bwd), static_args, left=True)\n else:\n f_, dyn_args = lu.wrap_init(self.fun), args\n fwd, bwd = lu.wrap_init(self.fwd), lu.wrap_init(self.bwd)\n args_flat, in_tree = tree_flatten(dyn_args)\n flat_fun, out_tree = flatten_fun_nokwargs(f_, in_tree)\n flat_fwd, out_trees = _flatten_fwd(fwd, in_tree)\n flat_bwd = _flatten_bwd(bwd, in_tree, out_trees)\n if _initial_style_staging():\n out_flat = custom_vjp_call_jaxpr(flat_fun, flat_fwd, flat_bwd,\n *args_flat, out_trees=out_trees)\n out_tree = out_tree()\n else:\n out_flat = custom_vjp_call(flat_fun, flat_fwd, flat_bwd,\n *args_flat, out_trees=out_trees)\n fst, aux = lu.merge_linear_aux(out_tree, out_trees)\n out_tree = aux if fst else aux[0]\n return tree_unflatten(out_tree, out_flat)\n\n@lu.transformation_with_aux\ndef _flatten_fwd(in_tree, *args):\n py_args = tree_unflatten(in_tree, args)\n pair_out = yield py_args, {}\n if not isinstance(pair_out, (list, tuple)) or len(pair_out) != 2:\n msg = (\"Custom VJP fwd function must produce a pair (list or tuple of \"\n \"length two) representing primal outputs and residuals (values \"\n \"stored from the forward pass for use on the backward pass), \"\n \"got {}.\")\n raise TypeError(msg.format(pair_out))\n py_outs, res = pair_out\n out, out_tree = tree_flatten(py_outs)\n res, res_tree = tree_flatten(res)\n yield res + out, (out_tree, res_tree)\n\n@lu.transformation\ndef _flatten_bwd(in_tree, out_trees, *args):\n out_tree, res_tree = out_trees()\n res, cts_out = split_list(args, [res_tree.num_leaves])\n py_res = tree_unflatten(res_tree, res)\n py_cts_out = tree_unflatten(out_tree, cts_out)\n py_cts_in = yield (py_res, py_cts_out), {}\n cts_in, in_tree2 = tree_flatten(py_cts_in)\n if in_tree != in_tree2:\n msg = (\"Custom VJP rule must produce an output with the same container \"\n \"(pytree) structure as the args tuple of the primal function, \"\n \"and in particular must produce a tuple of length equal to the \"\n \"number of arguments to the primal function, but got VJP output \"\n \"structure {} for primal input structure {}.\")\n raise TypeError(msg.format(in_tree2, in_tree)) from None\n yield cts_in\n\n\nclass CustomVJPCallPrimitive(core.CallPrimitive):\n def bind(self, fun, fwd, bwd, *args, out_trees):\n args = map(core.full_lower, args)\n top_trace = core.find_top_trace(args)\n if top_trace is None:\n outs = fun.call_wrapped(*args)\n else:\n tracers = map(top_trace.full_raise, args)\n outs = top_trace.process_custom_vjp_call(self, fun, fwd, bwd, tracers,\n out_trees=out_trees)\n return map(core.full_lower, outs)\n\n def impl(self, fun, fwd, bwd, *args, out_trees):\n del fwd, bwd, out_trees\n return fun.call_wrapped(*args)\n\ncustom_vjp_call_p = CustomVJPCallPrimitive('custom_vjp_call')\ncustom_vjp_call = custom_vjp_call_p.bind\n\ndef custom_vjp_call_jaxpr(fun, fwd, bwd, *args, out_trees):\n in_avals = [raise_to_shaped(core.get_aval(x)) for x in args]\n fun_jaxpr = _initial_style_jaxpr(fun, in_avals)\n fwd_jaxpr_thunk = _memoize(lambda: _initial_style_jaxpr(fwd, in_avals))\n return custom_vjp_call_jaxpr_p.bind(*args, fun_jaxpr=fun_jaxpr,\n fwd_jaxpr_thunk=fwd_jaxpr_thunk, bwd=bwd,\n out_trees=out_trees)\n\ndef _custom_vjp_call_jaxpr_impl(*args, fun_jaxpr, **_):\n return core.jaxpr_as_fun(fun_jaxpr)(*args)\n\ndef _custom_vjp_call_jaxpr_abstract_eval(*_, fun_jaxpr, **__):\n return fun_jaxpr.out_avals\n\ncustom_vjp_call_jaxpr_p = core.Primitive('custom_vjp_call_jaxpr')\ncustom_vjp_call_jaxpr_p.multiple_results = True\ncustom_vjp_call_jaxpr_p.def_impl(_custom_vjp_call_jaxpr_impl)\ncustom_vjp_call_jaxpr_p.def_abstract_eval(_custom_vjp_call_jaxpr_abstract_eval)\n\ndef _custom_vjp_call_jaxpr_jvp(primals, tangents, *, fun_jaxpr, fwd_jaxpr_thunk,\n bwd, out_trees):\n tangents = map(ad.instantiate_zeros, tangents)\n fwd_jaxpr = fwd_jaxpr_thunk()\n out_tree, res_tree = out_trees()\n res_and_primals_out = core.jaxpr_as_fun(fwd_jaxpr)(*primals)\n res, primals_out = split_list(res_and_primals_out, [res_tree.num_leaves])\n avals_out = [raise_to_shaped(core.get_aval(x)) for x in primals_out]\n tangents_out = ad.custom_lin_p.bind(\n *res, *tangents, num_res=res_tree.num_leaves, bwd=bwd, avals_out=avals_out)\n return primals_out, tangents_out\nad.primitive_jvps[custom_vjp_call_jaxpr_p] = _custom_vjp_call_jaxpr_jvp\n\ndef _custom_vjp_call_jaxpr_vmap(args, in_dims, *, fun_jaxpr, fwd_jaxpr_thunk,\n bwd, out_trees):\n size, = {x.shape[d] for x, d in zip(args, in_dims) if d is not not_mapped}\n args = [batching.moveaxis(x, d, 0) if d is not not_mapped and d != 0\n else x for x, d in zip(args, in_dims)]\n\n in_batched = [d is not not_mapped for d in in_dims]\n batched_fun_jaxpr, out_batched = batching.batch_jaxpr(fun_jaxpr, size, in_batched, False)\n out_dims1 = [0 if b else not_mapped for b in out_batched]\n out_dims2 = []\n\n @_memoize\n def batched_fwd_jaxpr_thunk():\n fwd_jaxpr = fwd_jaxpr_thunk()\n batched_fwd_jaxpr, out_batched = batching.batch_jaxpr(fwd_jaxpr, size, in_batched, False)\n out_dims2.append([0 if b else not_mapped for b in out_batched])\n return batched_fwd_jaxpr\n\n fwd_in_dims = [0 if b else not_mapped for b in in_batched]\n fwd_out_dims = lambda: out_dims2[0]\n # TODO: Support collectives in custom_vjp?\n batched_bwd = batching.batch_fun(bwd, fwd_out_dims, fwd_in_dims,\n axis_name='__unused_axis_name', sum_match=True)\n\n batched_outs = custom_vjp_call_jaxpr_p.bind(\n *args, fun_jaxpr=batched_fun_jaxpr,\n fwd_jaxpr_thunk=batched_fwd_jaxpr_thunk, bwd=batched_bwd,\n out_trees=out_trees)\n out_dims = out_dims2[0] if out_dims2 else out_dims1\n return batched_outs, out_dims\nbatching.primitive_batchers[custom_vjp_call_jaxpr_p] = _custom_vjp_call_jaxpr_vmap\n\nxla.initial_style_translations[custom_vjp_call_jaxpr_p] = \\\n xla.lower_fun_initial_style(_custom_vjp_call_jaxpr_impl)\n\nbatching.primitive_batchers[ad.custom_lin_p] = ad._raise_custom_vjp_error_on_jvp\n\n\n# TODO(mattjj): remove when omnistaging fully lands\n@config.register_omnistaging_enabler\ndef omnistaging_enabler() -> None:\n global _initial_style_jaxpr\n\n def _initial_style_jaxpr(fun, in_avals):\n jaxpr, out_avals, consts = pe.trace_to_jaxpr_dynamic(fun, in_avals)\n typed_jaxpr = core.TypedJaxpr(jaxpr, consts, in_avals, out_avals)\n return typed_jaxpr\n\n def bind(self, fun, jvp, *args):\n args = map(core.full_lower, args)\n top_trace = core.find_top_trace(args)\n fun, env_trace_todo1 = core.process_env_traces(\n fun, self, top_trace and top_trace.level, ())\n jvp, env_trace_todo2 = core.process_env_traces(\n jvp, self, top_trace and top_trace.level, ())\n tracers = map(top_trace.full_raise, args) # type: ignore\n outs = top_trace.process_custom_jvp_call(self, fun, jvp, tracers) # type: ignore\n _, env_trace_todo = lu.merge_linear_aux(env_trace_todo1, env_trace_todo2)\n if env_trace_todo:\n raise core.UnexpectedTracerError\n return map(core.full_lower, outs)\n CustomJVPCallPrimitive.bind = bind # type: ignore\n", "sub_path": "jax/custom_derivatives.py", "file_name": "custom_derivatives.py", "file_ext": "py", "file_size_in_byte": 25384, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "util.safe_map", "line_number": 34, "usage_type": "name"}, {"api_name": "util.safe_zip", "line_number": 35, "usage_type": "name"}, {"api_name": "inspect.signature", "line_number": 41, "usage_type": "call"}, {"api_name": "api_util.wrap_hashably", "line_number": 49, "usage_type": "argument"}, {"api_name": "interpreters.partial_eval.PartialVal.unknown", "line_number": 72, "usage_type": "call"}, {"api_name": "interpreters.partial_eval.PartialVal", "line_number": 72, "usage_type": "attribute"}, {"api_name": "interpreters.partial_eval", "line_number": 72, "usage_type": "name"}, {"api_name": "interpreters.partial_eval.trace_to_jaxpr", "line_number": 73, "usage_type": "call"}, {"api_name": "interpreters.partial_eval", "line_number": 73, "usage_type": "name"}, {"api_name": "abstract_arrays.raise_to_shaped", "line_number": 76, "usage_type": "argument"}, {"api_name": "util.unzip2", "line_number": 76, "usage_type": "call"}, {"api_name": "config.config.omnistaging_enabled", "line_number": 81, "usage_type": "attribute"}, {"api_name": "config.config", "line_number": 81, "usage_type": "name"}, {"api_name": "functools.reduce", "line_number": 87, "usage_type": "call"}, {"api_name": "interpreters.ad.add_tangents", "line_number": 87, "usage_type": "attribute"}, {"api_name": "interpreters.ad", "line_number": 87, "usage_type": "name"}, {"api_name": "tree_util.tree_map", "line_number": 90, "usage_type": "call"}, {"api_name": "ad_util.Zero.from_value", "line_number": 90, "usage_type": "attribute"}, {"api_name": "ad_util.Zero", "line_number": 90, "usage_type": "name"}, {"api_name": "ad_util.stop_gradient_p.bind", "line_number": 95, "usage_type": "call"}, {"api_name": "ad_util.stop_gradient_p", "line_number": 95, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 92, "usage_type": "call"}, {"api_name": "tree_util.tree_map", "line_number": 92, "usage_type": "argument"}, {"api_name": "functools.update_wrapper", "line_number": 138, "usage_type": "call"}, {"api_name": "tree_util.tree_multimap", "line_number": 205, "usage_type": "call"}, {"api_name": "api_util.argnums_partial", "line_number": 220, "usage_type": "call"}, {"api_name": "tree_util.tree_flatten", "line_number": 226, "usage_type": "call"}, {"api_name": "api_util.flatten_fun_nokwargs", "line_number": 227, "usage_type": "call"}, {"api_name": "tree_util.tree_unflatten", "line_number": 235, "usage_type": "call"}, {"api_name": "util.split_list", "line_number": 239, "usage_type": "call"}, {"api_name": "tree_util.tree_unflatten", "line_number": 240, "usage_type": "call"}, {"api_name": "tree_util.tree_unflatten", "line_number": 241, "usage_type": "call"}, {"api_name": "tree_util.tree_flatten", "line_number": 248, "usage_type": "call"}, {"api_name": "tree_util.tree_flatten", "line_number": 249, "usage_type": "call"}, {"api_name": "abstract_arrays.raise_to_shaped", "line_number": 254, "usage_type": "call"}, {"api_name": "abstract_arrays.raise_to_shaped", "line_number": 255, "usage_type": "call"}, {"api_name": "abstract_arrays.raise_to_shaped", "line_number": 298, "usage_type": "call"}, {"api_name": "interpreters.ad.instantiate_zeros", "line_number": 317, "usage_type": "attribute"}, {"api_name": "interpreters.ad", "line_number": 317, "usage_type": "name"}, {"api_name": "util.split_list", "line_number": 319, "usage_type": "call"}, {"api_name": "interpreters.ad.primitive_jvps", "line_number": 320, "usage_type": "attribute"}, {"api_name": "interpreters.ad", "line_number": 320, "usage_type": "name"}, {"api_name": "interpreters.batching.not_mapped", "line_number": 323, "usage_type": "name"}, {"api_name": "interpreters.batching.not_mapped", "line_number": 324, "usage_type": "name"}, {"api_name": "interpreters.batching.moveaxis", "line_number": 324, "usage_type": "call"}, {"api_name": "interpreters.batching", "line_number": 324, "usage_type": "name"}, {"api_name": "interpreters.batching.not_mapped", "line_number": 328, "usage_type": "name"}, {"api_name": "interpreters.batching.batch_jaxpr", "line_number": 329, "usage_type": "call"}, {"api_name": "interpreters.batching", "line_number": 329, "usage_type": "name"}, {"api_name": "interpreters.batching.not_mapped", "line_number": 330, "usage_type": "name"}, {"api_name": "interpreters.batching.batch_jaxpr", "line_number": 336, "usage_type": "call"}, {"api_name": "interpreters.batching", "line_number": 336, "usage_type": "name"}, {"api_name": "util.split_list", "line_number": 337, "usage_type": "call"}, {"api_name": "operator.or_", "line_number": 338, "usage_type": "attribute"}, {"api_name": "interpreters.batching.not_mapped", "line_number": 339, "usage_type": "name"}, {"api_name": "interpreters.batching.batch_jaxpr", "line_number": 340, "usage_type": "call"}, {"api_name": "interpreters.batching", "line_number": 340, "usage_type": "name"}, {"api_name": "interpreters.batching.primitive_batchers", "line_number": 348, "usage_type": "attribute"}, {"api_name": "interpreters.batching", "line_number": 348, "usage_type": "name"}, {"api_name": "interpreters.xla.initial_style_translations", "line_number": 350, "usage_type": "attribute"}, {"api_name": "interpreters.xla", "line_number": 350, "usage_type": "name"}, {"api_name": "interpreters.xla.lower_fun_initial_style", "line_number": 351, "usage_type": "call"}, {"api_name": "interpreters.xla", "line_number": 351, "usage_type": "name"}, {"api_name": "interpreters.ad.backward_pass", "line_number": 358, "usage_type": "call"}, {"api_name": "interpreters.ad", "line_number": 358, "usage_type": "name"}, {"api_name": "interpreters.ad.primitive_transposes", "line_number": 359, "usage_type": "attribute"}, {"api_name": "interpreters.ad", "line_number": 359, "usage_type": "name"}, {"api_name": "functools.update_wrapper", "line_number": 404, "usage_type": "call"}, {"api_name": "api_util.argnums_partial", "line_number": 461, "usage_type": "call"}, {"api_name": "api_util.argnums_partial", "line_number": 463, "usage_type": "call"}, {"api_name": "tree_util.tree_flatten", "line_number": 468, "usage_type": "call"}, {"api_name": "api_util.flatten_fun_nokwargs", "line_number": 469, "usage_type": "call"}, {"api_name": "tree_util.tree_unflatten", "line_number": 481, "usage_type": "call"}, {"api_name": "tree_util.tree_unflatten", "line_number": 485, "usage_type": "call"}, {"api_name": "tree_util.tree_flatten", "line_number": 494, "usage_type": "call"}, {"api_name": "tree_util.tree_flatten", "line_number": 495, "usage_type": "call"}, {"api_name": "util.split_list", "line_number": 501, "usage_type": "call"}, {"api_name": "tree_util.tree_unflatten", "line_number": 502, "usage_type": "call"}, {"api_name": "tree_util.tree_unflatten", "line_number": 503, "usage_type": "call"}, {"api_name": "tree_util.tree_flatten", "line_number": 505, "usage_type": "call"}, {"api_name": "abstract_arrays.raise_to_shaped", "line_number": 536, "usage_type": "call"}, {"api_name": "interpreters.ad.instantiate_zeros", "line_number": 556, "usage_type": "attribute"}, {"api_name": "interpreters.ad", "line_number": 556, "usage_type": "name"}, {"api_name": "util.split_list", "line_number": 560, "usage_type": "call"}, {"api_name": "abstract_arrays.raise_to_shaped", "line_number": 561, "usage_type": "call"}, {"api_name": "interpreters.ad.custom_lin_p.bind", "line_number": 562, "usage_type": "call"}, {"api_name": "interpreters.ad.custom_lin_p", "line_number": 562, "usage_type": "attribute"}, {"api_name": "interpreters.ad", "line_number": 562, "usage_type": "name"}, {"api_name": "interpreters.ad.primitive_jvps", "line_number": 565, "usage_type": "attribute"}, {"api_name": "interpreters.ad", "line_number": 565, "usage_type": "name"}, {"api_name": "interpreters.batching.not_mapped", "line_number": 569, "usage_type": "name"}, {"api_name": "interpreters.batching.not_mapped", "line_number": 570, "usage_type": "name"}, {"api_name": "interpreters.batching.moveaxis", "line_number": 570, "usage_type": "call"}, {"api_name": "interpreters.batching", "line_number": 570, "usage_type": "name"}, {"api_name": "interpreters.batching.not_mapped", "line_number": 573, "usage_type": "name"}, {"api_name": "interpreters.batching.batch_jaxpr", "line_number": 574, "usage_type": "call"}, {"api_name": "interpreters.batching", "line_number": 574, "usage_type": "name"}, {"api_name": "interpreters.batching.not_mapped", "line_number": 575, "usage_type": "name"}, {"api_name": "interpreters.batching.batch_jaxpr", "line_number": 581, "usage_type": "call"}, {"api_name": "interpreters.batching", "line_number": 581, "usage_type": "name"}, {"api_name": "interpreters.batching.not_mapped", "line_number": 582, "usage_type": "name"}, {"api_name": "interpreters.batching.not_mapped", "line_number": 585, "usage_type": "name"}, {"api_name": "interpreters.batching.batch_fun", "line_number": 588, "usage_type": "call"}, {"api_name": "interpreters.batching", "line_number": 588, "usage_type": "name"}, {"api_name": "interpreters.batching.primitive_batchers", "line_number": 597, "usage_type": "attribute"}, {"api_name": "interpreters.batching", "line_number": 597, "usage_type": "name"}, {"api_name": "interpreters.xla.initial_style_translations", "line_number": 599, "usage_type": "attribute"}, {"api_name": "interpreters.xla", "line_number": 599, "usage_type": "name"}, {"api_name": "interpreters.xla.lower_fun_initial_style", "line_number": 600, "usage_type": "call"}, {"api_name": "interpreters.xla", "line_number": 600, "usage_type": "name"}, {"api_name": "interpreters.batching.primitive_batchers", "line_number": 602, "usage_type": "attribute"}, {"api_name": "interpreters.batching", "line_number": 602, "usage_type": "name"}, {"api_name": "interpreters.ad.custom_lin_p", "line_number": 602, "usage_type": "attribute"}, {"api_name": "interpreters.ad", "line_number": 602, "usage_type": "name"}, {"api_name": "interpreters.ad._raise_custom_vjp_error_on_jvp", "line_number": 602, "usage_type": "attribute"}, {"api_name": "interpreters.partial_eval.trace_to_jaxpr_dynamic", "line_number": 611, "usage_type": "call"}, {"api_name": "interpreters.partial_eval", "line_number": 611, "usage_type": "name"}, {"api_name": "config.config.register_omnistaging_enabler", "line_number": 606, "usage_type": "attribute"}, {"api_name": "config.config", "line_number": 606, "usage_type": "name"}]}
+{"seq_id": "298241369", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Aug 11 16:58:17 2018\n\n@author: danpal\n\"\"\"\n\nfrom Bio import SeqIO\nfrom Bio.SeqRecord import SeqRecord\n\n\nclass Concatemer:\n def __init__(self, name, mers):\n self.name = name\n self.mers = mers\n\n @property\n def name(self):\n return self._name\n\n @name.setter\n def name(self, new):\n if type(new) == str:\n self._name = new\n else:\n raise ValueError('The name must be a string.')\n\n @property\n def mers(self):\n return self._mers\n\n @mers.setter\n def mers(self, new):\n if (type(new) == dict\n and all([type(v) == SeqRecord for v in new.values()])\n and all([type(k) == str for k in new.keys()])):\n self._mers = new\n else:\n raise ValueError('\"mers\" must be a dict with the format:'\n + ' {str: SeqRecor}')\n\n def containsMers(self, mers_list):\n check = True\n for mer in mers_list:\n if mer not in self.mers:\n print(f'There is no \"{mer}\" mer in \"{self.name}\"')\n check = False\n return check\n\n def getConcatemer(self, mers_list=None, return_mers=False):\n if mers_list is None:\n mers = list(self.mers.keys())\n else:\n if self.containsMers(mers_list):\n mers = mers_list\n else:\n raise ValueError('Some specified mers are not present.')\n conc = ''\n for mer in mers:\n conc += str(self.mers[mer].seq)\n if return_mers:\n return conc, mers\n else:\n return conc\n\n def printConcatemer(self, mers_list=None):\n print(self.getConcatemer(mers_list))\n\n def writeConcatemer(self, file, mers_list=None, mode='a'):\n conc, mers = self.getConcatemer(mers_list, return_mers=True)\n with open(file, mode) as f:\n f.write(f'>{self.name} {\" \".join(mers)}\\n{conc}\\n')\n\n def __repr__(self):\n return f'Concatemer({self.name}, {self.mers})'\n\n def __str__(self):\n return f''\n\n def __len__(self):\n return len(self.mers)\n\n\ndef getSequences(file):\n return SeqIO.parse(file, \"fasta\")\n\n\ndef clusterDescriptions(sequences, fields=slice(1, -1)):\n clusters = {}\n for seq in sequences:\n name = ' '.join(seq.description.split()[fields])\n if name not in clusters:\n clusters[name] = [seq]\n else:\n clusters[name].append(seq)\n return clusters\n\n\ndef makeConcatemer(clusters, name, mers_list):\n mers = {}\n for mer in mers_list:\n for seq in clusters[name]:\n if seq.description.split()[-1] == mer:\n mers[mer] = seq\n if mer not in mers:\n print(f'WARNING: \"{mer}\" could not be found for \"{name}\".')\n return Concatemer(name, mers)\n\n\ndef getConcatemers(file,\n mers_list=None,\n fields=slice(1, -1)):\n \"\"\"mers_list (default: ['PB2', 'PB1', 'PA', 'HA', 'NP',\n 'NA', 'M1', 'M2', 'NS1', 'NS2']\"\"\"\n seq_iter = getSequences(file)\n clusters = clusterDescriptions(seq_iter, fields)\n if mers_list is None:\n mers_list = ['PB2', 'PB1', 'PA', 'HA', 'NP',\n 'NA', 'M1', 'M2', 'NS1', 'NS2']\n return [makeConcatemer(clusters, name, mers_list) for name in clusters]\n\n\ndef writeConcatemers(in_file,\n out_file,\n mers_list=None,\n fields=slice(1, -1)):\n concatemers = getConcatemers(in_file, mers_list, fields)\n for conc in concatemers:\n try:\n conc.writeConcatemer(out_file, mers_list)\n except ValueError as error:\n print(f'The concatemer of is \"{conc.name}\" is not been written'\n + f' because: {error}')\n\n\nif __name__ == '__main__':\n path = \"/home/danpal/Unidad/05_virus/influenza/\"\n in_file = path + \"inf.fa\"\n mers_list = ['PB2', 'PB1', 'HA']\n out_file = path + \"test2.fa\"\n# c = getConcatemers(in_file)\n# a = c[0]\n writeConcatemers(in_file, out_file, mers_list)\n", "sub_path": "python/cluster/fasta_desc.py", "file_name": "fasta_desc.py", "file_ext": "py", "file_size_in_byte": 4180, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "Bio.SeqRecord.SeqRecord", "line_number": 36, "usage_type": "name"}, {"api_name": "Bio.SeqIO.parse", "line_number": 86, "usage_type": "call"}, {"api_name": "Bio.SeqIO", "line_number": 86, "usage_type": "name"}]}
+{"seq_id": "353725084", "text": "#!/usr/bin/env python\n\n##################################################\n#\n# howdoi - a code search tool.\n# written by Benjamin Gleitzman (gleitz@mit.edu)\n# inspired by Rich Jones (rich@anomos.info)\n#\n##################################################\n\nimport urllib.request, urllib.parse, urllib.error\nimport sys\nimport json\nimport argparse\nimport re\nimport lxml.html\n\n#from pyquery import PyQuery as pq\n\nGOOGLE_SEARCH_URL = \"https://www.google.com/search?q=site:stackoverflow.com%20{0}\"\nDUCK_SEARCH_URL = \"http://duckduckgo.com/html?q=site%3Astackoverflow.com%20{0}\"\nUSER_AGENT = \"Mozilla/5.0 (Macintosh; Intel Mac OS X 10_8_2) AppleWebKit/537.17 (KHTML, like Gecko) Chrome/24.0.1309.0 Safari/537.17\"\n\ndef get_result(url):\n opener = urllib.request.build_opener()\n opener.addheaders = [('User-agent', USER_AGENT)]\n result = opener.open(url)\n return result.read()\n\ndef is_question(link):\n return re.search(\"^http\\:\\/\\/stackoverflow\\.com\\/questions/\\d+/\", link)\n\ndef get_google_links(query):\n links = []\n url = GOOGLE_SEARCH_URL.format(urllib.parse.quote(query))\n result = get_result(url)\n html = lxml.html.document_fromstring(result)\n for l in html.iterlinks():\n if is_question(l[2]):\n links.append(l[2])\n \n return links\n\ndef get_duck_links(query):\n url = DUCK_SEARCH_URL.format(urllib.parse.quote(query))\n result = get_result(url)\n html = lxml.html.document_fromstring(result)\n links = html.xpath(\"//a[@href]\")\n return [l.get('href', None) for l in links]\n\ndef get_link_at_pos(links, pos):\n pos = int(pos) - 1\n for link in links:\n if is_question(link):\n if pos == 0:\n break\n else:\n pos = pos - 1\n continue\n return link\n\ndef get_instructions(args):\n text = []\n links = get_google_links(args['query'])\n if not links:\n return ''\n\n link = get_link_at_pos(links, args['pos'])\n if args.get('link'):\n return link\n\n link = link + '?answertab=votes'\n page = get_result(link)\n html = lxml.html.document_fromstring(page)\n first_answer = html.xpath(\"//td[@class='answercell']\")\n tags = first_answer[0].xpath(\"code\") or first_answer[0].xpath(\"//pre\")\n if tags:\n for t in tags[0]:\n text.append(t.text_content())\n else:\n post_text = first_answer[0].xpath(\"div[@class='post-text']/p\")\n if post_text:\n for t in post_text:\n text.append(t.text_content())\n return text\n\n return \"\\n\".join(text)\n\ndef howdoi(args):\n args['query'] = ' '.join(args['query']).replace('?', '')\n instructions = get_instructions(args) or 'Sorry, couldn\\'t find any help with that topic'\n print(instructions)\n\ndef command_line_runner():\n parser = argparse.ArgumentParser(description='code search tool')\n parser.add_argument('query', metavar='QUERY', type=str, nargs=argparse.REMAINDER,\n help='the question to answer')\n parser.add_argument('-p','--pos', help='select answer in specified position (default: 1)', default=1)\n parser.add_argument('-a','--all', help='display the full text of the answer',\n action='store_true')\n parser.add_argument('-l','--link', help='display only the answer link',\n action='store_true')\n args = vars(parser.parse_args())\n howdoi(args)\n\nif __name__ == '__main__':\n command_line_runner()\n", "sub_path": "howdoi/howdoi.py", "file_name": "howdoi.py", "file_ext": "py", "file_size_in_byte": 3456, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "urllib.request.request.build_opener", "line_number": 25, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 25, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 25, "usage_type": "name"}, {"api_name": "re.search", "line_number": 31, "usage_type": "call"}, {"api_name": "urllib.request.parse.quote", "line_number": 35, "usage_type": "call"}, {"api_name": "urllib.request.parse", "line_number": 35, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 35, "usage_type": "name"}, {"api_name": "lxml.html.html.document_fromstring", "line_number": 37, "usage_type": "call"}, {"api_name": "lxml.html.html", "line_number": 37, "usage_type": "attribute"}, {"api_name": "lxml.html", "line_number": 37, "usage_type": "name"}, {"api_name": "urllib.request.parse.quote", "line_number": 45, "usage_type": "call"}, {"api_name": "urllib.request.parse", "line_number": 45, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 45, "usage_type": "name"}, {"api_name": "lxml.html.html.document_fromstring", "line_number": 47, "usage_type": "call"}, {"api_name": "lxml.html.html", "line_number": 47, "usage_type": "attribute"}, {"api_name": "lxml.html", "line_number": 47, "usage_type": "name"}, {"api_name": "lxml.html.html.document_fromstring", "line_number": 74, "usage_type": "call"}, {"api_name": "lxml.html.html", "line_number": 74, "usage_type": "attribute"}, {"api_name": "lxml.html", "line_number": 74, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 95, "usage_type": "call"}, {"api_name": "argparse.REMAINDER", "line_number": 96, "usage_type": "attribute"}]}
+{"seq_id": "443892259", "text": "# Copyright 2018 Tensorforce Team. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ==============================================================================\n\nfrom collections import OrderedDict\n\nimport tensorflow as tf\n\nfrom tensorforce import util\nfrom tensorforce.core import Module\n\n\nclass CircularBuffer(Module):\n \"\"\"\n Circular buffer.\n\n Args:\n name (string): Buffer name\n (internal use).\n capacity (int > 0): Buffer capacity\n (required).\n values_spec (specification): Values specification\n (internal use).\n return_overwritten (bool): Whether to return overwritten values\n (default: false).\n initializers (dict[values]): Buffer initializers\n (internal use).\n device (string): Device name\n (default: inherit value of parent module).\n summary_labels ('all' | iter[string]): Labels of summaries to record\n (default: inherit value of parent module).\n \"\"\"\n\n def __init__(\n self, name, capacity, values_spec, return_overwritten=False, initializers=None,\n device=None, summary_labels=None\n ):\n super().__init__(name=name, device=device, summary_labels=summary_labels)\n\n self.values_spec = values_spec\n self.capacity = capacity\n self.return_overwritten = return_overwritten\n self.initializers = OrderedDict() if initializers is None else initializers\n\n def tf_initialize(self):\n super().tf_initialize()\n\n # Value buffers\n self.buffers = OrderedDict()\n for name, spec in self.values_spec.items():\n if util.is_nested(name=name):\n self.buffers[name] = OrderedDict()\n for inner_name, spec in spec.items():\n shape = (self.capacity,) + spec['shape']\n initializer = self.initializers.get(inner_name, 'zeros')\n self.buffers[name][inner_name] = self.add_variable(\n name=(inner_name + '-buffer'), dtype=spec['type'], shape=shape,\n is_trainable=False, initializer=initializer\n )\n else:\n shape = (self.capacity,) + spec['shape']\n initializer = self.initializers.get(name, 'zeros')\n self.buffers[name] = self.add_variable(\n name=(name + '-buffer'), dtype=spec['type'], shape=shape, is_trainable=False,\n initializer=initializer\n )\n\n # Buffer index (modulo capacity, next index to write to)\n self.buffer_index = self.add_variable(\n name='buffer-index', dtype='long', shape=(), is_trainable=False, initializer='zeros'\n )\n\n def tf_reset(self):\n # Constants\n zero = tf.constant(value=0, dtype=util.tf_dtype(dtype='long'))\n capacity = tf.constant(value=self.capacity, dtype=util.tf_dtype(dtype='long'))\n\n if not self.return_overwritten:\n # Reset buffer index\n assignment = self.buffer_index.assign(value=zero, read_value=False)\n\n # Return no-op\n with tf.control_dependencies(control_inputs=(assignment,)):\n return util.no_operation()\n\n # Overwritten buffer indices\n num_values = tf.minimum(x=self.buffer_index, y=capacity)\n indices = tf.range(start=(self.buffer_index - num_values), limit=self.buffer_index)\n indices = tf.mod(x=indices, y=capacity)\n\n # Get overwritten values\n values = OrderedDict()\n for name, buffer in self.buffers.items():\n if util.is_nested(name=name):\n values[name] = OrderedDict()\n for inner_name, buffer in buffer.items():\n values[name][inner_name] = tf.gather(params=buffer, indices=indices)\n else:\n values[name] = tf.gather(params=buffer, indices=indices)\n\n # Reset buffer index\n with tf.control_dependencies(control_inputs=util.flatten(xs=values)):\n assignment = self.buffer_index.assign(value=zero, read_value=False)\n\n # Return overwritten values\n with tf.control_dependencies(control_inputs=(assignment,)):\n return util.fmap(function=util.identity_operation, xs=values)\n\n def tf_enqueue(self, **values):\n # Constants\n zero = tf.constant(value=0, dtype=util.tf_dtype(dtype='long'))\n capacity = tf.constant(value=self.capacity, dtype=util.tf_dtype(dtype='long'))\n\n # Get number of values\n for value in values.values():\n if not isinstance(value, dict):\n break\n elif len(value) > 0:\n value = next(iter(value.values()))\n break\n if util.tf_dtype(dtype='long') in (tf.int32, tf.int64):\n num_values = tf.shape(input=value, out_type=util.tf_dtype(dtype='long'))[0]\n else:\n num_values = tf.dtypes.cast(\n x=tf.shape(input=value)[0], dtype=util.tf_dtype(dtype='long')\n )\n\n # Check whether instances fit into buffer\n assertion = tf.debugging.assert_less_equal(x=num_values, y=capacity)\n\n if self.return_overwritten:\n # Overwritten buffer indices\n with tf.control_dependencies(control_inputs=(assertion,)):\n start = tf.maximum(x=self.buffer_index, y=capacity)\n limit = tf.maximum(x=(self.buffer_index + num_values), y=capacity)\n num_overwritten = limit - start\n indices = tf.range(start=start, limit=limit)\n indices = tf.mod(x=indices, y=capacity)\n\n # Get overwritten values\n with tf.control_dependencies(control_inputs=(indices,)):\n overwritten_values = OrderedDict()\n for name, buffer in self.buffers.items():\n if util.is_nested(name=name):\n overwritten_values[name] = OrderedDict()\n for inner_name, buffer in buffer.items():\n overwritten_values[name][inner_name] = tf.gather(\n params=buffer, indices=indices\n )\n else:\n overwritten_values[name] = tf.gather(params=buffer, indices=indices)\n\n else:\n overwritten_values = (assertion,)\n\n # Buffer indices to (over)write\n with tf.control_dependencies(control_inputs=util.flatten(xs=overwritten_values)):\n indices = tf.range(start=self.buffer_index, limit=(self.buffer_index + num_values))\n indices = tf.mod(x=indices, y=capacity)\n indices = tf.expand_dims(input=indices, axis=1)\n\n # Write new values\n with tf.control_dependencies(control_inputs=(indices,)):\n assignments = list()\n for name, buffer in self.buffers.items():\n if util.is_nested(name=name):\n for inner_name, buffer in buffer.items():\n assignment = buffer.scatter_nd_update(\n indices=indices, updates=values[name][inner_name]\n )\n assignments.append(assignment)\n else:\n assignment = buffer.scatter_nd_update(indices=indices, updates=values[name])\n assignments.append(assignment)\n\n # Increment buffer index\n with tf.control_dependencies(control_inputs=assignments):\n assignment = self.buffer_index.assign_add(delta=num_values, read_value=False)\n\n # Return overwritten values or no-op\n with tf.control_dependencies(control_inputs=(assignment,)):\n if self.return_overwritten:\n any_overwritten = tf.math.greater(x=num_overwritten, y=zero)\n overwritten_values = util.fmap(\n function=util.identity_operation, xs=overwritten_values\n )\n return any_overwritten, overwritten_values\n else:\n return util.no_operation()\n", "sub_path": "tensorforce/core/utils/circular_buffer.py", "file_name": "circular_buffer.py", "file_ext": "py", "file_size_in_byte": 8898, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "tensorforce.core.Module", "line_number": 24, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 54, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 60, "usage_type": "call"}, {"api_name": "tensorforce.util.is_nested", "line_number": 62, "usage_type": "call"}, {"api_name": "tensorforce.util", "line_number": 62, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 63, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorforce.util.tf_dtype", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorforce.util", "line_number": 86, "usage_type": "name"}, {"api_name": "tensorflow.constant", "line_number": 87, "usage_type": "call"}, {"api_name": "tensorforce.util.tf_dtype", "line_number": 87, "usage_type": "call"}, {"api_name": "tensorforce.util", "line_number": 87, "usage_type": "name"}, {"api_name": "tensorflow.control_dependencies", "line_number": 94, "usage_type": "call"}, {"api_name": "tensorforce.util.no_operation", "line_number": 95, "usage_type": "call"}, {"api_name": "tensorforce.util", "line_number": 95, "usage_type": "name"}, {"api_name": "tensorflow.minimum", "line_number": 98, "usage_type": "call"}, {"api_name": "tensorflow.range", "line_number": 99, "usage_type": "call"}, {"api_name": "tensorflow.mod", "line_number": 100, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 103, "usage_type": "call"}, {"api_name": "tensorforce.util.is_nested", "line_number": 105, "usage_type": "call"}, {"api_name": "tensorforce.util", "line_number": 105, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 106, "usage_type": "call"}, {"api_name": "tensorflow.gather", "line_number": 108, "usage_type": "call"}, {"api_name": "tensorflow.gather", "line_number": 110, "usage_type": "call"}, {"api_name": "tensorflow.control_dependencies", "line_number": 113, "usage_type": "call"}, {"api_name": "tensorforce.util.flatten", "line_number": 113, "usage_type": "call"}, {"api_name": "tensorforce.util", "line_number": 113, "usage_type": "name"}, {"api_name": "tensorflow.control_dependencies", "line_number": 117, "usage_type": "call"}, {"api_name": "tensorforce.util.fmap", "line_number": 118, "usage_type": "call"}, {"api_name": "tensorforce.util", "line_number": 118, "usage_type": "name"}, {"api_name": "tensorforce.util.identity_operation", "line_number": 118, "usage_type": "attribute"}, {"api_name": "tensorflow.constant", "line_number": 122, "usage_type": "call"}, {"api_name": "tensorforce.util.tf_dtype", "line_number": 122, "usage_type": "call"}, {"api_name": "tensorforce.util", "line_number": 122, "usage_type": "name"}, {"api_name": "tensorflow.constant", "line_number": 123, "usage_type": "call"}, {"api_name": "tensorforce.util.tf_dtype", "line_number": 123, "usage_type": "call"}, {"api_name": "tensorforce.util", "line_number": 123, "usage_type": "name"}, {"api_name": "tensorforce.util.tf_dtype", "line_number": 132, "usage_type": "call"}, {"api_name": "tensorforce.util", "line_number": 132, "usage_type": "name"}, {"api_name": "tensorflow.int32", "line_number": 132, "usage_type": "attribute"}, {"api_name": "tensorflow.int64", "line_number": 132, "usage_type": "attribute"}, {"api_name": "tensorflow.shape", "line_number": 133, "usage_type": "call"}, {"api_name": "tensorforce.util.tf_dtype", "line_number": 133, "usage_type": "call"}, {"api_name": "tensorforce.util", "line_number": 133, "usage_type": "name"}, {"api_name": "tensorflow.dtypes.cast", "line_number": 135, "usage_type": "call"}, {"api_name": "tensorflow.dtypes", "line_number": 135, "usage_type": "attribute"}, {"api_name": "tensorflow.shape", "line_number": 136, "usage_type": "call"}, {"api_name": "tensorforce.util.tf_dtype", "line_number": 136, "usage_type": "call"}, {"api_name": "tensorforce.util", "line_number": 136, "usage_type": "name"}, {"api_name": "tensorflow.debugging.assert_less_equal", "line_number": 140, "usage_type": "call"}, {"api_name": "tensorflow.debugging", "line_number": 140, "usage_type": "attribute"}, {"api_name": "tensorflow.control_dependencies", "line_number": 144, "usage_type": "call"}, {"api_name": "tensorflow.maximum", "line_number": 145, "usage_type": "call"}, {"api_name": "tensorflow.maximum", "line_number": 146, "usage_type": "call"}, {"api_name": "tensorflow.range", "line_number": 148, "usage_type": "call"}, {"api_name": "tensorflow.mod", "line_number": 149, "usage_type": "call"}, {"api_name": "tensorflow.control_dependencies", "line_number": 152, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 153, "usage_type": "call"}, {"api_name": "tensorforce.util.is_nested", "line_number": 155, "usage_type": "call"}, {"api_name": "tensorforce.util", "line_number": 155, "usage_type": "name"}, {"api_name": "collections.OrderedDict", "line_number": 156, "usage_type": "call"}, {"api_name": "tensorflow.gather", "line_number": 158, "usage_type": "call"}, {"api_name": "tensorflow.gather", "line_number": 162, "usage_type": "call"}, {"api_name": "tensorflow.control_dependencies", "line_number": 168, "usage_type": "call"}, {"api_name": "tensorforce.util.flatten", "line_number": 168, "usage_type": "call"}, {"api_name": "tensorforce.util", "line_number": 168, "usage_type": "name"}, {"api_name": "tensorflow.range", "line_number": 169, "usage_type": "call"}, {"api_name": "tensorflow.mod", "line_number": 170, "usage_type": "call"}, {"api_name": "tensorflow.expand_dims", "line_number": 171, "usage_type": "call"}, {"api_name": "tensorflow.control_dependencies", "line_number": 174, "usage_type": "call"}, {"api_name": "tensorforce.util.is_nested", "line_number": 177, "usage_type": "call"}, {"api_name": "tensorforce.util", "line_number": 177, "usage_type": "name"}, {"api_name": "tensorflow.control_dependencies", "line_number": 188, "usage_type": "call"}, {"api_name": "tensorflow.control_dependencies", "line_number": 192, "usage_type": "call"}, {"api_name": "tensorflow.math.greater", "line_number": 194, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 194, "usage_type": "attribute"}, {"api_name": "tensorforce.util.fmap", "line_number": 195, "usage_type": "call"}, {"api_name": "tensorforce.util", "line_number": 195, "usage_type": "name"}, {"api_name": "tensorforce.util.identity_operation", "line_number": 196, "usage_type": "attribute"}, {"api_name": "tensorforce.util", "line_number": 196, "usage_type": "name"}, {"api_name": "tensorforce.util.no_operation", "line_number": 200, "usage_type": "call"}, {"api_name": "tensorforce.util", "line_number": 200, "usage_type": "name"}]}
+{"seq_id": "470660896", "text": "import os\nimport sys\nimport openalea.lpy as lpy\nimport openalea.plantgl.all as pgl\nfrom flask import Flask\nfrom flask import request, render_template, url_for, redirect, jsonify, session\nfrom flask import Markup\nfrom flask_socketio import SocketIO\nfrom threading import Lock\n\t\napp = Flask(__name__)\napp.config['SEND_FILE_MAX_AGE_DEFAULT'] = 0\napp.secret_key = 'Iamasecretkey'\nsocketio = SocketIO(app)\nLSystem = None\nlock = Lock()\n\n@app.before_request\ndef make_session_permanent():\n session.permanent = False\n\nif __name__ == '__main__':\n\tapp.run(debug=True)\n\t#host=140.77.193.122\n\n@socketio.on('disconnect')\ndef disconnect():\n session.pop('step', None)\n session.pop('currentStep', None)\n session.pop('code', None)\n\n@app.route('/')\ndef home():\n\treturn render_template('home.html')\n\n@app.route('/editor.html')\ndef editor():\n return render_template('result.html')\n\n\n#When Run is clicked, the server receives the code, derive and interpret the LString and sends the result.\n\n@app.route('/run', methods=['POST'])\ndef run():\n\tout = sys.stdout\n\toutlog = open('outlog.txt', 'w')\n\tsys.stdout = outlog\n\n\tdisconnect()\n\tl = lpy.Lsystem()\n\tcode = request.form['code']\n\tcode = code.encode('ascii', 'ignore')\n\n\ttry:\n\t\tl.set(code)\n\texcept:\n\t\toutlog.close()\n\t\tsys.stdout = out\n\t\toutlog = open('outlog.txt', 'r')\n\t\toutput = outlog.read()\n\t\treturn jsonify({'error': \"Syntax Error\", 'output': output})\n\n\tlstring = l.derive()\n\tilstring = l.interpret(lstring)\n\ttxtlstring = str(ilstring)\n\t# txtlstring = lstringtojson(ilstring)\n\n\toutlog.close()\n\tsys.stdout = out\n\n\toutlog = open('outlog.txt', 'r')\n\toutput = outlog.read()\n\treturn jsonify({'LString' : txtlstring, 'output': output})\n\n#When Step is clicked, the server receives the code, derive and interpret one time the LString, sends the result and keeps the new LString in the session array.\n\n@app.route('/step', methods=['POST'])\ndef step():\n\tglobal LSystem\n\tcode = request.form['code']\n\tcode = code.encode('ascii', 'ignore')\n\tstep = 0\n\n\tif session.get('step') is not None:\n\t\tif code != session['code']:\n\t\t\tl = lpy.Lsystem()\n\t\t\tsession['code'] = code\n\t\t\ttry:\n\t\t\t\tl.set(code)\n\t\t\texcept:\n\t\t\t\treturn jsonify({'error' : 'Syntax error'})\n\n\t\t\tsession['step'] = l.derivationLength\n\t\t\tsession['currentStep'] = 1\n\t\t\tlstring = l.derive(session['currentStep'])\n\t\t\tilstring = l.interpret(lstring)\n\t\t\ttxtlstring = str(ilstring)\n\t\t\tLSystem = l\n\t\t\treturn jsonify({'LString' : txtlstring, 'currentStep' : session['currentStep'], 'step' : session['step']})\n\n\t\telse:\n\t\t\twith lock:\n\t\t\t\tsession['currentStep'] += 1\n\t\t\t\tlstring = LSystem.derive(session['currentStep'])\n\t\t\t\tilstring = LSystem.interpret(lstring)\n\t\t\t\ttxtlstring = str(ilstring)\n\t\t\t\tif session['currentStep'] < session['step']:\n\t\t\t\t\treturn jsonify({'LString' : txtlstring, 'currentStep' : session['currentStep'], 'step' : session['step']})\n\t\t\t\telse:\n\t\t\t\t\tstep = session['step']\n\t\t\t\t\tdisconnect()\n\t\t\t\t\treturn jsonify({'LString' : txtlstring, 'currentStep' : step, 'step' : step})\n\t\t\t\t\n\telse:\n\t\tl = lpy.Lsystem()\n\t\tsession['code'] = code\n\t\ttry:\n\t\t\tl.set(code)\n\t\texcept:\n\t\t\treturn jsonify({'error' : 'Syntax error'})\n\n\t\tsession['step'] = l.derivationLength\n\t\tsession['currentStep'] = 1\n\t\tlstring = l.derive(session['currentStep'])\n\t\tilstring = l.interpret(lstring)\n\t\ttxtlstring = str(ilstring)\n\t\tLSystem = l\n\t\treturn jsonify({'LString' : txtlstring, 'currentStep' : session['currentStep'], 'step' : session['step']})\n\n#When Rewind is clicked, the server receives the code, interpret only the Axiom of the LString and sends the result.\n@app.route('/rewind', methods=['POST'])\ndef rewind():\n\tdisconnect()\n\tl = lpy.Lsystem()\n\tcode = request.form['code']\n\tcode = code.encode('ascii', 'ignore')\n\ttry:\n\t\tl.set(code)\n\texcept:\n\t\treturn jsonify({'error' : 'Syntax error'})\n\tlstring = l.axiom\n\tilstring = l.interpret(lstring)\n\ttxtlstring = str(ilstring)\n\treturn jsonify({'LString' : txtlstring})\n\n@app.route('/about.html')\ndef about():\n return render_template('about.html')\n\n#Sometimes Flask doesn't interpret static files updates.\n#This function clears the issue.\n\n@app.context_processor\ndef override_url_for():\n return dict(url_for=dated_url_for)\n\ndef dated_url_for(endpoint, **values):\n if endpoint == 'static':\n filename = values.get('filename', None)\n if filename:\n file_path = os.path.join(app.root_path,\n endpoint, filename)\n values['q'] = int(os.stat(file_path).st_mtime)\n return url_for(endpoint, **values)\n\n#Fred's code\n#Converts curves and LStrings into JSON\n\n# def iscurve(c):\n# if isinstance(c, pgl.Curve2D) : return True\n# return isinstance(c, pgl.LineicModel)\n# \n# def curvetojson(c):\n# # calcule des points et les transforme en chaine de caractere\n# deltau = (c.lastKnot - c.firstKnot) / c.stride\n# return '['+','.join([str(tuple(c.getPointAt(c.firstKnot + i * deltau))) for i in range(c.stride+1)])+']'\n# \n# def lstringtojson(lstring):\n# lstrrepr = ''\n# for module in lstring:\n# modrepr = module.name +'(' + ','.join([str(param) if not iscurve(param) else curvetojson(param) for param in module])\n# lstrrepr += ' '+modrepr\n# return lstrrepr", "sub_path": "lpyweb.py", "file_name": "lpyweb.py", "file_ext": "py", "file_size_in_byte": 5162, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "flask.Flask", "line_number": 11, "usage_type": "call"}, {"api_name": "flask_socketio.SocketIO", "line_number": 14, "usage_type": "call"}, {"api_name": "threading.Lock", "line_number": 16, "usage_type": "call"}, {"api_name": "flask.session.permanent", "line_number": 20, "usage_type": "attribute"}, {"api_name": "flask.session", "line_number": 20, "usage_type": "name"}, {"api_name": "flask.session.pop", "line_number": 28, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 28, "usage_type": "name"}, {"api_name": "flask.session.pop", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 29, "usage_type": "name"}, {"api_name": "flask.session.pop", "line_number": 30, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 30, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 34, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 38, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 45, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 47, "usage_type": "attribute"}, {"api_name": "openalea.lpy.Lsystem", "line_number": 50, "usage_type": "call"}, {"api_name": "openalea.lpy", "line_number": 50, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 51, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 51, "usage_type": "name"}, {"api_name": "sys.stdout", "line_number": 58, "usage_type": "attribute"}, {"api_name": "flask.jsonify", "line_number": 61, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 69, "usage_type": "attribute"}, {"api_name": "flask.jsonify", "line_number": 73, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 80, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 80, "usage_type": "name"}, {"api_name": "flask.session.get", "line_number": 84, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 84, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 85, "usage_type": "name"}, {"api_name": "openalea.lpy.Lsystem", "line_number": 86, "usage_type": "call"}, {"api_name": "openalea.lpy", "line_number": 86, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 87, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 91, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 93, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 94, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 95, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 99, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 99, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 103, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 104, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 107, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 108, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 108, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 110, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 112, "usage_type": "call"}, {"api_name": "openalea.lpy.Lsystem", "line_number": 115, "usage_type": "call"}, {"api_name": "openalea.lpy", "line_number": 115, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 116, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 120, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 122, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 123, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 124, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 128, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 128, "usage_type": "name"}, {"api_name": "openalea.lpy.Lsystem", "line_number": 134, "usage_type": "call"}, {"api_name": "openalea.lpy", "line_number": 134, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 135, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 135, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 140, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 144, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 148, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 161, "usage_type": "call"}, {"api_name": "os.path", "line_number": 161, "usage_type": "attribute"}, {"api_name": "os.stat", "line_number": 163, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 164, "usage_type": "call"}]}
+{"seq_id": "15608717", "text": "\"\"\"\nSupport for SageTV players.\n\nFor more details about this platform, please refer to the documentation at\nhttps://home-assistant.io/components/SageTV/\n\"\"\"\nfrom datetime import timedelta\nimport logging\nimport requests\nimport voluptuous as vol\n\nfrom homeassistant.components.media_player import (\n MediaPlayerEntity, PLATFORM_SCHEMA)\nfrom homeassistant.components.media_player.const import (\n SUPPORT_PAUSE, SUPPORT_PLAY, SUPPORT_STOP,\n SUPPORT_TURN_OFF, SUPPORT_TURN_ON, SUPPORT_NEXT_TRACK, SUPPORT_PREVIOUS_TRACK, SUPPORT_SEEK)\nfrom homeassistant.const import (\n CONF_HOST, CONF_NAME, STATE_IDLE, STATE_OFF, STATE_PLAYING)\nimport homeassistant.helpers.config_validation as cv\nfrom homeassistant.util.dt import utcnow\n\nREQUIREMENTS = []\nCONF_EXTENDER = 'extender'\nCONF_SAGEX = 'sagex'\nDEFAULT_NAME = \"SageTV\"\nSCAN_INTERVAL = timedelta(seconds=30)\n\n_LOGGER = logging.getLogger(__name__)\n\nSUPPORT_SAGETV = (SUPPORT_PLAY | SUPPORT_STOP | SUPPORT_PAUSE | SUPPORT_NEXT_TRACK | SUPPORT_PREVIOUS_TRACK | SUPPORT_SEEK)\n# No host is needed for configuration, however it can be set.\nPLATFORM_SCHEMA = PLATFORM_SCHEMA.extend({\n vol.Required(CONF_SAGEX): cv.string,\n vol.Required(CONF_EXTENDER): cv.string,\n vol.Optional(CONF_NAME, default=DEFAULT_NAME): cv.string,\n})\n\n\ndef setup_platform(hass, config, add_entities, discovery_info=None):\n conf = discovery_info if discovery_info else config\n\n # Register configured device with Home Assistant.\n add_entities([SageTV( conf[CONF_NAME], conf[CONF_SAGEX], conf[CONF_EXTENDER])])\n\n\nclass SageTV(MediaPlayerEntity):\n\n def __init__(self, name, sagex, extender):\n\n # Default name value, only to be overridden by user.\n self._name = name\n self._extender = extender\n self._baseurl = sagex\n # Assume we're off to start with\n self._state = STATE_IDLE\n self._position = 0\n self._duration = 0\n self._position_valid = 0\n self._media_title = ''\n self._title = ''\n self._poster = ''\n self.update()\n\n @property\n def name(self):\n \"\"\"Return the display name of this device.\"\"\"\n return self._name\n\n @property\n def state(self):\n \"\"\"Return _state variable, containing the appropriate constant.\"\"\"\n return self._state\n\n @property\n def supported_features(self):\n \"\"\"Flag media player features that are supported.\"\"\"\n return SUPPORT_SAGETV\n\n @property\n def media_duration(self):\n \"\"\"Duration of current playing media in seconds.\"\"\"\n return self._duration\n\n @property\n def media_position(self):\n \"\"\"Position of current playing media in seconds.\"\"\"\n return self._position\n\n @property\n def media_position_updated_at(self):\n \"\"\"When was the position of the current playing media valid.\"\"\"\n return self._position_valid\n\n def update(self):\n \"\"\"Update the internal state by querying the device.\"\"\"\n # This can take 5+ seconds to complete\n url = self._baseurl + 'sagex/api?c=ha:GetCurrentShow&1=' + self._extender + '&encoder=json'\n r = requests.get(url)\n rawJson = r.json()\n self._state = rawJson[\"Result\"][\"isPlaying\"]\n if self._state == 'idle':\n self._media_title = ''\n self._poster = ''\n self._duration = 0\n self._position = 0\n else:\n if self._media_title != rawJson[\"Result\"][\"title\"]:\n self._media_title = rawJson[\"Result\"][\"title\"]\n self._poster = rawJson[\"Result\"][\"poster\"]\n self._duration = rawJson[\"Result\"][\"duration\"]\n self._position = rawJson[\"Result\"][\"watchedDuration\"]\n self._position_valid = utcnow()\n\n def media_play(self):\n \"\"\"Send play command.\"\"\"\n url = self._baseurl + 'sagex/api?c=ha:Command&1=play&2=' + self._extender\n r = requests.get(url)\n\n def media_pause(self):\n \"\"\"Send pause command.\"\"\"\n url = self._baseurl + 'sagex/api?c=ha:Command&1=pause&2=' + self._extender\n r = requests.get(url)\n\n def media_stop(self):\n \"\"\"Send stop command.\"\"\"\n url = self._baseurl + 'sagex/api?c=ha:Command&1=stop&2=' + self._extender\n r = requests.get(url)\n\n def media_play_pause(self):\n url = self._baseurl + 'sagex/api?c=ha:PlayPause&1=' + self._extender\n r = requests.get(url)\n\n def media_next_track(self):\n url = self._baseurl + 'sagex/api?c=ha:Command&1=Right&2=' + self._extender\n r = requests.get(url)\n\n def media_previous_track(self):\n url = self._baseurl + 'sagex/api?c=ha:Command&1=Left&2=' + self._extender\n r = requests.get(url)\n\n def async_media_seek(self,position):\n url = self._baseurl + 'sagex/api?c=ha:Seek&1=' + self._extender + '&2=' + position;\n r = requests.get(url)\n\n @property\n def media_title(self):\n \"\"\"Title of current playing media.\"\"\"\n # find a string we can use as a title\n return self._media_title\n\n @property\n def media_image_url(self):\n return self._poster;\n", "sub_path": "sagetv/media_player.py", "file_name": "media_player.py", "file_ext": "py", "file_size_in_byte": 5101, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "datetime.timedelta", "line_number": 26, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 28, "usage_type": "call"}, {"api_name": "homeassistant.components.media_player.const.SUPPORT_PLAY", "line_number": 30, "usage_type": "name"}, {"api_name": "homeassistant.components.media_player.const.SUPPORT_STOP", "line_number": 30, "usage_type": "name"}, {"api_name": "homeassistant.components.media_player.const.SUPPORT_PAUSE", "line_number": 30, "usage_type": "name"}, {"api_name": "homeassistant.components.media_player.const.SUPPORT_NEXT_TRACK", "line_number": 30, "usage_type": "name"}, {"api_name": "homeassistant.components.media_player.const.SUPPORT_PREVIOUS_TRACK", "line_number": 30, "usage_type": "name"}, {"api_name": "homeassistant.components.media_player.const.SUPPORT_SEEK", "line_number": 30, "usage_type": "name"}, {"api_name": "homeassistant.components.media_player.PLATFORM_SCHEMA", "line_number": 32, "usage_type": "name"}, {"api_name": "homeassistant.components.media_player.PLATFORM_SCHEMA.extend", "line_number": 32, "usage_type": "call"}, {"api_name": "voluptuous.Required", "line_number": 33, "usage_type": "call"}, {"api_name": "voluptuous.Required", "line_number": 34, "usage_type": "call"}, {"api_name": "voluptuous.Optional", "line_number": 35, "usage_type": "call"}, {"api_name": "homeassistant.const.CONF_NAME", "line_number": 35, "usage_type": "argument"}, {"api_name": "homeassistant.helpers.config_validation.string", "line_number": 33, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 33, "usage_type": "name"}, {"api_name": "homeassistant.helpers.config_validation.string", "line_number": 34, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 34, "usage_type": "name"}, {"api_name": "homeassistant.helpers.config_validation.string", "line_number": 35, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 35, "usage_type": "name"}, {"api_name": "homeassistant.const.CONF_NAME", "line_number": 43, "usage_type": "name"}, {"api_name": "homeassistant.components.media_player.MediaPlayerEntity", "line_number": 46, "usage_type": "name"}, {"api_name": "homeassistant.const.STATE_IDLE", "line_number": 55, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 98, "usage_type": "call"}, {"api_name": "homeassistant.util.dt.utcnow", "line_number": 112, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 117, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 122, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 127, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 131, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 135, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 139, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 143, "usage_type": "call"}]}
+{"seq_id": "357024388", "text": "# IMPORTS AND SETUP\n\n# General imports\nimport multiprocessing\nimport numpy as np\nimport pandas as pd\nimport time\nimport sys\nimport dill\nimport warnings\nimport matplotlib.pyplot as plt\nimport seaborn as sns\nfrom scipy.stats import pearsonr\nimport collections\nimport os\n\n# Sklearn imports\nfrom sklearn.preprocessing import StandardScaler\nfrom sklearn.ensemble import RandomForestRegressor\nfrom sklearn import model_selection\nfrom sklearn.pipeline import Pipeline\nfrom sklearn import metrics\nfrom sklearn.dummy import DummyRegressor\nfrom sklearn.linear_model import Lasso, ElasticNet\nfrom stability_selection import StabilitySelection\n\n# Add directory to sys.path in order to import custom modules from there.\nsys.path.insert(0, \"/home/kkoras/Documents/Projects/Doktorat - Modelling drug efficacy in cancer/Projects/Created Modules\")\nfrom gdsc_projects_module import DrugGenomeWide, Experiment, Modeling, ModelingResults\n\n# LOAD DATA\n# Initialize proper file pathways\ndrug_annotations = \"/home/kkoras/Documents/Projects/Doktorat - Modelling drug efficacy in cancer/Data/Original Data/Genomics of Drug Sensitivity in Cancer/Original GDSC Data/Drug annotations/Screened_Compounds-March_27th_2018.xlsx\"\ncell_line_list = \"/home/kkoras/Documents/Projects/Doktorat - Modelling drug efficacy in cancer/Data/Original Data/Genomics of Drug Sensitivity in Cancer/Original GDSC Data/Cell line list (directly from website)/Cell_listThu Aug 16 22_06_49 2018.csv\"\ngene_expr = \"/home/kkoras/Documents/Projects/Doktorat - Modelling drug efficacy in cancer/Data/Original Data/Genomics of Drug Sensitivity in Cancer/Original GDSC Data/Gene expression/sanger1018_brainarray_ensemblgene_rma-March_2nd_2017.txt\"\ncnv1 = \"/home/kkoras/Documents/Projects/Doktorat - Modelling drug efficacy in cancer/Data/Original Data/Genomics of Drug Sensitivity in Cancer/Original GDSC Data/Copy number variations/cnv_binary_1.csv\"\ncnv2 = \"/home/kkoras/Documents/Projects/Doktorat - Modelling drug efficacy in cancer/Data/Original Data/Genomics of Drug Sensitivity in Cancer/Original GDSC Data/Copy number variations/PANCANCER_Genetic_feature_cna_Mon Aug 6 16_18_51 2018 (kopia).csv\"\ncoding_variants = \"/home/kkoras/Documents/Projects/Doktorat - Modelling drug efficacy in cancer/Data/Original Data/Genomics of Drug Sensitivity in Cancer/Original GDSC Data/Mutation calls/PANCANCER_Genetic_feature_variant_Mon Aug 6 15_45_44 2018.csv\"\ndrug_response = \"/home/kkoras/Documents/Projects/Doktorat - Modelling drug efficacy in cancer/Data/Original Data/Genomics of Drug Sensitivity in Cancer/Original GDSC Data/Sensitivity profiles/v17.3_fitted_dose_response-March_27th_2018.xlsx\"\n\n# Call loading function from DrugGenomeWide class\n(drug_annotations_df, cell_lines_list_df, gene_expression_df, cnv_binary_df, \n coding_variants_df, drug_response_df) = DrugGenomeWide.load_data(\n drug_annotations, cell_line_list, gene_expr, \n cnv1, cnv2, coding_variants, drug_response)\n\n# Load gene mappings\nfilepath1 = \"/home/kkoras/Documents/Projects/Doktorat - Modelling drug efficacy in cancer/Projects/GDSC - Prediction only with data related to nominal drug targets (minimal approach)/Created data/mapping_from_ensembl_id_to_hgnc_symbol.p\"\nfilepath2 = \"/home/kkoras/Documents/Projects/Doktorat - Modelling drug efficacy in cancer/Projects/GDSC - Prediction only with data related to nominal drug targets (minimal approach)/Created data/mapping_from_hgnc_symbol_to_ensembl_id.p\"\nDrugGenomeWide.load_mappings(filepath2, filepath1) # Initialize class variables\n\n# Print shapes of created DataFrames\nprint(\"Loading summary:\")\nprint(\"Drug annotations:\", drug_annotations_df.shape)\nprint(\"Cell line list\", cell_lines_list_df.shape)\nprint(\"Gene expression\", gene_expression_df.shape)\nprint(\"CNV binary:\", cnv_binary_df.shape)\nprint(\"Coding variants:\", coding_variants_df.shape)\nprint(\"Drug response:\", drug_response_df.shape)\n\n# CREATE A DICTIONARY WITH DRUG OBJECTS\ndrugs = DrugGenomeWide.create_drugs(drug_annotations_df)\nprint(len(drugs))\n\n# MODELING SETUP\n\n# Hyperparameter space to search on\nparam_grid = {\"estimator__alpha\": [0.001, 0.01, 0.1, 1., 5., 10., 30., 50., 100., 300.],\n \"estimator__l1_ratio\": [0.0, 0.1, 0.25, 0.5, 0.75, 0.9, 1.]}\n\n\n# Create ModelingWithFeatureSelecction object\nenet_seeds = [22, 37, 44, 55, 78]\nsplit_seeds = [11, 37, 52, 71, 98]\n\n\nenet_exp = Modeling(name=\"ElasticNet genome wide without feature selection\",\n param_grid=param_grid,\n estimator_seeds=enet_seeds,\n split_seeds=split_seeds,\n n_combinations=30,\n kfolds=3,\n tuning_jobs=12,\n max_iter=2500)\n\n# Initialize new ModelingResults object\nexp_results = ModelingResults(enet_exp)\n\n# Load current ModelingResults object\n#with open(\"../Created data/Results/genome_wide-enet_over_few_data_splits_without_feature_selection.pkl\", \"rb\") as f:\n #exp_results = dill.load(f)\n\n# ITERATE OVER DRUGS\n\n# Get rid of warnings\nimport warnings\nwarnings.filterwarnings(\"ignore\")\n\ndrug_counter = 0\nlog = True # Controls verbosity during iterating over drugs\n\nfor drug_id in drugs:\n # Current drug object\n drug = drugs[drug_id]\n if (drug.name, drug_id) in exp_results.cv_results_dict:\n continue\n if drug.name == \"Camptothecin\":\n continue\n \n \n # Extract full data\n data = drug.return_full_data(drug_response_df, gene_expression_df, data_combination=[\"expression\"])\n \n if log:\n print(drug.name, data.shape)\n\n # Extract features and labels\n y = data[\"AUC\"]\n X = data.drop([\"cell_line_id\", \"AUC\"], axis=1)\n assert X.shape[1] == 17737 and X.shape[0] == y.shape[0]\n \n # Add data shapes to corresponding dictionary field in ModelingResults\n exp_results.data_shapes[(drug.name, drug_id)] = X.shape\n \n # Compute the results\n (test_results_for_splits, cv_results_for_splits, \n best_parameters_for_splits, \n dummy_for_splits, tuning_seeds_for_splits) = enet_exp.enet_model_over_data_splits(X, y, verbose=2, log=True)\n \n # Put results into appropriate fields of ModelingResults object\n exp_results.performance_dict[(drug.name, drug_id)] = test_results_for_splits\n exp_results.dummy_performance_dict[(drug.name, drug_id)] = dummy_for_splits\n exp_results.best_params_dict[(drug.name, drug_id)] = best_parameters_for_splits\n exp_results.tuning_seeds_dict[(drug.name, drug_id)] = tuning_seeds_for_splits\n exp_results.cv_results_dict[(drug.name, drug_id)] = cv_results_for_splits\n \n # Save the results\n #res_name = exp_results.name.replace(\" \", \"_\").lower() + \".pkl\n with open(\"../Created data/Results/genome_wide-enet_over_few_data_splits_without_feature_selection.pkl\", \"wb\") as f:\n dill.dump(exp_results, f)\n \n drug_counter +=1 \n print(drug_counter, \"drugs done\")\n print()\n print(\"*\" * 50)\n print()\n \nprint()\nprint(\"SCRIPT FINISHED, ALL DRUGS DONE\")\nprint()\n", "sub_path": "GDSC - Prediction with genome wide gene expression/Scripts/enet_modeling_without_feat_selection_run_script.py", "file_name": "enet_modeling_without_feat_selection_run_script.py", "file_ext": "py", "file_size_in_byte": 6921, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sys.path.insert", "line_number": 28, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "gdsc_projects_module.DrugGenomeWide.load_data", "line_number": 43, "usage_type": "call"}, {"api_name": "gdsc_projects_module.DrugGenomeWide", "line_number": 43, "usage_type": "name"}, {"api_name": "gdsc_projects_module.DrugGenomeWide.load_mappings", "line_number": 50, "usage_type": "call"}, {"api_name": "gdsc_projects_module.DrugGenomeWide", "line_number": 50, "usage_type": "name"}, {"api_name": "gdsc_projects_module.DrugGenomeWide.create_drugs", "line_number": 62, "usage_type": "call"}, {"api_name": "gdsc_projects_module.DrugGenomeWide", "line_number": 62, "usage_type": "name"}, {"api_name": "gdsc_projects_module.Modeling", "line_number": 77, "usage_type": "call"}, {"api_name": "gdsc_projects_module.ModelingResults", "line_number": 87, "usage_type": "call"}, {"api_name": "warnings.filterwarnings", "line_number": 97, "usage_type": "call"}, {"api_name": "dill.dump", "line_number": 140, "usage_type": "call"}]}
+{"seq_id": "288769865", "text": "# %matplotlib inline\nimport sys\nimport math\n\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport matplotlib as mpl\nimport matplotlib\nimport numpy as np\nimport os\nfrom matplotlib.backends.backend_pdf import PdfPages\nfrom matplotlib.ticker import MultipleLocator\nfrom random import randint\nfrom matplotlib.backends.backend_pdf import PdfPages\nfrom matplotlib.backends.backend_pdf import PdfPages\nfrom palettable.colorbrewer.sequential import YlGnBu_5\n\nmodel_names = ['mobilenet','xception','nasnetamobile']\nmodel_name = model_names[1]\n# model_name = 'pdp'\ntotal_mac = '800'\ndata_folder = './gen_data/plot_data/' + model_name + '/' + total_mac + '/latency.csv'\nresult_dir = './gen_data/plot_data/' + model_name + '/' + total_mac + '/'\nif not os.path.exists(result_dir):\n os.makedirs(result_dir)\nresult_file = result_dir + '/latency_scatter_' + model_name + '.pdf'\n# Plot type\nplt.style.use('ggplot')\n\n#Loading data\ndf = pd.read_csv(data_folder)\n\ndf['total_mac_cycles'] = df['total_mac_cycles'] / df['total_mac_cycles'].max()\ndf['total_dma_accesses'] = df['total_dma_accesses'] / df['total_dma_accesses'].max()\ndf['total_memory_mb'] = df['total_memory_mb'] / df['total_memory_mb'].max()\ndf['total_memory_mb'] = [math.pow(x,6) for x in df['total_memory_mb']]\n\n# delete rows with column name 'schedule_name' = pdp\nindexNames = df[df['schedule_name'] == 'pdp'].index\n# Delete these row indexes from dataFrame\ndf.drop(indexNames, inplace=True)\ndf['schedule_name'] = df['schedule_name'].replace(['per_layer','dp','pdp'],['Base','DP-PT','PT-DP-PT'])\n\nfig, ax = plt.subplots(figsize=(12, 8))\nplt.gcf().subplots_adjust(bottom=0.16, right=0.99, top=0.88)\nplt.xticks(rotation=90)\n\n\n# plt.xlim(-0.5, 10)\n\nN = len(df[model_name])\nind = np.arange(N)\n# width = 0.8\n\ndf_sorted = df.sort_values(by=['total_dma_accesses'])\n\n\nmajorLocator = MultipleLocator(0.2)\nminorLocator = MultipleLocator(0.2)\n# majorFormatter = FormatStrFormatter('%d')\nax.xaxis.set_major_locator(majorLocator)\nax.xaxis.set_minor_locator(minorLocator)\n# ax.xaxis.set_major_formatter(majorFormatter)\n\n\n# cl = {'PT-DP-PT':'blue','DP-PT':'orange','Base':'green'}\ncl = {'PT-DP-PT':'blue','DP-PT':'red','Base':'green'}\nmarker = {'DP-PT':'x','Base':'+'}\n\nfor sch in list(set(df_sorted['schedule_name'])):\n vals = df_sorted[df_sorted['schedule_name'] == sch]\n ax.scatter(vals['total_dma_accesses'],vals['total_mac_cycles'],\n s=1000, marker = marker[sch],edgecolor='k', color=cl[sch],label=sch,\n alpha=1, edgecolors='k')\n\n # ax.scatter(vals['total_dma_accesses'], vals['total_mac_cycles'],\n # s=vals['total_memory_mb']*1024*3, marker = 'o',edgecolor='k', color=cl[sch],label=sch,\n # alpha=0.3, edgecolors='none')\n\n# for n in np.arange(len(df['name'])):\n# ax.text(df['Area'][n]-0.06, df['Power'][n]-0.07, df['name'][n], fontsize=16)\n\n# Put limit on Y axis\nplt.ylim(0, 1.1)\nplt.xlim(0.01, 1.1)\n\n# Set X label values\nax.set_ylabel('Latency (Lower is better)',fontsize=38, color='black')\nax.set_xlabel('DRAM Energy (Lower is better)',fontsize=38, color='black')\n\n# ax.set_xticks(ind+0.8);\n\n# Put the labels from 'app' coulmn\n# ax.set_xticklabels(u.rename(df['name']))\nax.set_facecolor('whitesmoke')\n\nplt.gca().xaxis.grid(True, color='black')\nplt.gca().yaxis.grid(True, color='black')\n\nplt.tick_params( axis='x', which='both', bottom=False, top=False, colors='black')\nplt.tick_params( axis='y', which='both', right=False, colors='black' )\nplt.tick_params(axis='both', which='major', direction='in', \n length=6, width=3,color='black', labelsize=28)\nplt.grid(linestyle='--')\n\n# plt.legend(numpoints=1)\nplt.legend(bbox_to_anchor=(0.90, 1.20), ncol=3, fontsize=35)\n\n\n\nax.spines['bottom'].set_color('gray')\nax.spines['top'].set_color('gray')\nax.spines['right'].set_color('gray')\nax.spines['left'].set_color('gray')\n\nplt.show()\n# Saving the plot\nfig.savefig(result_file,facecolor=fig.get_facecolor(), bbox_inches='tight')", "sub_path": "TB-scheduler/deprecated/plot_dp_latency_scatter.py", "file_name": "plot_dp_latency_scatter.py", "file_ext": "py", "file_size_in_byte": 3946, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.path.exists", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 28, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 31, "usage_type": "call"}, {"api_name": "math.pow", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.ticker.MultipleLocator", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.ticker.MultipleLocator", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 97, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tick_params", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tick_params", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tick_params", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 116, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 116, "usage_type": "name"}]}
+{"seq_id": "270109451", "text": "from telegram.ext import CallbackContext\n\nfrom .string import strings\n\nget_string = strings.get_string\nget_languages = strings.get_languages\n\n\ndef lang_(func):\n def wrapper(update, context):\n return func(\n update,\n context,\n context.bot_data.get(\n update.effective_chat.id, {}\n ).get(\n \"lang\", \"en\"\n )\n )\n return wrapper\n\n\ndef _lang(context: CallbackContext, chat_id: int):\n return context.bot_data.get(\n chat_id, {}\n ).get(\n \"lang\", \"en\"\n )\n", "sub_path": "strings/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 570, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "string.strings.get_string", "line_number": 5, "usage_type": "attribute"}, {"api_name": "string.strings", "line_number": 5, "usage_type": "name"}, {"api_name": "string.strings.get_languages", "line_number": 6, "usage_type": "attribute"}, {"api_name": "string.strings", "line_number": 6, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 23, "usage_type": "name"}]}
+{"seq_id": "159778518", "text": "from flask import Flask, render_template, request\nfrom models.models import PostIt\nfrom models.database import db_session\nfrom datetime import datetime\n\napp = Flask(__name__)\n\n@app.route('/')\ndef index():\n all_post_it = PostIt.query.all()\n return render_template('index.html', postits=all_post_it)\n\n@app.route('/add', methods=['post'])\ndef add():\n title = request.form['title']\n content = PostIt(title, datetime.now())\n print(f'{title}のPOSTを確認')\n db_session.add(content)\n db_session.commit()\n return index()\n\nif __name__ == '__main__':\n app.debug = True\n app.run(host='0.0.0.0', port=8080)", "sub_path": "TrelloLike/server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 627, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "flask.Flask", "line_number": 6, "usage_type": "call"}, {"api_name": "models.models.PostIt.query.all", "line_number": 10, "usage_type": "call"}, {"api_name": "models.models.PostIt.query", "line_number": 10, "usage_type": "attribute"}, {"api_name": "models.models.PostIt", "line_number": 10, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 11, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 15, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 15, "usage_type": "name"}, {"api_name": "models.models.PostIt", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 16, "usage_type": "name"}, {"api_name": "models.database.db_session.add", "line_number": 18, "usage_type": "call"}, {"api_name": "models.database.db_session", "line_number": 18, "usage_type": "name"}, {"api_name": "models.database.db_session.commit", "line_number": 19, "usage_type": "call"}, {"api_name": "models.database.db_session", "line_number": 19, "usage_type": "name"}]}
+{"seq_id": "603802426", "text": "# coding=utf-8\n# --------------------------------------------------------------------------\n# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License. See License.txt in the project root for\n# license information.\n#\n# Code generated by Microsoft (R) AutoRest Code Generator.\n# Changes may cause incorrect behavior and will be lost if the code is\n# regenerated.\n# --------------------------------------------------------------------------\n\nfrom msrest.serialization import Model\n\n\nclass DeletedVaultProperties(Model):\n \"\"\"Properties of the deleted vault.\n\n Variables are only populated by the server, and will be ignored when\n sending a request.\n\n :ivar vault_id: The resource id of the original vault.\n :vartype vault_id: str\n :ivar location: The location of the original vault.\n :vartype location: str\n :ivar deletion_date: The deleted date.\n :vartype deletion_date: datetime\n :ivar scheduled_purge_date: The scheduled purged date.\n :vartype scheduled_purge_date: datetime\n :ivar tags: Tags of the original vault.\n :vartype tags: dict[str, str]\n \"\"\"\n\n _validation = {\n 'vault_id': {'readonly': True},\n 'location': {'readonly': True},\n 'deletion_date': {'readonly': True},\n 'scheduled_purge_date': {'readonly': True},\n 'tags': {'readonly': True},\n }\n\n _attribute_map = {\n 'vault_id': {'key': 'vaultId', 'type': 'str'},\n 'location': {'key': 'location', 'type': 'str'},\n 'deletion_date': {'key': 'deletionDate', 'type': 'iso-8601'},\n 'scheduled_purge_date': {'key': 'scheduledPurgeDate', 'type': 'iso-8601'},\n 'tags': {'key': 'tags', 'type': '{str}'},\n }\n\n def __init__(self, **kwargs) -> None:\n super(DeletedVaultProperties, self).__init__(**kwargs)\n self.vault_id = None\n self.location = None\n self.deletion_date = None\n self.scheduled_purge_date = None\n self.tags = None\n", "sub_path": "azext_keyvault/mgmt/keyvault/models/deleted_vault_properties_py3.py", "file_name": "deleted_vault_properties_py3.py", "file_ext": "py", "file_size_in_byte": 1976, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "msrest.serialization.Model", "line_number": 15, "usage_type": "name"}]}
+{"seq_id": "418347187", "text": "import telebot\r\nimport pyowm \r\n\r\nowm = pyowm.OWM('affe09dd9ba0dc635779ed17bdf11f38', language = \"ru\")\r\nbot = telebot.TeleBot(\"1089654967:AAGuzKfKrH1SEEstj3uIPbBMLOf3CYV38Io\")\r\n\r\n@bot.message_handler(content_types=['text'])\r\ndef send_echo(message):\r\n\tobservation = owm.weather_at_place( message.text )\r\n\tw = observation.get_weather()\r\n\ttemp = w.get_temperature('celsius')[\"temp\"]\r\n\r\n\tanswer = \"В городе \" + message.text + \" сейчас \" + w.get_detailed_status() + \"\\n\"\r\n\tanswer += \"Температура сейчас в районе \" + str(temp) + \"\\n\\n\"\r\n\tif temp < -5:\r\n\t\tanswer += \"Сейчас очень холодно, одевайся как танк!\"\r\n\telif (temp < 10)&(temp > 0):\r\n\t\tanswer += \"Сейчас холодно, оденься потеплее. \"\r\n\telif (temp < -5)&(temp > -1):\r\n\t\tanswer += \"Сейчас холодно, но не так холодно как при меньше нуля, но я не знаю как это у тебя там работает. \"\r\n\telif temp < 20:\r\n\t\tanswer += \"Специальное веселое значение\"\r\n\telse:\r\n\t\tanswer += \"Температура норм, одевай платье. \"\r\n\r\n\tbot.send_message(message.chat.id, answer)\r\n\r\nbot.polling( none_stop= True )", "sub_path": "telegapogoda.py", "file_name": "telegapogoda.py", "file_ext": "py", "file_size_in_byte": 1250, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pyowm.OWM", "line_number": 4, "usage_type": "call"}, {"api_name": "telebot.TeleBot", "line_number": 5, "usage_type": "call"}]}
+{"seq_id": "515525541", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('emailmanager', '0009_auto_20150604_0057'),\n ]\n\n operations = [\n migrations.AddField(\n model_name='email',\n name='email_type',\n field=models.CharField(default='Action', max_length=80),\n preserve_default=False,\n ),\n migrations.AlterField(\n model_name='email',\n name='button',\n field=models.CharField(default=b'Take Email', max_length=50, verbose_name=b'button text'),\n ),\n migrations.AlterField(\n model_name='email',\n name='email_url',\n field=models.CharField(max_length=150),\n ),\n migrations.AlterField(\n model_name='email',\n name='full_name',\n field=models.CharField(max_length=80, verbose_name=b'full name', blank=True),\n ),\n migrations.AlterField(\n model_name='email',\n name='img_credit',\n field=models.CharField(max_length=80, verbose_name=b'photo credit', blank=True),\n ),\n migrations.AlterField(\n model_name='email',\n name='job_title',\n field=models.CharField(max_length=80, verbose_name=b'job title', blank=True),\n ),\n migrations.AlterField(\n model_name='email',\n name='lift_note',\n field=models.TextField(blank=True),\n ),\n migrations.AlterField(\n model_name='email',\n name='office',\n field=models.CharField(max_length=80, blank=True),\n ),\n migrations.AlterField(\n model_name='email',\n name='teaser',\n field=models.CharField(max_length=200, null=True, blank=True),\n ),\n migrations.AlterField(\n model_name='email',\n name='title',\n field=models.CharField(default=b'', max_length=150),\n ),\n migrations.AlterField(\n model_name='email',\n name='valedic',\n field=models.CharField(default=b'Thank you for speaking out,', max_length=80, verbose_name=b'valediction'),\n ),\n ]\n", "sub_path": "emailmanager/migrations/0010_auto_20150604_1519.py", "file_name": "0010_auto_20150604_1519.py", "file_ext": "py", "file_size_in_byte": 2287, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 28, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 28, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 33, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 35, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 35, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 38, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 38, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 40, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 40, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 43, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 43, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 45, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 45, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 48, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 48, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 50, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 50, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 53, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 53, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 55, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 55, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 58, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 58, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 60, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 60, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 63, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 63, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 65, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 65, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 68, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 68, "usage_type": "name"}]}
+{"seq_id": "187182517", "text": "from xml.etree import ElementTree as et\r\n'''Чтение из xml'''\r\ntree = et.parse('test_xml.xml')\r\nroot = tree.getroot()\r\n'''children = root.getchildren()'''\r\n\r\nfor group in root:\r\n print(\"Group: \", group.attrib)\r\n for student in group:\r\n print('{}: {}'.format(student.tag, student.text))\r\n'''Создание xml'''\r\nroot = et.Element('MAIN_ELEM')\r\nfor i in range(10):\r\n sub_element = et.SubElement(root, 'value{}'.format(i))\r\n sub_element.text = str(i * 10)\r\nprint(et.dump(root))\r\n\r\ndata = [\r\n {'x': 10, 'y': 20, 'z': 30},\r\n {'x': 'zello', 'y': 40, 'z': True}\r\n]\r\nroot = et.Element('records')\r\nfor item in data:\r\n record = et.SubElement(root, 'record')\r\n for key, value in item.items():\r\n e = et.SubElement(record, key)\r\n e.text = str(value)\r\ntree = et.ElementTree(root)\r\ntree.write('test_xml_created.xml', encoding='utf-8')", "sub_path": "Base_learning/pyproj_18_xml.py", "file_name": "pyproj_18_xml.py", "file_ext": "py", "file_size_in_byte": 875, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "xml.etree.ElementTree.parse", "line_number": 3, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 3, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 12, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 12, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 14, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 14, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.dump", "line_number": 16, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 16, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 22, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 22, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 24, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 24, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.SubElement", "line_number": 26, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 26, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.ElementTree", "line_number": 28, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 28, "usage_type": "name"}]}
+{"seq_id": "399095767", "text": "import pymysql\n\n\n\nhost = 'rm-2ze1w9epp05wvc4j4po.mysql.rds.aliyuncs.com'\nuser = 'test_admin'\npwd = '1qaz@WSX'\ndb = 'mc_client246'\nsql = 'select * from t_activity WHERE id = \\'8a69c7976e35b645016eac6a2ca5247c\\''\n\n\n\ndef test_mydb(host,user,pwd,db,sql,port=3306,charset='utf8'):\n con = pymysql.connect(host=host,port=port,user=user,passwd=pwd,db=db,charset = charset)\n cur = con.cursor()\n cur.execute(sql)\n\n if sql.strip()[:6].upper() == 'SELECT':\n res = cur.fetchall()\n\n else:\n con.commit()\n res ='ok'\n\n cur.close()\n con.close()\n print(res)\n return res\n\nprint(test_mydb(host,user,pwd,db,sql))", "sub_path": "db_fixture/test_db.py", "file_name": "test_db.py", "file_ext": "py", "file_size_in_byte": 638, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pymysql.connect", "line_number": 14, "usage_type": "call"}]}
+{"seq_id": "118699490", "text": "from flask import Flask\nfrom flask_cors import CORS\n\nfrom config import config\nfrom routes import users\n\n\ndef create_app(config_name):\n app = Flask(\n __name__, static_folder=\"../build/static\", template_folder=\"../build\"\n )\n app.config.from_object(config[config_name])\n config[config_name].init_app(app)\n register_additional_extensions(app)\n register_blueprint(app)\n\n return app\n\n\ndef register_additional_extensions(app):\n \"\"\"Register additional Flask extensions\"\"\"\n CORS(app)\n\ndef register_blueprint(app):\n \"\"\"Register Flask blueprints.\"\"\"\n app.register_blueprint(users.usersprint, url_prefix=\"/api/users\")\n return None\n", "sub_path": "labellab-flask/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 663, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "flask.Flask", "line_number": 9, "usage_type": "call"}, {"api_name": "config.config", "line_number": 12, "usage_type": "name"}, {"api_name": "config.config", "line_number": 13, "usage_type": "name"}, {"api_name": "flask_cors.CORS", "line_number": 22, "usage_type": "call"}, {"api_name": "routes.users.usersprint", "line_number": 26, "usage_type": "attribute"}, {"api_name": "routes.users", "line_number": 26, "usage_type": "name"}]}
+{"seq_id": "291449333", "text": "import numpy as np\nfrom sklearn import datasets\nboston=datasets.load_boston()\nx=boston.data\ny=boston.target\nx=x[y<50]\ny=y[y<50]\n\n\n#随机梯度下降法 多元线性回归训练波士顿房价数据\n\nfrom sklearn.model_selection import train_test_split\nx_train ,x_test,y_train,y_test=train_test_split(x,y,random_state=666)\n\n#数据预处理 数据标准化\nfrom sklearn.preprocessing import StandardScaler\nstandscaler=StandardScaler()\nstandscaler.fit(x_train)\nx_train_standard=standscaler.transform(x_train)\nx_test_standard=standscaler.transform(x_test)\n\n\n\n\n#导入线性模型中 随机梯度法 线性回归\n\nfrom sklearn.linear_model import SGDRegressor\nsgd_reg=SGDRegressor(n_iter=500)\nsgd_reg.fit(x_train_standard,y_train)\nscore=sgd_reg.score(x_test_standard,y_test)\nprint(score)", "sub_path": "SGD.py", "file_name": "SGD.py", "file_ext": "py", "file_size_in_byte": 787, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sklearn.datasets.load_boston", "line_number": 3, "usage_type": "call"}, {"api_name": "sklearn.datasets", "line_number": 3, "usage_type": "name"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 13, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 17, "usage_type": "call"}, {"api_name": "sklearn.linear_model.SGDRegressor", "line_number": 28, "usage_type": "call"}]}
+{"seq_id": "328477057", "text": "#!/usr/bin/python3\n\nfrom collections import namedtuple\nimport json\nimport logging\nimport math\nimport nltk\nimport numpy as np\n\nimport anagram\nimport clue as c\nimport definition\nimport synonym\n\n\nFeature = namedtuple(\"Feature\", [\"defn\", \"anagrind\", \"anagrist\", \"synonym\", \"lingo\"])\n\n\nEPSILON = 0.0001\n\nDEFINITION_INDEX = 0\nANAGRIND_INDEX = 1\nANAGRIST_INDEX = 2\nSYNONYM_INDEX = 3\nCHAFF_INDEX = 4\n\n\ndef rate_anagrind(word):\n ANAGRIND_BASES = [\"change\", \"destroy\", \"confuse\"]\n matches = [definition.word_similarity(word, grind)\n for grind in ANAGRIND_BASES]\n return max(matches)\n\n\ndef rate_lingo(word, solution):\n best_freq = 0\n with open(\"xword_lingo.freqs\", \"r\") as f:\n for line in f:\n lingo, char, freq = line.split()\n if word.lower() == lingo:\n if char in solution:\n best_freq = max([float(freq), best_freq])\n return best_freq\n\n\ndef get_features_word(word, solution, index, clue_length):\n f_defn = definition.word_similarity(word, solution)\n f_anagrind = rate_anagrind(word)\n f_anagrist, _, a_bitmask = anagram.is_possible_anagrist(word, solution)\n f_synonym, _, s_bitmask = synonym.is_synonym_in_solution(word, solution)\n f_lingo = rate_lingo(word, solution)\n f_length = len(word)\n f_position = (4/pow(clue_length, 2)) * pow((index - clue_length/2), 2)\n \n f = Feature(\n f_defn * f_position,\n f_anagrind,\n int(f_anagrist) * f_length,\n int(f_synonym) * f_length,\n f_lingo)\n #print(word, f.defn, f.anagrind, f.anagrist, f.synonym)\n return f, a_bitmask, s_bitmask\n\n\ndef get_features_clue(clue, solution):\n #import pdb; pdb.set_trace()\n words = nltk.word_tokenize(clue)\n index = 0\n record = \"\"\n for word in words:\n f, a_bitmask, s_bitmask = get_features_word(word, solution, index, len(words))\n index += 1\n record += \"{} {} {} {} {} {} {} {} \".format(\n word, f.defn, f.anagrind, f.anagrist, a_bitmask, f.synonym, s_bitmask, f.lingo)\n record += solution.replace(\" \",\"\")\n\n return record\n\n\ndef extract_features():\n with open(\"clues.txt\", \"r\") as f:\n with open(\"stuff.txt\", \"a\") as stuff:\n for line in f:\n record = json.loads(line)\n clue = record['clue']\n solution = record['solution']\n record = get_features_clue(clue, solution)\n stuff.write(record + \"\\n\")\n\n\nif __name__ == \"__main__\":\n extract_features()\n", "sub_path": "parser/features.py", "file_name": "features.py", "file_ext": "py", "file_size_in_byte": 2537, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "collections.namedtuple", "line_number": 16, "usage_type": "call"}, {"api_name": "definition.word_similarity", "line_number": 30, "usage_type": "call"}, {"api_name": "definition.word_similarity", "line_number": 47, "usage_type": "call"}, {"api_name": "anagram.is_possible_anagrist", "line_number": 49, "usage_type": "call"}, {"api_name": "synonym.is_synonym_in_solution", "line_number": 50, "usage_type": "call"}, {"api_name": "nltk.word_tokenize", "line_number": 67, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 84, "usage_type": "call"}]}
+{"seq_id": "527870085", "text": "\"\"\"server URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n https://docs.djangoproject.com/en/1.11/topics/http/urls/\nExamples:\nFunction views\n 1. Add an import: from my_app import views\n 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')\nClass-based views\n 1. Add an import: from other_app.views import Home\n 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')\nIncluding another URLconf\n 1. Import the include() function: from django.conf.urls import url, include\n 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls'))\n\"\"\"\nfrom django.conf.urls import url\nfrom django.contrib import admin\nfrom data_workings import views\n\nurlpatterns = [\n url(r'^admin/', admin.site.urls),\n url(r'^coordinates/',views.retrCoord),\n url(r'^create/',views.createCoord),\n url(r'^radius/',views.bathroomsInRadiusView),\n url(r'^upvote/',views.tryToUpvote),\n url(r'^signIn/',views.accountService),\n url(r'^updateHandle/', views.updateHandle)\n]\n", "sub_path": "BackEnd/server/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1058, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.conf.urls.url", "line_number": 21, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 21, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 21, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 22, "usage_type": "call"}, {"api_name": "data_workings.views.retrCoord", "line_number": 22, "usage_type": "attribute"}, {"api_name": "data_workings.views", "line_number": 22, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 23, "usage_type": "call"}, {"api_name": "data_workings.views.createCoord", "line_number": 23, "usage_type": "attribute"}, {"api_name": "data_workings.views", "line_number": 23, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 24, "usage_type": "call"}, {"api_name": "data_workings.views.bathroomsInRadiusView", "line_number": 24, "usage_type": "attribute"}, {"api_name": "data_workings.views", "line_number": 24, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 25, "usage_type": "call"}, {"api_name": "data_workings.views.tryToUpvote", "line_number": 25, "usage_type": "attribute"}, {"api_name": "data_workings.views", "line_number": 25, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 26, "usage_type": "call"}, {"api_name": "data_workings.views.accountService", "line_number": 26, "usage_type": "attribute"}, {"api_name": "data_workings.views", "line_number": 26, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 27, "usage_type": "call"}, {"api_name": "data_workings.views.updateHandle", "line_number": 27, "usage_type": "attribute"}, {"api_name": "data_workings.views", "line_number": 27, "usage_type": "name"}]}
+{"seq_id": "448947861", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits.mplot3d import Axes3D\n\ndef axis_setup_3d():\n fig = plt.figure()\n ax = fig.gca(projection=\"3d\")\n ax.set_aspect(\"equal\")\n return ax\n\n# https://stackoverflow.com/questions/13685386/matplotlib-equal-unit-length-with-equal-aspect-ratio-z-axis-is-not-equal-to\n\ndef set_axes_radius(ax, origin, radius):\n ax.set_xlim3d([origin[0] - radius, origin[0] + radius])\n ax.set_ylim3d([origin[1] - radius, origin[1] + radius])\n ax.set_zlim3d([origin[2] - radius, origin[2] + radius])\n\ndef set_axes_equal(ax):\n limits = np.array([\n ax.get_xlim3d(),\n ax.get_ylim3d(),\n ax.get_zlim3d(),\n ])\n origin = np.mean(limits, axis=1)\n radius = 0.5 * np.max(np.abs(limits[:, 1] - limits[:, 0]))\n set_axes_radius(ax, origin, radius)\n\ndef scatter_3d(obj, **kwargs):\n ax = axis_setup_3d()\n ax.scatter(obj[:, 0], obj[:, 1], obj[:, 2], **kwargs)\n set_axes_equal(ax)\n plt.show()\n", "sub_path": "flagelsim/plot.py", "file_name": "plot.py", "file_ext": "py", "file_size_in_byte": 981, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 6, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 6, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}]}
+{"seq_id": "20867916", "text": "#!/usr/bin/env python\nimport collections\nimport errno\nimport filecmp\nimport os\nimport shlex\nimport shutil\nimport subprocess\n\nimport f90nml\n\n#NPROCS_MAX = 576\nNPROCS_MAX = 480\n#NPROCS_MAX = 32\nDOC_LAYOUT = 'MOM_parameter_doc.layout'\nverbose = False\n\n\ndef regressions():\n base_path = os.getcwd()\n regressions_path = os.path.join(base_path, 'regressions')\n regression_tests = get_regression_tests(regressions_path)\n\n # Check output\n if (verbose):\n for compiler in regression_tests:\n print('{}: ['.format(compiler))\n for config in regression_tests[compiler]:\n print(' {}:'.format(config))\n for reg, test in regression_tests[compiler][config]:\n print(' {}'.format(reg))\n print(' {}'.format(test))\n print(']')\n\n n_tests = sum(len(t) for t in regression_tests['pgi'].values())\n print('Number of tests: {}'.format(n_tests))\n\n for compiler in regression_tests:\n for mode in ('repro',):\n running_tests = []\n for config in regression_tests[compiler]:\n for reg_path, test_path in regression_tests[compiler][config]:\n test = RegressionTest()\n test.refpath = reg_path\n test.runpath = test_path\n\n prefix = os.path.join(base_path, 'regressions', config)\n test.name = reg_path[len(prefix + os.sep):]\n\n mom_layout_path = os.path.join(test.runpath, 'MOM_layout')\n if os.path.isfile(mom_layout_path):\n layout_params = parse_mom6_param(mom_layout_path)\n layout = layout_params['LAYOUT']\n ocean_ni, ocean_nj = (int(n) for n in layout.split(','))\n\n masktable = layout_params.get('MASKTABLE')\n if masktable:\n n_mask = int(masktable.split('.')[1])\n else:\n n_mask = 0\n else:\n layout_path = os.path.join(test.runpath, DOC_LAYOUT)\n params = parse_mom6_param(layout_path)\n\n # If a run crashes, its proc count may be incorrect\n # TODO: Re-checkout the files?\n if not any(p in params for p in ('NIPROC', 'NJPROC')):\n print('ERROR: {} missing CPU layout'.format(test.name))\n continue\n\n ocean_ni = int(params['NIPROC'])\n ocean_nj = int(params['NJPROC'])\n\n n_mask = 0\n\n input_nml_path = os.path.join(test.runpath, 'input.nml')\n input_nml = f90nml.read(input_nml_path)\n coupler_nml = input_nml.get('coupler_nml', {})\n atmos_npes = coupler_nml.get('atmos_npes', 0)\n ocean_npes = coupler_nml.get('ocean_npes')\n\n if ocean_npes:\n assert(ocean_npes == ocean_ni * ocean_nj)\n else:\n ocean_npes = ocean_ni * ocean_nj\n\n nprocs = (ocean_npes - n_mask) + atmos_npes\n\n # XXX: I am running both in the same directory!!\n #for grid in ('dynamic', 'dynamic_symmetric'):\n for grid in ('dynamic_symmetric', ):\n # OBC tests require symmetric grids\n if (os.path.basename(test_path) == 'circle_obcs' and\n grid != 'dynamic_symmetric'):\n continue\n\n exe_path = os.path.join(\n base_path, 'MOM6-examples', 'build', compiler,\n mode, grid, config, 'MOM6'\n )\n\n if nprocs > NPROCS_MAX:\n print('{}: skipping {} ({} ranks)'.format(\n compiler, test.name, nprocs\n ))\n continue\n\n # Set up output directories\n # TODO: Ditch logpath, keep paths to stats file\n test.logpath = os.path.join(\n base_path, 'output', config, grid, test.name\n )\n mkdir_p(test.logpath)\n\n stdout_path = os.path.join(test.logpath, compiler + '.out')\n stderr_path = os.path.join(test.logpath, compiler + '.err')\n\n test.stdout = open(stdout_path, 'w')\n test.stderr = open(stderr_path, 'w')\n\n # FMS requires an existing RESTART directory\n os.chdir(test_path)\n mkdir_p('RESTART')\n\n # Stage the Slurm command\n srun_flags = ' '.join([\n '-mblock',\n '--exclusive',\n '-n {}'.format(nprocs),\n ])\n\n cmd = '{launcher} {flags} {exe}'.format(\n launcher='srun',\n flags=srun_flags,\n exe=exe_path\n )\n\n if (verbose):\n print(' Starting {}...'.format(test.name))\n\n proc = subprocess.Popen(\n shlex.split(cmd),\n stdout=test.stdout,\n stderr=test.stderr,\n )\n test.process = proc\n\n running_tests.append(test)\n\n print('{}: Running {} tests.'.format(compiler, len(running_tests)))\n\n # Wait for processes to complete\n # TODO: Cycle through and check them all, not just the first slow one\n for test in running_tests:\n test.process.wait()\n\n # Check if any runs exited with an error\n if all(test.process.returncode == 0 for test in running_tests):\n print('{}: Tests finished, no errors!'.format(compiler))\n else:\n for test in running_tests:\n if test.process.returncode != 0:\n print('{}: Test {} failed with code {}'.format(\n compiler, test.name, test.process.returncode\n ))\n\n # Process cleanup\n # TODO: Make a class method\n for test in running_tests:\n # Store the stats files\n stat_files = [\n f for f in os.listdir(test.runpath)\n if f.endswith('.stats')\n ]\n for fname in stat_files:\n src = os.path.join(test.runpath, fname)\n dst = os.path.join(test.logpath, fname) + '.' + compiler\n shutil.copy(src, dst)\n\n # Add to logs\n test.stats.append(dst)\n\n test.stdout.close()\n test.stderr.close()\n\n # Compare stats to reference\n test_results = {}\n for test in running_tests:\n test_results[test.name] = test.check_stats()\n\n if any(result == False for result in test_results.values()):\n for test in test_results:\n if test_results[test] == False:\n print('FAIL: {}'.format(test))\n else:\n print('{}: No regressions, test passed!'.format(compiler))\n\n\ndef get_regression_tests(reg_path, test_dirname='MOM6-examples'):\n regression_tests = {}\n\n model_configs = os.listdir(reg_path)\n for config in model_configs:\n config_path = os.path.join(reg_path, config)\n for path, _, files in os.walk(config_path):\n # TODO: symmetric and static support\n compilers = tuple(\n os.path.splitext(f)[1].lstrip('.')\n for f in files if f.startswith('ocean.stats')\n )\n if compilers:\n reg_dirname = os.path.basename(reg_path.rstrip(os.sep))\n r_s = path.index(reg_dirname)\n r_e = r_s + len(reg_dirname)\n test_path = path[:r_s] + test_dirname + path[r_e:]\n\n for compiler in compilers:\n if not compiler in regression_tests:\n regression_tests[compiler] = collections.defaultdict(list)\n\n test_record = path, test_path\n regression_tests[compiler][config].append(test_record)\n\n return regression_tests\n\n\ndef parse_mom6_param(path):\n params = {}\n with open(path) as param_file:\n for line in param_file:\n param_stmt = line.split('!')[0].strip()\n if param_stmt:\n key, val = [s.strip() for s in param_stmt.split('=')]\n params[key] = val\n return params\n\n\ndef mkdir_p(path):\n try:\n os.makedirs(path)\n except EnvironmentError as exc:\n if exc.errno != errno.EEXIST:\n raise\n\n\nclass RegressionTest(object):\n def __init__(self):\n self.runpath = None\n self.logpath = None\n self.refpath = None\n\n self.stats = []\n\n self.process = None\n\n self.stdout = None\n self.stderr = None\n\n def check_stats(self):\n \"\"\"Compare test stat results with regressions.\"\"\"\n\n ref_stats = [\n os.path.join(self.refpath, os.path.basename(stat))\n for stat in self.stats\n ]\n\n if self.stats:\n match = all(\n filecmp.cmp(ref, stat)\n for ref, stat in zip(ref_stats, self.stats)\n )\n else:\n match = False\n\n return match\n\n\nif __name__ == '__main__':\n regressions()\n", "sub_path": "run_regressions.py", "file_name": "run_regressions.py", "file_ext": "py", "file_size_in_byte": 10064, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.getcwd", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.sep", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path", "line_number": 62, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "f90nml.read", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 93, "usage_type": "call"}, {"api_name": "os.path", "line_number": 93, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 97, "usage_type": "call"}, {"api_name": "os.path", "line_number": 97, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 110, "usage_type": "call"}, {"api_name": "os.path", "line_number": 110, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path", "line_number": 115, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 116, "usage_type": "call"}, {"api_name": "os.path", "line_number": 116, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 122, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 141, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 142, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 172, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 176, "usage_type": "call"}, {"api_name": "os.path", "line_number": 176, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 177, "usage_type": "call"}, {"api_name": "os.path", "line_number": 177, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 178, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 202, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 204, "usage_type": "call"}, {"api_name": "os.path", "line_number": 204, "usage_type": "attribute"}, {"api_name": "os.walk", "line_number": 205, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 208, "usage_type": "call"}, {"api_name": "os.path", "line_number": 208, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 212, "usage_type": "call"}, {"api_name": "os.path", "line_number": 212, "usage_type": "attribute"}, {"api_name": "os.sep", "line_number": 212, "usage_type": "attribute"}, {"api_name": "collections.defaultdict", "line_number": 219, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 240, "usage_type": "call"}, {"api_name": "errno.EEXIST", "line_number": 242, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 263, "usage_type": "call"}, {"api_name": "os.path", "line_number": 263, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 263, "usage_type": "call"}, {"api_name": "filecmp.cmp", "line_number": 269, "usage_type": "call"}]}
+{"seq_id": "504155994", "text": "#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\n'''\n[env]\nTo activate this environment, use :: conda activate pandas_ga_1\nTo deactivate an active environment, use :: conda deactivate\n# you have created a env with all the required packages\nsource activate pandas_ga_1\n\n\n[path]\ncd /Users/brunoflaven/Documents/01_work/blog_articles/start_stopping_stop_starting/google_analytics_api_pandas_reporting/\n\n\n[file]\npython 008_google_analytics_api_pandas_reporting.py\n\n[source]\nhttps://developers.google.com/analytics/devguides/reporting/core/v4/quickstart/service-py\n\nhttps://www.themarketingtechnologist.co/getting-started-with-the-google-analytics-reporting-api-in-python/\n\n\n[install]\nhttps://pypi.org/project/google-analytics-data/\nhttps://anaconda.org/anaconda/pandas\nhttps://anaconda.org/conda-forge/google-api-python-client\n\npip install google-analytics-data\nconda install -c conda-forge analytics-python\nconda install -c conda-forge google-auth-oauthlib\nconda install -c anaconda pandas\nconda install -c anaconda google\n\ngoogle-analytics-data==0.8.0\ngoogle-auth-oauthlib==0.4.6\n\npip install --upgrade google-api-python-client\nconda install -c conda-forge google-api-python-client\n\npip install google-analytics-data\npip install --upgrade google-api-python-client\n\nconda install -c conda-forge oauth2client\n\n\n\n'''\n\n\n\"\"\"Hello Analytics Reporting API V4.\"\"\"\n\nimport pandas as pd\nfrom apiclient.discovery import build\nfrom oauth2client.service_account import ServiceAccountCredentials\nSCOPES = ['https://www.googleapis.com/auth/analytics.readonly']\n\n\n# just change it with your own stuff\nKEY_FILE_LOCATION = ''\n# download from https://console.cloud.google.com/apis/credentials?project=\n\n\n# just change it with your own stuff\nVIEW_ID = ''\n# xxxo45678765456nyuygfhvf65675 is a fake id\n# get at https://analytics.google.com/analytics/web/#/xxxo45678765456nyuygfhvf65675/admin/view/settings\n\n\n\ndef initialize_analyticsreporting():\n \"\"\"Initializes an Analytics Reporting API V4 service object.\n\n Returns:\n An authorized Analytics Reporting API V4 service object.\n \"\"\"\n credentials = ServiceAccountCredentials.from_json_keyfile_name(\n KEY_FILE_LOCATION, SCOPES)\n\n # Build the service object.\n analytics = build('analyticsreporting', 'v4', credentials=credentials)\n\n return analytics\n\n\ndef get_report(analytics):\n \"\"\"Queries the Analytics Reporting API V4.\n\n Args:\n analytics: An authorized Analytics Reporting API V4 service object.\n Returns:\n The Analytics Reporting API V4 response.\n \"\"\"\n return analytics.reports().batchGet(\n body={\n 'reportRequests': [\n {\n 'viewId': VIEW_ID,\n 'dateRanges': [{'startDate': '7daysAgo', 'endDate': 'today'}],\n 'metrics': [{'expression': 'ga:sessions'}],\n 'dimensions': [{'name': 'ga:country'}]\n }]\n }\n ).execute()\n\n\ndef print_response(response):\n \"\"\"Parses and prints the Analytics Reporting API V4 response.\n\n Args:\n response: An Analytics Reporting API V4 response.\n \"\"\"\n for report in response.get('reports', []):\n columnHeader = report.get('columnHeader', {})\n dimensionHeaders = columnHeader.get('dimensions', [])\n metricHeaders = columnHeader.get(\n 'metricHeader', {}).get('metricHeaderEntries', [])\n\n for row in report.get('data', {}).get('rows', []):\n dimensions = row.get('dimensions', [])\n dateRangeValues = row.get('metrics', [])\n\n for header, dimension in zip(dimensionHeaders, dimensions):\n print(header + ': ', dimension)\n\n for i, values in enumerate(dateRangeValues):\n print('Date range:', str(i))\n for metricHeader, value in zip(metricHeaders, values.get('values')):\n print(metricHeader.get('name') + ':', value)\n\ndef print_response_new (response):\n list = []\n # get report data\n for report in response.get('reports', []):\n # set column headers\n columnHeader = report.get('columnHeader', {})\n dimensionHeaders = columnHeader.get('dimensions', [])\n metricHeaders = columnHeader.get(\n 'metricHeader', {}).get('metricHeaderEntries', [])\n rows = report.get('data', {}).get('rows', [])\n\n for row in rows:\n # create dict for each row\n dict = {}\n dimensions = row.get('dimensions', [])\n dateRangeValues = row.get('metrics', [])\n\n # fill dict with dimension header (key) and dimension value (value)\n for header, dimension in zip(dimensionHeaders, dimensions):\n dict[header] = dimension\n\n # fill dict with metric header (key) and metric value (value)\n for i, values in enumerate(dateRangeValues):\n for metric, value in zip(metricHeaders, values.get('values')):\n #set int as int, float a float\n if ',' in value or '.' in value:\n dict[metric.get('name')] = float(value)\n else:\n dict[metric.get('name')] = int(value)\n\n list.append(dict)\n\n df = pd.DataFrame(list)\n return df\n\ndef main():\n analytics = initialize_analyticsreporting()\n response = get_report(analytics)\n\n # print_response(response)\n df = print_response_new(response)\n # show me the money\n print(df)\n\n\nif __name__ == '__main__':\n main()\n\n\n", "sub_path": "stop_starting_start_stopping/google_analytics_api_pandas_reporting/008_google_analytics_api_pandas_reporting.py", "file_name": "008_google_analytics_api_pandas_reporting.py", "file_ext": "py", "file_size_in_byte": 5269, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "oauth2client.service_account.ServiceAccountCredentials.from_json_keyfile_name", "line_number": 78, "usage_type": "call"}, {"api_name": "oauth2client.service_account.ServiceAccountCredentials", "line_number": 78, "usage_type": "name"}, {"api_name": "apiclient.discovery.build", "line_number": 82, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 164, "usage_type": "call"}]}
+{"seq_id": "503120277", "text": "import time\n\nimport matplotlib.pyplot as plt\n\nfrom c2 import c2\n\n\ndef c3():\n n_list = []\n t_list = []\n for n in range(2, 100):\n n_list.append(n)\n t0 = time.clock()\n c2(n)\n t_list.append(time.clock() - t0)\n plt.plot(n_list, t_list)\n plt.show()\n\n\nif __name__ == '__main__':\n c3()\n", "sub_path": "Project 1/c3.py", "file_name": "c3.py", "file_ext": "py", "file_size_in_byte": 324, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "time.clock", "line_number": 13, "usage_type": "call"}, {"api_name": "c2.c2", "line_number": 14, "usage_type": "call"}, {"api_name": "time.clock", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}]}
+{"seq_id": "98368811", "text": "from django.shortcuts import render\nfrom django.forms import formset_factory,modelformset_factory\nfrom . import forms\nfrom .models import ModelSetPost\nfrom django.core.files.storage import FileSystemStorage\nimport os \n\ndef index(request):\n return render(request,'formapp/index.html')\n\ndef form_page(request):\n form = forms.UserInfo()\n if request.method == 'POST':\n form=forms.UserInfo(request.POST)\n if form.is_valid():\n print('バリデーション成功')\n # print(\n # f\"name: {form.cleaned_data['name']}, mail: {form.cleaned_data['mail']},age: {form.cleaned_data['age']}\"\n # )\n print(form.cleaned_data)\n return render(\n request, 'formapp/form_page.html', \n context={ 'form': form\n }\n )\n\ndef form_post(request):\n form = forms.PostModelForm()\n if request.method == 'POST':\n form = forms.PostModelForm(request.POST)\n if form.is_valid():\n form.save()\n return render(\n request, 'formapp/form_post.html',\n context={'form':form}\n )\n\ndef form_set_post(request):\n TestFormset = formset_factory(forms.FormSetPost,extra=3)\n formset = TestFormset(request.POST or None)\n if formset.is_valid():\n for form in formset:\n print(form.cleaned_data)\n return render(\n request, 'formapp/form_set_post.html',\n context = {'formset':formset }\n )\n\ndef modelform_set_post(request):\n # TestFormSet = modelformset_factory(ModelSetPost, fields='__all__',extra=3)\n TestFormSet = modelformset_factory(ModelSetPost,form=forms.ModelFormSetPost,extra=3)\n formset = TestFormSet(request.POST or None, queryset=ModelSetPost.objects.filter(id__gt=3))\n if formset.is_valid():\n formset.save()\n return render(\n request, 'formapp/modelform_set_post.html',\n context = {'formset':formset }\n )\n\ndef upload_sample(request):\n if request.method == 'POST' and request.FILES['upload_file']:\n #送られたファイルの取り出し\n upload_file = request.FILES['upload_file']\n fs = FileSystemStorage() #ファイルを保存する\n file_path = os.path.join('upload', upload_file.name)\n file = fs.save(file_path, upload_file)\n uploaded_file_url = fs.url(file)\n return render(request, 'formapp/upload_file.html',\n context = {'uploaded_file_url': uploaded_file_url}\n )\n return render(request, 'formapp/upload_file.html')\n\ndef upload_model_form(request):\n user = None\n if request.method == 'POST':\n form = forms.UserForm(request.POST,request.FILES)\n if form.is_valid():\n user = form.save()\n else:\n form = forms.UserForm()\n return render(request,'formapp/upload_model_form.html',\n context = {'form':form, 'user': user}\n )", "sub_path": "django/Udemy/FormSample/FormProject/FormApp/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2842, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.shortcuts.render", "line_number": 9, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 21, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 33, "usage_type": "call"}, {"api_name": "django.forms.formset_factory", "line_number": 39, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 44, "usage_type": "call"}, {"api_name": "django.forms.modelformset_factory", "line_number": 51, "usage_type": "call"}, {"api_name": "models.ModelSetPost", "line_number": 51, "usage_type": "argument"}, {"api_name": "models.ModelSetPost.objects.filter", "line_number": 52, "usage_type": "call"}, {"api_name": "models.ModelSetPost.objects", "line_number": 52, "usage_type": "attribute"}, {"api_name": "models.ModelSetPost", "line_number": 52, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 55, "usage_type": "call"}, {"api_name": "django.core.files.storage.FileSystemStorage", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path", "line_number": 65, "usage_type": "attribute"}, {"api_name": "django.shortcuts.render", "line_number": 68, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 71, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 81, "usage_type": "call"}]}
+{"seq_id": "203710165", "text": "#coding: utf-8\nimport logging\n\nfrom sqlalchemy.exc import IntegrityError\nimport transaction\n\nimport models\nimport excepts\n\n\nlogger = logging.getLogger('balaio.checkin')\n\n\ndef get_attempt(package, Session=models.Session):\n \"\"\"\n Returns a brand new models.Attempt instance, bound to a models.ArticlePkg\n instance.\n\n A package is valid when it has at least one valid xml file, according to\n SPS or rSPS xsd, and one pdf file.\n\n Case 1: Package is valid and has all needed metadata:\n A :class:`models.Attempt` is returned, bound to a :class:`models.ArticlePkg`.\n Case 2: Package is valid and doesn't have all needed metadata:\n A :class:`models.Attempt` is returned, with :attr:`models.Attempt.is_valid==False`.\n Case 3: Package is invalid\n A :class:`models.Attempt` is returned, with :attr:`models.Attempt.is_valid==False`.\n Case 4: Package is duplicated\n raises :class:`excepts.DuplicatedPackage`.\n\n :param package: Instance of SafePackage.\n :param Session: (optional) Reference to a Session class.\n \"\"\"\n logger.info('Analyzing package: %s' % package)\n\n with package.analyzer as pkg:\n try:\n logger.debug('Creating a transactional session scope')\n session = Session()\n\n # Building a new Attempt\n attempt = models.Attempt.get_from_package(pkg)\n session.add(attempt)\n\n # Trying to bind a ArticlePkg\n savepoint = transaction.savepoint()\n try:\n article_pkg = models.ArticlePkg.get_or_create_from_package(pkg, session)\n if article_pkg not in session:\n session.add(article_pkg)\n\n attempt.articlepkg = article_pkg\n attempt.is_valid = True\n\n #checkin_notifier.tell('Attempt is valid.', models.Status.ok, 'Checkin')\n\n except Exception as e:\n savepoint.rollback()\n #checkin_notifier.tell('Failed to load an ArticlePkg for %s.' % package, models.Status.error, 'Checkin')\n\n logger.error('Failed to load an ArticlePkg for %s.' % package)\n logger.debug('---> Traceback: %s' % e)\n\n transaction.commit()\n return attempt\n\n except IOError as e:\n transaction.abort()\n logger.error('The package %s had been deleted during analysis' % package)\n logger.debug('---> Traceback: %s' % e)\n raise ValueError('The package %s had been deleted during analysis' % package)\n\n except IntegrityError as e:\n transaction.abort()\n logger.error('The package has no integrity. Aborting.')\n logger.debug('---> Traceback: %s' % e)\n\n if 'violates not-null constraint' in e.message:\n raise ValueError('An integrity error was cast as ValueError.')\n else:\n raise excepts.DuplicatedPackage('The package %s already exists.' % package)\n\n except Exception as e:\n transaction.abort()\n\n logger.error('Unexpected error! The package analysis for %s was aborted.' % (\n package))\n logger.debug('---> Traceback: %s' % e)\n raise ValueError('Unexpected error! The package analysis for %s was aborted.' % package)\n\n finally:\n logger.debug('Closing the transactional session scope')\n session.close()\n\n", "sub_path": "balaio/checkin.py", "file_name": "checkin.py", "file_ext": "py", "file_size_in_byte": 3443, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "models.Session", "line_number": 14, "usage_type": "attribute"}, {"api_name": "models.Attempt.get_from_package", "line_number": 42, "usage_type": "call"}, {"api_name": "models.Attempt", "line_number": 42, "usage_type": "attribute"}, {"api_name": "transaction.savepoint", "line_number": 46, "usage_type": "call"}, {"api_name": "models.ArticlePkg.get_or_create_from_package", "line_number": 48, "usage_type": "call"}, {"api_name": "models.ArticlePkg", "line_number": 48, "usage_type": "attribute"}, {"api_name": "transaction.commit", "line_number": 64, "usage_type": "call"}, {"api_name": "transaction.abort", "line_number": 68, "usage_type": "call"}, {"api_name": "sqlalchemy.exc.IntegrityError", "line_number": 73, "usage_type": "name"}, {"api_name": "transaction.abort", "line_number": 74, "usage_type": "call"}, {"api_name": "excepts.DuplicatedPackage", "line_number": 81, "usage_type": "call"}, {"api_name": "transaction.abort", "line_number": 84, "usage_type": "call"}]}
+{"seq_id": "164451879", "text": "from urllib import request\nfrom bs4 import BeautifulSoup\nfrom selenium import webdriver\nfrom selenium.webdriver.chrome.options import Options\nimport pandas as pd\nimport os\n\ndef get_pages(url):\n\tres = request.urlopen(url)\n\tsoup = BeautifulSoup(res,\"html.parser\")\n\tres.close()\n\treturn soup\n\ndef get_tables(url):\n\tdata = pd.read_html(url)\n\treturn data\n\ndef get_driver(opt=None):\n\tdp = r\"C:\\driver\\94\\chromedriver.exe\"\n\toptions = Options()\n\tif opt:\n\t\toptions.add_argument('--headless')\n\treturn webdriver.Chrome(dp, options=options)\n\ndef set_login(d):\n\td.get(r\"https://regist.netkeiba.com/account/?pid=login\")\n\t# d.switch_to_window(d.window_handles[1])\n\tloginid = d.find_element_by_name(\"login_id\")\n\tpswd = d.find_element_by_name('pswd')\n\tlogbtn = d.find_element_by_class_name('loginBtn__wrap').find_element_by_tag_name('input')\n\tloginid.send_keys(os.environ['email'])\n\tpswd.send_keys(os.environ['passwd'])\n\tlogbtn.click()\n\treturn d\n", "sub_path": "scraping/HorseRace/BK/scrapUtil.py", "file_name": "scrapUtil.py", "file_ext": "py", "file_size_in_byte": 928, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "urllib.request.urlopen", "line_number": 9, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 9, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 10, "usage_type": "call"}, {"api_name": "pandas.read_html", "line_number": 15, "usage_type": "call"}, {"api_name": "selenium.webdriver.chrome.options.Options", "line_number": 20, "usage_type": "call"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 23, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 23, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 32, "usage_type": "attribute"}]}
+{"seq_id": "366277671", "text": "# Script will stop neato lidar.\n# If the ros neato driver fails the lidar continues to\n# spin. This scrip is used to stop the spinning, especially\n# during development.\n\n# Author: Brannon Vann brannon.vann@gmail.com\n# License: MIT\n\n# Run this script: python stop_neato_lidar.py\n\nimport serial\n\ntry:\n serial = serial.Serial('/dev/ttyACM0', timeout=1)\n serial.write('setldsrotation off\\n')\n serial.write('setled buttonoff\\n')\n serial.write('testmode off\\n')\n\n # close serial port\n serial.close()\n print(\"Done stopping down neato.\")\nexcept:\n print(\"an error occurred while stoping neato lidar.\")\n", "sub_path": "scripts/stop_neato.py", "file_name": "stop_neato.py", "file_ext": "py", "file_size_in_byte": 617, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "serial.Serial", "line_number": 14, "usage_type": "call"}, {"api_name": "serial.write", "line_number": 15, "usage_type": "call"}, {"api_name": "serial.write", "line_number": 16, "usage_type": "call"}, {"api_name": "serial.write", "line_number": 17, "usage_type": "call"}, {"api_name": "serial.close", "line_number": 20, "usage_type": "call"}]}
+{"seq_id": "255257404", "text": "from torch.testing._internal.common_utils import TestCase, run_tests\nimport torch\nfrom torch import vmap\nimport warnings\n\nclass TestVmapAPI(TestCase):\n def test_non_tensor_output_raises(self):\n with self.assertRaisesRegex(ValueError, \"got type as the return\"):\n output = vmap(lambda x: 3.14)(torch.ones(3))\n\n def multiple_outputs(x):\n return x, 3\n\n with self.assertRaisesRegex(ValueError, \"got type for return 1\"):\n vmap(multiple_outputs)(torch.ones(3))\n\n def test_different_map_dim_size_raises(self):\n x = torch.randn(2)\n y = torch.randn(3)\n expected_msg = 'Expected all tensors to have the same size in the mapped dimension'\n with self.assertRaisesRegex(ValueError, expected_msg):\n vmap(torch.mul)(x, y)\n\n def test_func_with_no_inputs(self):\n expected_msg = 'got no inputs'\n\n def foo():\n return torch.randn(3)\n\n def bar(x):\n return torch.randn(3)\n\n with self.assertRaisesRegex(ValueError, expected_msg):\n vmap(foo)()\n\n with self.assertRaisesRegex(ValueError, expected_msg):\n vmap(bar)()\n\n def test_constant_function(self):\n output = vmap(lambda x: torch.tensor(3.14))(torch.ones(3))\n self.assertEqual(output, torch.tensor([3.14, 3.14, 3.14]))\n\n def test_single_input(self):\n x = torch.randn(2, 3)\n\n def square(x):\n return x * x\n\n output = vmap(square)(x)\n self.assertEqual(output, x * x)\n\n def test_multiple_inputs(self):\n x = torch.randn(2, 3)\n y = torch.randn(2, 3)\n output = vmap(torch.mul)(x, y)\n self.assertEqual(output, x * y)\n\n def test_multiple_outputs(self):\n def foo(x):\n return x * x, x * x * x\n\n x = torch.randn(3)\n outputs = vmap(foo)(x)\n self.assertEqual(outputs[0], x * x)\n self.assertEqual(outputs[1], x * x * x)\n\n def test_multiple_outputs_error_cases(self):\n # This is the same thing as\n # def returns_tuple_of_tensors(x):\n # return x, x\n def returns_tuple_of_tensors(x):\n return (x, x)\n\n def returns_list_of_two_tensors(x):\n return [x, x]\n\n def returns_list_of_one_tensor(x):\n return [x]\n\n x = torch.randn(3)\n\n # should not throw\n vmap(returns_tuple_of_tensors)(x)\n\n # jax supports these, but we don't yet\n msg = \"must only return Tensors, got type \"\n with self.assertRaisesRegex(ValueError, msg):\n vmap(returns_list_of_two_tensors)(x)\n with self.assertRaisesRegex(ValueError, msg):\n vmap(returns_list_of_one_tensor)(x)\n\n def test_nested_with_same_map_dim(self):\n x = torch.randn(2, 3, 5)\n y = torch.randn(2, 3, 5)\n output = vmap(vmap(torch.mul))(x, y)\n self.assertEqual(output, x * y)\n\n output = vmap(vmap(vmap(torch.mul)))(x, y)\n self.assertEqual(output, x * y)\n\n def test_nested_with_different_map_dim(self):\n x = torch.randn(2, 3)\n y = torch.randn(5, 3)\n output = vmap(lambda x: vmap(lambda y: x * y)(y))(x)\n self.assertEqual(output.shape, (2, 5, 3))\n self.assertEqual(output, x.view(2, 1, 3) * y)\n\n z = torch.randn(7, 3)\n output = vmap(lambda x: vmap(lambda y: vmap(lambda z: x * y * z)(z))(y))(x)\n self.assertEqual(output.shape, (2, 5, 7, 3))\n self.assertEqual(output, x.view(2, 1, 1, 3) * y.view(5, 1, 3) * z)\n\n def test_noop_in_inner_vmap(self):\n x = torch.randn(3)\n y = torch.randn(5)\n output = vmap(lambda x: vmap(lambda y: x)(y))(x)\n self.assertEqual(output, x.view(3, 1).expand(3, 5))\n\n def test_unsupported_op_err_msg(self):\n # Unsupported view op\n tensor = torch.randn(2, 3)\n with self.assertRaisesRegex(RuntimeError, \"doesn't work on in-place or view ops\"):\n vmap(torch.as_strided, (0, None, None))(tensor, [2, 3], [0, 0])\n\n # We don't support multiple returns yet\n with self.assertRaisesRegex(RuntimeError, 'multiple returns'):\n vmap(torch.var_mean)(tensor)\n\n # The fallback doesn't support TensorList\n with self.assertRaisesRegex(RuntimeError, 'Batching rule not implemented'):\n vmap(lambda t: torch.stack([t]))(tensor)\n\n # Don't support non-tensor returns. This is a limitation of vmap;\n # functions that don't return tensors must be special cased\n with self.assertRaisesRegex(RuntimeError, 'Batching rule not implemented'):\n vmap(torch.Tensor.item)(tensor)\n\n def test_unsupported_inplace_op_err_msg(self):\n def foo(x):\n return x.cos_()\n\n x = torch.randn(3)\n with self.assertRaisesRegex(\n RuntimeError, 'Batching rule not implemented'):\n vmap(foo)(x)\n\n def test_nonzero_out_dims(self):\n # Basic test\n tensor = torch.randn(2, 3)\n result = vmap(lambda x: x, out_dims=1)(tensor)\n self.assertEqual(result, tensor.permute(1, 0))\n self.assertEqual(result.data_ptr(), tensor.data_ptr())\n\n # Test that the batch dimension gets permuted to dim 2\n tensor = torch.randn(2, 3, 5, 7)\n result = vmap(lambda x: x, out_dims=2)(tensor)\n self.assertEqual(result, tensor.permute(1, 2, 0, 3))\n self.assertEqual(result.data_ptr(), tensor.data_ptr())\n\n # negative out_dim\n tensor = torch.randn(2, 3, 5, 7)\n result = vmap(lambda x: x, out_dims=-1)(tensor)\n self.assertEqual(result, tensor.permute(1, 2, 3, 0))\n self.assertEqual(result.data_ptr(), tensor.data_ptr())\n\n # check that out_dims works on ALL outputs\n tensor = torch.randn(2, 3, 5, 7)\n other = torch.randn(2, 3, 5, 7)\n result = vmap(lambda x, y: (x, y), out_dims=2)(tensor, other)\n self.assertEqual(result, (tensor.permute(1, 2, 0, 3), other.permute(1, 2, 0, 3)))\n\n # use out_dims with the maximum vmap-able tensor dims (64 dims)\n ndims = 64\n shape = [2] + [1] * (ndims - 1)\n expected_shape = [1, 1, 2] + [1] * (ndims - 3)\n tensor = torch.randn(shape)\n result = vmap(lambda x: x, out_dims=2)(tensor)\n self.assertEqual(result.shape, expected_shape)\n\n # test something that is not the identity function\n def foo(x, y):\n return x, x * y, x * y * y\n x = torch.randn(2, 3, 5)\n y = torch.randn(2, 3, 5)\n result = vmap(foo, out_dims=1)(x, y)\n self.assertEqual(\n result,\n (x.permute(1, 0, 2), (x * y).permute(1, 0, 2), (x * y * y).permute(1, 0, 2)))\n\n def test_multiple_out_dims(self):\n def foo(x):\n return x, x\n\n def bar(x, y):\n return x, x, x, x * y\n\n x = torch.randn(2, 3, 5)\n y = torch.randn(2, 3, 5)\n result = vmap(foo, out_dims=(0, 1))(x)\n self.assertEqual(result, (x, x.permute(1, 0, 2)))\n\n result = vmap(bar, out_dims=(-1, 0, 1, 2))(x, y)\n expected = (\n x.permute(1, 2, 0),\n x,\n x.permute(1, 0, 2),\n (x * y).permute(1, 2, 0),\n )\n self.assertEqual(result, expected)\n\n def test_nested_out_dims(self):\n y = torch.randn(2, 3, 5, 7)\n\n # Inner vmap has non-zero out_dim\n result = vmap(lambda y: vmap(lambda x: x, out_dims=1)(y))(y)\n self.assertEqual(result.shape, (2, 5, 3, 7))\n self.assertEqual(result, y.permute(0, 2, 1, 3))\n\n # all vmaps have non-zero out_dim\n result = vmap(lambda y: vmap(lambda x: x, out_dims=1)(y), out_dims=1)(y)\n self.assertEqual(result.shape, (5, 2, 3, 7))\n self.assertEqual(result, y.permute(2, 0, 1, 3))\n\n # throwing in some negative out_dims\n result = vmap(lambda y: vmap(lambda x: x, out_dims=-1)(y), out_dims=-1)(y)\n self.assertEqual(result.shape, (5, 7, 3, 2))\n self.assertEqual(result, y.permute(2, 3, 1, 0))\n\n # testing fn that isn't the identity\n x = torch.randn(2, 3)\n y = torch.randn(5, 3)\n result = vmap(lambda y: vmap(lambda x: x * y, out_dims=1)(x), out_dims=-1)(y)\n self.assertEqual(result.shape, (3, 2, 5))\n self.assertEqual(result, (y.view(5, 1, 3) * x).permute(2, 1, 0))\n\n def test_out_dims_edge_case(self):\n def foo(x):\n return x\n\n # Test that we accept out_dims=(1,) for a function with one output.\n tensor = torch.randn(2, 3)\n expected = vmap(foo, out_dims=1)(tensor)\n result = vmap(foo, out_dims=(1,))(tensor)\n self.assertEqual(result, expected)\n\n def test_out_dims_must_be_int_or_tuple_of_int_err_msg(self):\n msg = '`out_dims` must be an int or a tuple of int'\n tensor = torch.randn(2, 3)\n with self.assertRaisesRegex(ValueError, msg):\n vmap(lambda x: x, out_dims='lol')(tensor)\n with self.assertRaisesRegex(ValueError, msg):\n vmap(lambda x: x, out_dims=('lol',))(tensor)\n with self.assertRaisesRegex(ValueError, msg):\n vmap(lambda x: x, out_dims=None)(tensor)\n with self.assertRaisesRegex(ValueError, msg):\n vmap(lambda x: x, out_dims=(None,))(tensor)\n\n def test_out_dims_and_num_outputs_mismatch_err_msg(self):\n msg = '`out_dims` must have one dim per output'\n x = torch.randn(2, 3, 5)\n\n # Too many out_dims\n with self.assertRaisesRegex(ValueError, msg):\n vmap(lambda x: x, out_dims=(0, 0))(x)\n with self.assertRaisesRegex(ValueError, msg):\n vmap(lambda x: (x, x, x), out_dims=(0, 0, 0, 0))(x)\n\n # Too few out_dims\n with self.assertRaisesRegex(ValueError, msg):\n vmap(lambda x: (x, x), out_dims=(0,))(x)\n with self.assertRaisesRegex(ValueError, msg):\n vmap(lambda x: (x, x, x), out_dims=(0, 0))(x)\n\n def test_out_dim_out_of_bounds_err_msg(self):\n # TODO(rzou): This error message isn't that great. It comes straight\n # from maybe_wrap_dim. Consider doing a try-catch-(add some context) to\n # the error message in the future in C++\n msg = 'Dimension out of range'\n x = torch.randn(2, 3, 5)\n with self.assertRaisesRegex(IndexError, msg):\n vmap(lambda x: x, out_dims=3)(x)\n with self.assertRaisesRegex(IndexError, msg):\n vmap(lambda x: x, out_dims=-4)(x)\n\n def test_non_zero_in_dims(self):\n tensor = torch.randn(2, 3, 5)\n\n # Implicit out_dims = 0; vmap will move the batch dim to the front.\n output = vmap(lambda x: x, (1,))(tensor)\n self.assertEqual(output, tensor.permute(1, 0, 2))\n self.assertEqual(output.data_ptr(), tensor.data_ptr())\n\n x = torch.randn(2, 3)\n y = torch.randn(3, 2)\n output = vmap(torch.mul, (0, 1))(x, y)\n self.assertEqual(output, x * y.t())\n output = vmap(torch.mul, (1, 0))(x, y)\n self.assertEqual(output, x.t() * y)\n\n def test_none_in_dims(self):\n x = torch.randn(2, 3)\n y = torch.randn(2, 3)\n\n # None in_dim for a Tensor means we don't map over it\n output = vmap(torch.mul, (0, None))(x, y)\n self.assertEqual(output.shape, (2, 2, 3))\n self.assertEqual(output, x.view(2, 1, 3) * y)\n\n # None in_dim for non-tensor arguments\n output = vmap(torch.mul, (0, None))(x, 2)\n self.assertEqual(output, x * 2)\n\n def test_nested_non_default_in_dims(self):\n x = torch.rand(5, 2, 3)\n y = torch.rand(3, 5, 2)\n result = vmap(vmap(vmap(torch.mul), (1, 0)), (1, 2))(x, y)\n self.assertEqual(result, x.permute(1, 2, 0) * y.permute(2, 0, 1))\n\n def test_non_default_in_dims_out_dims(self):\n x = torch.randn(2, 3, 5)\n\n # Same in_dim as out_dim, vmap over identity\n result = vmap(lambda x: x, in_dims=1, out_dims=1)(x)\n self.assertEqual(result, x)\n self.assertEqual(result.data_ptr(), x.data_ptr())\n\n # Different in_dim from out_dim, vmap over identity\n result = vmap(lambda x: x, in_dims=2, out_dims=1)(x)\n self.assertEqual(result.shape, (2, 5, 3))\n self.assertEqual(result, x.transpose(1, 2))\n self.assertEqual(result.data_ptr(), x.data_ptr())\n\n def foo(x):\n return x * 2\n\n # Same in_dim as out_dim, vmap over operation\n result = vmap(foo, in_dims=1, out_dims=1)(x)\n self.assertEqual(result, x * 2)\n\n # Different in_dim as out_dim, vmap over operation\n result = vmap(foo, in_dims=2, out_dims=1)(x)\n self.assertEqual(result.shape, (2, 5, 3))\n self.assertEqual(result, (x * 2).transpose(1, 2))\n\n # Basic nested test.\n result = vmap(vmap(foo, 1, 1), 1, 1)(x)\n self.assertEqual(result, x * 2)\n\n def test_in_dims_wrong_type_err_msg(self):\n x = torch.randn(3)\n y = torch.randn(3)\n msg = 'expected `in_dims` to be int or tuple'\n with self.assertRaisesRegex(ValueError, msg):\n vmap(torch.mul, [0, 0])(x, y)\n with self.assertRaisesRegex(ValueError, msg):\n vmap(torch.mul, set({0, 0}))(x, y)\n with self.assertRaisesRegex(ValueError, msg):\n vmap(torch.mul, 'lol')(x, y)\n # The following should not throw\n vmap(torch.mul, (0, 0))(x, y)\n\n def test_not_enough_in_dims_err_msg(self):\n x = torch.randn(3)\n y = torch.randn(3)\n msg = r'expected one `in_dim` per input \\(got \\w+ inputs\\)'\n\n with self.assertRaisesRegex(ValueError, msg):\n vmap(torch.mul, (0,))(x, y)\n with self.assertRaisesRegex(ValueError, msg):\n vmap(torch.mul, (0, 0, 0))(x, y)\n # The following should not throw\n vmap(torch.mul, (0, 0))(x, y)\n\n def test_in_dims_must_be_flat_tuple_err_msg(self):\n msg = 'in_dims must be a flat tuple containing ints and/or Nones'\n\n x = torch.randn(3)\n y = torch.randn(3)\n z = torch.randn(3)\n\n def foo(xy):\n return xy[0] * xy[1]\n\n def bar(x, yz):\n return x * yz[0] * yz[1]\n\n # NB: jax supports all of the following, we don't yet.\n with self.assertRaisesRegex(ValueError, msg):\n vmap(foo, ((0, 0),))((x, y))\n with self.assertRaisesRegex(ValueError, msg):\n vmap(bar, (0, (0, 0)))(x, (y, z))\n with self.assertRaisesRegex(ValueError, msg):\n vmap(foo, ({0: 0, 1: 0},))({0: x, 1: y})\n\n def test_integer_in_dim_but_not_tensor_input_err_msg(self):\n def foo(xy):\n return xy[0] * xy[1]\n\n def bar(x, yz):\n return x * yz[0] * yz[1]\n\n x = torch.randn(2, 3)\n y = torch.randn(2, 3)\n\n # jax supports these, we too can in the future.\n msg = 'Got in_dim=0 for input 0, but input 0 is not a Tensor'\n with self.assertRaisesRegex(ValueError, msg):\n vmap(foo)((x, y))\n with self.assertRaisesRegex(ValueError, msg):\n vmap(foo, (0,))((x, y))\n\n # jax supports these as well, we too can in the future.\n msg = 'Got in_dim=0 for input 1, but input 1 is not a Tensor'\n with self.assertRaisesRegex(ValueError, msg):\n vmap(foo)(x, (x, y))\n with self.assertRaisesRegex(ValueError, msg):\n vmap(foo, (0, 0))(x, (x, y))\n\n # the following are errors in jax (and will always be errors)\n msg = 'Got in_dim=0 for input 1, but input 1 is not a Tensor'\n with self.assertRaisesRegex(ValueError, msg):\n vmap(torch.sum)(x, 0)\n with self.assertRaisesRegex(ValueError, msg):\n vmap(torch.sum, (0, 0))(x, 0)\n # The following should not throw\n vmap(torch.sum, (0, None))(x, 0)\n\n def test_in_dim_not_in_tensor_err_msg(self):\n def foo(x):\n return x * x\n\n msg = r'Got in_dim=-?\\w for input 0, but input 0 is a Tensor of dimensionality \\w'\n with self.assertRaisesRegex(ValueError, msg):\n vmap(foo)(torch.randn([]))\n with self.assertRaisesRegex(ValueError, msg):\n vmap(foo, in_dims=(0,))(torch.randn([]))\n with self.assertRaisesRegex(ValueError, msg):\n vmap(foo, in_dims=(-1,))(torch.randn(2, 3))\n with self.assertRaisesRegex(ValueError, msg):\n vmap(foo, in_dims=(2,))(torch.randn(2, 3))\n # the following should not throw\n vmap(foo, in_dims=(0,))(torch.randn(2, 3))\n vmap(foo, in_dims=(1,))(torch.randn(2, 3))\n\n def test_fallback_sub(self):\n # NB: One day we will implement a batching rule for torch.sub.\n # If/when we do, this test should be replaced to test the fallback\n # path on another operator to avoid bitrot.\n x = torch.randn(5, 7, 11)\n y = torch.randn(5, 7, 11)\n\n # Test the fallback path raises a warning\n with warnings.catch_warnings(record=True) as wa:\n result = vmap(torch.sub)(x, y)\n self.assertEqual(len(wa), 2)\n self.assertRegex(str(wa[-1].message),\n r'falling back to slow \\(for loop and stack\\) implementation')\n self.assertEqual(result, x - y)\n\n # fallback on torch.sub\n x = torch.randn(7, 11, 5)\n y = torch.randn(5, 7, 11)\n result = vmap(torch.sub, (2, 0))(x, y)\n self.assertEqual(result, x.permute(2, 0, 1) - y)\n\n # fallback on torch.sub, nested vmap\n x = torch.randn(7, 11, 5)\n y = torch.randn(5, 7, 11)\n result = vmap(vmap(torch.sub), (2, 0))(x, y)\n self.assertEqual(result, x.permute(2, 0, 1) - y)\n\n # big batch size (total 10000)\n x = torch.randn(100, 10, 10, 5)\n y = torch.randn(100, 10, 10)\n result = vmap(vmap(vmap(torch.sub)))(x, y)\n self.assertEqual(result, x - y.view(100, 10, 10, 1))\n\n def test_fallback_masked_fill(self):\n # NB: One day we will implement a batching rule for masked_fill\n # If/when we do, this test should be replaced to test the fallback\n # path on another operator to avoid bitrot.\n def run_test(batch_size):\n B0 = batch_size\n x = torch.randn(B0, 7, 11, 13)\n dim = 0\n index = torch.tensor([0, 4, 2])\n values = torch.randn(B0, 3, 13)\n\n with warnings.catch_warnings(record=True) as wa:\n result = vmap(torch.index_add, (0, None, None, 0))(x, dim, index, values)\n self.assertEqual(len(wa), 2)\n self.assertRegex(str(wa[-1].message),\n r'falling back to slow \\(for loop and stack\\) implementation')\n expected = torch.index_add(\n x, dim + 1, index, values.view(B0, 3, 1, 13))\n self.assertEqual(result, expected)\n\n run_test(batch_size=5)\n run_test(batch_size=1237)\n\n\ndef slice_inputs(inputs, bdims, i):\n result = []\n for inp, bdim in zip(inputs, bdims):\n if bdim is None:\n result.append(inp)\n else:\n result.append(inp.select(bdim, i))\n return tuple(result)\n\n\ndef reference_vmap(op, inputs, in_dims=0, out_dims=0):\n if isinstance(in_dims, int):\n in_dims = (in_dims,) * len(inputs)\n bdim_sizes = [inp.size(dim) for inp, dim in zip(inputs, in_dims) if dim is not None]\n assert all(bdim_size == bdim_sizes[0] for bdim_size in bdim_sizes)\n bdim_size = bdim_sizes[0]\n results = tuple(op(*slice_inputs(inputs, in_dims, i)) for i in range(bdim_size))\n # reference_vmap only supports functions that return a single Tensor output\n assert all(isinstance(result, torch.Tensor) for result in results)\n if isinstance(out_dims, int):\n out_dims = (out_dims,) * 1\n return torch.stack(results, dim=out_dims[0])\n\n\nclass TestVmapOperators(TestCase):\n def _vmap_view_test(self, op, inputs, in_dims=0, out_dims=0):\n result = vmap(op, in_dims, out_dims)(*inputs)\n reference_result = reference_vmap(op, inputs, in_dims, out_dims)\n self.assertEqual(result, reference_result)\n self.assertEqual(result.data_ptr() - result.storage_offset() * result.element_size(),\n inputs[0].data_ptr(),\n msg=\"result was not a view of the first input!\")\n\n # Assuming input[0] is a floating-point tensor. Check if the vmap\n # operation propagates the requires_grad flag. Some vmap operators are\n # implemented in a way that assumes that they are composite with respect\n # to autograd. If the operator ever is changed to not be composite with\n # respect to autograd, then the following check should fail.\n inputs_clone = list(inputs)\n inputs_clone[0] = inputs[0].clone().requires_grad_()\n result = vmap(op, in_dims, out_dims)(*inputs_clone)\n self.assertTrue(result.requires_grad)\n\n def test_diagonal(self):\n tensor = torch.randn(3, 5, 7, 11, 13)\n test = self._vmap_view_test\n op = torch.diagonal\n test(op, (tensor, 1, 0, 1), in_dims=(0, None, None, None))\n test(op, (tensor, 0, 2, -1), in_dims=(0, None, None, None))\n test(op, (tensor, 2, 1, 2), in_dims=(1, None, None, None))\n test(op, (tensor, 0, -2, -1), in_dims=(1, None, None, None), out_dims=1)\n test(vmap(lambda t: op(t, 0, 0, -1)), (tensor,), in_dims=1, out_dims=1)\n test(vmap(vmap(lambda t: op(t, 0, 0, 1), in_dims=1), in_dims=3),\n (tensor,), in_dims=1, out_dims=1)\n\n def test_expand_as(self):\n op = torch.Tensor.expand_as\n test = self._vmap_view_test\n B0, B1, B2 = 7, 11, 13\n test(op, (torch.rand(B0, 1, 5), torch.rand(B0, 2, 3, 5)))\n test(op, (torch.rand(B0, 1, 5), torch.rand(2, 3, 5)), in_dims=(0, None))\n test(op, (torch.rand(1, 5), torch.rand(B0, 2, 3, 5)), in_dims=(None, 0))\n test(vmap(op), (torch.rand(B0, B1, 1, 5), torch.rand(B0, B1, 2, 3, 5)))\n test(vmap(op), (torch.rand(B0, B1, 1, 5), torch.rand(B1, B0, 2, 3, 5)), in_dims=(0, 1))\n test(vmap(op), (torch.rand(B0, B1), torch.rand(B1, 2, 3, 5)), in_dims=(0, None))\n test(vmap(vmap(op)), (torch.rand(B0, B1, B2), torch.rand(B0, B1, B2, 2, 3, 5)))\n\n def test_select(self):\n op = torch.select\n test = self._vmap_view_test\n B0, B1, B2 = 7, 11, 13\n test(op, (torch.rand(B0, 2, 5), 0, 0), in_dims=(0, None, None))\n test(op, (torch.rand(2, B0, 5), 1, 1), in_dims=(1, None, None))\n test(vmap(lambda t: op(t, 1, 1)), (torch.rand(B1, 2, B0, 5),), in_dims=2)\n test(vmap(vmap(lambda t: op(t, 1, 1), in_dims=1)), (torch.rand(B1, 2, B0, B2, 5),), in_dims=2)\n\n def test_slice(self):\n test = self._vmap_view_test\n B0, B1, B2 = 7, 11, 13\n test(lambda t: t[0:1], (torch.rand(B0, 3, 5),))\n test(lambda t: t[:, 1:3], (torch.rand(3, 5, B0),), in_dims=2)\n test(vmap(lambda t: t[:, 0:1], in_dims=2), (torch.rand(3, 5, B0, B1),), in_dims=2)\n test(vmap(vmap(lambda t: t[0:1], in_dims=2), in_dims=2),\n (torch.rand(3, 5, B0, B1, B2),), in_dims=2)\n\n def test_t(self):\n op = torch.t\n test = self._vmap_view_test\n B0, B1, B2 = 7, 11, 13\n test(op, (torch.rand(B0, 2, 5),))\n test(op, (torch.rand(2, B0, 5),), in_dims=1)\n test(vmap(op), (torch.rand(B1, 2, B0, 5),), in_dims=2)\n test(vmap(vmap(op, in_dims=2)), (torch.rand(B1, 2, B0, 5, B2),), in_dims=2)\n\n\nif __name__ == '__main__':\n run_tests()\n", "sub_path": "test/test_vmap.py", "file_name": "test_vmap.py", "file_ext": "py", "file_size_in_byte": 23510, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "torch.testing._internal.common_utils.TestCase", "line_number": 6, "usage_type": "name"}, {"api_name": "torch.vmap", "line_number": 9, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 9, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 18, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 22, "usage_type": "attribute"}, {"api_name": "torch.randn", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 54, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 55, "usage_type": "attribute"}, {"api_name": "torch.randn", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 80, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 93, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 95, "usage_type": "attribute"}, {"api_name": "torch.vmap", "line_number": 98, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 98, "usage_type": "attribute"}, {"api_name": "torch.randn", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 103, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 104, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 108, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 109, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 114, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 115, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 116, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 121, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 123, "usage_type": "call"}, {"api_name": "torch.as_strided", "line_number": 123, "usage_type": "attribute"}, {"api_name": "torch.vmap", "line_number": 127, "usage_type": "call"}, {"api_name": "torch.var_mean", "line_number": 127, "usage_type": "attribute"}, {"api_name": "torch.vmap", "line_number": 131, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 131, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 136, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 136, "usage_type": "attribute"}, {"api_name": "torch.randn", "line_number": 142, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 145, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 149, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 150, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 155, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 156, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 161, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 162, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 167, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 168, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 169, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 176, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 177, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 183, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 184, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 185, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 197, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 198, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 199, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 202, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 212, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 215, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 220, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 225, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 230, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 231, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 232, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 241, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 242, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 243, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 248, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 250, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 252, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 254, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 256, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 260, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 264, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 266, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 270, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 272, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 279, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 281, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 283, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 286, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 289, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 293, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 294, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 295, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 295, "usage_type": "attribute"}, {"api_name": "torch.vmap", "line_number": 297, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 297, "usage_type": "attribute"}, {"api_name": "torch.randn", "line_number": 301, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 302, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 305, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 305, "usage_type": "attribute"}, {"api_name": "torch.vmap", "line_number": 310, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 310, "usage_type": "attribute"}, {"api_name": "torch.rand", "line_number": 314, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 315, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 316, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 316, "usage_type": "attribute"}, {"api_name": "torch.randn", "line_number": 320, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 323, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 328, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 337, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 341, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 346, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 350, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 351, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 354, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 354, "usage_type": "attribute"}, {"api_name": "torch.vmap", "line_number": 356, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 356, "usage_type": "attribute"}, {"api_name": "torch.vmap", "line_number": 358, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 358, "usage_type": "attribute"}, {"api_name": "torch.vmap", "line_number": 360, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 360, "usage_type": "attribute"}, {"api_name": "torch.randn", "line_number": 363, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 364, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 368, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 368, "usage_type": "attribute"}, {"api_name": "torch.vmap", "line_number": 370, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 370, "usage_type": "attribute"}, {"api_name": "torch.vmap", "line_number": 372, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 372, "usage_type": "attribute"}, {"api_name": "torch.randn", "line_number": 377, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 378, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 379, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 389, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 391, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 393, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 402, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 403, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 408, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 410, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 415, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 417, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 422, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 422, "usage_type": "attribute"}, {"api_name": "torch.vmap", "line_number": 424, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 424, "usage_type": "attribute"}, {"api_name": "torch.vmap", "line_number": 426, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 426, "usage_type": "attribute"}, {"api_name": "torch.vmap", "line_number": 434, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 434, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 436, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 436, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 438, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 438, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 440, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 440, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 442, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 442, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 443, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 443, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 449, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 450, "usage_type": "call"}, {"api_name": "warnings.catch_warnings", "line_number": 453, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 454, "usage_type": "call"}, {"api_name": "torch.sub", "line_number": 454, "usage_type": "attribute"}, {"api_name": "torch.randn", "line_number": 461, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 462, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 463, "usage_type": "call"}, {"api_name": "torch.sub", "line_number": 463, "usage_type": "attribute"}, {"api_name": "torch.randn", "line_number": 467, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 468, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 469, "usage_type": "call"}, {"api_name": "torch.sub", "line_number": 469, "usage_type": "attribute"}, {"api_name": "torch.randn", "line_number": 473, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 474, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 475, "usage_type": "call"}, {"api_name": "torch.sub", "line_number": 475, "usage_type": "attribute"}, {"api_name": "torch.randn", "line_number": 484, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 486, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 487, "usage_type": "call"}, {"api_name": "warnings.catch_warnings", "line_number": 489, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 490, "usage_type": "call"}, {"api_name": "torch.index_add", "line_number": 490, "usage_type": "attribute"}, {"api_name": "torch.index_add", "line_number": 494, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 520, "usage_type": "attribute"}, {"api_name": "torch.stack", "line_number": 523, "usage_type": "call"}, {"api_name": "torch.testing._internal.common_utils.TestCase", "line_number": 526, "usage_type": "name"}, {"api_name": "torch.vmap", "line_number": 528, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 542, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 546, "usage_type": "call"}, {"api_name": "torch.diagonal", "line_number": 548, "usage_type": "attribute"}, {"api_name": "torch.vmap", "line_number": 553, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 554, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 558, "usage_type": "attribute"}, {"api_name": "torch.rand", "line_number": 561, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 562, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 563, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 564, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 564, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 565, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 565, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 566, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 566, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 567, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 567, "usage_type": "call"}, {"api_name": "torch.select", "line_number": 570, "usage_type": "attribute"}, {"api_name": "torch.rand", "line_number": 573, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 574, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 575, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 575, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 576, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 576, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 581, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 582, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 583, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 583, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 584, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 585, "usage_type": "call"}, {"api_name": "torch.t", "line_number": 588, "usage_type": "attribute"}, {"api_name": "torch.rand", "line_number": 591, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 592, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 593, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 593, "usage_type": "call"}, {"api_name": "torch.vmap", "line_number": 594, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 594, "usage_type": "call"}, {"api_name": "torch.testing._internal.common_utils.run_tests", "line_number": 598, "usage_type": "call"}]}
+{"seq_id": "198890985", "text": "from app.newHope import bp, helpers\nfrom app.core.parser.copart.helpers.models_dict import cars_models\nfrom app.core.helpers.copart_helpers import CopartHelpers\nfrom app.forms.header_index_search_form import HeaderIndexForm\nfrom app.newHope.form.main import HomePageSearchForm\nfrom flask import request, render_template, redirect, jsonify\nfrom app.models.PostsModel import PostsModel\n\n@bp.route('/about')\ndef about_page():\n \"\"\"Main route for About Page\"\"\"\n header_form = HeaderIndexForm()\n context = {\n 'self_url': helpers.get_full_url(request),\n 'header_form': header_form\n }\n print(context['self_url'])\n return render_template('pages/about.html', context=context)\n\n\n@bp.route('/page/')\ndef pages_routes(id):\n \"\"\" Simple page routes \"\"\"\n page = PostsModel.query.get(id)\n header_form = HeaderIndexForm()\n context = {\n 'title': page.title,\n 'text': page.text,\n 'seo_title': page.seo_name,\n 'seo_description': page.seo_description,\n 'header_form': header_form\n }\n return render_template('pages/simple-text-page.html', context=context)\n\n\n@bp.route('/detail-page')\ndef detail_page():\n \"\"\" Detail page route \"\"\"\n header_form = HeaderIndexForm()\n context = {\n 'self_url': helpers.get_full_url(request),\n 'header_form': header_form\n }\n print(context['self_url'])\n return render_template('pages/detail-page.html', context=context)\n\n@bp.route('/faq')\ndef faq():\n \"\"\" FAQ route\"\"\"\n header_form = HeaderIndexForm()\n context = {\n 'self_url': helpers.get_full_url(request),\n 'header_form': header_form\n }\n print(context['self_url'])\n return render_template('pages/faq.html', context=context)\n#\n#\n# @bp.route('/text-page')\n# def text_page():\n# \"\"\" Simple text page\"\"\"\n# header_form = HeaderIndexForm()\n# context = {\n# 'self_url': helpers.get_full_url(request),\n# 'header_form': header_form\n# }\n# print(context['self_url'])\n# return render_template('pages/text-page.html', context=context)\n", "sub_path": "app/newHope/routes/SimpleSiteRoutes.py", "file_name": "SimpleSiteRoutes.py", "file_ext": "py", "file_size_in_byte": 2068, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "app.forms.header_index_search_form.HeaderIndexForm", "line_number": 12, "usage_type": "call"}, {"api_name": "app.newHope.helpers.get_full_url", "line_number": 14, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 14, "usage_type": "argument"}, {"api_name": "app.newHope.helpers", "line_number": 14, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 18, "usage_type": "call"}, {"api_name": "app.newHope.bp.route", "line_number": 9, "usage_type": "call"}, {"api_name": "app.newHope.bp", "line_number": 9, "usage_type": "name"}, {"api_name": "app.models.PostsModel.PostsModel.query.get", "line_number": 24, "usage_type": "call"}, {"api_name": "app.models.PostsModel.PostsModel.query", "line_number": 24, "usage_type": "attribute"}, {"api_name": "app.models.PostsModel.PostsModel", "line_number": 24, "usage_type": "name"}, {"api_name": "app.forms.header_index_search_form.HeaderIndexForm", "line_number": 25, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 33, "usage_type": "call"}, {"api_name": "app.newHope.bp.route", "line_number": 21, "usage_type": "call"}, {"api_name": "app.newHope.bp", "line_number": 21, "usage_type": "name"}, {"api_name": "app.forms.header_index_search_form.HeaderIndexForm", "line_number": 39, "usage_type": "call"}, {"api_name": "app.newHope.helpers.get_full_url", "line_number": 41, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 41, "usage_type": "argument"}, {"api_name": "app.newHope.helpers", "line_number": 41, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 45, "usage_type": "call"}, {"api_name": "app.newHope.bp.route", "line_number": 36, "usage_type": "call"}, {"api_name": "app.newHope.bp", "line_number": 36, "usage_type": "name"}, {"api_name": "app.forms.header_index_search_form.HeaderIndexForm", "line_number": 50, "usage_type": "call"}, {"api_name": "app.newHope.helpers.get_full_url", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 52, "usage_type": "argument"}, {"api_name": "app.newHope.helpers", "line_number": 52, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 56, "usage_type": "call"}, {"api_name": "app.newHope.bp.route", "line_number": 47, "usage_type": "call"}, {"api_name": "app.newHope.bp", "line_number": 47, "usage_type": "name"}]}
+{"seq_id": "239245221", "text": "import sys, os\nsys.path.append(os.getcwd())\nfrom dataset.mnist import load_mnist\nimport numpy as np\nfrom PIL import Image\nimport random\n\ndef image_show(img):\n pil_img = Image.fromarray(np.uint8(img))\n pil_img.show()\n\n(x_train, t_train), (x_test, t_test) = \\\nload_mnist(flatten=True, normalize=False)\n\nr = random.randrange(10000)\nimg = x_train[r]\nlabel = t_train[r]\nprint(label)\n\nprint(img.shape)\nimg = img.reshape(28, 28)\nprint(img.shape)\n\nimage_show(img)", "sub_path": "ch/mnist_show.py", "file_name": "mnist_show.py", "file_ext": "py", "file_size_in_byte": 461, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sys.path.append", "line_number": 2, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 2, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 2, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 9, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 9, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 9, "usage_type": "call"}, {"api_name": "dataset.mnist.load_mnist", "line_number": 13, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 15, "usage_type": "call"}]}
+{"seq_id": "63792832", "text": "#!/usr/bin/python3\n\"\"\" creates new view for place objects \"\"\"\n\nfrom api.v1.views import app_views\nfrom flask import jsonify, abort, request\nfrom models import storage\nfrom models.city import City\nfrom models.place import Place\n\n\n@app_views.route('/cities//places', methods=['GET'],\n strict_slashes=False)\ndef list_all_places(city_id):\n \"\"\"Reviews list of all Places in a City\"\"\"\n city = storage.get('City', city_id)\n if city is None:\n abort(404)\n all_places = storage.all('Place').values()\n city_places = [p.to_dict() for p in all_places if p.city_id == city_id]\n return jsonify(city_places)\n\n\n@app_views.route('/places/', methods=['GET'],\n strict_slashes=False)\ndef list_place(place_id):\n \"\"\"Retrieves place object \"\"\"\n place = storage.get('Place', place_id)\n if place is None:\n abort(404)\n return jsonify(place.to_dict())\n\n\n@app_views.route('/places/', methods=['DELETE'],\n strict_slashes=False)\ndef delete_place(place_id):\n \"\"\"Delete a place object \"\"\"\n place = storage.get('Place', place_id)\n if place is None:\n abort(404)\n storage.delete(place)\n storage.save()\n return jsonify({}), 200\n\n\n@app_views.route('/cities//places', methods=['POST'],\n strict_slashes=False)\ndef create_place(city_id):\n \"\"\"Adds another object to the storage\"\"\"\n city = storage.get(\"City\", city_id)\n if city is None:\n abort(404)\n new_place_dict = request.get_json(silent=True)\n if new_place_dict is None:\n return jsonify({\"error\": \"Not a JSON\"}), 400\n elif 'name' not in request.json:\n return jsonify({\"error\": \"Missing name\"}), 400\n elif 'user_id' not in request.json:\n return jsonify({\"error\": \"Missing user_id\"}), 400\n user_id = new_place_dict['user_id']\n user = storage.get(\"User\", user_id)\n if user is None:\n abort(404)\n new_place_dict['city_id'] = city_id\n new_place = Place(**new_place_dict)\n storage.new(new_place)\n storage.save()\n return jsonify(new_place.to_dict()), 201\n\n\n@app_views.route('/places_search', methods=['POST'])\ndef search_places():\n \"\"\" search places depending on state and city and amenity id\"\"\"\n form = request.get_json(force=True)\n place_list = []\n all_cities = []\n all_amenities = []\n all_places = []\n if len(form) == 0:\n all_places = storage.all('Place')\n for place in all_places.values():\n place_list.append(place.to_dict())\n return jsonify(place_list), 200\n if 'cities' in request.json:\n for city_id in form['cities']:\n all_cities.append(storage.get('City', 'city_id'))\n if 'states' in request.json:\n for state_id in form['states']:\n for city in storage.get('State', state_id).cities:\n if city not in all_cities:\n all_cities.append(city)\n for city in all_cities:\n for place in city.places:\n all_places.append(place)\n\n if 'amenities' in request.json and len(all_places) != 0:\n for amenity_id in form['amenities']:\n all_amenities.append(storage.get('Amenity', amenity_id))\n for amenity in all_amenities:\n for place in all_places:\n if place not in amenity.place_amenities:\n all_places.remove(place)\n if 'amenities' in request.json and len(all_places) == 0:\n for amenity_id in form['amenities']:\n all_amenities.append(storage.get('Amenity', amenity_id))\n for amenity in all_amenities:\n for place in amenity.place_amenities:\n place_list.append(place.to_dict())\n return jsonify(place_list), 200\n\n for place in all_places:\n place_list.append(place.to_dict())\n return jsonify(place_list), 200\n\n\n@app_views.route('/places/', methods=['PUT'],\n strict_slashes=False)\ndef update_place(place_id):\n \"\"\"Updates an instance of Place\"\"\"\n update_place_json = request.get_json(silent=True)\n if update_place_json is None:\n return jsonify({'error': 'Not a JSON'}), 400\n place = storage.get('Place', place_id)\n if place is None:\n abort(404)\n ignore = ['id', 'created_at', 'updated_at', 'city_id', 'user_id']\n for k, v in update_place_json.items():\n if k not in ignore:\n setattr(place, k, v)\n storage.save()\n return jsonify(place.to_dict()), 200\n", "sub_path": "api/v1/views/places.py", "file_name": "places.py", "file_ext": "py", "file_size_in_byte": 4465, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "models.storage.get", "line_number": 15, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 15, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 17, "usage_type": "call"}, {"api_name": "models.storage.all", "line_number": 18, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 18, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 20, "usage_type": "call"}, {"api_name": "api.v1.views.app_views.route", "line_number": 11, "usage_type": "call"}, {"api_name": "api.v1.views.app_views", "line_number": 11, "usage_type": "name"}, {"api_name": "models.storage.get", "line_number": 27, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 27, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 30, "usage_type": "call"}, {"api_name": "api.v1.views.app_views.route", "line_number": 23, "usage_type": "call"}, {"api_name": "api.v1.views.app_views", "line_number": 23, "usage_type": "name"}, {"api_name": "models.storage.get", "line_number": 37, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 37, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 39, "usage_type": "call"}, {"api_name": "models.storage.delete", "line_number": 40, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 40, "usage_type": "name"}, {"api_name": "models.storage.save", "line_number": 41, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 41, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 42, "usage_type": "call"}, {"api_name": "api.v1.views.app_views.route", "line_number": 33, "usage_type": "call"}, {"api_name": "api.v1.views.app_views", "line_number": 33, "usage_type": "name"}, {"api_name": "models.storage.get", "line_number": 49, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 49, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 51, "usage_type": "call"}, {"api_name": "flask.request.get_json", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 52, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 54, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 55, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 55, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 56, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 57, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 57, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 58, "usage_type": "call"}, {"api_name": "models.storage.get", "line_number": 60, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 60, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 62, "usage_type": "call"}, {"api_name": "models.place.Place", "line_number": 64, "usage_type": "call"}, {"api_name": "models.storage.new", "line_number": 65, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 65, "usage_type": "name"}, {"api_name": "models.storage.save", "line_number": 66, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 66, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 67, "usage_type": "call"}, {"api_name": "api.v1.views.app_views.route", "line_number": 45, "usage_type": "call"}, {"api_name": "api.v1.views.app_views", "line_number": 45, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 73, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 73, "usage_type": "name"}, {"api_name": "models.storage.all", "line_number": 79, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 82, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 83, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 83, "usage_type": "name"}, {"api_name": "models.storage.get", "line_number": 85, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 85, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 86, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 86, "usage_type": "name"}, {"api_name": "models.storage.get", "line_number": 88, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 88, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 95, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 95, "usage_type": "name"}, {"api_name": "models.storage.get", "line_number": 97, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 97, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 102, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 102, "usage_type": "name"}, {"api_name": "models.storage.get", "line_number": 104, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 104, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 108, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 112, "usage_type": "call"}, {"api_name": "api.v1.views.app_views.route", "line_number": 70, "usage_type": "call"}, {"api_name": "api.v1.views.app_views", "line_number": 70, "usage_type": "name"}, {"api_name": "flask.request.get_json", "line_number": 119, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 119, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 121, "usage_type": "call"}, {"api_name": "models.storage.get", "line_number": 122, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 122, "usage_type": "name"}, {"api_name": "flask.abort", "line_number": 124, "usage_type": "call"}, {"api_name": "models.storage.save", "line_number": 129, "usage_type": "call"}, {"api_name": "models.storage", "line_number": 129, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 130, "usage_type": "call"}, {"api_name": "api.v1.views.app_views.route", "line_number": 115, "usage_type": "call"}, {"api_name": "api.v1.views.app_views", "line_number": 115, "usage_type": "name"}]}
+{"seq_id": "78170356", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon May 24 15:08:21 2021\n\n@author: Mark Barbet\n\"\"\"\n\n\nimport os\n#from typing import final\nimport numpy as np\nimport re\nimport cantera as ct\nimport copy\nimport pandas as pd\n#import progressbar, time, sys\nimport time, sys\nimport enlighten\nimport yaml\nimport doe_object as dobj\n\nclass ranking():\n \n def __init__(self,doe_obj:dobj.doe_object):\n self.inputs=doe_obj.input_options\n \n # def __init__(self,module0=None,module1=None,settings={'batch-size':10,\n # 'total_new_exp':10,\n # 'output-csv':'output.csv'}):\n \n self.subloopBool=True\n #self.module0=module0\n #self.module1=module1\n self.doe_obj=doe_obj\n self.settings={}\n self.settings['batch-size']=doe_obj.input_options['batch-size']\n self.settings['total_new_exp']=doe_obj.input_options['total_new_exp']\n self.settings['output-csv']=doe_obj.input_options['output-csv']\n self.excluded_yamls=[]\n total_iters=int(np.ceil(self.settings['total_new_exp']/self.settings['batch-size']))\n #print(total_iters)\n #total_exps=0\n final_exp_dataframe=pd.DataFrame(columns=['experiment','ratio'])\n #current_yamls=copy.deepcopy(self.module1.yaml_file_list)\n S_countH=0\n S_countV=0\n new_Z=copy.deepcopy(self.doe_obj.Z_original)\n new_Y=copy.deepcopy(self.doe_obj.Y_original)\n new_X_list=list(self.doe_obj.X_original['value'])\n self.updated_S=copy.deepcopy(self.doe_obj.S_original)\n \n #self.printProgressBar(0, total_iters, prefix = 'Finding Best Experiments:', suffix = 'Complete', length = 70)\n #self.down()\n #total = progressbar.ProgressBar(maxval=total_iters)\n #total.start()\n self.manager = enlighten.get_manager()\n self.mainloop=self.manager.counter(total=total_iters,desc='Overall Progress',unit='batches',color='green')\n \n for i in np.arange(total_iters):\n #print(i,\"wtf\")\n if i==0:\n self.ranking=self.get_rankings(self.excluded_yamls,self.updated_S,i,countH=S_countH, countV=S_countV)\n elif i>0:\n self.ranking=self.get_rankings(self.excluded_yamls,self.updated_S,i,\n countH=S_countH, countV=S_countV,\n Z_prev=new_Z,Y_prev=new_Y,X_prev=new_X_list)\n if i+10:\n S_countV=self.experiment_length+S_countV+len(X_to_add)\n #elif i==0:\n # S_countV=S_countV+S_countH\n #print('New Count H: '+str(S_countH)+', New Count V: '+str(S_countV)+', Exp Length: '+str(self.experiment_length))\n return (S_proposed,new_Y,new_Z,new_X_list,S_countH,S_countV)\n \n \n def get_S_current_columns(self,doe_obj=None):\n colnames=list(doe_obj.X_original['value'])\n rownames=list(doe_obj.Y_original['value'])\n return (colnames,rownames)\n \n \n \n def get_Z(self,file,indexer):\n data=self.load_to_obj(file)\n previous_exp_index=int(self.rownames_nominal[-1].split('_')[-1])\n current_exp_index=previous_exp_index+1\n exclude_list=[]\n A=[]\n N=[]\n Ea=[]\n for i in np.arange(self.num_rxns):\n A=A+['A_'+str(i)]\n for i in np.arange(self.num_rxns): \n N=N+['n_'+str(i)]\n for i in np.arange(self.num_rxns):\n Ea=Ea+['Ea_'+str(i)]\n exclude_list=A+N+Ea\n #print(exclude_list)\n Z_to_add=self.doe_obj.experiment_matrices[indexer]['Z'][~self.doe_obj.experiment_matrices[indexer]['Z']['value'].isin(exclude_list)].copy()\n new_Z_names=[]\n for j,item in enumerate(list(Z_to_add['value'])):\n temp=item.split('_')\n if 'experiment' in temp[-1]:\n temp2=re.split('(\\d+)',temp[-1])\n #print(temp2)\n temp2[-2]=str(current_exp_index)\n temp[-1]=''.join(temp2)\n \n else:\n temp[-1]=str(current_exp_index)\n new_Z_names=new_Z_names+['_'.join(temp)]\n Z_to_add['value']=new_Z_names\n #Y_to_add=self.module1.matrices[indexer]['Z'][~self.module1.matrices[indexer]['Z']['value'].isin(exclude_list)]\n \n \n return Z_to_add\n \n def get_Y(self,file,indexer):\n data=self.load_to_obj(file)\n previous_exp_index=int(self.rownames_nominal[-1].split('_')[-1])\n current_exp_index=previous_exp_index+1\n exclude_list=[]\n A=[]\n N=[]\n Ea=[]\n for i in np.arange(self.num_rxns):\n A=A+['A_'+str(i)]\n for i in np.arange(self.num_rxns): \n N=N+['n_'+str(i)]\n for i in np.arange(self.num_rxns):\n Ea=Ea+['Ea_'+str(i)]\n exclude_list=A+N+Ea\n #print(exclude_list)\n Y_to_add=self.doe_obj.experiment_matrices[indexer]['Y'][~self.doe_obj.experiment_matrices[indexer]['Y']['value'].isin(exclude_list)].copy()\n new_Y_names=[]\n for j,item in enumerate(list(Y_to_add['value'])):\n temp=item.split('_')\n if 'experiment' in temp[-1]:\n temp2=re.split('(\\d+)',temp[-1])\n #print(temp2)\n temp2[-2]=str(current_exp_index)\n temp[-1]=''.join(temp2)\n \n else:\n temp[-1]=str(current_exp_index)\n new_Y_names=new_Y_names+['_'.join(temp)]\n Y_to_add['value']=new_Y_names\n #Y_to_add=self.module1.matrices[indexer]['Z'][~self.module1.matrices[indexer]['Z']['value'].isin(exclude_list)]\n \n \n return Y_to_add\n \n def get_X_names(self, Z):\n for i, item in enumerate(list(Z['value'])):\n if 'T_experiment' in item:\n start_exp=i\n break\n \n X_values=list(Z['value'])[start_exp:]\n return X_values\n \n def construct_Z_new(self,add_Z,Z):\n new_Z=pd.concat([Z,add_Z],ignore_index=True)\n return new_Z\n \n def construct_Y_new(self,add_Y,Y):\n new_Y=pd.concat([Y,add_Y],ignore_index=True)\n return new_Y\n \n # def get_S_block(self,col_min,col_max,row_min,row_max, S):\n # #print(np.shape(S))\n # #return S[row_min:row_max,col_min:col_max]\n # return S[col_min:col_max,row_min:row_max]\n def get_S_block(self,col_min,col_max,row_min,row_max, S,flag=False,X_add=None):\n #print(np.shape(S))\n #return S[row_min:row_max,col_min:col_max]\n y_len=3*self.num_rxns+len(self.inputs['observables'])+X_add\n if flag:\n #y_len=len(self.doe_obj.Y_original)\n #print(\"Fuckshit is : \"+str(y_len))\n temp_S=np.zeros((3*self.num_rxns+len(self.inputs['observables'])+X_add,3*self.num_rxns+X_add))\n indices=np.arange(y_len-len(self.inputs['observables']))\n temp_S[indices+len(self.inputs['observables']),indices]=1.0\n temp_S[0:len(self.inputs['observables']),0:np.shape(S)[1]]=S\n #print(temp_S)\n S=temp_S\n #print(col_min,col_max,row_min,row_max)\n #print(np.shape(temp_S))\n #print(np.shape(S))\n return S[col_min:col_max,row_min:row_max]\n\n # def get_new_S_chunks(self,rxn_count,X_add,Y_add,add_Z,S_current,Y,Z,S_new):\n # #print(np.shape(S_current),np.shape(S_new))\n \n \n # S1_new=self.get_S_block(0,self.experiment_length,0,3*self.num_rxns,S_new)\n # S2_new=self.get_S_block(0, self.experiment_length, 3*self.num_rxns, 3*self.num_rxns+len(X_add)+1, S_new)\n # #print(np.shape(S1_new))\n # #print(np.shape(S2_new))\n # S3_new=self.get_S_block(self.experiment_length,self.experiment_length+3*self.num_rxns,\n # 3*self.num_rxns,3*self.num_rxns+len(X_add)+1,S_new)\n # #print(np.shape(S3_new))\n # S4_new=self.get_S_block(self.experiment_length+3*self.num_rxns,\n # self.experiment_length+3*self.num_rxns+len(X_add)+1,\n # 0,3*self.num_rxns,S_new)\n # #print(np.shape(S4_new))\n # S5_new=self.get_S_block(self.experiment_length+3*self.num_rxns,\n # self.experiment_length+3*self.num_rxns+len(X_add)+1,\n # 3*self.num_rxns,3*self.num_rxns+len(X_add)+1,S_new)\n # #print(np.shape(S5_new))\n # #print(S5_new)\n # return (S1_new,S2_new,S3_new,S4_new,S5_new)\n def get_new_S_chunks(self,rxn_count,X_add,Y_add,add_Z,S_current,Y,Z,S_new):\n #print(np.shape(S_current),np.shape(S_new))\n #time.sleep(5)\n #print(np.shape(S_new))\n with open(os.path.join(os.getcwd(),'log.txt'),'a') as f:\n f.write(str(np.shape(S_new)))\n f.write('\\n')\n #print(S_new)\n #print(\"Length is: \"+str(self.experiment_length))\n S1_new=self.get_S_block(0,self.experiment_length,0,3*self.num_rxns,S_new,X_add=len(X_add))\n #print('EXP Length',self.experiment_length)\n S2_new=self.get_S_block(0, self.experiment_length, 3*self.num_rxns, 3*self.num_rxns+len(X_add)+1, S_new,X_add=len(X_add))\n #print(np.shape(S1_new))\n #print(np.shape(S2_new))\n #print(self.experiment_length,self.experiment_length+3*self.num_rxns,\n # 3*self.num_rxns,3*self.num_rxns+len(X_add)+1)\n S3_new=self.get_S_block(self.experiment_length,self.experiment_length+3*self.num_rxns,\n 3*self.num_rxns,3*self.num_rxns+len(X_add)+1,S_new,flag=True,X_add=len(X_add))\n #print(np.shape(S3_new))\n S4_new=self.get_S_block(self.experiment_length+3*self.num_rxns,\n self.experiment_length+3*self.num_rxns+len(X_add)+1,\n 0,3*self.num_rxns,S_new,flag=True,X_add=len(X_add))\n #print(np.shape(S4_new))\n S5_new=self.get_S_block(self.experiment_length+3*self.num_rxns,\n self.experiment_length+3*self.num_rxns+len(X_add)+1,\n 3*self.num_rxns,3*self.num_rxns+len(X_add)+1,S_new,flag=True,X_add=len(X_add))\n #print(np.shape(S5_new))\n #print(S5_new)\n return (S1_new,S2_new,S3_new,S4_new,S5_new)\n \n def get_exp_length(self,file):\n data=self.load_to_obj(file)\n mol_length=0\n conc_length=0\n for i,item in enumerate(data['datapoints']['mole-fraction']):\n if item['csvfile'] != None:\n mol_length=mol_length+1\n for i,item in enumerate(data['datapoints']['concentration']):\n if item['csvfile'] != None:\n conc_length=conc_length+1\n \n return mol_length+conc_length\n \n def get_prior_exp_len(self,priorY):\n #print(len(priorY))\n for i,item in enumerate(list(priorY['value'])):\n if 'A_' in item:\n end_exp_index=i\n break\n return len(list(priorY['value'])[0:end_exp_index])\n \n def get_prior_phys_param_len(self,priorY):\n #print(priorY)\n for i,item in enumerate(list(priorY['value'])):\n if 'Ea_' in item:\n final_Ea=i\n if self.inputs['MSI_settings']['rate_constant_targets'] != '':\n targets=pd.read_csv(os.path.join(self.inputs['working_dir'],\n self.inputs['MSI_settings']['rate_constant_targets']))\n target_length=len(targets['Reaction'])\n \n elif self.inputs['MSI_settings']['rate_constant_targets'] == '':\n target_length=0\n return len(list(priorY['value'])[final_Ea+1:])-target_length\n \n \n def construct_zeroes(self,S3,S1,S4,countV=0,countH=0):\n prior_exp_len=self.get_prior_exp_len(self.doe_obj.Y_original)\n Z1=np.zeros((prior_exp_len,np.shape(S3)[1]))\n prior_phys_params=self.get_prior_phys_param_len(self.doe_obj.Y_original)\n #Need to get length of targets for Z2\n if self.inputs['MSI_settings']['rate_constant_targets'] != '':\n targets=pd.read_csv(os.path.join(self.inputs['working_dir'],\n self.inputs['MSI_settings']['rate_constant_targets']))\n target_length=len(targets['Reaction'])\n \n elif self.inputs['MSI_settings']['rate_constant_targets'] == '':\n target_length=0\n \n #print(prior_phys_params,target_length)\n Z2=np.zeros((prior_phys_params+target_length+countV,np.shape(S3)[1]))\n Z3=np.zeros((np.shape(S1)[0],prior_phys_params+countH))\n Z4=np.zeros((np.shape(S4)[0],prior_phys_params+countH))\n \n return (Z1,Z2,Z3,Z4)\n \n \n \n def build_S(self,S1,S2,S3,S4,S5,S_old,countV=0,countH=0):\n \n Z1,Z2,Z3,Z4=self.construct_zeroes(S3,S1,S4,countV=countV,countH=countH)\n #print(np.shape(Z1))\n #print(np.shape(S3))\n #print(np.shape(Z2))\n block1=np.block([[Z1],[S3],[Z2]])\n block2=np.block([[S1,Z3,S2],[S4,Z4,S5]])\n #print(np.shape(Z2),np.shape(Z3),np.shape(Z4))\n \n #print('S',np.shape(S_old))\n #print('Z1',np.shape(Z1))\n # print('S3',np.shape(S3))\n # print('Z2',np.shape(Z2))\n # print('S1',np.shape(S1))\n # print('Z3',np.shape(Z3))\n # print('S2',np.shape(S2))\n # print('S4',np.shape(S4))\n # print('Z4',np.shape(Z4))\n # print('S5',np.shape(S5))\n #print(countV)\n #print(np.shape(block1))\n #print(np.shape(block2),'s')\n #print(np.shape(S_old))\n S=np.block([[S_old,block1],[block2]])\n return S\n \n def get_covariance(self,s):\n c = np.dot(np.transpose(s),s)\n c = np.linalg.inv(c)\n \n \n return (c)\n def return_posteriors(self,c,Z):\n covariance_posterior_df = pd.DataFrame(c)\n covariance_posterior_df.columns = Z\n covariance_posterior_df.reindex(labels = Z)\n posterior_diag = np.diag(c)\n posterior_sigmas = np.sqrt(posterior_diag)\n posterior_sigmas_df = pd.DataFrame({'parameter': Z,'value': posterior_sigmas.reshape((posterior_sigmas.shape[0],))})\n posterior_diag_df = pd.DataFrame({'parameter': Z,'value': posterior_diag.reshape((posterior_diag.shape[0],))})\n sorted_posterior_diag = posterior_diag_df.sort_values(by=['value'])\n \n def get_normalized_S(self,S,Z,Y):\n \n s=S/np.array(Z['Uncertainty'])[:,None]\n return s\n \n def calculate_sigmas_for_rate_constants(self,k_target_value_S_matrix,\n k_target_values_parsed_csv,\n unique_reactions,\n gas,\n covariance):\n\n \n reaction_list_from_mechanism = gas.reaction_equations()\n sigma_list_for_target_ks = [[] for reaction in range(len(unique_reactions))]\n shape = k_target_value_S_matrix.shape\n #print(unique_reactions)\n for row in range(shape[0]):\n #print(k_target_value_S_matrix,'blahasfjhbjhb')\n SC = np.dot(k_target_value_S_matrix[row,:],covariance)\n sigma_k = np.dot(SC,np.transpose(k_target_value_S_matrix[row,:]))\n sigma_k = np.sqrt(sigma_k)\n #print(row)\n #print(k_target_values_parsed_csv['Reaction'][row])\n indx = reaction_list_from_mechanism.index(k_target_values_parsed_csv['Reaction'][row])\n sigma_list_for_target_ks[unique_reactions.index(indx)].append(sigma_k)\n \n return sigma_list_for_target_ks\n \n def get_k_block(self,S):\n if self.inputs['MSI_settings']['rate_constant_targets'] != '':\n targets=pd.read_csv(os.path.join(self.inputs['working_dir'],\n self.inputs['MSI_settings']['rate_constant_targets']))\n target_length=len(targets['Reaction'])\n \n return S[len(S[:,0])-target_length:,:]\n \n def get_k_block_proposed(self,S,S_old):\n if self.inputs['MSI_settings']['rate_constant_targets'] != '':\n targets=pd.read_csv(os.path.join(self.inputs['working_dir'],\n self.inputs['MSI_settings']['rate_constant_targets']))\n target_length=len(targets['Reaction'])\n \n return S[len(S_old[:,0])-target_length:len(S_old[:,0]),:]\n \n def get_unique_elements(self,l,gas):\n \n uniques=[]\n for i,item in enumerate(l):\n index=list(gas.reaction_equations()).index(item)\n if index not in uniques:\n uniques=uniques+[index]\n return uniques\n\n\n\n\n def calculate_sigma(self,S_row,C):\n SC = np.dot(S_row,C)\n sigma = np.dot(SC,np.transpose(S_row))\n sigma = np.sqrt(sigma)\n return sigma\n\n\n def get_ignition_block(self,S):\n exp_len=self.get_prior_exp_len(self.doe_obj.Y_original)\n\n ig_block=S[exp_len-1,:]\n\n return ig_block\n\n\n def load_to_obj(self, path:str = ''):\n \"\"\"\n Takes in a file path for a yaml file and returns a dictionary of \n simulation information.\n\n Parameters\n ----------\n path : str, optional\n The path to where the yaml file is stored. The default is ''.\n\n Returns\n -------\n config : dictionary\n An unorganized dictionary that contains information reguarding\n the experiment the yaml input file was written for.\n\n \"\"\"\n with open(path) as f:\n config = yaml.load(f,Loader=yaml.FullLoader)\n return config\n\n\n def get_yaml_conditions(self,fname):\n\n template=self.load_to_obj(fname)\n outputs={}\n outputs['temperature']=float(template['common-properties']['temperature']['value-list'][0])\n outputs['pressure']=float(template['common-properties']['pressure']['value'])\n outputs['residence-time']=float(template['apparatus']['residence-time']['value'])\n for i,spec in enumerate(template['common-properties']['composition']):\n outputs[spec['species']]=spec['mole-fraction']\n \n return outputs\n \n \n def load_to_obj(self, path:str = ''):\n \"\"\"\n Takes in a file path for a yaml file and returns a dictionary of \n simulation information.\n\n Parameters\n ----------\n path : str, optional\n The path to where the yaml file is stored. The default is ''.\n\n Returns\n -------\n config : dictionary\n An unorganized dictionary that contains information reguarding\n the experiment the yaml input file was written for.\n\n \"\"\"\n with open(path) as f:\n config = yaml.load(f,Loader=yaml.FullLoader)\n return config\n \n \n def get_rankings(self,excluded_yamls,S,iteration,countH=0,countV=0,Z_prev=None,Y_prev=None,X_prev=None):\n #self.up()\n #sub = progressbar.ProgressBar(maxval=len(self.module1.yaml_file_list))\n #sub.start()\n #if self.subloopBool:\n self.subloop=self.manager.counter(total=len(self.doe_obj.experiment_matrices),desc='Possible experiments',unit='Experiments',color='red',leave=True)\n #self.subloopBool=False\n ranking_list=[]\n self.rownames_nominal,self.colnames_nominal=self.get_S_current_columns(doe_obj=self.doe_obj)\n self.mech=os.path.join(self.inputs['working_dir'],self.inputs['MSI_settings']['chemical_model'])\n gas=ct.Solution(self.mech)\n self.num_rxns=len(gas.reactions())\n if re.match('[Rr]ate[-_ ][Cc]onstant',self.inputs['quantity_of_interest']):\n k_block_og=self.get_k_block(self.doe_obj.S_original)\n targets=pd.read_csv(os.path.join(self.inputs['working_dir'],\n self.inputs['MSI_settings']['rate_constant_targets']))\n sigma_list_og=self.calculate_sigmas_for_rate_constants(k_block_og,\n targets,\n self.get_unique_elements(list(targets['Reaction']),gas),\n gas,\n self.doe_obj.covar_original)\n target_reaction=self.inputs['target_reaction']['equation']\n target_index=list(gas.reaction_equations()).index(target_reaction)\n sigma_list_index=self.get_unique_elements(list(targets['Reaction']),gas).index(target_index)\n #print(sigma_list[sigma_list_index][0])\n original_posterior=sigma_list_og[sigma_list_index][0]\n\n elif re.match('[Ii]gnition[-_ ][Dd]elay',self.inputs['quantity_of_interest']):\n '''This block collects the original uncertainty in the ignition delay quantity of interest. Gets the final experiment line\n and estimates the uncertainty in the ignition delay.'''\n\n ig_block_og=self.get_ignition_block(self.doe_obj.S_original)\n print(ig_block_og)\n #print(self.module0.)\n sigma_ig=self.calculate_sigma(ig_block_og,self.doe_obj.covar_original)\n original_posterior=sigma_ig\n\n final_yamls=[]\n for i,file in enumerate(self.doe_obj.experiment_yamls):\n \n if file not in excluded_yamls:\n final_yamls.append(file)\n print('File is:'+str(file))\n data=self.load_to_obj(os.path.join(self.inputs['working_dir'],file))\n #parametersX,parametersY,parametersZ=self.get_experiment_columns(os.path.join(self.module1.input_options['working_dir'],file),i)\n parametersZ=self.get_Z(os.path.join(self.inputs['working_dir'],file),i)\n parametersY=self.get_Y(os.path.join(self.inputs['working_dir'],file),i)\n if iteration==0:\n \n #print(parametersZ)\n self.experiment_length=self.get_exp_length(os.path.join(self.inputs['working_dir'],file))\n #print(self.experiment_length)\n X_to_add=self.get_X_names(parametersZ)\n \n new_Z=self.construct_Z_new(parametersZ,self.doe_obj.Z_original)\n \n new_Y=self.construct_Y_new(parametersY,self.doe_obj.Y_original)\n \n elif iteration>0:\n new_Z=self.construct_Z_new(parametersZ,Z_prev)\n X_to_add=self.get_X_names(parametersZ)\n new_Y=self.construct_Y_new(parametersY,Y_prev)\n self.experiment_length=self.get_exp_length(os.path.join(self.inputs['working_dir'],file))\n S1_new,S2_new,S3_new,S4_new,S5_new=self.get_new_S_chunks(self.num_rxns,X_to_add,parametersY,parametersZ,\n self.doe_obj.S_original,\n self.doe_obj.Y_original,\n self.doe_obj.Z_original,\n self.doe_obj.experiment_matrices[i]['S'])\n #print(self.module0.initial_optimization.Y_data_frame['value'][635:])\n #print('CountH: '+str(countH)+', countV: '+str(countV))\n S_proposed=self.build_S(S1_new,S2_new,S3_new,S4_new,S5_new,S,countH=countH,countV=countV)\n\n if iteration==0:\n new_X_list=list(self.doe_obj.X_original['value'])+X_to_add\n elif iteration>0:\n new_X_list=X_prev+X_to_add\n #print(list(self.module0.initial_optimization.z_data_frame['value']))\n #print(np.shape(S_proposed))\n #print(self.module0.initial_optimization.z_data_frame)\n #print(len(new_Z),np.shape(S_proposed))\n \n s=self.get_normalized_S(S_proposed, new_Z, new_Y)\n c=self.get_covariance(s)\n #print(c)\n self.updated_c=c\n #print(np.shape(c))\n if re.match('[Rr]ate[-_ ][Cc]onstant',self.inputs['quantity_of_interest']):\n k_block=self.get_k_block_proposed(S_proposed,self.doe_obj.S_original)\n targets=pd.read_csv(os.path.join(self.inputs['working_dir'],\n self.inputs['MSI_settings']['rate_constant_targets']))\n sigma_list=self.calculate_sigmas_for_rate_constants(k_block,\n targets,\n self.get_unique_elements(list(targets['Reaction']),gas),\n gas,\n c)\n #print(sigma_list)\n target_reaction=self.inputs['target_reaction']['equation']\n target_index=list(gas.reaction_equations()).index(target_reaction)\n sigma_list_index=self.get_unique_elements(list(targets['Reaction']),gas).index(target_index)\n #print(sigma_list[sigma_list_index][0],'text')\n ranking_list=ranking_list+[sigma_list[sigma_list_index][0]/original_posterior]\n #print(sigma_list,'poo')\n elif re.match('[Ii]gnition[-_ ][Dd]elay',self.inputs['quantity_of_interest']):\n print('Entered correct statement')\n ig_block=self.get_ignition_block(S_proposed)\n if i==len(self.doe_obj.experiment_yamls)-1:\n print(ig_block)\n sigma=self.calculate_sigma(ig_block,c)\n ranking_list=ranking_list+[sigma/original_posterior]\n #elif re.match('[Ii]gnition[_ -][Dd]elay',self.module0.startup_data['quantity_of_interest']):\n #if np.remainder(i,10)==0:\n #sub.update(i)\n self.subloop.update()\n #sub.finish()\n #print(ranking_list)\n output_ranking=pd.DataFrame(columns=['experiment','ratio'])\n output_ranking['experiment']=final_yamls\n temps=[]\n pres=[]\n restime=[]\n conds=self.get_yaml_conditions(os.path.join(self.inputs['working_dir'],final_yamls[0]))\n species_index=copy.deepcopy(list(conds.keys()))\n #print(list(conds.keys()).remove('temperature'))\n species_index.remove('temperature')\n species_index.remove('pressure')\n species_index.remove('residence-time')\n specs={}\n for i,spec in enumerate(species_index):\n specs[spec]=[]\n for i,fname in enumerate(final_yamls):\n conds=self.get_yaml_conditions(os.path.join(self.inputs['working_dir'],fname))\n temps.append(conds['temperature'])\n pres.append(conds['pressure'])\n restime.append(conds['residence-time'])\n for k,spec in enumerate(species_index):\n specs[spec].append(conds[spec])\n #print(ranking_list)\n output_ranking['ratio']=ranking_list\n output_ranking['temperature']=temps\n output_ranking['pressure']=pres\n output_ranking['residence-time']=restime\n for i,spec in enumerate(species_index):\n output_ranking[spec]=specs[spec]\n output_ranking.sort_values(by='ratio',ascending=True,inplace=True)\n\n #output_ranking.to_csv(os.path.join(self.module0.startup_data['working_dir'],\n #'output_rankings.csv'),index=False)\n #print(output_ranking)\n #print(output_ranking)\n return output_ranking\n #posteriors=self.return_posteriors(c,new_Z)\n #posteriors.to_csv('test.csv')\n #print(c)\n #print(new_S[687:])\n #print(X_to_add)\n #print(parametersY)\n #print('x',parametersX)\n #print('y',parametersY)\n #print('z',parametersZ)", "sub_path": "ranking_method1.py", "file_name": "ranking_method1.py", "file_ext": "py", "file_size_in_byte": 34968, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "doe_object.doe_object", "line_number": 24, "usage_type": "attribute"}, {"api_name": "numpy.ceil", "line_number": 40, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 43, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 47, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 48, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 50, "usage_type": "call"}, {"api_name": "enlighten.get_manager", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.remainder", "line_number": 67, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.remainder", "line_number": 70, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.remainder", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path", "line_number": 94, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path", "line_number": 96, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 106, "usage_type": "call"}, {"api_name": "os.path", "line_number": 106, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path", "line_number": 112, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path", "line_number": 115, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 120, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 120, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 121, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 121, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 125, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 125, "usage_type": "attribute"}, {"api_name": "sys.stdout.flush", "line_number": 126, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 126, "usage_type": "attribute"}, {"api_name": "copy.deepcopy", "line_number": 157, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 160, "usage_type": "call"}, {"api_name": "os.path", "line_number": 160, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 162, "usage_type": "call"}, {"api_name": "os.path", "line_number": 162, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 164, "usage_type": "call"}, {"api_name": "os.path", "line_number": 164, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 207, "usage_type": "call"}, {"api_name": "re.split", "line_number": 216, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 238, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 240, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 242, "usage_type": "call"}, {"api_name": "re.split", "line_number": 251, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 275, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 279, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 293, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 294, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 296, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 329, "usage_type": "call"}, {"api_name": "os.path", "line_number": 329, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 329, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 330, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 382, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 382, "usage_type": "call"}, {"api_name": "os.path", "line_number": 382, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 393, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 393, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 397, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 397, "usage_type": "call"}, {"api_name": "os.path", "line_number": 397, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 405, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 405, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 406, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 406, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 407, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 407, "usage_type": "call"}, {"api_name": "numpy.block", "line_number": 419, "usage_type": "call"}, {"api_name": "numpy.block", "line_number": 420, "usage_type": "call"}, {"api_name": "numpy.block", "line_number": 437, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 441, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 441, "usage_type": "call"}, {"api_name": "numpy.linalg.inv", "line_number": 442, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 442, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 447, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 450, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 451, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 452, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 453, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 458, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 474, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 475, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 475, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 476, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 486, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 486, "usage_type": "call"}, {"api_name": "os.path", "line_number": 486, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 494, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 494, "usage_type": "call"}, {"api_name": "os.path", "line_number": 494, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 513, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 514, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 514, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 515, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 545, "usage_type": "call"}, {"api_name": "yaml.FullLoader", "line_number": 545, "usage_type": "attribute"}, {"api_name": "yaml.load", "line_number": 580, "usage_type": "call"}, {"api_name": "yaml.FullLoader", "line_number": 580, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 593, "usage_type": "call"}, {"api_name": "os.path", "line_number": 593, "usage_type": "attribute"}, {"api_name": "cantera.Solution", "line_number": 594, "usage_type": "call"}, {"api_name": "re.match", "line_number": 596, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 598, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 598, "usage_type": "call"}, {"api_name": "os.path", "line_number": 598, "usage_type": "attribute"}, {"api_name": "re.match", "line_number": 611, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 627, "usage_type": "call"}, {"api_name": "os.path", "line_number": 627, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 629, "usage_type": "call"}, {"api_name": "os.path", "line_number": 629, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 630, "usage_type": "call"}, {"api_name": "os.path", "line_number": 630, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 634, "usage_type": "call"}, {"api_name": "os.path", "line_number": 634, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 646, "usage_type": "call"}, {"api_name": "os.path", "line_number": 646, "usage_type": "attribute"}, {"api_name": "re.match", "line_number": 670, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 672, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 672, "usage_type": "call"}, {"api_name": "os.path", "line_number": 672, "usage_type": "attribute"}, {"api_name": "re.match", "line_number": 686, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 699, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 704, "usage_type": "call"}, {"api_name": "os.path", "line_number": 704, "usage_type": "attribute"}, {"api_name": "copy.deepcopy", "line_number": 705, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 714, "usage_type": "call"}, {"api_name": "os.path", "line_number": 714, "usage_type": "attribute"}]}
+{"seq_id": "125262609", "text": "# #############################################################################\n# Author: 2017 Rostislav Spinar #\n# IQRF Tech s.r.o. #\n# #############################################################################\n\n# #############################################################################\n# #\n# sudo apt-get install python-dev build-essential #\n# sudo apt-get install python-pip #\n# sudo pip install paho-mqtt # \n# #\n# #############################################################################\n\nimport sys\nimport getopt\nimport time\nimport json\n\nimport paho.mqtt.client as paho\n \ndef on_connect(client, userdata, flags, rc):\n print('CONNACK received with code %d.' % (rc))\n\ndef on_publish(client, userdata, mid):\n print('mid: ' + str(mid))\n\ndef on_subscribe(client, userdata, mid, granted_qos):\n print('Subscribed: ' + str(mid) + ' ' + str(granted_qos))\n \ndef on_message(client, userdata, msg):\n print(msg.topic + ' ' + str(msg.qos) + ' ' + str(msg.payload))\n\ndef on_log(mqttc, userdata, level, string):\n print(string)\n\ndef create_dpa_frame(node_id, pnum, pcmd, hwpid, data=[]):\n\n byte_str = '%02x.%02x.%02x.%02x.%02x.%02x' % (node_id & 0xFF,\n node_id >> 8,\n pnum, pcmd,\n hwpid & 0xFF,\n hwpid >> 8)\n for i in data:\n byte_str += '.%02x' % i\n\n return byte_str\n\ndef create_dpa_json(msg_id, dpa_frame):\n request = {}\n request['ctype'] = 'dpa'\n request['type'] = 'raw'\n request['msgid'] = msg_id\n request['request'] = dpa_frame\n request['request_ts'] = ''\n request['confirmation'] = ''\n request['confirmation_ts'] = ''\n request['response'] = ''\n request['response_ts'] = ''\n\n return json.dumps(request)\n\ndef print_usage():\n print('iqrf_daemon_mqtt.py [-d] [-h hostname] [-p port] [-tp topic_pub] [-ts topic_sub]')\n\ndef main(argv):\n #IQRF\n # default hwpid\n hwpid = 0xffff\n # default DPA timeout (in miliseconds)\n timeout = 1000\n\n #MQTT\n host = 'localhost'\n port = 1883\n keepalive = 60\n client_id = str(time.time())\n #password = None\n #username = None\n debug=False\n\n topic_pub = 'Iqrf/DpaRequest'\n topic_sub = 'Iqrf/DpaResponse'\n\n try:\n opts, args = getopt.getopt(argv, 'd:h:p:tp:ts', ['debug', 'host=', 'port=', 'topic_pub=', 'topic_sub='])\n except getopt.GetoptError as s:\n print_usage()\n sys.exit(2)\n \n for opt, arg in opts:\n if opt in ('-d', '--debug'):\n host = arg\n elif opt in ('-h', '--host'):\n host = arg\n elif opt in ('-p', '--port'):\n port = int(arg)\n elif opt in ('-tp', '--topic_pub'):\n topic_pub = arg\n print(topic_pub)\n elif opt in ('-ts', '--topic_sub'):\n topic_sub = arg\n print(topic_sub)\n\n # client\n client = paho.Client(client_id=client_id, clean_session=True, userdata=None, protocol=paho.MQTTv31)\n #client.username_pw_set(username, password)\n #client.tls_set(“/path/to/ca.crt”)\n\n # client callbacks\n client.on_connect = on_connect\n client.on_publish = on_publish\n client.on_subscribe = on_subscribe\n client.on_message = on_message\n\n # debug\n if debug:\n client.on_log = on_log\n\n # connect\n client.connect(host=host, port=port, keepalive=keepalive, bind_address='')\n\n # subscribe\n client.subscribe(topic_sub)\n\n # blocking, good for sub only\n #client.loop_forever()\n\n # not blocking, background thread, returns\n client.loop_start()\n #client.loop_stop()\n\n # dpa frame\n dpa_frame = create_dpa_frame(0x0f, 0x06, 0x03, hwpid)\n\n while True:\n # json dpa\n msg_id = str(time.time())\n json_dpa = create_dpa_json(msg_id, dpa_frame)\n \n # publish\n (rc, mid) = client.publish(topic_pub, json_dpa, qos=1)\n \n # sleep\n time.sleep(10)\n\nif __name__ == \"__main__\":\n main(sys.argv[1:])\n", "sub_path": "python/mqtt/iqrf_daemon_mqtt.py", "file_name": "iqrf_daemon_mqtt.py", "file_ext": "py", "file_size_in_byte": 4209, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "json.dumps", "line_number": 60, "usage_type": "call"}, {"api_name": "time.time", "line_number": 76, "usage_type": "call"}, {"api_name": "getopt.getopt", "line_number": 85, "usage_type": "call"}, {"api_name": "getopt.GetoptError", "line_number": 86, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 88, "usage_type": "call"}, {"api_name": "paho.mqtt.client.Client", "line_number": 105, "usage_type": "call"}, {"api_name": "paho.mqtt.client", "line_number": 105, "usage_type": "name"}, {"api_name": "paho.mqtt.client.MQTTv31", "line_number": 105, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 137, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 144, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 147, "usage_type": "attribute"}]}
+{"seq_id": "402945923", "text": "import os\nfrom functools import partial\nfrom multiprocessing.dummy import Pool\nfrom subprocess import call\nfrom termcolor import colored\n\n\nbase_dir = '/data/SUNCG'\nwith open(os.path.join(base_dir, 'train_sceneId.txt')) as file:\n train_ids = file.read().splitlines()\nwith open(os.path.join(base_dir, 'test_sceneId.txt')) as file:\n test_ids = file.read().splitlines()\nscene_ids = train_ids + test_ids\ncommands = ['./process_houses.sh %s' % os.path.join(base_dir, 'house', scene_id) for scene_id in scene_ids]\n\nnproc = 8\npool = Pool(nproc)\nfor idx, return_code in enumerate(pool.imap(partial(call, shell=True), commands)):\n if return_code != 0:\n print(colored('Command \\\"%s\\\" failed' % commands[idx], 'yellow'))\n else:\n print(colored('-- Processed %d/%d' % (idx + 1, len(commands)), 'white', 'on_blue'))\n", "sub_path": "parallel_process.py", "file_name": "parallel_process.py", "file_ext": "py", "file_size_in_byte": 827, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.path.join", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "multiprocessing.dummy.Pool", "line_number": 17, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 18, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 18, "usage_type": "argument"}, {"api_name": "termcolor.colored", "line_number": 20, "usage_type": "call"}, {"api_name": "termcolor.colored", "line_number": 22, "usage_type": "call"}]}
+{"seq_id": "498316658", "text": "# Select *subsample* of broadly-defined LRGs with photo-z's in DR8 south\n# require NOBS>=1 for grz\n# Extinction correction is applied to grzW1W2\n\nfrom __future__ import division, print_function\nimport sys, os, glob, time, warnings\nimport numpy as np\nfrom astropy.table import Table, vstack, hstack\nimport fitsio\nimport gc\n# from multiprocessing import Pool\n\nsys.path.append(os.path.expanduser('~/git/Python/user_modules'))\nfrom match_coord import match_coord\n\ntime_start = time.time()\n\n# Select bricks with 300) & (cat_north['FLUX_IVAR_Z']>0)\n mask &= (cat_north['FLUX_Z_EC'] > 10**(0.4*(22.5-21.5))) | (cat_north['FIBERFLUX_Z_EC'] > 10**(0.4*(22.5-22.)))\n mask &= (cat_north['RA']>100) & (cat_north['RA']<300)\n if np.sum(mask)==0:\n continue\n\n idx = np.where(mask)[0]\n\n cat_north = fitsio.read(os.path.join(sweep_dir, sweep_fn), columns=columns, rows=idx)\n cat_north = Table(cat_north)\n cat_north['TYPE'] = cat_north['TYPE'].astype(str)\n\n pz_path = os.path.join(pz_dir, sweep_fn[:-5]+'-pz.fits')\n pz = Table(fitsio.read(pz_path, rows=idx))\n pz['survey'] = pz['survey'].astype(str)\n\n cat_north = hstack([cat_north, pz])\n\n # Apply extinction correction\n cat_north['FLUX_G_EC'] = cat_north['FLUX_G']/cat_north['MW_TRANSMISSION_G']\n cat_north['FLUX_R_EC'] = cat_north['FLUX_R']/cat_north['MW_TRANSMISSION_R']\n cat_north['FLUX_Z_EC'] = cat_north['FLUX_Z']/cat_north['MW_TRANSMISSION_Z']\n cat_north['FLUX_W1_EC'] = cat_north['FLUX_W1']/cat_north['MW_TRANSMISSION_W1']\n cat_north['FLUX_W2_EC'] = cat_north['FLUX_W2']/cat_north['MW_TRANSMISSION_W2']\n cat_north['FIBERFLUX_G_EC'] = cat_north['FIBERFLUX_G']/cat_north['MW_TRANSMISSION_G']\n cat_north['FIBERFLUX_R_EC'] = cat_north['FIBERFLUX_R']/cat_north['MW_TRANSMISSION_R']\n cat_north['FIBERFLUX_Z_EC'] = cat_north['FIBERFLUX_Z']/cat_north['MW_TRANSMISSION_Z']\n cat_north['FIBERTOTFLUX_G_EC'] = cat_north['FIBERTOTFLUX_G']/cat_north['MW_TRANSMISSION_G']\n cat_north['FIBERTOTFLUX_R_EC'] = cat_north['FIBERTOTFLUX_R']/cat_north['MW_TRANSMISSION_R']\n cat_north['FIBERTOTFLUX_Z_EC'] = cat_north['FIBERTOTFLUX_Z']/cat_north['MW_TRANSMISSION_Z']\n\n with warnings.catch_warnings():\n warnings.simplefilter(\"ignore\")\n\n # Quality cuts\n mask = (cat_north['NOBS_G']>=1) & (cat_north['NOBS_R']>=1) & (cat_north['NOBS_Z']>=1)\n mask &= (cat_north['TYPE']!='DUP') & (cat_north['TYPE']!='DUP ')\n \n # # Quality in r: SNR_R > 0 && RFLUX > 0\n # mask &= (cat_north['FLUX_R_EC']>0) & (cat_north['FLUX_IVAR_R']>0)\n\n # Quality in z: SNR_Z > 0 && ZFLUX > 0\n mask &= (cat_north['FLUX_Z_EC']>0) & (cat_north['FLUX_IVAR_Z']>0)\n\n # Quality in W1: FLUX_IVAR_W1 > 0 && W1FLUX > 0\n mask &= (cat_north['FLUX_W1_EC']>0) & (cat_north['FLUX_IVAR_W1']>0)\n\n # # None-stellar color: (z-w1) > 0.8*(r-z) - 1.0 => -0.8*r + 1.8*z - W1 > -1.0\n # mask_stellar = (cat_north['FLUX_R_EC']**(-0.8) * cat_north['FLUX_Z_EC']**1.8 / cat_north['FLUX_W1_EC'] < 10**(-0.4*(-1.0)))\n # # Include non-point sources\n # mask_stellar |= ((cat_north['TYPE']!='PSF') & (cat_north['TYPE']!='PSF '))\n\n # mask &= mask_stellar\n\n if np.sum(mask)>0:\n cat_north = cat_north[mask]\n else:\n continue\n\n # Remove unwanted columns\n cat_north = cat_north[columns_to_keep]\n\n ##################################### South #####################################\n\n field = 'south'\n sweep_dir = '/global/project/projectdirs/cosmo/data/legacysurvey/dr8/'+field+'/sweep/8.0'\n pz_dir = '/global/cscratch1/sd/rongpu/dr8_lrg_photoz/sweep_'+field\n cat_south = fitsio.read(os.path.join(sweep_dir, sweep_fn),\n columns=['RA', 'FLUX_Z', 'FIBERFLUX_Z', 'FLUX_IVAR_Z', 'MW_TRANSMISSION_Z'])\n cat_south = Table(cat_south)\n # Apply the (magnitude cut zmag<21.5) OR (fibermag cut zfibermag<22)\n cat_south['FLUX_Z_EC'] = cat_south['FLUX_Z']/cat_south['MW_TRANSMISSION_Z']\n cat_south['FIBERFLUX_Z_EC'] = cat_south['FIBERFLUX_Z']/cat_south['MW_TRANSMISSION_Z']\n mask = (cat_south['FLUX_Z_EC']>0) & (cat_south['FLUX_IVAR_Z']>0)\n mask &= (cat_south['FLUX_Z_EC'] > 10**(0.4*(22.5-21.5))) | (cat_south['FIBERFLUX_Z_EC'] > 10**(0.4*(22.5-22.)))\n mask &= (cat_south['RA']>100) & (cat_south['RA']<300)\n if np.sum(mask)==0:\n continue\n\n idx = np.where(mask)[0]\n\n cat_south = fitsio.read(os.path.join(sweep_dir, sweep_fn), columns=columns, rows=idx)\n cat_south = Table(cat_south)\n cat_south['TYPE'] = cat_south['TYPE'].astype(str)\n\n pz_path = os.path.join(pz_dir, sweep_fn[:-5]+'-pz.fits')\n pz = Table(fitsio.read(pz_path, rows=idx))\n pz['survey'] = pz['survey'].astype(str)\n\n cat_south = hstack([cat_south, pz])\n\n # Apply extinction correction\n cat_south['FLUX_G_EC'] = cat_south['FLUX_G']/cat_south['MW_TRANSMISSION_G']\n cat_south['FLUX_R_EC'] = cat_south['FLUX_R']/cat_south['MW_TRANSMISSION_R']\n cat_south['FLUX_Z_EC'] = cat_south['FLUX_Z']/cat_south['MW_TRANSMISSION_Z']\n cat_south['FLUX_W1_EC'] = cat_south['FLUX_W1']/cat_south['MW_TRANSMISSION_W1']\n cat_south['FLUX_W2_EC'] = cat_south['FLUX_W2']/cat_south['MW_TRANSMISSION_W2']\n cat_south['FIBERFLUX_G_EC'] = cat_south['FIBERFLUX_G']/cat_south['MW_TRANSMISSION_G']\n cat_south['FIBERFLUX_R_EC'] = cat_south['FIBERFLUX_R']/cat_south['MW_TRANSMISSION_R']\n cat_south['FIBERFLUX_Z_EC'] = cat_south['FIBERFLUX_Z']/cat_south['MW_TRANSMISSION_Z']\n cat_south['FIBERTOTFLUX_G_EC'] = cat_south['FIBERTOTFLUX_G']/cat_south['MW_TRANSMISSION_G']\n cat_south['FIBERTOTFLUX_R_EC'] = cat_south['FIBERTOTFLUX_R']/cat_south['MW_TRANSMISSION_R']\n cat_south['FIBERTOTFLUX_Z_EC'] = cat_south['FIBERTOTFLUX_Z']/cat_south['MW_TRANSMISSION_Z']\n\n with warnings.catch_warnings():\n warnings.simplefilter(\"ignore\")\n\n # Quality cuts\n mask = (cat_south['NOBS_G']>=1) & (cat_south['NOBS_R']>=1) & (cat_south['NOBS_Z']>=1)\n mask &= (cat_south['TYPE']!='DUP') & (cat_south['TYPE']!='DUP ')\n \n # # Quality in r: SNR_R > 0 && RFLUX > 0\n # mask &= (cat_south['FLUX_R_EC']>0) & (cat_south['FLUX_IVAR_R']>0)\n\n # Quality in z: SNR_Z > 0 && ZFLUX > 0\n mask &= (cat_south['FLUX_Z_EC']>0) & (cat_south['FLUX_IVAR_Z']>0)\n\n # Quality in W1: FLUX_IVAR_W1 > 0 && W1FLUX > 0\n mask &= (cat_south['FLUX_W1_EC']>0) & (cat_south['FLUX_IVAR_W1']>0)\n\n # # None-stellar color: (z-w1) > 0.8*(r-z) - 1.0 => -0.8*r + 1.8*z - W1 > -1.0\n # mask_stellar = (cat_south['FLUX_R_EC']**(-0.8) * cat_south['FLUX_Z_EC']**1.8 / cat_south['FLUX_W1_EC'] < 10**(-0.4*(-1.0)))\n # # Include non-point sources\n # mask_stellar |= ((cat_south['TYPE']!='PSF') & (cat_south['TYPE']!='PSF '))\n\n # mask &= mask_stellar\n\n if np.sum(mask)>0:\n cat_south = cat_south[mask]\n else:\n continue\n\n # Remove unwanted columns\n cat_south = cat_south[columns_to_keep]\n\n ##################################### Cross-matching #####################################\n \n idx1, idx2, _, _, _ = match_coord(cat_north['RA'], cat_north['DEC'], cat_south['RA'], cat_south['DEC'], search_radius=0.2)\n if len(idx1)==0:\n continue\n cat_north = cat_north[idx1]\n cat_south = cat_south[idx2]\n\n cat_north_stack.append(cat_north)\n cat_south_stack.append(cat_south)\n\n # clear cache\n gc.collect()\n\ncat_north_stack = vstack(cat_north_stack)\ncat_south_stack = vstack(cat_south_stack)\n\n# \"Fix\" bug in vstack that creates excessively long strings\ncat_north_stack['TYPE'] = cat_north_stack['TYPE'].astype('a4')\ncat_south_stack['TYPE'] = cat_south_stack['TYPE'].astype('a4')\n\nprint('Final combined catalog:', len(cat_north_stack))\n\noutput_path_north = '/global/cscratch1/sd/rongpu/dr8_lrg_photoz/lrg_extended_overlap_20190724_north.fits'\noutput_path_south = '/global/cscratch1/sd/rongpu/dr8_lrg_photoz/lrg_extended_overlap_20190724_south.fits'\ncat_north_stack.write(output_path_north)\ncat_south_stack.write(output_path_south)\n\nprint(time.strftime(\"%H:%M:%S\", time.gmtime(time.time() - time_start)))\n\n", "sub_path": "sv/dr8/extended_lrg_sample_north_south_overlap.py", "file_name": "extended_lrg_sample_north_south_overlap.py", "file_ext": "py", "file_size_in_byte": 12950, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sys.path.append", "line_number": 13, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.expanduser", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.intersect1d", "line_number": 21, "usage_type": "call"}, {"api_name": "fitsio.read", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "astropy.table.Table", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 79, "usage_type": "call"}, {"api_name": "fitsio.read", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "astropy.table.Table", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "astropy.table.Table", "line_number": 86, "usage_type": "call"}, {"api_name": "fitsio.read", "line_number": 86, "usage_type": "call"}, {"api_name": "astropy.table.hstack", "line_number": 89, "usage_type": "call"}, {"api_name": "warnings.catch_warnings", "line_number": 104, "usage_type": "call"}, {"api_name": "warnings.simplefilter", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 127, "usage_type": "call"}, {"api_name": "fitsio.read", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path", "line_number": 140, "usage_type": "attribute"}, {"api_name": "astropy.table.Table", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 152, "usage_type": "call"}, {"api_name": "fitsio.read", "line_number": 154, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 154, "usage_type": "call"}, {"api_name": "os.path", "line_number": 154, "usage_type": "attribute"}, {"api_name": "astropy.table.Table", "line_number": 155, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 158, "usage_type": "call"}, {"api_name": "os.path", "line_number": 158, "usage_type": "attribute"}, {"api_name": "astropy.table.Table", "line_number": 159, "usage_type": "call"}, {"api_name": "fitsio.read", "line_number": 159, "usage_type": "call"}, {"api_name": "astropy.table.hstack", "line_number": 162, "usage_type": "call"}, {"api_name": "warnings.catch_warnings", "line_number": 177, "usage_type": "call"}, {"api_name": "warnings.simplefilter", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 200, "usage_type": "call"}, {"api_name": "match_coord.match_coord", "line_number": 210, "usage_type": "call"}, {"api_name": "gc.collect", "line_number": 220, "usage_type": "call"}, {"api_name": "astropy.table.vstack", "line_number": 222, "usage_type": "call"}, {"api_name": "astropy.table.vstack", "line_number": 223, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 236, "usage_type": "call"}, {"api_name": "time.gmtime", "line_number": 236, "usage_type": "call"}, {"api_name": "time.time", "line_number": 236, "usage_type": "call"}]}
+{"seq_id": "4287580", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n# Copyright (c) 2014 Jean-Louis Fuchs\n# Copyright (c) 2013 Ulrich Mierendorff\n#\n# Permission is hereby granted, free of charge, to any person obtaining a\n# copy of this software and associated documentation files (the \"Software\"),\n# to deal in the Software without restriction, including without limitation\n# the rights to use, copy, modify, merge, publish, distribute, sublicense,\n# and/or sell copies of the Software, and to permit persons to whom the\n# Software is furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL\n# THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING\n# FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER\n# DEALINGS IN THE SOFTWARE.\n\n\n\"\"\"Binary protocol of Kyoto Tycoon with asyncio for io batching.\n\nKyoto Tycoon is a lightweight database server with impressive performance. It\ncan be accessed via several protocols, including an efficient binary protocol\nwhich is used in this Python library.\n\nThe current implementation of this library provides access to the following\ncommands: set_bulk, get_bulk, remove_bulk (plus some wrapper functions to\neasily use these commands if you only need to access a single item) and\nplay_script.\n\nThe library is implemented in pure Python and requires the module asyncio\nand other Python standard library modules. Therefore, it is possible to use\nthe library with other interpreters than the standard CPython. The code has\nbeen tested with python 3.4 since it is based on the asyncio module\nintroduced in 3.4. If pypy will implement asyncio in can be ported to\npypy and possibly other implementation.\n\n\"\"\"\n\n# TODO PEP8\n# TODO Move documentation from homepage to code\n# TODO Add some logging (tornados nice output?):w\n# TODO embed factory method that creates a ktserver and client\n# -> find free random port automatically (within range)\n# -> with keep alive and logging of failures\n# TODO compare original / asyincio wo batch / asyncio with batch\n# TODO Write tests\n# TODO remove that lazy stuff, since we are always lazy (batching)\n# TODO sphinx doc setup (take snippets from freeze)\n# TODO travis setup\n# TODO github badge setup\n# TODO stackoverflow question for promotion\n# TODO adsy blogging\n\n\nimport socket\nimport random\nimport struct\nimport logging\nimport threading\nimport subprocess\nimport sys\nimport time\nimport atexit\ntry:\n import asyncio\nexcept ImportError:\n import trollius as asyncio\n from trollius import From, Return\n\nMB_SET_BULK = 0xb8\nMB_GET_BULK = 0xba\nMB_REMOVE_BULK = 0xb9\nMB_ERROR = 0xbf\nMB_PLAY_SCRIPT = 0xb4\n\nDEFAULT_HOST = 'localhost'\nDEFAULT_PORT = 1978\nDEFAULT_EXPIRE = 0x7FFFFFFFFFFFFFFF\nMAX_CONNECTIONS = 4\n\nFLAG_NOREPLY = 0x01\n\nRANGE_FROM = 2 ** 15 - 2 ** 14\nRANGE_TO = 2 ** 15 - 1\n\n\ndef _l():\n \"\"\"Get the logger\"\"\"\n return logging.getLogger(\"ktasync\")\n\n\nclass KyotoTycoonError(Exception):\n \"\"\"Class for Exceptions in this module\"\"\"\n\n\nclass KyotoTycoon(object):\n \"\"\"New connections are created using the constructor. A connection is\n automatically closed when the object is destroyed. There is the factory\n method embedded which creates a server and client connected to it.\n\n Keys and values of database entries are python bytes. You can pickle\n objects to bytes strings. The encoding is handled by the user when\n converting to bytes. Usually bytes(bla, encoding=\"UTF-8\") is safe.\n\n \"\"\"\n\n _client = None\n\n @staticmethod\n def embedded(\n args=None,\n timeout=None,\n max_connections=MAX_CONNECTIONS,\n range_from=RANGE_FROM,\n range_to=RANGE_TO\n ):\n \"\"\"Start an embedded Kyoto Tycoon server and return a client conencted\n to it.\n\n :param args: Additional arguments for the Kyoto Tycoon server.\n\n :param timeout: Optional timeout for the socket. None means no timeout\n (please also look at the Python socket manual).\n\n :param max_connections: Maximum connections for io batching.\n\n :param range_from: Port range to select a random port from (from).\n\n :param range_to: Port range to select a random port from (to).\n\n :rtype: KyotoTycoon\n\n \"\"\"\n if KyotoTycoon._client:\n return KyotoTycoon._client\n if not args:\n args = []\n tries = 0\n sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n while tries < 20:\n tries += 1\n port = random.randint(range_from, range_to)\n try:\n sock.bind((\"127.0.0.1\", port))\n sock.listen(1)\n tries = 21\n except OSError:\n pass\n finally:\n sock.close()\n time.sleep(0.2)\n\n def keep_alive():\n \"\"\"Helper\"\"\"\n while True:\n proc = subprocess.Popen(\n [\n \"ktserver\",\n \"-le\",\n \"-host\",\n \"127.0.0.1\",\n \"-port\",\n str(port)\n ] + args,\n stderr=sys.__stderr__.fileno(),\n stdout=sys.__stdout__.fileno(),\n )\n cleanup_done = [False]\n\n def cleanup():\n \"\"\"Helper\"\"\"\n try:\n cleanup_done[0] = True\n proc.terminate()\n proc.wait()\n except ProcessLookupError:\n pass\n\n atexit.register(cleanup)\n proc.wait()\n time.sleep(10)\n if not cleanup_done[0]:\n _l().critical(\"ktserver died!\")\n\n thr = threading.Thread(target=keep_alive)\n thr.setDaemon(True)\n thr.start()\n tries = 0\n while tries < 20:\n tries += 1\n try:\n KyotoTycoon._client = KyotoTycoon(\n host=\"127.0.0.1\",\n port=port,\n probe=True,\n timeout=timeout,\n max_connections=max_connections\n )\n return KyotoTycoon._client\n except ConnectionRefusedError:\n # pypy #except Exception:\n time.sleep(0.2)\n raise KyotoTycoonError(\"Embedded server not started!\")\n\n def __init__(\n self,\n host=DEFAULT_HOST,\n port=DEFAULT_PORT,\n probe=False,\n timeout=None,\n max_connections=MAX_CONNECTIONS,\n ):\n \"\"\"\n :param host: The hostname or IP to connect to, defaults to\n 'localhost'.\n\n :param port: The port number, defaults to 1978 which is the default\n port of Kyoto Tycoon.\n\n :param probe: If set to True, the server connection is checked. The\n connections are taken from a pool. This option helps\n to prevent late failures.\n\n :param timeout: Optional timeout for the socket. None means no timeout\n (please also look at the Python socket manual).\n \"\"\"\n self.host = host\n self.port = port\n self.timeout = timeout\n self.socket = None\n self.loop = asyncio.get_event_loop()\n self.max_connections = max_connections\n self.free_streams = []\n self.semaphore = asyncio.Semaphore(max_connections)\n if probe:\n self._probe()\n\n @asyncio.coroutine\n def set(self, key, val, db=0, expire=DEFAULT_EXPIRE, flags=0):\n \"\"\"Wrapper function around set_bulk for easily storing a single item\n in the database.\n\n :param key: The key of the entry,\n\n :type key: bytes\n\n :param val: The value of the entry\n\n :type val: bytes\n\n :param db: Database index to store the record in. Default to 0.\n\n :type db: int\n\n :param expire: Expiration time for all entries.\n kyototycoon.DEFAULT_EXPIRE is 0x7FFFFFFFFFFFFFFF which\n means that the records should never expire in the\n (near) future.\n\n :param flags: If set to kyototycoon.FLAG_NOREPLY, function will not\n wait for an answer of the server.\n\n :return: The number of actually stored records, or None if flags was\n set to kyototycoon.FLAG_NOREPLY.\n \"\"\"\n return (yield from self.set_bulk(\n # pypy #raise Return((yield From(self.set_bulk(\n ((key, val, db, expire),), flags\n ))\n # pypy #))))\n\n @asyncio.coroutine\n def set_bulk_kv(self, kv, db=0, expire=DEFAULT_EXPIRE, flags=0):\n \"\"\"Wrapper function around set_bulk for simplifying the process of\n storing multiple records with equal expiration times in the same\n database.\n\n :param kv: dict of key/value pairs.\n\n :param db: database index to store the values in. defaults to 0.\n\n :param expire: Expiration time for all entries.\n kyototycoon.DEFAULT_EXPIRE is 0x7FFFFFFFFFFFFFFF which\n means that the records should never expire in the\n (near) future.\n\n :param flags: If set to kyototycoon.FLAG_NOREPLY, function will not\n wait for an answer of the server.\n\n :return: The number of actually stored records, or None if flags was\n set to kyototycoon.FLAG_NOREPLY.\n \"\"\"\n recs = ((key, val, db, expire) for key, val in kv.items())\n return (yield from self.set_bulk(recs, flags))\n # pypy #raise Return((yield From(self.set_bulk(recs, flags))))\n\n @asyncio.coroutine\n def set_bulk(self, recs, flags=0):\n \"\"\"Stores multiple records at once.\n\n :param recs: iterable (e.g. list) of records. Each record is a\n list or tuple of 4 entries: key, val, db, expire\n\n :param flags: If set to kyototycoon.FLAG_NOREPLY, function will not\n wait for an answer of the server.\n\n :return: The number of actually stored records, or None if flags was\n set to kyototycoon.FLAG_NOREPLY.\n \"\"\"\n sr, sw = yield from self._pop_streams()\n # pypy #sr, sw = yield From(self._pop_streams())\n try:\n request = [struct.pack('!BI', MB_SET_BULK, flags), None]\n\n cnt = 0\n for key, val, db, xt in recs:\n assert isinstance(key, bytes), \"Please pass bytes as key\"\n assert isinstance(val, bytes), \"Please pass bytes as value\"\n request.append(\n struct.pack('!HIIq', db, len(key), len(val), xt)\n )\n request.append(key)\n request.append(val)\n cnt += 1\n\n request[1] = struct.pack('!I', cnt)\n\n sw.write(b''.join(request))\n\n if flags & FLAG_NOREPLY:\n self._push_streams(sr, sw)\n return None\n # pypy #raise Return(None)\n\n magic, = struct.unpack('!B', (\n yield from sr.readexactly(1))\n # pypy # yield From(sr.readexactly(1)))\n )\n if magic == MB_SET_BULK:\n recs_cnt, = struct.unpack('!I', (\n yield from sr.readexactly(4))\n # pypy # yield From(sr.readexactly(4)))\n )\n self._push_streams(sr, sw)\n return recs_cnt\n # pypy #raise Return(recs_cnt)\n elif magic == MB_ERROR:\n raise KyotoTycoonError(\n 'Internal server error 0x%02x' % MB_ERROR\n )\n else:\n raise KyotoTycoonError('Unknown server error')\n finally:\n self._release_connection()\n\n @asyncio.coroutine\n def get(self, key, db=0, flags=0):\n \"\"\"Wrapper function around get_bulk for easily retrieving a single\n item from the database.\n\n\n :param key: The key of the entry\n :type key: bytes\n\n :param db: The database index. Defaults to 0.\n\n :param flags: reserved and not used now. (defined by protocol)\n\n :return: The value of the record, or None if the record could not be\n found in the database.\n\n \"\"\"\n recs = yield from self.get_bulk(((key, db),), flags)\n # pypy #recs = yield From(self.get_bulk(((key, db),), flags))\n if not recs:\n return None\n # pypy #raise Return(None)\n return recs[0][1]\n # pypy #raise Return(recs[0][1])\n\n @asyncio.coroutine\n def get_bulk_keys(self, keys, db=0, flags=0):\n \"\"\"Wrapper function around get_bulk for simplifying the process of\n retrieving multiple records from the same database.\n\n :param keys: iterable (e.g. list) of keys.\n\n :param db: database index to store the values in. defaults to 0.\n\n :param flags: reserved and not used now. (defined by protocol)\n\n :return: dict of key/value pairs.\n \"\"\"\n recs = ((key, db) for key in keys)\n recs = yield from self.get_bulk(recs, flags)\n return dict(((key, val) for key, val, db, xt in recs))\n # pypy #recs = yield From(self.get_bulk(recs, flags))\n # pypy #raise Return(dict(((key, val) for key, val, db, xt in recs)))\n\n @asyncio.coroutine\n def get_bulk(self, recs, flags=0):\n \"\"\"Retrieves multiple records at once.\n\n :param recs: iterable (e.g. list) of record descriptions. Each\n record is a list or tuple of 2 entries: key,db\n\n :param flags: reserved and not used now. (defined by protocol)\n\n :return: A list of records. Each record is a tuple of 4 entries: (key,\n val, db, expire)\n \"\"\"\n sr, sw = yield from self._pop_streams()\n # pypy #sr, sw = yield From(self._pop_streams())\n try:\n request = [struct.pack('!BI', MB_GET_BULK, flags), None]\n\n cnt = 0\n for key, db in recs:\n assert isinstance(key, bytes), \"Please pass bytes as key\"\n request.append(struct.pack('!HI', db, len(key)))\n request.append(key)\n cnt += 1\n\n request[1] = struct.pack('!I', cnt)\n\n sw.write(b''.join(request))\n res = yield from self._read_keys(sr, MB_GET_BULK)\n # pypy #res = yield From(self._read_keys(sr, MB_GET_BULK))\n self._push_streams(sr, sw)\n return res\n # pypy #raise Return(res)\n finally:\n self._release_connection()\n\n @asyncio.coroutine\n def _read_keys(self, sr, magic_expect):\n \"\"\"Internal function for reading key from get_bulk or play_script\"\"\"\n data = yield from sr.readexactly(5)\n # pypy #data = yield From(sr.readexactly(5))\n magic, = struct.unpack('!B', data[:1])\n if magic == magic_expect:\n recs_cnt, = struct.unpack('!I', data[1:])\n recs_cnt -= 1\n recs = []\n # Reduce yields be reading key and next header at once\n if recs_cnt >= 0:\n data = yield from sr.readexactly(18)\n # pypy #data = yield From(sr.readexactly(18))\n pre_data = 0\n for _ in range(recs_cnt):\n db, key_len, val_len, xt = struct.unpack(\n '!HIIq', data[pre_data:]\n )\n pre_data = key_len + val_len\n data = yield from sr.readexactly(pre_data + 18)\n # pypy #data = yield From(sr.readexactly(pre_data + 18))\n recs.append((data[:key_len], data[key_len:], db, xt))\n db, key_len, val_len, xt = struct.unpack(\n '!HIIq', data[pre_data:]\n )\n pre_data = key_len + val_len\n data = yield from sr.readexactly(pre_data)\n # pypy #data = yield From(sr.readexactly(pre_data))\n recs.append((data[:key_len], data[key_len:], db, xt))\n return recs\n # pypy #raise Return(recs)\n elif magic == MB_ERROR:\n raise KyotoTycoonError(\n 'Internal server error 0x%02x' % MB_ERROR\n )\n else:\n raise KyotoTycoonError('Unknown server error')\n\n @asyncio.coroutine\n def remove(self, key, db, flags=0):\n \"\"\"Wrapper function around remove_bulk for easily removing a single\n item from the database.\n\n :param key: The key of the entry.\n :type key: bytes\n\n :param db: database index to store the values in. defaults to 0.\n\n :param flags: If set to kyototycoon.FLAG_NOREPLY, function will not\n wait for an answer of the server.\n\n :return: The number of removed records, or None if flags was set to\n kyototycoon.FLAG_NOREPLY\n \"\"\"\n return (yield from self.remove_bulk(((key, db),), flags))\n # pypy #raise Return((yield From(self.remove_bulk(((key, db),), flags))))\n\n @asyncio.coroutine\n def remove_bulk_keys(self, keys, db, flags=0):\n \"\"\"Wrapper function around remove_bulk for simplifying the process of\n removing multiple records from the same database.\n\n :param keys: iterable (e.g. list) of keys.\n\n :param db: database index to store the values in. defaults to 0.\n\n :param flags: If set to kyototycoon.FLAG_NOREPLY, function will not\n wait for an answer of the server.\n\n :return: The number of removed records, or None if flags was set to\n kyototycoon.FLAG_NOREPLY\n \"\"\"\n recs = ((key, db) for key in keys)\n return (yield from self.remove_bulk(recs, flags))\n # pypy #raise Return((yield From(self.remove_bulk(recs, flags))))\n\n @asyncio.coroutine\n def remove_bulk(self, recs, flags=0):\n \"\"\"Remove multiple records at once.\n\n :param recs: iterable (e.g. list) of record descriptions. Each\n record is a list or tuple of 2 entries: key,db\n\n :param flags: If set to kyototycoon.FLAG_NOREPLY, function will not\n wait for an answer of the server.\n\n :return: The number of removed records, or None if flags was set to\n kyototycoon.FLAG_NOREPLY\n \"\"\"\n sr, sw = yield from self._pop_streams()\n # pypy #sr, sw = yield From(self._pop_streams())\n try:\n\n request = [struct.pack('!BI', MB_REMOVE_BULK, flags), None]\n\n cnt = 0\n for key, db in recs:\n request.append(struct.pack('!HI', db, len(key)))\n request.append(key)\n cnt += 1\n\n request[1] = struct.pack('!I', cnt)\n\n sw.write(''.join(request))\n\n if flags & FLAG_NOREPLY:\n self._push_streams(sr, sw)\n return None\n # pypy #raise Return(None)\n\n magic, = struct.unpack(\n '!B', (yield from sr.readexactly(1))\n # pypy # '!B', (yield From(sr.readexactly(1)))\n )\n if magic == MB_REMOVE_BULK:\n recs_cnt, = struct.unpack(\n '!I', (yield from sr.readexactly(4))\n # pypy # '!I', (yield From(sr.readexactly(4)))\n )\n self._push_streams(sr, sw)\n return recs_cnt\n # pypy #raise Return(recs_cnt)\n elif magic == MB_ERROR:\n raise KyotoTycoonError(\n 'Internal server error 0x%02x' % MB_ERROR\n )\n else:\n raise KyotoTycoonError('Unknown server error')\n finally:\n self._release_connection()\n\n @asyncio.coroutine\n def play_script(self, name, recs, flags=0):\n \"\"\"Calls a procedure of the LUA scripting language extension.\n\n :param name: The name of the LUA function.\n\n :param recs: iterable (e.g. list) of records. Each record is a list or\n tuple of 2 entries: key, val\n\n :param flags: If set to kyototycoon.FLAG_NOREPLY, function will not\n wait for an answer of the server.\n\n :return: A list of records. Each record is a tuple of 2 entries: (key,\n val). Or None if flags was set to kyototycoon.FLAG_NOREPLY.\n \"\"\"\n sr, sw = yield from self._pop_streams()\n # pypy #sr, sw = yield From(self._pop_streams())\n try:\n\n request = [\n struct.pack(\n '!BII', MB_PLAY_SCRIPT, flags, len(name)\n ), None, name\n ]\n\n cnt = 0\n for key, val in recs:\n request.append(struct.pack('!II', len(key), len(val)))\n request.append(key)\n request.append(val)\n cnt += 1\n\n request[1] = struct.pack('!I', cnt)\n\n yield from sw.write(''.join(request))\n # pypy #yield From(sw.write(''.join(request)))\n\n if flags & FLAG_NOREPLY:\n self._push_streams(sr, sw)\n return None\n # pypy #raise Return(None)\n\n magic, = struct.unpack(\n '!B', (yield from sr.readexactly(1))\n # pypy # '!B', (yield From(sr.readexactly(1)))\n )\n res = yield from self._read_keys(sr, MB_PLAY_SCRIPT)\n # pypy #res = yield From(self._read_keys(sr, MB_PLAY_SCRIPT))\n self._push_streams(sr, sw)\n return res\n # pypy #raise Return(res)\n finally:\n self._release_connection()\n\n def close(self):\n \"\"\"Close the sockets\"\"\"\n while self.free_streams:\n _, sw = self.free_streams.pop()\n sw.close()\n\n def _probe(self):\n \"\"\"Probe the server\"\"\"\n sock = socket.create_connection(\n (self.host, self.port),\n self.timeout\n )\n sock.close()\n\n def __del__(self):\n \"\"\"Cleanup on the delete\"\"\"\n self.close()\n\n @asyncio.coroutine\n def _pop_streams(self):\n \"\"\"Get a new stream. It will block (async) when max_connections is\n reached\"\"\"\n yield from self.semaphore.acquire()\n # pypy #yield From(self.semaphore.acquire())\n if self.free_streams:\n return self.free_streams.pop()\n # pypy #raise Return(self.free_streams.pop())\n else:\n return (yield from asyncio.open_connection(\n # pypy #raise Return((yield From(asyncio.open_connection(\n self.host,\n self.port,\n ))\n # pypy #))))\n\n def _release_connection(self):\n \"\"\"Release the semaphore\n\n If a connection dies, we won't return it. Therefore release is an\n extra method.\"\"\"\n self.semaphore.release()\n\n def _push_streams(self, sr, sw):\n \"\"\"Return used stream.\"\"\"\n self.free_streams.append((sr, sw))\n", "sub_path": "ktasync.py", "file_name": "ktasync.py", "file_ext": "py", "file_size_in_byte": 23715, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.getLogger", "line_number": 96, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 146, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 146, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 146, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 149, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 158, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 163, "usage_type": "call"}, {"api_name": "sys.__stderr__.fileno", "line_number": 172, "usage_type": "call"}, {"api_name": "sys.__stderr__", "line_number": 172, "usage_type": "attribute"}, {"api_name": "sys.__stdout__.fileno", "line_number": 173, "usage_type": "call"}, {"api_name": "sys.__stdout__", "line_number": 173, "usage_type": "attribute"}, {"api_name": "atexit.register", "line_number": 186, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 188, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 192, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 209, "usage_type": "call"}, {"api_name": "trollius.get_event_loop", "line_number": 238, "usage_type": "call"}, {"api_name": "trollius.Semaphore", "line_number": 241, "usage_type": "call"}, {"api_name": "trollius.coroutine", "line_number": 245, "usage_type": "attribute"}, {"api_name": "trollius.coroutine", "line_number": 279, "usage_type": "attribute"}, {"api_name": "struct.pack", "line_number": 320, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 327, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 333, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 342, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 347, "usage_type": "call"}, {"api_name": "trollius.coroutine", "line_number": 304, "usage_type": "attribute"}, {"api_name": "trollius.coroutine", "line_number": 363, "usage_type": "attribute"}, {"api_name": "trollius.coroutine", "line_number": 388, "usage_type": "attribute"}, {"api_name": "struct.pack", "line_number": 422, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 427, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 431, "usage_type": "call"}, {"api_name": "trollius.coroutine", "line_number": 407, "usage_type": "attribute"}, {"api_name": "struct.unpack", "line_number": 447, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 449, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 458, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 465, "usage_type": "call"}, {"api_name": "trollius.coroutine", "line_number": 442, "usage_type": "attribute"}, {"api_name": "trollius.coroutine", "line_number": 481, "usage_type": "attribute"}, {"api_name": "trollius.coroutine", "line_number": 500, "usage_type": "attribute"}, {"api_name": "struct.pack", "line_number": 536, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 540, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 544, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 553, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 558, "usage_type": "call"}, {"api_name": "trollius.coroutine", "line_number": 519, "usage_type": "attribute"}, {"api_name": "struct.pack", "line_number": 594, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 601, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 606, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 616, "usage_type": "call"}, {"api_name": "trollius.coroutine", "line_number": 574, "usage_type": "attribute"}, {"api_name": "socket.create_connection", "line_number": 636, "usage_type": "call"}, {"api_name": "trollius.open_connection", "line_number": 656, "usage_type": "call"}, {"api_name": "trollius.coroutine", "line_number": 646, "usage_type": "attribute"}]}
+{"seq_id": "563236354", "text": "import os #importing operating system information.\r\nimport bpy #importing blender info.\r\nimport sys\r\n\r\nfor env in os.environ:\r\n print(env)\r\n \r\nfor arg in sys.argv:\r\n print(arg)\r\n \r\nbpy.ops.object.select_all(action='DESELECT')\r\nfor obj in bpy.data.objects:\r\n if obj.type == 'MESH':\r\n obj.select_set(True)\r\nbpy.ops.object.delete()\r\n\r\npath = curuthers/'curuthers.obj'\r\nbpy.ops.import_scene.obj(filepath=path)\r\nobj = bpy.context.selected_objects[0]\r\n", "sub_path": "Lab_11.py", "file_name": "Lab_11.py", "file_ext": "py", "file_size_in_byte": 452, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.environ", "line_number": 5, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 8, "usage_type": "attribute"}, {"api_name": "bpy.ops.object.select_all", "line_number": 11, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 11, "usage_type": "attribute"}, {"api_name": "bpy.data", "line_number": 12, "usage_type": "attribute"}, {"api_name": "bpy.ops.object.delete", "line_number": 15, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 15, "usage_type": "attribute"}, {"api_name": "bpy.ops.import_scene.obj", "line_number": 18, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 18, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 19, "usage_type": "attribute"}]}
+{"seq_id": "494633483", "text": "#!/usr/bin/env python2\nfrom __future__ import print_function\nimport os\nimport sys\nimport time\nimport base64\nfrom urllib2 import urlopen\nfrom urllib2 import Request\nfrom urllib2 import HTTPError\nfrom urllib import urlencode\nfrom urllib import quote\nfrom exceptions import Exception\nfrom email.mime.multipart import MIMEMultipart\n\nfrom email.mime.base import MIMEBase\nfrom email.mime.application import MIMEApplication\n\nfrom email.encoders import encode_noop\n\nimport json\ndef json2python(data):\n try:\n return json.loads(data)\n except:\n pass\n return None\npython2json = json.dumps\n\nclass MalformedResponse(Exception):\n pass\nclass RequestError(Exception):\n pass\n\nclass Client(object):\n default_url = 'http://nova.astrometry.net/api/'\n\n def __init__(self,\n apiurl = default_url):\n self.session = None\n self.apiurl = apiurl\n\n def get_url(self, service):\n return self.apiurl + service\n\n def send_request(self, service, args={}, file_args=None):\n '''\n service: string\n args: dict\n '''\n if self.session is not None:\n args.update({ 'session' : self.session })\n print('Python:', args)\n json = python2json(args)\n print('Sending json:', json)\n url = self.get_url(service)\n print('Sending to URL:', url)\n\n # If we're sending a file, format a multipart/form-data\n if file_args is not None:\n m1 = MIMEBase('text', 'plain')\n m1.add_header('Content-disposition', 'form-data; name=\"request-json\"')\n m1.set_payload(json)\n\n m2 = MIMEApplication(file_args[1],'octet-stream',encode_noop)\n m2.add_header('Content-disposition',\n 'form-data; name=\"file\"; filename=\"%s\"' % file_args[0])\n\n mp = MIMEMultipart('form-data', None, [m1, m2])\n\n # Make a custom generator to format it the way we need.\n from cStringIO import StringIO\n from email.generator import Generator\n\n class MyGenerator(Generator):\n def __init__(self, fp, root=True):\n Generator.__init__(self, fp, mangle_from_=False,\n maxheaderlen=0)\n self.root = root\n def _write_headers(self, msg):\n # We don't want to write the top-level headers;\n # they go into Request(headers) instead.\n if self.root:\n return\n # We need to use \\r\\n line-terminator, but Generator\n # doesn't provide the flexibility to override, so we\n # have to copy-n-paste-n-modify.\n for h, v in msg.items():\n print(('%s: %s\\r\\n' % (h,v)), end='', file=self._fp)\n # A blank line always separates headers from body\n print('\\r\\n', end='', file=self._fp)\n\n # The _write_multipart method calls \"clone\" for the\n # subparts. We hijack that, setting root=False\n def clone(self, fp):\n return MyGenerator(fp, root=False)\n\n fp = StringIO()\n g = MyGenerator(fp)\n g.flatten(mp)\n data = fp.getvalue()\n headers = {'Content-type': mp.get('Content-type')}\n\n else:\n # Else send x-www-form-encoded\n data = {'request-json': json}\n print('Sending form data:', data)\n data = urlencode(data)\n print('Sending data:', data)\n headers = {}\n\n request = Request(url=url, headers=headers, data=data)\n\n try:\n f = urlopen(request)\n txt = f.read()\n print('Got json:', txt)\n result = json2python(txt)\n print('Got result:', result)\n stat = result.get('status')\n print('Got status:', stat)\n if stat == 'error':\n errstr = result.get('errormessage', '(none)')\n raise RequestError('server error message: ' + errstr)\n return result\n except HTTPError as e:\n print('HTTPError', e)\n txt = e.read()\n open('err.html', 'wb').write(txt)\n print('Wrote error text to err.html')\n\n def login(self, apikey):\n args = { 'apikey' : apikey }\n result = self.send_request('login', args)\n sess = result.get('session')\n print('Got session:', sess)\n if not sess:\n raise RequestError('no session in result')\n self.session = sess\n\n def _get_upload_args(self, **kwargs):\n args = {}\n for key,default,typ in [('allow_commercial_use', 'd', str),\n ('allow_modifications', 'd', str),\n ('publicly_visible', 'y', str),\n ('scale_units', None, str),\n ('scale_type', None, str),\n ('scale_lower', None, float),\n ('scale_upper', None, float),\n ('scale_est', None, float),\n ('scale_err', None, float),\n ('center_ra', None, float),\n ('center_dec', None, float),\n ('radius', None, float),\n ('downsample_factor', None, int),\n ('tweak_order', None, int),\n ('crpix_center', None, bool),\n ('x', None, list),\n ('y', None, list),\n # image_width, image_height\n ]:\n if key in kwargs:\n val = kwargs.pop(key)\n val = typ(val)\n args.update({key: val})\n elif default is not None:\n args.update({key: default})\n print('Upload args:', args)\n return args\n\n def url_upload(self, url, **kwargs):\n args = dict(url=url)\n args.update(self._get_upload_args(**kwargs))\n result = self.send_request('url_upload', args)\n return result\n\n def upload(self, fn=None, **kwargs):\n args = self._get_upload_args(**kwargs)\n file_args = None\n if fn is not None:\n try:\n f = open(fn, 'rb')\n file_args = (fn, f.read())\n except IOError:\n print('File %s does not exist' % fn)\n raise\n return self.send_request('upload', args, file_args)\n\n def submission_images(self, subid):\n result = self.send_request('submission_images', {'subid':subid})\n return result.get('image_ids')\n\n def overlay_plot(self, service, outfn, wcsfn, wcsext=0):\n from astrometry.util import util as anutil\n wcs = anutil.Tan(wcsfn, wcsext)\n params = dict(crval1 = wcs.crval[0], crval2 = wcs.crval[1],\n crpix1 = wcs.crpix[0], crpix2 = wcs.crpix[1],\n cd11 = wcs.cd[0], cd12 = wcs.cd[1],\n cd21 = wcs.cd[2], cd22 = wcs.cd[3],\n imagew = wcs.imagew, imageh = wcs.imageh)\n result = self.send_request(service, {'wcs':params})\n print('Result status:', result['status'])\n plotdata = result['plot']\n plotdata = base64.b64decode(plotdata)\n open(outfn, 'wb').write(plotdata)\n print('Wrote', outfn)\n\n def sdss_plot(self, outfn, wcsfn, wcsext=0):\n return self.overlay_plot('sdss_image_for_wcs', outfn,\n wcsfn, wcsext)\n\n def galex_plot(self, outfn, wcsfn, wcsext=0):\n return self.overlay_plot('galex_image_for_wcs', outfn,\n wcsfn, wcsext)\n\n def myjobs(self):\n result = self.send_request('myjobs/')\n return result['jobs']\n\n def job_status(self, job_id, justdict=False):\n result = self.send_request('jobs/%s' % job_id)\n if justdict:\n return result\n stat = result.get('status')\n if stat == 'success':\n result = self.send_request('jobs/%s/calibration' % job_id)\n print('Calibration:', result)\n result = self.send_request('jobs/%s/tags' % job_id)\n print('Tags:', result)\n result = self.send_request('jobs/%s/machine_tags' % job_id)\n print('Machine Tags:', result)\n result = self.send_request('jobs/%s/objects_in_field' % job_id)\n print('Objects in field:', result)\n result = self.send_request('jobs/%s/annotations' % job_id)\n print('Annotations:', result)\n result = self.send_request('jobs/%s/info' % job_id)\n print('Calibration:', result)\n\n return stat\n\n def annotate_data(self,job_id):\n \"\"\"\n :param job_id: id of job\n :return: return data for annotations\n \"\"\"\n result = self.send_request('jobs/%s/annotations' % job_id)\n return result\n\n def sub_status(self, sub_id, justdict=False):\n result = self.send_request('submissions/%s' % sub_id)\n if justdict:\n return result\n return result.get('status')\n\n def jobs_by_tag(self, tag, exact):\n exact_option = 'exact=yes' if exact else ''\n result = self.send_request(\n 'jobs_by_tag?query=%s&%s' % (quote(tag.strip()), exact_option),\n {},\n )\n return result\n\nif __name__ == '__main__':\n print(\"Running with args %s\"%sys.argv)\n import optparse\n parser = optparse.OptionParser()\n parser.add_option('--server', dest='server', default=Client.default_url,\n help='Set server base URL (eg, %default)')\n parser.add_option('--apikey', '-k', dest='apikey',\n help='API key for Astrometry.net web service; if not given will check AN_API_KEY environment variable')\n parser.add_option('--upload', '-u', dest='upload', help='Upload a file')\n parser.add_option('--upload-xy', dest='upload_xy', help='Upload a FITS x,y table as JSON')\n parser.add_option('--wait', '-w', dest='wait', action='store_true', help='After submitting, monitor job status')\n parser.add_option('--wcs', dest='wcs', help='Download resulting wcs.fits file, saving to given filename; implies --wait if --urlupload or --upload')\n parser.add_option('--newfits', dest='newfits', help='Download resulting new-image.fits file, saving to given filename; implies --wait if --urlupload or --upload')\n parser.add_option('--kmz', dest='kmz', help='Download resulting kmz file, saving to given filename; implies --wait if --urlupload or --upload')\n parser.add_option('--annotate','-a',dest='annotate',help='store information about annotations in give file, JSON format; implies --wait if --urlupload or --upload')\n parser.add_option('--urlupload', '-U', dest='upload_url', help='Upload a file at specified url')\n parser.add_option('--scale-units', dest='scale_units',\n choices=('arcsecperpix', 'arcminwidth', 'degwidth', 'focalmm'), help='Units for scale estimate')\n #parser.add_option('--scale-type', dest='scale_type',\n # choices=('ul', 'ev'), help='Scale bounds: lower/upper or estimate/error')\n parser.add_option('--scale-lower', dest='scale_lower', type=float, help='Scale lower-bound')\n parser.add_option('--scale-upper', dest='scale_upper', type=float, help='Scale upper-bound')\n parser.add_option('--scale-est', dest='scale_est', type=float, help='Scale estimate')\n parser.add_option('--scale-err', dest='scale_err', type=float, help='Scale estimate error (in PERCENT), eg \"10\" if you estimate can be off by 10%')\n parser.add_option('--ra', dest='center_ra', type=float, help='RA center')\n parser.add_option('--dec', dest='center_dec', type=float, help='Dec center')\n parser.add_option('--radius', dest='radius', type=float, help='Search radius around RA,Dec center')\n parser.add_option('--downsample', dest='downsample_factor', type=int, help='Downsample image by this factor')\n parser.add_option('--parity', dest='parity', choices=('0','1'), help='Parity (flip) of image')\n parser.add_option('--tweak-order', dest='tweak_order', type=int, help='SIP distortion order (default: 2)')\n parser.add_option('--crpix-center', dest='crpix_center', action='store_true', default=None, help='Set reference point to center of image?')\n parser.add_option('--sdss', dest='sdss_wcs', nargs=2, help='Plot SDSS image for the given WCS file; write plot to given PNG filename')\n parser.add_option('--galex', dest='galex_wcs', nargs=2, help='Plot GALEX image for the given WCS file; write plot to given PNG filename')\n parser.add_option('--jobid', '-i', dest='solved_id', type=int,help='retrieve result for jobId instead of submitting new image')\n parser.add_option('--substatus', '-s', dest='sub_id', help='Get status of a submission')\n parser.add_option('--jobstatus', '-j', dest='job_id', help='Get status of a job')\n parser.add_option('--jobs', '-J', dest='myjobs', action='store_true', help='Get all my jobs')\n parser.add_option('--jobsbyexacttag', '-T', dest='jobs_by_exact_tag', help='Get a list of jobs associated with a given tag--exact match')\n parser.add_option('--jobsbytag', '-t', dest='jobs_by_tag', help='Get a list of jobs associated with a given tag')\n parser.add_option( '--private', '-p',\n dest='public',\n action='store_const',\n const='n',\n default='y',\n help='Hide this submission from other users')\n parser.add_option('--allow_mod_sa','-m',\n dest='allow_mod',\n action='store_const',\n const='sa',\n default='d',\n help='Select license to allow derivative works of submission, but only if shared under same conditions of original license')\n parser.add_option('--no_mod','-M',\n dest='allow_mod',\n action='store_const',\n const='n',\n default='d',\n help='Select license to disallow derivative works of submission')\n parser.add_option('--no_commercial','-c',\n dest='allow_commercial',\n action='store_const',\n const='n',\n default='d',\n help='Select license to disallow commercial use of submission')\n opt,args = parser.parse_args()\n\n if opt.apikey is None:\n # try the environment\n opt.apikey = os.environ.get('AN_API_KEY', None)\n if opt.apikey is None:\n parser.print_help()\n print()\n print('You must either specify --apikey or set AN_API_KEY')\n sys.exit(-1)\n\n args = {}\n args['apiurl'] = opt.server\n c = Client(**args)\n c.login(opt.apikey)\n\n if opt.upload or opt.upload_url or opt.upload_xy:\n if opt.wcs or opt.kmz or opt.newfits or opt.annotate:\n opt.wait = True\n\n kwargs = dict(\n allow_commercial_use=opt.allow_commercial,\n allow_modifications=opt.allow_mod,\n publicly_visible=opt.public)\n if opt.scale_lower and opt.scale_upper:\n kwargs.update(scale_lower=opt.scale_lower,\n scale_upper=opt.scale_upper,\n scale_type='ul')\n elif opt.scale_est and opt.scale_err:\n kwargs.update(scale_est=opt.scale_est,\n scale_err=opt.scale_err,\n scale_type='ev')\n elif opt.scale_lower or opt.scale_upper:\n kwargs.update(scale_type='ul')\n if opt.scale_lower:\n kwargs.update(scale_lower=opt.scale_lower)\n if opt.scale_upper:\n kwargs.update(scale_upper=opt.scale_upper)\n\n for key in ['scale_units', 'center_ra', 'center_dec', 'radius',\n 'downsample_factor', 'tweak_order', 'crpix_center',]:\n if getattr(opt, key) is not None:\n kwargs[key] = getattr(opt, key)\n if opt.parity is not None:\n kwargs.update(parity=int(opt.parity))\n\n if opt.upload:\n upres = c.upload(opt.upload, **kwargs)\n if opt.upload_xy:\n from astrometry.util.fits import fits_table\n T = fits_table(opt.upload_xy)\n kwargs.update(x=[float(x) for x in T.x], y=[float(y) for y in T.y])\n upres = c.upload(**kwargs)\n if opt.upload_url:\n upres = c.url_upload(opt.upload_url, **kwargs)\n\n stat = upres['status']\n if stat != 'success':\n print('Upload failed: status', stat)\n print(upres)\n sys.exit(-1)\n\n opt.sub_id = upres['subid']\n\n if opt.wait:\n if opt.solved_id is None:\n if opt.sub_id is None:\n print(\"Can't --wait without a submission id or job id!\")\n sys.exit(-1)\n\n while True:\n stat = c.sub_status(opt.sub_id, justdict=True)\n print('Got status:', stat)\n jobs = stat.get('jobs', [])\n if len(jobs):\n for j in jobs:\n if j is not None:\n break\n if j is not None:\n print('Selecting job id', j)\n opt.solved_id = j\n break\n time.sleep(5)\n\n while True:\n stat = c.job_status(opt.solved_id, justdict=True)\n print('Got job status:', stat)\n if stat.get('status','') in ['success']:\n success = (stat['status'] == 'success')\n break\n time.sleep(5)\n\n if opt.solved_id:\n # we have a jobId for retrieving results\n retrieveurls = []\n if opt.wcs:\n # We don't need the API for this, just construct URL\n url = opt.server.replace('/api/', '/wcs_file/%i' % opt.solved_id)\n retrieveurls.append((url, opt.wcs))\n if opt.kmz:\n url = opt.server.replace('/api/', '/kml_file/%i/' % opt.solved_id)\n retrieveurls.append((url, opt.kmz))\n if opt.newfits:\n url = opt.server.replace('/api/', '/new_fits_file/%i/' % opt.solved_id)\n retrieveurls.append((url, opt.newfits))\n\n for url,fn in retrieveurls:\n print('Retrieving file from', url, 'to', fn)\n f = urlopen(url)\n txt = f.read()\n w = open(fn, 'wb')\n w.write(txt)\n w.close()\n print('Wrote to', fn)\n\n if opt.annotate:\n result = c.annotate_data(opt.solved_id)\n with open(opt.annotate,'w') as f:\n f.write(python2json(result))\n\n if opt.wait:\n # behaviour as in old implementation\n opt.sub_id = None\n\n if opt.sdss_wcs:\n (wcsfn, outfn) = opt.sdss_wcs\n c.sdss_plot(outfn, wcsfn)\n if opt.galex_wcs:\n (wcsfn, outfn) = opt.galex_wcs\n c.galex_plot(outfn, wcsfn)\n\n if opt.sub_id:\n print(c.sub_status(opt.sub_id))\n if opt.job_id:\n print(c.job_status(opt.job_id))\n\n if opt.jobs_by_tag:\n tag = opt.jobs_by_tag\n print(c.jobs_by_tag(tag, None))\n if opt.jobs_by_exact_tag:\n tag = opt.jobs_by_exact_tag\n print(c.jobs_by_tag(tag, 'yes'))\n\n if opt.myjobs:\n jobs = c.myjobs()\n print(jobs)\n\n\n", "sub_path": "astrometry/astrometry_client.py", "file_name": "astrometry_client.py", "file_ext": "py", "file_size_in_byte": 19399, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "json.loads", "line_number": 23, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 27, "usage_type": "attribute"}, {"api_name": "exceptions.Exception", "line_number": 29, "usage_type": "name"}, {"api_name": "exceptions.Exception", "line_number": 31, "usage_type": "name"}, {"api_name": "email.mime.base.MIMEBase", "line_number": 60, "usage_type": "call"}, {"api_name": "email.mime.application.MIMEApplication", "line_number": 64, "usage_type": "call"}, {"api_name": "email.encoders.encode_noop", "line_number": 64, "usage_type": "argument"}, {"api_name": "email.mime.multipart.MIMEMultipart", "line_number": 68, "usage_type": "call"}, {"api_name": "cStringIO.StringIO", "line_number": 97, "usage_type": "call"}, {"api_name": "urllib.urlencode", "line_number": 107, "usage_type": "call"}, {"api_name": "urllib2.Request", "line_number": 111, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 114, "usage_type": "call"}, {"api_name": "urllib2.HTTPError", "line_number": 125, "usage_type": "name"}, {"api_name": "astrometry.util.util.Tan", "line_number": 194, "usage_type": "call"}, {"api_name": "astrometry.util.util", "line_number": 194, "usage_type": "name"}, {"api_name": "base64.b64decode", "line_number": 203, "usage_type": "call"}, {"api_name": "urllib.quote", "line_number": 257, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 263, "usage_type": "attribute"}, {"api_name": "optparse.OptionParser", "line_number": 265, "usage_type": "call"}, {"api_name": "{'StringIO': 'cStringIO.StringIO', 'Generator': 'email.generator.Generator', 'anutil': 'astrometry.util.util'}.default_url", "line_number": 266, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 329, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 329, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 334, "usage_type": "call"}, {"api_name": "{'StringIO': 'cStringIO.StringIO', 'Generator': 'email.generator.Generator', 'anutil': 'astrometry.util.util'}", "line_number": 338, "usage_type": "call"}, {"api_name": "astrometry.util.fits.fits_table", "line_number": 375, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 385, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 393, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 407, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 415, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 433, "usage_type": "call"}]}
+{"seq_id": "308774118", "text": "import json\nimport os\n\nimport requests\nfrom dotenv import load_dotenv\n\nload_dotenv()\n\nos.environ[\"PYTHONWARNINGS\"] = \"ignore:Unverified HTTPS request\"\nCALIBRATE = os.getenv(\"CALIBRATE\", \"false\")\n\n\nclass DeviceRegistry:\n def __init__(self, tenant, url) -> None:\n self.tenant = tenant\n self.base_url = url\n self.calibrate_url = f\"{self.base_url}calibrate\"\n\n def insert_events(self, data):\n\n data_dict = list(data)\n measurements = []\n for row in data_dict:\n row_dict = dict(row)\n\n if f\"{CALIBRATE}\".strip().lower() == \"true\":\n time = row_dict.get(\"time\")\n device = row_dict.get(\"device\")\n pm2_5 = dict(row_dict.get(\"pm2_5\")).get(\"value\")\n pm10 = dict(row_dict.get(\"pm10\")).get(\"value\")\n temp = dict(row_dict.get(\"externalTemperature\")).get(\"value\")\n hum = dict(row_dict.get(\"externalHumidity\")).get(\"value\")\n\n row_dict[\"pm2_5\"][\"calibratedValue\"] = self.get_calibrated_value(\n device=device,\n time=time,\n humidity=hum,\n pm2_5=pm2_5,\n pm10=pm10,\n temperature=temp\n )\n\n measurements.append(row_dict)\n\n try:\n\n headers = {'Content-Type': 'application/json'}\n url = self.base_url + \"devices/events?tenant=\" + self.tenant\n json_data = json.dumps(measurements)\n\n response = requests.post(url, json_data, headers=headers, verify=False)\n\n if response.status_code == 200:\n print(response.json())\n else:\n print(\"Device registry failed to insert values. Status Code : \" + str(response.status_code))\n print(response.content)\n print(response.request.url)\n print(response.request.body)\n\n except Exception as ex:\n print(\"Error Occurred while inserting measurements: \" + str(ex))\n\n def get_calibrated_value(self, device, time, pm2_5, pm10, temperature, humidity):\n print(\"getting calibrated value\")\n\n data = {\n \"datetime\": time,\n \"raw_values\": [\n {\n \"device_id\": device,\n \"pm2.5\": pm2_5,\n \"pm10\": pm10,\n \"temperature\": temperature,\n \"humidity\": humidity\n }\n ]\n }\n\n try:\n headers = {'Content-Type': 'application/json'}\n post_request = requests.post(url=self.calibrate_url, data=json.dumps(data), timeout=60000, headers=headers)\n except Exception as ex:\n print(f\"Calibrate Url returned an error: {str(ex)}\")\n return None\n\n if post_request.status_code != 200:\n print('\\n')\n print(f\"Calibrate failed to return values. Status Code : \"\n f\"{str(post_request.status_code)}, Url : {self.calibrate_url}, Body: {data}\")\n print(post_request.content)\n print('\\n')\n return None\n\n try:\n response = post_request.json()\n\n calibrated_value = None\n for result in response:\n if \"calibrated_value\" in result:\n calibrated_value = result[\"calibrated_value\"]\n break\n return calibrated_value\n\n except Exception as ex:\n print(f\"Error processing calibrate response: {str(ex)}\")\n return None\n", "sub_path": "src/data-mgt/python/events-consumer/deviceRegistry.py", "file_name": "deviceRegistry.py", "file_ext": "py", "file_size_in_byte": 3577, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "dotenv.load_dotenv", "line_number": 7, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.getenv", "line_number": 10, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 49, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 51, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 82, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 82, "usage_type": "call"}]}
+{"seq_id": "93134680", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\n@author: Shane Caldwell, Steph Rivera\n\"\"\"\n\nimport dash\nimport dash_core_components as dcc\nimport dash_html_components as html\nimport pandas as pd\nfrom dash.dependencies import Input, Output\nimport plotly.graph_objs as go\nimport numpy as np\n\n# Import your dataframe from a csv with pandas\ndf = pd.read_csv('data/kiva_loans.csv')\n\ndef split_borrower_gender(l):\n m = 0\n f = 0\n if type(l) != list:\n return np.nan\n for i in l:\n if i== 'male':\n m += 1\n else:\n f += 1\n if m == 0:\n return 'female'\n elif f == 0:\n return 'male'\n else:\n return 'both'\n\ndf.borrower_genders = df.borrower_genders.str.split(', ').apply(split_borrower_gender)\n\ntop5 = df.groupby('activity').size().sort_values(ascending=False)[0:5] # lets look at top 5\n\ntop5_male = df.groupby('activity').size().sort_values(ascending=False)[0:5]\n\ntop5_female = df[df.borrower_genders == 'female'].groupby('activity').size().sort_values(ascending=False)[0:5]\n\n# Create a Dash object instance\napp = dash.Dash()\n\n# The layout attribute of the Dash object, app\n# is where you include the elements you want to appear in the\n# dashboard. Here, dcc.Graph and dcc.Slider are separate\n# graph objects. Most of Graph's features are defined\n# inside the function update_figure, but we set the id\n# here so we can reference it inside update_figure\napp.layout = html.Div(className='container', children=[\n html.H1(\n children='Top 5 activities for loans',\n style={\n 'textAlign': 'center', # center the header\n 'color': '#7F7F7F'\n # https://www.biotechnologyforums.com/thread-7742.html more color code options\n }\n ),\n html.Div(dcc.Graph( # add a bar graph to dashboard\n id='top5-by-gender',\n figure={\n 'data': [\n {\n 'x': top5_male.index,\n 'y': top5_male,\n 'type': 'bar',\n 'opacity': .6 # changes the bar chart's opacity\n }\n ]\n }\n )),\n \n html.Label('Gender'),\n dcc.RadioItems(\n id = 'gender-radio',\n options=[\n {'label': 'Male', 'value': 'male'},\n {'label': 'Female', 'value': 'female'},\n {'label': 'Both', 'value': 'both'}\n ],\n value='both'\n )\n ])\n\n@app.callback(\n dash.dependencies.Output('top5-by-gender', 'figure'),\n [dash.dependencies.Input('gender-radio', 'value')])\ndef update_barchart(gender):\n top5 = df[df.borrower_genders == gender].groupby('activity').size().sort_values(ascending=False)[0:5]\n return {\n 'data': [\n {\n 'x': top5.index,\n 'y': top5,\n 'type': 'bar',\n 'opacity': .6\n }\n ]\n }\n\nif __name__ == '__main__':\n app.run_server(debug=True)\n", "sub_path": "gender_bar_chart.py", "file_name": "gender_bar_chart.py", "file_ext": "py", "file_size_in_byte": 2981, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pandas.read_csv", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 22, "usage_type": "attribute"}, {"api_name": "dash.Dash", "line_number": 44, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 52, "usage_type": "call"}, {"api_name": "dash_html_components.H1", "line_number": 53, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 61, "usage_type": "call"}, {"api_name": "dash_core_components.Graph", "line_number": 61, "usage_type": "call"}, {"api_name": "dash_html_components.Label", "line_number": 75, "usage_type": "call"}, {"api_name": "dash_core_components.RadioItems", "line_number": 76, "usage_type": "call"}, {"api_name": "dash.dependencies.Output", "line_number": 88, "usage_type": "call"}, {"api_name": "dash.dependencies", "line_number": 88, "usage_type": "attribute"}, {"api_name": "dash.dependencies.Input", "line_number": 89, "usage_type": "call"}, {"api_name": "dash.dependencies", "line_number": 89, "usage_type": "attribute"}]}
+{"seq_id": "9468767", "text": "# Copyright (c) 2014-present PlatformIO \n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport os\nfrom fnmatch import fnmatch\n\nimport click\nimport serial\n\nfrom platformio.compat import IS_MACOS, IS_WINDOWS\nfrom platformio.device.list.util import list_logical_devices, list_serial_ports\nfrom platformio.fs import get_platformio_udev_rules_path\nfrom platformio.package.manager.platform import PlatformPackageManager\nfrom platformio.platform.factory import PlatformFactory\nfrom platformio.util import retry\n\nBLACK_MAGIC_HWIDS = [\n \"1D50:6018\",\n]\n\n\ndef parse_udev_rules_hwids(path):\n result = []\n with open(path, mode=\"r\", encoding=\"utf8\") as fp:\n for line in fp.readlines():\n line = line.strip()\n if not line or line.startswith(\"#\"):\n continue\n attrs = {}\n for attr in line.split(\",\"):\n attr = attr.replace(\"==\", \"=\").replace('\"', \"\").strip()\n if \"=\" not in attr:\n continue\n name, value = attr.split(\"=\", 1)\n attrs[name] = value\n hwid = \"%s:%s\" % (\n attrs.get(\"ATTRS{idVendor}\", \"*\"),\n attrs.get(\"ATTRS{idProduct}\", \"*\"),\n )\n if hwid != \"*:*\":\n result.append(hwid.upper())\n return result\n\n\ndef normalize_board_hwid(value):\n if isinstance(value, (list, tuple)):\n value = (\"%s:%s\" % (value[0], value[1])).replace(\"0x\", \"\")\n return value.upper()\n\n\ndef is_pattern_port(port):\n if not port:\n return False\n return set([\"*\", \"?\", \"[\", \"]\"]) & set(port)\n\n\ndef match_serial_port(pattern):\n for item in list_serial_ports():\n if fnmatch(item[\"port\"], pattern):\n return item[\"port\"]\n return None\n\n\ndef is_serial_port_ready(port, timeout=1):\n try:\n serial.Serial(port, timeout=timeout).close()\n return True\n except: # pylint: disable=bare-except\n pass\n return False\n\n\ndef find_serial_port( # pylint: disable=too-many-arguments\n initial_port,\n board_config=None,\n upload_protocol=None,\n ensure_ready=False,\n prefer_gdb_port=False,\n timeout=2,\n):\n if initial_port:\n if not is_pattern_port(initial_port):\n return initial_port\n return match_serial_port(initial_port)\n\n if upload_protocol and upload_protocol.startswith(\"blackmagic\"):\n return find_blackmagic_serial_port(prefer_gdb_port, timeout)\n if board_config and board_config.get(\"build.hwids\", []):\n return find_board_serial_port(board_config, timeout)\n port = find_known_uart_port(ensure_ready, timeout)\n if port:\n return port\n\n # pick the best PID:VID USB device\n best_port = None\n for item in list_serial_ports():\n if ensure_ready and not is_serial_port_ready(item[\"port\"]):\n continue\n port = item[\"port\"]\n if \"VID:PID\" in item[\"hwid\"]:\n best_port = port\n return best_port or port\n\n\ndef find_blackmagic_serial_port(prefer_gdb_port=False, timeout=0):\n try:\n\n @retry(timeout=timeout)\n def wrapper():\n candidates = []\n for item in list_serial_ports(filter_hwid=True):\n if (\n not any(hwid in item[\"hwid\"].upper() for hwid in BLACK_MAGIC_HWIDS)\n and not \"Black Magic\" in item[\"description\"]\n ):\n continue\n if (\n IS_WINDOWS\n and item[\"port\"].startswith(\"COM\")\n and len(item[\"port\"]) > 4\n ):\n item[\"port\"] = \"\\\\\\\\.\\\\%s\" % item[\"port\"]\n candidates.append(item)\n\n if not candidates:\n raise retry.RetryNextException()\n\n for item in candidates:\n if (\"GDB\" if prefer_gdb_port else \"UART\") in item[\"description\"]:\n return item[\"port\"]\n if IS_MACOS:\n # 1 - GDB, 3 - UART\n for item in candidates:\n if item[\"port\"].endswith(\"1\" if prefer_gdb_port else \"3\"):\n return item[\"port\"]\n\n candidates = sorted(candidates, key=lambda item: item[\"port\"])\n return (\n candidates[0] # first port is GDB?\n if len(candidates) == 1 or prefer_gdb_port\n else candidates[1]\n )[\"port\"]\n\n return wrapper()\n except retry.RetryStopException:\n pass\n return None\n\n\ndef find_board_serial_port(board_config, timeout=0):\n hwids = board_config.get(\"build.hwids\", [])\n try:\n\n @retry(timeout=timeout)\n def wrapper():\n for item in list_serial_ports(filter_hwid=True):\n hwid = item[\"hwid\"].upper()\n for board_hwid in hwids:\n if normalize_board_hwid(board_hwid) in hwid:\n return item[\"port\"]\n raise retry.RetryNextException()\n\n return wrapper()\n except retry.RetryStopException:\n pass\n\n click.secho(\n \"TimeoutError: Could not automatically find serial port \"\n \"for the `%s` board based on the declared HWIDs=%s\"\n % (board_config.get(\"name\", \"unknown\"), hwids),\n fg=\"yellow\",\n err=True,\n )\n\n return None\n\n\ndef find_known_uart_port(ensure_ready=False, timeout=0):\n known_hwids = list(BLACK_MAGIC_HWIDS)\n\n # load from UDEV rules\n udev_rules_path = get_platformio_udev_rules_path()\n if os.path.isfile(udev_rules_path):\n known_hwids.extend(parse_udev_rules_hwids(udev_rules_path))\n\n # load from installed dev-platforms\n for platform in PlatformPackageManager().get_installed():\n p = PlatformFactory.new(platform)\n for board_config in p.get_boards().values():\n for board_hwid in board_config.get(\"build.hwids\", []):\n board_hwid = normalize_board_hwid(board_hwid)\n if board_hwid not in known_hwids:\n known_hwids.append(board_hwid)\n\n try:\n\n @retry(timeout=timeout)\n def wrapper():\n for item in list_serial_ports(as_objects=True):\n if not item.vid or not item.pid:\n continue\n hwid = \"{:04X}:{:04X}\".format(item.vid, item.pid)\n for pattern in known_hwids:\n if fnmatch(hwid, pattern) and (\n not ensure_ready or is_serial_port_ready(item.device)\n ):\n return item.device\n raise retry.RetryNextException()\n\n return wrapper()\n except retry.RetryStopException:\n pass\n\n click.secho(\n \"TimeoutError: Could not automatically find serial port \"\n \"based on the known UART bridges\",\n fg=\"yellow\",\n err=True,\n )\n\n return None\n\n\ndef find_mbed_disk(initial_port):\n msdlabels = (\"mbed\", \"nucleo\", \"frdm\", \"microbit\")\n for item in list_logical_devices():\n if item[\"path\"].startswith(\"/net\"):\n continue\n if (\n initial_port\n and is_pattern_port(initial_port)\n and not fnmatch(item[\"path\"], initial_port)\n ):\n continue\n mbed_pages = [os.path.join(item[\"path\"], n) for n in (\"mbed.htm\", \"mbed.html\")]\n if any(os.path.isfile(p) for p in mbed_pages):\n return item[\"path\"]\n if item[\"name\"] and any(l in item[\"name\"].lower() for l in msdlabels):\n return item[\"path\"]\n return None\n", "sub_path": "platformio/device/finder.py", "file_name": "finder.py", "file_ext": "py", "file_size_in_byte": 8049, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "platformio.device.list.util.list_serial_ports", "line_number": 69, "usage_type": "call"}, {"api_name": "fnmatch.fnmatch", "line_number": 70, "usage_type": "call"}, {"api_name": "serial.Serial", "line_number": 77, "usage_type": "call"}, {"api_name": "platformio.device.list.util.list_serial_ports", "line_number": 107, "usage_type": "call"}, {"api_name": "platformio.device.list.util.list_serial_ports", "line_number": 122, "usage_type": "call"}, {"api_name": "platformio.compat.IS_WINDOWS", "line_number": 129, "usage_type": "name"}, {"api_name": "platformio.util.retry.RetryNextException", "line_number": 137, "usage_type": "call"}, {"api_name": "platformio.util.retry", "line_number": 137, "usage_type": "name"}, {"api_name": "platformio.compat.IS_MACOS", "line_number": 142, "usage_type": "name"}, {"api_name": "platformio.util.retry", "line_number": 119, "usage_type": "call"}, {"api_name": "platformio.util.retry.RetryStopException", "line_number": 156, "usage_type": "attribute"}, {"api_name": "platformio.util.retry", "line_number": 156, "usage_type": "name"}, {"api_name": "platformio.device.list.util.list_serial_ports", "line_number": 167, "usage_type": "call"}, {"api_name": "platformio.util.retry.RetryNextException", "line_number": 172, "usage_type": "call"}, {"api_name": "platformio.util.retry", "line_number": 172, "usage_type": "name"}, {"api_name": "platformio.util.retry", "line_number": 165, "usage_type": "call"}, {"api_name": "platformio.util.retry.RetryStopException", "line_number": 175, "usage_type": "attribute"}, {"api_name": "platformio.util.retry", "line_number": 175, "usage_type": "name"}, {"api_name": "click.secho", "line_number": 178, "usage_type": "call"}, {"api_name": "platformio.fs.get_platformio_udev_rules_path", "line_number": 193, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 194, "usage_type": "call"}, {"api_name": "os.path", "line_number": 194, "usage_type": "attribute"}, {"api_name": "platformio.package.manager.platform.PlatformPackageManager", "line_number": 198, "usage_type": "call"}, {"api_name": "platformio.platform.factory.PlatformFactory.new", "line_number": 199, "usage_type": "call"}, {"api_name": "platformio.platform.factory.PlatformFactory", "line_number": 199, "usage_type": "name"}, {"api_name": "platformio.device.list.util.list_serial_ports", "line_number": 210, "usage_type": "call"}, {"api_name": "fnmatch.fnmatch", "line_number": 215, "usage_type": "call"}, {"api_name": "platformio.util.retry.RetryNextException", "line_number": 219, "usage_type": "call"}, {"api_name": "platformio.util.retry", "line_number": 219, "usage_type": "name"}, {"api_name": "platformio.util.retry", "line_number": 208, "usage_type": "call"}, {"api_name": "platformio.util.retry.RetryStopException", "line_number": 222, "usage_type": "attribute"}, {"api_name": "platformio.util.retry", "line_number": 222, "usage_type": "name"}, {"api_name": "click.secho", "line_number": 225, "usage_type": "call"}, {"api_name": "platformio.device.list.util.list_logical_devices", "line_number": 237, "usage_type": "call"}, {"api_name": "fnmatch.fnmatch", "line_number": 243, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 246, "usage_type": "call"}, {"api_name": "os.path", "line_number": 246, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 247, "usage_type": "call"}, {"api_name": "os.path", "line_number": 247, "usage_type": "attribute"}]}
+{"seq_id": "233756002", "text": "import logging\nimport os\nfrom nameko.rpc import rpc\nfrom nameko.messaging import consume\nfrom nameko.extensions import DependencyProvider\nfrom kombu.messaging import Queue, Exchange\nimport bson.json_util\nfrom nameko_slack import web, rtm\n\n_log = logging.getLogger(__name__)\n\n\nclass ErrorHandler(DependencyProvider):\n\n def worker_result(self, worker_ctx, res, exc_info):\n if exc_info is None:\n return\n exc_type, exc, tb = exc_info\n _log.error(str(exc))\n\n\nclass NotifierServiceError(Exception):\n pass\n\n\nclass NotifierService(object):\n name = 'notifier'\n misaki = web.Slack('misaki')\n error = ErrorHandler()\n\n @property\n def channel(self):\n return f'#{os.getenv(\"NOTIFICATION_CHANNEL\", \"notifications\")}'\n\n @staticmethod\n def _format_notification(input_):\n keys = ('id', 'source', 'type', 'content')\n if not all(k in input_.keys() for k in keys):\n raise NotifierServiceError('Some keys are missing in the input dict')\n blocks = [{\n 'type': 'section',\n 'text': {\n 'type': 'mrkdwn',\n 'text': f'*{input_[\"content\"]}*',\n }\n },\n {\n 'type': 'context',\n 'elements': [{'type': 'mrkdwn', 'text': f'{k}: {input_[k]}'}\n for k in input_ if k != 'content']\n }]\n return blocks\n\n @consume(queue=Queue(name='evt_all_notifications',\n exchange=Exchange(name='all_notifications', type='topic', auto_delete=True)))\n def handle_all_notifications(self, payload):\n _log.info(f'Received {payload}')\n input_ = bson.json_util.loads(payload)\n self.misaki.api_call('chat.postMessage', \n channel=self.channel, blocks=self._format_notification(input_), text=input_['content'])\n\n @rpc\n def send_to_slack(self, channel, msg, image_url=None, context=None):\n _log.info(f'Sending message {msg} to slack channel {channel} ...')\n slack_msg = [\n {\n 'type': 'section',\n 'text': {\n 'type': 'mrkdwn',\n 'text': f'*{msg}*'\n }\n }\n ]\n if image_url:\n slack_msg.extend([\n {\n 'type': 'section',\n 'text': {\n 'type': 'mrkdwn',\n 'text': f'Please find your image at the following <{image_url}|link>'\n }\n },\n {\n 'type': 'image',\n 'image_url': image_url,\n 'alt_text': 'Can not be displayed here'\n }\n ])\n if context:\n slack_msg.append({\n 'type': 'context',\n 'elements': [{'type': 'mrkdwn', 'text': context}]\n })\n self.misaki.api_call('chat.postMessage', \n channel=channel, blocks=slack_msg, text=msg)\n\n @rtm.handle_message\n def handle_any_event(self, event, message):\n _log.info(event)\n _log.info(message)\n", "sub_path": "application/services/notifier.py", "file_name": "notifier.py", "file_ext": "py", "file_size_in_byte": 3102, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.getLogger", "line_number": 10, "usage_type": "call"}, {"api_name": "nameko.extensions.DependencyProvider", "line_number": 13, "usage_type": "name"}, {"api_name": "nameko_slack.web.Slack", "line_number": 28, "usage_type": "call"}, {"api_name": "nameko_slack.web", "line_number": 28, "usage_type": "name"}, {"api_name": "os.getenv", "line_number": 33, "usage_type": "call"}, {"api_name": "bson.json_util.json_util.loads", "line_number": 58, "usage_type": "call"}, {"api_name": "bson.json_util.json_util", "line_number": 58, "usage_type": "attribute"}, {"api_name": "bson.json_util", "line_number": 58, "usage_type": "name"}, {"api_name": "nameko.messaging.consume", "line_number": 54, "usage_type": "call"}, {"api_name": "kombu.messaging.Queue", "line_number": 54, "usage_type": "call"}, {"api_name": "kombu.messaging.Exchange", "line_number": 55, "usage_type": "call"}, {"api_name": "nameko.rpc.rpc", "line_number": 62, "usage_type": "name"}, {"api_name": "nameko_slack.rtm.handle_message", "line_number": 97, "usage_type": "attribute"}, {"api_name": "nameko_slack.rtm", "line_number": 97, "usage_type": "name"}]}
+{"seq_id": "59144344", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations, models\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ]\n\n operations = [\n migrations.CreateModel(\n name='Email',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('date_created', models.DateTimeField(auto_now_add=True, verbose_name='Cr\\xe9\\xe9 le')),\n ('name', models.CharField(max_length=254, verbose_name='Nom')),\n ('email', models.EmailField(max_length=254, verbose_name='Adresse email')),\n ('title', models.CharField(max_length=254, null=True, verbose_name='Title', blank=True)),\n ('template', models.TextField(verbose_name='Template')),\n ('is_sent', models.BooleanField(default=False, verbose_name='Envoy\\xe9 ?')),\n ],\n options={\n 'ordering': ('-date_created',),\n 'verbose_name': 'Email',\n 'verbose_name_plural': 'Emails',\n },\n ),\n ]\n", "sub_path": "Lib/site-packages/django_extended/contrib/emailing/migrations/0001_initial.py", "file_name": "0001_initial.py", "file_ext": "py", "file_size_in_byte": 1145, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.EmailField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}]}
+{"seq_id": "219683091", "text": "import pickle\ndef writePickle( Variable, fname):\n filename = fname +\".pkl\"\n f = open(\"pickle_vars/\"+filename, 'wb')\n pickle.dump(Variable, f)\n f.close()\ndef readPickle(fname):\n filename = \"../pickle_vars/\"+fname +\".pkl\" # notice the ../ addition to reach out to variables from the parent directory\n f = open(filename, 'rb')\n obj = pickle.load(f)\n f.close()\n return obj\n\n\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Dropout, Conv2D, Flatten, MaxPooling2D\n\n#import variables\nX_train_RID = readPickle(\"cnn_data_inputs/RID_Keras/X_train_RID\")\nX_dev_RID = readPickle(\"cnn_data_inputs/RID_Keras/X_dev_RID\")\nX_test_RID = readPickle(\"cnn_data_inputs/RID_Keras/X_test_RID\")\ny_train_RID = readPickle(\"cnn_data_inputs/RID_Keras/y_train_RID\")\ny_dev_RID = readPickle(\"cnn_data_inputs/RID_Keras/y_dev_RID\")\ny_test_RID = readPickle(\"cnn_data_inputs/RID_Keras/y_test_RID\")\nmax_line = 20\nmax_song = 100\nArtist2id = readPickle(\"indexing/Artist2id\")\n\n\n# create a sequential model\nmodel = Sequential()\n\n# add model layers\nmodel.add(Conv2D(32, kernel_size=3, padding = \"same\", activation=\"relu\", input_shape=(max_song,max_line,1)))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\nmodel.add(Conv2D(64, kernel_size=3, padding = \"same\", activation=\"relu\"))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\nmodel.add(Dropout(0.25))\nmodel.add(Flatten())\nmodel.add(Dense(len(list(Artist2id.keys())), activation=\"softmax\")) #here we need to find the length \\\n # of the potential labels\n\n# compile all the layers \nmodel.compile(optimizer='adam', loss='categorical_crossentropy', metrics=['accuracy'])\n\n# train the model using the development set\nmodel.fit(X_train_RID, y_train_RID, validation_data=(X_dev_RID, y_dev_RID), epochs=25, batch_size=128)\n\n\n# save the predictions on the test set\npredictions = model.predict(X_test_RID)\n\n# save the predictions in a pickle file that is later to be used for evaluation:\nwritePickle(predictions, \"predictions_RID_33\")\n\n# show the accuracy of the trained model on test set\nscore = model.evaluate(X_test_RID, y_test_RID, verbose=0)\nprint('Test loss:', score[0])\nprint('Test accuracy:', score[1])\n\n# show the model summary and save the model\nmodel.summary()\nmodel.save(\"../Saved_Models/RID_25ep_128bch_025Drop_33filter\")\n", "sub_path": "Modeling/Model Training/RID.py", "file_name": "RID.py", "file_ext": "py", "file_size_in_byte": 2354, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pickle.dump", "line_number": 5, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 10, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 31, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 34, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 35, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 36, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 37, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 38, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 39, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 40, "usage_type": "call"}]}
+{"seq_id": "326081100", "text": "import graphene\n\nimport graphql_social_auth\n\nfrom . import mixins\nfrom .testcases import RelaySchemaTestCase\n\n\nclass SocialAuthTests(mixins.SocialAuthCompleteMixin, RelaySchemaTestCase):\n query = '''\n mutation SocialAuthComplete($input: SocialAuthCompleteInput!) {\n socialAuthComplete(input: $input) {\n result {\n __typename\n ... on Social {\n social {\n uid\n extraData\n }\n }\n }\n clientMutationId\n }\n }'''\n\n class Mutations(graphene.ObjectType):\n social_auth_complete = graphql_social_auth.relay.SocialAuthComplete.Field()\n\n\nclass SocialAuthJWTTests(mixins.SocialAuthJWTCompleteMixin,\n RelaySchemaTestCase):\n\n query = '''\n mutation SocialAuthComplete($input: SocialAuthJWTCompleteInput!) {\n socialAuthComplete(input: $input) {\n result {\n __typename\n ... on JWT {\n social {\n uid\n extraData\n }\n token\n }\n }\n clientMutationId\n }\n }'''\n\n class Mutations(graphene.ObjectType):\n social_auth_complete = graphql_social_auth.relay.SocialAuthJWTComplete.Field()\n", "sub_path": "tests/test_relay.py", "file_name": "test_relay.py", "file_ext": "py", "file_size_in_byte": 1234, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "testcases.RelaySchemaTestCase", "line_number": 9, "usage_type": "name"}, {"api_name": "graphene.ObjectType", "line_number": 26, "usage_type": "attribute"}, {"api_name": "graphql_social_auth.relay.SocialAuthComplete.Field", "line_number": 27, "usage_type": "call"}, {"api_name": "graphql_social_auth.relay", "line_number": 27, "usage_type": "attribute"}, {"api_name": "testcases.RelaySchemaTestCase", "line_number": 31, "usage_type": "name"}, {"api_name": "graphene.ObjectType", "line_number": 50, "usage_type": "attribute"}, {"api_name": "graphql_social_auth.relay.SocialAuthJWTComplete.Field", "line_number": 51, "usage_type": "call"}, {"api_name": "graphql_social_auth.relay", "line_number": 51, "usage_type": "attribute"}]}
+{"seq_id": "64163626", "text": "# coding=utf-8\n# --------------------------------------------------------------------------\n# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License. See License.txt in the project root for\n# license information.\n#\n# Code generated by Microsoft (R) AutoRest Code Generator.\n# Changes may cause incorrect behavior and will be lost if the code is\n# regenerated.\n# --------------------------------------------------------------------------\n\nfrom msrest.serialization import Model\n\n\nclass EncryptionSettingsCollection(Model):\n \"\"\"Encryption settings for disk or snapshot.\n\n All required parameters must be populated in order to send to Azure.\n\n :param enabled: Required. Set this flag to true and provide\n DiskEncryptionKey and optional KeyEncryptionKey to enable encryption. Set\n this flag to false and remove DiskEncryptionKey and KeyEncryptionKey to\n disable encryption. If EncryptionSettings is null in the request object,\n the existing settings remain unchanged.\n :type enabled: bool\n :param encryption_settings: A collection of encryption settings, one for\n each disk volume.\n :type encryption_settings:\n list[~azure.mgmt.compute.v2018_09_30.models.EncryptionSettingsElement]\n \"\"\"\n\n _validation = {\n 'enabled': {'required': True},\n }\n\n _attribute_map = {\n 'enabled': {'key': 'enabled', 'type': 'bool'},\n 'encryption_settings': {'key': 'encryptionSettings', 'type': '[EncryptionSettingsElement]'},\n }\n\n def __init__(self, *, enabled: bool, encryption_settings=None, **kwargs) -> None:\n super(EncryptionSettingsCollection, self).__init__(**kwargs)\n self.enabled = enabled\n self.encryption_settings = encryption_settings\n", "sub_path": "azure-mgmt-compute/azure/mgmt/compute/v2018_09_30/models/encryption_settings_collection_py3.py", "file_name": "encryption_settings_collection_py3.py", "file_ext": "py", "file_size_in_byte": 1758, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "msrest.serialization.Model", "line_number": 15, "usage_type": "name"}]}
+{"seq_id": "242813811", "text": "\n\nfrom lib.Settings import Settings\nfrom lib.Misc.CheatCodes import CheatCodes\n\nfrom lib.Misc.Sounds import Sounds\n\nfrom lib.Boards.WelcomeBoard import WelcomeBoard\nfrom lib.Boards.Scoreboard import Scoreboard\nfrom lib.Maze.Maze import Maze\nfrom lib.Popups.Popups import Popups\n\nfrom lib.Actors.Player import Player\nfrom lib.Actors.Blinky import Blinky\nfrom lib.Actors.Pinky import Pinky\nfrom lib.Actors.Inky import Inky\nfrom lib.Actors.Clyde import Clyde\nfrom lib.Actors.Fruit import Fruit\n\nfrom lib.Maze.Portal import Portal\n\nimport pygame\n\nimport math\nimport random\nimport sys\n\n\nclass Pacman:\n\t\n\tdef __init__(self):\n\t\t\n\t\tself.settings = Settings()\n\t\t\n\t\tself.clock = None\n\t\t\n\t\tself.screen = None\n\t\tself.screen_rect = None\n\t\t\n\t\tself.initialize_pygame()\n\t\t\n\t\tself.welcome_board = None\n\t\tself.scoreboard = None\n\t\t\n\t\tself.sounds = None\n\t\t\n\t\tself.maze = None\n\t\tself.popups = None\n\t\t\n\t\tself.player = None\n\t\tself.ghosts = None\n\t\tself.fruits = []\n\t\tself.__fruits_remaining = self.settings.fruit_per_round\n\t\tself.__fruits_eaten = None\n\t\t\n\t\tself.portal = None\n\t\t\n\t\tself.cheater = None\n\t\t\n\t\tself.playing_round = False\n\t\tself.ghosts_eaten_since_last_power_pellet = 0\n\t\n\tdef initialize_pygame(self):\n\t\n\t\tprint(\"Initializing pygame\")\n\t\t\n\t\t# Initialize mixer, ***THEN*** pygame\n\t\t# I have no idea why, but calling pygame.init() before the mixer, introduces lag\n\t\tpygame.mixer.pre_init(self.settings.sound_sample_rate, -16, self.settings.sound_channels, 1024)\n\t\tpygame.mixer.init()\n\t\tpygame.init()\n\t\t\n\t\t# Clock\n\t\tself.clock = pygame.time.Clock()\n\t\t\n\t\t# Get the monitor's size\n\t\t# This fails horribly on a system with 2+ monitors\n\t\t# monitor_info = pygame.display.Info()\n\t\t# monitor_width = monitor_info.current_w\n\t\t# monitor_height = monitor_info.current_h\n\t\t# width = int(monitor_width * self.settings.screen_size_factor)\n\t\t# height = int(monitor_height * self.settings.screen_size_factor)\n\t\t\n\t\twidth = int(self.settings.screen_width * self.settings.screen_size_factor)\n\t\theight = int(self.settings.screen_height * self.settings.screen_size_factor)\n\t\t\n\t\t# Screen\n\t\tself.screen = pygame.display.set_mode((width, height))\n\t\tself.screen_rect = self.screen.get_rect()\n\t\tpygame.display.set_caption(self.settings.game_title)\n\t\n\tdef run(self):\n\t\t\n\t\tself.initialize_gamestuffs()\n\t\t\n\t\tself.sounds.play_theater_music()\n\t\t\n\t\tself.main_loop()\n\t\t\n\tdef main_loop(self):\n\t\t\n\t\twhile True:\n\t\t\t\n\t\t\tself.main_loop_one()\n\t\n\tdef main_loop_one(self):\n\t\t\n\t\telapsed_ms = self.clock.tick()\n\t\t\n\t\t# Keep elapsed time to a sane range\n\t\tif elapsed_ms < self.settings.minimum_elapsed_ms:\n\t\t\telapsed_ms = self.settings.minimum_elapsed_ms\n\t\telif elapsed_ms > self.settings.maximum_elapsed_ms:\n\t\t\telapsed_ms = self.settings.maximum_elapsed_ms\n\t\t\n\t\tself.handle_events()\n\t\t\n\t\tself.update(elapsed_ms)\n\t\t\n\t\tself.draw()\n\t\n\tdef handle_events(self):\n\t\t\n\t\t# Handle all events\n\t\tfor event in pygame.event.get():\n\t\t\t\n\t\t\t# Allow the cheater to peek\n\t\t\tself.cheater.peek_at_event(event)\n\t\t\t\n\t\t\t#\n\t\t\tif event.type == pygame.QUIT:\n\t\t\t\tsys.exit()\n\t\t\t\n\t\t\t# Allow the welcome board to sneak in\n\t\t\telif self.welcome_board.handle_event(event):\n\t\t\t\tcontinue\n\t\t\t\n\t\t\t#\n\t\t\telif event.type == pygame.KEYUP or event.type == pygame.KEYDOWN:\n\t\t\t\tself.handle_keyboard_event(event)\n\t\n\tdef handle_keyboard_event(self, event):\n\t\t\n\t\t# Q for Quit\n\t\tif event.key == pygame.K_q:\n\t\t\tsys.exit(0)\n\t\t\n\t\t# Arrows navigate the human player\n\t\telif (\n\t\t\tevent.key == pygame.K_UP\n\t\t\tor event.key == pygame.K_DOWN\n\t\t\tor event.key == pygame.K_LEFT\n\t\t\tor event.key == pygame.K_RIGHT\n\t\t):\n\t\t\tif event.key == pygame.K_UP:\n\t\t\t\tself.player.set_movement_input_up()\n\t\t\t\tself.player.set_movement_input_down(False)\n\t\t\telif event.key == pygame.K_DOWN:\n\t\t\t\tself.player.set_movement_input_up(False)\n\t\t\t\tself.player.set_movement_input_down()\n\t\t\tif event.key == pygame.K_LEFT:\n\t\t\t\tself.player.set_movement_input_left()\n\t\t\t\tself.player.set_movement_input_right(False)\n\t\t\telif event.key == pygame.K_RIGHT:\n\t\t\t\tself.player.set_movement_input_left(False)\n\t\t\t\tself.player.set_movement_input_right()\n\t\t\n\t\t# Space bar or \"P\" launches a portal\n\t\telif (\n\t\t\t(event.key == pygame.K_SPACE or event.key == pygame.K_p)\n\t\t\tand event.type == pygame.KEYUP\n\t\t):\n\t\t\tself.spawn_portal()\n\t\n\tdef handle_cheat_codes(self):\n\t\n\t\tcode = self.cheater.get_code()\n\t\tif code:\n\t\t\n\t\t\tif code == \"rects\":\n\t\t\t\n\t\t\t\tself.settings.maze_draw_debug_rects = (\n\t\t\t\t\tnot self.settings.maze_draw_debug_rects\n\t\t\t\t)\n\t\t\t\n\t\t\telif code == \"paths\":\n\t\t\t\t\n\t\t\t\tself.settings.debug_show_paths = (\n\t\t\t\t\tnot self.settings.debug_show_paths\n\t\t\t\t)\n\t\t\t\n\t\t\telif code == \"frames\":\n\t\t\t\t\n\t\t\t\tself.settings.animation_draw_frames = (\n\t\t\t\t\tnot self.settings.animation_draw_frames\n\t\t\t\t)\n\t\t\t\n\t\t\telif code == \"fruit\":\n\t\t\t\t\n\t\t\t\tself.spawn_fruit()\n\t\t\t\n\t\t\telif code == \"die\":\n\t\t\t\t\n\t\t\t\tif self.do_player_death():\n\t\t\t\t\tself.continue_round()\n\t\t\t\n\t\t\telif code == \"kill\":\n\t\t\t\t\n\t\t\t\tself.ghosts_eaten_since_last_power_pellet = 0\n\t\t\t\tfor ghost in self.ghosts:\n\t\t\t\t\tself.player_ate_ghost(ghost)\n\t\t\t\n\t\t\telif code == \"scare\":\n\t\t\t\n\t\t\t\tself.scare_all_ghosts()\n\t\t\t\n\t\t\telif code == \"clear\":\n\t\t\t\t\n\t\t\t\tbrick_me = self.player.get_closest_brick()\n\t\t\t\tbricks_me = self.maze.get_cluster_around_brick(brick_me, None)\n\t\t\t\tfor brick in self.maze.bricks:\n\t\t\t\t\tif brick not in bricks_me:\n\t\t\t\t\t\tif brick.is_pill() or brick.is_power_pellet():\n\t\t\t\t\t\t\tbrick.make_blank()\n\t\t\t\tself.maze.remember_power_pellets()\n\t\t\t\tself.maze.remember_pills()\n\t\t\t\tself.maybe_win_round()\n\t\t\t\n\t\t\telif code == \"win\":\n\t\t\t\t\n\t\t\t\tself.do_win_round()\n\t\t\t\n\t\t\telif code == \"5level\":\n\t\t\t\t\n\t\t\t\tfor i in range(5):\n\t\t\t\t\tself.increase_level()\n\t\t\t\n\t\t\telif code == \"level\":\n\t\t\t\t\n\t\t\t\tself.increase_level()\n\t\t\t\n\t\t\telse:\n\t\t\t\tprint(\"Invalid cheat code:\", code)\n\t\n\tdef initialize_gamestuffs(self):\n\t\t\n\t\tself.initialize_cheater()\n\t\t\n\t\tself.initialize_sounds()\n\t\tself.initialize_maze()\n\t\tself.initialize_boards()\n\t\tself.initialize_popups()\n\t\t\n\t\tself.initialize_player()\n\t\tself.initialize_ghosts()\n\t\t\n\tdef start_game(self):\n\t\t\n\t\tself.initialize_gamestuffs()\n\t\tprint(\"Starting game!!!\")\n\t\t\n\t\tself.start_round()\n\t\n\tdef start_round(self):\n\t\t\n\t\t#\n\t\tif self.welcome_board.is_activated():\n\t\t\treturn\n\t\t\n\t\t#\n\t\tself.playing_round = False\n\t\t\n\t\tself.__fruits_eaten = 0\n\t\tself.__fruits_remaining = self.settings.fruit_per_round\n\t\t\n\t\t# Increment level\n\t\tself.increase_level()\n\t\t\n\t\t# Load up the maze\n\t\tself.maze.load_maze_file(\"test\")\n\t\t\n\t\tself.continue_round()\n\t\n\tdef continue_round(self):\n\t\t\n\t\tself.playing_round = True\n\t\t\n\t\t# Reset player's animation\n\t\tself.player.set_default_animation()\n\t\t\n\t\t# Reinitialize the ghosts\n\t\tself.initialize_ghosts()\n\t\t\n\t\t# Place the player and ghosts\n\t\tself.place_player_at_random_pill()\n\t\tself.place_ghosts_in_pen()\n\t\t\n\t\tself.do_round_intro()\n\t\t\n\t\tself.sounds.play_ghost_sounds()\n\t\t\n\t\t# Reset ghost penned time\n\t\tfor ghost in self.ghosts:\n\t\t\tghost.set_penned_state(\"in\")\n\t\n\tdef do_round_intro(self):\n\t\t\n\t\tself.freeze_actors()\n\t\t\n\t\tself.sounds.play_intro_music()\n\t\t\n\t\tself.do_popup_game_state_message(\"Get Ready!\")\n\t\t\n\t\tself.unfreeze_actors()\n\t\n\tdef game_over(self):\n\t\t\n\t\tself.freeze_actors()\n\t\t\n\t\tself.sounds.play_game_over()\n\t\tself.do_popup_game_state_message(\"Game Over\")\n\t\t\n\t\tself.welcome_board.register_final_score(self.player.get_score())\n\t\tself.welcome_board.activate()\n\t\t\n\t\tself.welcome_board.show_high_scores()\n\t\t\n\t\tself.sounds.play_theater_music()\n\t\n\tdef do_popup_game_state_message(self, text):\n\t\t\n\t\t# Spawn a popup in the dead center\n\t\tpopup_x = int(self.screen_rect.width / 2)\n\t\tpopup_y = int(self.screen_rect.height / 2)\n\t\tpopup = self.popups.spawn_popup(text, popup_x, popup_y)\n\t\tpopup.set_spawn_time(self.settings.round_intro_title_spawn_time)\n\t\tpopup.set_hold_time(self.settings.round_intro_title_hold_time)\n\t\t\n\t\twhile self.popups.get_popups_count() > 0:\n\t\t\tself.main_loop_one()\n\t\n\tdef place_player_at_random_pill(self):\n\t\t\n\t\t# Make sure the maze knows about itself\n\t\tself.maze.draw(True)\n\t\t\n\t\tbrick = self.maze.get_random_pill_brick()\n\t\tif not brick:\n\t\t\tbrick = self.maze.get_random_power_pellet_brick()\n\t\tif not brick:\n\t\t\tbrick = self.maze.get_random_blank_brick()\n\t\tif not brick:\n\t\t\treturn\n\t\t\n\t\tself.player.move_to_maze_brick(brick)\n\t\n\tdef place_ghosts_in_pen(self):\n\t\t\n\t\t# Make sure maze knows about itself\n\t\tself.maze.draw(True)\n\t\t\n\t\tfor ghost in self.ghosts:\n\t\t\tself.place_ghost_in_pen(ghost)\n\t\n\tdef place_ghost_in_pen(self, ghost):\n\t\t\n\t\t# Locate a random pen\n\t\tbrick = self.maze.get_random_pen_brick()\n\t\tghost.move_to_maze_brick(brick)\n\t\tghost.set_penned_state(\"in\")\n\t\t\n\tdef increase_level(self):\n\t\t\n\t\t# Increment level\n\t\tself.scoreboard.adjust_level(1)\n\t\tself.settings.adjust_settings_to_level(self.scoreboard.get_level())\n\t\n\tdef get_level(self):\n\t\t\n\t\treturn self.scoreboard.get_level()\n\t\n\tdef initialize_sounds(self):\n\t\t\n\t\tself.sounds = Sounds()\n\t\n\tdef initialize_maze(self):\n\t\t\n\t\t# Use screen width\n\t\tmaze_width = None\n\t\t\n\t\t# Use screen height minus scoreboard height\n\t\tmaze_height = self.screen_rect.height - self.settings.scoreboard_height\n\t\t\n\t\tself.maze = Maze(\n\t\t\tself.settings, self.screen,\n\t\t\tmaze_width, maze_height\n\t\t)\n\t\n\tdef initialize_boards(self):\n\t\t\n\t\t# Welcome Board should only be initialized once\n\t\tif self.welcome_board is None:\n\t\t\tself.welcome_board = WelcomeBoard(\n\t\t\t\tself.settings, self.screen,\n\t\t\t\tNone, None,\n\t\t\t\tself\n\t\t\t)\n\t\t\n\t\t# Scoreboard\n\t\tself.scoreboard = Scoreboard(\n\t\t\tself.settings, self.screen, None, self.settings.scoreboard_height, self\n\t\t)\n\t\tself.scoreboard.set_blit_offset(\n\t\t\tNone,\n\t\t\tself.screen_rect.height - self.settings.scoreboard_height\n\t\t)\n\t\n\tdef initialize_popups(self):\n\t\t\n\t\tself.popups = Popups(\n\t\t\tself.settings, self.screen\n\t\t)\n\t\n\tdef initialize_player(self):\n\t\t\n\t\tself.player = Player(\n\t\t\tself.settings,\n\t\t\tself.maze.bricks_size,\n\t\t\tself, self.maze, self.screen,\n\t\t\tself.scoreboard\n\t\t)\n\t\n\tdef get_player(self):\n\t\t\n\t\treturn self.player\n\t\n\tdef initialize_ghosts(self):\n\t\t\n\t\tself.ghosts = []\n\t\t\n\t\tfor i in range(self.settings.ghost_duplicate_count):\n\t\t\t\n\t\t\tblinky = Blinky(self.settings, self.maze.bricks_size, self, self.maze, self.screen)\n\t\t\tself.ghosts.append(blinky)\n\t\t\t\n\t\t\tpinky = Pinky(self.settings, self.maze.bricks_size, self, self.maze, self.screen)\n\t\t\tself.ghosts.append(pinky)\n\t\t\t\n\t\t\tclyde = Clyde(self.settings, self.maze.bricks_size, self, self.maze, self.screen)\n\t\t\tself.ghosts.append(clyde)\n\t\t\t\n\t\t\tinky = Inky(self.settings, self.maze.bricks_size, self, self.maze, self.screen)\n\t\t\tself.ghosts.append(inky)\n\t\n\tdef initialize_cheater(self):\n\t\t\n\t\tself.cheater = CheatCodes(self.settings.cheat_codes)\n\t\n\tdef update(self, elapsed_ms):\n\t\t\n\t\tself.handle_cheat_codes()\n\t\t\n\t\t# Place this before the update loops,\n\t\t# so the gamestart happens after the board\n\t\t# is deactivated\n\t\tif self.welcome_board.wants_to_start():\n\t\t\tself.welcome_board.deactivate()\n\t\t\tself.start_game()\n\t\t\n\t\t# Update the welcome board?\n\t\tif self.welcome_board.is_activated():\n\t\t\t\n\t\t\tself.welcome_board.update(elapsed_ms)\n\t\t\t\n\t\telse:\n\t\t\t\n\t\t\tself.scoreboard.update(elapsed_ms)\n\t\t\t\n\t\t\tself.maze.clear_debug_rects()\n\t\t\tself.maze.update(elapsed_ms)\n\t\t\t\n\t\t\tself.update_board_actors(elapsed_ms)\n\t\t\n\t\tself.popups.update(elapsed_ms)\n\t\t\n\tdef update_board_actors(self, elapsed_ms):\n\t\t\n\t\tself.maybe_spawn_fruit(elapsed_ms)\n\t\t\n\t\tself.player.update(elapsed_ms)\n\t\t\n\t\tif not self.player.is_dying():\n\t\t\tfor ghost in self.ghosts:\n\t\t\t\tghost.update(elapsed_ms)\n\t\t\n\t\tfor fruit in self.fruits:\n\t\t\t\n\t\t\tfruit.update(elapsed_ms)\n\t\t\t\n\t\t\tif fruit.wants_to_die():\n\t\t\t\tself.fruits.remove(fruit)\n\t\t\n\t\tif self.portal:\n\t\t\tself.portal.update(elapsed_ms)\n\t\n\tdef ghost_ate_player(self, ghost):\n\t\t\n\t\tif self.player.is_dying() or (not self.playing_round):\n\t\t\treturn\n\t\t\n\t\t# HAHA JK STUPID GHOST\n\t\tif ghost.is_scared():\n\t\t\tself.player_ate_ghost(ghost)\n\t\t\treturn\n\t\t\n\t\tself.playing_round = True\n\t\t\n\t\tif self.do_player_death():\n\t\t\tself.continue_round()\n\t\n\tdef player_ate_ghost(self, ghost):\n\t\t\n\t\tpoints = self.settings.ghost_points_base\n\t\tpoints *= math.pow(\n\t\t\tself.settings.ghost_multiple_eaten_multiplier, self.ghosts_eaten_since_last_power_pellet\n\t\t)\n\t\tpoints = round(points)\n\t\t\n\t\tself.player.adjust_score(points)\n\t\tself.spawn_popup_at_actor(ghost, str(points))\n\t\t\n\t\tself.ghosts_eaten_since_last_power_pellet += 1\n\t\t\n\t\tghost.do_die()\n\t\t\n\t\tself.sounds.play_eat_ghost()\n\t\n\tdef player_ate_pill(self):\n\t\t\n\t\tself.sounds.play_eating_pill()\n\t\t\n\t\tself.maze.remember_pills()\n\t\t\n\t\tself.maybe_win_round()\n\t\t\n\tdef player_ate_power_pellet(self, brick, points):\n\t\t\n\t\tself.maze.remember_power_pellets()\n\t\t\n\t\tself.ghosts_eaten_since_last_power_pellet = 0\n\t\tself.scare_all_ghosts()\n\t\t\n\t\tself.spawn_popup_at_brick(brick, str(round(points)))\n\t\t\n\t\tself.sounds.play_eat_power_pellet()\n\t\t\n\t\tself.maybe_win_round()\n\t\n\tdef player_ate_fruit(self, fruit):\n\t\t\n\t\t# Determine points\n\t\tpoints = self.settings.fruit_points_base\n\t\tpoints *= pow(2, self.__fruits_eaten)\n\t\t\n\t\tself.player.adjust_score(points)\n\t\tself.spawn_popup_at_actor(fruit, str(round(points)))\n\t\t\n\t\tself.fruits.remove(fruit)\n\t\tself.__fruits_eaten += 1\n\t\t\n\t\tself.sounds.play_eat_fruit()\n\t\n\tdef player_hit_portal(self):\n\t\n\t\tself.actor_hit_portal(self.player)\n\t\t\n\t\tself.sounds.play_using_portal()\n\t\t\n\t\tself.portal = None\n\t\n\tdef ghost_hit_portal(self, ghost):\n\t\t\n\t\tself.actor_hit_portal(ghost)\n\t\n\tdef actor_hit_portal(self, actor):\n\t\t\n\t\t# Grab a random pill brick, or power pellet\n\t\tbrick = self.maze.get_random_pill_brick()\n\t\tif not brick:\n\t\t\tbrick = self.maze.get_random_power_pellet_brick()\n\t\t\tif not brick:\n\t\t\t\tprint(\"Unable to find any bricks to teleport to\")\n\t\t\t\treturn\n\t\t\n\t\t#\n\t\tactor.move_to_maze_brick(brick)\n\t\n\tdef maybe_win_round(self):\n\t\t\n\t\tpills_count = len(self.maze.bricks_pills)\n\t\tpower_pellets_count = len(self.maze.bricks_power_pellets)\n\t\t\n\t\t# Check for any remaining pills\n\t\tif pills_count > 0:\n\t\t\treturn\n\t\tif power_pellets_count > 0:\n\t\t\treturn\n\t\t\n\t\tself.do_win_round()\n\t\n\tdef do_win_round(self):\n\t\t\n\t\tself.freeze_actors()\n\t\t\n\t\tself.unscare_all_ghosts()\n\t\t\n\t\tself.sounds.stop_ghost_sounds()\n\t\tself.sounds.play_round_win()\n\t\t\n\t\t# Spawn a popup in the dead center\n\t\tpopup_x = int(self.screen_rect.width / 2)\n\t\tpopup_y = int(self.screen_rect.height / 2)\n\t\tpopup = self.popups.spawn_popup(\"Great Job! You Win !!!!\", popup_x, popup_y)\n\t\tpopup.set_spawn_time(self.settings.round_intro_title_spawn_time)\n\t\tpopup.set_hold_time(self.settings.round_intro_title_hold_time)\n\t\t\n\t\twhile self.popups.get_popups_count() > 0:\n\t\t\tself.main_loop_one()\n\t\t\n\t\tself.start_round()\n\t\n\tdef freeze_actors(self):\n\t\n\t\tself.player.set_can_move(False)\n\t\tfor ghost in self.ghosts:\n\t\t\tghost.set_can_move(False)\n\t\n\tdef unfreeze_actors(self):\n\t\t\n\t\tself.player.set_can_move(True)\n\t\tfor ghost in self.ghosts:\n\t\t\tghost.set_can_move(True)\n\t\n\t# Return True if the player has lives left\n\tdef do_player_death(self):\n\t\t\n\t\tself.playing_round = False\n\t\t\n\t\tself.freeze_actors()\n\t\tself.player.do_die()\n\t\t\n\t\tself.sounds.stop_ghost_sounds()\n\t\tself.sounds.play_player_death()\n\t\t\n\t\twhile self.player.is_dying():\n\t\t\t\n\t\t\tself.main_loop_one()\n\t\t\n\t\tself.unfreeze_actors()\n\t\t\n\t\tlives_remaining = self.scoreboard.get_lives_remaining()\n\t\thas_lives_left = lives_remaining > 0\n\t\t\n\t\tif has_lives_left:\n\t\t\t\n\t\t\tself.scoreboard.adjust_lives_remaining(-1)\n\t\t\n\t\telse:\n\t\t\t\n\t\t\tself.game_over()\n\t\t\t\n\t\treturn has_lives_left\n\t\n\tdef scare_all_ghosts(self):\n\t\t\n\t\tfor ghost in self.ghosts:\n\t\t\t\n\t\t\tghost.scare()\n\t\t\n\t\tself.sounds.play_ghost_sounds(True)\n\t\n\tdef unscare_all_ghosts(self):\n\t\t\n\t\tfor ghost in self.ghosts:\n\t\t\t\n\t\t\tghost.stop_being_scared()\n\t\t\n\t\tself.sounds.play_ghost_sounds(False)\n\t\n\tdef ghost_stopped_being_scared(self):\n\t\n\t\tfor ghost in self.ghosts:\n\t\t\tif ghost.is_scared():\n\t\t\t\treturn\n\t\t\n\t\tself.sounds.play_ghost_sounds(False)\n\t\n\tdef spawn_popup_at_actor(self, actor, text):\n\t\n\t\tx, y = actor.get_screen_location()\n\t\t\n\t\toffset_x, offset_y = self.maze.get_blit_offset()\n\t\tx += offset_x\n\t\ty += offset_y\n\t\t\n\t\tself.popups.spawn_popup(text, x, y)\n\t\n\tdef spawn_popup_at_brick(self, brick, text):\n\t\t\n\t\tx, y = brick.get_surface_coords()\n\t\t\n\t\toffset_x, offset_y = self.maze.get_blit_offset()\n\t\tx += offset_x\n\t\ty += offset_y\n\t\t\n\t\tself.popups.spawn_popup(text, x, y)\n\t\n\tdef maybe_spawn_fruit(self, elapsed_ms):\n\t\t\n\t\tif self.__fruits_remaining < 1:\n\t\t\treturn\n\t\t\n\t\t#\n\t\tif len(self.fruits) >= self.settings.fruit_max_concurrency:\n\t\t\treturn\n\t\t\n\t\t#\n\t\tchance = self.settings.fruit_chance_per_ms * elapsed_ms\n\t\tr = random.uniform(0, 1)\n\t\tif r <= chance:\n\t\t\tself.spawn_fruit()\n\t\n\tdef spawn_fruit(self):\n\t\t\n\t\tmax_time = random.gauss(\n\t\t\tself.settings.fruit_max_time_base,\n\t\t\tself.settings.fruit_max_time_std\n\t\t)\n\t\t\n\t\tcenter_x = int(self.screen_rect.width / 2)\n\t\t\n\t\tself.__fruits_remaining -= 1\n\t\t\n\t\tfruit = Fruit(\n\t\t\tself.settings, max_time, \"Fruit\", self.maze.bricks_size,\n\t\t\tself, self.maze, self.screen\n\t\t)\n\t\t\n\t\tbrick = self.maze.get_random_tunnel_brick()\n\t\tfruit.move_to_maze_brick(brick)\n\t\t\n\t\tfruit_x, fruit_y = fruit.get_screen_location()\n\t\t\n\t\t# Start off in the correct direction\n\t\tfruit.clear_movement_input()\n\t\tif fruit_x > center_x:\n\t\t\tfruit.set_movement_input_left()\n\t\telse:\n\t\t\tfruit.set_movement_input_right()\n\t\t\n\t\tself.fruits.append(fruit)\n\t\t\n\t\tself.scoreboard.set_fruit_count(self.__fruits_remaining)\n\t\n\tdef spawn_portal(self):\n\t\n\t\tportal = Portal(\n\t\t\tself.settings,\n\t\t\t\"Portal\", self.maze.bricks_size,\n\t\t\tself, self.maze, self.screen\n\t\t)\n\t\t\n\t\t# Get player's current brick,\n\t\t# and find a wall in the same direction as the player\n\t\t# is moving\n\t\tx = y = 0\n\t\tif self.player.get_movement_input_left() and not self.player.is_impeded_left():\n\t\t\tx = -1\n\t\telif self.player.get_movement_input_right() and not self.player.is_impeded_right():\n\t\t\tx = 1\n\t\telif self.player.get_movement_input_up() and not self.player.is_impeded_up():\n\t\t\ty = -1\n\t\telif self.player.get_movement_input_down() and not self.player.is_impeded_down():\n\t\t\ty = 1\n\t\tbrick_player = self.player.get_closest_brick()\n\t\tbrick_portal_destination = self.maze.get_brick_just_before_first_wall_in_direction(brick_player, x, y)\n\t\tif brick_portal_destination is None:\n\t\t\tprint(\"Unable to launch portal, because we couldn't find a suitable wall !\")\n\t\t\treturn\n\t\t\n\t\t# Spawn the portal at the player's position\n\t\tplayer_x, player_y = self.player.get_screen_location()\n\t\tportal.set_screen_location(player_x, player_y)\n\t\t\n\t\t# Tell the portal where to go\n\t\tbrick_x, brick_y = brick_portal_destination.get_surface_coords()\n\t\tportal.set_destination(brick_x, brick_y)\n\t\t\n\t\tself.sounds.play_creating_portal()\n\t\t\n\t\tself.portal = portal\n\t\n\tdef draw(self):\n\t\t\n\t\t# Fill the screen with black\n\t\tself.screen.fill(self.settings.screen_background_color)\n\t\t\n\t\t# Draw the welcome board?\n\t\tif self.welcome_board.is_activated():\n\t\t\t\n\t\t\tself.welcome_board.blitme()\n\t\t\n\t\telse:\n\t\t\t# Draw maze\n\t\t\tself.maze.blitme()\n\t\t\n\t\t\t# Draw the board_actors\n\t\t\tself.draw_board_actors()\n\t\t\n\t\t\t# Draw the boards\n\t\t\tself.draw_boards()\n\t\t\t\n\t\t# Draw popups\n\t\tself.popups.blitme()\n\t\t\n\t\t# Flip to screen\n\t\tpygame.display.flip()\n\n\tdef draw_board_actors(self):\n\t\t\n\t\tself.player.blitme()\n\t\t\n\t\tfor ghost in self.ghosts:\n\t\t\tghost.blitme()\n\t\t\n\t\tfor fruit in self.fruits:\n\t\t\tfruit.blitme()\n\t\t\n\t\tif self.portal:\n\t\t\tself.portal.blitme()\n\t\n\tdef draw_boards(self):\n\t\t\n\t\tself.scoreboard.blitme()\n\n\n#\nif __name__ == \"__main__\":\n\t\n\tpacman = Pacman()\n\tpacman.run()\n", "sub_path": "app/Pacman.py", "file_name": "Pacman.py", "file_ext": "py", "file_size_in_byte": 18920, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "lib.Settings.Settings", "line_number": 33, "usage_type": "call"}, {"api_name": "pygame.mixer.pre_init", "line_number": 69, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 69, "usage_type": "attribute"}, {"api_name": "pygame.mixer.init", "line_number": 70, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 70, "usage_type": "attribute"}, {"api_name": "pygame.init", "line_number": 71, "usage_type": "call"}, {"api_name": "pygame.time.Clock", "line_number": 74, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 74, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 88, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 88, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 90, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 90, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 125, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 125, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 131, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 132, "usage_type": "call"}, {"api_name": "pygame.KEYUP", "line_number": 139, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 139, "usage_type": "attribute"}, {"api_name": "pygame.K_q", "line_number": 145, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 146, "usage_type": "call"}, {"api_name": "pygame.K_UP", "line_number": 150, "usage_type": "attribute"}, {"api_name": "pygame.K_DOWN", "line_number": 151, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 152, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 153, "usage_type": "attribute"}, {"api_name": "pygame.K_UP", "line_number": 155, "usage_type": "attribute"}, {"api_name": "pygame.K_DOWN", "line_number": 158, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 161, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 164, "usage_type": "attribute"}, {"api_name": "pygame.K_SPACE", "line_number": 170, "usage_type": "attribute"}, {"api_name": "pygame.K_p", "line_number": 170, "usage_type": "attribute"}, {"api_name": "pygame.KEYUP", "line_number": 171, "usage_type": "attribute"}, {"api_name": "lib.Misc.Sounds.Sounds", "line_number": 384, "usage_type": "call"}, {"api_name": "lib.Maze.Maze.Maze", "line_number": 394, "usage_type": "call"}, {"api_name": "lib.Boards.WelcomeBoard.WelcomeBoard", "line_number": 403, "usage_type": "call"}, {"api_name": "lib.Boards.Scoreboard.Scoreboard", "line_number": 410, "usage_type": "call"}, {"api_name": "lib.Popups.Popups.Popups", "line_number": 420, "usage_type": "call"}, {"api_name": "lib.Actors.Player.Player", "line_number": 426, "usage_type": "call"}, {"api_name": "lib.Actors.Blinky.Blinky", "line_number": 443, "usage_type": "call"}, {"api_name": "lib.Actors.Pinky.Pinky", "line_number": 446, "usage_type": "call"}, {"api_name": "lib.Actors.Clyde.Clyde", "line_number": 449, "usage_type": "call"}, {"api_name": "lib.Actors.Inky.Inky", "line_number": 452, "usage_type": "call"}, {"api_name": "lib.Misc.CheatCodes.CheatCodes", "line_number": 457, "usage_type": "call"}, {"api_name": "math.pow", "line_number": 524, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 729, "usage_type": "call"}, {"api_name": "random.gauss", "line_number": 735, "usage_type": "call"}, {"api_name": "lib.Actors.Fruit.Fruit", "line_number": 744, "usage_type": "call"}, {"api_name": "lib.Maze.Portal.Portal", "line_number": 767, "usage_type": "call"}, {"api_name": "pygame.display.flip", "line_number": 827, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 827, "usage_type": "attribute"}]}
+{"seq_id": "389517262", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Fri Mar 15 17:37:34 2019\r\n\r\n@author: niranjans4\r\n\"\"\"\r\n\r\nprint(\"Hello WOrld\")\r\n\r\nimport matplotlib.pyplot as plt \r\n \r\n# x axis values \r\nx = [1,2,3,10,0,-5,6,11,55,36] \r\n# corresponding y axis values \r\ny = [1,5,3,25,56,0,-11,23,-65,75] \r\n \r\n# plotting the points \r\nplt.plot(x, y) \r\n \r\n# naming the x axis \r\nplt.xlabel('x - axis') \r\n# naming the y axis \r\nplt.ylabel('y - axis') \r\n \r\n# giving a title to my graph \r\nplt.title('My first graph!') \r\n \r\n# function to show the plot \r\nplt.show() \r\n\r\n\r\nleft = [0, 3, 6, 4, 5] \r\n \r\n# heights of bars \r\nheight = [10, 24, 36, 40, 5] \r\n \r\n# labels for bars \r\ntick_label = ['one', 'nir', 'three', 'four', 'five'] \r\n \r\n# plotting a bar chart \r\nplt.bar(left, height, tick_label = tick_label, \r\n width = 0.8, color = ['red', 'green']) \r\n \r\n# naming the x-axis \r\nplt.xlabel('x - axis') \r\n# naming the y-axis \r\nplt.ylabel('y - axis') \r\n# plot title \r\nplt.title('My bar chart!') \r\n \r\n# function to show the plot \r\nplt.show() \r\n\r\n \r\n# defining labels \r\nactivities = ['eat', 'sleep', 'work', 'play'] \r\n \r\n# portion covered by each label \r\nslices = [3, 7, 8, 6] \r\n \r\n# color for each label \r\ncolors = ['r', 'y', 'g', 'b'] \r\n \r\n# plotting the pie chart \r\nplt.pie(slices, labels = activities, colors=colors, \r\n startangle=90, shadow = False, explode = (0.05, 0, 0, 0), \r\n radius = 1, autopct = '%1.1f%%') \r\n \r\n# plotting legend \r\n#plt.legend() \r\n \r\n# showing the plot \r\n#plt.show() ", "sub_path": "Hello_World_graphs.py", "file_name": "Hello_World_graphs.py", "file_ext": "py", "file_size_in_byte": 1485, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "matplotlib.pyplot.plot", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.bar", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.pie", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}]}
+{"seq_id": "421125771", "text": "import csv, os\nimport cv2, h5py\nimport numpy as np\n\n# add path for csv of other data sets when using them \ncsv_location = './input/fixed_handheld.csv'\nimage_folder = './input/images_handheld/'\n\ndef read_image_data_pairs():\n data = []\n csv_file = open(csv_location) # setting up csv reading\n csv_read = csv.reader(csv_file)\n for row in csv_read: # read through the csv\n row_data = row[0].split() # break up info in row\n concise_row = row_data[2].split(\"/\") # isolate image ID and blight bool\n data.append(concise_row) # add to list for ease of manipulation\n return data\n\ndef get_all_Bools(start=None, end=None): # return list of blight bools in order\n to_bool = {\"True\":1, \"False\":0}\n bools = [to_bool[i[0]] for i in read_image_data_pairs()]\n return bools[start:end] # add slice\n\ndef read_all_images(x=None, y=None, start=None, end=None, mute=0): # returns all images, resized as specified\n image_ID_list = [image_folder + str(i[1]) for i in read_image_data_pairs()]\n image_ID_list = image_ID_list[start:end] # !!remove when using full dataset, this just saves time when iterating through and debugging!!\n image_list = []\n for image in image_ID_list:\n if mute == 0:\n print(\"Processing images: %.1f%% complete\" % (100*image_ID_list.index(image)/len(image_ID_list)), end=\"\\r\")\n image = cv2.imread(image)\n if x != None and y != None:\n image = cv2.resize(image, (y, x))\n image_list.append(image)\n if mute == 0:\n print(\"\\rProcessing images: 100% complete \") # possible future error debug, size given is that of first image,\n print(\"Processed {} images, sized {}\".format(len(image_list), image_list[0].shape)) # <----- not necessarily of all images\n return image_list\n\ndef get_features_and_labels(x=6000, y=4000, start=None, end=None): # Combined output of bools and images\n return np.array(read_all_images(x=x, y=y, start=start, end=end)), np.array(get_all_Bools(start=start, end=end))\n\ndef images_to_hdf5(file_name, data_name=\"pics\", start=None, end=None, mute=0): # write image data into an hdf5 file to save read time (<1/6 load time from observations)\n '''Use only when creating a new read file for crop data,\n read from existing file otherwise to save time and memory overhead'''\n filepath = './' + str(file_name) + '.hdf5'\n pictures = read_all_images(x=256, y=256, start=start, end=end, mute=mute)\n if os.path.isfile(filepath) == True:\n os.remove(filepath)\n os.mknod(filepath)\n pictures = np.array(pictures).tolist() # make into list\n with h5py.File(filepath, 'w') as f:\n f.create_dataset(data_name, data=pictures, dtype=\"uint8\")", "sub_path": "imagereader.py", "file_name": "imagereader.py", "file_ext": "py", "file_size_in_byte": 2765, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "csv.reader", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 31, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 49, "usage_type": "call"}, {"api_name": "os.mknod", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 51, "usage_type": "call"}, {"api_name": "h5py.File", "line_number": 52, "usage_type": "call"}]}
+{"seq_id": "540357011", "text": "from multiprocessing import Process,Event\nfrom urllib.request import urlretrieve,Request,urlopen\nfrom time import sleep\nimport os,re,sys,argparse,urllib\n\noptions = { #replace with settings file later\n \"concurrentDownloads\":1,\n \"downloadProgress\":False,\n \"retryDelay\":5\n}\n\ndownloadCount = 0\ndownloaders = []\ndownloaderArgs = []\ndownloadersDone = []\n\n\"\"\"\nDLProgressTracker = []\ndef DLProgress(blocks,blockSize,totalSize,percent=101): #display download progress every 25%\n global DLProgressTracker\n status = int((blocks*blockSize/totalSize)*100)\n if status == 0 and status not in DLProgressTracker:\n DLProgressTracker.append(clock())\n if status%percent == 0:\n if status not in DLProgressTracker:\n DLProgressTracker.append(status)\n #stdout.write(basename(argv[2])+\" progress: \"+str(status)+\"%% (%0.2fMB) of %0.2fMB\"%(blocks*blockSize/1024/1024,totalSize/1024/1024)+\"\\n\")\n #stdout.flush()\n if status == 100:\n stdout.write(basename(argv[2])+\" finished with avg download speed of %0.2fMB/s\"%((totalSize/1024/1024)/(clock()-DLProgressTracker[0]))+\"\\n\")\n stdout.flush()\n #DLProgressTracker = [] #not necessary, since the function is only used once\n\"\"\"\n\ndef downloader(url,target,e=None):\n req = Request(url,headers={\"User-agent\":\"Mozilla/5.0 (Windows NT 6.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/41.0.2228.0 Safari/537.36\"})\n if not os.path.exists(os.path.dirname(target)):\n os.mkdir(os.path.dirname(target))\n file = open(target,\"wb\")\n success = False\n retries = 0\n while not success and retries < 7:\n try:\n file.write(urlopen(req).read())\n success = True\n except urllib.error.HTTPError as e:\n print(\"HTTP Error:\")\n print(e)\n sleep(options[\"retryDelay\"])\n retries += 1\n \n file.close()\n if e != None:\n e.set()\n\n\n#DL([\"http://www.newgrounds.com/audio/download/626468\"],[\"file.mp3\"])\ndef DL(urlList, targetList):\n global downloadCount\n args = [(urlList[x], targetList[x]) for x in range(len(urlList))]\n while args != []:\n shift = 0\n for i in range(downloadCount):\n if downloaderArgs[i+shift][2].is_set():\n del(downloaders[i+shift])\n print(\"Finished downloading \"+os.path.basename(downloaderArgs[i+shift][1]))\n del(downloaderArgs[i+shift])\n downloadCount -= 1\n shift -= 1\n if downloadCount >= options[\"concurrentDownloads\"]:\n sleep(0.1)\n else:\n while downloadCount < options[\"concurrentDownloads\"] and args != []:\n arg = [x for x in args.pop()]\n arg.append(Event())\n arg = tuple(arg)\n downloaders.append(Process(target=downloader,args=arg))\n downloaders[-1].start()\n downloaderArgs.append(arg)\n print(\"Started downloading \"+os.path.basename(arg[1]))\n downloadCount += 1\n \n\ndef getFiles(username,folder=\".\\\\\",dlType=\"audio\"):\n matches = []\n matches_genres = []\n page_i = 1\n more = True\n while more:\n url = \"https://\"+username+\".newgrounds.com/\"+dlType+\"/page/\"+str(page_i)\n print(\"Fetching '{}'...\".format(url))\n req = Request(url, headers={\"User-agent\":\"Mozilla/5.0\", \"x-requested-with\": \"XMLHttpRequest\"})\n page = str(urlopen(req).read(),encoding=\"UTF-8\")\n matches += re.findall('', page)\n matches_genres += re.findall('detail-genre.*?(?:\\s)+([ \\w]+).*?div>', page)\n more = re.search(\"\\\"more\\\"\\:null\", page) is None\n page_i += 1\n print(\"Found {} songs.\".format(str(len(matches))))\n urls = [\"https://www.newgrounds.com/audio/download/{}/\".format(matches[i][0]) for i in range(len(matches))]\n files = [folder+username+\"\\\\\"+matches_genres[i].replace(\"Song\",\"\").strip()+\"\\\\\"+matches[i][1]+\".mp3\" for i in range(len(matches))]\n if not os.path.exists(folder+username):\n os.mkdir(folder+username)\n \n DL(urls,files)\n\n\ndef getArgParser():\n p = argparse.ArgumentParser(description=\"Newgrounds music downloader\")\n p.add_argument(\"-n\", type=int, default=4, dest=\"threads\", help=\"Sets the number of files to download concurrently. Should speed up downloads on fast connections.\")\n p.add_argument(\"-t\", type=float, default=5, dest=\"delay\", help=\"Delay (in seconds) to wait before retrying a failed download.\")\n p.add_argument(\"username\", nargs=\"+\", help=\"List of usernames whose songs you wish to download.\")\n return p\n\n\nif __name__ == '__main__':\n args = getArgParser().parse_args()\n options[\"concurrentDownloads\"] = args.threads\n options[\"retryDelay\"] = args.delay\n for user in args.username:\n print(\"\\n\\nParsing {}...\".format(user))\n getFiles(user)\n \n \n", "sub_path": "downloader.py", "file_name": "downloader.py", "file_ext": "py", "file_size_in_byte": 4957, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "urllib.request.Request", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 37, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "urllib.request.urlopen", "line_number": 44, "usage_type": "call"}, {"api_name": "urllib.error", "line_number": 46, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 71, "usage_type": "call"}, {"api_name": "multiprocessing.Event", "line_number": 75, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 80, "usage_type": "call"}, {"api_name": "os.path", "line_number": 80, "usage_type": "attribute"}, {"api_name": "urllib.request.Request", "line_number": 92, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 93, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 94, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 95, "usage_type": "call"}, {"api_name": "re.search", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path", "line_number": 101, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 102, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 108, "usage_type": "call"}]}
+{"seq_id": "225226051", "text": "from glob import glob\nfrom scipy.ndimage import imread\nimport scipy.misc\nfrom argparse import ArgumentParser\nimport os\nimport time\nfrom tqdm import tqdm\n\ndef main(in_path, out_path, folder_name_base, split_number):\n frames = sorted(glob(os.path.join(in_path, \"*\")))\n folder = []\n for i in range(split_number):\n folder_name = os.path.join(out_path, folder_name_base) + \"_\" + str(i)\n folder.append(folder_name)\n if not os.path.exists(folder_name):\n os.makedirs(folder_name)\n\n start_time = time.time()\n for index, frame in tqdm(enumerate(frames)):\n img = imread(frame)\n frame_name = frame.split('/',-1)[-1]\n scipy.misc.toimage(img).save(os.path.join(folder[index % split_number], frame_name))\n #if (time.time() - start_time) > 1:\n #print(str(index)+\"|\"+str(len(frames))+\": \"+frame+\" to \"+folder[index % split_number])\n #start_time = time.time()\n\nif __name__ == \"__main__\":\n parser = ArgumentParser()\n parser.add_argument(\"--inpath\", type=str, dest=\"in_path\",\n required=True, help=\"Path to directory of the images\")\n parser.add_argument(\"--outpath\", type=str, dest=\"out_path\",\n required=True, help=\"Path to directory where images should be saved (base directory)\")\n parser.add_argument(\"--basename\", type=str, dest=\"folder_name_base\",\n required=True, help=\"Base name of the folders in which the splitted images should be saved\")\n parser.add_argument(\"--splits\", type=int, dest=\"split_number\",\n required=True, help=\"In how many groups the images should be split up \")\n\n args = parser.parse_args()\n main(**vars(args))\n", "sub_path": "storage/preprocessing/split_frames.py", "file_name": "split_frames.py", "file_ext": "py", "file_size_in_byte": 1722, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "glob.glob", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 16, "usage_type": "call"}, {"api_name": "time.time", "line_number": 18, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 19, "usage_type": "call"}, {"api_name": "scipy.ndimage.imread", "line_number": 20, "usage_type": "call"}, {"api_name": "scipy.ndimage.misc.toimage", "line_number": 22, "usage_type": "call"}, {"api_name": "scipy.ndimage.misc", "line_number": 22, "usage_type": "attribute"}, {"api_name": "scipy.ndimage", "line_number": 22, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 28, "usage_type": "call"}]}
+{"seq_id": "568813910", "text": "import torch\nfrom torch import nn\nfrom torch.nn import functional as F\nfrom layer import GraphConvolution\n\nfrom config import args\n\nclass GCN(nn.Module):\n\n\n def __init__(self, input_dim, output_dim, num_features_nonzero):\n super(GCN, self).__init__()\n\n self.input_dim = input_dim # 1433\n self.output_dim = output_dim\n\n print('input dim:', input_dim)\n print('output dim:', output_dim)\n print('num_features_nonzero:', num_features_nonzero)\n\n\n self.layers1 = nn.Sequential(GraphConvolution(self.input_dim, args.hidden, num_features_nonzero,\n activation=F.relu,\n dropout=args.dropout,\n is_sparse_inputs=True)\n ).to('cuda:0')\n self.layers2 = nn.Sequential(GraphConvolution(args.hidden, output_dim, num_features_nonzero,\n activation=F.relu,\n dropout=args.dropout,\n is_sparse_inputs=False)\n ).to('cuda:1')\n\n def forward(self, inputs):\n # x, support = inputs\n # x, support = self.layers1((x, support))\n # x = self.layers2((x, support))\n\n x, support = self.layers1(inputs)\n x = self.layers2((x.to('cuda:1'), support.to('cuda:1')))\n\n return x\n\n def l2_loss(self):\n\n layer = self.layers1.children()\n loss = None\n\n for l in layer:\n for p in l.parameters():\n if loss is None:\n loss = p.pow(2).sum()\n else:\n loss += p.pow(2).sum()\n loss = loss.to('cuda:1')\n # print(loss.device)\n layer = self.layers2.children()\n for l in layer:\n for p in l.parameters():\n if loss is None:\n loss = p.pow(2).sum()\n else:\n loss += p.pow(2).sum()\n \n return loss\n", "sub_path": "pipe_model.py", "file_name": "pipe_model.py", "file_ext": "py", "file_size_in_byte": 2151, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "torch.nn.Module", "line_number": 8, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 8, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "name"}, {"api_name": "layer.GraphConvolution", "line_number": 22, "usage_type": "call"}, {"api_name": "config.args.hidden", "line_number": 22, "usage_type": "attribute"}, {"api_name": "config.args", "line_number": 22, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 23, "usage_type": "attribute"}, {"api_name": "torch.nn.functional", "line_number": 23, "usage_type": "name"}, {"api_name": "config.args.dropout", "line_number": 24, "usage_type": "attribute"}, {"api_name": "config.args", "line_number": 24, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}, {"api_name": "layer.GraphConvolution", "line_number": 27, "usage_type": "call"}, {"api_name": "config.args.hidden", "line_number": 27, "usage_type": "attribute"}, {"api_name": "config.args", "line_number": 27, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 28, "usage_type": "attribute"}, {"api_name": "torch.nn.functional", "line_number": 28, "usage_type": "name"}, {"api_name": "config.args.dropout", "line_number": 29, "usage_type": "attribute"}, {"api_name": "config.args", "line_number": 29, "usage_type": "name"}]}
+{"seq_id": "28658415", "text": "from functools import partial\nfrom copy import deepcopy\nfrom multiprocessing import Process, Queue\nfrom task import ProcessedTask, SchedulingResult\nfrom consts import TACT_SIZE\nfrom example import N\n\ntasks, activeTasks = list(), list()\n\ndef taskID(task):\n return f'{task.arrivalTime}{task.period}{task.deadline}'\n\ndef areTasksEqual(taskA, taskB):\n return taskID(taskA) == taskID(taskB)\n\ndef checkForNewTasks(time):\n global tasks, activeTasks\n futureTasks = lambda x: x.arrivalTime > time\n arrivedTasks = lambda x: x.arrivalTime <= time\n arrived = list(filter(arrivedTasks, tasks))\n if not arrived: return False\n activeTasks += arrived\n tasks = list(filter(futureTasks, tasks))\n return True\n\ndef waitForTasks(curTime, interval, clb):\n global activeTasks \n maxWaitTime = curTime + interval\n if not checkForNewTasks(maxWaitTime): return None\n clb(activeTasks)\n return activeTasks[0]\n\ndef checkMissedTasks(tasks, resTasks):\n diff = list()\n for task in tasks:\n processed = False\n for taskRes in resTasks:\n if task.arrivalTime == taskRes.arrival \\\n and task.deadline == taskRes.deadline:\n processed = True\n if not processed: diff.append(task)\n return diff\n\ndef BaseDynamicScheduler(queue, estimatePriority, allTasks):\n global tasks, activeTasks\n tasks = deepcopy(allTasks)\n tasks.sort(key=lambda x: x.arrivalTime)\n timestamps = dict()\n curTime, tactCount, idle = 0, 0, 0\n for task in tasks: timestamps[taskID(task)] = list()\n resTasks = list()\n while activeTasks or tasks:\n if tasks: checkForNewTasks(curTime)\n activeTasks = list(filter(lambda x: x.deadline > curTime, activeTasks))\n missed = list(filter(lambda x: x.deadline <= curTime, activeTasks))\n estimatePriority(activeTasks)\n for missedTask in missed:\n task = ProcessedTask(\n missedTask.arrivalTime,\n missedTask.deadline,\n [curTime, curTime],\n True\n )\n resTasks.append(task)\n \n remainedTime = tactCount or TACT_SIZE\n task = None\n if activeTasks:\n task = activeTasks[0]\n else:\n task = waitForTasks(curTime, remainedTime, estimatePriority)\n if not task:\n idle += remainedTime\n curTime += remainedTime\n if tactCount: tactCount = 0\n continue\n\n if task.arrivalTime > curTime:\n diff = task.arrivalTime - curTime\n idle += diff\n curTime = task.arrivalTime\n tactCount -= diff\n \n timestamps[taskID(task)].append(curTime)\n fitsIntoTact = task.wcet < remainedTime\n elapsedInTact = task.wcet if fitsIntoTact else remainedTime\n newTask = waitForTasks(curTime, elapsedInTact, estimatePriority)\n if newTask and \\\n not areTasksEqual(task, newTask) and \\\n not task.protected:\n task.wcet -= newTask.arrivalTime - curTime\n curTime = newTask.arrivalTime\n timestamps[taskID(task)].append(curTime)\n task = newTask\n timestamps[taskID(task)].append(curTime)\n fitsIntoTact = task.wcet < remainedTime\n elapsedInTact = task.wcet if fitsIntoTact else remainedTime\n curTime += elapsedInTact\n timestamps[taskID(task)].append(curTime)\n if not fitsIntoTact:\n task.wcet -= remainedTime\n tactCount = 0\n else:\n missed = task.deadline < curTime\n task = ProcessedTask(\n task.arrivalTime,\n task.deadline,\n timestamps[taskID(task)],\n missed\n )\n resTasks.append(task)\n activeTasks.pop(0)\n tactCount = remainedTime - elapsedInTact\n\n missed = checkMissedTasks(allTasks, resTasks)\n if missed:\n for task in missed: \n task = ProcessedTask(\n task.arrivalTime,\n task.deadline,\n [curTime, curTime],\n True\n )\n resTasks.append(task)\n queue.put(SchedulingResult(resTasks, idle, curTime))\n\ndef multiprocScheduler(estimatePriority, tasks):\n procs = list()\n queue = Queue()\n splitTasks = [tasks[i::N] for i in range(N)]\n for taskSet in splitTasks:\n procs.append(Process(\n target=BaseDynamicScheduler,\n args=(queue, estimatePriority, taskSet,)\n ))\n for proc in procs: proc.start()\n for proc in procs: proc.join()\n return [queue.get() for _ in splitTasks]\n\ndef estimatePriorityEDF(tasks):\n return tasks.sort(key=lambda x: (x.deadline, x.wcet)) \n\ndef estimatePriorityRM(tasks):\n return tasks.sort(key=lambda x: (x.period, x.wcet, x.deadline)) \n\nEDF = partial(multiprocScheduler, estimatePriorityEDF)\nRM = partial(multiprocScheduler, estimatePriorityRM)\n", "sub_path": "RTS_RGR/dynamicShedulers.py", "file_name": "dynamicShedulers.py", "file_ext": "py", "file_size_in_byte": 4503, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "task.arrivalTime", "line_number": 11, "usage_type": "attribute"}, {"api_name": "task.period", "line_number": 11, "usage_type": "attribute"}, {"api_name": "task.deadline", "line_number": 11, "usage_type": "attribute"}, {"api_name": "task.arrivalTime", "line_number": 38, "usage_type": "attribute"}, {"api_name": "task.deadline", "line_number": 39, "usage_type": "attribute"}, {"api_name": "copy.deepcopy", "line_number": 46, "usage_type": "call"}, {"api_name": "task.ProcessedTask", "line_number": 58, "usage_type": "call"}, {"api_name": "consts.TACT_SIZE", "line_number": 66, "usage_type": "name"}, {"api_name": "task.arrivalTime", "line_number": 78, "usage_type": "attribute"}, {"api_name": "task.arrivalTime", "line_number": 79, "usage_type": "attribute"}, {"api_name": "task.arrivalTime", "line_number": 81, "usage_type": "attribute"}, {"api_name": "task.wcet", "line_number": 85, "usage_type": "attribute"}, {"api_name": "task.wcet", "line_number": 86, "usage_type": "attribute"}, {"api_name": "task.protected", "line_number": 90, "usage_type": "attribute"}, {"api_name": "task.wcet", "line_number": 91, "usage_type": "attribute"}, {"api_name": "task.wcet", "line_number": 96, "usage_type": "attribute"}, {"api_name": "task.wcet", "line_number": 97, "usage_type": "attribute"}, {"api_name": "task.wcet", "line_number": 101, "usage_type": "attribute"}, {"api_name": "task.deadline", "line_number": 104, "usage_type": "attribute"}, {"api_name": "task.ProcessedTask", "line_number": 105, "usage_type": "call"}, {"api_name": "task.arrivalTime", "line_number": 106, "usage_type": "attribute"}, {"api_name": "task.deadline", "line_number": 107, "usage_type": "attribute"}, {"api_name": "task.ProcessedTask", "line_number": 118, "usage_type": "call"}, {"api_name": "task.arrivalTime", "line_number": 119, "usage_type": "attribute"}, {"api_name": "task.deadline", "line_number": 120, "usage_type": "attribute"}, {"api_name": "task.SchedulingResult", "line_number": 125, "usage_type": "call"}, {"api_name": "multiprocessing.Queue", "line_number": 129, "usage_type": "call"}, {"api_name": "example.N", "line_number": 130, "usage_type": "name"}, {"api_name": "multiprocessing.Process", "line_number": 132, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 146, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 147, "usage_type": "call"}]}
+{"seq_id": "345930785", "text": "import west.data_map\nimport west.data_management\nimport west.population\nimport numpy\nimport scipy\n\ndef evaluate_hex_capacity(p = 250, min_radius = 0.05, max_radius = 100, test_repack = False):\n\n areaarray = numpy.logspace(numpy.log(numpy.pi * 0.05 * 0.05)/numpy.log(10), numpy.log(numpy.pi * 100 * 100)/numpy.log(10), 65)\n\n def divide_function(this_value, other_value):\n if this_value == 0 and other_value == 0:\n return 0\n return float(this_value/other_value)\n\n def calculate_tower_area(latitude, longitude, latitude_index, longitude_index, current_value):\n return p / current_value\n\n def find_area_less_than_and_greater_than_towerarea(tower_area):\n for i in range(len(areaarray)):\n if areaarray[i] <= tower_area and areaarray[i + 1] > tower_area:\n if i == len(areaarray) - 1:\n return areaarray[i], None\n return areaarray[i], areaarray[i + 1]\n\n def interpolate(low_value, high_value, low_area, high_area, actual_area):\n low_value_log = numpy.log(low_value)/numpy.log(10)\n high_value_log = numpy.log(high_value)/numpy.log(10)\n actual_value_log = (high_value_log - low_value_log)/(high_area - low_area) * (actual_area - low_area)\n actual_value = scipy.power(10, actual_value_log)\n return actual_value\n\n\n with open (\"noisestrength_hex_data.pkl\", 'r') as f:\n interferencestrengthsdict = pickle.load(f)\n\n with open(\"signalstrength_hex_data.pkl\", 'r') as f:\n signalstrengthsdict = pickle.load(f)\n\n datamap_spec = west.data_management.SpecificationDataMap(west.data_map.DataMap2DContinentalUnitedStates, 400, 600)\n is_in_region_map_spec = west.data_management.SpecificationRegionMap(west.boundary.BoundaryContinentalUnitedStates, datamap_spec)\n is_in_region_map = is_in_region_map_spec.fetch_data()\n population_map_spec = west.data_management.SpecificationPopulationMap(is_in_region_map_spec, west.population.PopulationData)\n population_map = population_map_spec.fetch_data()\n region_area_map_spec = west.data_management.SpecificationRegionAreaMap(datamap_spec)\n region_area_map = region_area_map_spec.fetch_data()\n population_density_map = population_map.combine_datamaps_with_function(region_area_map, divide_function)\n areaoftower_map = west.data_map.DataMap2DContinentalUnitedStates.get_copy_of(population_density_map)\n\n #Note: Keep in mind what will happen if the population density is 0.\n areaoftower_map.update_all_values_via_function(calculate_tower_area)\n min_area = numpy.pi * min_radius * min_radius\n max_area = numpy.pi * max_radius * max_radius\n\n def clip_map(latitude, longitude, latitude_index, longitude_index, current_value):\n if current_value > max_area:\n return max_area\n if current_value < min_area:\n return min_area\n return current_value\n\n\n areaoftower_map.update_all_values_via_function(clip_map)\n\n channel_list = range(2, 52)\n channel_list.remove(37)\n\n if test_repack:\n exclusionfilename = \"15VHFFreeUSMinimumStationstoRemove0_allchannels.pcl\"\n noisefilename = \"noise_map_test_repack_withsubmaps.pcl\"\n else:\n exclusionfilename = \"original_map_fcccontours_withplmrs_allchannels.pcl\"\n noisefilename = \"noise_map_test_original_allocation_withsubmaps.pcl\"\n\n\n exclusion_map_3D = west.data_map.DataMap3D.from_pickle(exclusionfilename)\n noise_map_3D = west.data_map.DataMap3D.from_pickle(os.path.join(\"data\", \"Noise Maps\", noisefilename))\n\n\n fair_capacity_hex_map_3D = west.data_map.DataMap3D.from_DataMap2D(is_in_region_map, channel_list)\n avg_capacity_hex_map_3D = west.data_map.DataMap3D.from_DataMap2D(is_in_region_map, channel_list)\n min_capacity_hex_map_3D = west.data_map.DataMap3D.from_DataMap2D(is_in_region_map, channel_list)\n\n for c in channel_list:\n faircapacity_channelmap = fair_capacity_hex_map_3D.get_layer(c)\n avgcapacity_channelmap = avg_capacity_hex_map_3D.get_layer(c)\n mincapacity_channelmap = min_capacity_hex_map_3D.get_layer(c)\n for i in range(400):\n for j in range(600):\n if is_in_region_map.get_value_by_index(i, j) == 0:\n continue\n tower_area = areaoftower_map.get_value_by_index(i, j)\n low_area, high_area = find_area_less_than_and_greater_than_towerarea(tower_area)\n if high_area == None:\n high_area = areaarray[len(areaarray) - 1]\n low_signal = signalstrengthsdict[c][low_area]\n high_signal = signalstrengthsdict[c][high_area]\n low_interference = interferencestrengthsdict[c][low_area]\n high_interference = interferencestrengthsdict[c][high_area]\n\n signal = interpolate(low_signal, high_signal, low_area, high_area, tower_area)\n interference = interpolate(low_interference, high_interference, low_area, high_area, tower_area)\n total_noise = interference + noise_map_3D.get_layer(c).get_value_by_index(i, j)\n potential_capacity = exclusion_map_3D.get_layer(c).get_value_by_index(i, j) * 6e6 * numpy.log(1 + signal/total_noise)\n\n #Fair capacity: If each person is allowed to use the same rate, what would this rate be?\n fair_capacity_per_person = 1/sum(1/potential_capacity)\n\n total_fair_capacity = fair_capacity_per_person * p\n faircapacity_channelmap.set_value_by_index(i, j, total_fair_capacity)\n\n #Avg capacity\n avg_capacity = numpy.mean(potential_capacity)\n avgcapacity_channelmap.set_value_by_index(i, j, avg_capacity)\n\n #Min capacity\n min_capacity = min(potential_capacity)\n mincapacity_channelmap.set_value_by_index(i, j, min_capacity)\n\n fair_capacity_hex_map = fair_capacity_hex_map_3D.sum_all_layers()\n avg_capacity_hex_map = avg_capacity_hex_map_3D.sum_all_layers()\n min_capacity_hex_map = min_capacity_hex_map_3D.sum_all_layers()\n\n return fair_capacity_hex_map, avg_capacity_hex_map, min_capacity_hex_map\n\n\nfaircapacityhexmap, avgcapacityhexmap, mincapacityhexmap = evaluate_hex_capacity()\n\nfaircapacityhexmap.to_pickle(os.path.join(\"data\", \"Data Rate Maps\", \"hex_fair_capacity_original_allocation.pcl\"))\navgcapacityhexmap.to_pickle(os.path.join(\"data\", \"Data Rate Maps\", \"hex_avg_capacity_original_allocation.pcl\"))\nmincapacityhexmap.to_pickle(os.path.join(\"data\", \"Data Rate Maps\", \"hex_min_capacity_original_allocation.pcl\"))\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "evaluate_hex_capacity.py", "file_name": "evaluate_hex_capacity.py", "file_ext": "py", "file_size_in_byte": 6651, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "numpy.logspace", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 9, "usage_type": "attribute"}, {"api_name": "numpy.log", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 28, "usage_type": "call"}, {"api_name": "scipy.power", "line_number": 30, "usage_type": "call"}, {"api_name": "west.data_map.data_management.SpecificationDataMap", "line_number": 40, "usage_type": "call"}, {"api_name": "west.data_map.data_management", "line_number": 40, "usage_type": "attribute"}, {"api_name": "west.data_map", "line_number": 40, "usage_type": "name"}, {"api_name": "west.data_map.data_map", "line_number": 40, "usage_type": "attribute"}, {"api_name": "west.data_map.data_management.SpecificationRegionMap", "line_number": 41, "usage_type": "call"}, {"api_name": "west.data_map.data_management", "line_number": 41, "usage_type": "attribute"}, {"api_name": "west.data_map", "line_number": 41, "usage_type": "name"}, {"api_name": "west.data_map.boundary", "line_number": 41, "usage_type": "attribute"}, {"api_name": "west.data_map.data_management.SpecificationPopulationMap", "line_number": 43, "usage_type": "call"}, {"api_name": "west.data_map.data_management", "line_number": 43, "usage_type": "attribute"}, {"api_name": "west.data_map", "line_number": 43, "usage_type": "name"}, {"api_name": "west.data_map.population", "line_number": 43, "usage_type": "attribute"}, {"api_name": "west.data_map.data_management.SpecificationRegionAreaMap", "line_number": 45, "usage_type": "call"}, {"api_name": "west.data_map.data_management", "line_number": 45, "usage_type": "attribute"}, {"api_name": "west.data_map", "line_number": 45, "usage_type": "name"}, {"api_name": "west.data_map.data_map.DataMap2DContinentalUnitedStates.get_copy_of", "line_number": 48, "usage_type": "call"}, {"api_name": "west.data_map.data_map", "line_number": 48, "usage_type": "attribute"}, {"api_name": "west.data_map", "line_number": 48, "usage_type": "name"}, {"api_name": "numpy.pi", "line_number": 52, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 53, "usage_type": "attribute"}, {"api_name": "west.data_map.data_map.DataMap3D.from_pickle", "line_number": 76, "usage_type": "call"}, {"api_name": "west.data_map.data_map", "line_number": 76, "usage_type": "attribute"}, {"api_name": "west.data_map", "line_number": 76, "usage_type": "name"}, {"api_name": "west.data_map.data_map.DataMap3D.from_pickle", "line_number": 77, "usage_type": "call"}, {"api_name": "west.data_map.data_map", "line_number": 77, "usage_type": "attribute"}, {"api_name": "west.data_map", "line_number": 77, "usage_type": "name"}, {"api_name": "west.data_map.data_map.DataMap3D.from_DataMap2D", "line_number": 80, "usage_type": "call"}, {"api_name": "west.data_map.data_map", "line_number": 80, "usage_type": "attribute"}, {"api_name": "west.data_map", "line_number": 80, "usage_type": "name"}, {"api_name": "west.data_map.data_map.DataMap3D.from_DataMap2D", "line_number": 81, "usage_type": "call"}, {"api_name": "west.data_map.data_map", "line_number": 81, "usage_type": "attribute"}, {"api_name": "west.data_map", "line_number": 81, "usage_type": "name"}, {"api_name": "west.data_map.data_map.DataMap3D.from_DataMap2D", "line_number": 82, "usage_type": "call"}, {"api_name": "west.data_map.data_map", "line_number": 82, "usage_type": "attribute"}, {"api_name": "west.data_map", "line_number": 82, "usage_type": "name"}, {"api_name": "numpy.log", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 113, "usage_type": "call"}]}
+{"seq_id": "320613298", "text": "\"\"\"\nStartup Code\n\n@author: David Schote\n@reworked by: Sula Mueller\n@contact: david.schote@ovgu.de\n@version: 2.0.2\n@change: 02/11/2020\n\"\"\"\n\n# system includes\nimport sys\nfrom PyQt5.QtWidgets import QApplication\n\n# project includes\nfrom mainviewcontroller import MainViewController\n\nVERSION = \"2.0.2\"\nAUTHOR = \"David Schote, Sula Mueller\"\n\nif __name__ == '__main__':\n print(\"Graphical User Interface for Magnetic Resonance Imaging {} by {}\".format(VERSION, AUTHOR))\n app = QApplication(sys.argv)\n gui = MainViewController()\n gui.show()\n gui.connectiondialog.show()\n sys.exit(app.exec_())\n", "sub_path": "GOmri.py", "file_name": "GOmri.py", "file_ext": "py", "file_size_in_byte": 613, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 23, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 23, "usage_type": "attribute"}, {"api_name": "mainviewcontroller.MainViewController", "line_number": 24, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 27, "usage_type": "call"}]}
+{"seq_id": "653389361", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Mar 5 13:45:45 2021\n\n@author: vishakha\n\"\"\"\nimport numpy as np\nimport sys\nimport matplotlib.pyplot as plt\n\nsys.path.append(\".\")\nfrom Random import Random\n\nif __name__ == \"__main__\":\n\n\t#set default number of samples\n\tNsample = 1000\n\n\t# read the user-provided seed from the command line (if there)\n\tif '-Nsample' in sys.argv:\n\t\tp = sys.argv.index('-Nsample')\n\t\tNsample = int(sys.argv[p+1])\n\tif '-h' in sys.argv or '--help' in sys.argv:\n\t\tprint (\"Usage: %s -Nsample [number]\" % sys.argv[0])\n\t\tprint\n\t\tsys.exit(1) \n\n\tnAccept = 0\n\tnTotal = 0\n\n\t# accepted values\n\tXaccept = []\n\tYaccept = []\n\n\t# reject values\n\tXreject = []\n\tYreject = []\n\n\t# sample number\n\tisample = []\n\t# calculated values of Pi (per sample)\n\tcalcPi = []\n\n\trandom = Random()\n\n\tidraw = max(1,int(Nsample)/100000)\n\tfor i in range(0,Nsample):\n\t\tX = random.rand()\n\t\tY = random.rand()\n\n\t\tnTotal += 1\n\t\tif(X*X + Y*Y <= 1): #accept if inside\n\t\t\tnAccept += 1\n\t\t\tif(i % idraw == 0):\n\t\t\t\tXaccept.append(X)\n\t\t\t\tYaccept.append(Y)\n\t\telse: # reject if outside\n\t\t\tif(i % idraw == 0):\n\t\t\t\tXreject.append(X)\n\t\t\t\tYreject.append(Y)\n\t\tif(i % idraw == 0):\n\t\t\tisample.append(nTotal)\n\t\t\tcalcPi.append(4*nAccept/nTotal)\n\n\n\n\t#plot calculated pi vs sample number\n\tfig1 = plt.figure()\n\tplt.plot(isample,calcPi)\n\tplt.ylabel(r'Approximate $\\pi$')\n\tplt.xlabel(\"Sample number\")\n\tplt.xlim(0,isample[len(isample)-1])\n\tax = plt.gca()\n\tax.axhline(y=np.arccos(-1),color='green',label=r'true $\\pi$')\n\tplt.title(r'Approximation of $\\pi$ as a function of number of samples')\n\tplt.legend()\n\n\tfig1.savefig(\"calculatedPiPy.pdf\")\n\n\n\t#plot accept/reject points\n\tfig2 = plt.figure()\n\tplt.plot(Xaccept,Yaccept,marker='o',linestyle='',color='green',label='accept')\n\tplt.plot(Xreject,Yreject,marker='o',linestyle='',color='red',label='reject')\n\tplt.ylabel(\"Y\")\n\tplt.xlabel(\"X\")\n\tplt.legend()\n\n\n\tx_circle = np.arange(min(min(Xaccept),min(Xreject)),max(max(Xaccept),max(Xreject)),0.001)\n\ty_circle = [np.sqrt(1-i*i) for i in x_circle]\n\tplt.plot(x_circle,y_circle,color='blue',label=r'$x^2 + y^2 = 1$')\n\tplt.legend()\n\tplt.title('Sampled points')\n\tfig2.savefig(\"circleQuadPy.pdf\")", "sub_path": "untitled3.py", "file_name": "untitled3.py", "file_ext": "py", "file_size_in_byte": 2154, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sys.path.append", "line_number": 12, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 21, "usage_type": "attribute"}, {"api_name": "sys.argv.index", "line_number": 22, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 22, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 23, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 24, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 25, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 27, "usage_type": "call"}, {"api_name": "Random.Random", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "numpy.arccos", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}]}
+{"seq_id": "392369686", "text": "from django.shortcuts import render,redirect\nfrom.models import Product,Cart,CartItem\nfrom Orders.models import Order\nfrom accounts.forms import LoginForm,GuestForm\nfrom Billing.models import BillingProfile\nfrom accounts.models import GuestEmail\nfrom Payments.views import charge\n\n# Create your views here.\ndef cart_home(request):\n cart_obj, new_obj = Cart.objects.new_or_get(request)\n products=cart_obj.cartitem_set.all()\n total=0\n for x in products:\n line_total = float(x.products.price) * x.quantity\n total += line_total\n print(total)\n print(cart_obj.total)\n cart_obj.total=total\n cart_obj.save()\n\n return render(request,'cart_home.html',{'cart':cart_obj})\ndef cart_update(request,slug):\n try:\n qty=request.GET.get('qty')\n update_qty = True\n except:\n qty=None\n update_qty=False\n try:\n product_obj=Product.objects.get(slug=slug)\n except Product.DoesNotExist:\n print('product out of stock')\n return redirect('carts:cart_view')\n cart_obj, new_obj = Cart.objects.new_or_get(request)\n cart_item, created = CartItem.objects.get_or_create(cart=cart_obj,products=product_obj )\n if created:\n print('created')\n if update_qty and qty :\n if int(qty)==0 :\n cart_item.delete()\n else:\n cart_item.quantity=qty\n cart_item.save()\n else:\n pass\n # if product_obj in cart_obj.items.all():\n # cart_obj.items.remove(cart_item)\n # else:\n # cart_obj.items.add(cart_item)\n products = cart_obj.cartitem_set.all()\n request.session['cart_items'] = cart_obj.cartitem_set.count()\n\n\n return redirect('carts:cart_view')\n\ndef checkout_home(request):\n cart_obj,cart_created=Cart.objects.new_or_get(request)\n order_obj=None\n has_card=False\n if cart_created or cart_obj.cartitem_set.count()==0:\n redirect('carts:cart_view')\n login_form=LoginForm()\n guest_form=GuestForm()\n # billing model manager\n billing_profile, billing_profile_created = BillingProfile.objects.new_or_get(request)\n if billing_profile is not None:\n # order model manager\n order_obj, order_obj_created = Order.objects.new_or_get(billing_profile, cart_obj)\n has_card=billing_profile.has_card\n # finalize checkout\n if request.method == \"POST\":\n \"check the order is done\"\n is_done = order_obj.check_done()\n\n if is_done:\n order_obj.mark_paid()\n request.session['cart_items'] = 0\n del request.session['cart_id']\n\n\n return redirect('carts:success')\n context={\n 'billing_profile':billing_profile,\n 'object': order_obj,\n 'login_form':login_form,\n 'guest_form':guest_form,\n \"has_card\":has_card\n }\n return render(request,'checkout.html',context)\n\ndef checkout_done_view(request):\n return render(request,\"checkout_done.html\")", "sub_path": "carts/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2950, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "models.Cart.objects.new_or_get", "line_number": 11, "usage_type": "call"}, {"api_name": "models.Cart.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "models.Cart", "line_number": 11, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 22, "usage_type": "call"}, {"api_name": "models.Product.objects.get", "line_number": 31, "usage_type": "call"}, {"api_name": "models.Product.objects", "line_number": 31, "usage_type": "attribute"}, {"api_name": "models.Product", "line_number": 31, "usage_type": "name"}, {"api_name": "models.Product.DoesNotExist", "line_number": 32, "usage_type": "attribute"}, {"api_name": "models.Product", "line_number": 32, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 34, "usage_type": "call"}, {"api_name": "models.Cart.objects.new_or_get", "line_number": 35, "usage_type": "call"}, {"api_name": "models.Cart.objects", "line_number": 35, "usage_type": "attribute"}, {"api_name": "models.Cart", "line_number": 35, "usage_type": "name"}, {"api_name": "models.CartItem.objects.get_or_create", "line_number": 36, "usage_type": "call"}, {"api_name": "models.CartItem.objects", "line_number": 36, "usage_type": "attribute"}, {"api_name": "models.CartItem", "line_number": 36, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 55, "usage_type": "call"}, {"api_name": "models.Cart.objects.new_or_get", "line_number": 58, "usage_type": "call"}, {"api_name": "models.Cart.objects", "line_number": 58, "usage_type": "attribute"}, {"api_name": "models.Cart", "line_number": 58, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 62, "usage_type": "call"}, {"api_name": "accounts.forms.LoginForm", "line_number": 63, "usage_type": "call"}, {"api_name": "accounts.forms.GuestForm", "line_number": 64, "usage_type": "call"}, {"api_name": "Billing.models.BillingProfile.objects.new_or_get", "line_number": 66, "usage_type": "call"}, {"api_name": "Billing.models.BillingProfile.objects", "line_number": 66, "usage_type": "attribute"}, {"api_name": "Billing.models.BillingProfile", "line_number": 66, "usage_type": "name"}, {"api_name": "Orders.models.Order.objects.new_or_get", "line_number": 69, "usage_type": "call"}, {"api_name": "Orders.models.Order.objects", "line_number": 69, "usage_type": "attribute"}, {"api_name": "Orders.models.Order", "line_number": 69, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 82, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 90, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 93, "usage_type": "call"}]}
+{"seq_id": "19349836", "text": "import numpy as np\nimport matplotlib.pyplot as plt\n\ndef RK4(f, r0, tf, dt):\n \"\"\"Fourth-order Runge-Kutta integrator.\n\n :param f: Function to be integrated\n :param r0: Initial conditions\n :param tf: Integration duration\n :param dt: Timestep size\n :returns: time and trajectory vectors\n\n \"\"\"\n\n # generate an array of time steps\n ts = np.arange(0, tf, dt)\n # create an array to hold system state at each timestep\n traj = np.zeros((ts.shape[0], len(r0)))\n traj[0, :] = np.array(r0)\n # calculate system state at each time step, save it in the array\n for i in range(0, ts.shape[0]-1):\n t = ts[i]\n r = traj[i, :]\n\n k1 = dt * f(r, t)\n k2 = dt * f(r + k1/2, t + dt/2)\n k3 = dt * f(r + k2/2, t + dt/2)\n k4 = dt * f(r + k3, t + dt)\n K = (1.0/6)*(k1 + 2*k2 + 2*k3 + k4)\n\n traj[i+1, :] = r + K\n return (ts, traj)\n\ndef generateChain(r0, tf, dt, K,a_c,b_c,a_p,b_p,x0,y0,eps):\n \"\"\"Integrate a given Lorenz system.\"\"\"\n\n # define equations of lorenz system\n def Chain(r, t):\n x1 = r[0]; y1 = r[1]; z1 = r[2]\n x2 = r[3]; y2 = r[4]; z2 = r[5]\n u1 = x1*(1-x1/K) - (a_c*b_c*x1*y1/(x1+x0))\n v1 = a_c*y1*((b_c*x1/(x1+x0))-1)-(a_p*b_p*y1*z1/(y1+y0))+eps*(y2-y1)\n w1 = a_p*z1*(b_p*y1/(y1+y0)-1)+eps*(z2-z1)\n u2 = x2*(1-x2/K) - (a_c*b_c*x2*y2/(x2+x0))\n v2 = a_c*y2*((b_c*x2/(x2+x0))-1)-(a_p*b_p*y2*z2/(y2+y0))+eps*(y1-y2)\n w2 = a_p*z2*(b_p*y2/(y2+y0)-1)+eps*(z1-z2)\n return np.array([u1, v1, w1, u2, v2, w2])\n\n ts, traj = RK4(Chain, r0, tf, dt)\n return (ts, traj)\n\ndef get_chain_data(tf=250, dt=0.02, skip=25, split=0.8):\n _, traj = generateChain((1, 1, 1, 1.2,1.3,1.4), tf, dt, 10, 28, 2)\n \n skip_steps = int(25 / dt)\n traj = traj[skip_steps:]\n \n split_num = int(split * traj.shape[0])\n \n train_data = traj[:split_num]\n val_data = traj[split_num:]\n \n return train_data, val_data\n\ndef delta():\n delta_x = []\n epsilon = np.linspace(0.0, 0.012, 40)\n for j in range(40):\n T,traj=generateChain((1,1,1,1.2,1.3,1.4),tf,dt,0.99,0.4,2.009,0.08,2.876,0.16129,0.5,epsilon[j])\n delta_sum = 0\n Y1 = traj[500:,0]\n Y2 = traj[500:,3]\n T = T[500:]\n for i in range(len(T)):\n delta_sum += abs(Y1[i]-Y2[i])\n delta_ave = delta_sum/len(T)\n delta_x.append(delta_ave)\n delta_x = np.asarray(delta_x)\n epsilon = np.asarray(epsilon)\n return (epsilon, delta_x)\n\nif __name__==\"__main__\":\n tf, dt = 1000, 0.02\n T, traj = generateChain((1, 1, 1,1.2,1.3,1.4),tf,dt,0.99,0.4,2.009,0.08,2.876,0.16129,0.5,0.006)\n epsilon, delta_x = delta()\n plt.figure()\n plt.plot(T[500:], traj[500:,3])\n plt.xlim([200,400])\n plt.figure()\n plt.plot(T[500:], traj[500:,1])\n plt.xlim([200,400])\n plt.figure()\n plt.plot(epsilon, delta_x, 'r-',marker='s')\n plt.show()\n", "sub_path": "reservoir/reservoir-computer-lorenz-master/synchro/food_chains.py", "file_name": "food_chains.py", "file_ext": "py", "file_size_in_byte": 2920, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "numpy.arange", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 88, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 88, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}]}
+{"seq_id": "634619613", "text": "#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\nimport numpy\nimport sys\nimport log\nimport scipy\n\nfrom image3d import * \n#from skimage.filter import canny\nfrom PIL import Image#, ImageFilter\n\ndef get_image_channel(imarray, channelno):\n \"\"\"Returns array that is single channel of selected number from image array.\"\"\"\n width = imarray.shape[0]\n height = imarray.shape[1]\n #log.dbg(\"height=\"+str(height)+\" width=\"+str(width))\n channelarray = np.zeros((width, height))\n for x in xrange(height):\n for y in xrange(width):\n channelarray[y][x] = imarray[y][x][channelno]\n return channelarray \n\ndef get_image_dims(path):\n \"\"\"Returns width,height,depth of TIFF image.\"\"\"\n log.info(\"loading file \"+path)\n im = Image.open(path)\n\n height = im.size[0]\n width = im.size[1]\n for depth in xrange(10000): #find depth\n try: im.seek(depth)\n except: break\n log.dbg(\"height=\"+str(height)+\" width=\"+str(width)+\" depth=\"+str(depth))\n return width,height,depth\n \n\ndef yield_zstack(path):\n \"\"\"Generates triples of r,g,b-channels of consecutive layers from TIFF image.\"\"\"\n log.info(\"loading file \"+path)\n im = Image.open(path)\n\n height = im.size[0]\n width = im.size[1]\n for depth in xrange(10000): #find depth\n try: im.seek(depth)\n except: break\n log.dbg(\"height=\"+str(height)+\" width=\"+str(width)+\" depth=\"+str(depth))\n\n log.dbg(\"extracting pixel values\")\n for z in xrange(depth):\n im.seek(z)\n imarray = numpy.array(im) \n\n red_channel_array = get_image_channel(imarray, 0) \n green_channel_array = get_image_channel(imarray, 1)\n blue_channel_array = get_image_channel(imarray, 2)\n yield (red_channel_array, green_channel_array, blue_channel_array)\n\n\n\nif __name__==\"__main__\":\n log.info(\"--------------------------------------------------------------------------\")\n log.info(\"The program loads TIFF rgb file with pages and stores as separate .png files.\")\n log.info(\"Arguments: src-file, dst-file-prefix, [channel-names (rgb)], [dst-file-suffix], [resample-z times].\")\n\n try: path = sys.argv[1]\n except: log.err(\"Argument (TIFF path) expected!\"); sys.exit(-1);\n\n try: outpath = sys.argv[2]\n except: outpath = path \n\n try: channels = sys.argv[3].lower() \n except: channels = 'rgb'\n\n try: suffix = sys.argv[4].lower() \n except: suffix = ''\n\n try: resample_z = int(sys.argv[5])\n except: resample_z = 3\n\n\n log.info(\"src file = \"+path)\n log.info(\"dst prefix = \"+outpath)\n log.info(\"dst path suffix =\"+suffix)\n log.info(\"channels to be stored =\"+channels)\n log.info(\"resample_z =\"+str(resample_z))\n\n log.info(\"writing layers\")\n for z, (r,g,b) in enumerate(yield_zstack(path)):\n\n #FILTERS:\n #g = canny(g, 3, 0.3, 0.2)\n\n #STORING\n for offset in xrange(resample_z):\n nz = z*resample_z + offset\n log.dbg(\"writing layer with z=\"+str(z)+\" mapped to nz=\"+str(nz))\n if 'r' in channels:\n scipy.misc.imsave(outpath+\"_r\"+suffix+('%04d' % nz)+\".png\", r)\n if 'g' in channels:\n scipy.misc.imsave(outpath+\"_g\"+suffix+('%04d' % nz)+\".png\", g)\n if 'b' in channels:\n scipy.misc.imsave(outpath+\"_b\"+suffix+('%04d' % nz)+\".png\", b)\n\n\n", "sub_path": "stackdecomposition.py", "file_name": "stackdecomposition.py", "file_ext": "py", "file_size_in_byte": 3352, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "log.info", "line_number": 26, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 27, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 27, "usage_type": "name"}, {"api_name": "log.dbg", "line_number": 34, "usage_type": "call"}, {"api_name": "log.info", "line_number": 40, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 41, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 41, "usage_type": "name"}, {"api_name": "log.dbg", "line_number": 48, "usage_type": "call"}, {"api_name": "log.dbg", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 53, "usage_type": "call"}, {"api_name": "log.info", "line_number": 63, "usage_type": "call"}, {"api_name": "log.info", "line_number": 64, "usage_type": "call"}, {"api_name": "log.info", "line_number": 65, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 67, "usage_type": "attribute"}, {"api_name": "log.err", "line_number": 68, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 68, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 70, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 73, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 76, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 79, "usage_type": "attribute"}, {"api_name": "log.info", "line_number": 83, "usage_type": "call"}, {"api_name": "log.info", "line_number": 84, "usage_type": "call"}, {"api_name": "log.info", "line_number": 85, "usage_type": "call"}, {"api_name": "log.info", "line_number": 86, "usage_type": "call"}, {"api_name": "log.info", "line_number": 87, "usage_type": "call"}, {"api_name": "log.info", "line_number": 89, "usage_type": "call"}, {"api_name": "log.dbg", "line_number": 98, "usage_type": "call"}, {"api_name": "scipy.misc.imsave", "line_number": 100, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 100, "usage_type": "attribute"}, {"api_name": "scipy.misc.imsave", "line_number": 102, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 102, "usage_type": "attribute"}, {"api_name": "scipy.misc.imsave", "line_number": 104, "usage_type": "call"}, {"api_name": "scipy.misc", "line_number": 104, "usage_type": "attribute"}]}
+{"seq_id": "334442583", "text": "# standard python imports\nimport os\nimport time\nimport json\nimport hashlib\nimport datetime\n\n\n# external imports\nfrom sqlalchemy.orm.exc import NoResultFound, MultipleResultsFound\n\n\n# internal imports\nfrom main.app import db, message_queue, storage_manager\nfrom main.resources.models import Resource, ResourceRevision, Thumbnail\nfrom main.resources.file_conversion import compute_thumbnail\nfrom main.users.permissions import ACCESS_LEVEL_WRITE, ACCESS_TYPE_ORG_USERS, ACCESS_TYPE_ORG_CONTROLLERS\n\n\n\n# get the number corresponding to a resource type (given by a string name)\ndef resource_type_number(type_name):\n if type_name == 'basic_folder' or type_name == 'basicFolder':\n return Resource.BASIC_FOLDER\n elif type_name == 'organization_folder' or type_name == 'organizationFolder':\n return Resource.ORGANIZATION_FOLDER\n elif type_name == 'controller_folder' or type_name == 'controllerFolder':\n return Resource.CONTROLLER_FOLDER\n elif type_name == 'remote_folder' or type_name == 'remoteFolder':\n return Resource.REMOTE_FOLDER\n elif type_name == 'file':\n return Resource.FILE\n elif type_name == 'sequence':\n return Resource.SEQUENCE\n elif type_name == 'app':\n return Resource.APP\n\n\n# fix(soon): remove this\ndef _create_folders(path):\n parts = path.split('/')\n parent = None\n for part in parts:\n try:\n if parent:\n resource = Resource.query.filter(Resource.parent_id == parent.id, Resource.name == part, Resource.deleted == False).one()\n else:\n resource = Resource.query.filter(Resource.parent_id == None, Resource.name == part, Resource.deleted == False).one()\n except NoResultFound:\n resource = Resource()\n resource.parent_id = parent.id\n resource.organization_id = parent.organization_id\n resource.name = part\n resource.type = Resource.BASIC_FOLDER\n resource.creation_timestamp = datetime.datetime.utcnow()\n resource.modification_timestamp = resource.creation_timestamp\n db.session.add(resource)\n db.session.commit()\n except MultipleResultsFound:\n print('create_folders: duplicate folder at %s in %s' % (part, path))\n parent = resource\n return parent\n\n\n# create a file-type resource with the given contents and other attributes;\n# returns the newly created resource record object (or existing resource if already exists)\n# fix(soon): remove this\ndef _create_file(file_name, creation_timestamp, modification_timestamp, contents):\n last_slash = file_name.rfind('/')\n path = file_name[:last_slash]\n short_file_name = file_name[last_slash+1:]\n folder = _create_folders(path)\n\n # check for existing resource with same name\n try:\n resource = Resource.query.filter(Resource.parent_id == folder.id, Resource.name == short_file_name, Resource.deleted == False).one()\n new_resource = False\n except NoResultFound:\n\n # create new resource record\n resource = Resource()\n resource.parent_id = folder.id\n resource.organization_id = folder.organization_id\n resource.name = short_file_name\n resource.creation_timestamp = creation_timestamp\n resource.type = Resource.FILE\n new_resource = True\n\n # update or init resource record\n resource.deleted = False\n resource.modification_timestamp = modification_timestamp\n if resource.type != Resource.SEQUENCE:\n if resource.system_attributes:\n system_attributes = json.loads(resource.system_attributes)\n else:\n system_attributes = {}\n system_attributes['hash'] = hashlib.sha1(contents).hexdigest()\n system_attributes['size'] = len(contents)\n resource.system_attributes = json.dumps(system_attributes)\n if new_resource:\n db.session.add(resource)\n db.session.commit()\n\n # write file contents to a resource revision (possibly bulk storage)\n add_resource_revision(resource, modification_timestamp, contents)\n db.session.commit()\n\n # compute thumbnail for images\n if file_name.endswith('.png') or file_name.endswith('.jpg'): # fix(soon): handle more types, capitalizations\n for width in [120]: # fix(soon): what will be our standard sizes?\n (thumbnail_contents, thumbnail_width, thumbnail_height) = compute_thumbnail(contents, width) # fix(later): if this returns something other than requested width, we'll keep missing the cache\n thumbnail = Thumbnail()\n thumbnail.resource_id = resource.id\n thumbnail.width = thumbnail_width\n thumbnail.height = thumbnail_height\n thumbnail.format = 'jpg'\n thumbnail.data = thumbnail_contents\n db.session.add(thumbnail)\n db.session.commit()\n return resource\n\n\n# fix(soon): make leading slash required\n# find a resource given it's full name with path\ndef find_resource(file_name):\n file_name = file_name.strip('/')\n parts = file_name.split('/')\n parent = None\n for part in parts:\n try:\n if parent:\n resource = Resource.query.filter(Resource.parent_id == parent.id, Resource.name == part, Resource.deleted == False).one()\n else:\n resource = Resource.query.filter(Resource.parent_id == None, Resource.name == part, Resource.deleted == False).one()\n except NoResultFound:\n return None\n except MultipleResultsFound:\n print('find_resource/MultipleResultsFound: %s' % file_name)\n return None\n parent = resource\n return parent\n\n\n# split a camel case string into a space-separated string\ndef split_camel_case(name):\n result = ''\n for c in name:\n if c.isupper():\n result += ' '\n result += c\n return result\n\n\n# determine (guess) the mimetype of a file based on its file name extension\ndef mime_type_from_ext(file_name):\n file_ext = file_name.rsplit('.', 1)[-1]\n type = ''\n if file_ext == 'jpg':\n type = 'image/jpeg'\n elif file_ext == 'png':\n type = 'image/png'\n elif file_ext == 'txt':\n type = 'text/plain'\n elif file_ext == 'csv':\n type = 'text/csv'\n return type\n\n\n# this is a high-level function for setting the value of a sequence;\n# it (1) creates a sequence value record and (2) sends out a sequence_update message;\n# note that we don't commit resource here; outside code must commit\n# value should be a plain string (not unicode string), possibly containing binary data or encoded unicode data\ndef update_sequence_value(resource, resource_path, timestamp, value, emit_message=True):\n data_type = json.loads(resource.system_attributes)['data_type']\n\n # determine min interval between updates\n system_attributes = json.loads(resource.system_attributes) if resource.system_attributes else {}\n min_storage_interval = system_attributes.get('min_storage_interval')\n if min_storage_interval is None:\n if data_type == Resource.TEXT_SEQUENCE:\n min_storage_interval = 0\n else:\n min_storage_interval = 50\n\n # prep sequence update message data\n if emit_message:\n message_params = {\n 'id': resource.id,\n 'name': resource_path,\n 'timestamp': timestamp.isoformat() + 'Z',\n }\n if data_type != Resource.IMAGE_SEQUENCE: # for images we'll send revision IDs\n message_params['value'] = value # fix(soon): json.dumps crashes if this included binary data\n\n # if too soon since last update, don't store a new value (but do still send out an update message)\n if min_storage_interval == 0 or timestamp >= resource.modification_timestamp + datetime.timedelta(seconds=min_storage_interval):\n resource_revision = add_resource_revision(resource, timestamp, value)\n resource.modification_timestamp = timestamp\n\n # create thumbnails for image sequences\n if data_type == Resource.IMAGE_SEQUENCE:\n max_width = 240\n name = 'thumbnail-%d-x' % max_width\n (thumbnail_contents, thumbnail_width, thumbnail_height) = compute_thumbnail(value, max_width)\n try:\n thumbnail_resource = Resource.query.filter(Resource.parent_id == resource.id, Resource.name == name, Resource.deleted == False).one()\n except NoResultFound:\n thumbnail_resource = create_sequence(resource, name, Resource.IMAGE_SEQUENCE)\n thumbnail_revision = add_resource_revision(thumbnail_resource, timestamp, thumbnail_contents)\n if emit_message:\n message_params['revision_id'] = resource_revision.id\n message_params['thumbnail_revision_id'] = thumbnail_revision.id\n\n # create a short lived update message for subscribers to the folder containing this sequence\n if emit_message:\n message_queue.add(folder_id = resource.parent_id, type = 'sequence_update', parameters = message_params, timestamp = timestamp)\n\n\n# creates a resource revision record; places the data in the record (if it is small) or bulk storage (if it is large);\n# note that we don't commit resource here; outside code must commit\n# data should be a plain string (not unicode string), possibly containing binary data or encoded unicode data\ndef add_resource_revision(resource, timestamp, data):\n resource_revision = ResourceRevision()\n resource_revision.resource_id = resource.id\n resource_revision.timestamp = timestamp\n if len(data) < 1000 or not storage_manager:\n resource_revision.data = data\n bulk_storage = False\n else:\n bulk_storage = True\n db.session.add(resource_revision)\n db.session.commit()\n if bulk_storage:\n storage_manager.write(resource.storage_path(resource_revision.id), data)\n resource.last_revision_id = resource_revision.id # note that we don't commit here; outside code must commit\n return resource_revision\n\n\n# reads the most recent revision/value of a resource;\n# if check_timing is True, will display some timing diagnostics\ndef read_resource(resource, revision_id = None, check_timing = False):\n data = None\n if not revision_id:\n revision_id = resource.last_revision_id # if no last revision, this is a new resource with new data\n if revision_id:\n try:\n if check_timing:\n start_time = time.time()\n resource_revision = ResourceRevision.query.filter(ResourceRevision.id == revision_id).one()\n if check_timing:\n print('query time: %.4f' % (time.time() - start_time))\n data = resource_revision.data\n except NoResultFound:\n pass\n if data is None and storage_manager: # fix(later): move this inside try statement; we should always have a resource revision if we have data in storage\n if check_timing:\n start_time = time.time()\n data = storage_manager.read(resource.storage_path(revision_id))\n if check_timing:\n print('storage time: %.4f' % (time.time() - start_time))\n return data\n\n\n# create a new sequence resource; commits it to database and returns resource record\ndef create_sequence(parent_resource, name, data_type, max_history = 10000, units = None):\n r = Resource()\n r.parent_id = parent_resource.id\n r.organization_id = parent_resource.organization_id\n r.name = name\n r.type = Resource.SEQUENCE\n r.creation_timestamp = datetime.datetime.utcnow()\n r.modification_timestamp = r.creation_timestamp\n system_attributes = {\n 'data_type': data_type,\n 'max_history': max_history\n }\n if units:\n system_attributes['units'] = units\n r.system_attributes = json.dumps(system_attributes)\n db.session.add(r)\n db.session.commit()\n return r\n\n\n# create a new organization record\ndef create_organization(full_name, folder_name):\n r = Resource()\n r.name = folder_name\n r.type = Resource.ORGANIZATION_FOLDER\n r.creation_timestamp = datetime.datetime.utcnow()\n r.modification_timestamp = r.creation_timestamp\n r.system_attributes = json.dumps({\n 'full_name': full_name,\n 'timezone': 'US/Pacific',\n })\n db.session.add(r)\n db.session.commit()\n r.permissions = json.dumps([[ACCESS_TYPE_ORG_USERS, r.id, ACCESS_LEVEL_WRITE], [ACCESS_TYPE_ORG_CONTROLLERS, r.id, ACCESS_LEVEL_WRITE]])\n r.organization_id = r.id # the organization record has its own id as its organization\n db.session.commit()\n return r.id\n\n\n# create/update system resources\ndef create_system_resources():\n print('creating/updating system resources in database')\n\n # make sure system folder exists\n system_folder = find_resource('/system')\n if not system_folder:\n create_organization('System', 'system')\n system_folder = find_resource('/system')\n print('created system folder')\n\n # make sure home page exists\n home_page = find_resource('/system/home.md')\n if not home_page:\n resource = Resource()\n resource.parent_id = system_folder.id\n resource.type = Resource.FILE\n resource.name = 'home.md'\n db.session.add(resource)\n db.session.commit()\n home_contents = '''### Welcome\n\nIf you are logged in as a system admin, you can [edit this page](/system/home.md?edit=1).\n'''\n add_resource_revision(resource, datetime.datetime.utcnow(), home_contents)\n print('created home page')\n\n # fix(soon): create workers folder, log sequence, doc org, workers/log\n\n # add apps for system app templates\n file_names = os.listdir('main/templates/system')\n app_create_count = 0\n for file_name in file_names:\n if file_name.endswith('.html'):\n app_name = file_name.rsplit('.', 1)[0]\n app_title = split_camel_case(app_name).title().replace('_', ' ')\n try:\n resource = Resource.query.filter(Resource.parent_id == system_folder.id, Resource.name == app_title, Resource.deleted == False).one()\n except NoResultFound:\n print('creating: %s, %s' % (app_name, app_title))\n resource = Resource()\n resource.parent_id = system_folder.id\n resource.type = Resource.APP\n resource.name = app_title\n db.session.add(resource)\n db.session.commit()\n app_create_count += 1\n print('created %d apps' % app_create_count)\n", "sub_path": "main/resources/resource_util.py", "file_name": "resource_util.py", "file_ext": "py", "file_size_in_byte": 14543, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "main.resources.models.Resource.BASIC_FOLDER", "line_number": 24, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 24, "usage_type": "name"}, {"api_name": "main.resources.models.Resource.ORGANIZATION_FOLDER", "line_number": 26, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 26, "usage_type": "name"}, {"api_name": "main.resources.models.Resource.CONTROLLER_FOLDER", "line_number": 28, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 28, "usage_type": "name"}, {"api_name": "main.resources.models.Resource.REMOTE_FOLDER", "line_number": 30, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 30, "usage_type": "name"}, {"api_name": "main.resources.models.Resource.FILE", "line_number": 32, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 32, "usage_type": "name"}, {"api_name": "main.resources.models.Resource.SEQUENCE", "line_number": 34, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 34, "usage_type": "name"}, {"api_name": "main.resources.models.Resource.APP", "line_number": 36, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 36, "usage_type": "name"}, {"api_name": "main.resources.models.Resource.query.filter", "line_number": 46, "usage_type": "call"}, {"api_name": "main.resources.models.Resource.query", "line_number": 46, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 46, "usage_type": "name"}, {"api_name": "main.resources.models.Resource.parent_id", "line_number": 46, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource.name", "line_number": 46, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource.deleted", "line_number": 46, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource.query.filter", "line_number": 48, "usage_type": "call"}, {"api_name": "main.resources.models.Resource.query", "line_number": 48, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 48, "usage_type": "name"}, {"api_name": "main.resources.models.Resource.parent_id", "line_number": 48, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource.name", "line_number": 48, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource.deleted", "line_number": 48, "usage_type": "attribute"}, {"api_name": "sqlalchemy.orm.exc.NoResultFound", "line_number": 49, "usage_type": "name"}, {"api_name": "main.resources.models.Resource", "line_number": 50, "usage_type": "call"}, {"api_name": "main.resources.models.Resource.BASIC_FOLDER", "line_number": 54, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 54, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 55, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 55, "usage_type": "attribute"}, {"api_name": "main.app.db.session.add", "line_number": 57, "usage_type": "call"}, {"api_name": "main.app.db.session", "line_number": 57, "usage_type": "attribute"}, {"api_name": "main.app.db", "line_number": 57, "usage_type": "name"}, {"api_name": "main.app.db.session.commit", "line_number": 58, "usage_type": "call"}, {"api_name": "main.app.db.session", "line_number": 58, "usage_type": "attribute"}, {"api_name": "main.app.db", "line_number": 58, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.exc.MultipleResultsFound", "line_number": 59, "usage_type": "name"}, {"api_name": "main.resources.models.Resource.query.filter", "line_number": 76, "usage_type": "call"}, {"api_name": "main.resources.models.Resource.query", "line_number": 76, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 76, "usage_type": "name"}, {"api_name": "main.resources.models.Resource.parent_id", "line_number": 76, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource.name", "line_number": 76, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource.deleted", "line_number": 76, "usage_type": "attribute"}, {"api_name": "sqlalchemy.orm.exc.NoResultFound", "line_number": 78, "usage_type": "name"}, {"api_name": "main.resources.models.Resource", "line_number": 81, "usage_type": "call"}, {"api_name": "main.resources.models.Resource.FILE", "line_number": 86, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 86, "usage_type": "name"}, {"api_name": "main.resources.models.Resource.SEQUENCE", "line_number": 92, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 92, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 94, "usage_type": "call"}, {"api_name": "hashlib.sha1", "line_number": 97, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 99, "usage_type": "call"}, {"api_name": "main.app.db.session.add", "line_number": 101, "usage_type": "call"}, {"api_name": "main.app.db.session", "line_number": 101, "usage_type": "attribute"}, {"api_name": "main.app.db", "line_number": 101, "usage_type": "name"}, {"api_name": "main.app.db.session.commit", "line_number": 102, "usage_type": "call"}, {"api_name": "main.app.db.session", "line_number": 102, "usage_type": "attribute"}, {"api_name": "main.app.db", "line_number": 102, "usage_type": "name"}, {"api_name": "main.app.db.session.commit", "line_number": 106, "usage_type": "call"}, {"api_name": "main.app.db.session", "line_number": 106, "usage_type": "attribute"}, {"api_name": "main.app.db", "line_number": 106, "usage_type": "name"}, {"api_name": "main.resources.file_conversion.compute_thumbnail", "line_number": 111, "usage_type": "call"}, {"api_name": "main.resources.models.Thumbnail", "line_number": 112, "usage_type": "call"}, {"api_name": "main.app.db.session.add", "line_number": 118, "usage_type": "call"}, {"api_name": "main.app.db.session", "line_number": 118, "usage_type": "attribute"}, {"api_name": "main.app.db", "line_number": 118, "usage_type": "name"}, {"api_name": "main.app.db.session.commit", "line_number": 119, "usage_type": "call"}, {"api_name": "main.app.db.session", "line_number": 119, "usage_type": "attribute"}, {"api_name": "main.app.db", "line_number": 119, "usage_type": "name"}, {"api_name": "main.resources.models.Resource.query.filter", "line_number": 132, "usage_type": "call"}, {"api_name": "main.resources.models.Resource.query", "line_number": 132, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 132, "usage_type": "name"}, {"api_name": "main.resources.models.Resource.parent_id", "line_number": 132, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource.name", "line_number": 132, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource.deleted", "line_number": 132, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource.query.filter", "line_number": 134, "usage_type": "call"}, {"api_name": "main.resources.models.Resource.query", "line_number": 134, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 134, "usage_type": "name"}, {"api_name": "main.resources.models.Resource.parent_id", "line_number": 134, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource.name", "line_number": 134, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource.deleted", "line_number": 134, "usage_type": "attribute"}, {"api_name": "sqlalchemy.orm.exc.NoResultFound", "line_number": 135, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.exc.MultipleResultsFound", "line_number": 137, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 174, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 177, "usage_type": "call"}, {"api_name": "main.resources.models.Resource.TEXT_SEQUENCE", "line_number": 180, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 180, "usage_type": "name"}, {"api_name": "main.resources.models.Resource.IMAGE_SEQUENCE", "line_number": 192, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 192, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 196, "usage_type": "call"}, {"api_name": "main.resources.models.Resource.IMAGE_SEQUENCE", "line_number": 201, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 201, "usage_type": "name"}, {"api_name": "main.resources.file_conversion.compute_thumbnail", "line_number": 204, "usage_type": "call"}, {"api_name": "main.resources.models.Resource.query.filter", "line_number": 206, "usage_type": "call"}, {"api_name": "main.resources.models.Resource.query", "line_number": 206, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 206, "usage_type": "name"}, {"api_name": "main.resources.models.Resource.parent_id", "line_number": 206, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource.name", "line_number": 206, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource.deleted", "line_number": 206, "usage_type": "attribute"}, {"api_name": "sqlalchemy.orm.exc.NoResultFound", "line_number": 207, "usage_type": "name"}, {"api_name": "main.resources.models.Resource.IMAGE_SEQUENCE", "line_number": 208, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 208, "usage_type": "name"}, {"api_name": "main.app.message_queue.add", "line_number": 216, "usage_type": "call"}, {"api_name": "main.app.message_queue", "line_number": 216, "usage_type": "name"}, {"api_name": "main.resources.models.ResourceRevision", "line_number": 223, "usage_type": "call"}, {"api_name": "main.app.storage_manager", "line_number": 226, "usage_type": "name"}, {"api_name": "main.app.db.session.add", "line_number": 231, "usage_type": "call"}, {"api_name": "main.app.db.session", "line_number": 231, "usage_type": "attribute"}, {"api_name": "main.app.db", "line_number": 231, "usage_type": "name"}, {"api_name": "main.app.db.session.commit", "line_number": 232, "usage_type": "call"}, {"api_name": "main.app.db.session", "line_number": 232, "usage_type": "attribute"}, {"api_name": "main.app.db", "line_number": 232, "usage_type": "name"}, {"api_name": "main.app.storage_manager.write", "line_number": 234, "usage_type": "call"}, {"api_name": "main.app.storage_manager", "line_number": 234, "usage_type": "name"}, {"api_name": "time.time", "line_number": 248, "usage_type": "call"}, {"api_name": "main.resources.models.ResourceRevision.query.filter", "line_number": 249, "usage_type": "call"}, {"api_name": "main.resources.models.ResourceRevision.query", "line_number": 249, "usage_type": "attribute"}, {"api_name": "main.resources.models.ResourceRevision", "line_number": 249, "usage_type": "name"}, {"api_name": "main.resources.models.ResourceRevision.id", "line_number": 249, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 251, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.exc.NoResultFound", "line_number": 253, "usage_type": "name"}, {"api_name": "main.app.storage_manager", "line_number": 255, "usage_type": "name"}, {"api_name": "time.time", "line_number": 257, "usage_type": "call"}, {"api_name": "main.app.storage_manager.read", "line_number": 258, "usage_type": "call"}, {"api_name": "main.app.storage_manager", "line_number": 258, "usage_type": "name"}, {"api_name": "time.time", "line_number": 260, "usage_type": "call"}, {"api_name": "main.resources.models.Resource", "line_number": 266, "usage_type": "call"}, {"api_name": "main.resources.models.Resource.SEQUENCE", "line_number": 270, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 270, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 271, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 271, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 279, "usage_type": "call"}, {"api_name": "main.app.db.session.add", "line_number": 280, "usage_type": "call"}, {"api_name": "main.app.db.session", "line_number": 280, "usage_type": "attribute"}, {"api_name": "main.app.db", "line_number": 280, "usage_type": "name"}, {"api_name": "main.app.db.session.commit", "line_number": 281, "usage_type": "call"}, {"api_name": "main.app.db.session", "line_number": 281, "usage_type": "attribute"}, {"api_name": "main.app.db", "line_number": 281, "usage_type": "name"}, {"api_name": "main.resources.models.Resource", "line_number": 287, "usage_type": "call"}, {"api_name": "main.resources.models.Resource.ORGANIZATION_FOLDER", "line_number": 289, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 289, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 290, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 290, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 292, "usage_type": "call"}, {"api_name": "main.app.db.session.add", "line_number": 296, "usage_type": "call"}, {"api_name": "main.app.db.session", "line_number": 296, "usage_type": "attribute"}, {"api_name": "main.app.db", "line_number": 296, "usage_type": "name"}, {"api_name": "main.app.db.session.commit", "line_number": 297, "usage_type": "call"}, {"api_name": "main.app.db.session", "line_number": 297, "usage_type": "attribute"}, {"api_name": "main.app.db", "line_number": 297, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 298, "usage_type": "call"}, {"api_name": "main.users.permissions.ACCESS_TYPE_ORG_USERS", "line_number": 298, "usage_type": "name"}, {"api_name": "main.users.permissions.ACCESS_LEVEL_WRITE", "line_number": 298, "usage_type": "name"}, {"api_name": "main.users.permissions.ACCESS_TYPE_ORG_CONTROLLERS", "line_number": 298, "usage_type": "name"}, {"api_name": "main.app.db.session.commit", "line_number": 300, "usage_type": "call"}, {"api_name": "main.app.db.session", "line_number": 300, "usage_type": "attribute"}, {"api_name": "main.app.db", "line_number": 300, "usage_type": "name"}, {"api_name": "main.resources.models.Resource", "line_number": 318, "usage_type": "call"}, {"api_name": "main.resources.models.Resource.FILE", "line_number": 320, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 320, "usage_type": "name"}, {"api_name": "main.app.db.session.add", "line_number": 322, "usage_type": "call"}, {"api_name": "main.app.db.session", "line_number": 322, "usage_type": "attribute"}, {"api_name": "main.app.db", "line_number": 322, "usage_type": "name"}, {"api_name": "main.app.db.session.commit", "line_number": 323, "usage_type": "call"}, {"api_name": "main.app.db.session", "line_number": 323, "usage_type": "attribute"}, {"api_name": "main.app.db", "line_number": 323, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 328, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 328, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 334, "usage_type": "call"}, {"api_name": "main.resources.models.Resource.query.filter", "line_number": 341, "usage_type": "call"}, {"api_name": "main.resources.models.Resource.query", "line_number": 341, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 341, "usage_type": "name"}, {"api_name": "main.resources.models.Resource.parent_id", "line_number": 341, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource.name", "line_number": 341, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource.deleted", "line_number": 341, "usage_type": "attribute"}, {"api_name": "sqlalchemy.orm.exc.NoResultFound", "line_number": 342, "usage_type": "name"}, {"api_name": "main.resources.models.Resource", "line_number": 344, "usage_type": "call"}, {"api_name": "main.resources.models.Resource.APP", "line_number": 346, "usage_type": "attribute"}, {"api_name": "main.resources.models.Resource", "line_number": 346, "usage_type": "name"}, {"api_name": "main.app.db.session.add", "line_number": 348, "usage_type": "call"}, {"api_name": "main.app.db.session", "line_number": 348, "usage_type": "attribute"}, {"api_name": "main.app.db", "line_number": 348, "usage_type": "name"}, {"api_name": "main.app.db.session.commit", "line_number": 349, "usage_type": "call"}, {"api_name": "main.app.db.session", "line_number": 349, "usage_type": "attribute"}, {"api_name": "main.app.db", "line_number": 349, "usage_type": "name"}]}
+{"seq_id": "315924195", "text": "import os\nimport re\nimport cv2\nimport argparse\nimport functools\nimport subprocess\nimport numpy as np\nfrom PIL import Image\nimport moviepy.editor as mpy\n\nimport torchvision\nimport torch.nn.parallel\nimport torch.optim\nfrom ops.models import TSN\nfrom ops.transforms import * \nfrom torch.nn import functional as F\nimport uuid\ndef generateFileName():\n\tunique_filename = str(uuid.uuid4())\n\treturn uniunique_filename\ndef parse_shift_option_from_log_name(log_name):\n if 'shift' in log_name:\n strings = log_name.split('_')\n for i, s in enumerate(strings):\n if 'shift' in s:\n break\n return True, int(strings[i].replace('shift', '')), strings[i + 1]\n else:\n return False, None, None\n \n \n# options\nparser = argparse.ArgumentParser(description=\"test TSM on a single video\")\ngroup = parser.add_mutually_exclusive_group(required=True)\ngroup.add_argument('--video_file', type=str, default=None)\ngroup.add_argument('--frame_folder', type=str, default=None)\nparser.add_argument('--modality', type=str, default='RGB',\n choices=['RGB', 'Flow', 'RGBDiff'], )\nparser.add_argument('--dataset', type=str, default='moments',\n choices=['ucfcrime','something', 'jester', 'moments', 'somethingv2'])\nparser.add_argument('--rendered_output', type=str, default=None)\n#parser.add_argument('--arch', type=str, default=\"InceptionV3\")\nparser.add_argument('--input_size', type=int, default=224)\nparser.add_argument('--test_segments', type=int, default=8)\nparser.add_argument('--img_feature_dim', type=int, default=256)\nparser.add_argument('--consensus_type', type=str, default='avg')\nparser.add_argument('--weights', type=str)\n\n\nargs = parser.parse_args()\nthis_weights = args.weights\nis_shift, shift_div, shift_place = parse_shift_option_from_log_name(this_weights)\nmodality = 'RGB'\nif 'RGB' in this_weights:\n\tmodality = 'RGB'\n\n# Get dataset categories.\ncategories = ['Normal', #0\n\t'Abnormal', ] #10\nnum_class = len(categories)\nthis_arch = 'resnet50'\n\nnet = TSN(num_class, 1, modality,\n base_model=this_arch,\n consensus_type='avg',\n img_feature_dim=args.img_feature_dim,\n #pretrain=args.pretrain,\n is_shift=is_shift, shift_div=shift_div, shift_place=shift_place,\n non_local='_nl' in this_weights,\n )\n\ncheckpoint = torch.load(this_weights)\ncheckpoint = checkpoint['state_dict']\n\n# base_dict = {('base_model.' + k).replace('base_model.fc', 'new_fc'): v for k, v in list(checkpoint.items())}\nbase_dict = {'.'.join(k.split('.')[1:]): v for k, v in list(checkpoint.items())}\nreplace_dict = {'base_model.classifier.weight': 'new_fc.weight',\n 'base_model.classifier.bias': 'new_fc.bias',\n }\nfor k, v in replace_dict.items():\n if k in base_dict:\n base_dict[v] = base_dict.pop(k)\n\nnet.load_state_dict(base_dict)\nnet.cuda().eval()\n\ntransform=torchvision.transforms.Compose([\n Stack(roll=(this_arch in ['BNInception', 'InceptionV3'])),\n ToTorchFormatTensor(div=(this_arch not in ['BNInception', 'InceptionV3'])),\n GroupNormalize(net.input_mean, net.input_std),\n ])\n\n\ndef extract_frames(video_file, num_frames=8):\n try:\n os.makedirs(os.path.join(os.getcwd(), 'frames'))\n except OSError:\n pass\n\n output = subprocess.Popen(['ffmpeg', '-i', video_file],\n stderr=subprocess.PIPE).communicate()\n # Search and parse 'Duration: 00:05:24.13,' from ffmpeg stderr.\n re_duration = re.compile('Duration: (.*?)\\.')\n duration = re_duration.search(str(output[1])).groups()[0]\n\n seconds = functools.reduce(lambda x, y: x * 60 + y,\n map(int, duration.split(':')))\n rate = num_frames / float(seconds)\n print(rate)\n output = subprocess.Popen(['ffmpeg', '-i', video_file,\n '-vf', 'fps={}'.format(rate),\n '-vframes', str(15),\n '-loglevel', 'panic',\n 'frames/%d.jpg']).communicate()\n frame_paths = sorted([os.path.join('frames', frame)\n for frame in os.listdir('frames')])\n\n frames = load_frames(frame_paths)\n subprocess.call(['rm', '-rf', 'frames'])\n print(len(frames))\n return frames\n\n\ndef load_frames(frame_paths, num_frames=8):\n frames = [Image.open(frame).convert('RGB') for frame in frame_paths]\n if len(frames) >= num_frames:\n return frames[::int(np.ceil(len(frames) / float(num_frames)))]\n else:\n raise ValueError('Video must have at least {} frames'.format(num_frames))\n\n\ndef render_frames(frames, prediction):\n rendered_frames = []\n for frame in frames:\n img = np.array(frame)\n height, width, _ = img.shape\n cv2.putText(img, prediction,\n (0, int(height / 16)),\n cv2.FONT_HERSHEY_SIMPLEX,\n 0.7, (0, 255, 255), 2)\n rendered_frames.append(img)\n return rendered_frames\n\n\n# Obtain video frames\nif args.frame_folder is not None:\n print('Loading frames in {}'.format(args.frame_folder))\n import glob\n # Here, make sure after sorting the frame paths have the correct temporal order\n frame_paths = sorted(glob.glob(os.path.join(args.frame_folder, '*.jpg')))\n frames = load_frames(frame_paths)\nelse:\n print('Extracting frames using ffmpeg...')\n frames = extract_frames(args.video_file, args.test_segments)\n\n#print(frames)\n# Make video prediction.\ndata = transform(frames)\ninput = data.view(-1, 3, data.size(1), data.size(2)).unsqueeze(0).cuda()\n\nwith torch.no_grad():\n logits = net(input)\n h_x = torch.mean(F.softmax(logits, 1), dim=0).data\n probs, idx = h_x.sort(0, True)\n\n# Output the prediction.\nvideo_name = args.frame_folder if args.frame_folder is not None else args.video_file\nprint('RESULT ON ' + video_name)\nfor i in range(0, 2):\n print('{:.3f} -> {}'.format(probs[i], categories[idx[i]]))\n\n# Render output frames with prediction text.\nif args.rendered_output is not None:\n prediction = categories[idx[0]]\n rendered_frames = render_frames(frames, prediction)\n clip = mpy.ImageSequenceClip(rendered_frames, fps=4)\n clip.write_videofile(args.rendered_output)\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "tsm_model/SingleVideoTest.py", "file_name": "SingleVideoTest.py", "file_ext": "py", "file_size_in_byte": 6389, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "uuid.uuid4", "line_number": 19, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 33, "usage_type": "call"}, {"api_name": "ops.models.TSN", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.nn.parallel.load", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.nn.parallel", "line_number": 72, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 87, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 87, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path", "line_number": 96, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 96, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 100, "usage_type": "call"}, {"api_name": "subprocess.PIPE", "line_number": 101, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 103, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 106, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 110, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path", "line_number": 115, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 116, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 119, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 125, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 125, "usage_type": "name"}, {"api_name": "numpy.ceil", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 135, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 137, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 139, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 150, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 150, "usage_type": "call"}, {"api_name": "os.path", "line_number": 150, "usage_type": "attribute"}, {"api_name": "torch.nn.parallel.no_grad", "line_number": 161, "usage_type": "call"}, {"api_name": "torch.nn.parallel", "line_number": 161, "usage_type": "name"}, {"api_name": "torch.nn.parallel.mean", "line_number": 163, "usage_type": "call"}, {"api_name": "torch.nn.parallel", "line_number": 163, "usage_type": "name"}, {"api_name": "torch.nn.functional.softmax", "line_number": 163, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 163, "usage_type": "name"}, {"api_name": "moviepy.editor.ImageSequenceClip", "line_number": 176, "usage_type": "call"}, {"api_name": "moviepy.editor", "line_number": 176, "usage_type": "name"}]}
+{"seq_id": "578016536", "text": "import torch\nimport math\nimport torch.nn as nn\nimport torch.nn.functional as F\nfrom torch.nn import Parameter\n\nfrom efficientnet import EfficientNet\nfrom senet import se_resnext50_32x4d, se_resnext101_32x4d\n\n\nclass AdaptiveConcatPool2d(nn.Module):\n def __init__(self, sz=None):\n super().__init__()\n self.output_size = sz\n self.ap = nn.AdaptiveAvgPool2d(self.output_size)\n self.mp = nn.AdaptiveMaxPool2d(self.output_size)\n\n def forward(self, x):\n return torch.cat([self.mp(x), self.ap(x)], 1)\n\n\nclass Flatten(nn.Module):\n def __init__(self):\n super().__init__()\n\n def forward(self, x):\n return x.view(x.size(0), -1)\n\n\nclass PandaNet(nn.Module):\n def __init__(self, arch, pretrained=True, embedding_size=512):\n super().__init__()\n\n # load EfficientNet\n if 'efficientnet' in arch:\n if pretrained:\n self.base = EfficientNet.from_pretrained(model_name=arch)\n else:\n self.base = EfficientNet.from_name(model_name=arch)\n\n self.nc = self.base._fc.in_features\n self.extract_features = self.base.extract_features\n\n elif arch == 'se_resnext50_32x4d':\n if pretrained:\n self.base = se_resnext50_32x4d()\n else:\n self.base = se_resnext50_32x4d(pretrained=None)\n self.nc = self.base.last_linear.in_features\n self.extract_features = self.base.features\n\n elif arch == 'se_resnext101_32x4d':\n if pretrained:\n self.base = se_resnext101_32x4d()\n else:\n self.base = se_resnext101_32x4d(pretrained=None)\n self.nc = self.base.last_linear.in_features\n self.extract_features = self.base.features\n\n self.output = nn.Sequential(AdaptiveConcatPool2d(1),\n Flatten(),\n nn.BatchNorm1d(2 * self.nc),\n nn.Dropout(0.5),\n nn.Linear(2 * self.nc, 512))\n\n def forward(self, inputs):\n bs, num_tiles, c, h, w = inputs.size()\n inputs = inputs.view(-1, c, h, w)\n\n x = self.extract_features(inputs) # bs*N x c x h x w\n shape = x.shape\n\n # concatenate the output for tiles into a single map\n x = x.view(-1, num_tiles, shape[1], shape[2], shape[3]).permute(0, 2, 1, 3, 4).contiguous() \\\n .view(-1, shape[1], shape[2] * num_tiles, shape[3])\n\n # Pooling and final linear layer\n x = self.output(x)\n\n return x\n\n\nclass ArcMarginProduct(nn.Module):\n r\"\"\"Implement of large margin arc distance: :\n Args:\n in_features: size of each input sample\n out_features: size of each output sample\n s: norm of input feature\n m: margin\n cos(theta + m)\n \"\"\"\n\n def __init__(self, in_features, out_features, s=30.0, m=0.50, easy_margin=False):\n super().__init__()\n self.in_features = in_features\n self.out_features = out_features\n self.s = s\n self.m = m\n self.weight = Parameter(torch.FloatTensor(out_features, in_features))\n nn.init.xavier_uniform_(self.weight)\n\n self.easy_margin = easy_margin\n self.cos_m = math.cos(m)\n self.sin_m = math.sin(m)\n self.th = math.cos(math.pi - m)\n self.mm = math.sin(math.pi - m) * m\n\n def forward(self, input, label):\n # --------------------------- cos(theta) & phi(theta) ---------------------------\n cosine = F.linear(F.normalize(input), F.normalize(self.weight))\n sine = torch.sqrt((1.0 - torch.pow(cosine, 2)).clamp(0, 1))\n phi = cosine * self.cos_m - sine * self.sin_m\n if self.easy_margin:\n phi = torch.where(cosine > 0, phi, cosine)\n else:\n phi = torch.where(cosine > self.th, phi, cosine - self.mm)\n # --------------------------- convert label to one-hot ---------------------------\n # one_hot = torch.zeros(cosine.size(), requires_grad=True, device='cuda')\n one_hot = torch.zeros(cosine.size(), device='cuda')\n one_hot.scatter_(1, label.view(-1, 1).long(), 1)\n # -------------torch.where(out_i = {x_i if condition_i else y_i) -------------\n output = (one_hot * phi) + (\n (1.0 - one_hot) * cosine) # you can use torch.where if your torch.__version__ is 0.4\n output *= self.s\n # print(output)\n\n return output\n", "sub_path": "metric_learning/model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 4537, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "torch.nn.Module", "line_number": 11, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 11, "usage_type": "name"}, {"api_name": "torch.nn.AdaptiveAvgPool2d", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.nn.AdaptiveMaxPool2d", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 16, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 22, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 30, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 30, "usage_type": "name"}, {"api_name": "efficientnet.EfficientNet.from_pretrained", "line_number": 37, "usage_type": "call"}, {"api_name": "efficientnet.EfficientNet", "line_number": 37, "usage_type": "name"}, {"api_name": "efficientnet.EfficientNet.from_name", "line_number": 39, "usage_type": "call"}, {"api_name": "efficientnet.EfficientNet", "line_number": 39, "usage_type": "name"}, {"api_name": "senet.se_resnext50_32x4d", "line_number": 46, "usage_type": "call"}, {"api_name": "senet.se_resnext50_32x4d", "line_number": 48, "usage_type": "call"}, {"api_name": "senet.se_resnext101_32x4d", "line_number": 54, "usage_type": "call"}, {"api_name": "senet.se_resnext101_32x4d", "line_number": 56, "usage_type": "call"}, {"api_name": "torch.nn.Sequential", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 60, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 62, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 63, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 64, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 83, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 83, "usage_type": "name"}, {"api_name": "torch.nn.Parameter", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.nn.init.xavier_uniform_", "line_number": 100, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 100, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 100, "usage_type": "name"}, {"api_name": "math.cos", "line_number": 103, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 104, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 105, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 105, "usage_type": "attribute"}, {"api_name": "math.sin", "line_number": 106, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 106, "usage_type": "attribute"}, {"api_name": "torch.nn.functional.linear", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 110, "usage_type": "name"}, {"api_name": "torch.nn.functional.normalize", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.sqrt", "line_number": 111, "usage_type": "call"}, {"api_name": "torch.pow", "line_number": 111, "usage_type": "call"}, {"api_name": "torch.where", "line_number": 114, "usage_type": "call"}, {"api_name": "torch.where", "line_number": 116, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 119, "usage_type": "call"}]}
+{"seq_id": "161151299", "text": "import json\nfrom abc import abstractmethod\nfrom functools import reduce\nimport os\n\nimport numpy as np\nfrom scipy.stats import hypergeom\n\nfrom ..log import setup_logger\nfrom ..utils.tools import parse\nfrom ._unit import Unit\n\n\nthis_dir = os.path.dirname(__file__)\ndef join_root(path):\n return os.path.abspath(os.path.join(this_dir, path))\n\n\ndef _get_dict(path):\n \"\"\"\n Parameters\n __________\n\n path: string or array\n Path to json file. In case a list of paths is provided instead,\n read them all and merge then into a single dict. Assumes depth two.\n\n Returns\n _______\n d: dict\n Dictionary containing marker information.\n d = {\n key: {\n subkey: [...],\n ...\n },\n ...\n }\n \"\"\"\n # TODO straighten up the spaghetti\n\n if isinstance(path, str):\n with open(path, \"r\") as f:\n return json.load(f)\n else:\n d = {}\n for path in path:\n with open(path, \"r\") as f:\n d_part = json.load(f)\n for key in d_part:\n if key in d:\n for subkey in d_part[key]:\n if subkey in d[key]:\n # to remove duplicates\n d[key][subkey] = list(set().union(\n d[key][subkey], d_part[key][subkey]))\n else:\n d[key][subkey] = d_part[key][subkey]\n else:\n d[key] = d_part[key]\n return d\n\n\nclass Ide_HyperGeom(Unit):\n \"\"\"\n Runs hypergeom to find matching populations. Compute for every label\n in x, the pop in pops where x is most likely to have been drawn from.\n It is assumed that the dictionary that is passed has two levels of\n hierarchy of types. First determine the lvl1 type, then the lvl2 subtype.\n \"\"\"\n\n def __init__(self, path=join_root('../markers/cell_type_marker.json'), tissue='all'):\n self.logger = setup_logger(\"HyperGeom\")\n self.path = path\n self.tissue = tissue\n\n def get(self, x):\n \"\"\"\n Extended keys are: lvl1_type, lvl1_sv, lvl1_intersec, lvl1_total,\n lvl2_type, lvl2_sv, lvl2_intersec, lvl2_total\n type (string): identified type\n sv (float): survival value from Hypergeometric Test\n intersec (np.ndarray): array of names that overlap\n total (int): total number of names in dict[type]\n\n Returns the types of cells in x.\n\n Args:\n x (dict): x = {\n label_1: {\n outp_names: [name_1, ...],\n ...\n },\n ...\n }\n\n Returns:\n (dict): Extends x with new keys (returns copy).\n \"\"\"\n x = x.copy()\n lvl2 = _get_dict(self.path)\n\n # Construct lvl1 dict by merging all lvl2 dicts\n lvl1 = {}\n for pop in lvl2:\n lvl1[pop] = parse(\n np.array(reduce(lambda a, b: a+b, lvl2[pop].values()))\n )\n\n if self.tissue == 'all':\n self.process_level(x, lvl1, level=1)\n self.process_level(x, lvl2, level=2)\n else:\n self.logger.info(\n \"Running HyperGeom for {0} only.\".format(self.tissue))\n self.process_tissue(x, tissue=self.tissue, level_dict=lvl2)\n\n return x\n\n def process_level(self, x, level_dict, level):\n for key in x:\n if level > 1 and x[key]['lvl{0}_type'.format(level-1)] == 'None':\n tp, sv, intersec, total = \"None\", 1, np.array([]), 0\n all_pops = {'svs': np.array([]),\n 'intersecs': np.array([]),\n 'lens': np.array([])}\n else:\n if level > 1:\n tp, sv, intersec, total, all_pops = self.find_population(\n # x[key]['outp_names'],\n x[key]['lvl{0}_intersec'.format(level-1)],\n level_dict[x[key]['lvl{0}_type'.format(level-1)]]\n )\n else:\n tp, sv, intersec, total, all_pops = self.find_population(\n x[key]['outp_names'],\n level_dict\n )\n x[key]['lvl{0}_type'.format(level)] = tp\n x[key]['lvl{0}_sv'.format(level)] = sv\n x[key]['lvl{0}_intersec'.format(level)] = intersec\n x[key]['lvl{0}_total'.format(level)] = total\n x[key]['lvl{0}_all'.format(level)] = all_pops\n self.logger.info(\"Finished finding lvl{0} types.\".format(level))\n\n def process_tissue(self, x, tissue, level_dict):\n for key in x:\n tp, sv, intersec, total, all_pops = self.find_population(\n x[key]['outp_names'],\n level_dict[tissue]\n )\n x[key]['lvl1_type'] = \"User Defined\"\n x[key]['lvl1_sv'] = 1\n x[key]['lvl1_intersec'] = np.array([])\n x[key]['lvl1_total'] = 0\n x[key]['lvl1_all'] = {}\n\n x[key]['lvl2_type'] = tp\n x[key]['lvl2_sv'] = sv\n x[key]['lvl2_intersec'] = intersec\n x[key]['lvl2_total'] = total\n x[key]['lvl2_all'] = all_pops\n self.logger.info(\"Finished finding lvl2 types.\")\n\n def find_population(self, x, pops):\n \"\"\"\n See find_populations. Assumes x is a single list.\n\n Args:\n x (np.ndarray): 1D list of names.\n pops (dict): Dictionary of populations: pops = {\n type: [name_1, name_2, ...],\n ...\n }\n Returns:\n (string): population name\n (float): survival value\n (np.ndarray): common names\n (int): total number of names in matched population\n \"\"\"\n M = sum([len(pops[pop]) for pop in pops])\n N = len(x)\n\n survival_values = []\n intersections = []\n lens = []\n\n rsv, rpop, rk = 2, -1, 0\n\n for pop in pops:\n n = len(pops[pop])\n intersec = np.intersect1d(x, pops[pop])\n k = len(intersec)\n sv = hypergeom.sf(k-1, M=M, n=n, N=N) if k > 0 else 1\n\n survival_values.append(sv)\n intersections.append(intersec)\n lens.append(len(pops[pop]))\n\n if sv <= rsv or (rsv == 2 and k > 0):\n rsv, rpop, rk = sv, pop, k\n\n all_pops = {'svs': np.array(survival_values),\n 'intersecs': np.array(intersections),\n 'lens': np.array(lens)}\n\n if rk == 0: # in case of no intersection, return -1\n return \"None\", 1, np.array([]), 0, all_pops\n else:\n return rpop, rsv, np.intersect1d(x, pops[rpop]), len(pops[rpop]), all_pops\n", "sub_path": "src/units/_identificator.py", "file_name": "_identificator.py", "file_ext": "py", "file_size_in_byte": 6995, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.path.dirname", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "json.load", "line_number": 44, "usage_type": "call"}, {"api_name": "json.load", "line_number": 49, "usage_type": "call"}, {"api_name": "_unit.Unit", "line_number": 64, "usage_type": "name"}, {"api_name": "log.setup_logger", "line_number": 73, "usage_type": "call"}, {"api_name": "utils.tools.parse", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 107, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.intersect1d", "line_number": 192, "usage_type": "call"}, {"api_name": "scipy.stats.hypergeom.sf", "line_number": 194, "usage_type": "call"}, {"api_name": "scipy.stats.hypergeom", "line_number": 194, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 203, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 208, "usage_type": "call"}, {"api_name": "numpy.intersect1d", "line_number": 210, "usage_type": "call"}]}
+{"seq_id": "606385008", "text": "from rest_framework import serializers\nfrom api import models\n\n\nclass CourseSerializer(serializers.ModelSerializer):\n\n course_type = serializers.CharField(source='get_course_type_display')\n\n class Meta:\n model = models.Course\n fields = ['id', 'name', 'course_type']\n\n\nclass DegreeCourseSerializer(serializers.ModelSerializer):\n teachers = serializers.SerializerMethodField()\n scholarship = serializers.SerializerMethodField()\n\n class Meta:\n model = models.DegreeCourse\n fields = ['id', 'name', 'teachers', 'scholarship']\n\n def get_teachers(self, obj):\n obj_list = obj.teachers.all()\n return [item.name for item in obj_list]\n\n def get_scholarship(self, obj):\n obj_list = obj.scholarship_set.all()\n return [item.value for item in obj_list]\n\n\nclass DegreeCourseDetailSerializer(serializers.ModelSerializer):\n module_name = serializers.SerializerMethodField()\n\n class Meta:\n model = models.DegreeCourse\n fields = ['name', 'module_name']\n\n def get_module_name(self, obj):\n obj_list = obj.course_set.all()\n return [item.name for item in obj_list]\n\n\nclass CourseDetailSerializer(serializers.ModelSerializer):\n level = serializers.CharField(source='get_level_display')\n why_study = serializers.CharField(source='coursedetail.why_study')\n learn_content = serializers.CharField(source='coursedetail.what_to_study_brief')\n recommend_courses = serializers.SerializerMethodField()\n\n class Meta:\n model = models.Course\n fields = ['name', 'level', 'why_study', 'learn_content', 'recommend_courses']\n\n def get_recommend_courses(self,row):\n recommend_list = row.coursedetail.recommend_courses.all()\n return [{'id': item.id, 'name': item.name} for item in recommend_list]\n", "sub_path": "api/serializers.py", "file_name": "serializers.py", "file_ext": "py", "file_size_in_byte": 1830, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 5, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 5, "usage_type": "name"}, {"api_name": "rest_framework.serializers.CharField", "line_number": 7, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 7, "usage_type": "name"}, {"api_name": "api.models.Course", "line_number": 10, "usage_type": "attribute"}, {"api_name": "api.models", "line_number": 10, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 14, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 14, "usage_type": "name"}, {"api_name": "rest_framework.serializers.SerializerMethodField", "line_number": 15, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 15, "usage_type": "name"}, {"api_name": "rest_framework.serializers.SerializerMethodField", "line_number": 16, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 16, "usage_type": "name"}, {"api_name": "api.models.DegreeCourse", "line_number": 19, "usage_type": "attribute"}, {"api_name": "api.models", "line_number": 19, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 31, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 31, "usage_type": "name"}, {"api_name": "rest_framework.serializers.SerializerMethodField", "line_number": 32, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 32, "usage_type": "name"}, {"api_name": "api.models.DegreeCourse", "line_number": 35, "usage_type": "attribute"}, {"api_name": "api.models", "line_number": 35, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 43, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 43, "usage_type": "name"}, {"api_name": "rest_framework.serializers.CharField", "line_number": 44, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 44, "usage_type": "name"}, {"api_name": "rest_framework.serializers.CharField", "line_number": 45, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 45, "usage_type": "name"}, {"api_name": "rest_framework.serializers.CharField", "line_number": 46, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 46, "usage_type": "name"}, {"api_name": "rest_framework.serializers.SerializerMethodField", "line_number": 47, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 47, "usage_type": "name"}, {"api_name": "api.models.Course", "line_number": 50, "usage_type": "attribute"}, {"api_name": "api.models", "line_number": 50, "usage_type": "name"}]}
+{"seq_id": "452185877", "text": "from functools import reduce\n\ndef str2int(m):\n L = {'0':0, '1':1, '2':2, '3':3, '4':4, '5':5, '6':6, '7':7, '8':8, '9':9}\n return L[m]\n\ndef add1(x,y):\n return x * 10 + y\n\ndef add2(x,y):\n return x * 0.1 + y\n\ndef str2float(s):\n myindex = s.index('.')\n s = s[0:myindex] +s[myindex+1:]\n s = list(map(str2int,s))\n s1 = reduce(add1,s[:myindex])\n s2 = s[myindex:]\n s3 = (reduce(add2,s2[::-1])) * 0.1\n s = s1 +s3\n return s\n\n\nprint('str2float(\\'123.456\\') =', str2float('123.456'))\nif abs(str2float('123.456') - 123.456) < 0.00001:\n print('测试成功!')\nelse:\n print('测试失败!')\n", "sub_path": "Part 1 Learning Python/do_reduce_map.py", "file_name": "do_reduce_map.py", "file_ext": "py", "file_size_in_byte": 619, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "functools.reduce", "line_number": 17, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 19, "usage_type": "call"}]}
+{"seq_id": "211845425", "text": "from django.shortcuts import render\nfrom .forms import Trains\nfrom .api_handler import getStation,getTrains\n\ndef find_trains(request):\n if request.method == 'POST':\n filled_form = Trains(request.POST)\n if filled_form.is_valid():\n src_stations = getStation(filled_form.cleaned_data['source'])\n dest_stations = getStation(filled_form.cleaned_data['dest'])\n button_text = 'Find Trains'\n find_train_form = Trains()\n station_select = False\n return render(request,'rail_booking/find_trains.html',{'form':find_train_form,'src_stations':src_stations,'dest_stations':dest_stations,'button_text':button_text,'station_select':station_select})\n else:\n find_train_form = Trains()\n station_select = True\n button_text = 'Find city stations'\n return render(request,'rail_booking/find_trains.html',{'form':find_train_form,'button_text':button_text,'station_select':station_select})\n\ndef get_trains(request):\n if request.method == 'POST':\n filled_form = Trains(request.POST)\n if filled_form.is_valid():\n print(\"Entered\")\n trains = getTrains(filled_form.cleaned_data['source'],filled_form.cleaned_data['dest'],'24/06/2017')\n return render(request,'rail_booking/get_trains.html',{'trains':trains})", "sub_path": "rail_booking/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1341, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "forms.Trains", "line_number": 7, "usage_type": "call"}, {"api_name": "api_handler.getStation", "line_number": 9, "usage_type": "call"}, {"api_name": "api_handler.getStation", "line_number": 10, "usage_type": "call"}, {"api_name": "forms.Trains", "line_number": 12, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 14, "usage_type": "call"}, {"api_name": "forms.Trains", "line_number": 16, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 19, "usage_type": "call"}, {"api_name": "forms.Trains", "line_number": 23, "usage_type": "call"}, {"api_name": "api_handler.getTrains", "line_number": 26, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 27, "usage_type": "call"}]}
+{"seq_id": "454416228", "text": "import numpy as np\nimport matplotlib.pyplot as plt \n\n\"\"\"\nBenjamin Klimko, PHYS 416 Spring 2018\nThe program has a user definable input hardcoded (N) and produces plots of starting value vs computed values and starting value\nvs iterations for the hailstone (3n+1) problem as performed in problem A of this problem set. \n\"\"\"\n\n# initialize the start vector s, computed number vector f, iteration vector g, input vector nums, and counter count\nN = 200\nnums = np.arange(1, N)\ns = []\nf = []\ng = np.zeros(N-1)\n\n# for each num in the input vector\nfor idx in nums:\n\tcount = 0\n\t#code borrowed from my chap1_problemA.py \n\t# make a copy of idx to keep track of what the starting number is\n\tn = int(np.copy(idx))\n\n\n\t# as long as n is not 1 continue the algorithm\n\twhile n != 1:\n\t\t# add the current value of n to vector f and the value of idx to vector s for each time through the while loop\n\t\ts.append(idx)\n\t\tf.append(n)\n\t\t# check if n is even or odd-- if even divide by 2; if odd multiply by 3 and add 1\n\t\tif n % 2 == 0:\n\t\t\tn /= 2\n\t\telse:\n\t\t\tn = (3*n) + 1\n\n\t\t# increment the count for each time through the while loop\n\t\tcount += 1\n\n\ts.append(idx)\n\tf.append(n)\n\tg[idx-1] = count\n\n# plot starting value vs computed value and starting value vs iterations\nplt.subplot(2, 1, 1)\nplt.plot(s, f, '.')\nstrfn = lambda x: str(x)\nvecfn = np.vectorize(strfn)\nplt.xticks(np.arange(1,N+1, 10), vecfn(np.arange(1,N+1, 10)))\nax = plt.gca()\nax.set_xlim(1, N)\nax.set_ylim(0, max(f))\nplt.tight_layout()\n\nplt.subplot(2, 1, 2)\nplt.plot(nums, g)\nplt.xticks(np.arange(1,N,20), vecfn(np.arange(1,N,20)))\nplt.yticks(np.arange(0,N,20), vecfn(np.arange(0,N,20)))\nax2 = plt.gca()\nax2.set_xlim(1, N)\nax2.set_ylim(0, max(g))\n\nplt.show()", "sub_path": "Chapter 1 Work/chap1_problemB.py", "file_name": "chap1_problemB.py", "file_ext": "py", "file_size_in_byte": 1693, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "numpy.arange", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "numpy.vectorize", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}]}
+{"seq_id": "619557426", "text": "# profile_analysis.py - Likelihood/Objective function\n# --------------------------------------------------\n# This file is a part of DeerLab. License is MIT (see LICENSE.md).\n# Copyright(c) 2019-2021: Luis Fabregas, Stefan Stoll and other contributors.\n\nimport numpy as np\nfrom scipy.stats import chi2 \nfrom deerlab import fit\nfrom deerlab import noiselevel,UQResult\nimport warnings \nfrom tqdm import tqdm\n\ndef profile_analysis(model,y, *args, parameters='all', samples=50, noiselvl=None, verbose=False,**kargs):\n r\"\"\" \n Profile likelihood analysis for uncertainty quantification\n\n Parameters\n ----------\n model : :ref:`Model`\n Model object describing the data. All non-linear model parameters are profiled by default.\n\n y : array_like or list of array_like\n Experimental dataset(s).\n\n args : positional arguments\n Any other positional arguments to be passed to the ``fit`` function. See the \n documentation of the ``fit`` function for further details. \n\n parameters : string or list thereof\n Model parameters to profile. If set to ``'all'`` all non-linear parameters in the model are analyzed. \n\n samples : integer scalar\n Number of points to take to estimate the profile function.\n\n noiselvl : float, optional\n Noise level(s) of the datasets. If set to ``None`` it is determined automatically.\n\n verbose : boolean, optional\n Specifies whether to print the progress of the bootstrap analysis on the \n command window, the default is false.\n\n kargs : keyword-argument pairs\n Any other keyword-argument pairs to be passed to the ``fit`` function. See the \n documentation of the ``fit`` function for further details. \n\n Returns\n -------\n profuq : :ref:`UQResult`\n Profile uncertainty quantification for each non-linear parameter in the model. \n \"\"\"\n\n if noiselvl is None: \n noiselvl = noiselevel(y)\n\n # Optimize the whole model to fit the data\n fitresult = fit(model, y, *args, **kargs)\n\n # Prepare the statistical threshold function\n threshold = lambda coverage: noiselvl**2*chi2.ppf(coverage, df=1) + fitresult.cost\n\n # Loop over all parameters in the model\n uqresults = {}\n if parameters=='all':\n parameters = model._parameter_list()\n elif not isinstance(parameters,list):\n parameters = [parameters]\n for parameter in parameters:\n if np.any(getattr(model,parameter).linear):\n uqresults[parameter] = None\n continue \n\n if verbose:\n tqdm.write(f'Profiling model parameter {parameter}:',end='')\n \n # Construct the values of the model parameter to profile\n start = np.maximum(getattr(model, parameter).lb, getattr(fitresult,parameter)-10*getattr(fitresult,f'{parameter}Uncert').std)\n stop = np.minimum(getattr(model, parameter).ub, getattr(fitresult,parameter)+10*getattr(fitresult,f'{parameter}Uncert').std)\n varvals = np.linspace(start,stop,samples)\n\n # Calculate the profile objective function for the parameter\n profile = np.zeros(samples)\n for n,value in enumerate(tqdm(varvals, disable=not verbose)): \n\n # Freeze the model parameter at current value\n getattr(model, parameter).freeze(value)\n\n # Optimize the rest\n with warnings.catch_warnings():\n warnings.simplefilter(\"ignore\")\n fitresult_ = fit(model, y, *args, **kargs)\n\n # Extract the objective function value\n profile[n] = fitresult_.cost\n\n # Unfreeze the parameter\n getattr(model, parameter).unfreeze()\n profile = {'x':np.squeeze(varvals),'y':profile}\n uqresults[parameter] = UQResult('profile', data=getattr(fitresult,parameter), profiles=profile, threshold=threshold, noiselvl=noiselvl)\n uqresults[parameter].profile = uqresults[parameter].profile[0]\n \n return uqresults\n", "sub_path": "deerlab/profile_analysis.py", "file_name": "profile_analysis.py", "file_ext": "py", "file_size_in_byte": 3965, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "deerlab.noiselevel", "line_number": 53, "usage_type": "call"}, {"api_name": "deerlab.fit", "line_number": 56, "usage_type": "call"}, {"api_name": "scipy.stats.chi2.ppf", "line_number": 59, "usage_type": "call"}, {"api_name": "scipy.stats.chi2", "line_number": 59, "usage_type": "name"}, {"api_name": "numpy.any", "line_number": 68, "usage_type": "call"}, {"api_name": "tqdm.tqdm.write", "line_number": 73, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 73, "usage_type": "name"}, {"api_name": "numpy.maximum", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 81, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 82, "usage_type": "call"}, {"api_name": "warnings.catch_warnings", "line_number": 88, "usage_type": "call"}, {"api_name": "warnings.simplefilter", "line_number": 89, "usage_type": "call"}, {"api_name": "deerlab.fit", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 97, "usage_type": "call"}, {"api_name": "deerlab.UQResult", "line_number": 98, "usage_type": "call"}]}
+{"seq_id": "5562459", "text": "import platform\nimport os\nimport logging\n\n\ndef setup_log_and_return_log_file():\n _log_file = ''\n platform_str = platform.platform()\n if platform_str.startswith('Linux'):\n _log_file = os.path.join('/media/leizhang/Data/workspace/', 'logs', 'my_log.log')\n\n logging.basicConfig(\n level=logging.DEBUG,\n format='%(asctime)s: %(levelname)s: %(message)s',\n filename=_log_file,\n filemode='+a',\n )\n logging.info('The log is written to:' + _log_file)\n return _log_file\n\n\nif __name__ == '__main__':\n logging.debug('this is a debug mode')\n logging.warning('Some problem may happen here')\n logging.error('A fatal error happened here')\n", "sub_path": "scripts/log_util.py", "file_name": "log_util.py", "file_ext": "py", "file_size_in_byte": 691, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "platform.platform", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 12, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 13, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 18, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 23, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 24, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 25, "usage_type": "call"}]}
+{"seq_id": "173493296", "text": "# ///\n# This command is used to query the API, compare the results with the DB and make appropriate changes.\n# \\\\\\\nimport logging\nfrom datetime import datetime\n\nfrom django.core import mail\nfrom django.core.management.base import BaseCommand\n\nfrom develop.management.commands.actions import *\nfrom develop.management.commands.emails import *\nfrom develop.models import *\n\nitems_with_changes = []\nnew_items_added = []\n\nlogger = logging.getLogger(\"django\")\n\n\nclass Command(BaseCommand):\n def handle(self, *args, **options):\n # Lets turn this into a model and put it in the database later ******\n datasets = [[\"Development\", \"https://data.raleighnc.gov/resource/ncz3-s64h.json?$limit=10000&$$app_token=j6m9w6X5dCOGZVoBMBxxte4kN\"],\n [\"Zoning\", \"https://data.raleighnc.gov/resource/wz64-sgmw.json?zpyear=2017&$$app_token=j6m9w6X5dCOGZVoBMBxxte4kN\"]]\n\n for dataset in datasets:\n # Development specific section\n if dataset[0] == \"Development\":\n all_dev_json = get_api_json(dataset[1])\n\n for dev_json in all_dev_json:\n item_number = int(dev_json[\"devplan_id\"])\n item_year = int(dev_json[\"submitted_yr\"])\n\n # If we already know about this item and there is at least one dev CI\n if Development.objects.filter(devplan_id=item_number, submitted_yr=item_year).exists():\n development = Development.objects.get(devplan_id=item_number, submitted_yr=item_year)\n devs_cis = Development_change_instance.objects.filter(development=development)\n\n if devs_cis.exists():\n devs_latest_ci = devs_cis.latest('date_created')\n\n # If a difference exists between the known development and the json response\n if item_scan_compare(dev_json, development, devs_latest_ci):\n create_ci(development, dev_json, initial=False)\n items_with_changes.append(development)\n\n # Account for the very unlikely scenario that a Development is created without any CIs\n # Typically when a user deletes a CI leaving a Development without any\n else:\n create_ci(development, dev_json, initial=True)\n new_items_added.append(development)\n\n # else create a new dev plan with initial change instance\n else:\n new_dev = Development(devplan_id=item_number, submitted_yr=item_year)\n new_dev.save()\n create_ci(new_dev, dev_json, initial=True)\n new_items_added.append(new_dev)\n\n # Zoning specific section\n elif dataset[0] == \"Zoning\":\n all_zon_json = get_api_json(dataset[1])\n\n for zon_json in all_zon_json:\n item_number = int(zon_json[\"zpnum\"])\n item_year = int(zon_json[\"zpyear\"])\n\n # If we already know about this item\n if Zoning.objects.filter(zpnum=item_number, zpyear=item_year).exists():\n zoning_case = Zoning.objects.get(zpnum=item_number, zpyear=item_year)\n\n try:\n zons_latest_ci = Zoning_change_instance.objects.filter(zoning_case=zoning_case).latest('date_created')\n except:\n n = datetime.datetime.now()\n logger.info(\"Problem with \" + str(zoning_case) + \" at \" + n.strftime(\"%H:%M %m-%d-%y\"))\n\n # If a difference exists between the known zoning case and the json response\n if item_scan_compare(zon_json, zoning_case, zons_latest_ci):\n create_ci(zoning_case, zon_json, initial=False)\n items_with_changes.append(zoning_case)\n\n # else create a new zoning case with initial change instance\n else:\n new_zon = Zoning(zpnum=item_number, zpyear=item_year)\n new_zon.save()\n create_ci(new_zon, zon_json, initial=True)\n new_items_added.append(new_zon)\n\n # Create our EmailMessage objects\n active_subscribers = Subscriber.objects.filter(send_emails=True)\n\n if (new_items_added or items_with_changes) and active_subscribers:\n connection = mail.get_connection()\n messages = get_emails(new_items_added, items_with_changes)\n connection.send_messages(messages)\n n = datetime.datetime.now()\n logger.info(\"Email sent to: \" + str(messages[0].recipients()) + \" at \" + n.strftime(\"%H:%M %m-%d-%y\"))\n", "sub_path": "develop/management/commands/scan.py", "file_name": "scan.py", "file_ext": "py", "file_size_in_byte": 4943, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.getLogger", "line_number": 17, "usage_type": "call"}, {"api_name": "django.core.management.base.BaseCommand", "line_number": 20, "usage_type": "name"}, {"api_name": "datetime.datetime.datetime.now", "line_number": 76, "usage_type": "call"}, {"api_name": "datetime.datetime.datetime", "line_number": 76, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 76, "usage_type": "name"}, {"api_name": "django.core.mail.get_connection", "line_number": 95, "usage_type": "call"}, {"api_name": "django.core.mail", "line_number": 95, "usage_type": "name"}, {"api_name": "datetime.datetime.datetime.now", "line_number": 98, "usage_type": "call"}, {"api_name": "datetime.datetime.datetime", "line_number": 98, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 98, "usage_type": "name"}]}
+{"seq_id": "138837877", "text": "import requests\nfrom datetime import datetime\nimport matplotlib.pyplot as plt\n\n\nclass AlphaPlot:\n def __init__(self, ticker, api_key):\n self.ticker = ticker\n self.key = api_key\n\n def get_json(self):\n api_site = \"https://www.alphavantage.co/query\"\n data = {\n \"function\": \"TIME_SERIES_MONTHLY_ADJUSTED\",\n \"symbol\": self.ticker,\n \"datatype\": \"json\",\n \"apikey\": self.key\n }\n response = requests.get(api_site, params=data)\n json_data = response.json()\n\n return json_data\n\n def create_tuples(self, json):\n del json['Meta Data']\n\n pairs = []\n # Create tuples\n for k, v in json.items():\n for k2, v2 in v.items():\n pairs.append([k2[:-3], float(v2.get(\"7. dividend amount\")),\n float(v2.get(\"4. close\"))])\n\n # Remove dividend of zero\n prune = [i for i in pairs if i[1] != '0.0000']\n\n # Change date format, remove unnecessary padding zeros\n for idx, elem in enumerate(prune):\n sp = elem[0].split(\"-\")\n con = sp[1] + \"/\" + sp[0]\n prune[idx][0] = con\n\n # Sort by date\n prune.sort(key=lambda L: datetime.strptime(L[0], '%m/%Y'))\n\n # Sum quarterly dividend reports\n stop = len(prune)\n idx = 0\n for elem in range(len(prune) - 1):\n if prune[idx][0][3:] == prune[idx+1][0][3:] and idx < stop:\n prune[idx+1][1] += prune[idx][1]\n del prune[idx]\n stop -= 1\n else:\n idx += 1\n\n # Round dividend and price to 2 decimal places\n for i, div in enumerate(prune):\n prune[i][1] = \"%03.2f\" % round(prune[i][1], 2)\n prune[i][2] = \"%03.2f\" % round(prune[i][2], 2)\n\n return prune\n\n def plot_figure(self, data):\n # Data to lists\n dates = [x[0] for x in data]\n div = [float(x[1]) for x in data]\n close = [float(x[2]) for x in data]\n\n fig = plt.figure()\n fig.set_size_inches(16, 7)\n\n # ax is to encompass other two subplots\n ax = fig.add_subplot(111)\n ax1 = fig.add_subplot(211)\n ax2 = fig.add_subplot(212)\n\n # turn off big subplot\n ax.spines['top'].set_color('none')\n ax.spines['bottom'].set_color('none')\n ax.spines['left'].set_color('none')\n ax.spines['right'].set_color('none')\n ax.tick_params(labelcolor='w', top=False, bottom=False, left=False,\n right=False)\n\n # Set shared y-axis label\n ax.tick_params(axis='y', which='major', pad=15)\n ax.set_ylabel('Amount (USD)')\n\n # Share x-axis label\n ax1.get_shared_x_axes().join(ax1, ax2)\n ax1.set_xticklabels([])\n\n # plot data\n ax1.plot(dates, div, 'xkcd:aqua', marker='o', alpha=.5)\n ax2.plot(dates, close, 'xkcd:tomato', marker='o', alpha=.5)\n\n ax1.set_title(\"Dividends\")\n ax2.set_title(\"Price\")\n\n plt.xticks(rotation=35, fontsize=8, ha=\"right\")\n plt.suptitle(self.ticker.upper(), x=0.51, y=.97, fontsize=18)\n\n ax1.spines['right'].set_color('none')\n ax1.spines['top'].set_color('none')\n ax2.spines['right'].set_color('none')\n ax2.spines['top'].set_color('none')\n\n # Label plot\n n = max(div)\n m = max(close)\n for i, txt in enumerate(div):\n if div[i] != n:\n ax1.annotate(txt, (dates[i], div[i]+.05), ha=\"center\",\n weight=\"bold\", fontsize=9)\n else:\n ax1.annotate(txt, (dates[i], div[i]-.1), ha=\"center\",\n weight=\"bold\", fontsize=9)\n\n for i, txt in enumerate(close):\n if close[i] != m:\n ax2.annotate(txt, (dates[i], close[i]+1), ha=\"center\",\n weight=\"bold\", fontsize=9)\n else:\n ax2.annotate(txt, (dates[i], close[i]-3), ha=\"center\",\n weight=\"bold\", fontsize=9)\n\n ax1.grid(linestyle='--')\n ax2.grid(linestyle='--')\n\n # For main.py\n # plt.show()\n\n return fig\n", "sub_path": "avplot.py", "file_name": "avplot.py", "file_ext": "py", "file_size_in_byte": 4227, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "requests.get", "line_number": 19, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 44, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.suptitle", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}]}
+{"seq_id": "288618039", "text": "import json\n\nfrom django.http import JsonResponse\nfrom django.shortcuts import render\n\nfrom tools.list_element_compare_size import list_element_compare_size\nfrom tools.list_element_type_change import list_element_type_change\nfrom tools.str_division import str_division\nfrom tools.user_put_verification import user_put_verification\n\n\ndef get_index(request):\n return render(request, 'index.html')\n\n\ndef version_compare(request):\n json_obj = json.loads(request.body)\n num1 = json_obj.get('num1')\n num2 = json_obj.get('num2')\n print(num1, num2)\n\n # 校验输入\n verification_res = user_put_verification(num1, num2)\n print(verification_res)\n if verification_res:\n return JsonResponse({'code': 1000, 'error': '版本号必须是阿拉伯数字,按字符.分割,且版本字符串不以点开始或结束,并且其中不能有两个连续的点'})\n\n # 按 . 分割\n division_res = str_division(num1, num2)\n print(division_res)\n if len(division_res[0]) > 4 or len(division_res[1]) > 4:\n return JsonResponse({'code': 1001, 'error': '请输入最多4级版本号,按字符.分割'})\n\n # 列表每个元素类型转换\n version1 = list_element_type_change(division_res[0], division_res[1])[0]\n version2 = list_element_type_change(division_res[0], division_res[1])[1]\n print(version1, version2)\n\n # 比较本版号大小\n res = list_element_compare_size(version1, version2)\n print(res)\n\n return JsonResponse({'code': 200, 'msg': res})\n", "sub_path": "version_number_compare/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1507, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.shortcuts.render", "line_number": 13, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 17, "usage_type": "call"}, {"api_name": "tools.user_put_verification.user_put_verification", "line_number": 23, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 26, "usage_type": "call"}, {"api_name": "tools.str_division.str_division", "line_number": 29, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 32, "usage_type": "call"}, {"api_name": "tools.list_element_type_change.list_element_type_change", "line_number": 35, "usage_type": "call"}, {"api_name": "tools.list_element_type_change.list_element_type_change", "line_number": 36, "usage_type": "call"}, {"api_name": "tools.list_element_compare_size.list_element_compare_size", "line_number": 40, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 43, "usage_type": "call"}]}
+{"seq_id": "24260198", "text": "import requests\n\nclass Mixlr:\n\tserviceId = 1\n\t\n\tapi = 'https://api.mixlr.com/users/'\n\t\n\tdef __init__(self, name):\n\t\tself.name = name\n\t\tself.isLive = False\n\t\tself.link = 'http://mixlr.com/' + self.name\n\t\tself.service = 'Mixlr'\n\n\tdef checkLive(self):\n\t\turl = Mixlr.api + self.name\n\t\tresp = requests.get(url)\n\t\t\n\t\tif resp.status_code == 200:\n\t\t\tresult = resp.json()\n\t\t\tself.isLive = result['is_live']\n\t\telse:\n\t\t\tself.isLive = False\n\n\t\treturn self.isLive\n\t\n\tdef __lt__(self, other):\n\t\tif (self.serviceId == other.serviceId):\n\t\t\treturn self.name < other.name\n\t\telse:\n\t\t\treturn self.serviceId < other.serviceId\n\nclass Twitch:\n\tserviceId = 0\n\t\n\tapi = 'https://api.twitch.tv/kraken/'\n\theaders = { 'Accept':'application/vnd.twitchtv.v3+json', 'Client-ID':'ewvlchtxgqq88ru9gmfp1gmyt6h2b93' }\n\n\tdef __init__(self, name):\n\t\tself.name = name\n\t\tself.isLive = False\n\t\tself.link = 'https://player.twitch.tv/?branding=false&showInfo=false&channel=' + self.name\n\t\tself.service = 'Twitch'\n\n\tdef checkLive(self):\n\t\turl = Twitch.api + 'streams/' + self.name\n\t\tresp = requests.get(url, headers=Twitch.headers)\n\n\t\tif resp.status_code == 200:\n\t\t\tresult = resp.json()\n\t\t\tif result['stream']:\n\t\t\t\tself.isLive = True\n\t\t\telse:\n\t\t\t\tself.isLive = False\n\t\telse:\n\t\t\tself.isLive = False\n\n\t\treturn self.isLive\n\t\t\n\tdef __lt__(self, other):\n\t\tif (self.serviceId == other.serviceId):\n\t\t\treturn self.name < other.name\n\t\telse:\n\t\t\treturn self.serviceId < other.serviceId\n", "sub_path": "services.py", "file_name": "services.py", "file_ext": "py", "file_size_in_byte": 1431, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "requests.get", "line_number": 16, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 46, "usage_type": "call"}]}
+{"seq_id": "127718793", "text": "import os\nimport dpkt\nimport socket\nimport datetime\nimport pandas as pd\nimport numpy as np\nimport ipaddress\n\n\nclass LoadPcap:\n\n\n def __init__(self, file_name, query_start, period, network_segment):\n \n self.file_name = file_name\n self.query_start = query_start\n self.period = period\n self.network_segment = network_segment\n\n\n def arpProcess(self, eth):\n arp = eth.arp\n src_ip = socket.inet_ntoa(arp.spa)\n src_port = np.nan\n dst_ip = socket.inet_ntoa(arp.tpa)\n dst_port = np.nan\n \n return [src_ip, src_port, dst_ip, dst_port]\n\n\n def normalProcess(self, eth):\n ip = eth.data\n if ip.v == 4: \n src_ip = socket.inet_ntoa(ip.src)\n dst_ip = socket.inet_ntoa(ip.dst)\n \n elif ip.v == 6:\n src_ip = socket.inet_ntop(socket.AF_INET6, ip.src)\n dst_ip = socket.inet_ntop(socket.AF_INET6, ip.dst)\n try:\n src_ip = IPNetwork(src_ip).ipv4()\n except:\n pass\n try:\n dst_ip = IPNetwork(dst_ip).ipv4()\n except:\n pass\n src_ip = str(src_ip).split('/')[0]\n dst_ip = str(dst_ip).split('/')[0]\n \n try:\n src_port = ip.data.sport\n dst_port = ip.data.dport\n except:\n src_port = np.nan\n dst_port = np.nan\n \n return [src_ip, src_port, dst_ip, dst_port]\n\n\n def pcapProcess(self, pcap):\n pcap_list = list()\n count = 0\n for ts, buf in pcap:\n\n count += 1\n if count % 200000 == 0:\n print(count/10000)\n \n time = datetime.datetime.fromtimestamp(ts).strftime(\"%Y-%m-%dT%H:%M:%S.%f+08:00\")\n \n try:\n eth = dpkt.ethernet.Ethernet(buf)\n except:\n print('pass')\n \n try:\n # 2054 是 ARP\n if eth.type == 2054:\n pcap_list.append([time]+self.arpProcess(eth))\n\n else:\n pcap_list.append([time]+self.normalProcess(eth))\n\n except:\n pass\n\n return pd.DataFrame(pcap_list, columns=[\"Time\",\"Source\",\"SrcPort\",\"Destination\",\"DstPort\"]) \n\n\n def readPcap(self, filename):\n df = pd.DataFrame()\n for f in filename:\n file = open(\"data/\" + f, \"rb\")\n pcap = dpkt.pcap.Reader(file)\n df_pcap = self.pcapProcess(pcap)\n df = pd.concat([df, df_pcap])\n \n # print(df)\n return df\n\n\n def timeFilter(self, df, start_time, end_time):\n # df過濾時間\n df_time = df.loc[(df.Time >= start_time) & (df.Time < end_time)]\n\n return df_time\n\n '''\n def dataPreprocess(self, df):\n\n # 這裡最後要改成撈內網全部的流量或改成可以動態調整要全內網或部分網段。\n # # 過濾屬於內網ip的流量\n # group = df.groupby([\"Source\", \"Destination\"]).size().reset_index(name='Count')\n \n # group['src_private'] = group['Source'].apply(lambda x: ipaddress.ip_address(x).is_private)\n # group['dst_private'] = group['Destination'].apply(lambda x: ipaddress.ip_address(x).is_private)\n # group = group.loc[(group['src_private']) & (group['dst_private'])]\n # group.drop(['src_private', 'dst_private'], axis=1, inplace=True)\n # df_omp = group.reset_index(drop=True)\n\n network_segment = self.network_segment.split(',')\n \n # 將相同[\"Srcip\", \"Dstip\"] 的groupby 在一起\n group = dict(df.groupby([\"Source\", \"Destination\"]).size())\n \n # 過濾出內網(77網段下)的流量\n src_dst = []\n for k in group.keys():\n\n segment_s = False\n segment_d = False\n \n for i in network_segment:\n if k[0].find(i) == 0:\n segment_s = True\n if k[1].find(i) == 0:\n segment_d = True\n \n if segment_s == True and segment_d == True:\n src_dst.append([k[0], k[1], group[k]])\n \n # 再次組成dataframe\n df_omp = pd.DataFrame(src_dst, columns = [\"Source\", \"Destination\", \"Count\"])\n\n\n return df_omp\n '''\n\n \n def dataPreprocess(self, df):\n\n # 這裡最後要改成撈內網全部的流量或改成可以動態調整要全內網或部分網段。\n # 過濾屬於內網ip的流量\n group = df.groupby([\"Source\", \"Destination\"]).size().reset_index(name='Count')\n \n group['src_private'] = group['Source'].apply(lambda x: ipaddress.ip_address(x).is_private)\n group['dst_private'] = group['Destination'].apply(lambda x: ipaddress.ip_address(x).is_private)\n group = group.loc[(group['src_private']) & (group['dst_private'])]\n group.drop(['src_private', 'dst_private'], axis=1, inplace=True)\n df_omp = group.reset_index(drop=True)\n\n\n # network_segment = self.network_segment.split(',')\n \n # # 將相同[\"Srcip\", \"Dstip\"] 的groupby 在一起\n # group = dict(df.groupby([\"Source\", \"Destination\"]).size())\n \n # # 過濾出內網(77網段下)的流量\n # src_dst = []\n # for k in group.keys():\n\n # segment_s = False\n # segment_d = False\n \n # for i in network_segment:\n # if k[0].find(i) == 0:\n # segment_s = True\n # if k[1].find(i) == 0:\n # segment_d = True\n \n # if segment_s == True and segment_d == True:\n # src_dst.append([k[0], k[1], group[k]])\n \n # # 再次組成dataframe\n # df_omp = pd.DataFrame(src_dst, columns = [\"Source\", \"Destination\", \"Count\"])\n\n\n return df_omp\n \n \n\n def run(self):\n print(self.query_start) \n print(\"========== load data ==========\")\n \n \n # filename\n filename = []\n # file_time_now = datetime.datetime.now()\n # file_time = file_time_now\n file_time = datetime.datetime.strptime(self.query_start, '%Y-%m-%dT%H:%M:%S.%f')\n file_time = file_time - datetime.timedelta(hours = 1)\n file_cnt = self.period + 1\n for _ in range(file_cnt):\n linux_filename = file_time - datetime.timedelta(hours = 8)\n filename.append(linux_filename.strftime('%m-%d-%H') + '.pcap')\n file_time = file_time + datetime.timedelta(hours = 1)\n\n print('filename: ' + str(filename))\n \n df = self.readPcap(filename)\n\n # df = self.readPcap(self.file_name)\n\n start_time = datetime.datetime.strptime(self.query_start, '%Y-%m-%dT%H:%M:%S.%f')\n end_time = start_time + datetime.timedelta(hours = self.period)\n now_end = end_time.strftime(\"%Y-%m-%dT%H:%M:%S.%f\") \n\n abnor_df = self.timeFilter(df, self.query_start, now_end)\n abnor_data = self.dataPreprocess(abnor_df)\n # print(abnor_data)\n abnor_iplist = list(abnor_data['Source']) + list(abnor_data['Destination'])\n abnor_ip_list = sorted(list(set(abnor_iplist)))\n\n # 控制多少比例的 IP 數量為上班/下班\n # 怎麼知道公司總共有多少 ip\n work_ip_cnt = 254 * 0.29527\n\n if len(abnor_ip_list) > work_ip_cnt:\n week_threshold = 'weekday'\n else:\n week_threshold = 'weekend'\n\n\n if int(self.query_start[11:13]) >= 8 and int(self.query_start[11:13]) < 20:\n work_break = 'work'\n elif int(self.query_start[11:13]) < 8 or int(self.query_start[11:13]) >= 20:\n work_break = 'break'\n\n\n folder_name = './' + week_threshold + '_' + work_break + '/'\n filepath = folder_name + 'omp_and_frequency.csv'\n print('folder name: ' + week_threshold + '_' + work_break)\n\n if os.path.isfile(filepath):\n # print('no empty')\n\n nor_data = []\n\n else:\n print('first time')\n\n\n # filename\n filename = []\n file_time = datetime.datetime.strptime(self.query_start, '%Y-%m-%dT%H:%M:%S.%f') - datetime.timedelta(hours = 24)\n file_time = file_time - datetime.timedelta(hours = 1)\n file_cnt = self.period + 1\n for _ in range(file_cnt):\n linux_filename = file_time - datetime.timedelta(hours = 8)\n filename.append(linux_filename.strftime('%m-%d-%H') + '.pcap')\n file_time = file_time + datetime.timedelta(hours = 1)\n \n print('normal filename: ' + str(filename))\n \n df = self.readPcap(filename)\n\n # 以前一個小時為比較時間 V\n # 與前一天同一個時間區間比較 X\n nor_start_time = start_time - datetime.timedelta(hours = 1)\n nor_query_start = nor_start_time.strftime(\"%Y-%m-%dT%H:%M:%S.%f\")\n nor_end_time = end_time - datetime.timedelta(hours = 1)\n nor_end_time = nor_end_time.strftime(\"%Y-%m-%dT%H:%M:%S.%f\")\n\n nor_df = self.timeFilter(df, nor_query_start, nor_end_time)\n nor_data = self.dataPreprocess(nor_df)\n # print(nor_data)\n \n print('start time: ' + str(self.query_start)[:19])\n\n return abnor_df, nor_data, abnor_data, folder_name\n\n", "sub_path": "SecBuzzerESM/AI/OMP/code/app/load_pcap.py", "file_name": "load_pcap.py", "file_ext": "py", "file_size_in_byte": 9561, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "socket.inet_ntoa", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 24, "usage_type": "attribute"}, {"api_name": "socket.inet_ntoa", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 26, "usage_type": "attribute"}, {"api_name": "socket.inet_ntoa", "line_number": 34, "usage_type": "call"}, {"api_name": "socket.inet_ntoa", "line_number": 35, "usage_type": "call"}, {"api_name": "socket.inet_ntop", "line_number": 38, "usage_type": "call"}, {"api_name": "socket.AF_INET6", "line_number": 38, "usage_type": "attribute"}, {"api_name": "socket.inet_ntop", "line_number": 39, "usage_type": "call"}, {"api_name": "socket.AF_INET6", "line_number": 39, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 55, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 56, "usage_type": "attribute"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 70, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 70, "usage_type": "attribute"}, {"api_name": "dpkt.ethernet.Ethernet", "line_number": 73, "usage_type": "call"}, {"api_name": "dpkt.ethernet", "line_number": 73, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 88, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 92, "usage_type": "call"}, {"api_name": "dpkt.pcap.Reader", "line_number": 95, "usage_type": "call"}, {"api_name": "dpkt.pcap", "line_number": 95, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 97, "usage_type": "call"}, {"api_name": "ipaddress.ip_address", "line_number": 157, "usage_type": "call"}, {"api_name": "ipaddress.ip_address", "line_number": 158, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 202, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 202, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 203, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 206, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 208, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 216, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 216, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 217, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 246, "usage_type": "call"}, {"api_name": "os.path", "line_number": 246, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 257, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 257, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 257, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 258, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 261, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 263, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 271, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 273, "usage_type": "call"}]}
+{"seq_id": "492319966", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations, models\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ]\n\n operations = [\n migrations.CreateModel(\n name='HomepageLine',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('line', models.TextField(verbose_name='Translatable line')),\n ('context', models.TextField(verbose_name='Optional context for translators', blank=True)),\n ('number', models.PositiveSmallIntegerField(unique=True, verbose_name='The token number of the line')),\n ],\n options={\n 'ordering': ('number',),\n 'verbose_name': 'Homepage line',\n 'verbose_name_plural': 'Homepage lines',\n },\n ),\n migrations.CreateModel(\n name='HomepageText',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('full_text', models.TextField(verbose_name='Translatable text')),\n ],\n options={\n 'ordering': ('id',),\n 'verbose_name': 'Homepage text',\n },\n ),\n ]\n", "sub_path": "projects/yearcompass_homepage_2015_2016/migrations/0001_initial.py", "file_name": "0001_initial.py", "file_ext": "py", "file_size_in_byte": 1351, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.PositiveSmallIntegerField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 27, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 27, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 31, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 31, "usage_type": "name"}]}
+{"seq_id": "461515464", "text": "\"\"\"\nThis component provides support for HP Printers.\nFor more details about this component, please refer to the documentation at\nhttps://home-assistant.io/components/hpprinter/\n\"\"\"\nimport voluptuous as vol\n\nfrom homeassistant.const import (CONF_HOST, CONF_NAME, CONF_DEVICES)\nfrom homeassistant.helpers import config_validation as cv\n\nfrom .const import *\nfrom .HPDeviceData import *\nfrom .home_assistant import HPPrinterHomeAssistant\n\n_LOGGER = logging.getLogger(__name__)\n\nDEVICE_SCHEMA = vol.Schema({\n vol.Required(CONF_HOST): cv.string,\n vol.Optional(CONF_NAME): cv.string\n})\n\nCONFIG_SCHEMA = vol.Schema({\n DOMAIN: vol.Schema({\n vol.Required(CONF_DEVICES, []): vol.All(cv.ensure_list, [vol.Any(DEVICE_SCHEMA)])\n }),\n}, extra=vol.ALLOW_EXTRA)\n\n\ndef setup(hass, config):\n \"\"\"Set up a Blue Iris component.\"\"\"\n _LOGGER.debug(f\"Loading HP Printer domain\")\n\n initialized = False\n devices_handlers = []\n\n conf = config[DOMAIN]\n scan_interval = SCAN_INTERVAL\n devices = conf.get(CONF_DEVICES, [])\n\n device_id = 0\n\n for device in devices:\n try:\n device_id = device_id + 1\n\n host = device.get(CONF_HOST)\n name = device.get(CONF_NAME, f\"{DEFAULT_NAME} #{device_id}\")\n hp_data = None\n\n if host is not None:\n hp_data = HPDeviceData(host, name)\n\n ha = HPPrinterHomeAssistant(hass, scan_interval, name, hp_data)\n\n if host is not None:\n ha.initialize()\n\n devices_handlers.append(ha)\n\n _LOGGER.debug(f\"{name} is loaded\")\n initialized = True\n\n else:\n ha.notify_error_message(f\"{name} was not configured correctly\")\n\n except Exception as ex:\n exc_type, exc_obj, tb = sys.exc_info()\n line_number = tb.tb_lineno\n\n _LOGGER.error(f\"Failed to create HA component, Error: {ex}, Line: {line_number}\")\n\n hass.data[DATA_HP_PRINTER] = devices_handlers\n\n return initialized\n", "sub_path": "custom_components/hpprinter/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 2025, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "voluptuous.Schema", "line_number": 17, "usage_type": "call"}, {"api_name": "voluptuous.Required", "line_number": 18, "usage_type": "call"}, {"api_name": "homeassistant.const.CONF_HOST", "line_number": 18, "usage_type": "argument"}, {"api_name": "voluptuous.Optional", "line_number": 19, "usage_type": "call"}, {"api_name": "homeassistant.const.CONF_NAME", "line_number": 19, "usage_type": "argument"}, {"api_name": "homeassistant.helpers.config_validation.string", "line_number": 18, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 18, "usage_type": "name"}, {"api_name": "homeassistant.helpers.config_validation.string", "line_number": 19, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 19, "usage_type": "name"}, {"api_name": "voluptuous.Schema", "line_number": 22, "usage_type": "call"}, {"api_name": "voluptuous.Schema", "line_number": 23, "usage_type": "call"}, {"api_name": "voluptuous.Required", "line_number": 24, "usage_type": "call"}, {"api_name": "homeassistant.const.CONF_DEVICES", "line_number": 24, "usage_type": "argument"}, {"api_name": "voluptuous.All", "line_number": 24, "usage_type": "call"}, {"api_name": "homeassistant.helpers.config_validation.ensure_list", "line_number": 24, "usage_type": "attribute"}, {"api_name": "homeassistant.helpers.config_validation", "line_number": 24, "usage_type": "name"}, {"api_name": "voluptuous.Any", "line_number": 24, "usage_type": "call"}, {"api_name": "voluptuous.ALLOW_EXTRA", "line_number": 26, "usage_type": "attribute"}, {"api_name": "homeassistant.const.CONF_DEVICES", "line_number": 38, "usage_type": "argument"}, {"api_name": "homeassistant.const.CONF_HOST", "line_number": 46, "usage_type": "argument"}, {"api_name": "homeassistant.const.CONF_NAME", "line_number": 47, "usage_type": "argument"}, {"api_name": "home_assistant.HPPrinterHomeAssistant", "line_number": 53, "usage_type": "call"}]}
+{"seq_id": "102754422", "text": "import os\nfrom graphviz import Digraph\nfrom tfcore.interfaces.IModel import IModel, IModel_Params\nfrom tfcore.core.layer import *\nfrom tfcore.core.activations import *\nfrom tfcore.core.loss import *\nfrom tfcore.utilities.params_serializer import ParamsSerializer\nfrom tfcore.utilities.utils import reduce_std\nfrom cnn.node import Node, Node_Params\nfrom cnn.cell import Cell\n\n\nclass Architecture(ParamsSerializer):\n\n def __init__(self, node_params=[]):\n\n self.node_params = node_params\n\n def load(self, path):\n \"\"\" Load Parameter\n\n # Arguments\n path: Path of json-file\n # Return\n Parameter class\n \"\"\"\n super().load(os.path.join(path))\n for list in range(len(self.node_params)):\n for params in range(len(self.node_params[list])):\n node_params = Node_Params()\n node_params.set_params(self.node_params[list][params])\n self.node_params[list][params] = node_params\n\n def save(self, path):\n \"\"\" Save parameter as json-file\n\n # Arguments\n path: Path to save\n \"\"\"\n if not os.path.exists(path):\n os.makedirs(path)\n super().save(os.path.join(path))\n return\n\nclass Network_Params(IModel_Params):\n \"\"\"\n Parameter class for ExampleModel\n \"\"\"\n\n def __init__(self,\n f_start=32,\n activation='relu',\n normalization='IN',\n multiplier=1.750,\n load_model=False,\n L2_weight=3e-4,\n model_name='',\n scope='Classifier',\n name='Classifier'):\n super().__init__(scope=scope, name=name)\n\n self.load_model = load_model\n self.model_name = model_name\n self.f_start = f_start\n self.activation = activation\n self.normalization = normalization\n self.L2_weight = L2_weight\n self.path = os.path.realpath(__file__)\n self.multiplier = multiplier\n\nclass Network(IModel):\n \"\"\"\n Example of a simple 3 layer generator model for super-resolution\n \"\"\"\n\n def __init__(self, sess, params, global_steps, is_training, is_eval):\n \"\"\"\n Init of Example Class\n\n # Arguments\n sess: Tensorflow-Session\n params: Instance of ExampleModel_Params\n global_steps: Globel steps for optimizer\n is_training: placeholder variable for switch between training end eval phase\n \"\"\"\n super().__init__(sess, params, global_steps)\n self.model_name = self.params.name\n self.activation = get_activation(name='relu')\n self.normalization = get_normalization(self.params.normalization)\n self.is_training = is_training\n self.is_eval = is_eval\n\n def build_model(self, input, is_train=False, reuse=False):\n \"\"\"\n Build model and create summary\n\n # Arguments\n input: Input-Tensor\n is_train: Bool\n reuse: Bool\n\n # Return\n Tensor of dimension 4D\n \"\"\"\n self.reuse = reuse\n if not self.params.load_model:\n super().build_model(input, is_train, reuse)\n else:\n self.load_model(input, self.params.model_name)\n\n return self.probs\n\n def model(self, net, is_train=False, reuse=False):\n \"\"\"\n Create generator model\n\n # Arguments\n input: Input-Tensor\n is_train: Bool\n reuse: Bool\n\n # Return\n Tensor of dimension 4D\n \"\"\"\n\n f_out = self.params.f_start\n with tf.variable_scope(self.params.scope, reuse=tf.AUTO_REUSE):\n\n cell_prev_prev = Node(net=net,\n f_out=f_out,\n stride=1,\n func_name='conv3x3',\n cell_id=0,\n layer=0,\n type='N',\n normalization=self.params.normalization,\n L2_weight=self.params.L2_weight,\n is_training=self.is_training,\n name='input-2')\n\n cell_prev = Node(net=net,\n f_out=f_out,\n stride=1,\n func_name='conv3x3',\n cell_id=0,\n layer=0,\n type='N',\n normalization=self.params.normalization,\n L2_weight=self.params.L2_weight,\n is_training=self.is_training,\n name='input-1')\n\n cell_types = ['N', 'R', 'N', 'R', 'N', 'R']\n self.nodes = [cell_prev_prev, cell_prev]\n for cell_id in range(len(cell_types)):\n if cell_types[cell_id] == 'R':\n f_out *= 2\n\n cell = Cell(layer=3,\n cell_prev_prev=self.nodes[-2],\n cell_prev=self.nodes[-1],\n f_out=f_out,\n type=cell_types[cell_id],\n cell_id=cell_id,\n activation=self.params.activation,\n normalization=self.params.normalization,\n L2_weight=self.params.L2_weight,\n is_training=self.is_training,\n multiplier=self.params.multiplier,\n summary_val=self.summary_val,\n summary_val_2=self.summary_val_2,\n summary_vis=self.summary_vis,\n summary_vis_2=self.summary_vis_2)\n self.nodes.append(cell.nodes[-1][-1])\n\n net = conv2d(self.nodes[-1].features,\n k_size=1,\n f_out=2,\n stride=1,\n activation=self.params.activation,\n normalization=self.params.normalization,\n use_pre_activation=True,\n L2_weight=self.params.L2_weight,\n is_training=self.is_training,\n name='conv_GAP')\n net = avg_pool(net, radius=net.shape[1], stride=1, padding='VALID', name='GAP')\n net = tf.reduce_mean(net, axis=[1, 2])\n\n self.logits = net\n self.probs = tf.nn.softmax(net)\n\n print(' [*] DARTS loaded...')\n return self.logits\n\n def load_model(self, net, model_name):\n\n architecture = Architecture()\n architecture.load(model_name)\n\n def find_params(name, params_list):\n for list in params_list:\n for node_params in list:\n if node_params.name_in_graph == name:\n return node_params\n return None\n\n def find_node(node_name, node_list):\n for node in node_list:\n if node_name == node.params.name_in_graph:\n return node\n return None\n\n def create_prev_node(node_params, node_list, params_list):\n concat = []\n for prev_node_name in node_params.prev_node_names:\n node = find_node(prev_node_name, node_list)\n prev_params = find_params(prev_node_name, params_list)\n\n if node is None and prev_params is not None:\n node = create_prev_node(prev_params, node_list, params_list)\n if node is not None:\n concat.append(node.features)\n\n if 'concat' in node_params.name_in_graph:\n net = tf.concat(concat, axis=-1)\n else:\n net = tf.add_n(concat) / len(concat)\n\n new_node = Node(net=net,\n activation=self.params.activation,\n normalization=self.params.normalization,\n params=node_params,\n L2_weight=self.params.L2_weight,\n is_training=self.is_training)\n node_list.append(new_node)\n return new_node\n\n f_out = self.params.f_start\n with tf.variable_scope(self.params.scope, reuse=tf.AUTO_REUSE):\n cell_prev_prev = Node(net=net,\n f_out=f_out,\n stride=1,\n func_name='conv3x3',\n cell_id=0,\n layer=0,\n type='N',\n normalization=self.params.normalization,\n L2_weight=self.params.L2_weight,\n is_training=self.is_training,\n name='input-2')\n\n cell_prev = Node(net=net,\n f_out=f_out,\n stride=1,\n func_name='conv3x3',\n cell_id=0,\n layer=0,\n type='N',\n normalization=self.params.normalization,\n L2_weight=self.params.L2_weight,\n is_training=self.is_training,\n name='input-1')\n\n node = create_prev_node(architecture.node_params[0][0],\n [cell_prev_prev, cell_prev],\n architecture.node_params)\n\n net = conv2d(node.features,\n k_size=1,\n f_out=2,\n stride=1,\n activation=self.params.activation,\n normalization=self.params.normalization,\n L2_weight=self.params.L2_weight,\n use_pre_activation=True,\n is_training=self.is_training,\n name='conv_GAP')\n net = avg_pool(net, radius=net.shape[1], stride=1, padding='VALID', name='GAP')\n net = tf.reduce_mean(net, axis=[1, 2])\n\n self.logits = net\n self.probs = tf.nn.softmax(net)\n\n print(' [*] DARTS loaded...')\n return self.logits\n\n def make_summary(self, cell):\n for weight in cell.weights:\n weight = tf.expand_dims(weight, axis=1)\n weight = tf.expand_dims(weight, axis=0)\n weight = tf.expand_dims(weight, axis=0)\n self.summary_vis.append(tf.summary.image(\"weight_{}\".format(cell.cell_id), weight))\n self.summary_vis_2.append(tf.summary.image(\"weight_{}_eval\".format(cell.cell_id), weight))\n\n def loss(self, Y, normalize=False, name='MSE'):\n\n loss = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits_v2(labels=Y, logits=self.logits))\n\n # Weight decay regularizer\n l2_loss = tf.losses.get_regularization_loss()\n\n # total loss by the mean of cross entropy loss and the weighted regularizer\n self.total_loss = loss + l2_loss\n\n # Accuracy for train and test set\n correct_prediction = tf.equal(tf.argmax(Y, 1), tf.argmax(self.probs, 1))\n accuracy = tf.reduce_mean(tf.cast(correct_prediction, tf.float32))\n\n # Summarys for tensorboard\n self.summary.append(tf.summary.scalar(\"accuracy_train\", accuracy))\n self.summary_val.append(tf.summary.scalar(\"accuracy_test\", accuracy))\n\n self.summary.append(tf.summary.scalar(\"cross_entropy_train\", loss))\n self.summary_val.append(tf.summary.scalar(\"cross_entropy_test\", loss))\n\n self.summary.append(tf.summary.scalar(\"total_loss_train\", self.total_loss))\n self.summary_val.append(tf.summary.scalar(\"total_loss_test\", self.total_loss))\n\n self.summary_val.append(tf.summary.scalar(\"l2_loss\", l2_loss))\n return self.total_loss\n\n def make_graph(self, path, filename):\n self.nodes[-1].active = True\n last_node = [[self.nodes[-1]]]\n found = True\n while found:\n nodes = []\n for node in last_node[0]:\n\n if len(node.prev_nodes) > 0:\n index = self.sess.run([node.prev_weights_eval], feed_dict={self.is_eval: True})[0]\n node.params.prev_node_names = []\n for idx in range(len(index)):\n if index[idx] == 1.0:\n node.prev_nodes[idx].active = True\n if node.prev_nodes[idx].params.name_in_graph not in node.params.prev_node_names:\n node.params.prev_node_names.append(node.prev_nodes[idx].params.name_in_graph)\n if node.prev_nodes[idx] not in nodes:\n nodes.append(node.prev_nodes[idx])\n else:\n node.prev_nodes[idx].active = False\n if len(nodes) > 0:\n last_node.insert(0, nodes)\n else:\n found = False\n\n graph = Digraph('G', filename='hello.gv', format='png')\n last_node = [[self.nodes[-1]]]\n found = True\n edges = []\n architecture_params = [[self.nodes[-1].params.__dict__]]\n\n def check_if_active(nodes):\n for node in nodes:\n if node.active:\n return True\n return False\n\n while found:\n nodes = []\n params = []\n for node in last_node[0]:\n if len(node.prev_nodes) == 0:\n found = False\n break\n\n for prev_node in node.prev_nodes:\n if prev_node.active and node.active and \\\n (check_if_active(prev_node.prev_nodes) or len(prev_node.prev_nodes) == 0):\n if [prev_node, node] not in edges:\n graph.edge(prev_node.name, node.name)\n edges.append([prev_node, node])\n if prev_node not in nodes:\n nodes.append(prev_node)\n params.append(prev_node.params.__dict__)\n if len(nodes) > 0:\n last_node.insert(0, nodes)\n architecture_params.append(params)\n else:\n found = False\n graph.render(filename=filename, directory=path)\n\n architecture = Architecture(node_params=architecture_params)\n architecture.save(path=os.path.join(path, filename))\n\n print (' [*] Graph drawed')\n\n", "sub_path": "darts/cnn/architecture.py", "file_name": "architecture.py", "file_ext": "py", "file_size_in_byte": 14858, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "tfcore.utilities.params_serializer.ParamsSerializer", "line_number": 13, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "cnn.node.Node_Params", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "tfcore.interfaces.IModel.IModel_Params", "line_number": 45, "usage_type": "name"}, {"api_name": "os.path.realpath", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path", "line_number": 68, "usage_type": "attribute"}, {"api_name": "tfcore.interfaces.IModel.IModel", "line_number": 71, "usage_type": "name"}, {"api_name": "cnn.node.Node", "line_number": 129, "usage_type": "call"}, {"api_name": "cnn.node.Node", "line_number": 141, "usage_type": "call"}, {"api_name": "cnn.cell.Cell", "line_number": 159, "usage_type": "call"}, {"api_name": "cnn.node.Node", "line_number": 229, "usage_type": "call"}, {"api_name": "cnn.node.Node", "line_number": 240, "usage_type": "call"}, {"api_name": "cnn.node.Node", "line_number": 252, "usage_type": "call"}, {"api_name": "graphviz.Digraph", "line_number": 347, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 384, "usage_type": "call"}, {"api_name": "os.path", "line_number": 384, "usage_type": "attribute"}]}
+{"seq_id": "408510586", "text": "#pylint: disable-all\n\nfrom aqt import mw\nfrom aqt.utils import showInfo\nfrom aqt.qt import *\nfrom anki.hooks import wrap\n\nfrom aqt.utils import showInfo\n\nimport os\nimport random\n\nmw.showdogs = {}\nmw.showdogs['card_count'] = 0\nmw.showdogs['interval'] = 10\n\nclass DogDialog(QDialog):\n\n def keyPressEvent(self, event):\n\n # why does it have to be hex? nobody knows.\n # 0x20 == spacebar\n if event.key() == 0x20:\n self.close()\n\n\ndef showDog():\n mw.showdogs['card_count'] = mw.showdogs['card_count'] + 1\n if mw.showdogs['card_count'] % mw.showdogs['interval'] != 0:\n return\n\n dialog = DogDialog(mw)\n\n layout = QVBoxLayout(dialog)\n dialog.setLayout(layout)\n\n dogs_dir = os.path.join(mw.pm.addonFolder(), 'showdogs')\n\n image_path = random.choice(os.listdir(dogs_dir))\n data = open(os.path.join(dogs_dir, image_path), 'r').read()\n\n image = QImage()\n image.loadFromData(data)\n\n label = QLabel()\n myPixmap = QPixmap(os.path.join(dogs_dir, image_path))\n myScaledPixmap = myPixmap.scaled(label.size(), Qt.KeepAspectRatio, Qt.SmoothTransformation)\n label.setPixmap(myScaledPixmap)\n label.show()\n layout.addWidget(label)\n\n dialog.exec_()\n\nmw.reviewer.nextCard = wrap(mw.reviewer.nextCard, showDog)\n", "sub_path": "showdogs.py", "file_name": "showdogs.py", "file_ext": "py", "file_size_in_byte": 1275, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "aqt.mw.showdogs", "line_number": 13, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 13, "usage_type": "name"}, {"api_name": "aqt.mw.showdogs", "line_number": 14, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 14, "usage_type": "name"}, {"api_name": "aqt.mw.showdogs", "line_number": 15, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 15, "usage_type": "name"}, {"api_name": "aqt.mw.showdogs", "line_number": 28, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 28, "usage_type": "name"}, {"api_name": "aqt.mw.showdogs", "line_number": 29, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 29, "usage_type": "name"}, {"api_name": "aqt.mw", "line_number": 32, "usage_type": "argument"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "aqt.mw.pm.addonFolder", "line_number": 37, "usage_type": "call"}, {"api_name": "aqt.mw.pm", "line_number": 37, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 37, "usage_type": "name"}, {"api_name": "random.choice", "line_number": 39, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "aqt.mw.reviewer", "line_number": 54, "usage_type": "attribute"}, {"api_name": "aqt.mw", "line_number": 54, "usage_type": "name"}, {"api_name": "anki.hooks.wrap", "line_number": 54, "usage_type": "call"}]}
+{"seq_id": "430885953", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# @Time : 2019-11-15 10:23\n# @Author : weihuchao\n\nfrom flask import Flask, render_template, request\n\nfrom analyzer import check_mid_data, init_mid_data\n\napp = Flask(__name__)\n\n\n@app.route('/', methods=['GET', 'POST'])\ndef index():\n has_min_data = check_mid_data()\n if request.method == \"POST\":\n file_path = request.form.get(\"file_path\")\n init_mid_data(file_path=file_path)\n return render_template('index.html', has_min_data=True)\n\n return render_template('index.html', has_min_data=has_min_data)\n\n\n@app.route('/api/list')\ndef api_list():\n list_data = []\n return render_template('api/list.html', list_data=list_data)\n", "sub_path": "web.py", "file_name": "web.py", "file_ext": "py", "file_size_in_byte": 702, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "flask.Flask", "line_number": 10, "usage_type": "call"}, {"api_name": "analyzer.check_mid_data", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 16, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 16, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 17, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 17, "usage_type": "name"}, {"api_name": "analyzer.init_mid_data", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 19, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 21, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 27, "usage_type": "call"}]}
+{"seq_id": "2650324", "text": "import logging\nimport os\n\nimport numpy as np\n\nlogging.basicConfig(format='%(asctime)s - %(levelname)s - %(name)s - %(message)s',\n datefmt='%m/%d/%Y %H:%M:%S',\n level=logging.DEBUG)\nlogger = logging.getLogger(__name__)\n\n\ndef accuracy(out, labels):\n outputs = np.argmax(out, axis=1)\n return np.sum(outputs == labels)\n\n\ndef warmup_linear(x, warmup=0.002):\n if x < warmup:\n return x / warmup\n return 1.0 - x\n\n\ndef create_dirs(dirpath):\n \"\"\"Creating directories.\n \"\"\"\n if not os.path.exists(dirpath):\n os.makedirs(dirpath)\n logger.info(\"==> 📂 Created {0}\".format(dirpath))\n\n\nclass InputExample(object):\n \"\"\"A single training/test example for simple sequence classification.\"\"\"\n\n def __init__(self, guid, text_a, text_b=None, label=None):\n \"\"\"Constructs a InputExample.\n\n Args:\n guid: Unique id for the example.\n text_a: string. The untokenized text of the first sequence. For single\n sequence tasks, only this sequence must be specified.\n text_b: (Optional) string. The untokenized text of the second sequence.\n Only must be specified for sequence pair tasks.\n label: (Optional) string. The label of the example. This should be\n specified for train and dev examples, but not for test examples.\n \"\"\"\n self.guid = guid\n self.text_a = text_a\n self.text_b = text_b\n self.label = label\n\n\nclass InputFeatures(object):\n \"\"\"A single set of features of data.\"\"\"\n\n def __init__(self, input_ids, input_mask, segment_ids, label_id):\n self.input_ids = input_ids\n self.input_mask = input_mask\n self.segment_ids = segment_ids\n self.label_id = label_id\n\n\ndef convert_examples_to_features(examples, label_list, max_seq_length, tokenizer):\n \"\"\"Loads a data file into a list of `InputBatch`s.\"\"\"\n\n label_map = {label: i for i, label in enumerate(label_list)}\n\n features = []\n for (ex_index, example) in enumerate(examples):\n tokens_a = tokenizer.tokenize(example.text_a)\n\n tokens_b = None\n if example.text_b:\n tokens_b = tokenizer.tokenize(example.text_b)\n # Modifies `tokens_a` and `tokens_b` in place so that the total\n # length is less than the specified length.\n # Account for [CLS], [SEP], [SEP] with \"- 3\"\n _truncate_seq_pair(tokens_a, tokens_b, max_seq_length - 3)\n else:\n # Account for [CLS] and [SEP] with \"- 2\"\n if len(tokens_a) > max_seq_length - 2:\n tokens_a = tokens_a[:(max_seq_length - 2)]\n\n # The convention in BERT is:\n # (a) For sequence pairs:\n # tokens: [CLS] is this jack ##son ##ville ? [SEP] no it is not . [SEP]\n # type_ids: 0 0 0 0 0 0 0 0 1 1 1 1 1 1\n # (b) For single sequences:\n # tokens: [CLS] the dog is hairy . [SEP]\n # type_ids: 0 0 0 0 0 0 0\n #\n # Where \"type_ids\" are used to indicate whether this is the first\n # sequence or the second sequence. The embedding vectors for `type=0` and\n # `type=1` were learned during pre-training and are added to the wordpiece\n # embedding vector (and position vector). This is not *strictly* necessary\n # since the [SEP] token unambigiously separates the sequences, but it makes\n # it easier for the model to learn the concept of sequences.\n #\n # For classification tasks, the first vector (corresponding to [CLS]) is\n # used as as the \"sentence vector\". Note that this only makes sense because\n # the entire model is fine-tuned.\n tokens = [\"[CLS]\"] + tokens_a + [\"[SEP]\"]\n segment_ids = [0] * len(tokens)\n\n if tokens_b:\n tokens += tokens_b + [\"[SEP]\"]\n segment_ids += [1] * (len(tokens_b) + 1)\n\n input_ids = tokenizer.convert_tokens_to_ids(tokens)\n\n # The mask has 1 for real tokens and 0 for padding tokens. Only real\n # tokens are attended to.\n input_mask = [1] * len(input_ids)\n\n # Zero-pad up to the sequence length.\n padding = [0] * (max_seq_length - len(input_ids))\n input_ids += padding\n input_mask += padding\n segment_ids += padding\n\n assert len(input_ids) == max_seq_length\n assert len(input_mask) == max_seq_length\n assert len(segment_ids) == max_seq_length\n\n label_id = label_map[example.label]\n features.append(\n InputFeatures(input_ids=input_ids,\n input_mask=input_mask,\n segment_ids=segment_ids,\n label_id=label_id))\n return features\n\n\ndef convert_example_to_feature(text_a, text_b, max_seq_length, tokenizer, ):\n \"\"\"Transform input text into `InputBatch`.\"\"\"\n\n example = InputExample(guid=None, text_a=text_a, text_b=text_b, label=None)\n tokens_a = tokenizer.tokenize(example.text_a)\n\n tokens_b = None\n if example.text_b:\n tokens_b = tokenizer.tokenize(example.text_b)\n # Modifies `tokens_a` and `tokens_b` in place so that the total\n # length is less than the specified length.\n # Account for [CLS], [SEP], [SEP] with \"- 3\"\n _truncate_seq_pair(tokens_a, tokens_b, max_seq_length - 3)\n else:\n # Account for [CLS] and [SEP] with \"- 2\"\n if len(tokens_a) > max_seq_length - 2:\n tokens_a = tokens_a[:(max_seq_length - 2)]\n\n tokens = [\"[CLS]\"] + tokens_a + [\"[SEP]\"]\n segment_ids = [0] * len(tokens)\n\n if tokens_b:\n tokens += tokens_b + [\"[SEP]\"]\n segment_ids += [1] * (len(tokens_b) + 1)\n\n input_ids = tokenizer.convert_tokens_to_ids(tokens)\n\n # The mask has 1 for real tokens and 0 for padding tokens. Only real\n # tokens are attended to.\n input_mask = [1] * len(input_ids)\n\n # Zero-pad up to the sequence length.\n padding = [0] * (max_seq_length - len(input_ids))\n input_ids += padding\n input_mask += padding\n segment_ids += padding\n\n assert len(input_ids) == max_seq_length\n assert len(input_mask) == max_seq_length\n assert len(segment_ids) == max_seq_length\n\n feature = InputFeatures(input_ids=input_ids,\n input_mask=input_mask,\n segment_ids=segment_ids,\n label_id=None)\n return feature\n\n\ndef _truncate_seq_pair(tokens_a, tokens_b, max_length):\n \"\"\"Truncates a sequence pair in place to the maximum length.\"\"\"\n\n # This is a simple heuristic which will always truncate the longer sequence\n # one token at a time. This makes more sense than truncating an equal percent\n # of tokens from each, since if one sequence is very short then each token\n # that's truncated likely contains more information than a longer sequence.\n while True:\n total_length = len(tokens_a) + len(tokens_b)\n if total_length <= max_length:\n break\n if len(tokens_a) > len(tokens_b):\n tokens_a.pop()\n else:\n tokens_b.pop()\n", "sub_path": "models/bert_model/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 7142, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.basicConfig", "line_number": 6, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 8, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 27, "usage_type": "call"}]}
+{"seq_id": "64750570", "text": "#!/usr/bin/env python\n# encoding: utf-8\n\"\"\"The module that handles the main interface of loklak\"\"\"\nfrom __future__ import (absolute_import, division,\n print_function, unicode_literals)\n\nimport json\nimport re\nimport requests\nfrom xmljson import badgerfish as bf\nfrom json import dumps\nimport csv\n\n\nclass Loklak(object):\n \"\"\"The fields for the Loklak object\"\"\"\n baseUrl = 'http://loklak.org/'\n name = None\n followers = None\n following = None\n query = None\n since = None\n until = None\n source = None\n count = None\n fields = None\n from_user = None\n fields = None\n limit = None\n action = None\n data = {}\n\n def __init__(self, baseUrl='http://loklak.org/'):\n baseUrl = re.findall('http[s]?://(?:[a-zA-Z]|[0-9]|[$-_@.&+]|[!*\\(\\),]|(?:%[0-9a-fA-F][0-9a-fA-F]))+', baseUrl)\n try:\n if baseUrl[0]:\n if baseUrl[0] != 'http://loklak.org/':\n url_test = self.hello()\n if url_test['status'] == 'ok':\n self.baseUrl = baseUrl[0]\n else:\n self.baseUrl = baseUrl[0]\n except IndexError:\n pass\n\n def getBaseUrl(self):\n return self.baseUrl\n\n def status(self):\n \"\"\"Returns the status of the server\"\"\"\n status_application = 'api/status.json'\n url_to_give = self.baseUrl+status_application\n return_to_user = requests.get(url_to_give)\n if return_to_user.status_code == 200:\n return return_to_user.json()\n else:\n return_to_user = {}\n return json.dumps(return_to_user)\n\n def xmlToJson(self, xmlData = None):\n \"\"\"Converts XML to JSON as the service\"\"\"\n jsonData = ''\n if xmlData:\n jsonData = dumps(bf.data(fromstring(xmlData)))\n return jsonData\n\n def csvToJson(self, csvData = None, fieldnamesList = None):\n \"\"\"Converts CSV to JSON as the service\"\"\"\n jsonData = ''\n if csvData:\n data = csv.DictReader( csvData, fieldnames = fieldnamesList)\n jsonData = json.dumps( [ row for row in data ] )\n return out\n\n def hello(self):\n \"\"\"Gives a hello\"\"\"\n hello_application = 'api/hello.json'\n url_to_give = self.baseUrl+hello_application\n return_to_user = requests.get(url_to_give)\n if return_to_user.status_code == 200:\n return return_to_user.json()\n else:\n return_to_user = {}\n return json.dumps(return_to_user)\n\n def geocode(self, places=None):\n \"\"\"Gives the geocode\"\"\"\n geo_application = 'api/geocode.json'\n url_to_give = self.baseUrl+geo_application\n params = {}\n params['places'] = places\n return_to_user = requests.get(url_to_give, params=params)\n if return_to_user.status_code == 200:\n return return_to_user.json()\n else:\n return_to_user = {}\n return json.dumps(return_to_user)\n\n def get_map(self, latitude, longitude, width=500, height=500,\n zoom=8, text=\"\"):\n \"\"\"Returns a map of size 500x500\"\"\"\n map_application = 'vis/map.png'\n params = {'text': text, 'mlat': latitude, 'mlon': longitude,\n 'width': width, 'height': height, 'zoom': zoom}\n return_to_user = requests.get(self.baseUrl + map_application,\n params=params, stream=True)\n if return_to_user.status_code == 200:\n return return_to_user.raw.read()\n else:\n return ''\n\n def get_markdown(self, text, color_text=\"000000\", color_bg=\"ffffff\",\n padding=\"10\", uppercase=\"true\"):\n \"\"\"Returns a map of size 500x500\"\"\"\n map_application = 'vis/markdown.png'\n params = {'text': text, 'color_text': color_text, 'color_background': color_bg,\n 'padding': padding, 'uppercase': uppercase}\n return_to_user = requests.get(self.baseUrl + map_application,\n params=params, stream=True)\n if return_to_user.status_code == 200:\n return return_to_user.raw.read()\n else:\n return ''\n\n def peers(self):\n \"\"\"Gives the peers of a user\"\"\"\n peers_application = 'api/peers.json'\n url_to_give = self.baseUrl+peers_application\n return_to_user = requests.get(url_to_give)\n if return_to_user.status_code == 200:\n return return_to_user.json()\n else:\n return_to_user = {}\n return json.dumps(return_to_user)\n\n def user(self, name=None, followers=None, following=None):\n \"\"\"User information, including who they are following, and\n who follows them\"\"\"\n user_application = 'api/user.json'\n url_to_give = self.baseUrl+user_application\n self.name = name\n self.followers = followers\n self.following = following\n if name:\n params = {}\n params['screen_name'] = self.name\n if followers is not None:\n params['followers'] = self.followers\n if following is not None:\n params['following'] = self.following\n\n return_to_user = requests.get(url_to_give, params=params)\n if return_to_user.status_code == 200:\n return return_to_user.json()\n else:\n return_to_user = {}\n return json.dumps(return_to_user)\n else:\n return_to_user = {}\n return_to_user['error'] = ('No user name given to query. Please'\n ' check and try again')\n return json.dumps(return_to_user)\n\n def settings(self):\n \"\"\"Gives the settings of the application\"\"\"\n settings_application = 'api/settings.json'\n url_to_give = self.baseUrl + settings_application\n return_to_user = requests.get(url_to_give)\n if return_to_user.status_code == 200:\n return return_to_user.json()\n else:\n return_to_user = {}\n return_to_user['error'] = ('This API has access restrictions:'\n ' only localhost clients are granted.')\n return json.dumps(return_to_user)\n\n def susi(self, query=None):\n \"\"\"Hits Susi with the required query and returns back the susi response\"\"\"\n susi_application = 'api/susi.json'\n url_to_give = self.baseUrl + susi_application\n self.query = query\n if query:\n params = {}\n params['q'] = self.query\n return_to_user = requests.get(url_to_give, params=params)\n if return_to_user.status_code == 200:\n return return_to_user.json()\n else:\n return_to_user = {}\n return_to_user['error'] = ('Looks like there is a problem in susi replying.')\n return json.dumps(return_to_user)\n else:\n return_to_user = {}\n return_to_user['error'] = ('Please ask susi something.')\n return json.dumps(return_to_user)\n\n def search(self, query=None, since=None, until=None, from_user=None, count=None):\n \"\"\"Handles the searching\"\"\"\n search_application = 'api/search.json'\n url_to_give = self.baseUrl+search_application\n self.query = query\n self.since = since\n self.until = until\n self.from_user = from_user\n self.count = count\n if query:\n params = {}\n params['query'] = self.query\n if since:\n params['query'] = params['query'] + ' since:'+self.since\n if until:\n params['query'] = params['query'] + ' until:'+self.until\n if from_user:\n params['query'] = params['query'] + ' from:'+self.from_user\n if count:\n params['count'] = self.count\n return_to_user = requests.get(url_to_give, params=params)\n if return_to_user.status_code == 200:\n return return_to_user.json()\n else:\n return_to_user = {}\n return_to_user['error'] = ('Something went wrong, looks like'\n ' the server is down.')\n return json.dumps(return_to_user)\n else:\n return_to_user = {}\n return_to_user['error'] = ('No Query string has been'\n ' given to query for an account')\n return json.dumps(return_to_user)\n\n def suggest(self, query=None, count=None, order=None, orderby=None,since=None, until=None):\n suggest_application = 'api/suggest.json'\n url_to_give = self.baseUrl+suggest_application\n params = {}\n if query:\n params['query'] = query\n if count:\n params['count'] = count\n if order:\n params['order'] = order\n if since:\n params['since'] = since\n if until:\n params['until'] = until\n print(params)\n return_to_user = requests.get(url_to_give, params=params)\n print(return_to_user.url)\n if return_to_user.status_code == 200:\n return return_to_user.json()\n else :\n return_to_user = {}\n return_to_user['error'] = ('Something went wrong,'\n ' looks like the server is down.')\n return json.dumps(return_to_user)\n\n def aggregations(self, query=None, since=None, until=None,\n fields=None, limit=6, count=0):\n \"\"\"Gives the aggregations of the application\"\"\"\n aggregations_application = 'api/search.json'\n url_to_give = self.baseUrl+aggregations_application\n self.query = query\n self.since = since\n self.until = until\n self.fields = fields\n self.limit = limit\n self.count = count\n if query:\n params = {}\n params['query'] = self.query\n if since:\n params['query'] = params['query']+' since:'+self.since\n if until:\n params['query'] = params['query']+' until:'+self.until\n if fields:\n if isinstance(fields, list):\n params['fields'] = ','.join(self.fields)\n else:\n params['fields'] = self.fields\n\n params['count'] = self.count\n params['source'] = 'cache'\n return_to_user = requests.get(url_to_give, params=params)\n if return_to_user.status_code == 200:\n return return_to_user\n else:\n return_to_user = {}\n return_to_user['error'] = ('Something went wrong,'\n ' looks like the server is down.')\n return json.dumps(return_to_user)\n else:\n return_to_user = {}\n return_to_user['error'] = ('No Query string has been given to run'\n 'query for aggregations')\n return json.dumps(return_to_user)\n\n def account(self, name=None, action=None, data=None):\n \"\"\"Displays users account\"\"\"\n account_application = 'account.json'\n url_to_give = 'http://localhost:9000/api/'+account_application\n self.name = name\n self.data = data\n self.action = action\n # Simple GET Query\n headers = {\n 'User-Agent': ('Mozilla/5.0 (Android 4.4; Mobile; rv:41.0)'\n ' Gecko/41.0 Firefox/41.0'),\n 'From': 'info@loklak.org'\n }\n if name:\n params = {}\n params['screen_name'] = self.name\n return_to_user = requests.get(url_to_give, params=params,\n headers=headers)\n if return_to_user.status_code == 200:\n return return_to_user.json()\n else:\n return_to_user = {}\n return_to_user['error'] = ('Something went wrong,'\n ' looks like the query is wrong.')\n return json.dumps(return_to_user)\n # if action = update and data is provided, then make request\n elif self.action == 'update' and data:\n params = {}\n params['action'] = self.action\n params['data'] = self.data\n return_to_user = requests.post(url_to_give,\n params=params, headers=headers)\n if return_to_user.status_code == 200:\n return return_to_user.json()\n else:\n return_to_user = {}\n return_to_user['error'] = ('Something went wrong,'\n ' looks like the query is wrong.')\n return json.dumps(return_to_user)\n else:\n return_to_user = {}\n return_to_user['error'] = ('No Query string has been given'\n ' given to query for an account')\n return json.dumps(return_to_user)\n", "sub_path": "loklak.py", "file_name": "loklak.py", "file_ext": "py", "file_size_in_byte": 13203, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "re.findall", "line_number": 34, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 53, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 58, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 64, "usage_type": "call"}, {"api_name": "xmljson.badgerfish.data", "line_number": 64, "usage_type": "call"}, {"api_name": "xmljson.badgerfish", "line_number": 64, "usage_type": "name"}, {"api_name": "csv.DictReader", "line_number": 71, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 72, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 79, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 84, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 92, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 97, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 105, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 118, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 129, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 134, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 152, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 157, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 162, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 168, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 175, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 185, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 191, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 195, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 217, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 224, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 229, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 246, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 254, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 282, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 289, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 294, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 312, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 320, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 326, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 334, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 339, "usage_type": "call"}]}
+{"seq_id": "221039259", "text": "import sys\nfrom IPython.display import display, HTML\n\n\nclass Displayer:\n _quiet = False\n\n def __init__(self, quiet=False):\n self._quiet = quiet\n\n def _display(self, game):\n if not self._quiet:\n print(game)\n\n def show(self, game):\n self._display(game)\n\n def win(self, game):\n self._display(game)\n print(\"You win! Score: %s\" % game.score)\n\n def lose(self, game):\n print(\"You lose! Score: %s\" % game.score)\n\n\nclass IPythonDisplayer(Displayer):\n\n def __init__(self, board_size=40, quiet=False):\n self._quiet = quiet\n self._size = board_size\n\n def _render(self, game):\n board = game.board\n html = '''Score: {}
'''.format(game.score)\n table = ''''''\n td = '''{} | ''' % (self._size, self._size)\n content = ''\n for row in range(game.size):\n content += ''''''\n for col in range(game.size):\n elem = int(board[row, col])\n content += td.format(elem if elem else \"\")\n content += '''
'''\n html += table.format(content)\n return html\n\n def _display(self, game):\n if 'ipykernel' in sys.modules:\n if not self._quiet:\n source = self._render(game)\n display(HTML(source))\n else:\n print(\"Warning: since it's not in ipykernel, \"\n \"it will show the command line version.\")\n super()._display(game)\n", "sub_path": "game2048/displayer.py", "file_name": "displayer.py", "file_ext": "py", "file_size_in_byte": 1636, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sys.modules", "line_number": 49, "usage_type": "attribute"}, {"api_name": "IPython.display.display", "line_number": 52, "usage_type": "call"}, {"api_name": "IPython.display.HTML", "line_number": 52, "usage_type": "call"}]}
+{"seq_id": "363812092", "text": "import scrapy\nimport re\n#to do: course_specialisation \n# instructors_list\n\nclass PrometheusSpider(scrapy.Spider):\n name = 'prometheus'\n custom_settings = {\n 'ITEM_PIPELINES': {\n 'tutorial.pipelines.TutorialPipeline': 400\n }\n }\n start_urls = ['https://prometheus.org.ua/courses/']\n\n def parse(self, response):\n for href in response.xpath('//h3[contains(@class, \"gdlr-item-title gdlr-skin-title gdlr-skin-border\")]/a[contains(@href, \"courses\")]/@href').extract():\n yield scrapy.Request(response.urljoin(href),\n callback=self.parse_course)\n\n\n def parse_course(self, response):\n instructors_list = {}\n # print(response.xpath('//article[contains(@class, \"teacher\")]/h3/text()'))\n # print(response.xpath('//article[contains(@class, \"teacher\")]/p').extract())\n for instructor, info in zip(response.xpath('//article[contains(@class, \"teacher\")]/h3/text()').extract(), response.xpath('//article[contains(@class, \"teacher\")]/p').extract()):\n instructors_list[instructor] = info.replace('', '').replace('
', '')\n weeks_duration = response.xpath('(//h2[contains(text(), \"Тривалість\") or contains(text(), \"ТРИВАЛІСТЬ КУРСУ\") or contains(text(), \"Довжина курсу\")]/following-sibling::p)[1]/text()').extract()\n course_provider = (response.url).split('/')[4]\n if 'course-v1' in course_provider:\n course_provider = None\n try:\n duration_filter = (re.findall('\\d+', weeks_duration[0]))[0]\n\n # duration_filter = int(weeks_duration[0][0])\n except:\n duration_filter = None\n # instructor_name = response.xpath('//article[contains(@class, \"teacher\")]/h3/text()').extract()\n # instructors_desc = response.xpath('//article[contains(@class, \"teacher\")]/p/text()').extract()\n yield {\n 'name': response.xpath('//div[contains(@class, \"heading-group\")]/h1/text()').extract()[0].strip(),\n 'source': 'prometheus',\n 'category': None,\n 'provider': course_provider,\n 'language': 'Українська',\n 'duration': None if not weeks_duration else weeks_duration[0],\n 'duration_filter': duration_filter,\n 'instructors_list': instructors_list,\n 'start_date': response.xpath('//span[contains(@class, \"important-dates-item-text start-date\")]/text()').extract()[0],\n 'description': response.xpath('//section[contains(@class, \"about\")]/p/text()').extract()[0],\n 'link': response.url,\n 'video': None,\n # 'price': None,\n 'img': 'https://edx.prometheus.org.ua' + response.xpath('//div[contains(@class, \"hero\")]/img/@src').extract()[0]\n # }\n }", "sub_path": "spiders2/tutorial/spiders/prometheus_spider.py", "file_name": "prometheus_spider.py", "file_ext": "py", "file_size_in_byte": 2857, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "scrapy.Spider", "line_number": 6, "usage_type": "attribute"}, {"api_name": "scrapy.Request", "line_number": 17, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 32, "usage_type": "call"}]}
+{"seq_id": "440398683", "text": "#coding: utf-8\r\nimport tensorflow as tf\r\nimport tools\r\nimport tensorflow.keras as kr\r\n\r\ndef create_onehot(one_data,vocab_data,max_length):\r\n\tvocab_list = tools.read_file(vocab_data)\r\n\tvocabulary_word2index = {} ### word :index\r\n\tvocabulary_index2word = {} ### index :word\r\n\tfor i, vocab in enumerate(vocab_list):\r\n\t\tvocabulary_word2index[vocab] = i\r\n\t\tvocabulary_index2word[i] = vocab\r\n\r\n\tif isinstance(one_data, str) and \".txt\" in one_data:###是字符串,即txt\r\n\t\tone_data_list = tools.read_file(one_data)\r\n\t\tsingleTest = False\r\n\telse:\r\n\t\tone_data_list = one_data ### 数组\r\n\t\tsingleTest = True\r\n\tX = []\r\n\tif not singleTest:\r\n\t\tfor data in one_data_list:\r\n\t\t\tcontent = [vocabulary_word2index.get(e, 0) for e in data]\r\n\t\t\tX.append(content)\r\n\t\tx_pad = kr.preprocessing.sequence.pad_sequences(X, max_length) #### 对数据进行定长处理\r\n\telse:\r\n\t\tcontent = [vocabulary_word2index.get(e, 0) for e in one_data_list]\r\n\t\tX.append(content)\r\n\t\tx_pad = kr.preprocessing.sequence.pad_sequences(X, max_length)\r\n\treturn x_pad\r\n\r\ndef decode_text(labels):\r\n\tcategories, cat_to_id = tools.label_dict()\r\n\twords = []\r\n\tfor word_num in labels:\r\n\t\tword = categories[word_num]\r\n\t\twords.append(word)\r\n\treturn words\r\n\r\nclass Textcnn_pred(object):\r\n\tdef __init__(self,pb_path):\r\n\t\twith tf.Graph().as_default():\r\n\t\t\tself.output_graph_def = tf.GraphDef()\r\n\t\t\twith open(pb_path, \"rb\") as f:\r\n\t\t\t\tself.output_graph_def.ParseFromString(f.read())\r\n\t\t\t\ttf.import_graph_def(self.output_graph_def, name=\"\")\r\n\t\t\tsess_config = tf.ConfigProto(allow_soft_placement=True)\r\n\t\t\tsess_config.gpu_options.per_process_gpu_memory_fraction = 0.8\r\n\t\t\tsess_config.gpu_options.allow_growth = True\r\n\t\t\tself.sess = tf.Session(config=sess_config)\r\n\t\t\tself.sess.run(tf.global_variables_initializer())\r\n\t\t\t# 定义输入的张量名称,对应网络结构的输入张量\r\n\t\t\t# input:0作为输入图像,keep_prob:0作为dropout的参数,测试时值为1,is_training:0训练参数\r\n\t\t\tself.input_image_tensor = self.sess.graph.get_tensor_by_name(\"input_x:0\")\r\n\t\t\t# 定义输出的张量名称\r\n\t\t\toutput_tensor_name = self.sess.graph.get_tensor_by_name(\"output:0\")\r\n\t\t\tself.textcnn_pred = tf.argmax(tf.nn.softmax(output_tensor_name), 1)\r\n\t\tprint(\"################ load TextCNN model down! ##########################\")\r\n\r\n\tdef _close(self):\r\n\t\tself.sess.close()\r\n\r\n\tdef text_pre(self,input):\r\n\t\ttest_X = create_onehot(one_data=input, vocab_data=\"./model_and_data/vocab.txt\", max_length=200)\r\n\t\tout=self.sess.run(self.textcnn_pred, feed_dict={self.input_image_tensor: test_X})\r\n\t\tresult = decode_text(out)\r\n\t\treturn result\r\n\r\nif __name__==\"__main__\":\r\n\ttextcnn = Textcnn_pred(pb_path=\"./model_and_data/textcnn_twoClass.pb\")\r\n\twhile 1:\r\n\t\t#inputs = ['龙岗区四联路30号路段多辆车辆违停,私设路障,严重影响车辆和行人通行。',\r\n\t\t# '馨荔苑业主群,13:44分报警人发来短信:对不起,拨错号了,歉意。',\r\n\t # ]\r\n\t\tprint(\"开始输入:\")\r\n\t\tinputs = input()\r\n\t\tpred = textcnn.text_pre(inputs)[0]\r\n\t\tprint(\"Result: \",pred)", "sub_path": "Release/textcnn_pb.py", "file_name": "textcnn_pb.py", "file_ext": "py", "file_size_in_byte": 3047, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "tools.read_file", "line_number": 7, "usage_type": "call"}, {"api_name": "tools.read_file", "line_number": 15, "usage_type": "call"}, {"api_name": "tensorflow.keras.preprocessing.sequence.pad_sequences", "line_number": 25, "usage_type": "call"}, {"api_name": "tensorflow.keras.preprocessing", "line_number": 25, "usage_type": "attribute"}, {"api_name": "tensorflow.keras", "line_number": 25, "usage_type": "name"}, {"api_name": "tensorflow.keras.preprocessing.sequence.pad_sequences", "line_number": 29, "usage_type": "call"}, {"api_name": "tensorflow.keras.preprocessing", "line_number": 29, "usage_type": "attribute"}, {"api_name": "tensorflow.keras", "line_number": 29, "usage_type": "name"}, {"api_name": "tools.label_dict", "line_number": 33, "usage_type": "call"}, {"api_name": "tensorflow.Graph", "line_number": 42, "usage_type": "call"}, {"api_name": "tensorflow.GraphDef", "line_number": 43, "usage_type": "call"}, {"api_name": "tensorflow.import_graph_def", "line_number": 46, "usage_type": "call"}, {"api_name": "tensorflow.ConfigProto", "line_number": 47, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 50, "usage_type": "call"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 51, "usage_type": "call"}, {"api_name": "tensorflow.argmax", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.nn.softmax", "line_number": 57, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 57, "usage_type": "attribute"}]}
+{"seq_id": "579420332", "text": "import uvicorn\nfrom fastapi import FastAPI\nfrom fastapi.middleware.cors import CORSMiddleware\nfrom fastapi.responses import JSONResponse\nfrom fastapi.staticfiles import StaticFiles\n\n\nrequest = {'product1': {'product_url': 'https://www.elgiganten.dk/product/gaming/konsoller/playstation-konsoller/220280/playstation-5-ps5-digital-edition', 'product_name': 'Digital', 'class': 'not-available', 'store': 'Elgiganten', 'stock': False, 'time': False},\n 'product2': {'product_url': 'https://www.elgiganten.dk/product/gaming/konsoller/playstation-konsoller/220276/playstation-5-ps5','product_name': 'Disc Standard', 'class': 'not-available', 'store': 'Elgiganten', 'stock': False, 'time': False},\n 'product3': {'product_url': 'https://www.bilka.dk/produkter/sony-playstation-5/100532624/','product_name': 'Disc Standard', 'class': 'purchase-button v-btn v-btn--block v-btn--contained v-btn--disabled theme--light v-size--large mt-5 flex-grow-0', 'store': 'Bilka', 'stock': False, 'time': False},\n 'product3': {'product_url': 'https://www.bilka.dk/produkter/sony-playstation-5-digital/100553322/','product_name': 'Digital', 'class': 'purchase-button v-btn v-btn--block v-btn--contained v-btn--disabled theme--light v-size--large mt-5 flex-grow-0', 'store': 'Bilka', 'stock': False, 'time': False},\n }\n\n\n\napp = FastAPI()\n\n\n@app.post(\"/add-email\")\nasync def add_email():\n\n return JSONResponse({'hello': 'world'})\n\n\n\n\n@app.get(\"/status\")\nasync def status():\n\n return request\n\n\napp.mount(\"/\", StaticFiles(directory=\"website\", html=True), name=\"website\")\napp.add_middleware(CORSMiddleware, allow_origins=[\"*\"])\n\nif __name__ == \"__main__\":\n uvicorn.run(app, host=\"localhost\", port=8008)", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1721, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "fastapi.FastAPI", "line_number": 16, "usage_type": "call"}, {"api_name": "fastapi.responses.JSONResponse", "line_number": 22, "usage_type": "call"}, {"api_name": "fastapi.staticfiles.StaticFiles", "line_number": 33, "usage_type": "call"}, {"api_name": "fastapi.middleware.cors.CORSMiddleware", "line_number": 34, "usage_type": "argument"}, {"api_name": "uvicorn.run", "line_number": 37, "usage_type": "call"}]}
+{"seq_id": "627246182", "text": "import os\nfrom setuptools import setup\n\n\ndef read(fname):\n \"\"\"\n Utility function to read the README file.\n Used for the long_description. It's nice, because now 1) we have a top\n level README file and 2) it's easier to type in the README file than to put\n a raw string in below ...\n \"\"\"\n with open(os.path.join(os.path.dirname(__file__), fname)) as f:\n return f.read()\n\n\nsetup(\n name=\"whatever-forever\",\n version=\"0.0.7\",\n author=\"Tony Fast\",\n author_email=\"tony.fast@gmail.com\",\n description=\"prototype whatever in the Jupyter notebook\",\n license=\"BSD\",\n keywords=\"IPython Magic Jupyter\",\n url=\"http://github.com/tonyfast/whatever-forever\",\n py_modules=[\"whatever-forever\"],\n long_description=read(\"README.rst\"),\n classifiers=[\n \"Topic :: Utilities\",\n \"Framework :: IPython\",\n \"Natural Language :: English\",\n \"Programming Language :: Python\",\n \"Intended Audience :: Developers\",\n \"Development Status :: 3 - Alpha\",\n \"Operating System :: OS Independent\",\n \"Programming Language :: Python :: 3\",\n \"License :: OSI Approved :: BSD License\",\n \"Topic :: Software Development :: Testing\",\n ],\n install_requires=[\n \"toolz\",\n ]\n)", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1269, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.path.join", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 12, "usage_type": "call"}, {"api_name": "setuptools.setup", "line_number": 16, "usage_type": "call"}]}
+{"seq_id": "231237037", "text": "import cv2\n\nimport RPi.GPIO as GPIO # RPi.GPIO 라이브러리를 GPIO로 사용\n\nfrom time import sleep # time 라이브러리의 sleep함수 사용\n\nservoPin = 12 # 서보 핀\n\nSERVO_MAX_DUTY = 12 # 서보의 최대(180도) 위치의 주기\n\nSERVO_MIN_DUTY = 3 # 서보의 최소(0도) 위치의 주기\n\nGPIO.setmode(GPIO.BOARD) # GPIO 설정\n\nGPIO.setup(servoPin, GPIO.OUT) # 서보핀 출력으로 설정\n\nservo = GPIO.PWM(servoPin, 50) # 서보핀을 PWM 모드 50Hz로 사용하기 (50Hz > 20ms)\n\nservo.start(0) # 서보 PWM 시작 duty = 0, duty가 0이면 서보는 동작하지 않는다.\n\n\ndef setServoPos(degree):\n # 각도는 180도를 넘을 수 없다.\n\n if degree > 180:\n degree = 180\n\n # 각도(degree)를 duty로 변경한다.\n\n duty = SERVO_MIN_DUTY + (degree * (SERVO_MAX_DUTY - SERVO_MIN_DUTY) / 180.0)\n\n # duty 값 출력\n\n print(\"Degree: {} to {}(Duty)\".format(degree, duty))\n\n # 변경된 duty값을 서보 pwm에 적용\n\n servo.ChangeDutyCycle(duty)\n\n\nmodel = 'models/dnn_face_detector/res10_300x300_ssd_iter_140000_fp16.caffemodel'\n\nconfig = 'models/dnn_face_detector/deploy.prototxt'\n\nimport sys\n\nimport numpy as np\n\nimport cv2\n\nimport face_recognition_predict\n\nmodel = 'models/dnn_face_detector/res10_300x300_ssd_iter_140000_fp16.caffemodel'\n\nconfig = 'models/dnn_face_detector/deploy.prototxt'\n\nface_predict = face_recognition_predict.FaceRecognitionPredict()\n\ncap = cv2.VideoCapture(0)\n\nif not cap.isOpened():\n print('Camera open failed!')\n\n sys.exit()\n\nnet = cv2.dnn.readNet(model, config)\n\nif net.empty():\n print('Net open failed!')\n\n sys.exit()\n\nwhile True:\n\n ret, frame = cap.read()\n\n if not ret:\n break\n\n blob = cv2.dnn.blobFromImage(frame, 1, (300, 300), (104, 177, 123))\n\n net.setInput(blob)\n\n out = net.forward()\n\n detect = out[0, 0, :, :]\n\n (h, w) = frame.shape[:2]\n\n for i in range(detect.shape[0]):\n\n confidence = detect[i, 2]\n\n if confidence < 0.6:\n break\n\n # detect값는 정규화가 되어있어 실제 들어온 영상의 w, h를 곱해야함.\n\n x1 = int(detect[i, 3] * w)\n\n y1 = int(detect[i, 4] * h)\n\n x2 = int(detect[i, 5] * w)\n\n y2 = int(detect[i, 6] * h)\n\n crop = frame[y1:y2, x1:x2]\n\n name, class_probability = face_predict.predict(crop)\n\n cv2.rectangle(frame, (x1, y1), (x2, y2), (0, 255, 0))\n\n if class_probability:\n\n label = f'Name: {name} Probability {class_probability:4.2f}'\n\n cv2.putText(frame, label, (x1, y1 - 1), cv2.FONT_HERSHEY_SIMPLEX, 0.8, (0, 255, 0), 1, cv2.LINE_AA)\n\n setServoPos(0)\n\n sleep(1) # 1초 대기\n\n # 90도에 위치\n\n setServoPos(90)\n\n sleep(5)\n\n setServoPos(0)\n\n sleep(1) # 1초 대기\n\n\n\n else:\n\n label = f'wait!!'\n\n cv2.putText(frame, label, (x1, y1 - 1), cv2.FONT_HERSHEY_SIMPLEX, 0.8, (0, 255, 0), 1, cv2.LINE_AA)\n\n cv2.imshow('frame', frame)\n\n if cv2.waitKey(1) == 27:\n break\n\n# 서보 PWM 정지\n\nservo.stop()\n\n# GPIO 모드 초기화\n\nGPIO.cleanup()\n\ncv2.destroyAllWindows()", "sub_path": "Face_Recognition/face_recognition_live.py", "file_name": "face_recognition_live.py", "file_ext": "py", "file_size_in_byte": 3176, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "RPi.GPIO.setmode", "line_number": 13, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 13, "usage_type": "name"}, {"api_name": "RPi.GPIO.BOARD", "line_number": 13, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.setup", "line_number": 15, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 15, "usage_type": "name"}, {"api_name": "RPi.GPIO.OUT", "line_number": 15, "usage_type": "attribute"}, {"api_name": "RPi.GPIO.PWM", "line_number": 17, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 17, "usage_type": "name"}, {"api_name": "face_recognition_predict.FaceRecognitionPredict", "line_number": 57, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 59, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 64, "usage_type": "call"}, {"api_name": "cv2.dnn.readNet", "line_number": 66, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 66, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 71, "usage_type": "call"}, {"api_name": "cv2.dnn.blobFromImage", "line_number": 80, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 80, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 111, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 117, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 117, "usage_type": "attribute"}, {"api_name": "cv2.LINE_AA", "line_number": 117, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 121, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 127, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 131, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 139, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 139, "usage_type": "attribute"}, {"api_name": "cv2.LINE_AA", "line_number": 139, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 141, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 143, "usage_type": "call"}, {"api_name": "RPi.GPIO.cleanup", "line_number": 152, "usage_type": "call"}, {"api_name": "RPi.GPIO", "line_number": 152, "usage_type": "name"}, {"api_name": "cv2.destroyAllWindows", "line_number": 154, "usage_type": "call"}]}
+{"seq_id": "519465632", "text": "# Copyright 2018 The YARL-Project, All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ==============================================================================\n\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nimport logging\nimport numpy as np\nimport unittest\n\nfrom yarl.components import Component\nfrom yarl.tests import ComponentTest\nfrom yarl.utils import root_logger\nfrom yarl.tests.dummy_components import Dummy1to1\n\n\nclass TestConnectionsWithLabels(unittest.TestCase):\n \"\"\"\n Tests for different ways to place different, but single sub-Components into the core.\n \"\"\"\n root_logger.setLevel(level=logging.INFO)\n\n def test_connecting_in1_and_1to1_to_1to1_no_labels(self):\n \"\"\"\n Adds two components (A, B) with 1-to-1 graph_fns to the core.\n Connects \"input1\" with A, A's \"output\" to \"output\".\n Connects \"input2\" with B and B's output to A.\n If we now pull out-Socket \"output\", it should know by the given input, which op we actually want.\n \"\"\"\n core = Component(inputs=[\"input1\", \"input2\"], outputs=\"output\", scope=\"container\")\n a = Dummy1to1(scope=\"A\")\n b = Dummy1to1(scope=\"B\")\n # Throw in the sub-components.\n core.add_components(a, b)\n # Connect them.\n core.connect(\"input1\", (a, \"input\"))\n core.connect(\"input2\", (b, \"input\"))\n core.connect((b, \"output\"), (a, \"input\")) # a/input now has two incoming connections.\n core.connect((a, \"output\"), \"output\")\n\n test = ComponentTest(component=core, input_spaces=dict(input1=float, input2=float))\n\n # Now pulling on the same Socket (given one of the input Sockets) should trigger the correct op.\n # Expected output: input1 + 1.0(a)\n test.test(out_socket_names=\"output\", inputs=dict(input1=np.array(1.1)), expected_outputs=2.1)\n # Expected output: input2 + 1.0(b) + 1.0(a)\n test.test(out_socket_names=\"output\", inputs=dict(input2=np.float32(1.1)), expected_outputs=3.1)\n\n def test_connecting_in1_to_1to1_no_labels(self):\n \"\"\"\n Adds one component (A) with 1-to-1 graph_fn to the core which has two in-Sockets and 1 out-Socket.\n Connects \"input_c\" with A, A's \"output\" to \"output\".\n Connects \"input_a\" with A, A's \"output\" to \"output\".\n No labels whatsoever.\n If we now pull \"output\", and provide input1 AND input2, it should use the in-Socket that comes\n first in alphabetic order (\"input_a\" even if it's defined second).\n \"\"\"\n core = Component(inputs=[\"input_c\", \"input_a\"], outputs=\"output\", scope=\"container\")\n a = Dummy1to1(scope=\"A\")\n # Throw in the sub-component.\n core.add_components(a)\n # Connect correctly.\n core.connect(\"input_c\", (a, \"input\"))\n core.connect(\"input_a\", (a, \"input\"))\n core.connect((a, \"output\"), \"output\")\n\n test = ComponentTest(component=core, input_spaces=dict(input_c=float, input_a=float))\n\n # Now pulling on \"output\" and providing both inputs should cause disambiguity, but it should chose\n # the one fist in alphabetic order (input_a).\n # Expected output: input_a + 1.0\n test.test(out_socket_names=\"output\", inputs=dict(input_c=np.array(1.5), input_a=np.array(2.1)),\n expected_outputs=3.1)\n\n def test_connecting_in1_and_in2_to_1to1_to_out1_and_out2_with_labels(self):\n \"\"\"\n Same as `test_connecting_in1_to_1to1_no_labels` but with labels.\n So if we provide both inputs, it should know which one to take (instead of using alphabetic order).\n \"\"\"\n core = Component(inputs=[\"input_c\", \"input_a\"], outputs=\"output\", scope=\"container\")\n dummy = Dummy1to1(scope=\"dummy\")\n # Throw in the sub-component.\n core.add_components(dummy)\n # Connect correctly (with labels).\n # [input_c->dummy/input] is now labelled as \"from_in_c\" (op_records passing to dummy will be labelled so).\n core.connect(\"input_c\", (dummy, \"input\"), label=\"from_in_c\")\n # [input_a->dummy/input] is now labelled as \"from_in_a\" (op_records passing to dummy will be labelled so).\n core.connect(\"input_a\", (dummy, \"input\"), label=\"from_in_a\")\n # Force using the input_c path (over the alphabetically favored input_a).\n # [dummy/output->output] is now labelled as \"from_in_c\" and will thus only allow ops that have this label to\n # be passed through to \"output\" (all other op_records will be filtered).\n core.connect((dummy, \"output\"), \"output\", label=\"from_in_c\")\n\n test = ComponentTest(component=core, input_spaces=dict(input_c=float, input_a=float))\n\n # Now pulling on \"output\" and providing both inputs should not cause disambiguity since out out-Socket\n # was connected to A's \"output\" through the label \"from_in_c\", so it should always use \"input_c\".\n # Expected output: input_c + 1.0\n test.test(out_socket_names=\"output\", inputs=dict(input_c=np.array(1.5), input_a=np.array(2.1)),\n expected_outputs=2.5)\n test.test(out_socket_names=\"output\", inputs=dict(input_c=np.array(1.5), input_a=np.array(2.1)),\n expected_outputs=2.5)\n\n def test_connecting_in_to_2x_to_different_1to1_then_2x_to_1to1_to_out1_and_out2_with_labels(self):\n \"\"\"\n\n \"\"\"\n core = Component(inputs=\"input\", outputs=[\"output1\", \"output2\"], scope=\"container\")\n pre1 = Dummy1to1(scope=\"pre1\")\n pre2 = Dummy1to1(scope=\"pre2\", constant_value=2.0) # make it different from pre1\n hub = Dummy1to1(scope=\"hub\")\n\n # Throw in the sub-components.\n core.add_components(pre1, pre2, hub)\n\n # Connect correctly (with labels).\n core.connect(\"input\", (pre1, \"input\"))\n core.connect(\"input\", (pre2, \"input\"))\n # [pre1/input->hub/input] is now labelled as \"from_pre1\" (op_records passing to dummy will be labelled so).\n core.connect((pre1, \"output\"), (hub, \"input\"), label=\"from_pre1\")\n # [pre2/input->hub/input] is now labelled as \"from_pre2\" (op_records passing to dummy will be labelled so).\n core.connect((pre2, \"output\"), (hub, \"input\"), label=\"from_pre2\")\n # Provide both paths (via pre1 or pre2) via labels for the two different out-Sockets.\n core.connect((hub, \"output\"), \"output1\", label=\"from_pre1\")\n core.connect((hub, \"output\"), \"output2\", label=\"from_pre2\")\n\n test = ComponentTest(component=core, input_spaces=dict(input=float))\n\n # Now pulling on \"output1\", should take the op over pre1. Pulling \"output2\" should take the op over pre2.\n # Expected output: (input + 1.0) + 1.0\n test.test(out_socket_names=\"output1\", inputs=dict(input=np.array(187.5)), expected_outputs=189.5)\n # Expected output: (input + 2.0) + 1.0\n test.test(out_socket_names=\"output2\", inputs=dict(input=np.array(87.5)), expected_outputs=90.5)\n", "sub_path": "yarl/tests/test_core/test_connections_with_labels.py", "file_name": "test_connections_with_labels.py", "file_ext": "py", "file_size_in_byte": 7503, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "unittest.TestCase", "line_number": 30, "usage_type": "attribute"}, {"api_name": "yarl.utils.root_logger.setLevel", "line_number": 34, "usage_type": "call"}, {"api_name": "yarl.utils.root_logger", "line_number": 34, "usage_type": "name"}, {"api_name": "logging.INFO", "line_number": 34, "usage_type": "attribute"}, {"api_name": "yarl.components.Component", "line_number": 43, "usage_type": "call"}, {"api_name": "yarl.tests.dummy_components.Dummy1to1", "line_number": 44, "usage_type": "call"}, {"api_name": "yarl.tests.dummy_components.Dummy1to1", "line_number": 45, "usage_type": "call"}, {"api_name": "yarl.tests.ComponentTest", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 60, "usage_type": "call"}, {"api_name": "yarl.components.Component", "line_number": 71, "usage_type": "call"}, {"api_name": "yarl.tests.dummy_components.Dummy1to1", "line_number": 72, "usage_type": "call"}, {"api_name": "yarl.tests.ComponentTest", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 85, "usage_type": "call"}, {"api_name": "yarl.components.Component", "line_number": 93, "usage_type": "call"}, {"api_name": "yarl.tests.dummy_components.Dummy1to1", "line_number": 94, "usage_type": "call"}, {"api_name": "yarl.tests.ComponentTest", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 114, "usage_type": "call"}, {"api_name": "yarl.components.Component", "line_number": 121, "usage_type": "call"}, {"api_name": "yarl.tests.dummy_components.Dummy1to1", "line_number": 122, "usage_type": "call"}, {"api_name": "yarl.tests.dummy_components.Dummy1to1", "line_number": 123, "usage_type": "call"}, {"api_name": "yarl.tests.dummy_components.Dummy1to1", "line_number": 124, "usage_type": "call"}, {"api_name": "yarl.tests.ComponentTest", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 146, "usage_type": "call"}]}
+{"seq_id": "363552213", "text": "from django.conf.urls import url\nfrom .views import *\n\nurlpatterns = [\n url(r'^index',index,name='index'),\n url(r'^login', login, name='login'),\n\n url(r'^register', register, name='register'),\n url(r'^check', check, name='check'),\n\n]", "sub_path": "python1808/Django/Day06/Day06Django/CookApp/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 245, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.conf.urls.url", "line_number": 5, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 6, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}]}
+{"seq_id": "186460344", "text": "\r\n# coding: utf-8\r\nimport _init_paths\r\n\r\nimport sys\r\nimport os\r\nimport os.path\r\nimport random\r\nimport collections\r\nimport shutil\r\nimport time\r\nimport glob\r\nimport csv\r\nimport numpy as np\r\n\r\nimport torch\r\nimport torch.backends.cudnn as cudnn\r\nimport torch.nn as nn\r\nimport torch.nn.parallel\r\nimport torch.optim as optim\r\nimport torch.utils.data as data\r\nimport torchvision.datasets as datasets\r\nimport torchvision.models as models\r\nimport torchvision.transforms as transforms\r\nimport utils\r\n#from utils.Tensor import to_variable,to_numpy\r\nfrom utils.metric import AccuracyMeter,MovingAverageMeter,AverageMeter\r\nfrom torch.autograd import Variable\r\nimport torch.nn.functional as F\r\nimport image_loader\r\nfrom image_loader import image_loader\r\n\r\nfrom PIL import Image\r\n\r\n\r\nclass Trainer(object):\r\n def __init__(self, model, optimizer, train_loader, valid_loader, use_cuda=True):\r\n self.model = model\r\n self.optimizer = optimizer\r\n self.train_loader = train_loader\r\n self.valid_loader = valid_loader\r\n self.use_cuda = use_cuda\r\n\r\n def train(self, epoch):\r\n self.model.train()\r\n\r\n train_loss = AverageMeter()\r\n train_acc = AccuracyMeter()\r\n\r\n for i, (x, y) in enumerate(self.train_loader):\r\n x = Variable(x)\r\n y = Variable(y)\r\n if self.use_cuda:\r\n x = x.cuda()\r\n y = y.cuda()\r\n output = self.model(x)\r\n loss = F.cross_entropy(output, y)\r\n\r\n self.optimizer.zero_grad()\r\n loss.backward()\r\n self.optimizer.step()\r\n\r\n train_loss.update(float(loss.data))\r\n\r\n y_pred = output.data.max(dim=1)[1]\r\n correct = int(y_pred.eq(y.data).cpu().sum())\r\n train_acc.update(correct, x.size(0))\r\n if i % 10 ==0:\r\n print('\\nTrain Epoch/batch| [{}/{}]: Average batch loss: {:.6f}\\n'.format(epoch,i,train_acc.accuracy))\r\n return train_loss.average, train_acc.accuracy\r\n\r\n def validate(self):\r\n self.model.eval()\r\n\r\n valid_loss = AverageMeter()\r\n valid_acc = AccuracyMeter()\r\n\r\n for i, (x, y) in enumerate(self.valid_loader):\r\n x = Variable(x, volatile=True)\r\n y = Variable(y).long()\r\n if self.use_cuda:\r\n x = x.cuda()\r\n y = y.cuda()\r\n output = self.model(x)\r\n loss = F.cross_entropy(output, y)\r\n\r\n valid_loss.update(float(loss.data))\r\n\r\n y_pred = output.data.max(dim=1)[1]\r\n correct = int(y_pred.eq(y.data).cpu().sum())\r\n valid_acc.update(correct, x.size(0))\r\n print('\\nTrain Epoch [{}]: Average batch loss: {:.6f}\\n'.format(epoch,valid_acc.accuracy))\r\n return valid_loss.average, valid_acc.accuracy\r\n\r\ndef print_msg(proc,epoch,loss,acc):\r\n print('proc={},epoch={}:loss={},acc={}'.format(proc,epoch,loss,acc))\r\n\r\nif __name__ == '__main__':\r\n train_batch_size=32\r\n\r\n ROOT_DIR = os.getcwd()\r\n DATA_HOME_DIR = ROOT_DIR + '/data/cat_dog'\r\n\r\n # paths\r\n data_path = DATA_HOME_DIR \r\n train_path = data_path + '/train/'\r\n valid_path = data_path + '/test/'\r\n train_loader,valid_loader=image_loader(train_path,valid_path,\r\n train_batch_size,\r\n valid_batch_size=None,\r\n train_shuffle=True,\r\n valid_shuffle=False,\r\n train_num_workers=0,\r\n valid_num_workers=0)\r\n\r\n #model = models.resnet34(pretrained=True) \r\n model = models.resnet50(pretrained=True) \r\n for param in model.parameters():\r\n param.requires_grad = False\r\n #model.fc = nn.Linear(512,2) #resnet34\r\n model.fc = nn.Linear(2048,2)\r\n #optimizer = optim.Adam(model.parameters(), lr=1e-4)\r\n optimizer = optim.Adam(model.fc.parameters(), lr=1e-4, weight_decay=1e-4)\r\n if torch.cuda.is_available():\r\n model.cuda()\r\n \r\n #optimizer = optim.Adam(model.module.fc.parameters(), lr=1e-3)\r\n scheduler = torch.optim.lr_scheduler.StepLR(optimizer, step_size=3, gamma=0.85)\r\n trainer = Trainer(model=model,optimizer=optimizer,train_loader=train_loader,valid_loader=valid_loader,use_cuda=True)\r\n epochs = 50\r\n train_loss_epochs=np.zeros((epochs,))\r\n train_acc_epochs=np.zeros((epochs,))\r\n test_loss_epochs=np.zeros((epochs,))\r\n test_acc_epochs=np.zeros((epochs,))\r\n i=0\r\n for epoch in range(1, 10):\r\n scheduler.step()\r\n train_loss, train_acc=trainer.train(epoch)\r\n train_loss_epochs[i]=train_loss\r\n train_acc_epochs[i]=train_acc\r\n print_msg('train',epoch,train_loss,train_acc)\r\n test_loss, test_acc=trainer.validate()\r\n test_loss_epochs[i]=test_loss\r\n test_acc_epochs[i]=test_acc\r\n print_msg('test',epoch,test_loss,test_acc)\r\n i+=1\r\n data = dict()\r\n data['train_loss']=train_loss_epochs\r\n data['train_acc']=train_acc_epochs\r\n data['test_loss']=test_loss_epochs\r\n data['test_acc']=test_acc_epochs\r\n torch.save(data,'data.pt')\r\n print('finished')\r\n ", "sub_path": "Test/classifier_dog_cat.py", "file_name": "classifier_dog_cat.py", "file_ext": "py", "file_size_in_byte": 5093, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "utils.metric.AverageMeter", "line_number": 47, "usage_type": "call"}, {"api_name": "utils.metric.AccuracyMeter", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.nn.functional.cross_entropy", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 57, "usage_type": "name"}, {"api_name": "utils.metric.AverageMeter", "line_number": 75, "usage_type": "call"}, {"api_name": "utils.metric.AccuracyMeter", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 80, "usage_type": "call"}, {"api_name": "torch.nn.functional.cross_entropy", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 85, "usage_type": "name"}, {"api_name": "os.getcwd", "line_number": 101, "usage_type": "call"}, {"api_name": "image_loader.image_loader", "line_number": 108, "usage_type": "call"}, {"api_name": "torchvision.models.resnet50", "line_number": 117, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 117, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 121, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 121, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 123, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 123, "usage_type": "name"}, {"api_name": "torch.cuda.is_available", "line_number": 124, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 124, "usage_type": "attribute"}, {"api_name": "torch.optim.lr_scheduler.StepLR", "line_number": 128, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 128, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 134, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 147, "usage_type": "name"}, {"api_name": "torch.utils.data", "line_number": 148, "usage_type": "name"}, {"api_name": "torch.utils.data", "line_number": 149, "usage_type": "name"}, {"api_name": "torch.utils.data", "line_number": 150, "usage_type": "name"}, {"api_name": "torch.utils.data", "line_number": 151, "usage_type": "name"}, {"api_name": "torch.save", "line_number": 152, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 152, "usage_type": "argument"}]}
+{"seq_id": "373078498", "text": "import glob\nimport os\n\nimport numpy as np\nimport pyasdf\nfrom obspy.io.sac.sactrace import SACTrace\n\n\"\"\"\nthis script outputs the stacked cross-correlation functions into SAC traces\n\nadd an option to output the CCFs into csv files for image transform analysis\n\"\"\"\n\n# ------absolute path to output data-------\nSTACKDIR = \"/Users/chengxin/Documents/SCAL/STACK\"\nALLFILES = glob.glob(os.path.join(STACKDIR, \"*/*.h5\"))\nCOMP_OUT = [\"ZZ\", \"RR\", \"TT\"]\n# COMP_OUT = ['ZR','ZT','ZZ','TR','TT','TZ','RR','RT','RZ']\n# COMP_OUT = ['EE','EN','EZ','NE','NN','NZ','ZE','ZN','ZZ']\ndtype = \"Allstack_linear\"\n\n# ---output file format-----\nout_SAC = True\nout_CSV = False\n\nif (not out_SAC) and (not out_CSV):\n raise ValueError(\"out_SAC and out_CSV cannot be False at the same time\")\n\nnfiles = len(ALLFILES)\nif not os.path.isdir(os.path.join(STACKDIR, \"STACK_SAC\")):\n os.mkdir(os.path.join(STACKDIR, \"STACK_SAC\"))\n\n# ----loop through station pairs----\nfor ii in range(nfiles):\n with pyasdf.ASDFDataSet(ALLFILES[ii], mode=\"r\") as ds:\n # -----get station info from file name-----\n fname = ALLFILES[ii].split(\"/\")[-1].split(\"_\")\n staS = fname[0].split(\".\")[1]\n netS = fname[0].split(\".\")[0]\n staR = fname[1].split(\".\")[1]\n netR = fname[1].split(\".\")[0]\n\n # -----read data information-------\n slist = ds.auxiliary_data.list()\n rlist = ds.auxiliary_data[slist[0]].list()\n maxlag = ds.auxiliary_data[slist[0]][rlist[0]].parameters[\"maxlag\"]\n dt = ds.auxiliary_data[slist[0]][rlist[0]].parameters[\"dt\"]\n slat = ds.auxiliary_data[slist[0]][rlist[0]].parameters[\"latS\"]\n slon = ds.auxiliary_data[slist[0]][rlist[0]].parameters[\"lonS\"]\n rlat = ds.auxiliary_data[slist[0]][rlist[0]].parameters[\"latR\"]\n rlon = ds.auxiliary_data[slist[0]][rlist[0]].parameters[\"lonR\"]\n\n # ----make sure data exists------\n if dtype in slist:\n for icomp in range(len(COMP_OUT)):\n comp = COMP_OUT[icomp]\n\n if comp in rlist:\n if out_SAC:\n # --------read the correlations---------\n corr = ds.auxiliary_data[dtype][comp].data[:]\n temp = netS + \".\" + staS + \"_\" + netR + \".\" + staR + \"_\" + comp + \".SAC\"\n\n # -------check whether folder exists-------\n if not os.path.isdir(os.path.join(STACKDIR, \"STACK_SAC/\" + netS + \".\" + staS)):\n os.mkdir(os.path.join(STACKDIR, \"STACK_SAC/\" + netS + \".\" + staS))\n filename = os.path.join(STACKDIR, \"STACK_SAC/\" + netS + \".\" + staS, temp)\n\n # --------write into SAF format----------\n sac = SACTrace(\n nzyear=2000,\n nzjday=1,\n nzhour=0,\n nzmin=0,\n nzsec=0,\n nzmsec=0,\n b=-maxlag,\n delta=dt,\n stla=rlat,\n stlo=rlon,\n evla=slat,\n evlo=slon,\n data=corr,\n )\n sac.write(filename, byteorder=\"big\")\n\n if out_CSV:\n # -----------output name and read data-------------\n temp = netS + \".\" + staS + \"_\" + netR + \".\" + staR + \"_\" + comp + \".dat\"\n if not os.path.isdir(os.path.join(STACKDIR, \"STACK_DAT\")):\n os.mkdir(os.path.join(STACKDIR, \"STACK_DAT\"))\n filename = os.path.join(STACKDIR, \"STACK_DAT\", temp)\n corr = ds.auxiliary_data[dtype][comp].data[:]\n\n # -------make an array for output-------\n npts = len(corr)\n indx = npts // 2\n data = np.zeros((3, indx + 2), dtype=np.float32)\n data[0, 0] = slon\n data[1, 0] = slat\n data[2, 0] = 0\n data[0, 1] = rlon\n data[1, 1] = rlat\n data[2, 1] = 0\n tt = 0\n for jj in range(indx):\n data[0, 2 + jj] = tt\n data[1, 2 + jj] = corr[indx + jj]\n data[2, 2 + jj] = corr[indx - jj]\n tt = tt + dt\n\n np.savetxt(filename + \".csv\", np.transpose(data), delimiter=\",\")\n", "sub_path": "src/noisepy/seis/application_modules/write_sac.py", "file_name": "write_sac.py", "file_ext": "py", "file_size_in_byte": 4721, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "glob.glob", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "pyasdf.ASDFDataSet", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path", "line_number": 65, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 65, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "obspy.io.sac.sactrace.SACTrace", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path", "line_number": 90, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 90, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path", "line_number": 91, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path", "line_number": 92, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 98, "usage_type": "attribute"}, {"api_name": "numpy.savetxt", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 112, "usage_type": "call"}]}
+{"seq_id": "280848521", "text": "import Blockchain\nfrom Blockchain import blockchain\n\nfrom flask import Flask\nfrom flask import request\nfrom flask import Blueprint\n\nimport json\nimport Util\nfrom Util import ComplexEncoder\n\n#Exporting blueprint\nconsensus_api = Blueprint('consensus_api', __name__)\n\n@consensus_api.route('/chain', methods=['GET'])\ndef get_chain():\n #Blocks become dictionaries\n chain_to_send = []\n\n for block in blockchain:\n shit = block.index\n chain_to_send.append(block.reprJSON())\n\n #Send our requested chain\n return json.dumps(chain_to_send, cls=ComplexEncoder)\n\ndef find_new_chains():\n #Get others nodes blockchains\n other_chains = []\n for node_url in peer_nodes:\n block = requests.get(node_url + \"/blocks\").content\n\n #Converting Json to dictionary for easy manipulation\n block = json.loads(block)\n\n #add to chains list\n other_chains.append(block)\n return other_chains\n\ndef consensus():\n #Get blocks from other nodes\n other_chains = find_new_chains()\n\n #If this node's chain is not the longest, store the longest\n longest_chain = blockchain\n for chain in other_chains:\n if len(longest_chain) < len(chain):\n longest_chain = chain\n\n blockchain = longest_chain\n", "sub_path": "Tiny-Blockchain/Consensus.py", "file_name": "Consensus.py", "file_ext": "py", "file_size_in_byte": 1256, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "flask.Blueprint", "line_number": 13, "usage_type": "call"}, {"api_name": "Blockchain.blockchain", "line_number": 20, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 25, "usage_type": "call"}, {"api_name": "Util.ComplexEncoder", "line_number": 25, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 34, "usage_type": "call"}, {"api_name": "Blockchain.blockchain", "line_number": 45, "usage_type": "name"}, {"api_name": "Blockchain.blockchain", "line_number": 50, "usage_type": "name"}]}
+{"seq_id": "169946606", "text": "from copy import copy\n\nfrom django.http import HttpResponse\nfrom django.shortcuts import render\nfrom PyPDF2 import PdfFileReader, PdfFileWriter\n\n# Create your views here.\ndef makeHalfPagePDF(request):\n if \"GET\" == request.method:\n return render(request, 'pdfTools/makeHalfPagePDF.html', {})\n else:\n file = request.FILES[\"pdf_file\"]\n with open('original_pdf', 'wb+') as pdfFileObj:\n for chunk in file.chunks():\n pdfFileObj.write(chunk)\n pdf_reader = PdfFileReader(pdfFileObj)\n\n pdf_writer = PdfFileWriter()\n for i in range(pdf_reader.numPages):\n page = pdf_reader.getPage(i)\n width = page.mediaBox.getUpperRight_x() - page.mediaBox.getUpperLeft_x()\n height = page.mediaBox.getUpperLeft_y() - page.mediaBox.getLowerLeft_y()\n\n # 왼쪽 반\n for_left = copy(page)\n for_left.cropBox.setLowerLeft((0, 0))\n for_left.cropBox.setUpperRight((width/2, height))\n pdf_writer.addPage(for_left)\n\n # 오른쪽 반\n for_right = copy(page)\n for_right.cropBox.setLowerLeft((width / 2, 0))\n for_right.cropBox.setUpperRight((width, height))\n pdf_writer.addPage(for_right)\n\n with open('result.pdf', 'wb') as pdfResultFile:\n pdf_writer.write(pdfResultFile)\n\n with open('result.pdf', 'rb') as pdfResult:\n response = HttpResponse(pdfResult.read(), content_type=\"application/pdf\")\n response['Content-Disposition'] = 'attachment; filename=result.pdf'\n return response\n", "sub_path": "pdfTools/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1704, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.shortcuts.render", "line_number": 10, "usage_type": "call"}, {"api_name": "PyPDF2.PdfFileReader", "line_number": 16, "usage_type": "call"}, {"api_name": "PyPDF2.PdfFileWriter", "line_number": 18, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 25, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 31, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 40, "usage_type": "call"}]}
+{"seq_id": "164776967", "text": "import logging\nimport os\nimport shutil\nimport sys\n\nlogging.basicConfig(stream=sys.stdout, level=logging.INFO)\nLOG = logging.getLogger(__name__)\n\nPATH = \"../../../EVA2_8Sensoren_CronosFlex_Conv_Clean/\"\nSAVE_PATH = \"../../combined\"\nHEALTHY_INTER_SAVE_PATH = os.path.join(SAVE_PATH, 'healthy')\nBAD_INTER_SAVE_PATH = os.path.join(SAVE_PATH, 'bad')\n\n\ndef join_combined_into_one(data_path, condition):\n save_path = os.path.join(SAVE_PATH, condition)\n os.makedirs(save_path, exist_ok=True)\n concatted = open(os.path.join(save_path, \"{}-data.csv\".format(condition)), \"a\")\n files = os.listdir(data_path)\n # first file:\n print(\"Writing first csv\")\n for line in open(os.path.join(data_path, files[0])):\n concatted.write(line)\n # now the rest:\n print(\"Writing other csvs\")\n for file in files[1:]:\n print(\"Writing {}\".format(file))\n combined_csv = open(os.path.join(data_path, file))\n combined_csv.__next__() # skip the header\n for line in combined_csv:\n concatted.write(line)\n combined_csv.close() # not really needed\n print(\"Done\\n\\n\")\n concatted.close()\n\n\nFOLDERS = [\"2018-07-12_08-53-42\", \"2018-07-12_13-28-58\", \"2018-07-12_09-14-40\", '2018-07-12_13-32-39',\n \"2018-07-12_08-33-50\", \"2018-07-12_09-33-37\", \"2018-07-12_13-48-22\"]\n\nhealthy_folders = [\"2018-07-12_08-53-42\", \"2018-07-12_09-14-40\", \"2018-07-12_08-33-50\", \"2018-07-12_09-33-37\"]\nbad_folders = [\"2018-07-12_13-28-58\", '2018-07-12_13-32-39', \"2018-07-12_13-48-22\"]\n\n\ndef copy_folders(folder_group, label):\n for folder in folder_group:\n LOG.info(\"Copying combined.csv in folder: %s...\", folder)\n shutil.copyfile(os.path.join(PATH, folder, \"combined.csv\"),\n os.path.join(SAVE_PATH, label, \"{}.csv\".format(folder)))\n\n\nif __name__ == '__main__':\n copy_folders(healthy_folders, \"healthy\")\n copy_folders(bad_folders, \"bad\")\n\n join_combined_into_one(HEALTHY_INTER_SAVE_PATH, \"healthy\")\n join_combined_into_one(BAD_INTER_SAVE_PATH, \"bad\")\n", "sub_path": "1DCNN_John/data_util/concat_combined.py", "file_name": "concat_combined.py", "file_ext": "py", "file_size_in_byte": 2034, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.basicConfig", "line_number": 6, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 6, "usage_type": "attribute"}, {"api_name": "logging.INFO", "line_number": 6, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "shutil.copyfile", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}]}
+{"seq_id": "440099278", "text": "from collections import Counter\r\n\r\ndef count_words(path):\r\n \r\n txt = open(path, 'r')\r\n #txt = open('E:\\Pyth0n\\python\\modules\\soome.txt', 'r')\r\n text = txt.read()\r\n \r\n word_list = []\r\n for word in text.split(): #в переменную ворд записать каждое слово с txt\r\n clear_word = \"\" #сюда будут записаны только буквы\r\n for letter in word: #каждое слово разбиваем на символы\r\n if letter.isalpha(): #если символ с алфавита, записываем его в letter\r\n clear_word += letter.lower() #записываем буквы в clear_word\r\n word_list.append(clear_word) #записываем наши \"чистые\" слова в массив\r\n txt.close\r\n print(Counter(word_list))\r\n \r\n\r\ncount_words(input(\"Введите путь файла\"))", "sub_path": "Repeat_word_counter.py", "file_name": "Repeat_word_counter.py", "file_ext": "py", "file_size_in_byte": 924, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "collections.Counter", "line_number": 17, "usage_type": "call"}]}
+{"seq_id": "511669873", "text": "#!/usr/bin/python3\n# -*- coding: utf-8 -*-\nimport datetime\nimport json\n\nfrom elasticsearch import Elasticsearch, helpers\nfrom data_transmission.application.helper.file_operator import FileOperator\nfrom data_transmission.application.helper.mysql_helper import MysqlHelper\nfrom data_transmission.application.helper import LOG_WARNING\nfrom data_transmission.application.helper.run import wrapper\n\n\nclass ElasticsearchHelper(object):\n def __init__(self, es_config, db_url):\n if not es_config:\n es_config = {\"host\": \"127.0.0.1\", \"port\": 9200}\n self.es = self._init_es_connect(es_config)\n self.file_operator = FileOperator()\n self.mysql_obj = MysqlHelper(db_url)\n\n @staticmethod\n def _init_es_connect(es_config):\n username = es_config.pop('username') if 'username' in es_config else None\n password = es_config.pop('password') if 'password' in es_config else None\n if username and password:\n return Elasticsearch(hosts=[es_config], http_auth=(username, password))\n else:\n return Elasticsearch(hosts=[es_config])\n\n def query(self, es_query, index=\"s_cmm*\", current_node_code=None):\n result = list()\n if current_node_code:\n query_temp = {\"term\": {\"belong.deploy_level.keyword\": {\"value\": current_node_code}}}\n es_query[\"query\"][\"bool\"][\"must\"].append(query_temp)\n try:\n search_result = self.es.search(index=index, body=es_query)\n if \"hits\" in search_result:\n temp_data = search_result.get(\"hits\")\n if \"hits\" in temp_data:\n data = temp_data.get(\"hits\")\n # 取出主要数据\n if not isinstance(data, list):\n data = [data]\n for each in data:\n if \"_source\" in each:\n _source = each.get(\"_source\")\n _index = each.get(\"_index\")\n _doc_type = each.get(\"_type\")\n format_data = {\"_index\": _index, \"_type\": _doc_type, \"_source\": _source}\n result.append(format_data)\n except Exception as ex:\n LOG_WARNING.error(u\"ES查询数据时错误,查询条件为:{}\".format(es_query))\n LOG_WARNING.error(u\"错误的原因为:{}\".format(str(ex)))\n return result\n\n def save_data_to_file(self, file_name, file_path, history_times, current_node_code=None, limit_size=1000):\n result = list()\n es_queries = self.calculate_query_sentence(history_times)\n data_list = list()\n for each_es_query in es_queries:\n json_data = self.query(each_es_query, current_node_code=current_node_code)\n if json_data:\n data_list += json_data\n # 如果查询的数据不存在,不做处理\n if data_list:\n # 将10分钟的数据进行拆分\n split_data_list = self.split_data(data=data_list, limit_size=limit_size)\n for index, data in enumerate(split_data_list):\n temp_file_name = \"{}_-{}.txt\".format(file_name, index + 1)\n self.file_operator.save_file(data=data, file_name=temp_file_name, file_path=file_path)\n result.append(temp_file_name)\n return result\n\n @staticmethod\n def calculate_query_sentence(history_times):\n # 查询当前ES的数据,history_times为空的时候,查询当前所有ES的数据,不为空的时候,根据history_times查询往后的10min\n # 计算时间差,10min的数据以每分钟来查询,防止一次查询10min的数据超过10000条\n result = list()\n if history_times:\n now_time = datetime.datetime.now()\n for _ in range(0, 10):\n temp_time = now_time - datetime.timedelta(minutes=1)\n start_time = temp_time.strftime(\"%Y-%m-%d %H:%M:%S\")\n end_time = now_time.strftime(\"%Y-%m-%d %H:%M:%S\")\n es_query = {\n \"query\": {\n \"bool\": {\n \"must\": [\n {\n \"range\": {\n \"public_field.datatime.keyword\": {\n \"gte\": start_time,\n \"lte\": end_time\n }\n }\n }\n ]\n }\n },\n \"size\": 10000\n }\n result.append(es_query)\n now_time = temp_time\n else:\n now_time = datetime.datetime.now()\n start_time = now_time.strftime(\"%Y-%m-%d %H:%M:%S\")\n es_query = {\n \"query\": {\n \"bool\": {\n \"must\": [\n {\n \"range\": {\n \"public_field.datatime.keyword\": {\n \"lte\": start_time\n }\n }\n }\n ]\n }\n },\n \"size\": 10000\n }\n result.append(es_query)\n return result\n\n @staticmethod\n def split_data(data, limit_size=1000):\n \"\"\"\n 将每10分钟从ES查询的数据分隔开,避免文件过大.\n 100条数据,大概1.2M.\n :param data: ES查询出的所有数据.\n :param limit_size: 每个文件的数据量.\n :return:\n \"\"\"\n len_data = len(data)\n number = len_data // limit_size\n result = list()\n if data:\n if number == 0:\n result.append(data)\n else:\n index = 0\n for index in range(0, number):\n result.append(data[index * limit_size: (index + 1) * limit_size])\n last_data = data[(index + 1) * limit_size:]\n if last_data:\n result.append(last_data)\n return result\n\n def insert_into_es(self, task_id, data, send_params, typeof, source, es_handler=None):\n try:\n if data:\n self.mysql_obj.insert_record(task_id, send_params=send_params, typeof=typeof, success_flag=0, source=source)\n distinguish_data = self.distinguish_index(data=data)\n flag = None\n for es_index, es_data in distinguish_data.items():\n if es_data:\n flag = wrapper(self.bulk_insert, es_index, es_data, es_handler)\n # self.bulk_insert(es_index=es_index, data=es_data, es_handler=es_handler)\n if flag:\n self.mysql_obj.set_except_flag(task_id, 1)\n except Exception as ex:\n LOG_WARNING.error(u\"读取文件数据,并将数据插入ES时出错,错误原因为:{}\".format(str(ex)))\n\n def bulk_insert(self, es_index, data, es_handler=None):\n \"\"\"\n 批量插入数据到ES。\n :param str es_index: 待插入的ES的index.\n :param list data: 待批量插入的数据.\n :param es_handler: 测试时用.\n 数据格式:[{\"_index\": index, \"_type: type, \"_source\": source}, ...]\n :return:\n \"\"\"\n try:\n if es_handler:\n es = self._init_es_connect(es_handler)\n else:\n es = self.es\n self.create_es_index(index=es_index, es=es)\n # 先去除重复数据\n # data = self.remove_duplicates(data)\n helpers.bulk(es, data)\n return True\n except Exception as ex:\n LOG_WARNING.error(u\"批量插入数据到ES时发生错误,错误原因为:{}\".format(str(ex)))\n return False\n\n @staticmethod\n def distinguish_index(data):\n \"\"\"\n 一次读取的文件内容中可能包含不同天的数据,要将数据区分出来.\n :param data: 一次读取的文件数据\n :return: 整理后的数据.\n \"\"\"\n result = dict()\n if data:\n try:\n if isinstance(data, (bytes, str)):\n data = json.loads(data)\n except json.JSONDecodeError:\n LOG_WARNING.error(u\"ES数据格式有问题,丢弃该次数据.\")\n return result\n if not isinstance(data, list):\n data = [data]\n for each in data:\n index = each.get(\"_index\")\n if index in result:\n result[index].append(each)\n else:\n result.setdefault(index, [each])\n return result\n\n def create_es_index(self, index, es=None):\n # 为index设置容错率\n mapping = {\n \"settings\": {\n \"index.mapping.ignore_malformed\": True\n },\n \"mappings\": {\n \"DEVICE\": {\n \"properties\": {\n \"GGQ\": {\"properties\": {\"FSSJ\": {\"type\": \"date\", \"format\": \"yyyy-MM-dd HH:mm:ss\"}}}}\n },\n \"CHANNEL\": {\n \"properties\": {\n \"GGQ\": {\"properties\": {\"FSSJ\": {\"type\": \"date\", \"format\": \"yyyy-MM-dd HH:mm:ss\"}}}}\n },\n \"DATA\": {\n \"properties\": {\n \"GGQ\": {\"properties\": {\"FSSJ\": {\"type\": \"date\", \"format\": \"yyyy-MM-dd HH:mm:ss\"}}}}\n },\n \"BUSINESS\": {\n \"properties\": {\n \"GGQ\": {\"properties\": {\"FSSJ\": {\"type\": \"date\", \"format\": \"yyyy-MM-dd HH:mm:ss\"}}}}\n },\n \"SOFTWARE\": {\n \"properties\": {\n \"GGQ\": {\"properties\": {\"FSSJ\": {\"type\": \"date\", \"format\": \"yyyy-MM-dd HH:mm:ss\"}}}}\n }\n\n }\n }\n try:\n if index:\n if es:\n if not es.indices.exists(index=index):\n es.indices.create(index=index, body=mapping)\n else:\n if not self.es.indices.exists(index=index):\n self.es.indices.create(index=index, body=mapping)\n except Exception as e:\n LOG_WARNING.warning('为{}的index设置settings出错,错误原因:{}'.format(index, str(e)))\n\n def remove_duplicates(self, es_data):\n es_query = {\n\n }\n self.es.search()\n", "sub_path": "data_transmission/data_transmission/application/helper/elasticsearch_operator.py", "file_name": "elasticsearch_operator.py", "file_ext": "py", "file_size_in_byte": 10738, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "data_transmission.application.helper.file_operator.FileOperator", "line_number": 18, "usage_type": "call"}, {"api_name": "data_transmission.application.helper.mysql_helper.MysqlHelper", "line_number": 19, "usage_type": "call"}, {"api_name": "elasticsearch.Elasticsearch", "line_number": 26, "usage_type": "call"}, {"api_name": "elasticsearch.Elasticsearch", "line_number": 28, "usage_type": "call"}, {"api_name": "data_transmission.application.helper.LOG_WARNING.error", "line_number": 52, "usage_type": "call"}, {"api_name": "data_transmission.application.helper.LOG_WARNING", "line_number": 52, "usage_type": "name"}, {"api_name": "data_transmission.application.helper.LOG_WARNING.error", "line_number": 53, "usage_type": "call"}, {"api_name": "data_transmission.application.helper.LOG_WARNING", "line_number": 53, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 80, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 80, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 82, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 105, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 105, "usage_type": "attribute"}, {"api_name": "data_transmission.application.helper.run.wrapper", "line_number": 158, "usage_type": "call"}, {"api_name": "data_transmission.application.helper.LOG_WARNING.error", "line_number": 163, "usage_type": "call"}, {"api_name": "data_transmission.application.helper.LOG_WARNING", "line_number": 163, "usage_type": "name"}, {"api_name": "elasticsearch.helpers.bulk", "line_number": 182, "usage_type": "call"}, {"api_name": "elasticsearch.helpers", "line_number": 182, "usage_type": "name"}, {"api_name": "data_transmission.application.helper.LOG_WARNING.error", "line_number": 185, "usage_type": "call"}, {"api_name": "data_transmission.application.helper.LOG_WARNING", "line_number": 185, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 199, "usage_type": "call"}, {"api_name": "json.JSONDecodeError", "line_number": 200, "usage_type": "attribute"}, {"api_name": "data_transmission.application.helper.LOG_WARNING.error", "line_number": 201, "usage_type": "call"}, {"api_name": "data_transmission.application.helper.LOG_WARNING", "line_number": 201, "usage_type": "name"}, {"api_name": "data_transmission.application.helper.LOG_WARNING.warning", "line_number": 252, "usage_type": "call"}, {"api_name": "data_transmission.application.helper.LOG_WARNING", "line_number": 252, "usage_type": "name"}]}
+{"seq_id": "433464734", "text": "import os\nimport configparser\n\nconfigParser = configparser.ConfigParser()\nconfigFile = 'config.ini'\nconfigDefault = 'config.default'\nif os.path.isfile(configFile):\n configParser.read(configFile)\nelse:\n configParser.read(configDefault)\n with open(configFile, 'w') as conf:\n configParser.write(conf)\n\nclass Config(object):\n SECRET_KEY = configParser.get('Security', 'SECRET_KEY')\n BASIC_AUTH_USERNAME = configParser.get('Security', 'BASIC_AUTH_USERNAME')\n BASIC_AUTH_PASSWORD = configParser.get('Security', 'BASIC_AUTH_PASSWORD')\n BASIC_AUTH_FORCE = configParser.get('Security', 'BASIC_AUTH_FORCE')\n\n CONTENT_PATH = configParser.get('Content', 'CONTENT_PATH')\n CONTENT_FOLDERS = eval(configParser.get('Content', 'CONTENT_FOLDERS'))\n\n def make_path():\n path = configParser.get('Content', 'CONTENT_PATH')\n folders = eval(configParser.get('Content', 'CONTENT_FOLDERS'))\n for folder in folders:\n if not os.path.isdir(os.path.join(path,folder)):\n os.makedirs(os.path.join(path,folder))\n", "sub_path": "config.py", "file_name": "config.py", "file_ext": "py", "file_size_in_byte": 1013, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "configparser.ConfigParser", "line_number": 4, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}]}
+{"seq_id": "161621829", "text": "\"\"\"\nBootstrapping\n-------------\nFunctions for bootstrap calculations.\n\nAuthors: Petter Lind\nCreated: Autumn 2016\nUpdates:\n May 2020\n\"\"\"\nimport numpy as np\nimport multiprocessing as mp\n\n\ndef block_bootstr(data, block=5, nrep=500, nproc=1):\n \"\"\"\n Calculate block bootstrap samples.\n\n This is a block boostrap function, converted from R into python, based on:\n http://stat.wharton.upenn.edu/~buja/STAT-541/time-series-bootstrap.R\n\n Parameters\n ----------\n data: list/array\n 1D data array on which to perform the block bootstrap.\n block: int\n the block length to be used. Default is 5.\n nrep: int\n the number of resamples produced in the bootstrap. Default is 500.\n nproc: int\n Number of processors, default 1. If larger than 1, multiple processors\n are used in parallell using the multiprocessing module.\n\n Returns\n -------\n arrBt: Array\n 2D array with bootstrap samples; rows are the samples, columns the\n values.\n \"\"\"\n\n # Make sure the data is a numpy array\n data = np.array(data)\n\n error_message = \"*** ERROR ***\\n Data array should be 1D\"\n assert np.ndim(data) == 1, error_message\n\n if nproc > 1:\n # Number of cores to be used.\n # Available cores on system is a constraint\n nr_procs_set = np.minimum(nproc, mp.cpu_count())\n\n pool = mp.Pool(processes=nr_procs_set)\n computations = [pool.apply_async(_get_bootsample,\n args=(data, block)) for j in range(nrep)]\n\n arrBt = [k.get() for k in computations]\n else:\n arrBt = [_get_bootsample(data, block) for irep in range(nrep)]\n\n return np.array(arrBt)\n\n\ndef _get_bootsample(data, block):\n \"\"\"\n Sample one-dimensional data with replacement.\n\n Function to sample 1D input data by filling a vector\n with random blocks extracted from data.\n \"\"\"\n N = data.size # size of data series\n k = block # size of moving blocks\n nk = int(np.ceil(N/k)) # number of blocks\n\n dataBt = np.repeat(np.nan, N) # local vector for a bootstrap replication\n\n # fill the vector with random blocks by\n # randomly sampling endpoints and copying blocks\n for i in range(nk):\n endpoint = np.random.randint(k, N+1, size=1)\n dataBt[(i-1)*k + np.array(range(k))] = \\\n data[endpoint-np.array(range(k))[::-1]-1]\n\n return dataBt[0:N]\n\n\ndef _mproc_get_bootsamples(data, nx, j, block):\n \"\"\"Return samples from bootstrapping using multi-processing module\"\"\"\n bs = np.array([_get_bootsample(data[:, i], block=block)\n for i in range(nx)])\n return bs\n", "sub_path": "rcat/stats/bootstrap.py", "file_name": "bootstrap.py", "file_ext": "py", "file_size_in_byte": 2685, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "numpy.array", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.ndim", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 50, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 50, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 74, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 79, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 88, "usage_type": "call"}]}
+{"seq_id": "441574573", "text": "# -*- coding: utf-8 -*-\nimport scrapy\nfrom scrapy.http import Request\nimport urllib.request\nfrom jgdqedu.items import JgdqeduItem \nimport configparser\nconfig = configparser.ConfigParser()\nconfig.read('/opt/spider-exam/scrapy/jgdqedu/config.ini')\n\n\nclass KejiSpider(scrapy.Spider):\n name = \"keji\"\n allowed_domains = [\"jgdqedu.cn\"]\n start_urls = ['http://jgdqedu.cn/']\n\n def parse(self, response):\n listpage='http://www.jgdqedu.cn/science/'\n yield Request(url=listpage,callback=self.next)\n\n def next(self,response):\n print('处理列表页地址:')\n #list_page = response.xpath(\"//div[@class='list-text']/ul/li/a/@href\").extract()\n ini_list=config.get(\"jgdqedu\",\"list_page\")\n list_page =response.xpath(ini_list).extract()\n for i in range(0,len(list_page)-10):\n thisurl=('http://jgdqedu.cn'+list_page[i])\n #print(thisurl)\n yield Request(url=thisurl,callback=self.page)\n def page(self,response):\n item = JgdqeduItem()\n #item['title'] = response.xpath('//h1/text()').extract()\n ini_title=config.get(\"jgdqedu\",\"title\")\n item['title'] =response.xpath(ini_title).extract()\n item['url']=response.url\n item['catalog']='科技'\n #item['content']=response.xpath(\"//div[@class='content']\").extract()\n ini_content=config.get(\"jgdqedu\",\"content\")\n item['content']=response.xpath(ini_content).extract()\n yield item\n", "sub_path": "scrapy/jgdqedu_keji/jgdqedu/spiders/keji.py", "file_name": "keji.py", "file_ext": "py", "file_size_in_byte": 1471, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "configparser.ConfigParser", "line_number": 7, "usage_type": "call"}, {"api_name": "scrapy.Spider", "line_number": 11, "usage_type": "attribute"}, {"api_name": "scrapy.http.Request", "line_number": 18, "usage_type": "call"}, {"api_name": "scrapy.http.Request", "line_number": 28, "usage_type": "call"}, {"api_name": "jgdqedu.items.JgdqeduItem", "line_number": 30, "usage_type": "call"}]}
+{"seq_id": "511036853", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\nimport struct\nfrom enum import Enum, unique\nimport copy\n\nfrom python_c_bytes_trans import StructType\nfrom crc16 import *\n\n@unique\nclass PyType(Enum):\n\n _Normal = \"_normal\"\n _String = \"_string\"\n _Array = \"_array\"\n\nclass SwNormalDataUnit():\n\n def __init__(self, _ctype=None, _pytype=None, size=0, len=0, data=None):\n self._ctype = _ctype\n self._pytype = _pytype\n self.size = size\n self.len = len\n self.data = data\n self.new_data = data\n\n def is_sw_normal(self):\n return self._pytype == PyType._Normal\n\n def is_sw_string(self):\n return self._pytype == PyType._String\n\n def is_sw_array(self):\n return self._pytype == PyType._Array\n\n def _ctype_str(self):\n return self._ctype.value\n\n def _pytype_str(self):\n return self._pytype.value\n\n def encode(self):\n if self.is_sw_normal():\n return struct.pack(self._ctype_str(), self.new_data)\n elif self.is_sw_array(): \n return struct.pack(self._ctype_str()*self.len,\n *(self.new_data + [0]*(self.len - len(self.new_data))))\n elif self.is_sw_string():\n return struct.pack(str(self.size) + self._ctype_str(),\n (self.new_data + '\\0'*(self.size - len(self.new_data))).encode(encoding=\"utf-8\"))\n\n def decode(self, data):\n if self.is_sw_normal():\n self.new_data = struct.unpack(self._ctype_str(), data)[0]\n elif self.is_sw_array(): \n self.new_data = list(struct.unpack(self._ctype_str()*self.len, data))\n elif self.is_sw_string():\n self.new_data = struct.unpack(str(self.size) + self._ctype_str(), data)[0].decode(encoding=\"utf-8\")\n\n def savedata(self):\n self.data = copy.deepcopy(self.new_data)\n\nclass SwMobileUnit(SwNormalDataUnit):\n\n def __init__(self, id=None, _ctype=None, _pytype=None, size=0, len=0, data=None):\n super(SwMobileUnit, self).__init__(_ctype, _pytype, size, len, data)\n self.id = id\n self.id_len = 2\n self.id_ctype = StructType.UShort2\n self.id_pytype = PyType._Normal\n\n self.len_len = 1\n self.len_ctype = StructType.UChar1\n self.len_pytype = PyType._Normal\n\n self.unit_len = self.len_len + self.id_len + self.size\n\n def encode_unit(self):\n unit = bytes()\n unit += struct.pack(self.len_ctype.value, self.unit_len)\n unit += struct.pack(self.id_ctype.value, self.id)\n unit += self.encode()\n return unit\n\n def decode_unit(self, data):\n #self.decode_id(data)\n #self.decode_unit_len(data)\n\n start = self.len_len + self.id_len\n end = self.unit_len\n self.decode(data[start:end])\n\n def decode_id(self, data):\n start = self.len_len\n end = self.len_len + self.id_len\n self.id = struct.unpack(self.id_ctype.value, data[start:end])[0]\n\n def decode_unit_len(self, data):\n start = 0\n end = self.len_len\n self.unit_len = struct.unpack(self.len_ctype.value, data[start:end])[0]\n\n def print_data(self):\n print(\"id = 0x%.4x, size = %-10.d\" % (self.id, self.size), end=' ')\n print(\"data : \", end='')\n print(self.new_data or self.data)\n\nclass SwHeadPacket():\n\n def __init__(self, ap_id=0x03, vp_id=0x01, sttn_id=0x00, dev_id=0x00, pkt_id=0x01, vp_ack=0x80, mcp_id=0x01, cmd_id=0x02, cmd_ack=0xff):\n self.data = {\n 'ap_id' : SwNormalDataUnit(StructType.UChar1, PyType._Normal, 1, 1, ap_id),\n 'vp_id' : SwNormalDataUnit(StructType.UChar1, PyType._Normal, 1, 1, vp_id),\n 'sttn_id' : SwNormalDataUnit(StructType.UInt4, PyType._Normal, 4, 1, sttn_id), #站点编号\n 'dev_id' : SwNormalDataUnit(StructType.UChar1, PyType._Normal, 1, 1, dev_id), #设备编号\n 'pkt_id' : SwNormalDataUnit(StructType.UShort2, PyType._Normal, 2, 1, pkt_id),\n 'vp_ack' : SwNormalDataUnit(StructType.UChar1, PyType._Normal, 1, 1, vp_ack),\n 'mcp_id' : SwNormalDataUnit(StructType.UChar1, PyType._Normal, 1, 1, mcp_id),\n 'cmd_id' : SwNormalDataUnit(StructType.UChar1, PyType._Normal, 1, 1, cmd_id), #设置查询\n 'cmd_ack' : SwNormalDataUnit(StructType.UChar1, PyType._Normal, 1, 1, cmd_ack),\n }\n self.head_len = 13\n\n def packet(self):\n packet = bytes()\n for key in self.data:\n packet += self.data[key].encode()\n return packet\n\n def unpacket(self, data):\n start = 0\n end = 0\n for key in self.data:\n end = start + self.data[key].size\n self.data[key].decode(data[start:end])\n start = end\n self.head_len = end\n\n def print_data(self):\n for key in self.data:\n print(self.data[key].data)\n\nclass SwMobileBytes():\n\n def __init__(self):\n self.trans_bytes = None\n self.origin_bytes = None\n self.valid = True\n\n def get_trans_bytes(self):\n return self.trans_bytes\n\n def get_origin_bytes(self):\n return self.origin_bytes\n\n def del_trans_bytes(self):\n self.trans_bytes = None\n\n def del_origin_bytes(self):\n self.origin_bytes = None\n\n def set_mobile_crc_valid(self):\n crc = sumCrc16(0, self.origin_bytes[:-2])\n self.valid = crc == self.unpacket_crc()\n self.origin_bytes = self.origin_bytes[:-2]\n\n def is_mobile_crc_valid(self):\n return self.valid\n\n def add_head_tail(self):\n self.trans_bytes = b'~' + self.trans_bytes + b'~'\n\n def del_head_tail(self):\n self.origin_bytes = self.origin_bytes.strip(b'~')\n\n def add_crc16(self):\n crc = sumCrc16(0, self.trans_bytes)\n self.trans_bytes += struct.pack(StructType.UShort2.value, crc)\n\n def unpacket_crc(self):\n return struct.unpack(StructType.UShort2.value, self.origin_bytes[-2:])[0]\n\n def trans_pack7e(self):\n self.trans_bytes = self.trans_bytes.replace(b'^', b'^]')\n self.trans_bytes = self.trans_bytes.replace(b'~', b'^}')\n\n def trans_unpack7e(self):\n self.origin_bytes = self.origin_bytes.replace(b'^]', b'^')\n self.origin_bytes = self.origin_bytes.replace(b'^}', b'~')\n\n def transport_bytes(self, data):\n self.trans_bytes = data\n self.add_crc16()\n self.trans_pack7e()\n self.add_head_tail()\n\n def original_bytes(self, data):\n self.origin_bytes = data\n if self.get_valid_bytes() is False:\n return False\n self.del_head_tail()\n self.trans_unpack7e()\n self.set_mobile_crc_valid()\n return self.is_mobile_crc_valid()\n\n def get_valid_bytes(self):\n start = self.origin_bytes.find(b'~')\n #print(\"start = %d\" % start)\n if start == -1:\n return False\n end = self.origin_bytes[start+1:].find(b'~')\n #print(\"end = %d\" % end)\n if end == -1:\n return False\n self.origin_bytes = self.origin_bytes[start:end+2]\n return True\n\nif __name__ == \"__main__\":\n test = SwHeadPacket()\n str1 = test.packet()\n\n unit1 = SwMobileUnit(0xcccc, StructType.sArrayChar_Str, PyType._String, 20, 20, \"test-code\")\n unit2 = SwMobileUnit(0xccc1, StructType.sArrayChar_Str, PyType._String, 6, 6, \"python\")\n unit3 = SwMobileUnit(0xccc2, StructType.UInt4, PyType._Array, 32, 8, [12, 34])\n\n str1 += unit1.encode_unit()\n str1 += unit2.encode_unit()\n str1 += unit3.encode_unit()\n\n m_bytes = SwMobileBytes()\n m_bytes.transport_bytes(str1)\n str1 = m_bytes.get_trans_bytes()\n print(get_valid_bytes(str1))\n m_bytes.del_trans_bytes()\n\n print(\"len = %d\" % len(str1))\n for val in str1:\n print(hex(val), end=' ')\n\n m_bytes.original_bytes(str1)\n print(m_bytes.is_mobile_crc_valid())\n str1 = m_bytes.get_origin_bytes()\n m_bytes.del_origin_bytes()\n\n test2 = SwHeadPacket()\n test2.unpacket(str1)\n test2.print_data()\n\n unit1.decode_unit(str1[test2.head_len:])\n unit1.print_data()\n\n unit2.decode_unit(str1[test2.head_len+unit1.unit_len:])\n unit2.print_data()\n\n unit3.decode_unit(str1[test2.head_len+unit1.unit_len+unit2.unit_len:])\n unit3.print_data()\n\ndef packetHeadCommon():\n head = SwHeadPacket()\n send_bytes = head.packet()\n del head\n return send_bytes\n\ndef packetHeadOfSetCmd():\n pass\n\ndef packetHeadOfGetCmd():\n pass\n", "sub_path": "sw_protocol.py", "file_name": "sw_protocol.py", "file_ext": "py", "file_size_in_byte": 8426, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "enum.Enum", "line_number": 12, "usage_type": "name"}, {"api_name": "enum.unique", "line_number": 11, "usage_type": "name"}, {"api_name": "struct.pack", "line_number": 45, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 47, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 50, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 55, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 57, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 59, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 62, "usage_type": "call"}, {"api_name": "python_c_bytes_trans.StructType.UShort2", "line_number": 70, "usage_type": "attribute"}, {"api_name": "python_c_bytes_trans.StructType", "line_number": 70, "usage_type": "name"}, {"api_name": "python_c_bytes_trans.StructType.UChar1", "line_number": 74, "usage_type": "attribute"}, {"api_name": "python_c_bytes_trans.StructType", "line_number": 74, "usage_type": "name"}, {"api_name": "struct.pack", "line_number": 81, "usage_type": "call"}, {"api_name": "struct.pack", "line_number": 82, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 97, "usage_type": "call"}, {"api_name": "struct.unpack", "line_number": 102, "usage_type": "call"}, {"api_name": "python_c_bytes_trans.StructType.UChar1", "line_number": 113, "usage_type": "attribute"}, {"api_name": "python_c_bytes_trans.StructType", "line_number": 113, "usage_type": "name"}, {"api_name": "python_c_bytes_trans.StructType.UChar1", "line_number": 114, "usage_type": "attribute"}, {"api_name": "python_c_bytes_trans.StructType", "line_number": 114, "usage_type": "name"}, {"api_name": "python_c_bytes_trans.StructType.UInt4", "line_number": 115, "usage_type": "attribute"}, {"api_name": "python_c_bytes_trans.StructType", "line_number": 115, "usage_type": "name"}, {"api_name": "python_c_bytes_trans.StructType.UChar1", "line_number": 116, "usage_type": "attribute"}, {"api_name": "python_c_bytes_trans.StructType", "line_number": 116, "usage_type": "name"}, {"api_name": "python_c_bytes_trans.StructType.UShort2", "line_number": 117, "usage_type": "attribute"}, {"api_name": "python_c_bytes_trans.StructType", "line_number": 117, "usage_type": "name"}, {"api_name": "python_c_bytes_trans.StructType.UChar1", "line_number": 118, "usage_type": "attribute"}, {"api_name": "python_c_bytes_trans.StructType", "line_number": 118, "usage_type": "name"}, {"api_name": "python_c_bytes_trans.StructType.UChar1", "line_number": 119, "usage_type": "attribute"}, {"api_name": "python_c_bytes_trans.StructType", "line_number": 119, "usage_type": "name"}, {"api_name": "python_c_bytes_trans.StructType.UChar1", "line_number": 120, "usage_type": "attribute"}, {"api_name": "python_c_bytes_trans.StructType", "line_number": 120, "usage_type": "name"}, {"api_name": "python_c_bytes_trans.StructType.UChar1", "line_number": 121, "usage_type": "attribute"}, {"api_name": "python_c_bytes_trans.StructType", "line_number": 121, "usage_type": "name"}, {"api_name": "struct.pack", "line_number": 179, "usage_type": "call"}, {"api_name": "python_c_bytes_trans.StructType.UShort2", "line_number": 179, "usage_type": "attribute"}, {"api_name": "python_c_bytes_trans.StructType", "line_number": 179, "usage_type": "name"}, {"api_name": "struct.unpack", "line_number": 182, "usage_type": "call"}, {"api_name": "python_c_bytes_trans.StructType.UShort2", "line_number": 182, "usage_type": "attribute"}, {"api_name": "python_c_bytes_trans.StructType", "line_number": 182, "usage_type": "name"}, {"api_name": "python_c_bytes_trans.StructType.sArrayChar_Str", "line_number": 223, "usage_type": "attribute"}, {"api_name": "python_c_bytes_trans.StructType", "line_number": 223, "usage_type": "name"}, {"api_name": "python_c_bytes_trans.StructType.sArrayChar_Str", "line_number": 224, "usage_type": "attribute"}, {"api_name": "python_c_bytes_trans.StructType", "line_number": 224, "usage_type": "name"}, {"api_name": "python_c_bytes_trans.StructType.UInt4", "line_number": 225, "usage_type": "attribute"}, {"api_name": "python_c_bytes_trans.StructType", "line_number": 225, "usage_type": "name"}]}
+{"seq_id": "601668753", "text": "\"\"\"\nStarted during the summer period in 2014\n@author: Mustafa Erkan Basar\n\nThis script demostrates how event detection and time-to-event estimation can be applied on social media in the context of ADNEXT, INFITI, COMMIT project.\n\nThe event detection is a result of the research done by Florian Kunneman.\n\nThe time-to-event estimation module developed by Ali Hürriyetoglu.\n\nThe Program flow is as follows:\n 1- Connect to MongoDB (MangoLab)\n 2- Retrieve tweets from twiqs.nl every hour.\n 3- Call the event detection.\n 4- Clear the former results from the database.\n 5- Run the time-to-event estimation module.\n 6- Write the new results to database, after each time-to-event estimation. \n\n\nTrouble Shooting:\n - The url for twiqs.nl may change from time to time.\n - Twiqs.nl may not provide tweets in time. Therefore this particular hour will not be taken into account.\n\n\"\"\"\n\nimport configparser\n\nimport requests\nimport random\n\nimport time\nfrom datetime import date, datetime, timedelta\n\nimport pymongo\n\nimport DEvents.event_pairs as event_pairs\n\n\n#Get all the private configurations;\nconfig = configparser.ConfigParser()\nconfig.read('/home/ebasar/oauth.ini')\n\n#MongoLab OAuth;\nclient_host = config.get('LE_script_db', 'client_host')\nclient_port = int(config.get('LE_script_db', 'client_port'))\ndb_name = config.get('LE_script_db', 'db_name')\nuser_name = config.get('LE_script_db', 'user_name')\npasswd = config.get('LE_script_db', 'passwd')\n\n#Twiqs OAuth;\nuser_name2 = config.get('LE_script_twiqs', 'user_name')\npasswd2 = config.get('LE_script_twiqs', 'passwd')\n\n\n#!IDEA! = Add try-except block for the connection part;\n#MongoLab Connection;\nconnection = pymongo.MongoClient(client_host, client_port)\nledb = connection[db_name] #Database\nledb.authenticate(user_name, passwd)\nlecl = ledb.lecl #Collection\nprint(\"Connected to DB\")\n\n\nep = event_pairs.Event_pairs(\"all\",\"coco_out/\",\"tmp/\")\nprint(\"Event Detection Initialised\")\n\n\n#Get the cookie for twiqs.nl;\ns = requests.Session()\nr = s.post(\"http://145.100.57.182/cgi-bin/twitter\", data={\"NAME\":user_name2, \"PASSWD\":passwd2})\n\n\n#!IDEA! = Argparser can be used to get system parameters;\n#Twiqs.nl parameters;\npayload = {'SEARCH': 'echtalles', 'DATE': 'yyyymmddhh-yyyymmddhh', 'DOWNLOAD':True, 'SHOWTWEETS':True}\n#DATE = --> start and end should point to the same hour in order to get tweets about an hour\n\n\ndef RequestTweets():\n\t\"\"\"\n\tFetches the tweets from twiqs.nl\n\tWarning = The url may need to be updated from time to time!\n\t\"\"\"\n\toutput1st = requests.get(\"http://145.100.57.182/cgi-bin/twitter\", params=payload, cookies=s.cookies)\n\treturn output1st\n\n\n#If True, don't contain details of tweets except ids and users. Also don't contain the keyterms of events after keeping them in keylist.\nDeleteTweetDetails = True\n\n#If True, delete the former events from mongo db.\nDeleteFormerEvents = True\n\nwhile True:\n\ttime.sleep(120) #Check every two minutes if you are in the next hour.\n\n\t#Time Calculations;\n\tnowDate = datetime.now()\n\tnowDate_earlier = nowDate - timedelta(hours=1) #Get the previous hour. Because you can get tweets for the last hour from twiqs.nl.\n\ttweethour = nowDate_earlier.strftime(\"%H:00 %d-%m-%Y\") #Just for showing off the hour which tweets requested.\n\tnes = nowDate_earlier.strftime(\"%Y%m%d%H\") #'yyyymmddhh' twiqs.nl format.\n\tpDate = nes+'-'+nes #Twiqs.nl needs this format. Start and end time should be the same to retrieve tweets for one hour.\n\tcurrTime = nowDate.strftime(\"%H:%M\") #Just for showing off the minutes while waiting for the next hour.\n\n\t#Check if we are still in the same hour:\n\tif payload['DATE'] == pDate: #Continue waiting if you are in the same hour. Otherwise process the next hour.\n\t\tprint(currTime)\n\t\tcontinue\n\telse:\n\t\tpayload['DATE'] = pDate #It will remain the same until next hour.\n\t\tprint(\"Tweet hour : \" + tweethour)\n\n\t\t#Request to Twiqs;\n\t\toutput = RequestTweets()\n\t\tprint(\"Request Completed\")\n\n\t\t#Check the cookie;\n\t\twithoutcookie = '#user_id\\t#tweet_id\\t#DATE='+pDate+'\\t#SEARCHTOKEN=echtalles\\n'\n\t\tif output.text[:70] == withoutcookie: #if the cookie doesn't have access right to download the tweets, it will skip this hour.\n\t\t\tprint(\"Cookie is wrong. I'll skip tweets at \" + tweethour + \"You have to check your cookie configuration!\")\n\t\t\t#!IDEA! = If the cookie is wrong, write the code(call the relevant method) for getting a new one here.\n\t\t\tcontinue\n\t\telse:\n\t\t\tprint(\"Cookie is Fine.\")\n\n\t\t#Check the result of request;\n\t\tdumpoutput = '#user_id\\t#tweet_id\\t#date\\t#time\\t#reply_to_tweet_id\\t#retweet_to_tweet_id\\t#user_name\\t#tweet\\t#DATE='+pDate+'\\t#SEARCHTOKEN=echtalles\\n'\n\t\tif output.text[:1000] == dumpoutput: #If there isn't any tweet try the request again.\n\t\t\tprint(\"No tweet found at the first time! I'll try again\")\n\t\t\ttime.sleep(300)\n\t\t\toutput = RequestTweets()\n\t\t\tif output.text[:1000] == dumpoutput: #If there isn't any tweet again, it will skip this hour.\n\t\t\t\tprint(\"Still there is not any tweet! I'll skip tweets at \" + tweethour)\n\t\t\t\tcontinue\n\t\t\telse:\n\t\t\t\tprint(\"Tweets came at the second time\")\n\t\telse:\n\t\t\tprint(\"Tweets are O.K.\")\n\n\t\t#Event Detection; (refer to Florian Kunneman for any issue)\n\t\tEventDic = ep.detect_events(output.text[:-1]) # [:-1] = ignoring the last '\\n' at the bottom of the file.\n\t\tprint(\"Event Detection Completed\")\n\n\t\tif DeleteFormerEvents:\n\t\t\tlecl.remove({ }) #Delete the old events from database\n\t\t\tprint(\"Former events are deleted from the database\")\n\n\t\tfor k,v in EventDic.items(): #For every detected event\n\n\t\t\t#TimeToEventEstimation Calculations;\n\t\t\tcreateDate = datetime.now() #TTE Estimation will be added to the current time\n\t\t\trandomTTE = random.uniform(0.0, 193.0) #random number for estimation (for now)\n\t\t\thh, premm = divmod(randomTTE, 1)\n\t\t\tmm = (60*premm)*0.1\n\t\t\tv['Estimation'] = createDate + timedelta(hours=int(hh), minutes=int(mm))\n\n\t\t\t#Convert date formats to datetime format;\n\t\t\tv['date'] = datetime.combine(v['date'], datetime.min.time())\n\n\t\t\t#Writing keyterms in a list without keyterm scores; (In django using this list is more efficient)\n\t\t\tv['keylist'] = []\n\t\t\tfor m in v['keyterms']:\n\t\t\t\tmt = m[0].title() #capitalization\n\t\t\t\tv['keylist'].append(mt)\n\n\t\t\tif DeleteTweetDetails:\n\t\t\t\tdel v['keyterms']\n\t\t\t\tfor i in v['tweets']:\n\t\t\t\t\tdel i['date'], i['date references'], i['text'], i['entities']\n\t\t\telse:\n\t\t\t\t#If you don't delete details; convert date formats to datetime format;\n\t\t\t\tfor i in v['tweets']:\n\t\t\t\t\ti['date'] = datetime.combine(i['date'], datetime.min.time())\n\n\t\t\t#Write to database event by event;\n\t\t\tlecl.insert(v) \n\t\tprint(\"Written to Database\")\n\n\t\tcontinue\n\n\n", "sub_path": "LamaEvents.py", "file_name": "LamaEvents.py", "file_ext": "py", "file_size_in_byte": 6555, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "configparser.ConfigParser", "line_number": 40, "usage_type": "call"}, {"api_name": "pymongo.MongoClient", "line_number": 57, "usage_type": "call"}, {"api_name": "DEvents.event_pairs.Event_pairs", "line_number": 64, "usage_type": "call"}, {"api_name": "DEvents.event_pairs", "line_number": 64, "usage_type": "name"}, {"api_name": "requests.Session", "line_number": 69, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 84, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 95, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 98, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 98, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 99, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 130, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 151, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 151, "usage_type": "name"}, {"api_name": "random.uniform", "line_number": 152, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 155, "usage_type": "call"}, {"api_name": "datetime.datetime.combine", "line_number": 158, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 158, "usage_type": "name"}, {"api_name": "datetime.datetime.min.time", "line_number": 158, "usage_type": "call"}, {"api_name": "datetime.datetime.min", "line_number": 158, "usage_type": "attribute"}, {"api_name": "datetime.datetime.combine", "line_number": 173, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 173, "usage_type": "name"}, {"api_name": "datetime.datetime.min.time", "line_number": 173, "usage_type": "call"}, {"api_name": "datetime.datetime.min", "line_number": 173, "usage_type": "attribute"}]}
+{"seq_id": "224053149", "text": "import importlib\nimport json\nimport sys\nfrom os import listdir, environ, pathsep\nfrom os.path import isfile, join, splitext\nimport core.config.paths\nfrom core.helpers import list_apps\nfrom core.config.paths import keywords_path, graphviz_path\nfrom collections import OrderedDict\n\n\ndef load_config():\n global https\n self = sys.modules[__name__]\n with open(core.config.paths.config_path) as config_file:\n config = json.loads(config_file.read())\n for key, value in config.items():\n if value:\n if hasattr(core.config.paths, key):\n setattr(core.config.paths, key, value)\n elif hasattr(self, key):\n setattr(self, key, value)\n\n\ndef write_values_to_file(values=None):\n if values is None:\n values = [\"graphviz_path\", \"templates_path\", \"profile_visualizations_path\", \"keywords_path\", \"db_path\",\n \"tls_version\",\n \"certificate_path\", \"https\", \"private_key_path\", \"debug\", \"default_server\", \"host\", \"port\"]\n self = sys.modules[__name__]\n f = open(core.config.paths.config_path, \"r\")\n parsed = json.loads(f.read())\n f.close()\n for key in values:\n parsed[key] = getattr(self, key)\n\n with open(core.config.paths.config_path, \"w\") as f:\n json.dump(parsed, f)\n\n# Enables/Disables Browser Notifications\nnotifications = \"True\"\n\n# Path to graphviz location\nenviron[\"PATH\"] += (pathsep + graphviz_path)\n\n# Database Path\n\nreinitialize_case_db_on_startup = True\n\ntls_version = \"1.2\"\nhttps = \"false\"\n\ndebug = \"True\"\ndefault_server = \"True\"\nhost = \"127.0.0.1\"\nport = \"5000\"\n\n# Loads the keywords into the environment filter for use\nJINJA_GLOBALS = {splitext(fn)[0]: getattr(importlib.import_module(\"core.keywords.\" + splitext(fn)[0]), \"main\")\n for fn in listdir(keywords_path) if\n isfile(join(keywords_path, fn)) and not splitext(fn)[0] in [\"__init__\", \".\"]}\n\n# Active Execution (Workflows called from constant loop) settings.\n# secondsDelay - delay in seconds between execution loops\n# maxJobs - maximum number of jobs to be run at once\nexecution_settings = {\n \"secondsDelay\": 0.1,\n \"maxJobs\": 2\n}\n\nnum_threads = 5\nthreadpool_shutdown_timeout_sec = 3\n\n# Function Dict Paths/Initialization\n\nfunction_info = None\n\n\ndef load_function_info():\n global function_info\n try:\n with open(core.config.paths.function_info_path) as f:\n function_info = json.loads(f.read())\n app_funcs = {}\n for app in list_apps():\n with open(join(core.config.paths.apps_path, app, 'functions.json')) as function_file:\n app_funcs[app] = json.loads(function_file.read())\n function_info['apps'] = app_funcs\n\n except Exception as e:\n print(\"caught!\")\n print(e)\n\nload_config()\ntry:\n with open(core.config.paths.events_path) as f:\n possible_events = json.loads(f.read(), object_pairs_hook=OrderedDict)\nexcept (IOError, OSError):\n possible_events = {}\n\n\nload_function_info()\n\n\n# Function to set config value\ndef set(key, value):\n self = sys.modules[__name__]\n if hasattr(self, key):\n setattr(self, key, value)\n return True\n return False\n", "sub_path": "core/config/config.py", "file_name": "config.py", "file_ext": "py", "file_size_in_byte": 3215, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sys.modules", "line_number": 14, "usage_type": "attribute"}, {"api_name": "core.config.paths.config", "line_number": 15, "usage_type": "attribute"}, {"api_name": "core.config.paths", "line_number": 15, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 16, "usage_type": "call"}, {"api_name": "core.config.paths.config", "line_number": 19, "usage_type": "attribute"}, {"api_name": "core.config.paths", "line_number": 19, "usage_type": "name"}, {"api_name": "core.config.paths.config", "line_number": 20, "usage_type": "attribute"}, {"api_name": "core.config.paths", "line_number": 20, "usage_type": "name"}, {"api_name": "sys.modules", "line_number": 30, "usage_type": "attribute"}, {"api_name": "core.config.paths.config", "line_number": 31, "usage_type": "attribute"}, {"api_name": "core.config.paths", "line_number": 31, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 32, "usage_type": "call"}, {"api_name": "core.config.paths.config", "line_number": 37, "usage_type": "attribute"}, {"api_name": "core.config.paths", "line_number": 37, "usage_type": "name"}, {"api_name": "json.dump", "line_number": 38, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 44, "usage_type": "name"}, {"api_name": "os.pathsep", "line_number": 44, "usage_type": "name"}, {"api_name": "core.config.paths.graphviz_path", "line_number": 44, "usage_type": "name"}, {"api_name": "os.path.splitext", "line_number": 59, "usage_type": "call"}, {"api_name": "importlib.import_module", "line_number": 59, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 60, "usage_type": "call"}, {"api_name": "core.config.paths.keywords_path", "line_number": 60, "usage_type": "argument"}, {"api_name": "os.path.isfile", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 61, "usage_type": "call"}, {"api_name": "core.config.paths.keywords_path", "line_number": 61, "usage_type": "argument"}, {"api_name": "os.path.splitext", "line_number": 61, "usage_type": "call"}, {"api_name": "core.config.paths.config", "line_number": 82, "usage_type": "attribute"}, {"api_name": "core.config.paths", "line_number": 82, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 83, "usage_type": "call"}, {"api_name": "core.helpers.list_apps", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 86, "usage_type": "call"}, {"api_name": "core.config.paths.config", "line_number": 86, "usage_type": "attribute"}, {"api_name": "core.config.paths", "line_number": 86, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 87, "usage_type": "call"}, {"api_name": "core.config.paths.config", "line_number": 96, "usage_type": "attribute"}, {"api_name": "core.config.paths", "line_number": 96, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 97, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 97, "usage_type": "name"}, {"api_name": "sys.modules", "line_number": 107, "usage_type": "attribute"}]}
+{"seq_id": "388980124", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[101]:\n\n\nfrom numpy.random import seed\nimport csv\nimport sqlite3\nimport time\nimport numpy as np\nimport random\nimport pandas as pd\nfrom pandas import DataFrame\nimport scipy.sparse as sp\nimport math\nimport copy\n\nfrom sklearn.model_selection import KFold\nfrom sklearn.decomposition import PCA\nfrom sklearn.metrics import auc\nfrom sklearn.metrics import roc_auc_score\nfrom sklearn.metrics import accuracy_score\nfrom sklearn.metrics import recall_score\nfrom sklearn.metrics import f1_score\nfrom sklearn.metrics import precision_score\nfrom sklearn.metrics import precision_recall_curve\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.preprocessing import label_binarize\nfrom sklearn.svm import SVC\nfrom sklearn.model_selection import StratifiedKFold\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.ensemble import GradientBoostingClassifier\nfrom sklearn.decomposition import KernelPCA\n\nimport sys\nimport torch\nfrom torch import nn\nimport torch.optim as optim\nfrom torch.utils.data import Dataset\nfrom torch.utils.data import DataLoader\nfrom pytorchtools import EarlyStopping\nfrom pytorchtools import BalancedDataParallel\nfrom radam import RAdam\nimport torch.nn.functional as F\n\nimport networkx as nx\n\nimport warnings\nwarnings.filterwarnings(\"ignore\")\n\nimport os\nfrom tensorboardX import SummaryWriter\n\n\n# In[102]:\n\n\nseed=0\nrandom.seed(seed)\nos.environ['PYTHONHASHSEED'] = str(seed)\nnp.random.seed(seed)\ntorch.manual_seed(seed)\ntorch.cuda.manual_seed(seed)\ntorch.cuda.manual_seed_all(seed)\ntorch.backends.cudnn.deterministic = True\n\n\n# In[103]:\n\n\ndef prepare(df_drug, feature_list,mechanism,action,drugA,drugB):\n\n d_label = {}\n d_feature = {}\n\n # Transfrom the interaction event to number\n d_event=[]\n for i in range(len(mechanism)):\n d_event.append(mechanism[i]+\" \"+action[i])\n\n\n count={}\n for i in d_event:\n if i in count:\n count[i]+=1\n else:\n count[i]=1\n event_num=len(count)\n list1 = sorted(count.items(), key=lambda x: x[1],reverse=True)\n for i in range(len(list1)):\n d_label[list1[i][0]]=i\n\n\n vector = np.zeros((len(np.array(df_drug['name']).tolist()), 0), dtype=float) #vector=[]\n for i in feature_list:\n #vector = np.hstack((vector, feature_vector(i, df_drug, vector_size)))#1258*1258\n tempvec=feature_vector(i, df_drug)\n vector = np.hstack((vector,tempvec))\n # Transfrom the drug ID to feature vector\n for i in range(len(np.array(df_drug['name']).tolist())):\n d_feature[np.array(df_drug['name']).tolist()[i]] = vector[i]\n\n # Use the dictionary to obtain feature vector and label\n new_feature = []\n new_label = []\n\n for i in range(len(d_event)):\n temp=np.hstack((d_feature[drugA[i]],d_feature[drugB[i]]))\n new_feature.append(temp)\n new_label.append(d_label[d_event[i]])\n\n \n new_feature = np.array(new_feature) #323539*....\n new_label = np.array(new_label) #323539\n\n return new_feature, new_label, drugA,drugB,event_num\n\n\n# In[104]:\n\n\ndef feature_vector(feature_name, df):\n def Jaccard(matrix):\n matrix = np.mat(matrix)\n\n numerator = matrix * matrix.T\n\n denominator = np.ones(np.shape(matrix)) * matrix.T + matrix * np.ones(np.shape(matrix.T)) - matrix * matrix.T\n\n return numerator / denominator\n \n all_feature = []\n drug_list = np.array(df[feature_name]).tolist()\n # Features for each drug, for example, when feature_name is target, drug_list=[\"P30556|P05412\",\"P28223|P46098|……\"]\n for i in drug_list:\n for each_feature in i.split('|'):\n if each_feature not in all_feature:\n all_feature.append(each_feature) # obtain all the features\n feature_matrix = np.zeros((len(drug_list), len(all_feature)), dtype=float)\n df_feature = DataFrame(feature_matrix, columns=all_feature) # Consrtuct feature matrices with key of dataframe\n for i in range(len(drug_list)):\n for each_feature in df[feature_name].iloc[i].split('|'):\n df_feature[each_feature].iloc[i] = 1\n \n df_feature = np.array(df_feature)\n sim_matrix = np.array(Jaccard(df_feature))\n \n print(feature_name+\" len is:\"+ str(len(sim_matrix[0])))\n return sim_matrix\n\n\n# In[105]:\n\n\nclass DDIDataset(Dataset):\n def __init__(self,x,y):\n self.len=len(x)\n self.x_data=torch.from_numpy(x)\n\n self.y_data=torch.from_numpy(y)\n def __getitem__(self,index):\n return self.x_data[index],self.y_data[index]\n def __len__(self):\n return self.len\n\n\n# In[106]:\n\n\nclass MultiHeadAttention(torch.nn.Module):\n def __init__(self,input_dim,n_heads,ouput_dim=None):\n \n super(MultiHeadAttention, self).__init__()\n self.d_k=self.d_v=input_dim//n_heads\n self.n_heads = n_heads\n if ouput_dim==None:\n self.ouput_dim=input_dim\n else:\n self.ouput_dim=ouput_dim\n self.W_Q = torch.nn.Linear(input_dim, self.d_k * self.n_heads, bias=False)\n self.W_K = torch.nn.Linear(input_dim, self.d_k * self.n_heads, bias=False)\n self.W_V = torch.nn.Linear(input_dim, self.d_v * self.n_heads, bias=False)\n self.fc = torch.nn.Linear(self.n_heads * self.d_v, self.ouput_dim, bias=False)\n def forward(self,X):\n ## (S, D) -proj-> (S, D_new) -split-> (S, H, W) -trans-> (H, S, W)\n Q=self.W_Q(X).view( -1, self.n_heads, self.d_k).transpose(0,1)\n K=self.W_K(X).view( -1, self.n_heads, self.d_k).transpose(0,1)\n V=self.W_V(X).view( -1, self.n_heads, self.d_v).transpose(0,1)\n \n scores = torch.matmul(Q, K.transpose(-1, -2)) / np.sqrt(self.d_k)\n # context: [n_heads, len_q, d_v], attn: [n_heads, len_q, len_k]\n attn = torch.nn.Softmax(dim=-1)(scores)\n context = torch.matmul(attn, V)\n # context: [len_q, n_heads * d_v]\n context = context.transpose(1, 2).reshape(-1, self.n_heads * self.d_v)\n output = self.fc(context)\n return output\n\n\n# In[107]:\n\n\nclass EncoderLayer(torch.nn.Module):\n def __init__(self,input_dim,n_heads):\n super(EncoderLayer, self).__init__()\n self.attn = MultiHeadAttention(input_dim,n_heads)\n self.AN1=torch.nn.LayerNorm(input_dim)\n \n self.l1=torch.nn.Linear(input_dim, input_dim)\n self.AN2=torch.nn.LayerNorm(input_dim)\n def forward (self,X):\n \n output=self.attn(X)\n X=self.AN1(output+X)\n \n output=self.l1(X)\n X=self.AN2(output+X)\n \n return X\n\n\n# In[108]:\n\n\ndef gelu(x):\n return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0)))\n\n\n# In[109]:\n\n\nclass AE1(torch.nn.Module): #Joining together\n def __init__(self,vector_size):\n super(AE1,self).__init__()\n \n self.vector_size=vector_size\n \n self.l1 = torch.nn.Linear(self.vector_size,(self.vector_size+len_after_AE)//2)\n self.bn1 = torch.nn.BatchNorm1d((self.vector_size+len_after_AE)//2)\n \n self.att2=EncoderLayer((self.vector_size+len_after_AE)//2,bert_n_heads)\n self.l2 = torch.nn.Linear((self.vector_size+len_after_AE)//2,len_after_AE)\n \n self.l3 = torch.nn.Linear(len_after_AE,(self.vector_size+len_after_AE)//2)\n self.bn3 = torch.nn.BatchNorm1d((self.vector_size+len_after_AE)//2)\n \n self.l4 = torch.nn.Linear((self.vector_size+len_after_AE)//2,self.vector_size)\n \n self.dr = torch.nn.Dropout(drop_out_rating)\n self.ac=gelu\n \n def forward(self,X):\n \n X=self.dr(self.bn1(self.ac(self.l1(X))))\n \n X=self.att2(X)\n X=self.l2(X)\n \n X_AE=self.dr(self.bn3(self.ac(self.l3(X))))\n \n X_AE=self.l4(X_AE)\n \n return X,X_AE\n\n\n# In[110]:\n\n\nclass AE2(torch.nn.Module):# twin network\n def __init__(self,vector_size):\n super(AE2,self).__init__()\n \n self.vector_size=vector_size//2\n \n self.l1 = torch.nn.Linear(self.vector_size,(self.vector_size+len_after_AE//2)//2)\n self.bn1 = torch.nn.BatchNorm1d((self.vector_size+len_after_AE//2)//2)\n \n self.att2=EncoderLayer((self.vector_size+len_after_AE//2)//2,bert_n_heads)\n self.l2 = torch.nn.Linear((self.vector_size+len_after_AE//2)//2,len_after_AE//2)\n \n self.l3 = torch.nn.Linear(len_after_AE//2,(self.vector_size+len_after_AE//2)//2)\n self.bn3 = torch.nn.BatchNorm1d((self.vector_size+len_after_AE//2)//2)\n \n self.l4 = torch.nn.Linear((self.vector_size+len_after_AE//2)//2,self.vector_size)\n \n self.dr = torch.nn.Dropout(drop_out_rating)\n \n self.ac=gelu\n \n def forward(self,X):\n \n X1=X[:,0:self.vector_size]\n X2=X[:,self.vector_size:]\n \n X1=self.dr(self.bn1(self.ac(self.l1(X1))))\n X1=self.att2(X1)\n X1=self.l2(X1)\n X_AE1=self.dr(self.bn3(self.ac(self.l3(X1))))\n X_AE1=self.l4(X_AE1)\n \n X2=self.dr(self.bn1(self.ac(self.l1(X2))))\n X2=self.att2(X2)\n X2=self.l2(X2)\n X_AE2=self.dr(self.bn3(self.ac(self.l3(X2))))\n X_AE2=self.l4(X_AE2)\n \n X=torch.cat((X1,X2), 1)\n X_AE=torch.cat((X_AE1,X_AE2), 1)\n \n return X,X_AE\n\n\n# In[111]:\n\n\nclass cov(torch.nn.Module):\n def __init__(self,vector_size):\n super(cov,self).__init__()\n \n self.vector_size=vector_size\n \n self.co2_1=torch.nn.Conv2d(1, 1, kernel_size=(2,cov2KerSize))\n self.co1_1=torch.nn.Conv1d(1, 1, kernel_size=cov1KerSize)\n self.pool1=torch.nn.AdaptiveAvgPool1d(len_after_AE)\n \n self.ac=gelu\n \n \n def forward(self,X):\n \n X1=X[:,0:self.vector_size//2]\n X2=X[:,self.vector_size//2:]\n \n X=torch.cat((X1,X2), 0)\n \n X=X.view(-1,1,2,self.vector_size//2)\n \n X=self.ac(self.co2_1(X))\n \n X=X.view(-1,self.vector_size//2-cov2KerSize+1, 1) \n X=X.permute(0,2,1)\n X=self.ac(self.co1_1(X))\n \n X=self.pool1(X)\n \n X=X.contiguous().view(-1,len_after_AE)\n \n return X\n\n\n# In[112]:\n\n\nclass ADDAE(torch.nn.Module):\n def __init__(self,vector_size):\n super(ADDAE,self).__init__()\n \n self.vector_size=vector_size//2\n \n self.l1 = torch.nn.Linear(self.vector_size,(self.vector_size+len_after_AE)//2)\n self.bn1 = torch.nn.BatchNorm1d((self.vector_size+len_after_AE)//2)\n \n self.att1=EncoderLayer((self.vector_size+len_after_AE)//2,bert_n_heads)\n self.l2 = torch.nn.Linear((self.vector_size+len_after_AE)//2,len_after_AE)\n #self.att2=EncoderLayer(len_after_AE//2,bert_n_heads)\n \n self.l3 = torch.nn.Linear(len_after_AE,(self.vector_size+len_after_AE)//2)\n self.bn3 = torch.nn.BatchNorm1d((self.vector_size+len_after_AE)//2)\n \n self.l4 = torch.nn.Linear((self.vector_size+len_after_AE)//2,self.vector_size)\n \n self.dr = torch.nn.Dropout(drop_out_rating)\n \n self.ac=gelu\n \n def forward(self,X):\n \n X1=X[:,0:self.vector_size]\n X2=X[:,self.vector_size:]\n X=X1+X2\n \n X=self.dr(self.bn1(self.ac(self.l1(X))))\n \n X=self.att1(X)\n X=self.l2(X)\n \n X_AE=self.dr(self.bn3(self.ac(self.l3(X))))\n \n X_AE=self.l4(X_AE)\n X_AE=torch.cat((X_AE,X_AE), 1)\n \n return X,X_AE\n\n\n# In[113]:\n\n\nclass BERT(torch.nn.Module):\n def __init__(self,input_dim,n_heads,n_layers,event_num):\n super(BERT, self).__init__()\n \n self.ae1=AE1(input_dim) #Joining together\n self.ae2=AE2(input_dim)#twin loss\n self.cov=cov(input_dim)#cov \n self.ADDAE=ADDAE(input_dim)\n \n self.dr = torch.nn.Dropout(drop_out_rating)\n self.input_dim=input_dim\n \n self.layers = torch.nn.ModuleList([EncoderLayer(len_after_AE*5,n_heads) for _ in range(n_layers)])\n self.AN=torch.nn.LayerNorm(len_after_AE*5)\n \n self.l1=torch.nn.Linear(len_after_AE*5,(len_after_AE*5+event_num)//2)\n self.bn1=torch.nn.BatchNorm1d((len_after_AE*5+event_num)//2)\n \n self.l2=torch.nn.Linear((len_after_AE*5+event_num)//2,event_num)\n \n self.ac=gelu\n \n def forward(self, X):\n X1,X_AE1=self.ae1(X)\n X2,X_AE2=self.ae2(X)\n \n X3=self.cov(X)\n \n X4,X_AE4=self.ADDAE(X)\n \n X5=X1+X2+X3+X4\n \n X=torch.cat((X1,X2,X3,X4,X5), 1)\n \n for layer in self.layers:\n X = layer(X)\n X=self.AN(X)\n \n X=self.dr(self.bn1(self.ac(self.l1(X))))\n \n X=self.l2(X)\n \n return X,X_AE1,X_AE2,X_AE4\n\n\n# In[114]:\n\n\nclass focal_loss(nn.Module):\n def __init__(self, gamma=2):\n \n super(focal_loss,self).__init__()\n \n self.gamma = gamma\n\n def forward(self, preds, labels):\n \n # assert preds.dim() == 2 and labels.dim()==1\n labels = labels.view(-1, 1) # [B * S, 1]\n preds = preds.view(-1, preds.size(-1)) # [B * S, C]\n \n preds_logsoft = F.log_softmax(preds, dim=1) # 先softmax, 然后取log\n preds_softmax = torch.exp(preds_logsoft) # softmax\n\n preds_softmax = preds_softmax.gather(1, labels) # 这部分实现nll_loss ( crossempty = log_softmax + nll )\n preds_logsoft = preds_logsoft.gather(1, labels)\n \n loss = -torch.mul(torch.pow((1-preds_softmax), self.gamma), preds_logsoft) # torch.pow((1-preds_softmax), self.gamma) 为focal loss中 (1-pt)**γ\n \n loss = loss.mean()\n \n return loss\nclass my_loss1(nn.Module):\n def __init__(self):\n \n super(my_loss1,self).__init__()\n \n self.criteria1 = torch.nn.CrossEntropyLoss()\n self.criteria2=torch.nn.MSELoss()\n\n def forward(self, X, target,inputs,X_AE1,X_AE2,X_AE4):\n\n\n \n loss=calssific_loss_weight*self.criteria1(X,target)+\\\n self.criteria2(inputs.float(),X_AE1)+\\\n self.criteria2(inputs.float(),X_AE2)+\\\n self.criteria2(inputs.float(),X_AE4)\n return loss\nclass my_loss2(nn.Module):\n def __init__(self):\n \n super(my_loss2,self).__init__()\n \n self.criteria1 = focal_loss()\n self.criteria2=torch.nn.MSELoss()\n\n def forward(self, X, target,inputs,X_AE1,X_AE2,X_AE4):\n\n loss=calssific_loss_weight*self.criteria1(X,target)+\\\n self.criteria2(inputs.float(),X_AE1)+\\\n self.criteria2(inputs.float(),X_AE2)+\\\n self.criteria2(inputs.float(),X_AE4)\n return loss\n\n\n\n\ndef mixup(x1, x2, y1, y2, alpha):\n beta = np.random.beta(alpha, alpha)\n x = beta * x1 + (1 - beta) * x2\n y = beta * y1 + (1 - beta) * y2\n return x, y\n\n#writer = SummaryWriter(\"./tbx\")\ndef BERT_train(model,x_train,y_train,x_test,y_test,event_num):\n\n model_optimizer=RAdam(model.parameters(),lr=learn_rating,weight_decay=weight_decay_rate)\n model=torch.nn.DataParallel(model)\n model=model.to(device)\n\n x_train=np.vstack((x_train,np.hstack((x_train[:,len(x_train[0])//2:],x_train[:,:len(x_train[0])//2]))))\n y_train = np.hstack((y_train, y_train))\n np.random.seed(seed)\n np.random.shuffle(x_train)\n np.random.seed(seed)\n np.random.shuffle(y_train)\n\n len_train=len(y_train)\n len_test=len(y_test)\n print(\"arg train len\", len(y_train))\n print(\"test len\", len(y_test))\n\n\n train_dataset = DDIDataset(x_train,np.array(y_train))\n test_dataset = DDIDataset(x_test,np.array(y_test))\n train_loader=DataLoader(dataset=train_dataset,batch_size=batch_size,shuffle=True)\n test_loader=DataLoader(dataset=test_dataset,batch_size=batch_size,shuffle=False)\n\n\n for epoch in range(epo_num):\n if epoch Epoch: {} Average loss: {:.4f}'.format(epoch, train_loss / len(train_loader.dataset)))\r\n torch.save(VAEs[cat].state_dict(), 'vae' + str(cat) + '.pt')\r\n\r\n\r\ndef test():\r\n for i in range(10):\r\n VAEs[i].eval()\r\n test_loss = 0\r\n with torch.no_grad():\r\n for data, _ in test_loader:\r\n data = data.cuda()\r\n recon, mu, log_var = VAEs[i](data)\r\n\r\n # sum up batch loss\r\n test_loss += loss_function(recon, data, mu, log_var).item()\r\n\r\n test_loss /= len(test_loader.dataset)\r\n print('====> Test set loss: {:.4f}'.format(test_loss))\r\n\r\n#uncomment if you have trained models\r\n# for i in range(10):\r\n# VAEs[i].load_state_dict(torch.load('vae' + str(i) + '.pt'))\r\n\r\n#comment if you don't want to train\r\nfor e in range(30):\r\n for c in range(10):\r\n train(e,c)\r\n # test()\r\nimport matplotlib.pyplot as plt\r\nimport cv2\r\nimport numpy as np\r\nimport torchvision\r\nplt.ion()\r\nimport datetime\r\n\r\ncolonIm = np.zeros((28,8))\r\ncolonIm[7:11,2:6] = 1\r\ncolonIm[17:21,2:6] = 1\r\n\r\nOldZ = [None]*10\r\nOldZ2 = [None]*6\r\nwith torch.no_grad():\r\n z = torch.randn(6, 2).cuda()\r\n while(True):\r\n for batch_idx, (data, label) in enumerate(train_loader):\r\n\r\n now = datetime.datetime.now()\r\n h = now.hour\r\n m = now.minute\r\n s = now.second\r\n DIGITS = [h/10,h%10,m/10,m%10,s/10,s%10]\r\n # print(DIGITS)\r\n DIGITS_IMS = []\r\n\r\n\r\n\r\n for d in range(len(DIGITS)):\r\n # print(z)\r\n if (mode == 'interpolation'):\r\n inds = (label==DIGITS[d]).nonzero()\r\n ind = inds[np.random.randint(len(inds)),0]\r\n ind2 = inds[np.random.randint(len(inds)),0]\r\n mean,std = VAEs[int(DIGITS[d])].encoder(data[ind].reshape((-1,28*28)).cuda())\r\n mean2,std2 = VAEs[int(DIGITS[d])].encoder(data[ind2].reshape((-1,28*28)).cuda())\r\n zreal2 = VAEs[int(DIGITS[d])].sampling(mean2, std2)\r\n if (OldZ[int(DIGITS[d])] is None):\r\n\r\n zreal = VAEs[int(DIGITS[d])].sampling(mean, std)\r\n OldZ[int(DIGITS[d])] = zreal2\r\n else:\r\n zreal = OldZ[int(DIGITS[d])]\r\n OldZ[int(DIGITS[d])] = zreal2\r\n # z = torch.randn(1, 2).cuda()\r\n\r\n else:\r\n zreal2 = torch.randn(1, 10).cuda()\r\n if (OldZ2[d] is None):\r\n zreal = torch.randn(1, 10).cuda()\r\n\r\n else:\r\n zreal = OldZ2[d]\r\n\r\n\r\n n = (torch.randn(1, 10) * 0.1).cuda()\r\n rand0_1 = 0.2 #the smaller it is the shorter the random walk step\r\n z = (1-rand0_1)*zreal + rand0_1*zreal2\r\n # z = zreal + (rand0_1 * zreal2)\r\n OldZ2[d] = z\r\n sample = VAEs[int(DIGITS[d])].decoder(z).cuda()\r\n IM = sample.view(28, 28).cpu().numpy()\r\n DIGITS_IMS.append(IM)\r\n if (len(DIGITS_IMS) in [2,5,6]):\r\n DIGITS_IMS.append(colonIm)\r\n mosaic = np.concatenate(DIGITS_IMS,1)\r\n IM = cv2.resize(mosaic, dsize=(250*6, 250), interpolation=cv2.INTER_CUBIC)\r\n cv2.imshow('image', IM)\r\n cv2.waitKey(1)\r\n\r\n", "sub_path": "VAE_CLOCK.py", "file_name": "VAE_CLOCK.py", "file_ext": "py", "file_size_in_byte": 6517, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "torchvision.datasets.MNIST", "line_number": 11, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 11, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 11, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 11, "usage_type": "name"}, {"api_name": "torchvision.datasets.MNIST", "line_number": 12, "usage_type": "call"}, {"api_name": "torchvision.datasets", "line_number": 12, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 12, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 12, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 15, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 16, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 19, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 24, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 25, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 29, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 30, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 31, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 34, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.exp", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.randn_like", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.nn.functional.relu", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 44, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 45, "usage_type": "name"}, {"api_name": "torch.nn.functional.sigmoid", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 46, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 62, "usage_type": "name"}, {"api_name": "torch.nn.functional.binary_cross_entropy", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 65, "usage_type": "name"}, {"api_name": "torch.sum", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ion", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 121, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 124, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 130, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 131, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 135, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 135, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 149, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 150, "usage_type": "attribute"}, {"api_name": "torch.randn", "line_number": 164, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 166, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 182, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 183, "usage_type": "call"}, {"api_name": "cv2.INTER_CUBIC", "line_number": 183, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 184, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 185, "usage_type": "call"}]}
+{"seq_id": "442670280", "text": "#!/usr/bin/python\n\n\nfrom __future__ import (absolute_import, division, print_function)\n__metaclass__ = type\n\nANSIBLE_METADATA = {\n 'metadata_version': '1.1',\n 'status': ['preview'],\n 'supported_by': 'community'\n}\n\nDOCUMENTATION = '''\n---\nmodule: read_resources\n\nshort_description: Import OpenStack network\n\nversion_added: \"2.9.0\"\n\nauthor: \"OpenStack tenant migration tools (@os-migrate)\"\n\ndescription:\n - \"Read an OS-Migrate YAML resources file structure\"\n\noptions:\n path:\n description:\n - Resources YAML file to read.\n required: true\n type: str\n'''\n\nEXAMPLES = '''\n- name: Read resources from /opt/os-migrate/networks.yml\n os_migrate.os_migrate.read_resources:\n path: /opt/os-migrate/networks.yml\n register: read_networks\n\n- name: Debug-print resources\n debug:\n msg: \"{{ read_networks.resources }}\"\n'''\n\nRETURN = '''\nresources:\n description: List of resources deserialized from YAML file\n returned: success\n type: complex\n contains:\n type:\n description: Type of the resource.\n returned: success\n type: str\n params:\n description: Resource parameters important for import.\n returned: success\n type: dict\n info:\n description: Additional resource information, not needed for import.\n returned: success\n type: dict\n'''\n\nfrom ansible.module_utils.basic import AnsibleModule\n\nfrom ansible_collections.os_migrate.os_migrate.plugins.module_utils import filesystem\n\n\ndef run_module():\n module_args = dict(\n path=dict(type='str', required=True),\n )\n\n result = dict(\n # This module doesn't change anything.\n changed=False,\n )\n\n module = AnsibleModule(\n argument_spec=module_args,\n # Module doesn't change anything, we can let it run as-is in\n # check mode.\n supports_check_mode=True,\n )\n\n struct = filesystem.load_resources_file(module.params['path'])\n result['resources'] = struct['resources']\n\n module.exit_json(**result)\n\n\ndef main():\n run_module()\n\n\nif __name__ == '__main__':\n main()\n", "sub_path": "os_migrate/plugins/modules/read_resources.py", "file_name": "read_resources.py", "file_ext": "py", "file_size_in_byte": 2126, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "ansible.module_utils.basic.AnsibleModule", "line_number": 80, "usage_type": "call"}, {"api_name": "ansible_collections.os_migrate.os_migrate.plugins.module_utils.filesystem.load_resources_file", "line_number": 87, "usage_type": "call"}, {"api_name": "ansible_collections.os_migrate.os_migrate.plugins.module_utils.filesystem", "line_number": 87, "usage_type": "name"}]}
+{"seq_id": "318525496", "text": "import os\nimport sys\nimport random\nimport time\n\nimport torch\nfrom torch import nn\nimport torch.nn.functional as F\n\nsys.path.insert(0, '.')\nsys.path.insert(0, '..')\n\nfrom torch_geometric.datasets import ModelNet10PC # noqa\nfrom torch_geometric.utils import DataLoader # noqa\nfrom torch_geometric.transform import NormalizeScale, CartesianAdj # noqa\nfrom torch_geometric.nn.modules import SplineConv # noqa\nfrom torch_geometric.nn.functional import (sparse_voxel_max_pool,\n dense_voxel_max_pool) # noqa\n\npath = os.path.dirname(os.path.realpath(__file__))\npath = os.path.join(path, '..', 'data', 'ModelNet10PC')\n\ntrain_dataset = ModelNet10PC(path, True, transform=NormalizeScale())\ntest_dataset = ModelNet10PC(path, False, transform=NormalizeScale())\ntrain_loader = DataLoader(train_dataset, batch_size=32, shuffle=True)\ntest_loader = DataLoader(test_dataset, batch_size=32)\n\n\nclass Net(nn.Module):\n def __init__(self):\n super(Net, self).__init__()\n self.conv1 = SplineConv(1, 64, dim=3, kernel_size=5)\n self.conv2 = SplineConv(64, 64, dim=3, kernel_size=5)\n self.conv3 = SplineConv(64, 128, dim=3, kernel_size=5)\n self.fc1 = nn.Linear(8 * 128, 256)\n self.fc2 = nn.Linear(256, 10)\n\n def pool_args(self, mean, x):\n if not self.training:\n return 1 / mean, 0\n size = 1 / random.uniform(mean - x, mean + x)\n start = random.uniform(-1 / (mean - x), 0)\n return size, start\n\n def forward(self, data):\n # data.input = F.elu(self.conv1(data.adj, data.input))\n size, start = self.pool_args(13, 5)\n data, _ = sparse_voxel_max_pool(data, size, start, CartesianAdj())\n\n data.input = F.elu(self.conv1(data.adj, data.input))\n size, start = self.pool_args(7, 3)\n data, _ = sparse_voxel_max_pool(data, size, start, CartesianAdj())\n\n data.input = F.elu(self.conv2(data.adj, data.input))\n size, start = self.pool_args(5, 2)\n data, _ = sparse_voxel_max_pool(data, size, start, CartesianAdj())\n\n data.input = F.elu(self.conv3(data.adj, data.input))\n data, _ = dense_voxel_max_pool(data, 1 / 2, 0, 1)\n\n x = data.input.view(-1, self.fc1.weight.size(1))\n x = F.elu(self.fc1(x))\n x = F.dropout(x, training=self.training)\n x = self.fc2(x)\n return F.log_softmax(x, dim=1)\n\n\nmodel = Net()\nif torch.cuda.is_available():\n model.cuda()\n\noptimizer = torch.optim.Adam(model.parameters(), lr=0.001)\n\n\ndef train(epoch):\n model.train()\n\n if epoch == 8:\n for param_group in optimizer.param_groups:\n param_group['lr'] = 0.0001\n\n for data in train_loader:\n data = data.cuda().to_variable()\n optimizer.zero_grad()\n t_forward = time.perf_counter()\n output = model(data)\n loss = F.nll_loss(output, data.target)\n torch.cuda.synchronize()\n t_forward = time.perf_counter() - t_forward\n t_backward = time.perf_counter()\n loss.backward()\n torch.cuda.synchronize()\n t_backward = time.perf_counter() - t_backward\n optimizer.step()\n\n\ndef test(epoch, loader, dataset, str):\n model.eval()\n correct = 0\n\n for data in loader:\n data = data.cuda().to_variable(['input'])\n pred = model(data).data.max(1)[1]\n correct += pred.eq(data.target).sum()\n\n print('Epoch:', epoch, str, 'Accuracy:', correct / len(dataset))\n\n\nfor epoch in range(1, 21):\n train(epoch)\n test(epoch, train_loader, train_dataset, 'Train')\n test(epoch, test_loader, test_dataset, 'Test')\n", "sub_path": "alpha/modelnet10_pc.py", "file_name": "modelnet10_pc.py", "file_ext": "py", "file_size_in_byte": 3600, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sys.path.insert", "line_number": 10, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "sys.path.insert", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "torch_geometric.datasets.ModelNet10PC", "line_number": 23, "usage_type": "call"}, {"api_name": "torch_geometric.transform.NormalizeScale", "line_number": 23, "usage_type": "call"}, {"api_name": "torch_geometric.datasets.ModelNet10PC", "line_number": 24, "usage_type": "call"}, {"api_name": "torch_geometric.transform.NormalizeScale", "line_number": 24, "usage_type": "call"}, {"api_name": "torch_geometric.utils.DataLoader", "line_number": 25, "usage_type": "call"}, {"api_name": "torch_geometric.utils.DataLoader", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 29, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 29, "usage_type": "name"}, {"api_name": "torch_geometric.nn.modules.SplineConv", "line_number": 32, "usage_type": "call"}, {"api_name": "torch_geometric.nn.modules.SplineConv", "line_number": 33, "usage_type": "call"}, {"api_name": "torch_geometric.nn.modules.SplineConv", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn.Linear", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 36, "usage_type": "name"}, {"api_name": "random.uniform", "line_number": 41, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 42, "usage_type": "call"}, {"api_name": "torch_geometric.nn.functional.sparse_voxel_max_pool", "line_number": 48, "usage_type": "call"}, {"api_name": "torch_geometric.transform.CartesianAdj", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.nn.functional.elu", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 50, "usage_type": "name"}, {"api_name": "torch_geometric.nn.functional.sparse_voxel_max_pool", "line_number": 52, "usage_type": "call"}, {"api_name": "torch_geometric.transform.CartesianAdj", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.nn.functional.elu", "line_number": 54, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 54, "usage_type": "name"}, {"api_name": "torch_geometric.nn.functional.sparse_voxel_max_pool", "line_number": 56, "usage_type": "call"}, {"api_name": "torch_geometric.transform.CartesianAdj", "line_number": 56, "usage_type": "call"}, {"api_name": "torch.nn.functional.elu", "line_number": 58, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 58, "usage_type": "name"}, {"api_name": "torch_geometric.nn.functional.dense_voxel_max_pool", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.nn.functional.elu", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 62, "usage_type": "name"}, {"api_name": "torch.nn.functional.dropout", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 63, "usage_type": "name"}, {"api_name": "torch.nn.functional.log_softmax", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 65, "usage_type": "name"}, {"api_name": "torch.cuda.is_available", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 69, "usage_type": "attribute"}, {"api_name": "torch.optim.Adam", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 72, "usage_type": "attribute"}, {"api_name": "time.perf_counter", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.nn.functional.nll_loss", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 87, "usage_type": "name"}, {"api_name": "torch.cuda.synchronize", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 88, "usage_type": "attribute"}, {"api_name": "time.perf_counter", "line_number": 89, "usage_type": "call"}, {"api_name": "time.perf_counter", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.cuda.synchronize", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 92, "usage_type": "attribute"}, {"api_name": "time.perf_counter", "line_number": 93, "usage_type": "call"}]}
+{"seq_id": "313088224", "text": "import time\n\nfrom selenium import webdriver\nfrom selenium.webdriver import ActionChains\nfrom selenium.webdriver.chrome.service import Service\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.common.keys import Keys\n\n# driver = webdriver.Chrome(executable_path='C:\\\\_teach\\\\Silenium\\\\chromedriver.exe')\ns = Service('C:/_teach/Silenium/chromedriver.exe')\ndriver = webdriver.Chrome(service=s)\nbase_url = 'https://demoqa.com/buttons/'\n# login_standard_user = \"standard_user\"\n# login_password_user = \"secret_sauce\"\ndriver.get(base_url)\n# driver.maximize_window()\n\naction = ActionChains(driver)\nelements_button = driver.find_element(By.XPATH, \"//button[@id='doubleClickBtn']\")\naction.double_click(elements_button).perform()\nresult_duble_click = driver.find_element(By.XPATH, \"//p[@id='doubleClickMessage']\").text\nassert result_duble_click == 'You have done a double click'\nprint('Двжды кликнули')\ntime.sleep(2)\n\naction = ActionChains(driver)\nelements_button = driver.find_element(By.XPATH, \"//button[@id='rightClickBtn']\")\naction.context_click(elements_button).perform()\nresult_context_click = driver.find_element(By.XPATH, \"//p[@id='rightClickMessage']\").text\nassert result_context_click == 'You have done a right click'\nprint('Правой кликнули')\ntime.sleep(2)\n", "sub_path": "test_9_double_click.py", "file_name": "test_9_double_click.py", "file_ext": "py", "file_size_in_byte": 1301, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "selenium.webdriver.chrome.service.Service", "line_number": 10, "usage_type": "call"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 11, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 11, "usage_type": "name"}, {"api_name": "selenium.webdriver.ActionChains", "line_number": 18, "usage_type": "call"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 19, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 19, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 21, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 21, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 24, "usage_type": "call"}, {"api_name": "selenium.webdriver.ActionChains", "line_number": 26, "usage_type": "call"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 27, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 27, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.XPATH", "line_number": 29, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 29, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 32, "usage_type": "call"}]}
+{"seq_id": "521514788", "text": "# gabbar\n\n\nfrom sklearn import svm\nfrom sklearn.externals import joblib\nimport os\n\nhas_legs = False\n\ndef changeset_to_data(changeset):\n \"\"\"Convert changeset dictionary into an array with required features.\n\n Parameters\n ----------\n changeset: dict\n\n Returns\n -------\n data: tuple\n Tuple of data items\n \"\"\"\n return [\n changeset['create'],\n changeset['modify'],\n changeset['delete']\n ]\n\ndef load_model():\n directory = os.path.dirname(os.path.realpath(__file__))\n filename = 'models/gabbar.pkl'\n model = os.path.join(directory, filename)\n return joblib.load(model)\n\ndef predict(model, data):\n \"\"\"Returns model prediction for data.\n\n Parameters\n ----------\n model: object\n Trained machine learning classifier\n data: tuple\n Tuple of data items\n Returns\n -------\n prediction: int\n -1 for outlier, +1 for inlier\n \"\"\"\n prediction = model.predict(data)\n return prediction[0]\n", "sub_path": "gabbar/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 990, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.path.dirname", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path", "line_number": 31, "usage_type": "attribute"}, {"api_name": "sklearn.externals.joblib.load", "line_number": 32, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib", "line_number": 32, "usage_type": "name"}]}
+{"seq_id": "618634136", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Dec 23 17:18:29 2020\n\n.. note:: Need to keep these groups together, if you split them into files you\n get a circular import.\n\n:copyright: \n Jared Peacock (jpeacock@usgs.gov)\n\n:license: MIT\n\n\"\"\"\n\n# =============================================================================\n# Imports\n# =============================================================================\nimport inspect\nimport weakref\n\nimport h5py\nimport numpy as np\nimport pandas as pd\nimport xarray as xr\n\nfrom mt_metadata import timeseries as metadata\nfrom mt_metadata.utils.mttime import MTime\nfrom mt_metadata.base import Base\nfrom mt_metadata.timeseries.filters import ChannelResponseFilter\n\nfrom mth5 import CHUNK_SIZE, CHANNEL_DTYPE\nfrom mth5.groups.base import BaseGroup\nfrom mth5.groups import FiltersGroup, TransferFunctionGroup\nfrom mth5.utils.exceptions import MTH5Error\nfrom mth5.helpers import (\n to_numpy_type,\n from_numpy_type,\n inherit_doc_string,\n validate_name,\n)\n\nfrom mth5.timeseries import ChannelTS, RunTS\nfrom mth5.timeseries.channel_ts import make_dt_coordinates\nfrom mth5.utils.mth5_logger import setup_logger\n\nmeta_classes = dict(inspect.getmembers(metadata, inspect.isclass))\n# =============================================================================\n# Standards Group\n# =============================================================================\n\n\nclass MasterStationGroup(BaseGroup):\n \"\"\"\n Utility class to holds information about the stations within a survey and\n accompanying metadata. This class is next level down from Survey for\n stations ``/Survey/Stations``. This class provides methods to add and\n get stations. A summary table of all existing stations is also provided\n as a convenience look up table to make searching easier.\n\n To access MasterStationGroup from an open MTH5 file:\n\n >>> from mth5 import mth5\n >>> mth5_obj = mth5.MTH5()\n >>> mth5_obj.open_mth5(r\"/test.mth5\", mode='a')\n >>> stations = mth5_obj.stations_group\n\n To check what stations exist\n\n >>> stations.groups_list\n ['summary', 'MT001', 'MT002', 'MT003']\n\n To access the hdf5 group directly use `SurveyGroup.hdf5_group`.\n\n >>> stations.hdf5_group.ref\n \n\n .. note:: All attributes should be input into the metadata object, that\n way all input will be validated against the metadata standards.\n If you change attributes in metadata object, you should run the\n `SurveyGroup.write_metadata()` method. This is a temporary\n solution, working on an automatic updater if metadata is changed.\n\n >>> stations.metadata.existing_attribute = 'update_existing_attribute'\n >>> stations.write_metadata()\n\n If you want to add a new attribute this should be done using the\n `metadata.add_base_attribute` method.\n\n >>> stations.metadata.add_base_attribute('new_attribute',\n >>> ... 'new_attribute_value',\n >>> ... {'type':str,\n >>> ... 'required':True,\n >>> ... 'style':'free form',\n >>> ... 'description': 'new attribute desc.',\n >>> ... 'units':None,\n >>> ... 'options':[],\n >>> ... 'alias':[],\n >>> ... 'example':'new attribute\n\n To add a station:\n\n >>> new_station = stations.add_station('new_station')\n >>> stations\n /Survey/Stations:\n ====================\n --> Dataset: summary\n ......................\n |- Group: new_station\n ---------------------\n --> Dataset: summary\n ......................\n\n Add a station with metadata:\n\n >>> from mth5.metadata import Station\n >>> station_metadata = Station()\n >>> station_metadata.id = 'MT004'\n >>> station_metadata.time_period.start = '2020-01-01T12:30:00'\n >>> station_metadata.location.latitude = 40.000\n >>> station_metadata.location.longitude = -120.000\n >>> new_station = stations.add_station('Test_01', station_metadata)\n >>> # to look at the metadata\n >>> new_station.metadata\n {\n \"station\": {\n \"acquired_by.author\": null,\n \"acquired_by.comments\": null,\n \"id\": \"MT004\",\n ...\n }\n }\n\n\n .. seealso:: `mth5.metadata` for details on how to add metadata from\n various files and python objects.\n\n To remove a station:\n\n >>> stations.remove_station('new_station')\n >>> stations\n /Survey/Stations:\n ====================\n --> Dataset: summary\n ......................\n\n .. note:: Deleting a station is not as simple as del(station). In HDF5\n this does not free up memory, it simply removes the reference\n to that station. The common way to get around this is to\n copy what you want into a new file, or overwrite the station.\n\n To get a station:\n\n >>> existing_station = stations.get_station('existing_station_name')\n >>> existing_station\n /Survey/Stations/existing_station_name:\n =======================================\n --> Dataset: summary\n ......................\n |- Group: run_01\n ----------------\n --> Dataset: summary\n ......................\n --> Dataset: Ex\n ......................\n --> Dataset: Ey\n ......................\n --> Dataset: Hx\n ......................\n --> Dataset: Hy\n ......................\n --> Dataset: Hz\n ......................\n\n A summary table is provided to make searching easier. The table\n summarized all stations within a survey. To see what names are in the\n summary table:\n\n >>> stations.summary_table.dtype.descr\n [('id', ('|S5', {'h5py_encoding': 'ascii'})),\n ('start', ('|S32', {'h5py_encoding': 'ascii'})),\n ('end', ('|S32', {'h5py_encoding': 'ascii'})),\n ('components', ('|S100', {'h5py_encoding': 'ascii'})),\n ('measurement_type', ('|S12', {'h5py_encoding': 'ascii'})),\n ('sample_rate', '>> stations.summary_table\n index | id | start | end\n | components | measurement_type | sample_rate\n -------------------------------------------------------------------------\n --------------------------------------------------\n 0 | Test_01 | 1980-01-01T00:00:00+00:00 | 1980-01-01T00:00:00+00:00\n | Ex,Ey,Hx,Hy,Hz | BBMT | 100\n\n \"\"\"\n\n def __init__(self, group, **kwargs):\n\n super().__init__(group, **kwargs)\n\n @property\n def channel_summary(self):\n \"\"\"\n Summary of all channels in the file.\n \"\"\"\n ch_list = []\n for station in self.groups_list:\n s_group = StationGroup(self.hdf5_group[station])\n for run in s_group.groups_list:\n r_group = RunGroup(s_group.hdf5_group[run])\n for ch in r_group.groups_list:\n ds_type = r_group.hdf5_group[ch].attrs[\"mth5_type\"]\n if ds_type.lower() in [\"electric\"]:\n ch_dataset = ElectricDataset(r_group.hdf5_group[ch])\n elif ds_type.lower() in [\"magnetic\"]:\n ch_dataset = MagneticDataset(r_group.hdf5_group[ch])\n elif ds_type.lower() in [\"auxiliary\"]:\n ch_dataset = AuxiliaryDataset(r_group.hdf5_group[ch])\n ch_list.append(ch_dataset.channel_entry)\n ch_list = np.array(ch_list)\n return pd.DataFrame(ch_list.flatten())\n\n @property\n def station_summary(self):\n \"\"\"\n Summary of stations in the file\n\n :return: DESCRIPTION\n :rtype: TYPE\n\n \"\"\"\n st_list = []\n for key, group in self.hdf5_group.items():\n entry = {\n \"station\": key,\n \"start\": group.attrs[\"time_period.start\"],\n \"end\": group.attrs[\"time_period.end\"],\n \"latitude\": group.attrs[\"location.latitude\"],\n \"longitude\": group.attrs[\"location.longitude\"],\n }\n st_list.append(entry)\n\n df = pd.DataFrame(st_list)\n df.start = pd.to_datetime(df.start)\n df.end = pd.to_datetime(df.end)\n\n return df\n\n def add_station(self, station_name, station_metadata=None):\n \"\"\"\n Add a station with metadata if given with the path:\n ``/Survey/Stations/station_name``\n\n If the station already exists, will return that station and nothing\n is added.\n\n :param station_name: Name of the station, should be the same as\n metadata.id\n :type station_name: string\n :param station_metadata: Station metadata container, defaults to None\n :type station_metadata: :class:`mth5.metadata.Station`, optional\n :return: A convenience class for the added station\n :rtype: :class:`mth5_groups.StationGroup`\n\n :Example: ::\n\n >>> from mth5 import mth5\n >>> mth5_obj = mth5.MTH5()\n >>> mth5_obj.open_mth5(r\"/test.mth5\", mode='a')\n >>> # one option\n >>> stations = mth5_obj.stations_group\n >>> new_station = stations.add_station('MT001')\n >>> # another option\n >>> new_staiton = mth5_obj.stations_group.add_station('MT001')\n\n .. todo:: allow dictionaries, json string, xml elements as metadata\n input.\n\n \"\"\"\n if station_name is None:\n raise Exception(\n \"station name is None, do not know what to name it\"\n )\n station_name = validate_name(station_name)\n try:\n station_group = self.hdf5_group.create_group(station_name)\n self.logger.debug(\"Created group %s\", station_group.name)\n\n if station_metadata is None:\n station_metadata = metadata.Station(id=station_name)\n else:\n if validate_name(station_metadata.id) != station_name:\n msg = (\n f\"Station group name {station_name} must be same as \"\n + f\"station id {station_metadata.id}\"\n )\n self.logger.error(msg)\n raise MTH5Error(msg)\n station_obj = StationGroup(\n station_group,\n station_metadata=station_metadata,\n **self.dataset_options,\n )\n station_obj.initialize_group()\n\n # be sure to add a table entry\n # self.summary_table.add_row(station_obj.table_entry)\n except ValueError:\n msg = \"Station %s already exists, returning existing group.\"\n self.logger.info(msg, station_name)\n station_obj = self.get_station(station_name)\n return station_obj\n\n def get_station(self, station_name):\n \"\"\"\n Get a station with the same name as station_name\n\n :param station_name: existing station name\n :type station_name: string\n :return: convenience station class\n :rtype: :class:`mth5.mth5_groups.StationGroup`\n :raises MTH5Error: if the station name is not found.\n\n :Example:\n\n >>> from mth5 import mth5\n >>> mth5_obj = mth5.MTH5()\n >>> mth5_obj.open_mth5(r\"/test.mth5\", mode='a')\n >>> # one option\n >>> stations = mth5_obj.stations_group\n >>> existing_station = stations.get_station('MT001')\n >>> # another option\n >>> existing_staiton = mth5_obj.stations_group.get_station('MT001')\n MTH5Error: MT001 does not exist, check station_list for existing names\n\n \"\"\"\n station_name = validate_name(station_name)\n try:\n return StationGroup(\n self.hdf5_group[station_name], **self.dataset_options\n )\n except KeyError:\n msg = (\n f\"{station_name} does not exist, \"\n + \"check station_list for existing names\"\n )\n self.logger.debug(\"Error\" + msg)\n raise MTH5Error(msg)\n\n def remove_station(self, station_name):\n \"\"\"\n Remove a station from the file.\n\n .. note:: Deleting a station is not as simple as del(station). In HDF5\n this does not free up memory, it simply removes the reference\n to that station. The common way to get around this is to\n copy what you want into a new file, or overwrite the station.\n\n :param station_name: existing station name\n :type station_name: string\n\n :Example: ::\n\n >>> from mth5 import mth5\n >>> mth5_obj = mth5.MTH5()\n >>> mth5_obj.open_mth5(r\"/test.mth5\", mode='a')\n >>> # one option\n >>> stations = mth5_obj.stations_group\n >>> stations.remove_station('MT001')\n >>> # another option\n >>> mth5_obj.stations_group.remove_station('MT001')\n\n \"\"\"\n\n station_name = validate_name(station_name)\n try:\n del self.hdf5_group[station_name]\n self.logger.info(\n \"Deleting a station does not reduce the HDF5\"\n + \"file size it simply remove the reference. If \"\n + \"file size reduction is your goal, simply copy\"\n + \" what you want into another file.\"\n )\n except KeyError:\n msg = (\n f\"{station_name} does not exist, \"\n + \"check station_list for existing names\"\n )\n self.logger.debug(\"Error\" + msg)\n raise MTH5Error(msg)\n\n\n# =============================================================================\n# Station Group\n# =============================================================================\nclass StationGroup(BaseGroup):\n \"\"\"\n StationGroup is a utility class to hold information about a single station\n and accompanying metadata. This class is the next level down from\n Stations --> ``/Survey/Stations/station_name``.\n\n This class provides methods to add and get runs. A summary table of all\n existing runs in the station is also provided as a convenience look up\n table to make searching easier.\n\n :param group: HDF5 group for a station, should have a path\n ``/Survey/Stations/station_name``\n :type group: :class:`h5py.Group`\n :param station_metadata: metadata container, defaults to None\n :type station_metadata: :class:`mth5.metadata.Station`, optional\n\n :Usage:\n\n :Access StationGroup from an open MTH5 file:\n\n >>> from mth5 import mth5\n >>> mth5_obj = mth5.MTH5()\n >>> mth5_obj.open_mth5(r\"/test.mth5\", mode='a')\n >>> station = mth5_obj.stations_group.get_station('MT001')\n\n :Check what runs exist:\n\n >>> station.groups_list\n ['MT001a', 'MT001b', 'MT001c', 'MT001d']\n\n To access the hdf5 group directly use `StationGroup.hdf5_group`.\n\n >>> station.hdf5_group.ref\n \n\n .. note:: All attributes should be input into the metadata object, that\n way all input will be validated against the metadata standards.\n If you change attributes in metadata object, you should run the\n `SurveyGroup.write_metadata()` method. This is a temporary\n solution, working on an automatic updater if metadata is changed.\n\n >>> station.metadata.existing_attribute = 'update_existing_attribute'\n >>> station.write_metadata()\n\n If you want to add a new attribute this should be done using the\n `metadata.add_base_attribute` method.\n\n >>> station.metadata.add_base_attribute('new_attribute',\n >>> ... 'new_attribute_value',\n >>> ... {'type':str,\n >>> ... 'required':True,\n >>> ... 'style':'free form',\n >>> ... 'description': 'new attribute desc.',\n >>> ... 'units':None,\n >>> ... 'options':[],\n >>> ... 'alias':[],\n >>> ... 'example':'new attribute\n\n :To add a run:\n\n >>> new_run = stations.add_run('MT001e')\n >>> new_run\n /Survey/Stations/Test_01:\n =========================\n |- Group: MT001e\n -----------------\n --> Dataset: summary\n ......................\n --> Dataset: summary\n ......................\n\n :Add a run with metadata:\n\n >>> from mth5.metadata import Run\n >>> run_metadata = Run()\n >>> run_metadata.time_period.start = '2020-01-01T12:30:00'\n >>> run_metadata.time_period.end = '2020-01-03T16:30:00'\n >>> run_metadata.location.latitude = 40.000\n >>> run_metadata.location.longitude = -120.000\n >>> new_run = runs.add_run('Test_01', run_metadata)\n >>> # to look at the metadata\n >>> new_run.metadata\n {\n \"run\": {\n \"acquired_by.author\": \"new_user\",\n \"acquired_by.comments\": \"First time\",\n \"channels_recorded_auxiliary\": ['T'],\n ...\n }\n }\n\n\n .. seealso:: `mth5.metadata` for details on how to add metadata from\n various files and python objects.\n\n :Remove a run:\n\n >>> station.remove_run('new_run')\n >>> station\n /Survey/Stations/Test_01:\n =========================\n --> Dataset: summary\n ......................\n\n .. note:: Deleting a station is not as simple as del(station). In HDF5\n this does not free up memory, it simply removes the reference\n to that station. The common way to get around this is to\n copy what you want into a new file, or overwrite the station.\n\n :Get a run:\n\n >>> existing_run = stations.get_station('existing_run')\n >>> existing_run\n /Survey/Stations/MT001/MT001a:\n =======================================\n --> Dataset: summary\n ......................\n --> Dataset: Ex\n ......................\n --> Dataset: Ey\n ......................\n --> Dataset: Hx\n ......................\n --> Dataset: Hy\n ......................\n --> Dataset: Hz\n ......................\n\n :summary Table:\n\n A summary table is provided to make searching easier. The table\n summarized all stations within a survey. To see what names are in the\n summary table:\n\n >>> new_run.summary_table.dtype.descr\n [('id', ('|S20', {'h5py_encoding': 'ascii'})),\n ('start', ('|S32', {'h5py_encoding': 'ascii'})),\n ('end', ('|S32', {'h5py_encoding': 'ascii'})),\n ('components', ('|S100', {'h5py_encoding': 'ascii'})),\n ('measurement_type', ('|S12', {'h5py_encoding': 'ascii'})),\n ('sample_rate', '>> station.summary_table\n index | id | start | end | components | measurement_type | sample_rate |\n hdf5_reference\n --------------------------------------------------------------------------\n -------------\n \"\"\"\n\n def __init__(self, group, station_metadata=None, **kwargs):\n super().__init__(group, group_metadata=station_metadata, **kwargs)\n\n self._defaults_summary_keys = [\n \"id\",\n \"start\",\n \"end\",\n \"components\",\n \"measurement_type\",\n \"sample_rate\",\n \"hdf5_reference\",\n \"mth5_type\",\n ]\n\n self._default_subgroup_names = [\n \"Transfer_Functions\",\n ]\n\n def initialize_group(self, **kwargs):\n \"\"\"\n Initialize group by making a summary table and writing metadata\n\n \"\"\"\n for key, value in kwargs.items():\n setattr(self, key, value)\n self.write_metadata()\n\n for group_name in self._default_subgroup_names:\n self.hdf5_group.create_group(f\"{group_name}\")\n m5_grp = getattr(self, f\"{group_name.lower()}_group\")\n m5_grp.initialize_group()\n\n @property\n def master_station_group(self):\n \"\"\"shortcut to master station group\"\"\"\n return MasterStationGroup(self.hdf5_group.parent)\n\n @property\n def transfer_functions_group(self):\n \"\"\"Convinience method for /Station/Transfer_Functions\"\"\"\n return TransferFunctionsGroup(\n self.hdf5_group[\"Transfer_Functions\"], **self.dataset_options\n )\n\n @BaseGroup.metadata.getter\n def metadata(self):\n \"\"\"Overwrite get metadata to include run information in the station\"\"\"\n\n self._metadata.runs = []\n for key in self.groups_list:\n if key.lower() == \"transfer_functions\":\n continue\n try:\n key_group = self.get_run(key)\n self._metadata.runs.append(key_group.metadata)\n except MTH5Error:\n self.logger.warning(f\"Could not find run {key}\")\n return self._metadata\n\n @property\n def name(self):\n return self.metadata.id\n\n @name.setter\n def name(self, name):\n self.metadata.id = name\n\n @property\n def run_summary(self):\n \"\"\"\n Summary of runs in the station\n\n :return: DESCRIPTION\n :rtype: TYPE\n\n \"\"\"\n\n run_list = []\n for key, group in self.hdf5_group.items():\n if group.attrs[\"mth5_type\"].lower() in [\"run\"]:\n comps = \",\".join(\n [\n ii.decode()\n for ii in group.attrs[\n \"channels_recorded_auxiliary\"\n ].tolist()\n + group.attrs[\"channels_recorded_electric\"].tolist()\n + group.attrs[\"channels_recorded_magnetic\"].tolist()\n ]\n )\n run_list.append(\n (\n group.attrs[\"id\"],\n group.attrs[\"time_period.start\"].split(\"+\")[0],\n group.attrs[\"time_period.end\"].split(\"+\")[0],\n comps,\n group.attrs[\"data_type\"],\n group.attrs[\"sample_rate\"],\n self.hdf5_group.ref,\n )\n )\n run_summary = np.array(\n run_list,\n dtype=np.dtype(\n [\n (\"id\", \"U20\"),\n (\"start\", \"datetime64[ns]\"),\n (\"end\", \"datetime64[ns]\"),\n (\"components\", \"U100\"),\n (\"measurement_type\", \"U12\"),\n (\"sample_rate\", float),\n (\"hdf5_reference\", h5py.ref_dtype),\n ]\n ),\n )\n\n return pd.DataFrame(run_summary)\n\n def make_run_name(self, alphabet=False):\n \"\"\"\n Make a run name that will be the next alphabet letter extracted from\n the run list. Expects that all runs are labled as id{a-z}.\n\n :return: metadata.id + next letter\n :rtype: string\n\n >>> station.metadata.id = 'MT001'\n >>> station.make_run_name()\n 'MT001a'\n\n \"\"\"\n\n run_list = sorted(\n [group[-1:] for group in self.groups_list if self.name in group]\n )\n\n next_letter = None\n if len(run_list) == 0:\n if alphabet:\n next_letter = \"a\"\n else:\n next_letter = \"001\"\n else:\n try:\n next_letter = chr(ord(run_list[-1]) + 1)\n except TypeError:\n try:\n next_letter = f\"{int(run_list[-1]) + 1}\"\n except ValueError:\n self.logger.info(\"Could not create a new run name\")\n return next_letter\n\n def locate_run(self, sample_rate, start):\n \"\"\"\n Locate a run based on sample rate and start time from the summary table\n\n :param sample_rate: sample rate in samples/seconds\n :type sample_rate: float\n :param start: start time\n :type start: string or :class:`mth5.utils.mttime.MTime`\n :return: appropriate run name, None if not found\n :rtype: string or None\n\n \"\"\"\n\n if not isinstance(start, MTime):\n start = MTime(start)\n if self.run_summary.size < 1:\n return None\n sr_find = self.run_summary[\n (self.run_summary.sample_rate == sample_rate)\n & (self.run_summary.start == start)\n ]\n if sr_find.size < 1:\n return None\n return sr_find\n\n def add_run(self, run_name, run_metadata=None):\n \"\"\"\n Add a run to a station.\n\n :param run_name: run name, should be id{a-z}\n :type run_name: string\n :param metadata: metadata container, defaults to None\n :type metadata: :class:`mth5.metadata.Station`, optional\n\n need to be able to fill an entry in the summary table.\n\n .. todo:: auto fill run name if none is given.\n\n .. todo:: add ability to add a run with data.\n\n \"\"\"\n\n run_name = validate_name(run_name)\n try:\n run_group = self.hdf5_group.create_group(run_name)\n if run_metadata is None:\n run_metadata = metadata.Run(id=run_name)\n elif validate_name(run_metadata.id) != run_name:\n msg = \"Run name %s must be the same as run_metadata.id %s\"\n self.logger.error(msg, run_name, run_metadata.id)\n raise MTH5Error(msg % (run_name, run_metadata.id))\n run_obj = RunGroup(\n run_group, run_metadata=run_metadata, **self.dataset_options\n )\n run_obj.initialize_group()\n except ValueError:\n msg = \"run %s already exists, returning existing group.\"\n self.logger.info(msg, run_name)\n run_obj = self.get_run(run_name)\n return run_obj\n\n def get_run(self, run_name):\n \"\"\"\n get a run from run name\n\n :param run_name: existing run name\n :type run_name: string\n :return: Run object\n :rtype: :class:`mth5.mth5_groups.RunGroup`\n\n >>> existing_run = station.get_run('MT001')\n\n \"\"\"\n\n run_name = validate_name(run_name)\n try:\n return RunGroup(self.hdf5_group[run_name], **self.dataset_options)\n except KeyError:\n msg = (\n f\"{run_name} does not exist, \"\n + \"check groups_list for existing names\"\n )\n self.logger.debug(\"Error\" + msg)\n raise MTH5Error(msg)\n\n def remove_run(self, run_name):\n \"\"\"\n Remove a run from the station.\n\n .. note:: Deleting a station is not as simple as del(station). In HDF5\n this does not free up memory, it simply removes the reference\n to that station. The common way to get around this is to\n copy what you want into a new file, or overwrite the station.\n\n :param station_name: existing station name\n :type station_name: string\n\n :Example: ::\n\n >>> from mth5 import mth5\n >>> mth5_obj = mth5.MTH5()\n >>> mth5_obj.open_mth5(r\"/test.mth5\", mode='a')\n >>> # one option\n >>> stations = mth5_obj.stations_group\n >>> stations.remove_station('MT001')\n >>> # another option\n >>> mth5_obj.stations_group.remove_station('MT001')\n\n \"\"\"\n\n run_name = validate_name(run_name)\n try:\n del self.hdf5_group[run_name]\n self.logger.info(\n \"Deleting a run does not reduce the HDF5\"\n + \"file size it simply remove the reference. If \"\n + \"file size reduction is your goal, simply copy\"\n + \" what you want into another file.\"\n )\n except KeyError:\n msg = (\n f\"{run_name} does not exist, \"\n + \"check station_list for existing names\"\n )\n self.logger.debug(\"Error\" + msg)\n raise MTH5Error(msg)\n\n def update_station_metadata(self):\n \"\"\"\n Check metadata from the runs and make sure it matches the station metadata\n\n :return: DESCRIPTION\n :rtype: TYPE\n\n \"\"\"\n\n run_summary = self.run_summary.copy()\n self._metadata.time_period.start = run_summary.start.min().isoformat()\n self._metadata.time_period.end = run_summary.end.max().isoformat()\n self._metadata.channels_recorded = list(\n set(\",\".join(run_summary.components.to_list()).split(\",\"))\n )\n\n self.write_metadata()\n\n\n# =============================================================================\n# Transfer Functions Group\n# =============================================================================\nclass TransferFunctionsGroup(BaseGroup):\n \"\"\"\n Object to hold transfer functions\n\n The is the high level group, all transfer functions for the station are\n held here and each one will have its own TransferFunctionGroup.\n\n This has add, get, remove_transfer_function.\n \"\"\"\n\n def __init__(self, group, **kwargs):\n super().__init__(group, **kwargs)\n\n def tf_summary(self, as_dataframe=True):\n \"\"\"\n Summary of all transfer functions in this group\n\n :return: DESCRIPTION\n :rtype: TYPE\n\n \"\"\"\n\n tf_list = []\n for tf_id in self.groups_list:\n tf_group = self.get_transfer_function(tf_id)\n tf_entry = tf_group.tf_entry\n\n tf_entry[\"station_hdf5_reference\"][:] = self.hdf5_group.parent.ref\n tf_entry[\"station\"][:] = self.hdf5_group.parent.attrs[\"id\"]\n tf_entry[\"latitude\"][:] = self.hdf5_group.parent.attrs[\n \"location.latitude\"\n ]\n tf_entry[\"longitude\"][:] = self.hdf5_group.parent.attrs[\n \"location.longitude\"\n ]\n tf_entry[\"elevation\"][:] = self.hdf5_group.parent.attrs[\n \"location.elevation\"\n ]\n\n tf_list.append(tf_entry)\n tf_list = np.array(tf_list)\n\n if as_dataframe:\n return pd.DataFrame(tf_list.flatten())\n return tf_list\n\n def add_transfer_function(self, name, tf_object=None):\n \"\"\"\n Add a transfer function to the group\n\n :param name: name of the transfer function\n :type name: string\n :param tf_object: Transfer Function object\n :type tf_object: :class:`mt_metadata.transfer_function.core.TF`\n :return: DESCRIPTION\n :rtype: TYPE\n\n >>> from mth5.mth5 import MTH5\n >>> m = MTH5()\n >>> m.open_mth5(\"example.h5\", \"a\")\n >>> station_group = m.get_station(\"mt01\", survey=\"test\")\n >>> tf_group = station_group.transfer_functions_group\n >>> tf_group.add_transfer_function(\"mt01_4096\", tf_object)\n\n\n \"\"\"\n name = validate_name(name)\n\n tf_group = TransferFunctionGroup(\n self.hdf5_group.create_group(name), **self.dataset_options\n )\n\n if tf_object is not None:\n tf_group.from_tf_object(tf_object)\n return tf_group\n\n def get_transfer_function(self, tf_id):\n \"\"\"\n Get transfer function from id\n\n :param tf_id: name of transfer function\n :type tf_id: string\n :return: Transfer function group\n :rtype: :class:`mth5.groups.TransferFunctionGroup`\n\n >>> from mth5.mth5 import MTH5\n >>> m = MTH5()\n >>> m.open_mth5(\"example.h5\", \"a\")\n >>> station_group = m.get_station(\"mt01\", survey=\"test\")\n >>> tf_group = station_group.transfer_functions_group.get_transfer_function(\"mt01_4096\")\n\n\n \"\"\"\n\n tf_id = validate_name(tf_id)\n try:\n return TransferFunctionGroup(\n self.hdf5_group[tf_id], **self.dataset_options\n )\n except KeyError:\n msg = (\n f\"{tf_id} does not exist, \"\n + \"check station_list for existing names\"\n )\n self.logger.debug(\"Error\" + msg)\n raise MTH5Error(msg)\n\n def remove_transfer_function(self, tf_id):\n \"\"\"\n Remove a transfer function from the group\n\n :param tf_id: DESCRIPTION\n :type tf_id: TYPE\n :return: DESCRIPTION\n :rtype: TYPE\n\n >>> from mth5.mth5 import MTH5\n >>> m = MTH5()\n >>> m.open_mth5(\"example.h5\", \"a\")\n >>> station_group = m.get_station(\"mt01\", survey=\"test\")\n >>> tf_group = station_group.transfer_functions_group\n >>> tf_group.remove_transfer_function(\"mt01_4096\")\n\n \"\"\"\n\n tf_id = validate_name(tf_id)\n try:\n del self.hdf5_group[tf_id]\n self.logger.info(\n \"Deleting a station does not reduce the HDF5\"\n + \"file size it simply remove the reference. If \"\n + \"file size reduction is your goal, simply copy\"\n + \" what you want into another file.\"\n )\n except KeyError:\n msg = (\n f\"{tf_id} does not exist, \"\n + \"check station_list for existing names\"\n )\n self.logger.debug(\"Error\" + msg)\n raise MTH5Error(msg)\n\n def get_tf_object(self, tf_id):\n \"\"\"\n This is the function you want to use to get a proper\n :class:`mt_metadata.transfer_functions.core.TF` object with all the\n appropriate metadata.\n\n\n :param tf_id: name of the transfer function to get\n :type tf_id: string\n :return: Full transfer function with appropriate metadata\n :rtype: :class:`mt_metadata.transfer_functions.core.TF`\n\n >>> from mth5.mth5 import MTH5\n >>> m = MTH5()\n >>> m.open_mth5(\"example.h5\", \"a\")\n >>> station_group = m.get_station(\"mt01\", survey=\"test\")\n >>> tf_group = station_group.transfer_functions_group\n >>> tf_object = tf_group.get_tf_object(\"mt01_4096\")\n\n\n \"\"\"\n\n tf_group = self.get_transfer_function(tf_id)\n\n return tf_group.to_tf_object()\n\n\n# =============================================================================\n# Run Group\n# =============================================================================\nclass RunGroup(BaseGroup):\n \"\"\"\n RunGroup is a utility class to hold information about a single run\n and accompanying metadata. This class is the next level down from\n Stations --> ``/Survey/Stations/station/station{a-z}``.\n\n This class provides methods to add and get channels. A summary table of\n all existing channels in the run is also provided as a convenience look up\n table to make searching easier.\n\n :param group: HDF5 group for a station, should have a path\n ``/Survey/Stations/station_name/run_name``\n :type group: :class:`h5py.Group`\n :param station_metadata: metadata container, defaults to None\n :type station_metadata: :class:`mth5.metadata.Station`, optional\n\n :Access RunGroup from an open MTH5 file:\n\n >>> from mth5 import mth5\n >>> mth5_obj = mth5.MTH5()\n >>> mth5_obj.open_mth5(r\"/test.mth5\", mode='a')\n >>> run = mth5_obj.stations_group.get_station('MT001').get_run('MT001a')\n\n :Check what channels exist:\n\n >>> station.groups_list\n ['Ex', 'Ey', 'Hx', 'Hy']\n\n To access the hdf5 group directly use `RunGroup.hdf5_group`\n\n >>> station.hdf5_group.ref\n \n\n .. note:: All attributes should be input into the metadata object, that\n way all input will be validated against the metadata standards.\n If you change attributes in metadata object, you should run the\n `SurveyGroup.write_metadata()` method. This is a temporary\n solution, working on an automatic updater if metadata is changed.\n\n >>> run.metadata.existing_attribute = 'update_existing_attribute'\n >>> run.write_metadata()\n\n If you want to add a new attribute this should be done using the\n `metadata.add_base_attribute` method.\n\n >>> station.metadata.add_base_attribute('new_attribute',\n >>> ... 'new_attribute_value',\n >>> ... {'type':str,\n >>> ... 'required':True,\n >>> ... 'style':'free form',\n >>> ... 'description': 'new attribute desc.',\n >>> ... 'units':None,\n >>> ... 'options':[],\n >>> ... 'alias':[],\n >>> ... 'example':'new attribute\n\n :Add a channel:\n\n >>> new_channel = run.add_channel('Ex', 'electric',\n >>> ... data=numpy.random.rand(4096))\n >>> new_run\n /Survey/Stations/MT001/MT001a:\n =======================================\n --> Dataset: summary\n ......................\n --> Dataset: Ex\n ......................\n --> Dataset: Ey\n ......................\n --> Dataset: Hx\n ......................\n --> Dataset: Hy\n ......................\n\n\n :Add a channel with metadata:\n\n >>> from mth5.metadata import Electric\n >>> ex_metadata = Electric()\n >>> ex_metadata.time_period.start = '2020-01-01T12:30:00'\n >>> ex_metadata.time_period.end = '2020-01-03T16:30:00'\n >>> new_ex = run.add_channel('Ex', 'electric',\n >>> ... channel_metadata=ex_metadata)\n >>> # to look at the metadata\n >>> new_ex.metadata\n {\n \"electric\": {\n \"ac.end\": 1.2,\n \"ac.start\": 2.3,\n ...\n }\n }\n\n\n .. seealso:: `mth5.metadata` for details on how to add metadata from\n various files and python objects.\n\n :Remove a channel:\n\n >>> run.remove_channel('Ex')\n >>> station\n /Survey/Stations/MT001/MT001a:\n =======================================\n --> Dataset: summary\n ......................\n --> Dataset: Ey\n ......................\n --> Dataset: Hx\n ......................\n --> Dataset: Hy\n ......................\n\n .. note:: Deleting a station is not as simple as del(station). In HDF5\n this does not free up memory, it simply removes the reference\n to that station. The common way to get around this is to\n copy what you want into a new file, or overwrite the station.\n\n :Get a channel:\n\n >>> existing_ex = stations.get_channel('Ex')\n >>> existing_ex\n Channel Electric:\n -------------------\n data type: Ex\n data type: electric\n data format: float32\n data shape: (4096,)\n start: 1980-01-01T00:00:00+00:00\n end: 1980-01-01T00:32:+08:00\n sample rate: 8\n\n\n :summary Table:\n\n A summary table is provided to make searching easier. The table\n summarized all stations within a survey. To see what names are in the\n summary table:\n\n >>> run.summary_table.dtype.descr\n [('component', ('|S5', {'h5py_encoding': 'ascii'})),\n ('start', ('|S32', {'h5py_encoding': 'ascii'})),\n ('end', ('|S32', {'h5py_encoding': 'ascii'})),\n ('n_samples', '>> new_run.summary_table\n index | component | start | end | n_samples | measurement_type | units |\n hdf5_reference\n --------------------------------------------------------------------------\n -------------\n \"\"\"\n\n def __init__(self, group, run_metadata=None, **kwargs):\n super().__init__(group, group_metadata=run_metadata, **kwargs)\n\n # summary of channels in run\n self._defaults_summary_attrs = {\n \"name\": \"summary\",\n \"max_shape\": (20,),\n \"dtype\": np.dtype(\n [\n (\"component\", \"S20\"),\n (\"start\", \"S32\"),\n (\"end\", \"S32\"),\n (\"n_samples\", int),\n (\"measurement_type\", \"S12\"),\n (\"units\", \"S25\"),\n (\"hdf5_reference\", h5py.ref_dtype),\n ]\n ),\n }\n\n @property\n def station_group(self):\n \"\"\"shortcut to station group\"\"\"\n return StationGroup(self.hdf5_group.parent)\n\n @property\n def station_metadata(self):\n \"\"\"station metadata\"\"\"\n\n meta_dict = dict(self.hdf5_group.parent.attrs)\n for key, value in meta_dict.items():\n meta_dict[key] = from_numpy_type(value)\n station_metadata = metadata.Station()\n station_metadata.from_dict({\"station\": meta_dict})\n return station_metadata\n\n @property\n def master_station_group(self):\n \"\"\"shortcut to master station group\"\"\"\n return MasterStationGroup(self.hdf5_group.parent.parent)\n\n @property\n def survey_metadata(self):\n \"\"\"survey metadata\"\"\"\n\n meta_dict = dict(self.hdf5_group.parent.parent.parent.attrs)\n for key, value in meta_dict.items():\n meta_dict[key] = from_numpy_type(value)\n survey_metadata = metadata.Survey()\n survey_metadata.from_dict({\"survey\": meta_dict})\n return survey_metadata\n\n @BaseGroup.metadata.getter\n def metadata(self):\n \"\"\"Overwrite get metadata to include channel information in the runs\"\"\"\n\n self._metadata.channels = []\n for ch in self.groups_list:\n ch_group = self.get_channel(ch)\n self._metadata.channels.append(ch_group.metadata)\n self._metadata.hdf5_reference = self.hdf5_group.ref\n return self._metadata\n\n @property\n def channel_summary(self):\n \"\"\"\n\n summary of channels in run\n :return: DESCRIPTION\n :rtype: TYPE\n\n \"\"\"\n\n ch_list = []\n for key, group in self.hdf5_group.items():\n try:\n ch_type = group.attrs[\"type\"]\n if ch_type in [\"electric\", \"magnetic\", \"auxiliary\"]:\n ch_list.append(\n (\n group.attrs[\"component\"],\n group.attrs[\"time_period.start\"].split(\"+\")[0],\n group.attrs[\"time_period.end\"].split(\"+\")[0],\n group.size,\n group.attrs[\"sample_rate\"],\n group.attrs[\"type\"],\n group.attrs[\"units\"],\n group.ref,\n )\n )\n except KeyError:\n pass\n ch_summary = np.array(\n ch_list,\n dtype=np.dtype(\n [\n (\"component\", \"U20\"),\n (\"start\", \"datetime64[ns]\"),\n (\"end\", \"datetime64[ns]\"),\n (\"n_samples\", np.int64),\n (\"sample_rate\", np.int64),\n (\"measurement_type\", \"U12\"),\n (\"units\", \"U25\"),\n (\"hdf5_reference\", h5py.ref_dtype),\n ]\n ),\n )\n\n return pd.DataFrame(ch_summary)\n\n def write_metadata(self):\n \"\"\"\n Overwrite Base.write_metadata to include updating table entry\n Write HDF5 metadata from metadata object.\n\n \"\"\"\n\n for key, value in self.metadata.to_dict(single=True).items():\n value = to_numpy_type(value)\n self.hdf5_group.attrs.create(key, value)\n\n def add_channel(\n self,\n channel_name,\n channel_type,\n data,\n channel_dtype=\"int32\",\n max_shape=(None,),\n chunks=True,\n channel_metadata=None,\n **kwargs,\n ):\n \"\"\"\n add a channel to the run\n\n :param channel_name: name of the channel\n :type channel_name: string\n :param channel_type: [ electric | magnetic | auxiliary ]\n :type channel_type: string\n :raises MTH5Error: If channel type is not correct\n\n :param channel_metadata: metadata container, defaults to None\n :type channel_metadata: [ :class:`mth5.metadata.Electric` |\n :class:`mth5.metadata.Magnetic` |\n :class:`mth5.metadata.Auxiliary` ], optional\n :return: Channel container\n :rtype: [ :class:`mth5.mth5_groups.ElectricDatset` |\n :class:`mth5.mth5_groups.MagneticDatset` |\n :class:`mth5.mth5_groups.AuxiliaryDatset` ]\n\n >>> new_channel = run.add_channel('Ex', 'electric', None)\n >>> new_channel\n Channel Electric:\n -------------------\n component: None\n data type: electric\n data format: float32\n data shape: (1,)\n start: 1980-01-01T00:00:00+00:00\n end: 1980-01-01T00:00:00+00:00\n sample rate: None\n\n\n \"\"\"\n channel_name = validate_name(channel_name.lower())\n estimate_size = (1,)\n for key, value in kwargs.items():\n setattr(self, key, value)\n if data is not None:\n if data.size < 1024:\n chunks = None\n try:\n if data is not None:\n channel_group = self.hdf5_group.create_dataset(\n channel_name,\n data=data,\n dtype=data.dtype,\n chunks=chunks,\n maxshape=max_shape,\n **self.dataset_options,\n )\n # initialize an resizable data array\n # need to set the chunk size to something useful, if the chunk\n # size is 1 this causes performance issues and bloating of the\n # hdf5 file. Set to 8196 for now. Should add a parameter for\n # this\n else:\n if channel_metadata is not None:\n # can estimate a size, this will help with allocating\n # and set the chunk size to a realistic value\n if (\n channel_metadata.time_period.start\n != channel_metadata.time_period.end\n ):\n if channel_metadata.sample_rate > 0:\n estimate_size = (\n int(\n (\n channel_metadata.time_period._end_dt\n - channel_metadata.time_period._start_dt\n )\n * channel_metadata.sample_rate\n ),\n )\n else:\n estimate_size = (1,)\n chunks = CHUNK_SIZE\n else:\n estimate_size = (1,)\n chunks = CHUNK_SIZE\n\n if estimate_size[0] > 2**31:\n estimate_size = (1,)\n self.logger.warning(\n \"Estimated size is too large. Check start and end \"\n \"times, initializing with size (1,)\"\n )\n\n channel_group = self.hdf5_group.create_dataset(\n channel_name,\n shape=estimate_size,\n maxshape=max_shape,\n dtype=channel_dtype,\n chunks=chunks,\n **self.dataset_options,\n )\n\n if channel_metadata and channel_metadata.component is None:\n channel_metadata.component = channel_name\n if channel_type.lower() in [\"magnetic\"]:\n channel_obj = MagneticDataset(\n channel_group, dataset_metadata=channel_metadata\n )\n elif channel_type.lower() in [\"electric\"]:\n channel_obj = ElectricDataset(\n channel_group, dataset_metadata=channel_metadata\n )\n elif channel_type.lower() in [\"auxiliary\"]:\n channel_obj = AuxiliaryDataset(\n channel_group, dataset_metadata=channel_metadata\n )\n else:\n msg = (\n \"`channel_type` must be in [ electric | magnetic | \"\n + \"auxiliary ]. Input was {0}\".format(channel_type)\n )\n self.logger.error(msg)\n raise MTH5Error(msg)\n except (OSError, RuntimeError, ValueError):\n msg = f\"channel {channel_name} already exists, returning existing group.\"\n self.logger.debug(msg)\n channel_obj = self.get_channel(channel_name)\n\n if data is not None:\n self.logger.debug(\n \"Replacing data with new shape %s\", data.shape\n )\n channel_obj.replace_dataset(data)\n\n self.logger.debug(\"Updating metadata\")\n channel_obj.metadata.update(channel_metadata)\n channel_obj.write_metadata()\n self.logger.debug(\"Done with %s\", channel_name)\n # need to make sure the channel name is passed.\n if channel_obj.metadata.component is None:\n channel_obj.metadata.component = channel_name\n channel_obj.write_metadata()\n return channel_obj\n\n def get_channel(self, channel_name):\n \"\"\"\n\n Get a channel from an existing name. Returns the appropriate\n container.\n\n :param channel_name: name of the channel\n :type channel_name: string\n :return: Channel container\n :rtype: [ :class:`mth5.mth5_groups.ElectricDatset` |\n :class:`mth5.mth5_groups.MagneticDatset` |\n :class:`mth5.mth5_groups.AuxiliaryDatset` ]\n :raises MTH5Error: If no channel is found\n\n :Example:\n\n >>> existing_channel = run.get_channel('Ex')\n MTH5Error: Ex does not exist, check groups_list for existing names'\n\n >>> run.groups_list\n ['Ey', 'Hx', 'Hz']\n\n >>> existing_channel = run.get_channel('Ey')\n >>> existing_channel\n Channel Electric:\n -------------------\n component: Ey\n data type: electric\n data format: float32\n data shape: (4096,)\n start: 1980-01-01T00:00:00+00:00\n end: 1980-01-01T00:00:01+00:00\n sample rate: 4096\n\n\n \"\"\"\n\n channel_name = validate_name(channel_name.lower())\n try:\n ch_dataset = self.hdf5_group[channel_name]\n except KeyError:\n msg = (\n f\"{channel_name} does not exist, \"\n + \"check groups_list for existing names\"\n )\n self.logger.debug(\"Error\" + msg)\n raise MTH5Error(msg)\n\n if ch_dataset.attrs[\"mth5_type\"].lower() in [\"electric\"]:\n ch_metadata = meta_classes[\"Electric\"]()\n ch_metadata.from_dict({\"Electric\": ch_dataset.attrs})\n channel = ElectricDataset(\n ch_dataset,\n dataset_metadata=ch_metadata,\n write_metadata=False,\n )\n elif ch_dataset.attrs[\"mth5_type\"].lower() in [\"magnetic\"]:\n ch_metadata = meta_classes[\"Magnetic\"]()\n ch_metadata.from_dict({\"Magnetic\": ch_dataset.attrs})\n channel = MagneticDataset(\n ch_dataset,\n dataset_metadata=ch_metadata,\n write_metadata=False,\n )\n elif ch_dataset.attrs[\"mth5_type\"].lower() in [\"auxiliary\"]:\n ch_metadata = meta_classes[\"Auxiliary\"]()\n ch_metadata.from_dict({\"Auxiliary\": ch_dataset.attrs})\n channel = AuxiliaryDataset(\n ch_dataset,\n dataset_metadata=ch_metadata,\n write_metadata=False,\n )\n else:\n channel = ChannelDataset(ch_dataset)\n channel.read_metadata()\n\n return channel\n\n def remove_channel(self, channel_name):\n \"\"\"\n Remove a run from the station.\n\n .. note:: Deleting a channel is not as simple as del(channel). In HDF5\n this does not free up memory, it simply removes the reference\n to that channel. The common way to get around this is to\n copy what you want into a new file, or overwrite the channel.\n\n :param station_name: existing station name\n :type station_name: string\n\n :Example:\n\n >>> from mth5 import mth5\n >>> mth5_obj = mth5.MTH5()\n >>> mth5_obj.open_mth5(r\"/test.mth5\", mode='a')\n >>> run = mth5_obj.stations_group.get_station('MT001').get_run('MT001a')\n >>> run.remove_channel('Ex')\n\n .. todo:: Need to remove summary table entry as well.\n\n \"\"\"\n\n channel_name = validate_name(channel_name.lower())\n\n try:\n del self.hdf5_group[channel_name]\n self.logger.info(\n \"Deleting a channel does not reduce the HDF5\"\n + \"file size it simply remove the reference. If \"\n + \"file size reduction is your goal, simply copy\"\n + \" what you want into another file.\"\n )\n except KeyError:\n msg = (\n f\"{channel_name} does not exist, \"\n + \"check groups_list for existing names\"\n )\n self.logger.debug(\"Error\" + msg)\n raise MTH5Error(msg)\n\n def to_runts(self, start=None, end=None, n_samples=None):\n \"\"\"\n create a :class:`mth5.timeseries.RunTS` object from channels of the\n run\n\n :return: DESCRIPTION\n :rtype: TYPE\n\n \"\"\"\n ch_list = []\n for channel in self.groups_list:\n if channel in [\"summary\"]:\n continue\n ch_obj = self.get_channel(channel)\n\n if start is not None:\n ts_obj = ch_obj.time_slice(start, end=end, n_samples=n_samples)\n else:\n ts_obj = ch_obj.to_channel_ts()\n ch_list.append(ts_obj)\n return RunTS(\n ch_list,\n run_metadata=self.metadata,\n station_metadata=self.station_metadata,\n survey_metadata=self.survey_metadata,\n )\n\n def from_runts(self, run_ts_obj, **kwargs):\n \"\"\"\n create channel datasets from a :class:`mth5.timeseries.RunTS` object\n and update metadata.\n\n :parameter :class:`mth5.timeseries.RunTS` run_ts_obj: Run object with all\n the appropriate channels and metadata.\n\n Will create a run group and appropriate channel datasets.\n \"\"\"\n\n if not isinstance(run_ts_obj, RunTS):\n msg = f\"Input must be a mth5.timeseries.RunTS object not {type(run_ts_obj)}\"\n self.logger.error(msg)\n raise MTH5Error(msg)\n self.metadata.update(run_ts_obj.run_metadata)\n\n channels = []\n\n for comp in run_ts_obj.channels:\n ch = getattr(run_ts_obj, comp)\n\n if ch.station_metadata.id is not None:\n if ch.station_metadata.id != self.station_group.metadata.id:\n if ch.station_metadata.id not in [\"0\", None]:\n self.logger.warning(\n f\"Channel station.id {ch.station_metadata.id} != \"\n + f\" group station.id {self.station_group.metadata.id}\"\n )\n if ch.run_metadata.id is not None:\n if ch.run_metadata.id != self.metadata.id:\n if ch.run_metadata.id not in [\"0\", None]:\n self.logger.warning(\n f\"Channel run.id {ch.run_metadata.id} != \"\n + f\" group run.id {self.metadata.id}\"\n )\n\n channels.append(self.from_channel_ts(ch))\n\n self.update_run_metadata()\n return channels\n\n def from_channel_ts(self, channel_ts_obj):\n \"\"\"\n create a channel data set from a :class:`mth5.timeseries.ChannelTS` object and\n update metadata.\n\n :param channel_ts_obj: a single time series object\n :type channel_ts_obj: :class:`mth5.timeseries.ChannelTS`\n :return: new channel dataset\n :rtype: :class:`mth5.groups.ChannelDataset\n\n \"\"\"\n\n if not isinstance(channel_ts_obj, ChannelTS):\n msg = f\"Input must be a mth5.timeseries.ChannelTS object not {type(channel_ts_obj)}\"\n self.logger.error(msg)\n raise MTH5Error(msg)\n\n ## Need to add in the filters\n if channel_ts_obj.channel_response_filter.filters_list != []:\n from mth5.groups import FiltersGroup\n\n fg = FiltersGroup(self.hdf5_group.parent.parent.parent[\"Filters\"])\n for ff in channel_ts_obj.channel_response_filter.filters_list:\n fg.add_filter(ff)\n\n ch_obj = self.add_channel(\n channel_ts_obj.component,\n channel_ts_obj.channel_metadata.type,\n channel_ts_obj.ts,\n channel_metadata=channel_ts_obj.channel_metadata,\n )\n\n # need to update the channels recorded\n if channel_ts_obj.channel_metadata.type == \"electric\":\n if self.metadata.channels_recorded_electric is None:\n self.metadata.channels_recorded_electric = [\n channel_ts_obj.component\n ]\n elif (\n channel_ts_obj.component\n not in self.metadata.channels_recorded_electric\n ):\n self.metadata.channels_recorded_electric.append(\n channel_ts_obj.component\n )\n elif channel_ts_obj.channel_metadata.type == \"magnetic\":\n if self.metadata.channels_recorded_magnetic is None:\n self.metadata.channels_recorded_magnetic = [\n channel_ts_obj.component\n ]\n elif (\n channel_ts_obj.component\n not in self.metadata.channels_recorded_magnetic\n ):\n self.metadata.channels_recorded_magnetic.append(\n channel_ts_obj.component\n )\n elif channel_ts_obj.channel_metadata.type == \"auxiliary\":\n if self.metadata.channels_recorded_auxiliary is None:\n self.metadata.channels_recorded_auxiliary = [\n channel_ts_obj.component\n ]\n elif (\n channel_ts_obj.component\n not in self.metadata.channels_recorded_auxiliary\n ):\n self.metadata.channels_recorded_auxiliary.append(\n channel_ts_obj.component\n )\n return ch_obj\n\n def update_run_metadata(self):\n \"\"\"\n Update metadata and table entries to ensure consistency\n\n :return: DESCRIPTION\n :rtype: TYPE\n\n \"\"\"\n channel_summary = self.channel_summary.copy()\n\n self._metadata.time_period.start = (\n channel_summary.start.min().isoformat()\n )\n self._metadata.time_period.end = channel_summary.end.max().isoformat()\n self._metadata.sample_rate = channel_summary.sample_rate.unique()[0]\n self.write_metadata()\n\n def plot(self, start=None, end=None, n_samples=None):\n \"\"\"\n Produce a simple matplotlib plot using runts\n \"\"\"\n\n runts = self.to_runts(start=start, end=end, n_samples=n_samples)\n\n runts.plot()\n\n\nclass ChannelDataset:\n \"\"\"\n Holds a channel dataset. This is a simple container for the data to make\n sure that the user has the flexibility to turn the channel into an object\n they want to deal with.\n\n For now all the numpy type slicing can be used on `hdf5_dataset`\n\n :param dataset: dataset object for the channel\n :type dataset: :class:`h5py.Dataset`\n :param dataset_metadata: metadata container, defaults to None\n :type dataset_metadata: [ :class:`mth5.metadata.Electric` |\n :class:`mth5.metadata.Magnetic` |\n :class:`mth5.metadata.Auxiliary` ], optional\n :raises MTH5Error: If the dataset is not of the correct type\n\n Utilities will be written to create some common objects like:\n\n * xarray.DataArray\n * pandas.DataFrame\n * zarr\n * dask.Array\n\n The benefit of these other objects is that they can be indexed by time,\n and they have much more buit-in funcionality.\n\n .. code-block:: python\n\n >>> from mth5 import mth5\n >>> mth5_obj = mth5.MTH5()\n >>> mth5_obj.open_mth5(r\"/test.mth5\", mode='a')\n >>> run = mth5_obj.stations_group.get_station('MT001').get_run('MT001a')\n >>> channel = run.get_channel('Ex')\n >>> channel\n Channel Electric:\n -------------------\n component: Ey\n data type: electric\n data format: float32\n data shape: (4096,)\n start: 1980-01-01T00:00:00+00:00\n end: 1980-01-01T00:00:01+00:00\n sample rate: 4096\n\n \"\"\"\n\n def __init__(\n self, dataset, dataset_metadata=None, write_metadata=True, **kwargs\n ):\n for key, value in kwargs.items():\n setattr(self, key, value)\n if dataset is not None and isinstance(dataset, (h5py.Dataset)):\n self.hdf5_dataset = weakref.ref(dataset)()\n self.logger = setup_logger(f\"{__name__}.{self._class_name}\")\n\n # set metadata to the appropriate class. Standards is not a\n # Base object so should be skipped. If the class name is not\n # defined yet set to Base class.\n self.metadata = Base()\n try:\n self.metadata = meta_classes[self._class_name]()\n except KeyError:\n self.metadata = Base()\n if not hasattr(self.metadata, \"mth5_type\"):\n self._add_base_attributes()\n self.metadata.hdf5_reference = self.hdf5_dataset.ref\n self.metadata.mth5_type = self._class_name\n # if the input data set already has filled attributes, namely if the\n # channel data already exists then read them in with our writing back\n if \"mth5_type\" in list(self.hdf5_dataset.attrs.keys()):\n self.metadata.from_dict(\n {self.hdf5_dataset.attrs[\"mth5_type\"]: self.hdf5_dataset.attrs}\n )\n # if metadata is input, make sure that its the same class type amd write\n # to the hdf5 dataset\n if dataset_metadata is not None:\n if not isinstance(dataset_metadata, type(self.metadata)):\n msg = \"metadata must be type metadata.%s not %s\"\n self.logger.error(\n msg, self._class_name, type(dataset_metadata)\n )\n raise MTH5Error(\n msg % (self._class_name, type(dataset_metadata))\n )\n # load from dict because of the extra attributes for MTH5\n self.metadata.from_dict(dataset_metadata.to_dict())\n self.metadata.hdf5_reference = self.hdf5_dataset.ref\n self.metadata.mth5_type = self._class_name\n\n # write out metadata to make sure that its in the file.\n if write_metadata:\n self.write_metadata()\n # if the attrs don't have the proper metadata keys yet write them\n if not \"mth5_type\" in list(self.hdf5_dataset.attrs.keys()):\n self.write_metadata()\n\n def _add_base_attributes(self):\n # add 2 attributes that will help with querying\n # 1) the metadata class name\n self.metadata.add_base_attribute(\n \"mth5_type\",\n self._class_name,\n {\n \"type\": str,\n \"required\": True,\n \"style\": \"free form\",\n \"description\": \"type of group\",\n \"units\": None,\n \"options\": [],\n \"alias\": [],\n \"example\": \"group_name\",\n \"default\": None,\n },\n )\n\n # 2) the HDF5 reference that can be used instead of paths\n self.metadata.add_base_attribute(\n \"hdf5_reference\",\n self.hdf5_dataset.ref,\n {\n \"type\": \"h5py_reference\",\n \"required\": True,\n \"style\": \"free form\",\n \"description\": \"hdf5 internal reference\",\n \"units\": None,\n \"options\": [],\n \"alias\": [],\n \"example\": \"\",\n \"default\": None,\n },\n )\n\n def __str__(self):\n try:\n lines = [\"Channel {0}:\".format(self._class_name)]\n lines.append(\"-\" * (len(lines[0]) + 2))\n info_str = \"\\t{0:<18}{1}\"\n lines.append(\n info_str.format(\"component:\", self.metadata.component)\n )\n lines.append(info_str.format(\"data type:\", self.metadata.type))\n lines.append(\n info_str.format(\"data format:\", self.hdf5_dataset.dtype)\n )\n lines.append(\n info_str.format(\"data shape:\", self.hdf5_dataset.shape)\n )\n lines.append(\n info_str.format(\"start:\", self.metadata.time_period.start)\n )\n lines.append(\n info_str.format(\"end:\", self.metadata.time_period.end)\n )\n lines.append(\n info_str.format(\"sample rate:\", self.metadata.sample_rate)\n )\n return \"\\n\".join(lines)\n except ValueError:\n return \"MTH5 file is closed and cannot be accessed.\"\n\n def __repr__(self):\n return self.__str__()\n\n @property\n def _class_name(self):\n return self.__class__.__name__.split(\"Dataset\")[0]\n\n @property\n def run_group(self):\n \"\"\"shortcut to run group\"\"\"\n return RunGroup(self.hdf5_dataset.parent)\n\n @property\n def run_metadata(self):\n \"\"\"run metadata\"\"\"\n\n meta_dict = dict(self.hdf5_dataset.parent.attrs)\n for key, value in meta_dict.items():\n meta_dict[key] = from_numpy_type(value)\n run_metadata = metadata.Run()\n run_metadata.from_dict({\"run\": meta_dict})\n return run_metadata\n\n @property\n def station_group(self):\n \"\"\"shortcut to station group\"\"\"\n\n return StationGroup(self.hdf5_dataset.parent.parent)\n\n @property\n def station_metadata(self):\n \"\"\"station metadata\"\"\"\n\n meta_dict = dict(self.hdf5_dataset.parent.parent.attrs)\n for key, value in meta_dict.items():\n meta_dict[key] = from_numpy_type(value)\n station_metadata = metadata.Station()\n station_metadata.from_dict({\"station\": meta_dict})\n return station_metadata\n\n @property\n def master_station_group(self):\n \"\"\"shortcut to master station group\"\"\"\n\n return MasterStationGroup(self.hdf5_dataset.parent.parent.parent)\n\n @property\n def survey_metadata(self):\n \"\"\"survey metadata\"\"\"\n\n meta_dict = dict(self.hdf5_dataset.parent.parent.parent.parent.attrs)\n for key, value in meta_dict.items():\n meta_dict[key] = from_numpy_type(value)\n survey_metadata = metadata.Survey()\n survey_metadata.from_dict({\"survey\": meta_dict})\n return survey_metadata\n\n @property\n def survey_id(self):\n \"\"\"shortcut to survey group\"\"\"\n\n return self.hdf5_dataset.parent.parent.parent.parent.attrs[\"id\"]\n\n @property\n def channel_response_filter(self):\n # get the filters to make a channel response\n filters_group = FiltersGroup(\n self.hdf5_dataset.parent.parent.parent.parent[\"Filters\"]\n )\n f_list = []\n for name in self.metadata.filter.name:\n name = name.replace(\"/\", \" per \").lower()\n try:\n f_list.append(filters_group.to_filter_object(name))\n except KeyError:\n self.logger.warning(\"Could not locate filter %s\", name)\n continue\n return ChannelResponseFilter(filters_list=f_list)\n\n @property\n def start(self):\n return self.metadata.time_period._start_dt\n\n @start.setter\n def start(self, value):\n \"\"\"set start time and validate through metadata validator\"\"\"\n if isinstance(value, MTime):\n self.metadata.time_period.start = value.iso_str\n else:\n self.metadata.time_period.start = value\n\n @property\n def end(self):\n \"\"\"return end time based on the data\"\"\"\n return self.start + (self.n_samples / self.sample_rate)\n\n @property\n def sample_rate(self):\n return self.metadata.sample_rate\n\n @sample_rate.setter\n def sample_rate(self, value):\n \"\"\"set sample rate through metadata validator\"\"\"\n self.metadata.sample_rate = value\n\n @property\n def n_samples(self):\n return self.hdf5_dataset.size\n\n @property\n def time_index(self):\n \"\"\"\n Create a time index based on the metadata. This can help when asking\n for time windows from the data\n\n :return: DESCRIPTION\n :rtype: TYPE\n\n \"\"\"\n\n return make_dt_coordinates(\n self.start, self.sample_rate, self.n_samples, self._logger\n )\n\n def read_metadata(self):\n \"\"\"\n Read metadata from the HDF5 file into the metadata container, that\n way it can be validated.\n\n \"\"\"\n meta_dict = dict(self.hdf5_dataset.attrs)\n for key, value in meta_dict.items():\n meta_dict[key] = from_numpy_type(value)\n self.metadata.from_dict({self._class_name: meta_dict})\n\n def write_metadata(self):\n \"\"\"\n Write metadata from the metadata container to the HDF5 attrs\n dictionary.\n\n \"\"\"\n meta_dict = self.metadata.to_dict()[self.metadata._class_name.lower()]\n for key, value in meta_dict.items():\n value = to_numpy_type(value)\n self.hdf5_dataset.attrs.create(key, value)\n\n def replace_dataset(self, new_data_array):\n \"\"\"\n replace the entire dataset with a new one, nothing left behind\n\n :param new_data_array: new data array shape (npts, )\n :type new_data_array: :class:`numpy.ndarray`\n\n \"\"\"\n if not isinstance(new_data_array, np.ndarray):\n try:\n new_data_array = np.array(new_data_array)\n except (ValueError, TypeError) as error:\n msg = f\"{error} Input must be a numpy array not {type(new_data_array)}\"\n self.logger.exception(msg)\n raise TypeError(msg)\n if new_data_array.shape != self.hdf5_dataset.shape:\n self.hdf5_dataset.resize(new_data_array.shape)\n self.hdf5_dataset[...] = new_data_array\n\n def extend_dataset(\n self,\n new_data_array,\n start_time,\n sample_rate,\n fill=None,\n max_gap_seconds=1,\n fill_window=10,\n ):\n \"\"\"\n Append data according to how the start time aligns with existing\n data. If the start time is before existing start time the data is\n prepended, similarly if the start time is near the end data will be\n appended.\n\n If the start time is within the existing time range, existing data\n will be replace with the new data.\n\n If there is a gap between start or end time of the new data with\n the existing data you can either fill the data with a constant value\n or an error will be raise depending on the value of fill.\n\n :param new_data_array: new data array with shape (npts, )\n :type new_data_array: :class:`numpy.ndarray`\n :param start_time: start time of the new data array in UTC\n :type start_time: string or :class:`mth5.utils.mttime.MTime`\n :param sample_rate: Sample rate of the new data array, must match\n existing sample rate\n :type sample_rate: float\n :param fill: If there is a data gap how do you want to fill the gap\n * None: will raise an :class:`mth5.utils.exceptions.MTH5Error`\n * 'mean': will fill with the mean of each data set within\n the fill window\n * 'median': will fill with the median of each data set\n within the fill window\n * value: can be an integer or float to fill the gap\n * 'nan': will fill the gap with NaN\n :type fill: string, None, float, integer\n :param max_gap_seconds: sets a maximum number of seconds the gap can\n be. Anything over this number will raise\n a :class:`mth5.utils.exceptions.MTH5Error`.\n :type max_gap_seconds: float or integer\n :param fill_window: number of points from the end of each data set\n to estimate fill value from.\n :type fill_window: integer\n\n :raises: :class:`mth5.utils.excptions.MTH5Error` if sample rate is\n not the same, or fill value is not understood,\n\n :Append Example:\n\n >>> ex = mth5_obj.get_channel('MT001', 'MT001a', 'Ex')\n >>> ex.n_samples\n 4096\n >>> ex.end\n 2015-01-08T19:32:09.500000+00:00\n >>> t = timeseries.ChannelTS('electric',\n ... data=2*np.cos(4 * np.pi * .05 * \\\n ... np.linspace(0,4096l num=4096) *\n ... .01),\n ... channel_metadata={'electric':{\n ... 'component': 'ex',\n ... 'sample_rate': 8,\n ... 'time_period.start':(ex.end+(1)).iso_str}})\n >>> ex.extend_dataset(t.ts, t.start, t.sample_rate, fill='median',\n ... max_gap_seconds=2)\n 2020-07-02T18:02:47 - mth5.groups.Electric.extend_dataset - INFO -\n filling data gap with 1.0385180759767025\n >>> ex.n_samples\n 8200\n >>> ex.end\n 2015-01-08T19:40:42.500000+00:00\n\n \"\"\"\n fw = fill_window\n # check input parameters\n if sample_rate != self.sample_rate:\n msg = (\n \"new data must have the same sampling rate as existing data.\\n\"\n + f\"\\tnew sample rate = {sample_rate}\\n\"\n + f\"\\texisting sample rate = {self.sample_rate}\"\n )\n self.logger.error(msg)\n raise MTH5Error(msg)\n if not isinstance(new_data_array, np.ndarray):\n try:\n new_data_array = np.array(new_data_array)\n except (ValueError, TypeError) as error:\n msg = f\"{error} Input must be a numpy array not {type(new_data_array)}\"\n self.logger.exception(msg)\n raise TypeError(msg)\n if not isinstance(start_time, MTime):\n start_time = MTime(start_time)\n # get end time will need later\n end_time = start_time + (new_data_array.size / sample_rate)\n\n # check start time\n start_t_diff = self._get_diff_new_array_start(start_time)\n end_t_diff = self._get_diff_new_array_end(end_time)\n\n self.logger.info(\"Extending data.\")\n self.logger.info(f\"Existing start time {self.start}\")\n self.logger.info(f\"New start time {start_time}\")\n self.logger.info(f\"Existing end time {self.end}\")\n self.logger.info(f\"New end time {end_time}\")\n\n # prepend data\n if start_t_diff < 0:\n self.logger.info(\"Prepending: \")\n self.logger.info(\n f\"new start time {start_time} is before existing {self.start}\"\n )\n if end_time.iso_no_tz not in self.time_index:\n gap = abs(end_time - self.start)\n if gap > 0:\n if gap > max_gap_seconds:\n msg = (\n f\"Time gap of {gap} seconds \"\n + f\"is more than max_gap_seconds = {max_gap_seconds}.\"\n + \" Consider making a new run.\"\n )\n self.logger.error(msg)\n raise MTH5Error(msg)\n if fill is None:\n msg = (\n f\"A time gap of {gap} seconds is found \"\n + \"between new and existing data sets. \\n\"\n + f\"\\tnew end time: {end_time}\\n\"\n + f\"\\texisting start time: {self.start}\"\n )\n self.logger.error(msg)\n raise MTH5Error(msg)\n # set new start time\n old_slice = self.time_slice(self.start, end_time=self.end)\n old_start = self.start.copy()\n self.start = start_time\n\n # resize the existing data to make room for new data\n self.hdf5_dataset.resize(\n (\n int(\n new_data_array.size\n + self.hdf5_dataset.size\n + gap * sample_rate\n ),\n )\n )\n\n # fill based on time, refill existing data first\n self.hdf5_dataset[\n self.get_index_from_time(old_start) :\n ] = old_slice.ts.values\n self.hdf5_dataset[\n 0 : self.get_index_from_time(end_time)\n ] = new_data_array\n\n if fill == \"mean\":\n fill_value = np.mean(\n np.array(\n [\n new_data_array[-fw:].mean(),\n float(old_slice.ts[0:fw].mean()),\n ]\n )\n )\n elif fill == \"median\":\n fill_value = np.median(\n np.array(\n [\n np.median(new_data_array[-fw:]),\n np.median(old_slice.ts[0:fw]),\n ]\n )\n )\n elif fill == \"nan\":\n fill_value = np.nan\n elif isinstance(fill, (int, float)):\n fill_value = fill\n else:\n msg = f\"fill value {fill} is not understood\"\n self.logger.error(msg)\n raise MTH5Error(msg)\n self.logger.info(f\"filling data gap with {fill_value}\")\n self.hdf5_dataset[\n self.get_index_from_time(\n end_time\n ) : self.get_index_from_time(old_start)\n ] = fill_value\n else:\n new_size = (\n self.n_samples + int(abs(start_t_diff) * sample_rate),\n )\n overlap = abs(end_time - self.start)\n self.logger.warning(\n f\"New data is overlapping by {overlap} s.\"\n + \" Any overlap will be overwritten.\"\n )\n # set new start time\n old_slice = self.time_slice(self.start, end_time=self.end)\n old_start = self.start.copy()\n self.start = start_time\n self.logger.debug(\n f\"resizing data set from {self.n_samples} to {new_size}\"\n )\n self.hdf5_dataset.resize(new_size)\n\n # put back the existing data, which any overlapping times\n # will be overwritten\n self.hdf5_dataset[\n self.get_index_from_time(old_start) :\n ] = old_slice.ts.values\n self.hdf5_dataset[\n 0 : self.get_index_from_time(end_time)\n ] = new_data_array\n # append data\n elif start_t_diff > 0:\n old_end = self.end.copy()\n if start_time.iso_no_tz not in self.time_index:\n gap = abs(self.end - start_time)\n if gap > 0:\n if gap > max_gap_seconds:\n msg = (\n f\"Time gap of {gap} seconds \"\n + f\"is more than max_gap_seconds = {max_gap_seconds}.\"\n + \" Consider making a new run.\"\n )\n self.logger.error(msg)\n raise MTH5Error(msg)\n if fill is None:\n msg = (\n f\"A time gap of {gap} seconds is found \"\n + \"between new and existing data sets. \\n\"\n + f\"\\tnew start time: {start_time}\\n\"\n + f\"\\texisting end time: {self.end}\"\n )\n self.logger.error(msg)\n raise MTH5Error(msg)\n # resize the existing data to make room for new data\n self.hdf5_dataset.resize(\n (\n int(\n new_data_array.size\n + self.hdf5_dataset.size\n + gap * sample_rate\n ),\n )\n )\n\n self.hdf5_dataset[\n self.get_index_from_time(start_time) :\n ] = new_data_array\n old_index = self.get_index_from_time(old_end)\n if fill == \"mean\":\n fill_value = np.mean(\n np.array(\n [\n new_data_array[0:fw].mean(),\n np.mean(\n self.hdf5_dataset[old_index - fw :]\n ),\n ]\n )\n )\n elif fill == \"median\":\n fill_value = np.median(\n np.array(\n [\n np.median(new_data_array[0:fw]),\n np.median(\n self.hdf5_dataset[old_index - fw :]\n ),\n ]\n )\n )\n elif fill == \"nan\":\n fill_value = np.nan\n elif isinstance(fill, (int, float)):\n fill_value = fill\n else:\n msg = f\"fill value {fill} is not understood\"\n self.logger.error(msg)\n raise MTH5Error(msg)\n self.logger.info(f\"filling data gap with {fill_value}\")\n self.hdf5_dataset[\n self.get_index_from_time(\n old_end\n ) : self.get_index_from_time(start_time)\n ] = fill_value\n else:\n # if the new data fits within the extisting time span\n if end_t_diff < 0:\n self.logger.debug(\n \"New data fits within existing time span\"\n + \" all data in the window : \"\n f\"{start_time} -- {end_time} \" + \"will be overwritten.\"\n )\n self.hdf5_dataset[\n self.get_index_from_time(\n start_time\n ) : self.get_index_from_time(end_time)\n ] = new_data_array\n else:\n new_size = (\n self.n_samples + int(abs(start_t_diff) * sample_rate),\n )\n overlap = abs(self.end - start_time)\n self.logger.warning(\n f\"New data is overlapping by {overlap} s.\"\n + \" Any overlap will be overwritten.\"\n )\n\n self.logger.debug(\n f\"resizing data set from {self.n_samples} to {new_size}\"\n )\n self.hdf5_dataset.resize(new_size)\n\n # put back the existing data, which any overlapping times\n # will be overwritten\n self.hdf5_dataset[\n self.get_index_from_time(start_time) :\n ] = new_data_array\n\n def to_channel_ts(self):\n \"\"\"\n :return: a Timeseries with the appropriate time index and metadata\n :rtype: :class:`mth5.timeseries.ChannelTS`\n\n loads from memory (nearly half the size of xarray alone, not sure why)\n\n \"\"\"\n return ChannelTS(\n channel_type=self.metadata.type,\n data=self.hdf5_dataset[()],\n channel_metadata=self.metadata,\n run_metadata=self.run_metadata.copy(),\n station_metadata=self.station_metadata.copy(),\n survey_metadata=self.survey_metadata.copy(),\n channel_response_filter=self.channel_response_filter,\n )\n\n def to_xarray(self):\n \"\"\"\n :return: an xarray DataArray with appropriate metadata and the\n appropriate time index.\n :rtype: :class:`xarray.DataArray`\n\n .. note:: that metadta will not be validated if changed in an xarray.\n\n loads from memory\n \"\"\"\n\n return xr.DataArray(\n self.hdf5_dataset[()],\n coords=[(\"time\", self.time_index)],\n attrs=self.metadata.to_dict(single=True),\n )\n\n def to_dataframe(self):\n \"\"\"\n\n :return: a dataframe where data is stored in the 'data' column and\n attributes are stored in the experimental attrs attribute\n :rtype: :class:`pandas.DataFrame`\n\n .. note:: that metadta will not be validated if changed in an xarray.\n\n loads into RAM\n \"\"\"\n\n df = pd.DataFrame(\n {\"data\": self.hdf5_dataset[()]}, index=self.time_index\n )\n df.attrs.update(self.metadata.to_dict(single=True))\n\n return df\n\n def to_numpy(self):\n \"\"\"\n :return: a numpy structured array with 2 columns (time, channel_data)\n :rtype: :class:`numpy.core.records`\n\n .. note:: data is a builtin to numpy and cannot be used as a name\n\n loads into RAM\n\n \"\"\"\n\n return np.core.records.fromarrays(\n [self.time_index.to_numpy(), self.hdf5_dataset[()]],\n names=\"time,channel_data\",\n )\n\n def from_channel_ts(\n self,\n channel_ts_obj,\n how=\"replace\",\n fill=None,\n max_gap_seconds=1,\n fill_window=10,\n ):\n \"\"\"\n fill data set from a :class:`mth5.timeseries.ChannelTS` object.\n\n Will check for time alignement, and metadata.\n\n :param channel_ts_obj: time series object\n :type channel_ts_obj: :class:`mth5.timeseries.ChannelTS`\n :param how: how the new array will be input to the existing dataset:\n\n - 'replace' -> replace the entire dataset nothing is left over.\n - 'extend' -> add onto the existing dataset, any overlapping\n values will be rewritten, if there are gaps between data sets\n those will be handled depending on the value of fill.\n\n :param fill: If there is a data gap how do you want to fill the gap:\n\n - None -> will raise an :class:`mth5.utils.exceptions.MTH5Error`\n - 'mean'-> will fill with the mean of each data set within\n the fill window\n - 'median' -> will fill with the median of each data set\n within the fill window\n - value -> can be an integer or float to fill the gap\n - 'nan' -> will fill the gap with NaN\n\n :type fill: string, None, float, integer\n :param max_gap_seconds: sets a maximum number of seconds the gap can\n be. Anything over this number will raise\n a :class:`mth5.utils.exceptions.MTH5Error`.\n\n :type max_gap_seconds: float or integer\n :param fill_window: number of points from the end of each data set\n to estimate fill value from.\n\n :type fill_window: integer\n\n \"\"\"\n\n if not isinstance(channel_ts_obj, ChannelTS):\n msg = (\n f\"Input must be a ChannelTS object not {type(channel_ts_obj)}\"\n )\n self.logger.error(msg)\n raise TypeError(msg)\n if how == \"replace\":\n self.metadata.from_dict(channel_ts_obj.channel_metadata.to_dict())\n self.replace_dataset(channel_ts_obj.ts)\n # apparently need to reset these otherwise they get overwritten with None\n self.metadata.hdf5_reference = self.hdf5_dataset.ref\n self.metadata.mth5_type = self._class_name\n self.write_metadata()\n elif how == \"extend\":\n self.extend_dataset(\n channel_ts_obj.ts,\n channel_ts_obj.start,\n channel_ts_obj.sample_rate,\n fill=fill,\n )\n #\n # TODO need to check on metadata.\n\n def from_xarray(\n self,\n data_array,\n how=\"replace\",\n fill=None,\n max_gap_seconds=1,\n fill_window=10,\n ):\n \"\"\"\n fill data set from a :class:`xarray.DataArray` object.\n\n Will check for time alignement, and metadata.\n\n :param data_array_obj: Xarray data array\n :type channel_ts_obj: :class:`xarray.DataArray`\n :param how: how the new array will be input to the existing dataset:\n\n - 'replace' -> replace the entire dataset nothing is left over.\n - 'extend' -> add onto the existing dataset, any overlapping\n values will be rewritten, if there are gaps between data sets\n those will be handled depending on the value of fill.\n\n :param fill: If there is a data gap how do you want to fill the gap:\n\n - None -> will raise an :class:`mth5.utils.exceptions.MTH5Error`\n - 'mean'-> will fill with the mean of each data set within\n the fill window\n - 'median' -> will fill with the median of each data set\n within the fill window\n - value -> can be an integer or float to fill the gap\n - 'nan' -> will fill the gap with NaN\n\n :type fill: string, None, float, integer\n :param max_gap_seconds: sets a maximum number of seconds the gap can\n be. Anything over this number will raise a\n :class:`mth5.utils.exceptions.MTH5Error`.\n\n :type max_gap_seconds: float or integer\n :param fill_window: number of points from the end of each data set\n to estimate fill value from.\n\n :type fill_window: integer\n\n \"\"\"\n\n if not isinstance(data_array, xr.DataArray):\n msg = f\"Input must be a xarray.DataArray object not {type(data_array)}\"\n self.logger.error(msg)\n raise TypeError(msg)\n if how == \"replace\":\n self.metadata.from_dict(\n {self.metadata._class_name: data_array.attrs}\n )\n self.replace_dataset(data_array.values)\n self.write_metadata()\n elif how == \"extend\":\n self.extend_dataset(\n data_array.values,\n data_array.coords.indexes[\"time\"][0].isoformat(),\n 1e9 / data_array.coords.indexes[\"time\"][0].freq.nanos,\n fill=fill,\n )\n # TODO need to check on metadata.\n\n def _get_diff_new_array_start(self, start_time):\n \"\"\"\n Make sure the new array has the same start time if not return the\n time difference\n\n :param start_time: start time of the new array\n :type start_time: string, int or :class:`mth5.utils.MTime`\n :return: time difference in seconds as new start time minus old.\n\n * A positive number means new start time is later than old\n start time.\n * A negative number means the new start time is earlier than\n the old start time.\n\n :rtype: float\n\n \"\"\"\n if not isinstance(start_time, MTime):\n start_time = MTime(start_time)\n t_diff = 0\n if start_time != self.start:\n t_diff = start_time - self.start\n return t_diff\n\n def _get_diff_new_array_end(self, end_time):\n \"\"\"\n Make sure the new array has the same end time if not return the\n time difference\n\n :param end_time: end time of the new array\n :type end_time: string, int or :class:`mth5.utils.MTime`\n :return: time difference in seconds as new end time minus old.\n\n * A positive number means new end time is later than old\n end time.\n * A negative number means the new end time is earlier than\n the old end time.\n\n :rtype: float\n\n \"\"\"\n if not isinstance(end_time, MTime):\n end_time = MTime(end_time)\n t_diff = 0\n if end_time != self.end:\n t_diff = end_time - self.end\n return t_diff\n\n @property\n def table_entry(self):\n \"\"\"\n Creat a table entry to put into the run summary table.\n\n \"\"\"\n\n return np.array(\n [\n (\n self.metadata.component,\n self.metadata.time_period._start_dt.iso_no_tz,\n self.metadata.time_period._end_dt.iso_no_tz,\n self.hdf5_dataset.size,\n self.metadata.type,\n self.metadata.units,\n self.hdf5_dataset.ref,\n )\n ],\n dtype=np.dtype(\n [\n (\"component\", \"U20\"),\n (\"start\", \"datetime64[ns]\"),\n (\"end\", \"datetime64[ns]\"),\n (\"n_samples\", int),\n (\"measurement_type\", \"U12\"),\n (\"units\", \"U25\"),\n (\"hdf5_reference\", h5py.ref_dtype),\n ]\n ),\n )\n\n @property\n def channel_entry(self):\n \"\"\"\n channel entry that will go into a full channel summary of the entire survey\n\n \"\"\"\n return np.array(\n [\n (\n self.survey_id,\n self.hdf5_dataset.parent.parent.attrs[\"id\"],\n self.hdf5_dataset.parent.attrs[\"id\"],\n self.hdf5_dataset.parent.parent.attrs[\"location.latitude\"],\n self.hdf5_dataset.parent.parent.attrs[\n \"location.longitude\"\n ],\n self.hdf5_dataset.parent.parent.attrs[\n \"location.elevation\"\n ],\n self.metadata.component,\n self.metadata.time_period.start,\n self.metadata.time_period.end,\n self.hdf5_dataset.size,\n self.metadata.sample_rate,\n self.metadata.type,\n self.metadata.measurement_azimuth,\n self.metadata.measurement_tilt,\n self.metadata.units,\n self.hdf5_dataset.ref,\n self.hdf5_dataset.parent.ref,\n self.hdf5_dataset.parent.parent.ref,\n )\n ],\n dtype=CHANNEL_DTYPE,\n )\n\n def time_slice(\n self,\n start,\n end=None,\n n_samples=None,\n return_type=\"channel_ts\",\n ):\n \"\"\"\n Get a time slice from the channel and return the appropriate type\n\n * numpy array with metadata\n * pandas.Dataframe with metadata\n * xarray.DataFrame with metadata\n * :class:`mth5.timeseries.ChannelTS` 'default'\n * dask.DataFrame with metadata 'not yet'\n\n :param start: start time of the slice\n :type start: string or :class:`mth5.utils.mttime.MTime`\n :param end: end time of the slice\n :type end: string or :class:`mth5.utils.mttime.MTime`, optional\n :param n_samples: number of samples to read in\n :type n_samples: integer, optional\n :return: the correct container for the time series.\n :rtype: [ :class:`xarray.DataArray` | :class:`pandas.DataFrame` |\n :class:`mth5.timeseries.ChannelTS` | :class:`numpy.ndarray` ]\n :raises: ValueError if both end_time and n_samples are None or given.\n\n :Example with number of samples:\n\n .. code-block::\n\n >>> ex = mth5_obj.get_channel('FL001', 'FL001a', 'Ex')\n >>> ex_slice = ex.time_slice(\"2015-01-08T19:49:15\", n_samples=4096)\n >>> ex_slice\n \n array([0.93115046, 0.14233688, 0.87917119, ..., 0.26073634, 0.7137319 ,\n 0.88154395])\n Coordinates:\n * time (time) datetime64[ns] 2015-01-08T19:49:15 ... 2015-01-08T19:57:46.875000\n Attributes:\n ac.end: None\n ac.start: None\n ...\n\n >>> type(ex_slice)\n mth5.timeseries.ChannelTS\n\n # plot the time series\n >>> ex_slice.ts.plot()\n\n :Example with start and end time:\n\n >>> ex_slice = ex.time_slice(\"2015-01-08T19:49:15\",\n ... end_time=\"2015-01-09T19:49:15\")\n\n :Raises Example:\n\n >>> ex_slice = ex.time_slice(\"2015-01-08T19:49:15\",\n ... end_time=\"2015-01-09T19:49:15\",\n ... n_samples=4096)\n ValueError: Must input either end_time or n_samples, not both.\n\n \"\"\"\n\n if not isinstance(start, MTime):\n start = MTime(start)\n if end is not None:\n if not isinstance(end, MTime):\n end = MTime(end)\n if n_samples is not None:\n n_samples = int(n_samples)\n if n_samples is None and end is None:\n msg = \"Must input either end_time or n_samples.\"\n self.logger.error(msg)\n raise ValueError(msg)\n if n_samples is not None and end is not None:\n msg = \"Must input either end_time or n_samples, not both.\"\n self.logger.error(msg)\n raise ValueError(msg)\n # if end time is given\n if end is not None and n_samples is None:\n start_index = self.get_index_from_time(start)\n end_index = self.get_index_from_time(end)\n npts = int(end_index - start_index)\n # if n_samples are given\n elif end is None and n_samples is not None:\n start_index = self.get_index_from_time(start)\n end_index = start_index + (n_samples - 1)\n npts = n_samples\n if npts > self.hdf5_dataset.size or end_index > self.hdf5_dataset.size:\n msg = (\n \"Requested slice is larger than data. \"\n f\"Slice length = {npts}, data length = {self.hdf5_dataset.shape}. \"\n f\"Setting end_index to {self.hdf5_dataset.shape}\"\n )\n end_index = self.hdf5_dataset.size - 1\n self.logger.warning(msg)\n # create a regional reference that can be used, need +1 to be inclusive\n try:\n regional_ref = self.hdf5_dataset.regionref[\n start_index : end_index + 1\n ]\n except (OSError, RuntimeError):\n self.logger.debug(\n \"file is in read mode cannot set an internal reference, using index values\"\n )\n regional_ref = slice(start_index, end_index)\n dt_index = make_dt_coordinates(\n start, self.sample_rate, npts, self.logger\n )\n\n meta_dict = self.metadata.to_dict()[self.metadata._class_name]\n meta_dict[\"time_period.start\"] = dt_index[0].isoformat()\n meta_dict[\"time_period.end\"] = dt_index[-1].isoformat()\n\n data = None\n if return_type == \"xarray\":\n # need the +1 to be inclusive of the last point\n data = xr.DataArray(\n self.hdf5_dataset[regional_ref], coords=[(\"time\", dt_index)]\n )\n data.attrs.update(meta_dict)\n elif return_type == \"pandas\":\n data = pd.DataFrame(\n {\"data\": self.hdf5_dataset[regional_ref]}, index=dt_index\n )\n data.attrs.update(meta_dict)\n elif return_type == \"numpy\":\n data = self.hdf5_dataset[regional_ref]\n elif return_type == \"channel_ts\":\n data = ChannelTS(\n self.metadata.type,\n data=self.hdf5_dataset[regional_ref],\n channel_metadata={self.metadata.type: meta_dict},\n channel_response_filter=self.channel_response_filter,\n )\n else:\n msg = \"return_type not understood, must be [ pandas | numpy | channel_ts ]\"\n self.logger.error(msg)\n raise ValueError(msg)\n return data\n\n def get_index_from_time(self, given_time):\n \"\"\"\n get the appropriate index for a given time.\n\n :param given_time: time string\n :type given_time: string or MTime\n :return: index value\n :rtype: int\n\n \"\"\"\n\n if not isinstance(given_time, MTime):\n given_time = MTime(given_time)\n index = (\n given_time - self.metadata.time_period.start\n ) * self.metadata.sample_rate\n\n return int(round(index))\n\n\n@inherit_doc_string\nclass ElectricDataset(ChannelDataset):\n def __init__(self, group, **kwargs):\n\n super().__init__(group, **kwargs)\n\n\n@inherit_doc_string\nclass MagneticDataset(ChannelDataset):\n def __init__(self, group, **kwargs):\n\n super().__init__(group, **kwargs)\n\n\n@inherit_doc_string\nclass AuxiliaryDataset(ChannelDataset):\n def __init__(self, group, **kwargs):\n super().__init__(group, **kwargs)\n", "sub_path": "mth5/groups/master_station_run_channel.py", "file_name": "master_station_run_channel.py", "file_ext": "py", "file_size_in_byte": 104435, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "inspect.getmembers", "line_number": 46, "usage_type": "call"}, {"api_name": "mt_metadata.timeseries", "line_number": 46, "usage_type": "argument"}, {"api_name": "inspect.isclass", "line_number": 46, "usage_type": "attribute"}, {"api_name": "mth5.groups.base.BaseGroup", "line_number": 52, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 223, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 224, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 246, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 247, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 248, "usage_type": "call"}, {"api_name": "mth5.helpers.validate_name", "line_number": 287, "usage_type": "call"}, {"api_name": "mt_metadata.timeseries.Station", "line_number": 293, "usage_type": "call"}, {"api_name": "mt_metadata.timeseries", "line_number": 293, "usage_type": "name"}, {"api_name": "mth5.helpers.validate_name", "line_number": 295, "usage_type": "call"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 301, "usage_type": "call"}, {"api_name": "mth5.helpers.validate_name", "line_number": 340, "usage_type": "call"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 351, "usage_type": "call"}, {"api_name": "mth5.helpers.validate_name", "line_number": 378, "usage_type": "call"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 393, "usage_type": "call"}, {"api_name": "mth5.groups.base.BaseGroup", "line_number": 399, "usage_type": "name"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 607, "usage_type": "name"}, {"api_name": "mth5.groups.base.BaseGroup.metadata", "line_number": 596, "usage_type": "attribute"}, {"api_name": "mth5.groups.base.BaseGroup", "line_number": 596, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 653, "usage_type": "call"}, {"api_name": "numpy.dtype", "line_number": 655, "usage_type": "call"}, {"api_name": "h5py.ref_dtype", "line_number": 663, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 668, "usage_type": "call"}, {"api_name": "mt_metadata.utils.mttime.MTime", "line_number": 717, "usage_type": "argument"}, {"api_name": "mt_metadata.utils.mttime.MTime", "line_number": 718, "usage_type": "call"}, {"api_name": "mth5.helpers.validate_name", "line_number": 746, "usage_type": "call"}, {"api_name": "mt_metadata.timeseries.Run", "line_number": 750, "usage_type": "call"}, {"api_name": "mt_metadata.timeseries", "line_number": 750, "usage_type": "name"}, {"api_name": "mth5.helpers.validate_name", "line_number": 751, "usage_type": "call"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 754, "usage_type": "call"}, {"api_name": "mth5.helpers.validate_name", "line_number": 778, "usage_type": "call"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 787, "usage_type": "call"}, {"api_name": "mth5.helpers.validate_name", "line_number": 814, "usage_type": "call"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 829, "usage_type": "call"}, {"api_name": "mth5.groups.base.BaseGroup", "line_number": 853, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 893, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 896, "usage_type": "call"}, {"api_name": "mth5.helpers.validate_name", "line_number": 919, "usage_type": "call"}, {"api_name": "mth5.groups.TransferFunctionGroup", "line_number": 921, "usage_type": "call"}, {"api_name": "mth5.helpers.validate_name", "line_number": 947, "usage_type": "call"}, {"api_name": "mth5.groups.TransferFunctionGroup", "line_number": 949, "usage_type": "call"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 958, "usage_type": "call"}, {"api_name": "mth5.helpers.validate_name", "line_number": 978, "usage_type": "call"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 993, "usage_type": "call"}, {"api_name": "mth5.groups.base.BaseGroup", "line_number": 1025, "usage_type": "name"}, {"api_name": "numpy.dtype", "line_number": 1190, "usage_type": "call"}, {"api_name": "h5py.ref_dtype", "line_number": 1198, "usage_type": "attribute"}, {"api_name": "mth5.helpers.from_numpy_type", "line_number": 1214, "usage_type": "call"}, {"api_name": "mt_metadata.timeseries.Station", "line_number": 1215, "usage_type": "call"}, {"api_name": "mt_metadata.timeseries", "line_number": 1215, "usage_type": "name"}, {"api_name": "mth5.helpers.from_numpy_type", "line_number": 1230, "usage_type": "call"}, {"api_name": "mt_metadata.timeseries.Survey", "line_number": 1231, "usage_type": "call"}, {"api_name": "mt_metadata.timeseries", "line_number": 1231, "usage_type": "name"}, {"api_name": "mth5.groups.base.BaseGroup.metadata", "line_number": 1235, "usage_type": "attribute"}, {"api_name": "mth5.groups.base.BaseGroup", "line_number": 1235, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 1275, "usage_type": "call"}, {"api_name": "numpy.dtype", "line_number": 1277, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 1282, "usage_type": "attribute"}, {"api_name": "numpy.int64", "line_number": 1283, "usage_type": "attribute"}, {"api_name": "h5py.ref_dtype", "line_number": 1286, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 1291, "usage_type": "call"}, {"api_name": "mth5.helpers.to_numpy_type", "line_number": 1301, "usage_type": "call"}, {"api_name": "mth5.helpers.validate_name", "line_number": 1347, "usage_type": "call"}, {"api_name": "mth5.CHUNK_SIZE", "line_number": 1389, "usage_type": "name"}, {"api_name": "mth5.CHUNK_SIZE", "line_number": 1392, "usage_type": "name"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 1430, "usage_type": "call"}, {"api_name": "mth5.helpers.validate_name", "line_number": 1489, "usage_type": "call"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 1498, "usage_type": "call"}, {"api_name": "mth5.helpers.validate_name", "line_number": 1554, "usage_type": "call"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 1570, "usage_type": "call"}, {"api_name": "mth5.timeseries.RunTS", "line_number": 1592, "usage_type": "call"}, {"api_name": "mth5.timeseries.RunTS", "line_number": 1610, "usage_type": "argument"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 1613, "usage_type": "call"}, {"api_name": "mth5.timeseries.ChannelTS", "line_number": 1653, "usage_type": "argument"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 1656, "usage_type": "call"}, {"api_name": "mth5.groups.FiltersGroup", "line_number": 1662, "usage_type": "call"}, {"api_name": "h5py.Dataset", "line_number": 1790, "usage_type": "attribute"}, {"api_name": "weakref.ref", "line_number": 1791, "usage_type": "call"}, {"api_name": "mth5.utils.mth5_logger.setup_logger", "line_number": 1792, "usage_type": "call"}, {"api_name": "mt_metadata.base.Base", "line_number": 1797, "usage_type": "call"}, {"api_name": "mt_metadata.base.Base", "line_number": 1801, "usage_type": "call"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 1820, "usage_type": "call"}, {"api_name": "{'FiltersGroup': 'mth5.groups.FiltersGroup'}", "line_number": 1909, "usage_type": "call"}, {"api_name": "mth5.helpers.from_numpy_type", "line_number": 1917, "usage_type": "call"}, {"api_name": "mt_metadata.timeseries.Run", "line_number": 1918, "usage_type": "call"}, {"api_name": "mt_metadata.timeseries", "line_number": 1918, "usage_type": "name"}, {"api_name": "mth5.helpers.from_numpy_type", "line_number": 1934, "usage_type": "call"}, {"api_name": "mt_metadata.timeseries.Station", "line_number": 1935, "usage_type": "call"}, {"api_name": "mt_metadata.timeseries", "line_number": 1935, "usage_type": "name"}, {"api_name": "mth5.helpers.from_numpy_type", "line_number": 1951, "usage_type": "call"}, {"api_name": "mt_metadata.timeseries.Survey", "line_number": 1952, "usage_type": "call"}, {"api_name": "mt_metadata.timeseries", "line_number": 1952, "usage_type": "name"}, {"api_name": "mth5.groups.FiltersGroup", "line_number": 1965, "usage_type": "call"}, {"api_name": "mt_metadata.timeseries.filters.ChannelResponseFilter", "line_number": 1976, "usage_type": "call"}, {"api_name": "mt_metadata.utils.mttime.MTime", "line_number": 1985, "usage_type": "argument"}, {"api_name": "mth5.timeseries.channel_ts.make_dt_coordinates", "line_number": 2019, "usage_type": "call"}, {"api_name": "mth5.helpers.from_numpy_type", "line_number": 2031, "usage_type": "call"}, {"api_name": "mth5.helpers.to_numpy_type", "line_number": 2042, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 2053, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 2055, "usage_type": "call"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 2147, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 2148, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 2150, "usage_type": "call"}, {"api_name": "mt_metadata.utils.mttime.MTime", "line_number": 2155, "usage_type": "argument"}, {"api_name": "mt_metadata.utils.mttime.MTime", "line_number": 2156, "usage_type": "call"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 2186, "usage_type": "call"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 2195, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 2221, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2222, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 2230, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2231, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 2233, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 2234, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 2239, "usage_type": "attribute"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 2245, "usage_type": "call"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 2291, "usage_type": "call"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 2300, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 2317, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2318, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 2321, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 2328, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2329, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 2331, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 2332, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 2339, "usage_type": "attribute"}, {"api_name": "mth5.utils.exceptions.MTH5Error", "line_number": 2345, "usage_type": "call"}, {"api_name": "mth5.timeseries.ChannelTS", "line_number": 2394, "usage_type": "call"}, {"api_name": "xarray.DataArray", "line_number": 2415, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 2433, "usage_type": "call"}, {"api_name": "numpy.core.records.fromarrays", "line_number": 2451, "usage_type": "call"}, {"api_name": "numpy.core", "line_number": 2451, "usage_type": "attribute"}, {"api_name": "mth5.timeseries.ChannelTS", "line_number": 2501, "usage_type": "argument"}, {"api_name": "xarray.DataArray", "line_number": 2569, "usage_type": "attribute"}, {"api_name": "mt_metadata.utils.mttime.MTime", "line_number": 2605, "usage_type": "argument"}, {"api_name": "mt_metadata.utils.mttime.MTime", "line_number": 2606, "usage_type": "call"}, {"api_name": "mt_metadata.utils.mttime.MTime", "line_number": 2629, "usage_type": "argument"}, {"api_name": "mt_metadata.utils.mttime.MTime", "line_number": 2630, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 2643, "usage_type": "call"}, {"api_name": "numpy.dtype", "line_number": 2655, "usage_type": "call"}, {"api_name": "h5py.ref_dtype", "line_number": 2663, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 2674, "usage_type": "call"}, {"api_name": "mth5.CHANNEL_DTYPE", "line_number": 2701, "usage_type": "name"}, {"api_name": "mt_metadata.utils.mttime.MTime", "line_number": 2768, "usage_type": "argument"}, {"api_name": "mt_metadata.utils.mttime.MTime", "line_number": 2769, "usage_type": "call"}, {"api_name": "mt_metadata.utils.mttime.MTime", "line_number": 2771, "usage_type": "argument"}, {"api_name": "mt_metadata.utils.mttime.MTime", "line_number": 2772, "usage_type": "call"}, {"api_name": "mth5.timeseries.channel_ts.make_dt_coordinates", "line_number": 2811, "usage_type": "call"}, {"api_name": "xarray.DataArray", "line_number": 2822, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 2827, "usage_type": "call"}, {"api_name": "mth5.timeseries.ChannelTS", "line_number": 2834, "usage_type": "call"}, {"api_name": "mt_metadata.utils.mttime.MTime", "line_number": 2857, "usage_type": "argument"}, {"api_name": "mt_metadata.utils.mttime.MTime", "line_number": 2858, "usage_type": "call"}, {"api_name": "mth5.helpers.inherit_doc_string", "line_number": 2866, "usage_type": "name"}, {"api_name": "mth5.helpers.inherit_doc_string", "line_number": 2873, "usage_type": "name"}, {"api_name": "mth5.helpers.inherit_doc_string", "line_number": 2880, "usage_type": "name"}]}
+{"seq_id": "32743149", "text": "import numpy as np\nfrom pathlib import Path\n\nfrom opdxread import opdxtype\n\nfrom typing import Any, Dict\n\n\nclass OPDxFile(object):\n def __init__(self, path: Path):\n self.path = path\n self.filesize = path.stat().st_size\n\n self.data: Dict[str, Any] = {}\n\n self.read()\n\n def read(self):\n with self.path.open(\"rb\") as fp:\n assert fp.read(12) == b\"VCA DATA\\x01\\x00\\x00\\x55\"\n\n while fp.tell() < self.filesize:\n item = opdxtype.NamedValue(fp)\n if item.value is not None:\n self.data[item.name] = item.value\n\n def get_1d_linear_fit(self, r: float = None, m: float = None) -> np.ndarray:\n extent = self.data[\"1D_Data\"][\"Raw\"][\"Extent\"].value\n x = self.data[\"1D_Data\"][\"Raw\"][\"PositionFunction\"].data\n y = self.data[\"1D_Data\"][\"Raw\"][\"Array\"].array\n ydiv = extent / y.size\n\n r = 0.0 if r is None else (r if r >= 0.0 else extent + r)\n m = extent if m is None else (m if m >= 0.0 else extent + m)\n assert m > r\n\n ir = np.clip(int(r / ydiv), 0, y.size - 1)\n im = np.clip(int(m / ydiv), 0, y.size - 1)\n\n coefs = np.polynomial.polynomial.polyfit([r, m], [y[ir], y[im]], 1)\n return np.polynomial.polynomial.polyval(x, coefs)\n\n def get_1d_data(self, r: float = None, m: float = None) -> np.ndarray:\n scale = self.data[\"1D_Data\"][\"Raw\"][\"DataScale\"].value\n x = self.data[\"1D_Data\"][\"Raw\"][\"PositionFunction\"].data\n y = self.data[\"1D_Data\"][\"Raw\"][\"Array\"].array.copy()\n if r is not None or m is not None:\n y -= self.get_1d_linear_fit(r, m)\n return np.stack((x, y * scale), axis=1)\n", "sub_path": "opdxread/opdxfile.py", "file_name": "opdxfile.py", "file_ext": "py", "file_size_in_byte": 1703, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pathlib.Path", "line_number": 10, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 14, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 14, "usage_type": "name"}, {"api_name": "opdxread.opdxtype.NamedValue", "line_number": 23, "usage_type": "call"}, {"api_name": "opdxread.opdxtype", "line_number": 23, "usage_type": "name"}, {"api_name": "numpy.clip", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.polynomial.polynomial.polyfit", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.polynomial", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.polynomial.polynomial.polyval", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.polynomial", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 27, "usage_type": "attribute"}, {"api_name": "numpy.stack", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 43, "usage_type": "attribute"}]}
+{"seq_id": "432207922", "text": "import json\n\nimport pymongo\n\nclient = pymongo.MongoClient('mongodb://comp9900:z12345@ds161529.mlab.com:61529/comp9900_2019')\ndb = client[\"comp9900_2019\"]\n\ncustomers_collection = db.customers_collection\nproperty_collection = db.property_collection\norder_collection = db.order_collection\n\ndef update_data(collection_name,data):\n collection_name.insert_one(data)\n return \"successful\"\n'''''\nwith open(\"orders.json\", 'r') as ff:\n context = ff.read()\n json_data = json.loads(context)\nprint(json_data)\nkeys_set = json_data.keys()\nprint(keys_set)\nfor i in keys_set:\n data = json_data[i]\n print(data)\n update_data(order_collection,data)\n\n\ndict_use = {\"average_mark\": 4.5,\n\t\t\t\"cleanliness_mark\": 4.5,\n\t\t\t\"facility_mark\": 3,\n\t\t\t\"attitude_mark\": 4,\n\t\t\t\"text\": \"I think this house not very good for me, because the hot water is not available 24 hours.\",\n\t\t\t\"photo\": None,\n\t\t\t\"date\": \"09-06-2019\"}\n\n\ndatt = property_collection.find_one({\"property_id\": '5'})\n#property_collection.update({\"property_id\": str(4)},{\"$set\": {\"comments\": dict_use}})\ncomment = datt[\"comments\"]\ncomment.append(dict_use)\nproperty_collection.update_one({\"property_id\": str(5)},{\"$set\": {\"comments\": comment}})\nprint(comment)\n'''''\n#url = \"https://maps.google.com/maps/api/geocode/json?key=AIzaSyAANyBQ6ikIoa53iMdahFL99Bjt0oBmWpc&address={address}&sensor=false\".format(address=ad)\n#data = requests.request(\"GET\",url)\n#data.json()['results'][0]['geometry']['location']", "sub_path": "app/demo/anhao0522_client_v1/api/MongoDB.py", "file_name": "MongoDB.py", "file_ext": "py", "file_size_in_byte": 1447, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pymongo.MongoClient", "line_number": 5, "usage_type": "call"}]}
+{"seq_id": "164417865", "text": "#imports\n# from Loyal_Fox.codes.mailing import mail_company, mail_viwer\nimport os\nfrom datetime import datetime\nfrom enum import unique\nfrom logging import debug\n\nfrom flask import Flask, flash, redirect, render_template, request, make_response\nfrom flask_login import (LoginManager, UserMixin, current_user, login_required, login_user, logout_user)\nfrom flask_mail import Mail, Message\nfrom flask_sqlalchemy import SQLAlchemy\nfrom werkzeug.utils import secure_filename\n\n# from mailing import *\n\n\n#global\n# frontend_url = \"http://localhost:3000\"\nALLOWED_EXTENSIONS = set(['png', 'jpg', 'jpeg', 'gif'])\nUPLOAD_FOLDER = \"./static/blogs_images\"\n\n\n#configuring\napp = Flask(__name__)\napp.config[\"SECRET_KEY\"]= \"KEY\"\napp.config[\"SQLALCHEMY_DATABASE_URI\"] = \"sqlite:///blogs.db\"\n# app.config[\"SQLALCHEMY_DATABASE_URI\"] = \"sqlite:///User.db\"\n\napp.config[\"SQLALCHEMY_TRACK_MODIFICATIONS\"] = True\n\ndb = SQLAlchemy(app)\n\napp.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER\napp.config['MAX_CONTENT_LENGTH'] = 16 * 1024 * 1024\n\nlogin_manager = LoginManager()\nlogin_manager.init_app(app)\n\napp.config.update(\n DEBUG=True,\n \n # for actual code running on AWS\n MAIL_PORT = 587,\n MAIL_USE_TLS = True,\n \n MAIL_SERVER = \"email-smtp.ap-south-1.amazonaws.com\",\n MAIL_USERNAME = \"AKIAZXXKXWLK4PLL4OHQ\",\n MAIL_PASSWORD = \"BDl2EYN45O31YT2+n2o1P+RQ8A7dtn6i0GZFXWVqldQB\"\n)\nmail = Mail(app)\n\n\n#creating flask object models or database table\nclass blogs_table(db.Model):\n id = db.Column(\"Id\",db.Integer, primary_key=True)\n timestamp = db.Column(\"Timestamp\",db.DateTime,default=datetime.now(), nullable=False)\n title = db.Column(\"Title\",db.String(200),nullable=False)\n abstract = db.Column(\"Abstract\",db.Text)\n # author = db.Column(db.String(20))\n image = db.Column(\"Image\",db.String(500))\n content = db.Column(\"Content\",db.Text,nullable=False)\n\n\nclass User(UserMixin, db.Model):\n id = db.Column(\"ID\",db.Integer, primary_key=True)\n username = db.Column(\"Username\",db.String(100),unique=True, nullable=False)\n password = db.Column(\"Password\",db.String(100), nullable=False)\n\n\n#functions\ndef allowed_file(filename):\n return '.' in filename and filename.rsplit('.', 1)[1].lower() in ALLOWED_EXTENSIONS\n\n\n\n# login manager\n@login_manager.user_loader\ndef load_user(user_id):\n return User.query.get(int(user_id))\n\n\n#APIs routes\n\n#Home\n@app.route(\"/\")\n# @app.route(\"/home/\")\n@app.route(\"/home\")\ndef home():\n return render_template(\"home.html\")\n\n# -----------------------------------------------\n\n#About US\n# @app.route(\"/who_we_are/\")\n@app.route(\"/who_we_are\")\ndef who_we_are():\n return render_template(\"who_we_are.html\")\n\n# @app.route(\"/what_we_do/\")\n@app.route(\"/what_we_do\")\ndef what_we_do():\n return render_template(\"what_we_do.html\")\n\n# @app.route(\"/our_mission/\")\n@app.route(\"/our_mission\")\ndef our_mission():\n return render_template(\"our_mission.html\")\n\n# ------------------------------------------------------\n\n#Our Solutions\n# @app.route(\"/multi-tenant-loyalty/\")\n@app.route(\"/multi-tenant-loyalty\")\ndef multi_tenant_loyalty():\n return render_template(\"multi-tenant-loyalty.html\")\n\n# @app.route(\"/marketing-campaigns/\")\n@app.route(\"/marketing-campaigns\")\ndef marketing_campaigns():\n return render_template(\"marketing-campaigns.html\")\n\n# @app.route(\"/helpdesk-services/\")\n@app.route(\"/helpdesk-services\")\ndef helpdesk_services():\n return render_template(\"helpdesk-services.html\")\n\n# @app.route(\"/data-analytics/\")\n@app.route(\"/data-analytics\")\ndef data_analytics():\n return render_template(\"data-analytics.html\")\n\n# @app.route(\"/instant-rewards/\")\n@app.route(\"/instant-rewards\")\ndef instant_rewards():\n return render_template(\"instant-rewards-new.html\")\n #return redirect(\"/404\")\n\n# @app.route(\"/content-and-creatives/\")\n@app.route(\"/content-and-creatives\")\ndef content_and_creatives():\n return render_template(\"content-and-creatives.html\")\n\n# @app.route(\"/robust-tech/\")\n@app.route(\"/robust-tech\")\ndef robust_tech():\n return render_template(\"robust-tech.html\")\n\n\n\n\n\n#----------------------------------------------------\n\n# @app.route(\"/ourteam/\")\n@app.route(\"/ourteam\")\ndef ourteam():\n return render_template(\"ourteam.html\")\n\n# @app.route(\"/navbar/\")\n@app.route(\"/navbar\")\ndef navbar():\n return render_template(\"navbar.html\")\n \n#--------------------------------------------------------\n\n\n# @app.route(\"/productfulfillment/\")\n@app.route(\"/productfulfillment\")\ndef product():\n return render_template(\"product.html\")\n\n@app.route(\"/admin\")\ndef admin():\n return render_template(\"admin_login.html\")\n\n\n# @app.route(\"/experiences/\")\n@app.route(\"/experiences\")\ndef experiences():\n return render_template(\"experiences.html\")\n #return redirect(\"/404\")\n\n\n#--------------------------------------------------------\n\n# @app.route(\"/404/\")\n@app.route(\"/404\")\ndef func_404():\n return render_template(\"404.html\")\n\n@app.errorhandler(404)\ndef not_found(e):\n return redirect(\"/404\")\n\n\n#--------------------------------------------------------\n\n# @app.route(\"/rewards/\", methods=[\"GET\",\"POST\"])\n@app.route(\"/rewards\", methods=[\"GET\",\"POST\"])\ndef rewards():\n if request.method == \"POST\":\n # try\n name = request.form.get(\"user_name\")\n # name = request.form[\"name\"]\n phone = request.form.get(\"phone\")\n email = request.form.get(\"email\")\n note = request.form.get(\"note\")\n \n # mail_msg = \"\"\"\n # Response from 'Apply for Gift Vouchers': \\n\n # Name : \"\"\" + str(name) + \"\"\"\n # Phone : \"\"\" + str(phone) + \"\"\" \n # Email ID : \"\"\" + str(email) + \"\"\" \n # Any Note : \"\"\" + str(note) + \"\"\" \\n\n # \"\"\"\n # print(mail_msg)\n\n\n # if no entry in the form\n # if(not name or not phone or not email or not note):\n # err_msg = \"Give all the inputs\"\n # print(err_msg)\n # flash(err_msg,\"error\")\n # return redirect(\"/rewards\")\n\n #check whether phoneNO is 10 digits\n # if(len(str(phone))!=10 or str(phone)[0]=='0'):\n # err_msg = \"Enter 10 digits Phone no.\"\n # print(err_msg)\n # flash(err_msg,\"error\")\n # return redirect(\"/rewards\")\n\n\n mail_msg = \"\"\"\n Response from 'Apply for Gift Vouchers': \\n\n Name : \"\"\" + str(name) + \"\"\"\n Phone : \"\"\" + str(phone) + \"\"\" \n Email ID : \"\"\" + str(email) + \"\"\" \n Any Note : \"\"\" + str(note) + \"\"\" \\n\n \"\"\"\n\n print(mail_msg)\n\n print(\"\\n Mailing..........\")\n msg_company = Message(\n # sender = (\"Test\",\"testing0963@gmail.com\"),\n sender = (\"Website\",\"info@loyaltyfox.com\"),\n recipients=[\"info@loyaltyfox.com\",\"testing0963@gmail.com\"],\n body = mail_msg\n # subject = \"ContactUS form\"\n )\n # msg.body = \"Hey avi\"\n msg_company.subject = \"Loyalty Fox Website | Apply For Gift Vouchers Response\"\n mail.send(msg_company)\n\n # mail_company(mail_msg)\n print(\"mailed to loyality fox..\")\n \n\n #mail the the person whom want to be contacted\n msg_viewer = Message(\n sender = (\"Website\",\"info@loyaltyfox.com\"),\n recipients=[email],\n body = mail_msg\n # subject = \"ContactUS form\"\n )\n # msg.body = \"Hey avi\"\n msg_viewer.subject = \"Loyalty Fox | Apply For Gift Vouchers Response\"\n # mail.send(msg_viewer)\n\n # mail_viwer(email,mail_msg)\n print(\"mailed to user.\")\n flash(\"Thank you for your interest, we'll contact you soon. For any other query, you can call us on 8802065822 or write to us on info@loyaltyfox.com. \",\"success\")\n \n return render_template(\"rewards.html\")\n # return redirect(frontend_url)\n\n # except Exception as e:\n # print(e)\n # err_msg = \"Error in contactUS: \" + str(e)\n # print(err_msg)\n # flash(err_msg,\"error\")\n # return redirect(\"/rewards\")\n\n\n\n#contact us API\n# @app.route(\"/contactUS/\", methods=[\"GET\",\"POST\"])\n@app.route(\"/contactUS\", methods=[\"GET\",\"POST\"])\ndef contact():\n if request.method == \"POST\":\n # try\n name = request.form.get(\"user_name\")\n # name = request.form[\"name\"]\n phone = request.form.get(\"phone\")\n email = request.form.get(\"email\")\n company = request.form.get(\"company\")\n note = request.form.get(\"note\")\n \n # mail_msg = \"\"\"\n # Response from Contact Us: \\n\n # Name : \"\"\" + str(name) + \"\"\"\n # Phone : \"\"\" + str(phone) + \"\"\" \n # Email ID : \"\"\" + str(email) + \"\"\"\n # Company : \"\"\" + str(company) + \"\"\" \n # Any Note : \"\"\" + str(note) + \"\"\" \\n\n # \"\"\"\n # print(mail_msg)\n\n\n # if no entry in the form\n # if(not name or not phone or not email or not note):\n # err_msg = \"Give all the inputs\"\n # print(err_msg)\n # flash(err_msg,\"error\")\n # return redirect(\"/contactUS\")\n\n #check whether phoneNO is 10 digits\n # if(len(str(phone))!=10 or str(phone)[0]=='0'):\n # err_msg = \"Enter 10 digits Phone no.\"\n # print(err_msg)\n # flash(err_msg,\"error\")\n # return redirect(\"/contactUS\")\n\n\n mail_msg = \"\"\"\n Response from Contact Us: \\n\n Name : \"\"\" + str(name) + \"\"\"\n Phone : \"\"\" + str(phone) + \"\"\" \n Email ID : \"\"\" + str(email) + \"\"\"\n Company : \"\"\" + str(company) + \"\"\" \n Any Note : \"\"\" + str(note) + \"\"\" \\n\n \"\"\"\n\n print(mail_msg)\n\n print(\"\\n Mailing..........\")\n msg_company = Message(\n # sender = (\"Test\",\"testing0963@gmail.com\"),\n sender = (\"Website\",\"info@loyaltyfox.com\"),\n recipients=[\"info@loyaltyfox.com\",\"testing0963@gmail.com\"],\n body = mail_msg\n # subject = \"ContactUS form\"\n )\n # msg.body = \"Hey avi\"\n msg_company.subject = \"Loyalty Fox Website | Contact US Response\"\n mail.send(msg_company)\n\n # mail_company(mail_msg)\n print(\"mailed to loyality fox..\")\n \n\n #mail the the person whom want to be contacted\n msg_viewer = Message(\n sender = \"testing0963@gmail.com\",\n recipients=[email],\n body = mail_msg\n # subject = \"ContactUS form\"\n )\n # msg.body = \"Hey avi\"\n msg_viewer.subject = \"Loyalty Fox | Contact US Response\"\n # mail.send(msg_viewer)\n\n # mail_viwer(email,mail_msg)\n print(\"mailed to user.\")\n flash(\"Thank you for your interest, we'll contact you soon. For any other query, you can call us on 8802065822 or write to us on info@loyaltyfox.com. \",\"success\")\n \n return render_template(\"contactus.html\")\n # return redirect(frontend_url)\n\n # except Exception as e:\n # print(e)\n # err_msg = \"Error in contactUS: \" + str(e)\n # print(err_msg)\n # flash(err_msg,\"error\")\n # return redirect(\"/contactUS\")\n \n# ---------------------------------------------------------------------------------\n\n\n#ppc-channel landing page API\n@app.route(\"/ppc-channel-loyalty-program/\", methods=[\"GET\",\"POST\"])\n@app.route(\"/ppc-channel-loyalty-program\", methods=[\"GET\",\"POST\"])\ndef ppc():\n if request.method == \"POST\":\n # try\n name = request.form.get(\"name\")\n # name = request.form[\"name\"]\n phone = request.form.get(\"phone\")\n email = request.form.get(\"email\")\n note = request.form.get(\"message\")\n \n # mail_msg = \"\"\"\n # Response from Contact Us: \\n\n # Name : \"\"\" + str(name) + \"\"\"\n # Phone : \"\"\" + str(phone) + \"\"\" \n # Email ID : \"\"\" + str(email) + \"\"\"\n # Company : \"\"\" + str(company) + \"\"\" \n # Any Note : \"\"\" + str(note) + \"\"\" \\n\n # \"\"\"\n # print(mail_msg)\n\n\n # if no entry in the form\n # if(not name or not phone or not email or not note):\n # err_msg = \"Give all the inputs\"\n # print(err_msg)\n # flash(err_msg,\"error\")\n # return redirect(\"/contactUS\")\n\n #check whether phoneNO is 10 digits\n # if(len(str(phone))!=10 or str(phone)[0]=='0'):\n # err_msg = \"Enter 10 digits Phone no.\"\n # print(err_msg)\n # flash(err_msg,\"error\")\n # return redirect(\"/contactUS\")\n\n\n mail_msg = \"\"\"\n Response from PPC-channel-loyalty-program landing page: \\n\n Name : \"\"\" + str(name) + \"\"\"\n Phone : \"\"\" + str(phone) + \"\"\" \n Email ID : \"\"\" + str(email) + \"\"\" \n Any Note : \"\"\" + str(note) + \"\"\" \\n\n \"\"\"\n\n print(mail_msg)\n\n print(\"\\n Mailing..........\")\n msg_company = Message(\n # sender = (\"Test\",\"testing0963@gmail.com\"),\n sender = (\"Website\",\"info@loyaltyfox.com\"),\n recipients=[\"info@loyaltyfox.com\",\"testing0963@gmail.com\"],\n body = mail_msg\n # subject = \"ContactUS form\"\n )\n # msg.body = \"Hey avi\"\n msg_company.subject = \"Loyalty Fox Website | ppc-channel-loyalty-program Response\"\n mail.send(msg_company)\n\n # mail_company(mail_msg)\n print(\"mailed to loyality fox..\")\n \n\n #mail the the person whom want to be contacted\n msg_viewer = Message(\n sender = (\"Website\",\"info@loyaltyfox.com\"),\n recipients=[email],\n body = mail_msg\n # subject = \"ContactUS form\"\n )\n # msg.body = \"Hey avi\"\n msg_viewer.subject = \"Loyalty Fox | ppc-channel-loyalty-program Response\"\n # mail.send(msg_viewer)\n\n # mail_viwer(email,mail_msg)\n print(\"mailed to user.\")\n flash(\"Thank you for your interest, we'll contact you soon. For any other query, you can call us on 8802065822 or write to us on info@loyaltyfox.com. \",\"success\")\n \n return render_template(\"ppc-channel.html\")\n\n\n@app.route(\"/gift-vouchers/\", methods=[\"GET\",\"POST\"])\n@app.route(\"/gift-vouchers\", methods=[\"GET\",\"POST\"])\ndef gifts():\n if request.method == \"POST\":\n # try\n name = request.form.get(\"name\")\n # name = request.form[\"name\"]\n phone = request.form.get(\"phone\")\n email = request.form.get(\"email\")\n note = request.form.get(\"message\")\n \n # mail_msg = \"\"\"\n # Response from Contact Us: \\n\n # Name : \"\"\" + str(name) + \"\"\"\n # Phone : \"\"\" + str(phone) + \"\"\" \n # Email ID : \"\"\" + str(email) + \"\"\"\n # Company : \"\"\" + str(company) + \"\"\" \n # Any Note : \"\"\" + str(note) + \"\"\" \\n\n # \"\"\"\n # print(mail_msg)\n\n\n # if no entry in the form\n # if(not name or not phone or not email or not note):\n # err_msg = \"Give all the inputs\"\n # print(err_msg)\n # flash(err_msg,\"error\")\n # return redirect(\"/contactUS\")\n\n #check whether phoneNO is 10 digits\n # if(len(str(phone))!=10 or str(phone)[0]=='0'):\n # err_msg = \"Enter 10 digits Phone no.\"\n # print(err_msg)\n # flash(err_msg,\"error\")\n # return redirect(\"/contactUS\")\n\n\n mail_msg = \"\"\"\n Response from Gift-vouchers landing page: \\n\n Name : \"\"\" + str(name) + \"\"\"\n Phone : \"\"\" + str(phone) + \"\"\" \n Email ID : \"\"\" + str(email) + \"\"\" \n Any Note : \"\"\" + str(note) + \"\"\" \\n\n \"\"\"\n\n print(mail_msg)\n\n print(\"\\n Mailing..........\")\n msg_company = Message(\n # sender = (\"Test\",\"testing0963@gmail.com\"),\n sender = (\"Website\",\"info@loyaltyfox.com\"),\n recipients=[\"info@loyaltyfox.com\",\"testing0963@gmail.com\"],\n body = mail_msg\n # subject = \"ContactUS form\"\n )\n # msg.body = \"Hey avi, get lost\"\n msg_company.subject = \"Loyalty Fox Website | gift-vouchers page Response\"\n mail.send(msg_company)\n\n # mail_company(mail_msg)\n print(\"mailed to loyality fox..\")\n \n\n #mail the the person whom want to be contacted\n msg_viewer = Message(\n sender = (\"Website\",\"info@loyaltyfox.com\"),\n recipients=[email],\n body = mail_msg\n # subject = \"ContactUS form\"\n )\n # msg.body = \"Hey avi, get lost\"\n msg_viewer.subject = \"Loyalty Fox | gift-vouchers Response\"\n # mail.send(msg_viewer)\n\n # mail_viwer(email,mail_msg)\n print(\"mailed to user.\")\n flash(\"Thank you for your interest, we'll contact you soon. For any other query, you can call us on 8802065822 or write to us on info@loyaltyfox.com. \",\"success\")\n \n return render_template(\"gift-vouchers.html\")\n\n\n# --------------------------------------------------------------------------------\n\n# Blog page\n# @app.route(\"/add_blog/\", methods=[\"GET\",\"POST\"])\n@app.route(\"/add_blog\", methods=[\"GET\",\"POST\"])\n@login_required\ndef add_blog():\n # try:\n\n if(request.method == \"POST\"):\n title = request.form.get(\"title\")\n abstract = request.form.get(\"abstract\")\n file = request.files[\"image_file\"]\n content = request.form.get(\"content\")\n\n filename = secure_filename(file.filename)\n\n if not content:\n err_msg = \"Content/Main body field empty !\"\n flash(err_msg,\"error\")\n return redirect(\"/add_blog\")\n\n if(file and len(filename)>100):\n flash(\"Please upload file with shorter filename\",\"error\")\n return redirect(\"/add_blog\")\n\n imagename = str(filename) + \"_--_\" + str(title)\n imagename_secure = secure_filename(imagename)\n print(imagename_secure)\n\n \n if file:\n if allowed_file(file.filename):\n file.save(os.path.join(app.config['UPLOAD_FOLDER'], imagename_secure))\n\n # file_path = UPLOAD_FOLDER + \"/\" + str(imagename_secure)\n # file.save(file_path)\n\n #adding row in table\n new_blog_entry = blogs_table(title=title,abstract=abstract,content=content, image=imagename_secure)\n\n db.session.add(new_blog_entry)\n db.session.commit()\n\n flash(\"Blog added successfully\",\"success\")\n return redirect(\"/dashboard\")\n\n else:\n flash('Invalid, Upload only png, jpg, jpeg, gif')\n\n return render_template(\"add_blog.html\")\n\n # except Exception as e:\n # print(e)\n # err_msg = \"Error in add_blog: \" + str(e)\n # print(err_msg)\n # flash(err_msg,\"error\")\n # return redirect(\"/add_blog\")\n\n#----------------------------------------------------\n# @app.route(\"/blogs\")\n# @app.route(\"/blogs/\")\n# def blogsNew():\n# return render_template(\"blogs_new.html\")\n\n\n\n\n@app.route(\"/dashboard\", methods=[\"GET\",\"POST\"])\n@login_required\ndef blogs_dashboard():\n # try\n\n all_blogs_list = blogs_table.query.order_by(blogs_table.id.desc()).all()\n print(all_blogs_list)\n\n return render_template(\"blogs_dashboard.html\",all_blogs_list=all_blogs_list)\n\n\n@app.route(\"/delete/\")\n@login_required\ndef blog_delete(blog_id):\n\n image_obj = blogs_table.query.filter_by(id=blog_id).first()\n imagename_secure = image_obj.image\n print(imagename_secure)\n blogs_table.query.filter_by(id=blog_id).delete()\n os.remove(os.path.join(app.config['UPLOAD_FOLDER'], imagename_secure))\n # post.delete()\n db.session.commit()\n\n return redirect(\"/dashboard\")\n\n\n# @app.route(\"/blogs/\", methods=[\"GET\",\"POST\"])\n@app.route(\"/blogs\", methods=[\"GET\",\"POST\"])\ndef blogs():\n # try\n\n all_blogs_list = blogs_table.query.order_by(blogs_table.id.desc()).all()\n print(all_blogs_list)\n\n return render_template(\"blogsx.html\",all_blogs_list=all_blogs_list)\n\n\n# @app.route(\"/blog_post\")\n@app.route(\"/blog_post/\")\ndef blog_post(blog_id):\n post = blogs_table.query.filter_by(id=blog_id).first()\n\n return render_template(\"postx.html\",post=post)\n\n\n@app.route(\"/admin\", methods=[\"GET\",\"POST\"])\ndef admin_login():\n if(request.method==\"POST\"):\n username = request.form.get(\"user_name\")\n password = request.form.get(\"password\")\n\n user_obj = User.query.filter_by(username=username).first()\n \n if user_obj:\n if(username==user_obj.username and password==user_obj.password):\n print(user_obj.username)\n login_user(user_obj)\n print(current_user)\n return redirect(\"/dashboard\")\n \n else:\n err_msg = \"Username/Password Incorrect !\"\n flash(err_msg,\"error\")\n \n else:\n err_msg = \"Username/Password Incorrect !\"\n flash(err_msg,\"error\")\n\n \n return render_template(\"admin_login.html\")\n \n\n#logout API\n@app.route(\"/logout\")\n@login_required\ndef logout():\n id = current_user.id\n user = User.query.filter_by(id=id).first()\n # print(user)\n logout_user()\n flash(\"You are logged out\",\"success\")\n return redirect(\"/admin\")\n\n# --------------------------------------------------------------------------\n\n\n#API for sitemap.xml file\n@app.route(\"/sitemap.xml\")\ndef sitemap():\n template = render_template(\"sitemap.xml\")\n response = make_response(template)\n response.headers['Content-Type'] = 'application/xml'\n\n return response\n # return render_template(\"sitemap.xml\")\n\n\n@app.route(\"/abc.xml\")\ndef abc():\n template = render_template(\"abc.xml\")\n response = make_response(template)\n response.headers['Content-Type'] = 'application/xml'\n\n return response\n\n# --------------------------------------------------------------------------\n\n\nif __name__ == \"__main__\":\n db.create_all()\n # admin = User(username=\"admin\",password=\"admin\")\n # db.session.add(admin)\n # db.session.commit()\n app.run(debug=True, port=5000, host=\"0.0.0.0\")\n\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 22006, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "flask.Flask", "line_number": 24, "usage_type": "call"}, {"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 31, "usage_type": "call"}, {"api_name": "flask_login.LoginManager", "line_number": 36, "usage_type": "call"}, {"api_name": "flask_mail.Mail", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 56, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 56, "usage_type": "name"}, {"api_name": "flask_login.UserMixin", "line_number": 64, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 89, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 97, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 102, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 107, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 115, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 120, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 125, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 130, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 135, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 141, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 146, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 157, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 162, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 170, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 174, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 180, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 189, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 193, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 201, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 201, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 203, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 203, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 203, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 205, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 205, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 205, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 206, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 206, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 206, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 207, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 207, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 207, "usage_type": "name"}, {"api_name": "flask_mail.Message", "line_number": 245, "usage_type": "call"}, {"api_name": "flask_mail.Message", "line_number": 261, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 273, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 275, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 291, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 291, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 293, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 293, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 293, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 295, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 295, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 295, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 296, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 296, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 296, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 297, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 297, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 297, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 298, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 298, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 298, "usage_type": "name"}, {"api_name": "flask_mail.Message", "line_number": 338, "usage_type": "call"}, {"api_name": "flask_mail.Message", "line_number": 354, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 366, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 368, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 385, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 385, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 387, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 387, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 387, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 389, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 389, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 389, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 390, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 390, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 390, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 391, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 391, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 391, "usage_type": "name"}, {"api_name": "flask_mail.Message", "line_number": 430, "usage_type": "call"}, {"api_name": "flask_mail.Message", "line_number": 446, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 458, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 460, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 466, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 466, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 468, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 468, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 468, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 470, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 470, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 470, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 471, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 471, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 471, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 472, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 472, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 472, "usage_type": "name"}, {"api_name": "flask_mail.Message", "line_number": 511, "usage_type": "call"}, {"api_name": "flask_mail.Message", "line_number": 527, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 539, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 541, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 553, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 553, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 554, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 554, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 554, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 555, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 555, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 555, "usage_type": "name"}, {"api_name": "flask.request.files", "line_number": 556, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 556, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 557, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 557, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 557, "usage_type": "name"}, {"api_name": "werkzeug.utils.secure_filename", "line_number": 559, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 563, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 564, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 567, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 568, "usage_type": "call"}, {"api_name": "werkzeug.utils.secure_filename", "line_number": 571, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 577, "usage_type": "call"}, {"api_name": "os.path", "line_number": 577, "usage_type": "attribute"}, {"api_name": "flask.flash", "line_number": 588, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 589, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 592, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 594, "usage_type": "call"}, {"api_name": "flask_login.login_required", "line_number": 549, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 620, "usage_type": "call"}, {"api_name": "flask_login.login_required", "line_number": 613, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 631, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 631, "usage_type": "call"}, {"api_name": "os.path", "line_number": 631, "usage_type": "attribute"}, {"api_name": "flask.redirect", "line_number": 635, "usage_type": "call"}, {"api_name": "flask_login.login_required", "line_number": 624, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 646, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 654, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 659, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 659, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 660, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 660, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 660, "usage_type": "name"}, {"api_name": "flask.request.form.get", "line_number": 661, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 661, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 661, "usage_type": "name"}, {"api_name": "flask_login.login_user", "line_number": 668, "usage_type": "call"}, {"api_name": "flask_login.current_user", "line_number": 669, "usage_type": "argument"}, {"api_name": "flask.redirect", "line_number": 670, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 674, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 678, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 681, "usage_type": "call"}, {"api_name": "flask_login.current_user.id", "line_number": 688, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 688, "usage_type": "name"}, {"api_name": "flask_login.logout_user", "line_number": 691, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 692, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 693, "usage_type": "call"}, {"api_name": "flask_login.login_required", "line_number": 686, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 701, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 702, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 711, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 712, "usage_type": "call"}]}
+{"seq_id": "473652758", "text": "# -*- coding: utf-8 -*-\n\"\"\"\n===============================================================================\nHodge.WaterResources, LLC\nProject Number: AD006\nProject Name: FORTmod\nDeveloped by: Matt Hodge\nType: Program\n\nCreated on Thurs Oct 18 14:00:00 2018\nLast Updated: 10/18/2018\n\nPurpose: \nThe purpose of this program is to demonstrate how to use the .pyd file created\nwith f2py for the Streeter-Phelps model created in FORTRAN. \nNotes:\nThis program includes one application of the Streeter-Phelps model. The \napplication loads the same input file used in the FORTRAN program. The results \nof the model run are saved to a text file (DO_Curve.txt) and saved to a jpg.\nHWR Disclaimer:\nThis script was created by Hodge.WaterResources, LLC (HWR). HWR makes no warranty,\nexpressed or implied, as to its usefulness or correctness.\n===============================================================================\n\"\"\"\n# import necessary modules\nimport pdb\nimport os\nimport sys\nimport matplotlib.pyplot as plt\nimport numpy as np\n\nif __name__ == \"__main__\":\n # import Streeter-Phelps\n os.chdir(r'..\\fortran')\n sys.path.append(os.getcwd())\n import str_phps as sp\n ##--------Run From Input File-----------------\n # input file\n file_path = 'str_phps.inp'\n # load input file\n (rlen, rwid, rq, r_s, man_n, rint, wwq, rtmp, bod_k, rbod0, rdo0, wwbod0, wwdo0) = sp.str_phps_mod.read_inp(file_path)\n # change to python directory\n os.chdir(r'..\\python')\n # run FORTRAN model and return results\n (riv_x, riv_dos, riv_dod) = sp.str_phps_mod.run(rlen, rwid, rq, r_s, man_n, rint, wwq, rtmp, bod_k, rbod0, rdo0, wwbod0, wwdo0)\n riv_do = riv_dos - riv_dod\n # plot DO sag curve & DO sat\n plt.plot(riv_x, riv_do, color = 'b', linewidth = 1)\n plt.plot(riv_x, riv_dos, color = 'b', linestyle = '-.', linewidth = .5)\n plt.xlabel('Distance from WWTP (m)')\n plt.ylabel('DO (mg/L)')\n plt.title('FORTmod DO Sag Curve')\n plt.savefig('str-phps_SagCurve.png')\n plt.show()", "sub_path": "python/str_phps_py_program.py", "file_name": "str_phps_py_program.py", "file_ext": "py", "file_size_in_byte": 2002, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.chdir", "line_number": 34, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 35, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 35, "usage_type": "call"}, {"api_name": "str_phps.str_phps_mod.read_inp", "line_number": 41, "usage_type": "call"}, {"api_name": "str_phps.str_phps_mod", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.chdir", "line_number": 43, "usage_type": "call"}, {"api_name": "str_phps.str_phps_mod.run", "line_number": 45, "usage_type": "call"}, {"api_name": "str_phps.str_phps_mod", "line_number": 45, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}]}
+{"seq_id": "436672645", "text": "import socket,os,time,hashlib\nfrom conf import settings\nfrom modules import display\nfrom modules import log\n\n\nHome = ''\nFlagLogin = False\nUser = ''\nCommand = []\nsk = None\ndef run():\n print('Welcome!')\n print(display.HelpMenu)\n login()\n init()\n while not FlagLogin:\n print(Home)\n print(FlagLogin)\n print(User)\n print(sk)\n try:\n while True:\n command = input(\"[server@%s]:\"%User).strip()\n Command = command.split()\n if Command[0].startswith('rz'): # upload file request\n file_path = \"%s/%s\"%(Home,Command[1])\n if os.path.isfile(file_path):\n sk.send(command.encode())\n data = sk.recv(1024)\n FileSize = os.stat(file_path).st_size\n sk.send(str(FileSize).encode()) # send file size\n m5 = hashlib.md5()\n with open(file_path, 'rb') as f:\n SentSize = 0\n for line in f:\n sk.send(line)\n m5.update(line)\n SentSize += len(line)\n print(\"Transfer:%s%%\"%(100*SentSize/FileSize),end='\\r')\n print('\\rCompletely!')\n # md5 = m5.encode()\n sk.send(m5.hexdigest()) # send md5 value after finish\n else:\n print(\"Invalid file!\")\n elif Command[0].startswith('sz'): #download files,continu when receive 'sz'\n sk.send(command.encode())\n data = sk.recv(1024)\n if data.decode() == 'sz':\n UpName = time.strftime(\"%Y%m%d%H%M%S:\", time.localtime()) + 'upload'\n FileSize = int(sk.recv(1024).decode())\n RecvSize = 0\n m5 = hashlib.md5()\n with open('%s/%s' % (Home, UpName), 'w') as f:\n while RecvSize < FileSize:\n size = FileSize - RecvSize\n if size < 1024:\n data = sk.recv(size)\n else:\n data = sk.recv(1024)\n m5.update(data)\n f.write(data)\n RecvSize += len(data)\n print(\"Transfer:%s%%\" % (100*RecvSize/FileSize), end='\\r')\n else:\n print('\\rCompletely!')\n md5_recv = sk.recv(1024).decode() # receive md5 check\n if md5_recv == m5.hexdigest():\n os.rename(\"%s/%s\" % (Home, UpName),\n \"%s/%s\" % (Home, Command[1])) # rename filename\n print(\"File check passed!\")\n else:\n sk.send(b'file crash!!!')\n os.remove(\"%s/%s\" % (Home, UpName))\n print(\"Remove file because check failed...\")\n else:\n print(\"Invalid filename!\")\n\n elif data == 'logout':\n FlagLogin = False\n break\n print(data)\n except Exception:\n print(\"exist exception!!!\")\n else:\n print(\"offline, use 'login' to connect server!\")\n print(Home)\n print(FlagLogin)\n print(User)\n print(\"Disconnect to server...\")\n\n\ndef login():\n import json\n # global FlagLogin\n # global User\n while True:\n user = input(\"username:\").strip()\n psd = input(\"passwod:\").strip()\n with open('%s/users/users.db', 'r') as f:\n users = json.load(f)\n if user in users:\n if psd == users[user]:\n print(\"login successful!\")\n FlagLogin = True\n User = user\n\n else:\n print(\"Incorrect password!\")\n else:\n print(\"No such user!\")\n\ndef init():\n '''\n init users dir,socket\n :return:\n '''\n # global Home\n sk = socket.socket()\n sk.connect((settings.ADDR,settings.PORT))\n print(\"\\033[1;34;1mConnected to socket!\\033[0m\", )\n path = \"%s/client/%s\"%(settings.DATABASE,User)\n if not os.path.exists(path):\n os.makedirs(path)\n print(\"Created new dir for user!\")\n Home = path\n\n\n\n", "sub_path": "Senior FTP/modules/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 4708, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "modules.display.HelpMenu", "line_number": 14, "usage_type": "attribute"}, {"api_name": "modules.display", "line_number": 14, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.stat", "line_number": 31, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 33, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 50, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 50, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 53, "usage_type": "call"}, {"api_name": "os.rename", "line_number": 69, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 74, "usage_type": "call"}, {"api_name": "json.load", "line_number": 101, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 119, "usage_type": "call"}, {"api_name": "conf.settings.ADDR", "line_number": 120, "usage_type": "attribute"}, {"api_name": "conf.settings", "line_number": 120, "usage_type": "name"}, {"api_name": "conf.settings.PORT", "line_number": 120, "usage_type": "attribute"}, {"api_name": "conf.settings.DATABASE", "line_number": 122, "usage_type": "attribute"}, {"api_name": "conf.settings", "line_number": 122, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path", "line_number": 123, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 124, "usage_type": "call"}]}
+{"seq_id": "98129426", "text": "import requests\n\ndef get_bbquote():\n url = 'https://breaking-bad-quotes.herokuapp.com/v1/quotes'\n response = requests.get(url).json()[0]\n\n return f\"{response['quote']} \\n> {response['author']}\"\n\n\nif __name__ == '__main__':\n print(get_bbquote())", "sub_path": "bbquotes/lib.py", "file_name": "lib.py", "file_ext": "py", "file_size_in_byte": 248, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "requests.get", "line_number": 5, "usage_type": "call"}]}
+{"seq_id": "238103742", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\n# solvers for multidomain with fat\n\nimport sys\nimport numpy as np\nimport subprocess\nimport datetime\nimport time\nfrom cycler import cycler\nimport os\nimport pandas_utility\n\ninput_filename = \"logs/log_resolution1.csv\"\nlist_columns = False\nshow_plots = True\n\n# load matplotlib\nimport matplotlib\nif not show_plots:\n matplotlib.use('Agg')\n\nimport matplotlib.pyplot as plt\nimport pandas as pd\nimport numpy as np\n\n# define global plotting parameters\nmatplotlib.rcdefaults()\nplt.rcParams.update({'font.size': 16})\nplt.rcParams['lines.linewidth'] = 2\n\ndf = pandas_utility.load_df(input_filename)\n\n# filter data\ndf = df.loc[df['nIterations_multidomainLinearSolver'] != 0] # exclude runs where solver diverged (no number of iterations)\n#df = df.loc[df['nIterations_multidomainLinearSolver'] < 1000]\n\ntry:\n df['duration_init'] = df['totalUsertime'] - df['duration_total'] + df['durationParaview3DInit'] + df['durationParaview1DInit']\nexcept:\n df['duration_init'] = 0\n\n# Info about the data structure\n#print(\"df info:\")\n#df.info()\n#print(df.head())\n\n# set options for console display\npd.set_option('display.max_rows', 500)\npd.set_option('display.max_columns', 500)\npd.set_option('display.width', 1000)\npd.set_option('display.max_colwidth', 100)\n\ndef merge_dicts(x, y):\n z = x.copy() # start with x's keys and values\n z.update(y) # modifies z with y's keys and values & returns None\n return z\n\ndef plot(df, items):\n scenario_names = [s for s in df['scenarioName'].unique() if not isinstance(s,float) and not \"preonly_lu\" in s]\n \n print(scenario_names)\n df = df.groupby(['scenarioName', 'multidomainLinearSolver_preconditionerType']).agg(items)\n \n lines_runtime = {}\n lines_n_iterations = {}\n \n print(\"plot df\")\n # collect lines for plot\n for scenario_name in scenario_names:\n \n print(\"scenario_name: {}\".format(scenario_name))\n \n # example for scenario_name:\n # gmres_bjacobi_dt0.001_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse_10mus\n pos = scenario_name.rfind(\"_\")\n \n row_name = scenario_name[0:pos]\n n_mus = (int)(scenario_name[pos+1:scenario_name.find(\"mus\")])\n value = df.loc[scenario_name][\"duration_multidomain\"]\n n_iterations = (float)(df.loc[scenario_name][\"nIterations_multidomainLinearSolver\"])\n \n if row_name not in lines_runtime:\n lines_runtime[row_name] = [np.nan,np.nan,np.nan,np.nan]\n if row_name not in lines_n_iterations:\n lines_n_iterations[row_name] = [np.nan,np.nan,np.nan,np.nan]\n \n #if n_mus == 4: i = 0\n if n_mus == 6: i = 0\n elif n_mus == 8: i = 1\n elif n_mus == 10: i = 2\n elif n_mus == 12: i = 3\n lines_runtime[row_name][i] = value\n \n print(n_iterations)\n if n_iterations != 0:\n lines_n_iterations[row_name][i] = n_iterations\n \n print(\"lines\")\n print(lines_runtime.keys())\n label = {\n 'gmres_boomeramg_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse': \"BoomerAMG\",\n 'gmres_boomeramg_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symTrue_lumpFalse': \"BoomerAMG (symmetric)\",\n 'gmres_euclid_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse': \"Euclid\",\n 'gmres_euclid_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symTrue_lumpFalse': \"Euclid (symmetric)\",\n 'gmres_bjacobi_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse': \"Block Jacobi\",\n 'gmres_bjacobi_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symTrue_lumpFalse': \"Block Jacobi (symmetric)\",\n 'gmres_sor_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse': \"Parallel SOR\",\n 'gmres_sor_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symTrue_lumpFalse': \"Parallel SOR (symmetric)\", \n 'gmres_pilut_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse': \"Pilut\",\n 'gmres_pilut_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symTrue_lumpFalse': \"Pilut (symmetric)\",\n 'gmres_none_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse': \"No preconditioner\",\n 'gmres_none_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symTrue_lumpFalse': \"No preconditioner (symmetric)\",\n 'gmres_sor,ilu_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse': \"Block Gauss-Seidel with ILU\",\n 'gmres_bjacobi,gamg_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse': \"Block Jacobi with AMG\",\n 'gmres_bjacobi,cg-euclid_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse': \"Block Jacobi with Euclid\",\n 'gmres_boomeramg,hypre_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse': 'BoomerAMG w/ coord.',\n 'gmres_boomeramg,hypre_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symTrue_lumpFalse': 'BoomerAMG w/ coord (symmetric)',\n 'gmres_bjacobi,hypre_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse': 'Block Jacobi + AMG',\n }\n \n #order = sorted(lines.iteritems())\n \n order = [\n 'gmres_none_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse',# \"No preconditioner\",\n 'gmres_pilut_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse',# \"Pilut\",\n 'gmres_pilut_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symTrue_lumpFalse',# \"Pilut (symmetric)\",\n 'gmres_boomeramg_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse',# \"BoomerAMG\",\n 'gmres_boomeramg_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symTrue_lumpFalse',# \"BoomerAMG (symmetric)\",\n 'gmres_euclid_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse',# \"Euclid\",\n 'gmres_euclid_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symTrue_lumpFalse',# \"Euclid (symmetric)\",\n 'gmres_bjacobi_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse',# \"Block jacobi\",\n 'gmres_bjacobi_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symTrue_lumpFalse',# \"Block jacobi (symmetric)\",\n 'gmres_sor_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse',# \"Block Gauss-Seidel\",\n 'gmres_sor_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symTrue_lumpFalse',# \"Block Gauss-Seidel (symmetric)\",\n 'gmres_bjacobi,gamg_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse',# \"Block Jacobi with AMG\",\n 'gmres_bjacobi,cg-euclid_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse', #Block Jacobi with Euclid\",\n 'gmres_boomeramg,hypre_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse', #'BoomerAMG w/ coord.',\n 'gmres_boomeramg,hypre_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symTrue_lumpFalse', #'BoomerAMG w/ coord (symmetric)',\n 'gmres_bjacobi,hypre_dt0.0005_atol1e-15_rtol1e-15_theta1.0_symFalse_lumpFalse', #'Block Jacobi + AMG',\n ]\n \n mylabels = None\n \n \n #plt.style.use('ggplot')\n prop_cycle = plt.rcParams['axes.prop_cycle']\n colors = prop_cycle.by_key()['color']\n \n #colors = [\"r\",\"m\",\"c\",\"b\"]\n #linestyle_cycler = cycler.cycler('linestyle',['-','--',':','-.'])\n # (cycler.cycler('color', [\"k\",(0.3,0.3,0.7),(0.7,0.7,1.0), \"r\", \"y\"])+cycler.cycler('linestyle', ['-', '--', ':', '-', '-'])))\n plt.rc('axes', prop_cycle=(cycler('color', [colors[0],colors[1],colors[1],colors[2],colors[2],colors[3],colors[3],colors[4],colors[4],colors[5],colors[5],colors[6],colors[7],colors[8],colors[8],colors[9]]) +\n cycler('linestyle', ['-', '-', '--', '-', '--', '-', '--', '-', '--', '-', '--','-','-','-','--','-']) +\n cycler('marker', ['o', 'o', 'x', 'o', 'x', 'o', 'x', 'o', 'x', 'o', 'x', 'o', 'o', 'o', 'x', 'o'])\n ))\n #plt.rc('axes', prop_cycle=(\"cycler('color', 'rgb') + cycler('linestyle', ['-', '-', ':'])\"))\n \n # -------------------------------------\n # plot runtime\n fig = plt.figure(figsize=(12,5))\n for name in order:\n \n x_values = [6,8,10,12]\n y_values = lines_runtime[name]\n \n print(label[name])\n \n plt.plot(x_values, y_values, label=label[name])\n \n ax = plt.gca()\n ax.set_xticks([6,8,10,12])\n ax.set_xticklabels([\"6\\n36\",\"8\\n48\",\"10\\n60\",\"12\\n72\"])\n ax.set_yscale('log')\n ax.grid(which='major')\n ax.set_xlabel('number of motor units\\nnumber of processes')\n ax.set_ylabel('runtime of solver [s]')\n if mylabels is not None:\n ax.legend(labels=mylabels)\n ax.legend(bbox_to_anchor=(1.0, 1.0))\n \n fig.subplots_adjust(bottom=0.2, right=0.7)\n \n if show_plots:\n pass\n #plt.show()\n else:\n plt.tight_layout()\n plot_filename = \"{}_runtime.png\".format(title.replace(\" \", \"\").replace(\"/\", \"\"))\n plt.savefig(plot_filename)\n print(\"Created \\\"{}\\\".\".format(plot_filename))\n \n # -------------------------------------\n # plot number of iterations\n fig = plt.figure(figsize=(12,5))\n for name in order:\n \n x_values = [6,8,10,12]\n y_values = lines_n_iterations[name]\n print(name,y_values)\n \n plt.plot(x_values, y_values, label=label[name])\n \n ax = plt.gca()\n ax.set_xticks([6,8,10,12])\n ax.set_xticklabels([\"6\\n36\",\"8\\n48\",\"10\\n60\",\"12\\n72\"])\n ax.set_yscale('log')\n ax.grid(which='major')\n ax.set_xlabel('number of motor units\\nnumber of processes')\n ax.set_ylabel('number of iterations')\n if mylabels is not None:\n ax.legend(labels=mylabels)\n ax.legend(bbox_to_anchor=(1.0, 1.0))\n \n fig.subplots_adjust(bottom=0.2, right=0.7)\n \n if show_plots:\n plt.show()\n else:\n plt.tight_layout()\n plot_filename = \"{}_iterations.png\".format(title.replace(\" \", \"\").replace(\"/\", \"\"))\n plt.savefig(plot_filename)\n print(\"Created \\\"{}\\\".\".format(plot_filename))\n \n \n \ndef output(df, title, columns_to_print, columns_to_plot, plot_labels=None):\n \"\"\"\n print values to console and produce plot\n \"\"\"\n columns_to_extract = list(set(columns_to_plot + columns_to_print + [\"totalUsertime\", \"durationReadGeometry\", \"durationSetStiffnessMatrix\", \n \"durationOnlyWrite\", \"durationAssembleBoundaryConditions\", \"durationInitCellml\", \"durationComputeMappingBetweenMeshes\",\n \"durationMap\"]))\n\n # create auxiliary columns that will be computed\n if not \"memoryResidentSet\" in df:\n df[\"memoryResidentSet\"] = 0\n df[\"n\"] = 0\n\n # remove column that are not present in the df\n for column in list(columns_to_extract):\n if column not in df:\n print(\"Note: remove invalid column {} from columns_to_extract\".format(column))\n columns_to_extract.remove(column)\n \n for column in list(columns_to_print):\n if column not in df:\n print(\"Note: remove invalid column {} from columns_to_print\".format(column))\n columns_to_print.remove(column)\n \n for column in list(columns_to_plot):\n if column not in df:\n print(\"Note: remove column {} from columns_to_plot\".format(column))\n columns_to_plot.remove(column)\n\n # select the captions for the table\n table_shortnames = [column_shortnames[long_name] if long_name in column_shortnames else long_name for long_name in columns_to_print]\n \n # define items to be printed, the columns \"n\" and \"memoryResidentSet\" need to be already present in the df\n items = merge_dicts(\n {column_name: lambda v: (np.mean(v) if v.dtype == np.float64 else str(v.iloc[0]) ) for column_name in columns_to_extract},\n {'n': np.size, \"memoryResidentSet\": lambda v: \"{:.3f} GB\".format(np.mean(v)/(1024.**3))}\n )\n\n print(\"-\"*120)\n print(title)\n print(df.groupby(['scenarioName','multidomainLinearSolver_preconditionerType','nRanks']).agg(items).rename(columns = column_shortnames)[table_shortnames])\n print(\"-\"*120)\n\n # create plot\n plot(df, items)\n\n# ------------------------------------------------\n# define shortnames for the table, each line is\n# long_name : short_name\ncolumn_shortnames = {\n \"totalUsertime\": \"user\",\n \"duration_total\": \"total comp.\",\n \"duration_0D\": \"0D\",\n \"duration_1D\": \"1D\",\n \"duration_init\": \"duration_init\",\n \"duration_bidomain\": \"bidomain\",\n \"duration_mechanics\": \"mechanics\",\n \"duration_multidomain\": \"multidomain\",\n \"durationAssembleBoundaryConditions\": \"initBC\",\n \"durationSetStiffnessMatrix\": \"stiffness\",\n \"durationComputeMappingBetweenMeshes\": \"compMap\",\n \"durationMap\": \"map\",\n \"durationReadGeometry\": \"read\",\n \"durationOnlyWrite\": \"write\",\n \"durationInitCellml\": \"initCell\",\n \"memoryResidentSet\": \"mem\",\n \"meta_partitioning\": \"subdomains\",\n \"nIterations_multidomainLinearSolver\": \"niter\",\n}\n\n# define columns for table and plot (long names)\ncolumns_to_print = [\"meta_partitioning\", \"~nDofs3Dmesh\", \"duration_total\", \"duration_0D\", \"duration_1D\", \"duration_bidomain\", \"duration_multidomain\", \"duration_mechanics\", \"duration_init\", \"durationOnlyWrite\", \"nIterations_multidomainLinearSolver\", \"memoryResidentSet\", \"n\"]\ncolumns_to_plot = [\"duration_total\", \"duration_init\", \"durationOnlyWrite\", \"duration_0D\", \"duration_1D\", \"duration_bidomain\", \"duration_multidomain\", \"duration_mechanics\"]\n\nplot_labels = [\"total\", \"initialization\", \"write VTK files\", \"0D model\", \"1D model\", \"3D model\"]\n\ntitle = input_filename\noutput(df, title, columns_to_print, columns_to_plot, plot_labels)\n\n", "sub_path": "opendihu/07_multidomain_solver/plot_resolution1.py", "file_name": "plot_resolution1.py", "file_ext": "py", "file_size_in_byte": 12723, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "matplotlib.use", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.rcdefaults", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams.update", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 30, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 31, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "pandas_utility.load_df", "line_number": 33, "usage_type": "call"}, {"api_name": "pandas.set_option", "line_number": 50, "usage_type": "call"}, {"api_name": "pandas.set_option", "line_number": 51, "usage_type": "call"}, {"api_name": "pandas.set_option", "line_number": 52, "usage_type": "call"}, {"api_name": "pandas.set_option", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 85, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 87, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 148, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 148, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rc", "line_number": 154, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 154, "usage_type": "name"}, {"api_name": "cycler.cycler", "line_number": 154, "usage_type": "call"}, {"api_name": "cycler.cycler", "line_number": 155, "usage_type": "call"}, {"api_name": "cycler.cycler", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 162, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 162, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 170, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 170, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 172, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 172, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 189, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 191, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 191, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 196, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 196, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 203, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 203, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 205, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 205, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 219, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 219, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 221, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 221, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 223, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 223, "usage_type": "name"}, {"api_name": "numpy.float64", "line_number": 262, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 262, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 263, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 263, "usage_type": "call"}]}
+{"seq_id": "379145224", "text": "from flask import render_template, flash, redirect\nfrom app import app\n@app.route('/')\n@app.route('/index')\ndef index():\n user = {'nickname': 'Michael'}\n posts = [\n {\n 'author': {'nickname': 'Fred Post 1'},\n 'body': 'This is the body of the first post.'\n },\n {\n 'author': {'nickname': 'Craig Post 2'},\n 'body': 'This is the body of the second post.'\n }\n ]\n return render_template('index.html',\n title='Home',\n user=user,\n posts=posts)\n@app.route('/about')\ndef about():\n return render_template('about.html',\n title='About')\n", "sub_path": "app/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 574, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "flask.render_template", "line_number": 17, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 3, "usage_type": "call"}, {"api_name": "app.app", "line_number": 3, "usage_type": "name"}, {"api_name": "app.app.route", "line_number": 4, "usage_type": "call"}, {"api_name": "app.app", "line_number": 4, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 23, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 21, "usage_type": "call"}, {"api_name": "app.app", "line_number": 21, "usage_type": "name"}]}
+{"seq_id": "440875980", "text": "import os\nimport requests\nimport ConfigParser\nfrom sufa.settings import BASE_DIR\n\n\ndef verification_code_cer(request):\n\n def get_client_ip(request):\n x_forwarded_for = request.META.get('HTTP_X_FORWARDED_FOR')\n if x_forwarded_for:\n ip = x_forwarded_for.split(',')[0]\n else:\n ip = request.META.get('REMOTE_ADDR')\n return ip\n\n config_file = os.path.join(BASE_DIR, \"sufa.cnf\")\n cf = ConfigParser.ConfigParser()\n cf.read(config_file)\n\n aid = cf.get('tencent', 'aid')\n app_secret_key = cf.get('tencent', 'app_secret_key')\n\n ticket = request.data.get('ticket')\n randstr = request.dat.get('randstr')\n user_ip = get_client_ip(request)\n url = 'https://ssl.captcha.qq.com/ticket/verify'\n\n params = 'aid=' + aid + '&AppSecretKey=' + app_secret_key + '&Ticket=' + ticket + \\\n '&Randstr=' + randstr + '&UserIP=' + user_ip\n\n requests.get(url, params)\n", "sub_path": "utils/tencent_verification_code.py", "file_name": "tencent_verification_code.py", "file_ext": "py", "file_size_in_byte": 934, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "sufa.settings.BASE_DIR", "line_number": 17, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "ConfigParser.ConfigParser", "line_number": 18, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 32, "usage_type": "call"}]}
+{"seq_id": "625119638", "text": "import argparse\nimport os\nimport numpy as np\nimport pandas as pd \nimport torch\nimport pyro\nimport pyro.distributions as dist\nfrom torch.distributions import constraints\nfrom tqdm import tqdm\nimport sys\nimport helper\nimport itertools\n\n######## FOLLOWING WE DEFINE A CLASS THAT WILL KEEP TRACK OF THE SAMPLE POSTERIOR DATA ########\nclass Posterior_pi_log(object):\n\tdef __init__(self, ref_epig_name_list):\n\t\tself.ref_epig_name_list = ref_epig_name_list\n\t\tself.pi = torch.zeros(len(ref_epig_name_list))\n\tdef update(self, updated_pi):\n\t\tself.pi = updated_pi\n\tdef write_pi(self, output_folder):\n\t\tlast_pi = pd.Series(self.pi.detach().numpy())\n\t\tlast_pi.index = self.ref_epig_name_list\n\t\tsave_fn = os.path.join(output_folder, 'posterior_pi.txt')\n\t\tlast_pi.to_csv(save_fn, header = False, index = True, sep = '\\t')\n###########################################\n\ndef generate_tiny_toy_data(num_obs):\n\tnum_ref_epig = 5\n\tnum_state = 3\n\tnum_mark = 3\n\talpha = np.random.uniform(1, 6, num_ref_epig)\n\tpi = pyro.sample('pi', dist.Dirichlet(torch.tensor(alpha)))\n\tref_state_np = np.random.choice(num_state, size = num_obs * num_ref_epig, replace = True).reshape(num_obs, num_ref_epig)\n\temission_np = np.array([[0.1,0.1,0.8], [0.2,0.1,0.7], [0.3, 0.1,0.6]])\n\temission_df = pd.DataFrame(emission_np, columns = list(map(lambda x: 'M{}'.format(x), range(emission_np.shape[1])))) # columns: M0 --> M...\n\ttransition_mat = torch.tensor(np.array([[0.8,0.1,0.1], [0.3,0.5,0.2], [0.2,0.2,0.6]]))\n\tmark_data = np.zeros((num_obs, num_mark))\n\tZ = torch.zeros(num_obs)\n\tS = torch.zeros(num_obs)\n\tfor i in pyro.plate('genome_loop', num_obs):\n\t\tZ[i] = (pyro.sample('z_{}'.format(i), dist.Categorical(pi))).type(torch.long) # sample reference epig from pi at this genomic position \n\t\tR_i = ref_state_np[i,int(Z[i])] # index of state that is observed at the pick refernece epigenome at the current position\n\t\tS[i] = pyro.sample('S_{}'.format(i), dist.Categorical(transition_mat[R_i,:])) # We can get access to parameters by just using pyro.param('')\n\t\tfor j in pyro.plate('mark_loop', num_mark):\n\t\t\tmark_data[i,j] = pyro.sample('M_{}_{}'.format(i,j), dist.Bernoulli(emission_np[S[i].type(torch.long),j])).item()\n\tmark_data = pd.DataFrame(mark_data, columns = emission_df.columns)\n\treturn alpha, pi, ref_state_np, emission_df, transition_mat, mark_data\n\n# @pyro.infer.config_enumerate : this is never needed because we it is only used when all the hidden variables are discrete. In our case, pi is not discrete.\ndef model(alpha, transformed_emission_tt, ref_state_np, transformed_mark_data, num_obs, num_state, NUM_BINS_SAMPLE_PER_ITER): \n\tnum_ct = len(alpha)\n\tpi = pyro.sample('pi', dist.Dirichlet(alpha)) # sample mixture probabilities of reference epigenome\n\tfor i in pyro.plate('state_loop', num_state):\n\t\ttrans_from_state = pyro.param('beta_{}'.format(i), torch.randn(num_state).exp(), constraint = constraints.simplex) \n\tfor i in pyro.plate('genome_loop', num_obs):\n\t\tz_i = pyro.sample('z_{}'.format(i), dist.Categorical(pi))\n\t\tR_i = ref_state_np[i,z_i] \n\t\tS_i = pyro.sample('S_{}'.format(i), dist.Categorical(pyro.param('beta_{}'.format(R_i)))) # We can get access to parameters by just using pyro.param('')\n\t\tpyro.sample('M_{}'.format(i), dist.Categorical(transformed_emission_tt[S_i.type(torch.long)]), obs = transformed_mark_data[i])\n\n\ndef guide(alpha, transformed_emission_tt, ref_state_np, transformed_mark_data, num_obs, num_state, NUM_BINS_SAMPLE_PER_ITER):\n\t# in this guide, we assume that pi and z are independent\n\t# transformed_mark_data: a 1D tensor, each position corresponding to a genomic position\n\tnum_ct = len(alpha)\n\tq_lambda = pyro.param('q_lambda', alpha, constraint = constraints.positive)\n\tpi = pyro.sample('pi', dist.Dirichlet(q_lambda))\n\tfor i in pyro.plate('state_loop', num_state):\n\t\ttrans_from_state = pyro.param('beta_{}'.format(i), torch.randn(num_state).exp(), constraint = constraints.simplex) # sample transition from state i in ref_epig to other states in the sample of interest\t\n\tfor i in pyro.plate('genome_loop', num_obs, subsample_size = NUM_BINS_SAMPLE_PER_ITER): # for subsampling, we only need to specify the subsampling in guide function, not in the model function\n\t\tz_probs = pyro.param(\"q_z_{}\".format(i), torch.randn(num_ct).exp(), constraint=constraints.simplex) \n\t\t# i added .exp() as suggested by https://www.programcreek.com/python/example/123171/torch.distributions.constraints.positive, constraints.simplex is to guarantee that they sum up to 1, based on https://pytorch.org/docs/stable/distributions.html (search for simplex in this page)\n\t\tz_i = pyro.sample('z_{}'.format(i), dist.Categorical(z_probs))\n\t\tR_i = (ref_state_np[i,z_i]).astype(int) # Ha also checked that when doing subsampling, the model still got the exact data as expected\n\t\tS_i = pyro.sample('S_{}'.format(i), dist.Categorical(pyro.param('beta_{}'.format(R_i)))) \n\ndef train(alpha, ref_state_np, transformed_mark_data, transformed_emission_tt, num_state, num_obs, NUM_TRAIN_ITERATIONS, NUM_BINS_SAMPLE_PER_ITER):\n\tpyro.clear_param_store()\n\tloss_func = pyro.infer.TraceGraph_ELBO(max_plate_nesting=1)\n\tsvi = pyro.infer.SVI(model, guide, pyro.optim.Adam({\"lr\": 0.01}), loss=loss_func)\n\tlosses = []\n\tfor _ in tqdm(range(NUM_TRAIN_ITERATIONS)):\n\t\tloss = svi.step(alpha, transformed_emission_tt, ref_state_np, transformed_mark_data, num_obs, num_state, NUM_BINS_SAMPLE_PER_ITER)\n\t\tlosses.append(loss)\n\tposterior_params = {k: np.array(v.data) for k, v in pyro.get_param_store().items()}\n\treturn posterior_params\n\ndef read_chrom_mark_observed_signals(mark_data):\n\tchrom_mark_list = mark_data.columns\n\tmark_data = mark_data.apply(lambda x: x.astype(int).astype(str), axis = 0) # convert the data from 0.0, 1.0 to 0 and 1 integers\n\tmark_data['combined_obs_int'] = mark_data.apply(lambda x: int(''.join(x), 2), axis = 1) # apply function to each row\n\ttransformed_mark_data = torch.tensor(mark_data['combined_obs_int'].values) # 1D tensor each element is the observed data at each postion. If we have 3 marks, the the observed values can be 0-7.\n\treturn transformed_mark_data, chrom_mark_list \n\ndef calculate_join_emission_multiple_marks(row, binary_tuple, chrom_mark_list):\n\t# this will process each row in the emission matrix (each state)\n\t# binary tuple will be a tuple of length #num_mark, each element in the tuple is 0/1 --> presence/absence call of chromatin mark. The order of chromatin marks will be given in chrom_mark_list. Ex: binary_tuple = (0,0,1), chrom_mark_list = [m1, m2, m3] --> m3 is present and others are absent. This function will return the probability of observing binary_tuple given each of the state. \n\t# function tested on 08/03/2021\n\tbase = row[chrom_mark_list]\n\texponent = pd.Series(binary_tuple, index = chrom_mark_list)\n\treturn np.prod(base**exponent * (1-base)**(1-exponent))\n\ndef read_emission_matrix_into_categorical_prob(emission_df, chrom_mark_list):\n\tall_possible_obs_marks = list(itertools.product(range(2), repeat = len(chrom_mark_list))) # list of tuples, each of length # num_marks --> all possible observations of marks \n\tall_possible_obs_marks_str = list(map(lambda x: ''.join(list(map(str, x))), all_possible_obs_marks)) # convert (0,0,0) --> '000'\n\tfor obs_pattern in all_possible_obs_marks:\n\t\tobs_string = ''.join(list(map(str, obs_pattern)))\n\t\temission_df[obs_string] = emission_df.apply(lambda x: calculate_join_emission_multiple_marks(x, obs_pattern, chrom_mark_list), axis = 1) # apply function to each row\n\tresult_df = emission_df[all_possible_obs_marks_str].copy() # columns are all the possible chromatin mark sequences for the chrom_mark_list. Right now, we assume that the assays being profiled are a subset of the 12 marks in the 25-state roadmap model, We can care about the case where the profiled marks for sample of interest are not among the 12 marks later.\n\tresult_df.columns = list(map(lambda x: int(x, 2), result_df.columns))\n\tresult_df = result_df[np.arange(len(all_possible_obs_marks))] # rearrange so that if # marks = 3 --> columns will be 0 --> 7, correpsonding to the 8 possible combination of observed marks 000 --> 111\n\tresult_df = torch.tensor(result_df.values) # tensor with rows: states, columns: possible combinations of chromatin marks \n\treturn result_df\n\ndef evaluate(alpha, pi, transition_mat, num_state, posterior_params):\n\tprint(posterior_params)\n\tprint (\"alpha\")\n\tprint(alpha)\n\tq_lambda = posterior_params['q_lambda']\t\n\tprint(posterior_params['q_lambda'])\n\tprint(\"pi\")\n\tprint(pi)\n\tpred_p = q_lambda / q_lambda.sum() # expected valye of pi given q_lambda\n\tprint(pred_p)\n\tprint('beta')\n\tprint(transition_mat)\n\tfor i in range(num_state):\n\t\tprint(posterior_params['beta_{}'.format(i)])\n\treturn \n\ndef main(args):\n\tNUM_TRAIN_ITERATIONS = args.NUM_TRAIN_ITERATIONS\n\tNUM_BINS_SAMPLE_PER_ITER = args.NUM_BINS_SAMPLE_PER_ITER\n\tnum_obs = 10000\n\talpha, pi, ref_state_np, emission_df, transition_mat, mark_data = generate_tiny_toy_data(num_obs)\n\t# mark_data and emission_df are pd.DataFrame that share the same column names\n\tprint ('Done generating data')\n\talpha = torch.tensor(alpha) # to make it implementable for pyro\n\tnum_state = emission_df.shape[0]\n\ttransformed_mark_data, chrom_mark_list = read_chrom_mark_observed_signals(mark_data) \n\t# transformed_mark_data: 1D tensor, each element is the observed data at each postion.\n\t# If we have 3 marks, the the observed values can be 0-7.\n\ttransformed_emission_tt = read_emission_matrix_into_categorical_prob(emission_df, chrom_mark_list) # tested\n\t# 2D tensor with rows: states, columns: possible combinations of chromatin marks \n\tprint(transformed_emission_tt)\n\tposterior_params = train(alpha, ref_state_np, transformed_mark_data, transformed_emission_tt, num_state, num_obs, NUM_TRAIN_ITERATIONS, NUM_BINS_SAMPLE_PER_ITER)\n\tevaluate(alpha, pi, transition_mat, num_state, posterior_params)\n\nif __name__ == \"__main__\":\n assert pyro.__version__.startswith(\"1.7.0\")\n parser = argparse.ArgumentParser(description=\"Tiny toy example\")\n parser.add_argument(\"-n\", \"--NUM_TRAIN_ITERATIONS\", default=4000, type=int)\n parser.add_argument(\"-o\", \"--NUM_BINS_SAMPLE_PER_ITER\", default=1000, type=int)\n args = parser.parse_args()\n main(args)\n\n", "sub_path": "experiment_pyro/tiny_toy_example.py", "file_name": "tiny_toy_example.py", "file_ext": "py", "file_size_in_byte": 10168, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "torch.zeros", "line_number": 18, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 32, "usage_type": "attribute"}, {"api_name": "pyro.sample", "line_number": 33, "usage_type": "call"}, {"api_name": "pyro.distributions.Dirichlet", "line_number": 33, "usage_type": "call"}, {"api_name": "pyro.distributions", "line_number": 33, "usage_type": "name"}, {"api_name": "torch.tensor", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 34, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 35, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 40, "usage_type": "call"}, {"api_name": "pyro.plate", "line_number": 41, "usage_type": "call"}, {"api_name": "pyro.sample", "line_number": 42, "usage_type": "call"}, {"api_name": "pyro.distributions.Categorical", "line_number": 42, "usage_type": "call"}, {"api_name": "pyro.distributions", "line_number": 42, "usage_type": "name"}, {"api_name": "torch.long", "line_number": 42, "usage_type": "attribute"}, {"api_name": "pyro.sample", "line_number": 44, "usage_type": "call"}, {"api_name": "pyro.distributions.Categorical", "line_number": 44, "usage_type": "call"}, {"api_name": "pyro.distributions", "line_number": 44, "usage_type": "name"}, {"api_name": "pyro.plate", "line_number": 45, "usage_type": "call"}, {"api_name": "pyro.sample", "line_number": 46, "usage_type": "call"}, {"api_name": "pyro.distributions.Bernoulli", "line_number": 46, "usage_type": "call"}, {"api_name": "pyro.distributions", "line_number": 46, "usage_type": "name"}, {"api_name": "torch.long", "line_number": 46, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 47, "usage_type": "call"}, {"api_name": "pyro.sample", "line_number": 53, "usage_type": "call"}, {"api_name": "pyro.distributions.Dirichlet", "line_number": 53, "usage_type": "call"}, {"api_name": "pyro.distributions", "line_number": 53, "usage_type": "name"}, {"api_name": "pyro.plate", "line_number": 54, "usage_type": "call"}, {"api_name": "pyro.param", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.distributions.constraints.simplex", "line_number": 55, "usage_type": "attribute"}, {"api_name": "torch.distributions.constraints", "line_number": 55, "usage_type": "name"}, {"api_name": "pyro.plate", "line_number": 56, "usage_type": "call"}, {"api_name": "pyro.sample", "line_number": 57, "usage_type": "call"}, {"api_name": "pyro.distributions.Categorical", "line_number": 57, "usage_type": "call"}, {"api_name": "pyro.distributions", "line_number": 57, "usage_type": "name"}, {"api_name": "pyro.sample", "line_number": 59, "usage_type": "call"}, {"api_name": "pyro.distributions.Categorical", "line_number": 59, "usage_type": "call"}, {"api_name": "pyro.distributions", "line_number": 59, "usage_type": "name"}, {"api_name": "pyro.param", "line_number": 59, "usage_type": "call"}, {"api_name": "pyro.sample", "line_number": 60, "usage_type": "call"}, {"api_name": "pyro.distributions.Categorical", "line_number": 60, "usage_type": "call"}, {"api_name": "pyro.distributions", "line_number": 60, "usage_type": "name"}, {"api_name": "torch.long", "line_number": 60, "usage_type": "attribute"}, {"api_name": "pyro.param", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.distributions.constraints.positive", "line_number": 67, "usage_type": "attribute"}, {"api_name": "torch.distributions.constraints", "line_number": 67, "usage_type": "name"}, {"api_name": "pyro.sample", "line_number": 68, "usage_type": "call"}, {"api_name": "pyro.distributions.Dirichlet", "line_number": 68, "usage_type": "call"}, {"api_name": "pyro.distributions", "line_number": 68, "usage_type": "name"}, {"api_name": "pyro.plate", "line_number": 69, "usage_type": "call"}, {"api_name": "pyro.param", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.distributions.constraints.simplex", "line_number": 70, "usage_type": "attribute"}, {"api_name": "torch.distributions.constraints", "line_number": 70, "usage_type": "name"}, {"api_name": "pyro.plate", "line_number": 71, "usage_type": "call"}, {"api_name": "pyro.param", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.distributions.constraints.simplex", "line_number": 72, "usage_type": "attribute"}, {"api_name": "torch.distributions.constraints", "line_number": 72, "usage_type": "name"}, {"api_name": "pyro.sample", "line_number": 74, "usage_type": "call"}, {"api_name": "pyro.distributions.Categorical", "line_number": 74, "usage_type": "call"}, {"api_name": "pyro.distributions", "line_number": 74, "usage_type": "name"}, {"api_name": "pyro.sample", "line_number": 76, "usage_type": "call"}, {"api_name": "pyro.distributions.Categorical", "line_number": 76, "usage_type": "call"}, {"api_name": "pyro.distributions", "line_number": 76, "usage_type": "name"}, {"api_name": "pyro.param", "line_number": 76, "usage_type": "call"}, {"api_name": "pyro.clear_param_store", "line_number": 79, "usage_type": "call"}, {"api_name": "pyro.infer.TraceGraph_ELBO", "line_number": 80, "usage_type": "call"}, {"api_name": "pyro.infer", "line_number": 80, "usage_type": "attribute"}, {"api_name": "pyro.infer.SVI", "line_number": 81, "usage_type": "call"}, {"api_name": "pyro.infer", "line_number": 81, "usage_type": "attribute"}, {"api_name": "pyro.optim.Adam", "line_number": 81, "usage_type": "call"}, {"api_name": "pyro.optim", "line_number": 81, "usage_type": "attribute"}, {"api_name": "tqdm.tqdm", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 86, "usage_type": "call"}, {"api_name": "pyro.get_param_store", "line_number": 86, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 93, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.prod", "line_number": 102, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 112, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 113, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 139, "usage_type": "call"}, {"api_name": "pyro.__version__.startswith", "line_number": 151, "usage_type": "call"}, {"api_name": "pyro.__version__", "line_number": 151, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 152, "usage_type": "call"}]}
+{"seq_id": "123041933", "text": "#!/usr/bin/env python\n\nfrom utils import utils, inspector\nfrom bs4 import BeautifulSoup\nfrom datetime import datetime\nimport re\nimport logging\n\n# oldest year: 1979\n\ndef run(options):\n crawl_index(SEMIANNUAL_REPORTS_URL, options)\n crawl_index(AUDIT_REPORTS_URL, options, True)\n crawl_index(PEER_REVIEW_REPORTS_URL, options)\n crawl_index(MISCELLANEOUS_REPORTS_URL, options)\n\ndef crawl_index(base_url, options, is_meta_index=False):\n year_range = inspector.year_range(options)\n max_pages = options.get('pages')\n if max_pages:\n max_pages = int(max_pages)\n page = 1\n\n only_id = options.get('report_id')\n\n done = False\n while not done:\n url = url_for(base_url, page)\n body = utils.download(url)\n\n doc = BeautifulSoup(body)\n\n next_page = page + 1\n found_next_page = False\n page_links = doc.select(\"dl.moreResults a\")\n for page_link in page_links:\n if page_link.text == str(next_page):\n found_next_page = True\n break\n if not found_next_page:\n done = True\n if max_pages and next_page > max_pages:\n done = True\n\n results = doc.select(\"div#svPortal dl\")\n for result in results:\n if \"moreResults\" in result.get(\"class\"):\n continue\n if is_meta_index:\n url = \"http://www.gsaig.gov\" + result.a.get(\"href\")\n crawl_index(url, options, False)\n else:\n report = report_from(result, base_url)\n year = int(report['published_on'][:4])\n\n if only_id and (report['report_id'] != only_id):\n continue\n\n if year not in year_range:\n continue\n\n inspector.save_report(report)\n\n page = next_page\n if not done:\n logging.info('Moving to next page (%d)' % page)\n\ndef url_for(base_url, page = 1):\n return \"%s?startRow=%d\" % (base_url, page * 10 - 9)\n\ndef report_from(result, base_url):\n report = {\n 'inspector': 'gsa',\n 'inspector_url': 'http://gsaig.gov/',\n 'agency': 'gsa',\n 'agency_name': 'General Services Administration'\n }\n\n link = result.a\n title = link.text\n url = link.get('href')\n\n date_holders = result.find_all(\"dt\", class_=\"releaseDate\")\n if len(date_holders) > 0:\n published_date = date_holders[0].text\n date = datetime.strptime(published_date, \"%B %d, %Y\")\n elif title in HARDCODED_DATES:\n # This is an ugly solution, but there's no date information on the web page.\n # The next best solution would be to grab the PDF file and pull the file\n # creation date out of its metadata.\n published_date = HARDCODED_DATES[title]\n date = datetime.strptime(published_date, \"%B %d, %Y\")\n elif base_url == SEMIANNUAL_REPORTS_URL:\n # get last match\n match = None\n for match in DATE_RE.finditer(title):\n pass\n published_date = match.group(0)\n date = datetime.strptime(published_date, \"%B %d, %Y\")\n else:\n match = DATE_RE_MM_DD_YY.search(result.text)\n if match:\n published_date = match.group(0)\n date = datetime.strptime(published_date, \"%m/%d/%y\")\n else:\n raise Exception(\"Couldn't find date for %s\" % title)\n\n id = ID_RE.search(url).group(1)\n\n report_type = type_for(base_url)\n\n js_match = JS_RE.match(url)\n if js_match:\n url = \"http://www.gsaig.gov\" + js_match.group(1)\n elif url.startswith('/'):\n url = \"http://www.gsaig.gov\" + url\n\n report['type'] = report_type\n report['published_on'] = datetime.strftime(date, \"%Y-%m-%d\")\n report['url'] = url\n report['report_id'] = id\n report['title'] = title.strip()\n report['file_type'] = 'pdf'\n\n return report\n\ndef type_for(base_url):\n if base_url.find('special-reports') != -1:\n return \"audit\"\n if base_url.find('audit-reports') != -1:\n return \"audit\"\n return \"other\"\n\nSEMIANNUAL_REPORTS_URL = \"http://www.gsaig.gov/index.cfm/oig-reports/semiannual-reports-to-the-congress/\"\nAUDIT_REPORTS_URL = \"http://www.gsaig.gov/index.cfm/oig-reports/audit-reports/\"\nPEER_REVIEW_REPORTS_URL = \"http://www.gsaig.gov/index.cfm/oig-reports/peer-review-reports/\"\nMISCELLANEOUS_REPORTS_URL = \"http://www.gsaig.gov/index.cfm/oig-reports/miscellaneous-reports/\"\n\nID_RE = re.compile(\"LinkServID=([-0-9A-F]*)&showMeta=\")\nJS_RE = re.compile(\"\"\"javascript:newWin=window.open\\('/(\\?LinkServID=([-0-9A-F]*)&showMeta=0)','NewWin[0-9]*'\\);newWin.focus\\(\\);void\\(0\\)\"\"\")\nDATE_RE = re.compile(\"(January|February|March|April|May|June|July|August|\" +\n \"September|October|November|December) ([123]?[0-9]), \" +\n \"([12][0-9][0-9][0-9])\")\nDATE_RE_MM_DD_YY = re.compile(\"[0-9]?[0-9]/[0-9]?[0-9]/[0-9][0-9]\")\n\nHARDCODED_DATES = {\n \"Hats Off Program Investigative Report\": \"June 16, 2011\",\n \"Major Issues from Fiscal Year 2010 Multiple Award Schedule Preaward Audits\": \"September 26, 2011\",\n \"Review of Center for Information Security Services FTS\": \"March 23, 2001\",\n \"Audit of Procurement of Profesional Services from the FSS Multiple Award Schedules\": \"July 31, 2003\",\n \"Special Report: MAS Pricing Practices: Is FSS Observing Regulatory Provisions Regarding Pricing?\": \"August 24, 2001\",\n \"Updated Assessment of GSA's Most Serious Challenges\": \"December 8, 2004\",\n \"Limited Audit of FSS's Contracting for Services Under Multiple Award Schedule Contracts\": \"January 9, 2001\",\n \"Procurement Reform and the Multiple Award Schedule Program\": \"July 30, 2010\",\n \"FTS Alert Report\": \"March 6, 2003\",\n \"FTS CSC Audit Report\": \"January 8, 2004\",\n \"Compendium FTS CSC Audit Report\": \"December 14, 2004\",\n \"Compendium FTS CSC Controls Audit Report\": \"June 14, 2005\",\n \"Compendium FTS Client Support Center Controls Audit Report\": \"September 29, 2006\",\n \"Review of the Federal Acquisition Service's Client Support Center, Southeast Sunbelt Region - A090139-3\": \"June 4, 2010\"\n}\n\nutils.run(run) if (__name__ == \"__main__\") else None\n", "sub_path": "inspectors/gsa.py", "file_name": "gsa.py", "file_ext": "py", "file_size_in_byte": 5738, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "utils.inspector.year_range", "line_number": 18, "usage_type": "call"}, {"api_name": "utils.inspector", "line_number": 18, "usage_type": "name"}, {"api_name": "utils.utils.download", "line_number": 29, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 29, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 31, "usage_type": "call"}, {"api_name": "utils.inspector.save_report", "line_number": 62, "usage_type": "call"}, {"api_name": "utils.inspector", "line_number": 62, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 66, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 86, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 86, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 92, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 92, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 99, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 99, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 104, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 104, "usage_type": "name"}, {"api_name": "datetime.datetime.strftime", "line_number": 119, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 119, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 139, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 140, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 141, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 144, "usage_type": "call"}, {"api_name": "utils.utils.run", "line_number": 163, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 163, "usage_type": "name"}]}
+{"seq_id": "535121570", "text": "from flask import Flask, Response, request, jsonify, abort\n\n# from . import settings\nfrom .utils.database import (\n get_database,\n list_databases,\n list_database_templates,\n database_from_template)\n\n\napp = Flask(__name__)\n\n\n@app.route('/', methods=['get'])\ndef hello_world():\n return jsonify(\n status=\"Service is running\",\n templates=list_database_templates(),\n databases=list_databases())\n\n\n@app.route('/db', methods=['get'])\ndef show_databases():\n return jsonify(databases=list_databases())\n\n\n@app.route('/templates', methods=['get'])\ndef show_templates():\n return jsonify(templates=list_database_templates())\n\n\n@app.route('/db//', methods=['get'])\ndef inspect_database(template, name):\n db = get_database(template, name)\n if not db:\n abort(404)\n\n if request.args.get('all'):\n return jsonify(container=db.inspect())\n\n return jsonify(\n database=db.database,\n host=db.external_ip,\n port=db.external_port,\n user=db.user,\n password=db.password,\n running=db.running())\n\n\n@app.route('/db//', methods=['post'])\ndef create_database(template, name):\n db = get_database(template, name)\n if db:\n response = Response(status=304) # not modified\n del response.headers['content-type']\n return response\n\n db = database_from_template(template, name)\n if not db:\n abort(404)\n db.start()\n response = inspect_database(template, name)\n response.status_code = 201\n return response\n\n\n@app.route('/db//', methods=['delete'])\ndef remove_database(template, name):\n db = get_database(template, name)\n if not db:\n abort(404)\n db.purge()\n response = Response(status=204)\n del response.headers['content-type']\n return response\n\n\n@app.errorhandler(404)\ndef page_not_found(e):\n response = jsonify(status=\"Not found\")\n response.status_code = 404\n return response\n", "sub_path": "db/launcher/server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 1977, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "flask.Flask", "line_number": 11, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 16, "usage_type": "call"}, {"api_name": "utils.database.list_database_templates", "line_number": 18, "usage_type": "call"}, {"api_name": "utils.database.list_databases", "line_number": 19, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 24, "usage_type": "call"}, {"api_name": "utils.database.list_databases", "line_number": 24, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 29, "usage_type": "call"}, {"api_name": "utils.database.list_database_templates", "line_number": 29, "usage_type": "call"}, {"api_name": "utils.database.get_database", "line_number": 34, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 36, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 38, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 38, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 39, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 41, "usage_type": "call"}, {"api_name": "utils.database.get_database", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 54, "usage_type": "call"}, {"api_name": "utils.database.database_from_template", "line_number": 58, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 60, "usage_type": "call"}, {"api_name": "utils.database.get_database", "line_number": 69, "usage_type": "call"}, {"api_name": "flask.abort", "line_number": 71, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 73, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 80, "usage_type": "call"}]}
+{"seq_id": "106082190", "text": "from __future__ import division, print_function\nimport numpy as np\nimport cv2\nimport skfuzzy as fuzz\nimport matplotlib.pyplot as plt\n\ncap = cv2.VideoCapture(\"/home/veerav/PycharmProjects/fuzzy/Badminton_ce2/img/%04d.jpg\")\n\n # take first frame of the video\nret,frame = cap.read()\nrefPt = []\ncropping = False\nprint(frame.shape)\n\n\ndef click_and_crop(event, x, y, flags, param):\n # grab references to the global variables\n global refPt, cropping\n\n # if the left mouse button was clicked, record the starting\n # (x, y) coordinates and indicate that cropping is being\n # performed\n if event == cv2.EVENT_LBUTTONDOWN:\n refPt = [(x, y)]\n cropping = True\n\n # check to see if the left mouse button was released\n elif event == cv2.EVENT_LBUTTONUP:\n # record the ending (x, y) coordinates and indicate that\n # the cropping operation is finished\n refPt.append((x, y))\n cropping = False\n\n # draw a rectangle around the region of interest\n cv2.rectangle(frame, refPt[0], refPt[1], (0, 255, 0), 2)\n cv2.imshow(\"image\", frame)\n\n\n\n\n# load the image, clone it, and setup the mouse callback function\n\nclone = frame.copy()\ncv2.namedWindow(\"image\")\ncv2.setMouseCallback(\"image\", click_and_crop)\n\n# keep looping until the 'q' key is pressed\nwhile True:\n # display the image and wait for a keypress\n cv2.imshow(\"image\", frame)\n key = cv2.waitKey(1) & 0xFF\n\n # if the 'r' key is pressed, reset the cropping region\n if key == ord(\"r\"):\n image = clone.copy()\n\n # if the 'c' key is pressed, break from the loop\n elif key == ord(\"c\"):\n break\n\n# if there are two reference points, then crop the region of interest\n# from teh image and display it\nif len(refPt) == 2:\n roi = clone[refPt[0][1]:refPt[1][1], refPt[0][0]:refPt[1][0]]\n cv2.imshow(\"ROI\", roi)\n cv2.waitKey(0)\n\n# close all open windows\ncv2.destroyAllWindows()\ntrack_window = (refPt[0][0],refPt[0][1],refPt[1][0]-refPt[0][0],refPt[1][1]-refPt[0][1])\n # setup initial location of window\n#r,h,c,w = 250,90,400,125 # simply hardcoded the values\n#track_window = (c,r,w,h)\n\n # set up the ROI for tracking\n#roi = frame[r:r+h, c:c+w]\n\n\"\"\"\n========================\nFuzzy c-means clustering\n========================\n\n\"\"\"\n\n# Generate test data\n\nroi = cv2.cvtColor(roi, cv2.COLOR_BGR2HSV)\nprint ('shape of image is ',roi.shape)\nxpts = np.zeros(roi.shape[0]*roi.shape[1])\nypts = np.zeros(roi.shape[0]*roi.shape[1])\n#zpts = np.zeros(roi.shape[0]*roi.shape[1])\nm = 0\nfor y in range(0,roi.shape[0]):\n for x in range(0,roi.shape[1]):\n xpts[m] = roi[y][x][0]\n #ypts[m] = roi[y][x][1]\n #zpts[m] = roi[y][x][2]\n m = m+1\nalldata = xpts.reshape(1, roi.shape[0] * roi.shape[1])\n#alldata = np.vstack((xpts,ypts,zpts))\n#alldata = np.vstack((xpts,ypts))\nprint ('alldata shape is ',alldata.shape)\n\n\n\n#Clustering\n#----------\n\n\ncntr, u, u0, d, jm, p, fpc = fuzz.cluster.cmeans(alldata, 8, 2, error=0.005, maxiter=100, init=None)\n\n#print ('membership is ',u)\n\n#print ('shape of partition matrix is ', u.shape)\n\n#print ('fp is ',fpc)\n\n\ncluster_assign = np.argmax(u,axis =0) # Hardening for visualization\ncluster_assign = cluster_assign.astype(np.uint8)\n#print (cluster_assign.shape)\n#print (type(cluster_assign))\nroi_hist = cv2.calcHist([cluster_assign],[0],None,[8],[0,8])\n\n\n\n\n\n#############\n#hsv_roi = cv2.cvtColor(roi, cv2.COLOR_BGR2HSV)\n#mask = cv2.inRange(hsv_roi, np.array((0., 60.,32.)), np.array((180.,255.,255.)))\n#roi_hist = cv2.calcHist([hsv_roi],[0,1],None,[32,32],[0,180,0,256])\n#cv2.normalize(roi_hist,roi_hist,0,255,cv2.NORM_MINMAX)\n#cv2.normalize(roi_hist,roi_hist,0,255,cv2.NORM_MINMAX)\ncv2.normalize(roi_hist,roi_hist,alpha = 0,beta = 1,norm_type=cv2.NORM_MINMAX,dtype=cv2.CV_32F)\n\n# Setup the termination criteria, either 10 iteration or move by atleast 1 pt\nterm_crit = ( cv2.TERM_CRITERIA_EPS | cv2.TERM_CRITERIA_COUNT, 10, 1 )\n\nwhile(1):\n ret ,frame = cap.read()\n\n if ret == True:\n hsv = cv2.cvtColor(frame, cv2.COLOR_BGR2HSV)\n\n dst = cv2.calcBackProject([hsv],[0],roi_hist,[0,8],1)\n\n\n # apply meanshift to get the new location\n ret, track_window = cv2.meanShift(dst, track_window, term_crit)\n\n # Draw it on image\n x,y,w,h = track_window\n img2 = cv2.rectangle(frame, (x,y), (x+w,y+h), 255,2)\n cv2.imshow('img2',img2)\n\n k = cv2.waitKey(60) & 0xff\n if k == 27:\n break\n else:\n cv2.imwrite(chr(k)+\".jpg\",img2)\n\n else:\n break\n\n\ncv2.destroyAllWindows()\ncap.release()", "sub_path": "fuzzy_test.py", "file_name": "fuzzy_test.py", "file_ext": "py", "file_size_in_byte": 4550, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "cv2.VideoCapture", "line_number": 7, "usage_type": "call"}, {"api_name": "cv2.EVENT_LBUTTONDOWN", "line_number": 23, "usage_type": "attribute"}, {"api_name": "cv2.EVENT_LBUTTONUP", "line_number": 28, "usage_type": "attribute"}, {"api_name": "cv2.rectangle", "line_number": 35, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 36, "usage_type": "call"}, {"api_name": "cv2.namedWindow", "line_number": 44, "usage_type": "call"}, {"api_name": "cv2.setMouseCallback", "line_number": 45, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 50, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 51, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 65, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 66, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 69, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 87, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2HSV", "line_number": 87, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 90, "usage_type": "call"}, {"api_name": "skfuzzy.cluster.cmeans", "line_number": 110, "usage_type": "call"}, {"api_name": "skfuzzy.cluster", "line_number": 110, "usage_type": "attribute"}, {"api_name": "numpy.argmax", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 120, "usage_type": "attribute"}, {"api_name": "cv2.calcHist", "line_number": 123, "usage_type": "call"}, {"api_name": "cv2.normalize", "line_number": 135, "usage_type": "call"}, {"api_name": "cv2.NORM_MINMAX", "line_number": 135, "usage_type": "attribute"}, {"api_name": "cv2.CV_32F", "line_number": 135, "usage_type": "attribute"}, {"api_name": "cv2.TERM_CRITERIA_EPS", "line_number": 138, "usage_type": "attribute"}, {"api_name": "cv2.TERM_CRITERIA_COUNT", "line_number": 138, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 144, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2HSV", "line_number": 144, "usage_type": "attribute"}, {"api_name": "cv2.calcBackProject", "line_number": 146, "usage_type": "call"}, {"api_name": "cv2.meanShift", "line_number": 150, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 154, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 155, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 157, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 161, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 167, "usage_type": "call"}]}
+{"seq_id": "408694054", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\nimport sys\nsys.path.append(\"./examples/proto\")\n\nfrom pprint import pprint\nimport logging\nimport time\n\nfrom ledgerblue.comm import getDongle\nimport argparse\nfrom base import parse_bip32_path\n\nimport validateSignature\nimport binascii\nimport base58\n\nlogging.basicConfig(level=logging.DEBUG, format=\"%(asctime)s - %(levelname)s - %(message)s\")\nlogger = logging.getLogger()\n\n# Start Ledger\ndongle = getDongle(True)\n\n\ndef chunks(l, n):\n \"\"\"Yield successive n-sized chunks from l.\"\"\"\n for i in range(0, len(l), n):\n yield l[i:i + n]\n\ndef apduMessage(INS, P1, P2, PATH, MESSAGE):\n hexString = \"\"\n if PATH:\n hexString = \"E0{:02x}{:02x}{:02x}{:02x}{:02x}{}\".format(INS,P1,P2,(len(PATH)+len(MESSAGE))//2+1,len(PATH)//4//2,PATH+MESSAGE)\n else:\n hexString = \"E0{:02x}{:02x}{:02x}{:02x}{}\".format(INS,P1,P2,len(MESSAGE)//2,MESSAGE)\n print(hexString)\n return bytearray.fromhex(hexString)\n\ndef ledgerSign(PATH, tx, tokenSignature=[]):\n raw_tx = tx.raw_data.SerializeToString().hex()\n # Sign in chunks\n chunkList = list(chunks(raw_tx,420))\n if len(tokenSignature)>0:\n chunkList.extend(tokenSignature)\n\n # P1 = P1_FIRST = 0x00\n if len(chunkList)>1:\n result = dongle.exchange(apduMessage(0x04,0x00,0x00, PATH, chunkList[0]))\n else:\n result = dongle.exchange(apduMessage(0x04,0x10,0x00, PATH, chunkList[0]))\n\n for i in range(1,len(chunkList)-1-len(tokenSignature)):\n # P1 = P1_MODE = 0x80\n result = dongle.exchange(apduMessage(0x04,0x80,0x00, None, chunkList[i]))\n \n for i in range(0,len(tokenSignature)-1):\n result = dongle.exchange(apduMessage(0x04,0xA0 | (0x00+i), 0x00, None, tokenSignature[i]))\n \n # P1 = P1_LAST = 0x90\n if len(chunkList)>1:\n if len(tokenSignature)>0:\n result = dongle.exchange(apduMessage(0x04,0xA0 | 0x08 | (0x00+len(tokenSignature)-1),0x00, None, chunkList[len(chunkList)-1]))\n else:\n result = dongle.exchange(apduMessage(0x04,0x90,0x00, None, chunkList[len(chunkList)-1]))\n\n return raw_tx, result\n\ndef address_hex(address):\n return base58.b58decode_check(address).hex().upper()\n\n# Get Addresses\nlogger.debug('-= Vision Ledger =-')\n\n'''\nVision Protobuf\n'''\nfrom api import api_pb2 as api\nfrom api.api_pb2_grpc import WalletStub\nfrom core import Vision_pb2 as vision\nfrom google.protobuf.any_pb2 import Any\nimport grpc\n\nfrom core.contract import balance_contract_pb2 as contract\n\n# Start Channel and WalletStub\nchannel = grpc.insecure_channel(\"54.238.114.220:60061\")\nstub = WalletStub(channel)\n\nlogger.debug('''\n Vision MultiSign tests\n''')\n\ntx = stub.CreateTransaction2(\n contract.TransferContract(\n owner_address=bytes.fromhex(address_hex(\"VNsAHAYVSKfKLAkvN4pRRWYoXtpeH69iqk\")),\n to_address=bytes.fromhex(address_hex(\"VBkBvRzsUNzAtq4XdEVnzh15i4DDkUcM1X\")),\n amount=100000\n )\n ) \n# use permission 2\ntx.transaction.raw_data.contract[0].Permission_id=2\n\n\n\nraw_tx, sign1 = ledgerSign(parse_bip32_path(\"44'/195'/0'/0/0\"),tx.transaction)\nraw_tx, sign2 = ledgerSign(parse_bip32_path(\"44'/195'/1'/0/0\"),tx.transaction)\n\ntx.transaction.signature.extend([bytes(sign1[0:65])])\ntx.transaction.signature.extend([bytes(sign2[0:65])])\nprint(\"tx:\", tx)\nprint(\"txID:\", tx.txid.hex())\nr = stub.BroadcastTransaction(tx.transaction)\nprint(\"result:\", r)\nif r.result == True:\n\tprint(\"Success\")\nelse:\n\tprint(\"Fail\")\n", "sub_path": "examples/multisignTransfer.py", "file_name": "multisignTransfer.py", "file_ext": "py", "file_size_in_byte": 3466, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sys.path.append", "line_number": 4, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 4, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 18, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 18, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 19, "usage_type": "call"}, {"api_name": "ledgerblue.comm.getDongle", "line_number": 22, "usage_type": "call"}, {"api_name": "base58.b58decode_check", "line_number": 69, "usage_type": "call"}, {"api_name": "grpc.insecure_channel", "line_number": 86, "usage_type": "call"}, {"api_name": "api.api_pb2_grpc.WalletStub", "line_number": 87, "usage_type": "call"}, {"api_name": "core.contract.balance_contract_pb2.TransferContract", "line_number": 94, "usage_type": "call"}, {"api_name": "core.contract.balance_contract_pb2", "line_number": 94, "usage_type": "name"}, {"api_name": "base.parse_bip32_path", "line_number": 105, "usage_type": "call"}, {"api_name": "base.parse_bip32_path", "line_number": 106, "usage_type": "call"}]}
+{"seq_id": "137640873", "text": "import sys\nimport pygame\nfrom pygame.sprite import Sprite\n\n\nclass DrawDrops:\n\n def __init__(self):\n pygame.init()\n self.screen_width = 800\n self.screen_height = 600\n self.screen = pygame.display.set_mode((self.screen_width, self.screen_height))\n self.bg_color = (255, 255, 255)\n pygame.display.set_caption(\"Raining\")\n\n self.drops = pygame.sprite.Group()\n self.drop_spacing = 1.5\n self.dropping_speed = 2\n self._create_drops()\n\n def run_game(self):\n while True:\n self._check_events()\n self._update_screen()\n\n def _check_events(self):\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n pygame.quit()\n sys.exit()\n\n\n def _create_drops(self):\n drop = Drop(self)\n drop_width = drop.rect.width\n available_space_x = self.screen_width - (self.drop_spacing*drop_width)\n number_drop_x = int(available_space_x // (self.drop_spacing*drop_width))\n\n\n for drop_number in range(number_drop_x):\n drop = Drop(self)\n drop.x = drop_width + self.drop_spacing*drop_width*drop_number\n drop.rect.x = int(drop.x)\n self.drops.add(drop)\n\n def _update_drops(self):\n self.drops.update()\n for drop in self.drops.copy():\n if drop.rect.top > self.screen.get_rect().height:\n self.drops.remove(drop)\n #print(len(self.drops))\n \"\"\" when the drops disappeared, create new drops.\"\"\"\n if len(self.drops) < 30:\n self._create_drops()\n\n \n def _update_screen(self):\n self.screen.fill(self.bg_color)\n self.drops.draw(self.screen)\n\n self._update_drops()\n \n pygame.display.flip()\n\n\n \n\nclass Drop(Sprite):\n def __init__(self, d_game):\n super().__init__()\n self.screen = d_game.screen\n self.image = pygame.image.load('./drop.bmp')\n self.rect = self.image.get_rect()\n \n self.rect.x = self.rect.width\n self.rect.y = self.rect.height\n\n self.x = float(self.rect.x)\n\n self.dropping_speed = d_game.dropping_speed\n\n def update(self):\n self.rect.y += self.dropping_speed\n\n \n\n\nif __name__ == '__main__':\n #Make a game instance, and run the game.\n d = DrawDrops()\n d.run_game()\n \n \n \n", "sub_path": "Chapter 13/13-4.py", "file_name": "13-4.py", "file_ext": "py", "file_size_in_byte": 2414, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pygame.init", "line_number": 9, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 12, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 14, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pygame.sprite.Group", "line_number": 16, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 27, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 29, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 30, "usage_type": "call"}, {"api_name": "pygame.display.flip", "line_number": 63, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 63, "usage_type": "attribute"}, {"api_name": "pygame.sprite.Sprite", "line_number": 68, "usage_type": "name"}, {"api_name": "pygame.image.load", "line_number": 72, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 72, "usage_type": "attribute"}]}
+{"seq_id": "138776623", "text": "# Создание датасета, включает считывание исходных данных\n# с диска, их предобработку, аугментацию и перемешивание\nimport cv2\nimport json\nimport numpy as np\nimport pandas as pd\nfrom PIL import Image\n\nfrom pathlib import Path\n\nimport os\n#os.chdir(\"..\")\n#import sys\n#sys.path.append(\"..\") \n\n\nfrom lib import *\n\nimport torch\n\n\nfrom torch.utils.data import Dataset, DataLoader\n\nDATA_MODES = ['train', 'val', 'test']\n\nclass CigaretteButtDataset(Dataset):\n\n def __init__(self,\n img_dpath,\n img_fnames,\n img_transform,\n mask_encodings=None,\n mask_size=None,\n mask_transform=None):\n self.img_dpath = img_dpath\n self.img_fnames = img_fnames\n self.img_transform = img_transform\n\n self.mask_encodings = mask_encodings\n self.mask_size = mask_size\n self.mask_transform = mask_transform\n\n def __getitem__(self, i):\n\n seed = np.random.randint(2147483647)\n\n fname = self.img_fnames[i]\n fpath = os.path.join(self.img_dpath, fname)\n img = Image.open(fpath)\n if self.img_transform is not None:\n random.seed(seed)\n img = self.img_transform(img)\n\n if self.mask_encodings is None:\n return img, fname\n\n if self.mask_size is None or self.mask_transform is None:\n raise ValueError('If mask_dpath is not None, mask_size and mask_transform must not be None.')\n\n mask = np.zeros(self.mask_size, dtype=np.uint8)\n if self.mask_encodings[fname][0] == self.mask_encodings[fname][0]: # NaN doesn't equal to itself\n for encoding in self.mask_encodings[fname]:\n mask += rle_decode(encoding, self.mask_size)\n mask = np.clip(mask, 0, 1)\n\n mask = Image.fromarray(mask)\n\n random.seed(seed)\n mask = self.mask_transform(mask)\n\n return img, torch.from_numpy(np.array(mask, dtype=np.int64))\n\n def __len__(self):\n return len(self.img_fnames)\n\ndef prepare_datasets():\n\n TRAIN_DIR = Path('data/train/images/')\n VAL_DIR = Path('data/val/images/')\n TEST_DIR = Path('data/real_test/')\n\n #path = \"data/train\"\n train_images = os.listdir(f\"{path}/images\")\n val_images = os.listdir()\n train_annotations = json.load(open(f\"{TRAIN_DIR}/coco_annotations.json\", \"r\"))\n #img_id = int(np.random.choice(images).split(\".\")[0])\n\n img = np.array(Image.open(f\"{path}/images/{img_id:08}.jpg\"))\n mask = get_mask(img_id, annotations)\n show_img_with_mask(img, mask)\n\n train_images = sorted(list(TRAIN_DIR.rglob('*.jpg')))\n val_files = sorted(list(VAL_DIR.rglob('*.jpg')))\n\n train_loader = DataLoader(train_dataset, batch_size=batch_size, shuffle=True)\n val_loader = DataLoader(val_dataset, batch_size=batch_size, shuffle=False)\n test_loader = DataLoader(test_dataset, batch_size=batch_size, shuffle=False)\n\n data_loader = {'train': train_loader, 'val': val_loader, 'test': test_loader}\n\n return data_loader\n", "sub_path": "cigarette_butt_segmentation/source/datasets.py", "file_name": "datasets.py", "file_ext": "py", "file_size_in_byte": 3113, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "torch.utils.data.Dataset", "line_number": 26, "usage_type": "name"}, {"api_name": "numpy.random.randint", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 49, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 49, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 60, "usage_type": "attribute"}, {"api_name": "numpy.clip", "line_number": 64, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 66, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 66, "usage_type": "name"}, {"api_name": "torch.from_numpy", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 71, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 78, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 79, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 80, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 83, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 84, "usage_type": "call"}, {"api_name": "json.load", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 88, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 88, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 88, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 97, "usage_type": "call"}]}
+{"seq_id": "506304619", "text": "from django.core import mail\nfrom django.test import TestCase\nfrom django.shortcuts import resolve_url as r\n\n\nclass SubscribePostValid(TestCase):\n def setUp(self):\n data = dict(name='Marcelo Andriolli', cpf='12345678901',\n email='marcelorsa@gmail.com', phone='48-996274443')\n self.client.post(r('subscriptions:new'), data)\n self.email = mail.outbox[0]\n \n def test_send_subscribe_email_subject(self):\n expect = 'Confirmação de inscrição'\n\n self.assertEqual(expect, self.email.subject)\n \n def test_send_subscribe_email_from(self):\n expect = 'contato@eventex.com.br'\n\n self.assertEqual(expect, self.email.from_email)\n\n def test_send_subscribe_email_to(self):\n expect = ['contato@eventex.com.br', 'marcelorsa@gmail.com']\n\n self.assertEqual(expect, self.email.to)\n\n def test_send_subscribe_email_body(self):\n contents = [\n 'Marcelo Andriolli',\n '12345678901',\n 'marcelorsa@gmail.com',\n '48-996274443'\n ]\n for content in contents:\n with self.subTest():\n self.assertIn(content, self.email.body)\n", "sub_path": "eventex/subscriptions/tests/test_mail_subscribe.py", "file_name": "test_mail_subscribe.py", "file_ext": "py", "file_size_in_byte": 1186, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.test.TestCase", "line_number": 6, "usage_type": "name"}, {"api_name": "django.shortcuts.resolve_url", "line_number": 10, "usage_type": "call"}, {"api_name": "django.core.mail.outbox", "line_number": 11, "usage_type": "attribute"}, {"api_name": "django.core.mail", "line_number": 11, "usage_type": "name"}]}
+{"seq_id": "123919684", "text": "import json\nimport os\nimport pandas as pd\nimport warnings\nfrom datamart.metadata.global_metadata import GlobalMetadata\nfrom datamart.metadata.variable_metadata import VariableMetadata\nfrom datamart.es_managers.index_manager import IndexManager\nfrom datamart.utils import Utils\nfrom datamart.profilers.basic_profiler import BasicProfiler\nimport typing\nimport traceback\n\nGLOBAL_INDEX_INTERVAL = 10000\n\n\nclass IndexBuilder(object):\n def __init__(self) -> None:\n \"\"\"Init method of IndexBuilder.\n\n \"\"\"\n\n self.resources_path = os.path.join(os.path.dirname(__file__), \"resources\")\n with open(os.path.join(self.resources_path, 'index_info.json'), 'r') as index_info_f:\n self.index_config = json.load(index_info_f)\n self.current_global_index = None\n self.GLOBAL_INDEX_INTERVAL = GLOBAL_INDEX_INTERVAL\n self.basic_profiler = BasicProfiler\n self.im = IndexManager(es_host=self.index_config[\"es_host\"], es_port=self.index_config[\"es_port\"])\n\n def indexing(self,\n description_path: str,\n es_index: str,\n data_path: str = None,\n query_data_for_indexing: bool = False,\n save_to_file: str = None,\n save_to_file_mode: str = \"a+\",\n delete_old_es_index: bool = False\n ) -> dict:\n \"\"\"API for the index builder.\n\n By providing description file, index builder should be able to process it and create metadata json for the\n dataset, create index in our index store\n\n Args:\n description_path: Path to description json file.\n es_index: str, es index for this dataset\n data_path: Path to data csv file.\n query_data_for_indexing: Bool. If no data is presented, and query_data_for_indexing is False, will only\n create metadata according to the description json. If query_data_for_indexing is True and no data is\n presented, will use Materialize to query data for profiling and indexing\n save_to_file: str, a path to the json line file\n save_to_file_mode: str, mode for saving, default \"a+\"\n delete_old_es_index: bool, boolean if delete original es index if it exist\n\n Returns:\n metadata dictionary\n\n \"\"\"\n\n self._check_es_index(es_index=es_index, delete_old_es_index=delete_old_es_index)\n\n if not self.current_global_index or delete_old_es_index:\n self.current_global_index = self.im.current_global_datamart_id(index=es_index)\n\n description, data = self._read_data(description_path, data_path)\n if not data and query_data_for_indexing:\n try:\n data = Utils.materialize(metadata=description)\n except:\n traceback.print_exc()\n warnings.warn(\"Materialization Failed, index based on schema json only\")\n\n metadata = self.construct_global_metadata(description=description, data=data)\n Utils.validate_schema(metadata.value)\n\n if save_to_file:\n self._save_data(save_to_file=save_to_file, save_mode=save_to_file_mode, metadata=metadata)\n\n self.im.create_doc(index=es_index, doc_type='_doc', body=metadata.value, id=metadata.value['datamart_id'])\n\n return metadata.value\n\n def updating(self,\n description_path: str,\n es_index: str,\n document_id: int,\n data_path: str = None,\n query_data_for_updating: bool = False\n ) -> dict:\n\n \"\"\"Update document in elastic search.\n\n By providing description file, index builder should be able to process it and create metadata json for the\n dataset, update document in elastic search\n\n Args:\n description_path: Path to description json file.\n es_index: str, es index for this dataset\n document_id: int, document id of document which need to be updated\n data_path: Path to data csv file.\n query_data_for_updating: Bool. If no data is presented, and query_data_for_updating is False, will only\n create metadata according to the description json. If query_data_for_updating is True and no data is\n presented, will use Materialize to query data for profiling and indexing\n\n Returns:\n metadata dictionary\n\n \"\"\"\n\n self._check_es_index(es_index=es_index)\n\n description, data = self._read_data(description_path, data_path)\n if not data and query_data_for_updating:\n try:\n materializer_module = description[\"materialization\"][\"python_path\"]\n materializer = Utils.load_materializer(materializer_module)\n data = materializer.get(metadata=description)\n except:\n warnings.warn(\"Materialization Failed, index based on schema json only\")\n\n metadata = self.construct_global_metadata(description=description, data=data, overwrite_datamart_id=document_id)\n Utils.validate_schema(metadata.value)\n\n self.im.update_doc(index=es_index, doc_type='document', body={\"doc\": metadata.value},\n id=metadata.value['datamart_id'])\n\n return metadata.value\n\n def bulk_indexing(self,\n description_dir: str,\n es_index: str,\n data_dir: str = None,\n query_data_for_indexing: bool = False,\n save_to_file: str = None,\n save_to_file_mode: str = \"a+\",\n delete_old_es_index: bool = False\n ) -> None:\n \"\"\"Bulk indexing many dataset by providing a path\n\n Args:\n description_dir: dir of description json files.\n es_index: str, es index for this dataset\n data_dir: dir of data csv files.\n query_data_for_indexing: Bool. If no data is presented, and query_data_for_indexing is False, will only\n create metadata according to the description json. If query_data_for_indexing is True and no data is\n presented, will use Materialize to query data for profiling and indexing\n save_to_file: str, a path to the json line file\n save_to_file_mode: str, mode for saving, default \"a+\"\n delete_old_es_index: bool, boolean if delete original es index if it exist\n\n Returns:\n\n \"\"\"\n\n self._check_es_index(es_index=es_index, delete_old_es_index=delete_old_es_index)\n for description in os.listdir(description_dir):\n if description.endswith('.json'):\n description_path = os.path.join(description_dir, description)\n data_path = None\n if data_dir:\n data_path = os.path.join(data_dir, description.replace(\"_description.json\", \".csv\"))\n print(\"==== Creating metadata and indexing for \" + description)\n self.indexing(description_path=description_path,\n es_index=es_index,\n data_path=data_path,\n query_data_for_indexing=query_data_for_indexing,\n save_to_file=save_to_file,\n save_to_file_mode=save_to_file_mode)\n\n def _check_es_index(self, es_index: str, delete_old_es_index: bool = False) -> None:\n \"\"\"Check es index, delete or create if necessary\n\n Args:\n es_index: str, es index for this dataset\n delete_old_es_index: bool, boolean if delete original es index if it exist\n\n Returns:\n\n \"\"\"\n\n if not self.im.check_exists(index=es_index):\n self.im.create_index(index=es_index)\n elif delete_old_es_index:\n self.im.delete_index(index=[es_index])\n self.im.create_index(index=es_index)\n\n @staticmethod\n def _read_data(description_path: str, data_path: str = None) -> typing.Tuple[dict, pd.DataFrame]:\n \"\"\"Read dataset description json and dataset if present.\n\n Args:\n description_path: Path to description json file.\n data_path: Path to data csv file.\n\n Returns:\n Tuple of (description json, dataframe of data)\n \"\"\"\n\n description = json.load(open(description_path, 'r'))\n Utils.validate_schema(description)\n if data_path:\n data = pd.read_csv(open(data_path), 'r')\n else:\n data = None\n return description, data\n\n @staticmethod\n def _save_data(save_to_file: str, save_mode: str, metadata: GlobalMetadata) -> None:\n \"\"\"Save metadata json to file.\n\n Args:\n save_to_file: Path of the saving file.\n save_mode: save mode\n metadata: metadata instance.\n\n Returns:\n save to file with 2 lines for each metadata, first line is id, second line is metadata json\n \"\"\"\n\n with open(save_to_file, mode=save_mode) as out:\n out.write(str(metadata.datamart_id))\n out.write(\"\\n\")\n out.write(json.dumps(metadata.value))\n out.write(\"\\n\")\n\n def construct_global_metadata(self,\n description: dict,\n data: pd.DataFrame = None,\n overwrite_datamart_id: int = None\n ) -> GlobalMetadata:\n\n \"\"\"Construct global metadata.\n\n Args:\n description: description dict.\n data: dataframe of data.\n overwrite_datamart_id: integer id for over writing original one\n\n Returns:\n GlobalMetadata instance\n \"\"\"\n if not overwrite_datamart_id:\n self.current_global_index += self.GLOBAL_INDEX_INTERVAL\n datamart_id = self.current_global_index\n else:\n datamart_id = overwrite_datamart_id\n\n global_metadata = GlobalMetadata.construct_global(description, datamart_id=datamart_id)\n\n if data is not None:\n global_metadata = self._profiling_entire(global_metadata, data)\n\n if description.get(\"variables\", []):\n for col_offset, variable_description in enumerate(description[\"variables\"]):\n variable_metadata = self.construct_variable_metadata(description=variable_description,\n global_datamart_id=datamart_id,\n col_offset=col_offset,\n data=data)\n global_metadata.add_variable_metadata(variable_metadata)\n\n elif data is not None:\n for col_offset in range(data.shape[1]):\n variable_metadata = self.construct_variable_metadata(description={},\n global_datamart_id=datamart_id,\n col_offset=col_offset,\n data=data)\n global_metadata.add_variable_metadata(variable_metadata)\n\n else:\n warnings.warn(\n \"No data to profile for variable metadata. No variable description. Leave empty for variable metadata\")\n\n return global_metadata\n\n def construct_variable_metadata(self,\n description: dict,\n global_datamart_id: int,\n col_offset: int,\n data: pd.DataFrame = None\n ) -> VariableMetadata:\n\n \"\"\"Construct variable metadata.\n\n Args:\n description: description dict.\n global_datamart_id: integer of datamart id.\n col_offset: integer, the column index.\n data: dataframe of data.\n\n Returns:\n VariableMetadata instance\n \"\"\"\n\n variable_metadata = VariableMetadata.construct_variable(description,\n datamart_id=col_offset + global_datamart_id + 1)\n\n if data is not None:\n variable_metadata = self._profiling_column(description, variable_metadata, data.iloc[:, col_offset])\n\n return variable_metadata\n\n def _profiling_column(self,\n description: dict,\n variable_metadata: VariableMetadata,\n column: pd.Series\n ) -> VariableMetadata:\n \"\"\"Profiling single column for necessary fields of metadata, if data is present .\n\n Args:\n description: description dict about the column.\n variable_metadata: the original VariableMetadata instance.\n column: the column to profile.\n\n Returns:\n profiled VariableMetadata instance\n \"\"\"\n\n if not variable_metadata.name:\n variable_metadata.name = column.name\n\n if not variable_metadata.description:\n variable_metadata.description = self.basic_profiler.construct_variable_description(column)\n\n if variable_metadata.named_entity is None:\n variable_metadata.named_entity = self.basic_profiler.profile_named_entity(column)\n elif variable_metadata.named_entity is False and not description:\n named_entities = self.basic_profiler.named_entity_recognize(column)\n if named_entities:\n variable_metadata.named_entity = named_entities\n\n if variable_metadata.temporal_coverage is not False:\n if not variable_metadata.temporal_coverage['start'] or not variable_metadata.temporal_coverage['end']:\n variable_metadata.temporal_coverage = self.basic_profiler.profile_temporal_coverage(\n column=column, coverage=variable_metadata.temporal_coverage)\n\n elif not description:\n temporal_coverage = self.basic_profiler.profile_temporal_coverage(column=column)\n if temporal_coverage:\n variable_metadata.temporal_coverage = temporal_coverage\n\n return variable_metadata\n\n def _profiling_entire(self, global_metadata: GlobalMetadata, data: pd.DataFrame) -> GlobalMetadata:\n \"\"\"Profiling entire dataset for necessary fields of metadata, if data is present .\n\n Args:\n global_metadata: the original GlobalMetadata instance.\n data: dataframe of data.\n\n Returns:\n profiled GlobalMetadata instance\n \"\"\"\n\n if not global_metadata.title:\n global_metadata.title = self.basic_profiler.construct_global_title(data)\n\n if not global_metadata.description:\n global_metadata.description = self.basic_profiler.construct_global_description(data)\n\n if not global_metadata.keywords:\n global_metadata.keywords = self.basic_profiler.construct_global_keywords(data)\n\n return global_metadata\n\n def _bulk_load_metadata(self,\n metadata_out_file: str,\n es_index: str\n ) -> None:\n \"\"\"Internal method for bulk loading documents to elasticsearch.\n\n Args:\n metadata_out_file: file of metadata output file produced by index builder\n es_index: str of es index\n\n Returns:\n\n \"\"\"\n\n self.im.create_doc_bulk(file=metadata_out_file, index=es_index)\n", "sub_path": "datamart/index_builder.py", "file_name": "index_builder.py", "file_ext": "py", "file_size_in_byte": 15696, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 24, "usage_type": "call"}, {"api_name": "datamart.profilers.basic_profiler.BasicProfiler", "line_number": 27, "usage_type": "name"}, {"api_name": "datamart.es_managers.index_manager.IndexManager", "line_number": 28, "usage_type": "call"}, {"api_name": "datamart.utils.Utils.materialize", "line_number": 68, "usage_type": "call"}, {"api_name": "datamart.utils.Utils", "line_number": 68, "usage_type": "name"}, {"api_name": "traceback.print_exc", "line_number": 70, "usage_type": "call"}, {"api_name": "warnings.warn", "line_number": 71, "usage_type": "call"}, {"api_name": "datamart.utils.Utils.validate_schema", "line_number": 74, "usage_type": "call"}, {"api_name": "datamart.utils.Utils", "line_number": 74, "usage_type": "name"}, {"api_name": "datamart.utils.Utils.load_materializer", "line_number": 116, "usage_type": "call"}, {"api_name": "datamart.utils.Utils", "line_number": 116, "usage_type": "name"}, {"api_name": "warnings.warn", "line_number": 119, "usage_type": "call"}, {"api_name": "datamart.utils.Utils.validate_schema", "line_number": 122, "usage_type": "call"}, {"api_name": "datamart.utils.Utils", "line_number": 122, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 156, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 158, "usage_type": "call"}, {"api_name": "os.path", "line_number": 158, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 161, "usage_type": "call"}, {"api_name": "os.path", "line_number": 161, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 199, "usage_type": "call"}, {"api_name": "datamart.utils.Utils.validate_schema", "line_number": 200, "usage_type": "call"}, {"api_name": "datamart.utils.Utils", "line_number": 200, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 202, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 188, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 188, "usage_type": "attribute"}, {"api_name": "datamart.metadata.global_metadata.GlobalMetadata", "line_number": 208, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 223, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 228, "usage_type": "attribute"}, {"api_name": "datamart.metadata.global_metadata.GlobalMetadata.construct_global", "line_number": 248, "usage_type": "call"}, {"api_name": "datamart.metadata.global_metadata.GlobalMetadata", "line_number": 248, "usage_type": "name"}, {"api_name": "warnings.warn", "line_number": 270, "usage_type": "call"}, {"api_name": "datamart.metadata.global_metadata.GlobalMetadata", "line_number": 230, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 279, "usage_type": "attribute"}, {"api_name": "datamart.metadata.variable_metadata.VariableMetadata.construct_variable", "line_number": 294, "usage_type": "call"}, {"api_name": "datamart.metadata.variable_metadata.VariableMetadata", "line_number": 294, "usage_type": "name"}, {"api_name": "datamart.metadata.variable_metadata.VariableMetadata", "line_number": 280, "usage_type": "name"}, {"api_name": "datamart.metadata.variable_metadata.VariableMetadata", "line_number": 304, "usage_type": "name"}, {"api_name": "pandas.Series", "line_number": 305, "usage_type": "attribute"}, {"api_name": "datamart.metadata.variable_metadata.VariableMetadata", "line_number": 306, "usage_type": "name"}, {"api_name": "datamart.metadata.global_metadata.GlobalMetadata", "line_number": 343, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 343, "usage_type": "attribute"}]}
+{"seq_id": "5527136", "text": "import argparse\n\nfrom sqlalchemy import desc, func\nfrom database_code.models.models import LogInfo, RequestUrl, RequestIPAddress\nfrom database_code.mysql_client.client import MySQLConnection\nfrom database_code.tests.builder import MySQLBuilder\nfrom python_scripts.exceptions import ParamsException, CommandNotFoundException\n\n\ndef mysql_client(rebuild_db):\n return MySQLConnection('root', 'pass', 'nginx_logs', rebuild_db=rebuild_db)\n\n\nclass LogProcessing:\n\n def __init__(self, log_path, rebuild_db=True):\n self.mysql: MySQLConnection = mysql_client(rebuild_db)\n self.builder = MySQLBuilder(self.mysql, rebuild_db=rebuild_db)\n if rebuild_db:\n self.builder.fill_database(log_path)\n\n def run_cmd1(self):\n print(f'{len(self.mysql.session.query(LogInfo).all())} lines in log file')\n\n def run_cmd2(self, req_type):\n if not req_type:\n raise ParamsException('Invalid count of parameters!')\n print(f'Count of {req_type} requests is equal {len(self.mysql.session.query(LogInfo).filter_by(req_method=req_type).all())}')\n\n def run_cmd3(self):\n query = self.mysql.session.query(LogInfo).order_by(desc(LogInfo.req_size)).limit(10)\n req_urls = list(map(lambda req_url: req_url.url, self.mysql.session.query(RequestUrl).all()))\n req_ip_addresses = list(map(lambda req_id_address: req_id_address.ip, self.mysql.session.query(RequestIPAddress)))\n for line in query:\n print(f'{req_ip_addresses[line.req_ip_address_id - 1]} {req_urls[line.req_url_id - 1]} {line.req_method} {line.response_code} {line.req_size}')\n\n def run_cmd4(self):\n query = self.mysql.session.query(LogInfo.req_url_id, LogInfo.response_code, func.count()).filter(LogInfo.response_code.like('4%')).order_by(desc(func.count())).group_by(LogInfo.req_url_id, LogInfo.response_code).all()\n req_urls = list(map(lambda req_url: req_url.url, self.mysql.session.query(RequestUrl).all()))\n for line in query:\n print(f'{req_urls[line[0] - 1]} status_code:{line[1]} is repeated {line[2]} times')\n\n def run_cmd5(self):\n query = self.mysql.session.query(LogInfo.req_ip_address_id, LogInfo.req_url_id, LogInfo.response_code, LogInfo.req_size).filter(LogInfo.response_code.like('4%')).order_by(desc(LogInfo.req_size)).distinct().limit(10).all()\n req_urls = list(map(lambda req_url: req_url.url, self.mysql.session.query(RequestUrl).all()))\n req_ip_addresses = list(map(lambda req_id_address: req_id_address.ip, self.mysql.session.query(RequestIPAddress)))\n for line in query:\n print(f'{req_ip_addresses[line[0] - 1]} {req_urls[line[1] - 1]} status_code:{line[2]} {line[3]}')\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument('--log_path', type=str, dest='log_path')\n parser.add_argument('--cmd', type=int, dest='cmd')\n parser.add_argument('--req_type', type=str, dest='req_type')\n\n args = parser.parse_args()\n\n log_p = LogProcessing(args.log_path, rebuild_db=False)\n if args.cmd == 1:\n log_p.run_cmd1()\n elif args.cmd == 2:\n log_p.run_cmd2(args.req_type)\n elif args.cmd == 3:\n log_p.run_cmd3()\n elif args.cmd == 4:\n log_p.run_cmd4()\n elif args.cmd == 5:\n log_p.run_cmd5()\n else:\n raise CommandNotFoundException('Command not found!')", "sub_path": "hometask3/python_scripts/sql_log_processing.py", "file_name": "sql_log_processing.py", "file_ext": "py", "file_size_in_byte": 3374, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "database_code.mysql_client.client.MySQLConnection", "line_number": 11, "usage_type": "call"}, {"api_name": "database_code.mysql_client.client.MySQLConnection", "line_number": 17, "usage_type": "name"}, {"api_name": "database_code.tests.builder.MySQLBuilder", "line_number": 18, "usage_type": "call"}, {"api_name": "database_code.models.models.LogInfo", "line_number": 23, "usage_type": "argument"}, {"api_name": "python_scripts.exceptions.ParamsException", "line_number": 27, "usage_type": "call"}, {"api_name": "database_code.models.models.LogInfo", "line_number": 28, "usage_type": "argument"}, {"api_name": "database_code.models.models.LogInfo", "line_number": 31, "usage_type": "argument"}, {"api_name": "sqlalchemy.desc", "line_number": 31, "usage_type": "call"}, {"api_name": "database_code.models.models.LogInfo.req_size", "line_number": 31, "usage_type": "attribute"}, {"api_name": "database_code.models.models.RequestUrl", "line_number": 32, "usage_type": "argument"}, {"api_name": "database_code.models.models.RequestIPAddress", "line_number": 33, "usage_type": "argument"}, {"api_name": "database_code.models.models.LogInfo.req_url_id", "line_number": 38, "usage_type": "attribute"}, {"api_name": "database_code.models.models.LogInfo", "line_number": 38, "usage_type": "name"}, {"api_name": "database_code.models.models.LogInfo.response_code", "line_number": 38, "usage_type": "attribute"}, {"api_name": "sqlalchemy.func.count", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.func", "line_number": 38, "usage_type": "name"}, {"api_name": "database_code.models.models.LogInfo.response_code.like", "line_number": 38, "usage_type": "call"}, {"api_name": "sqlalchemy.desc", "line_number": 38, "usage_type": "call"}, {"api_name": "database_code.models.models.RequestUrl", "line_number": 39, "usage_type": "argument"}, {"api_name": "database_code.models.models.LogInfo.req_ip_address_id", "line_number": 44, "usage_type": "attribute"}, {"api_name": "database_code.models.models.LogInfo", "line_number": 44, "usage_type": "name"}, {"api_name": "database_code.models.models.LogInfo.req_url_id", "line_number": 44, "usage_type": "attribute"}, {"api_name": "database_code.models.models.LogInfo.response_code", "line_number": 44, "usage_type": "attribute"}, {"api_name": "database_code.models.models.LogInfo.req_size", "line_number": 44, "usage_type": "attribute"}, {"api_name": "database_code.models.models.LogInfo.response_code.like", "line_number": 44, "usage_type": "call"}, {"api_name": "sqlalchemy.desc", "line_number": 44, "usage_type": "call"}, {"api_name": "database_code.models.models.RequestUrl", "line_number": 45, "usage_type": "argument"}, {"api_name": "database_code.models.models.RequestIPAddress", "line_number": 46, "usage_type": "argument"}, {"api_name": "argparse.ArgumentParser", "line_number": 51, "usage_type": "call"}, {"api_name": "python_scripts.exceptions.CommandNotFoundException", "line_number": 70, "usage_type": "call"}]}
+{"seq_id": "99217454", "text": "import pygame as pg\nimport random\nfrom pygame.locals import *\n\n# essa função cria um ponto coordenado entre 0 e 59. \n# Usaremos o truque do //10*10 para o arredondamento.\n# como a tela terá 600 pixels, vamos até o tile 59 porque cada tile tem 10 pixels.\ndef lugarRandom():\n x = random.randint(0,590)\n y = random.randint(0,590)\n return (x//10 * 10, y//10 * 10)\n\n# detecta colisões meramente comparando as tuplas.\ndef detectaColisao(c1, c2):\n return (c1[0] == c2[0]) and (c1[1] == c2[1])\n\n#TODO: fazer uma consequência de colisão melhor.\ndef reInicializa():\n pg.quit()\n\n# essas serão as direções que a \"cobrinha\" pode se mover.\nUP = 0\nDOWN = 1\nLEFT = 2\nRIGHT = 3\n\n# inicializando a tela vazia.\npg.init()\ntela = pg.display.set_mode((600,600))\npg.display.set_caption(\"Snake Game no Python\")\n\n# inicializa o relógio do jogo.\nrelogio = pg.time.Clock()\n\n# cria a cobrinha, com três pedaços no corpo, cada um em uma tupla coordenada adjacente \n# snake skin, o tile que compõe a cobrinha, será um quadrado (surface, do pygame), \n# de 10 por 10 e preenchido com branco. Em cada tupla coordenada esse quadrado será desenhado.\ncobra = [(200, 200), (210, 200), (220, 200)]\ncobra_skin = pg.Surface((10, 10))\ncobra_skin.fill((255,255,255))\n\n# cria a maçã, no mesmo esquema\nmaca_pos = lugarRandom()\nmaca = pg.Surface((10,10))\nmaca.fill((255,0,0))\n\n# define uma direção inicial\ndirecao = RIGHT\n\nwhile True:\n # o tick modera a velocidade.\n velocidade = 20 \n relogio.tick(velocidade)\n \n # classificador de eventos.\n for event in pg.event.get():\n\n # quit do gerenciador.\n if event.type == QUIT:\n pg.quit()\n \n # detecta os eventos de pressionar de teclas.\n if event.type == KEYDOWN:\n\n # controle direcional\n if event.key == K_UP:\n direcao = UP\n if event.key == K_DOWN:\n direcao = DOWN\n if event.key == K_LEFT:\n direcao = LEFT\n if event.key == K_RIGHT:\n direcao = RIGHT\n \n # dá quit pelo teclado\n if event.key == K_ESCAPE: \n pg.quit()\n \n # esse evento coloca um segmento no final da cobra, quando ela pega a maçã. \n if detectaColisao(cobra[0], maca_pos):\n maca_pos = lugarRandom()\n cobra.append((0,0))\n\n #detecta colisão com o fim da tela\n if cobra[0][0]<0 or cobra[0][0]>590 or cobra[0][1]<0 or cobra[0][1]>590:\n reInicializa() \n\n # muda a direção de movimento da cobra baseado no evento de input.\n if direcao == UP:\n cobra[0] = (cobra[0][0], cobra[0][1] - 10)\n if direcao == DOWN:\n cobra[0] = (cobra[0][0], cobra[0][1] + 10)\n if direcao == LEFT:\n cobra[0] = (cobra[0][0] -10, cobra[0][1])\n if direcao == RIGHT:\n cobra[0] = (cobra[0][0] + 10, cobra[0][1])\n\n # isso faz a cobra \"andar\": dá, a cada i da cobra, a posição do i anterior.\n for i in range(len(cobra) -1, 0, -1):\n cobra[i] = (cobra[i-1][0], cobra[i-1][1])\n\n # renderiza tudo de novo, menos a cobra.\n tela.fill((0,0,0))\n tela.blit(maca, maca_pos)\n\n # renderiza a cobra, segmento por segmento.\n for pos in cobra:\n tela.blit(cobra_skin, pos)\n tela\n\n pg.display.update()", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3330, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "random.randint", "line_number": 9, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 10, "usage_type": "call"}, {"api_name": "pygame.quit", "line_number": 19, "usage_type": "call"}, {"api_name": "pygame.init", "line_number": 28, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 29, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 29, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 30, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 30, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 33, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 33, "usage_type": "attribute"}, {"api_name": "pygame.Surface", "line_number": 39, "usage_type": "call"}, {"api_name": "pygame.Surface", "line_number": 44, "usage_type": "call"}, {"api_name": "pygame.event.get", "line_number": 56, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 56, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 60, "usage_type": "call"}, {"api_name": "pygame.quit", "line_number": 77, "usage_type": "call"}, {"api_name": "pygame.display.update", "line_number": 111, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 111, "usage_type": "attribute"}]}
+{"seq_id": "606533556", "text": "from __future__ import division\nimport torch\nimport torch.nn.functional as F\nfrom utils import ensure_shared_grads\nfrom model import agentNET\nfrom torch.autograd import Variable\nfrom env import *\nimport random\n\nS_INFO = 6 # bit_rate, buffer_size, next_chunk_size, bandwidth_measurement(throughput and time), chunk_til_video_end\nS_LEN = 8 # take how many frames in the past\nA_DIM = 6\nTRAIN_SEQ_LEN = 100 # take as a train batch\nMODEL_SAVE_INTERVAL = 100\nVIDEO_BIT_RATE = [300,750,1200,1850,2850,4300] # Kbps\nHD_REWARD = [1, 2, 3, 12, 15, 20]\nBUFFER_NORM_FACTOR = 10.0\nCHUNK_TIL_VIDEO_END_CAP = 48.0\nM_IN_K = 1000.0\nREBUF_PENALTY = 4.3 # 1 sec rebuffering -> 3 Mbps\nSMOOTH_PENALTY = 1\nDEFAULT_QUALITY = 0 # default video quality without agent\n\ndef train(rank, args, shared_model, optimizer, all_cooked_time, all_cooked_bw):\n torch.manual_seed(args.seed + rank)\n env = Environment(all_cooked_time=all_cooked_time,\n all_cooked_bw=all_cooked_bw,\n random_seed=rank\n )\n\n model = agentNET()\n model.train()\n\n time_stamp = 0\n uploadtime = 0\n end_of_video = True\n last_bit_rate = DEFAULT_QUALITY\n bit_rate = DEFAULT_QUALITY\n entropy_weights = [5, 1, 1, 0.5, 0.5, 0.1, 0.1, 0.1, 0.01] + [0.01] * 1000\n entropy_weight = 0.01 # default entropy weight\n\n while True:\n model.load_state_dict(shared_model.state_dict())\n\n if args.gpu:\n model = model.cuda()\n cx = Variable(torch.zeros(1, 96).cuda())\n hx = Variable(torch.zeros(1, 96).cuda())\n else:\n cx = Variable(torch.zeros(1, 96))\n hx = Variable(torch.zeros(1, 96))\n\n state = np.zeros([S_INFO, S_LEN])\n for i in range(S_LEN):\n # do an default action\n bit_rate = random.randint(0,5)\n delay, sleep_time, buffer_size, rebuf, \\\n video_chunk_size, next_video_chunk_sizes, \\\n end_of_video, video_chunk_remain = \\\n env.get_video_chunk(bit_rate)\n\n time_stamp += delay # in ms\n time_stamp += sleep_time # in ms\n\n # get new state\n state[0][i] = VIDEO_BIT_RATE[last_bit_rate] / float(np.max(VIDEO_BIT_RATE)) # last quality\n state[1][i] = buffer_size / BUFFER_NORM_FACTOR # 10 sec\n state[2][i] = float(video_chunk_size) / float(delay) / M_IN_K # kilo byte / ms\n state[3][i] = float(delay) / M_IN_K / BUFFER_NORM_FACTOR # 10 sec\n state[4][i] = (np.array(next_video_chunk_sizes) / M_IN_K / M_IN_K)[DEFAULT_QUALITY] # mega byte\n state[5][i] = min(video_chunk_remain, CHUNK_TIL_VIDEO_END_CAP) / float(CHUNK_TIL_VIDEO_END_CAP)\n\n last_bit_rate = bit_rate\n state = torch.from_numpy(np.array([state, ])).float()\n\n values = []\n log_probs = []\n rewards = []\n entropies = []\n\n entropy_weight = 20000 / (uploadtime + 4000)\n # entropy_weight = 0.05\n # entropy_weight = entropy_weights[int(uploadtime // 1000)]\n\n while True:\n if args.gpu:\n value, logit, (hx, cx) = model((Variable(state.unsqueeze(0)).cuda(), (hx, cx)))\n else:\n value, logit, (hx, cx) = model((Variable(state.unsqueeze(0)), (hx, cx)))\n\n prob = F.softmax(logit)\n log_prob = F.log_softmax(logit)\n entropy = -(log_prob * prob).sum(1)\n entropies.append(entropy)\n\n if args.gpu:\n action = prob.multinomial().data.cpu()\n action.view(-1, 1)\n log_prob = log_prob.gather(1, Variable(action.cuda()))\n else:\n action = prob.multinomial().data\n action.view(-1, 1)\n log_prob = log_prob.gather(1, Variable(action))\n\n bit_rate = action.numpy()[0][0]\n\n # do an action\n delay, sleep_time, buffer_size, rebuf, \\\n video_chunk_size, next_video_chunk_sizes, \\\n end_of_video, video_chunk_remain = \\\n env.get_video_chunk(bit_rate)\n\n\n time_stamp += delay # in ms\n time_stamp += sleep_time # in ms\n\n # -- linear reward --\n # reward is video quality - rebuffer penalty - smoothness\n reward = VIDEO_BIT_RATE[bit_rate] / M_IN_K \\\n - REBUF_PENALTY * rebuf \\\n - SMOOTH_PENALTY * np.abs(VIDEO_BIT_RATE[bit_rate] -\n VIDEO_BIT_RATE[last_bit_rate]) / M_IN_K\n\n # get new state\n for i in range(S_INFO):\n for j in range(S_LEN - 1):\n state[0][i][j] = state[0][i][j + 1]\n # state = np.zeros(S_INFO)\n state[0][0][S_LEN - 1] = VIDEO_BIT_RATE[last_bit_rate] / float(np.max(VIDEO_BIT_RATE)) # last quality\n state[0][1][S_LEN - 1] = buffer_size / BUFFER_NORM_FACTOR # 10 sec\n state[0][2][S_LEN - 1] = float(video_chunk_size) / float(delay) / M_IN_K # kilo byte / ms\n state[0][3][S_LEN - 1] = float(delay) / M_IN_K / BUFFER_NORM_FACTOR # 10 sec\n state[0][4][S_LEN - 1] = (np.array(next_video_chunk_sizes) / M_IN_K / M_IN_K)[action.numpy()[0][0]] # mega byte\n state[0][5][S_LEN - 1] = min(video_chunk_remain, CHUNK_TIL_VIDEO_END_CAP) / float(CHUNK_TIL_VIDEO_END_CAP)\n # state = torch.from_numpy(np.array([state, ])).float()\n last_bit_rate = bit_rate\n\n values.append(value)\n log_probs.append(log_prob)\n rewards.append(reward)\n\n if end_of_video:\n last_bit_rate = DEFAULT_QUALITY\n bit_rate = DEFAULT_QUALITY\n break\n\n # update the network\n R = torch.zeros(1, 1)\n if not end_of_video:\n if args.gpu:\n value, _, _ = model((Variable(state.unsqueeze(0).cuda()), (hx, cx)))\n else:\n value, _, _ = model((Variable(state.unsqueeze(0)), (hx, cx)))\n R = value.data\n\n if args.gpu:\n values.append(Variable(R.cuda()))\n R = Variable(R.cuda())\n else:\n values.append(Variable(R))\n R = Variable(R)\n\n policy_loss = 0\n value_loss = 0\n gae = torch.zeros(1, 1)\n\n for i in reversed(range(len(rewards))):\n R = args.gamma * R + rewards[i]\n advantage = R - values[i]\n value_loss = value_loss + 0.5 * advantage.pow(2)\n\n if args.gpu:\n delta_t = rewards[i] + args.gamma * \\\n values[i + 1].data.cpu() - values[i].data.cpu()\n else:\n delta_t = rewards[i] + args.gamma * \\\n values[i + 1].data - values[i].data\n\n gae = gae * args.gamma * args.tau + delta_t\n\n if args.gpu:\n policy_loss = policy_loss - \\\n log_probs[i] * Variable(gae.cuda()) - entropy_weight * entropies[i]\n else:\n policy_loss = policy_loss - \\\n log_probs[i] * Variable(gae) - entropy_weight * entropies[i]\n\n optimizer.zero_grad()\n (policy_loss + 0.5 * value_loss).backward()\n\n if args.gpu:\n model = model.cpu()\n\n torch.nn.utils.clip_grad_norm(model.parameters(), 40)\n ensure_shared_grads(model, shared_model)\n optimizer.step()\n\n uploadtime += 1\n if uploadtime % 1000 == 0 and rank == 1:\n print('---> after {0} steps <---'.format(uploadtime * args.workers))", "sub_path": "train/train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 7649, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "torch.manual_seed", "line_number": 25, "usage_type": "call"}, {"api_name": "model.agentNET", "line_number": 31, "usage_type": "call"}, {"api_name": "model.train", "line_number": 32, "usage_type": "call"}, {"api_name": "model.load_state_dict", "line_number": 43, "usage_type": "call"}, {"api_name": "model.cuda", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 51, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 56, "usage_type": "call"}, {"api_name": "env.get_video_chunk", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 89, "usage_type": "call"}, {"api_name": "torch.nn.functional.softmax", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 91, "usage_type": "name"}, {"api_name": "torch.nn.functional.log_softmax", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 92, "usage_type": "name"}, {"api_name": "torch.autograd.Variable", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 103, "usage_type": "call"}, {"api_name": "env.get_video_chunk", "line_number": 111, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 148, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 151, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 153, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 157, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 158, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 160, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 161, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 165, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 183, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 186, "usage_type": "call"}, {"api_name": "model.cpu", "line_number": 192, "usage_type": "call"}, {"api_name": "torch.nn.utils.clip_grad_norm", "line_number": 194, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 194, "usage_type": "attribute"}, {"api_name": "model.parameters", "line_number": 194, "usage_type": "call"}, {"api_name": "utils.ensure_shared_grads", "line_number": 195, "usage_type": "call"}]}
+{"seq_id": "368683408", "text": "import time\nimport threading\n\nfrom appium import webdriver\nfrom selenium.webdriver.common.by import By\n\nfrom api import API\nfrom Email import EMail\nfrom util import Util\n\ndesired_caps1 = {\n \"platformName\": \"android\",\n \"appPackage\": \"com.alibaba.android.rimet\",\n \"appActivity\": \".biz.SplashActivity\",\n \"deviceName\": \"honor\",\n \"noReset\": \"true\",\n \"udid\": \"UEUNW17221002815\"\n}\n\ndesired_caps2 = {\n \"platformName\": \"android\",\n \"appPackage\": \"com.alibaba.android.rimet\",\n \"appActivity\": \".biz.SplashActivity\",\n \"deviceName\": \"sumsung\",\n \"noReset\": \"true\",\n \"udid\": \"64de8478\"\n}\n\n\n# email = EMail()\n# body = \"\"\"\n# 测试图片
\n#
# 引用图片\n# \"\"\"\n\n\ndef task1():\n email = EMail()\n util1 = Util()\n util1.random_time(900)\n body = \"\"\"\n 测试图片
\n
# 引用图片\n \"\"\"\n\n try:\n driver = webdriver.Remote('http://localhost:4723/wd/hub', desired_caps1)\n driver.implicitly_wait(10)\n api = API(driver)\n element = api.isElement(By.ID, 'et_pwd_login')\n if element:\n driver.find_element(by=By.ID, value='et_pwd_login').send_keys('lihang1988')\n driver.find_element(by=By.ID, value='btn_next').click()\n\n time.sleep(3)\n driver.save_screenshot('UEUNW17221002815.png')\n email.send_mail('打卡成功', content='打卡截图', image_body=body, image='UEUNW17221002815.png')\n\n except:\n email.send_mail('打开失败了,赶紧找人吧!')\n\n time.sleep(2)\n driver.quit()\n\n\ndef task2():\n email = EMail()\n util1 = Util()\n util1.random_time(900)\n body = \"\"\"\n 测试图片
\n
# 引用图片\n \"\"\"\n\n try:\n driver = webdriver.Remote('http://localhost:4725/wd/hub', desired_caps2)\n driver.implicitly_wait(10)\n api = API(driver)\n element = api.isElement(By.ID, 'et_pwd_login')\n if element:\n driver.find_element(by=By.ID, value='et_pwd_login').send_keys('qetuo13579')\n driver.find_element(by=By.ID, value='btn_next').click()\n\n time.sleep(3)\n driver.save_screenshot('64de8478.png')\n email.send_mail('打卡成功', content='打卡截图', to_addrs_in='409803186@qq.com', image_body=body, image='64de8478.png')\n\n except:\n email.send_mail('打开失败了,赶紧找人吧!')\n\n time.sleep(2)\n driver.quit()\n\n\nthreads = []\nt1 = threading.Thread(target=task1)\nthreads.append(t1)\n\nt2 = threading.Thread(target=task2)\nthreads.append(t2)\n\nif __name__ == '__main__':\n for i in threads:\n i.start()\n", "sub_path": "case005.py", "file_name": "case005.py", "file_ext": "py", "file_size_in_byte": 2649, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "Email.EMail", "line_number": 38, "usage_type": "call"}, {"api_name": "util.Util", "line_number": 39, "usage_type": "call"}, {"api_name": "appium.webdriver.Remote", "line_number": 47, "usage_type": "call"}, {"api_name": "appium.webdriver", "line_number": 47, "usage_type": "name"}, {"api_name": "api.API", "line_number": 49, "usage_type": "call"}, {"api_name": "api.isElement", "line_number": 50, "usage_type": "call"}, {"api_name": "selenium.webdriver.common.by.By.ID", "line_number": 50, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 50, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.ID", "line_number": 52, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 52, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.ID", "line_number": 53, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 53, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 55, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 62, "usage_type": "call"}, {"api_name": "Email.EMail", "line_number": 67, "usage_type": "call"}, {"api_name": "util.Util", "line_number": 68, "usage_type": "call"}, {"api_name": "appium.webdriver.Remote", "line_number": 76, "usage_type": "call"}, {"api_name": "appium.webdriver", "line_number": 76, "usage_type": "name"}, {"api_name": "api.API", "line_number": 78, "usage_type": "call"}, {"api_name": "api.isElement", "line_number": 79, "usage_type": "call"}, {"api_name": "selenium.webdriver.common.by.By.ID", "line_number": 79, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 79, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.ID", "line_number": 81, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 81, "usage_type": "name"}, {"api_name": "selenium.webdriver.common.by.By.ID", "line_number": 82, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.common.by.By", "line_number": 82, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 84, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 91, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 96, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 99, "usage_type": "call"}]}
+{"seq_id": "248150729", "text": "import itertools\ndef threeSum(list):\n res=[]\n list.sort()\n length=len(list)\n print(list)\n for i in range(length-2):\n if i>0 and list[i]==list[i-1]:\n continue\n l=i+1\n r=length-1\n while l0:\n r=r-1\n else:\n res.append([list[i],list[l],list[r]])\n while l\",end=\" \")\n for j in self.graph[i]:\n print(j,end=\",\")\n print(\"\")\n \n def checkRingTopology(self):\n flag = True\n for i in range(self.vertices):\n degree = len(self.graph[i])\n if self.vertices < 3 :\n flag = False\n break\n if self.totalEdges != self.vertices:\n flag = False\n break\n if degree !=2 :\n flag = False\n break\n return flag\n\nprint(\"Enter the total number of vertices \")\nn = int(input())\ng = Graph(n)\nprint(\"Enter the total number of edges\")\nedge = int(input())\nfor i in range(edge):\n u,v = map(int,input().split())\n g.addEdge(u,v)\ng.setIdentifier()\ng.showGraph()\nflag = g.checkRingTopology()\nif flag:\n print(\"yes\")\nelse:\n print(\"no\")", "sub_path": "graphTheory/CheckRingTopology.py", "file_name": "CheckRingTopology.py", "file_ext": "py", "file_size_in_byte": 1423, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "collections.defaultdict", "line_number": 13, "usage_type": "call"}]}
+{"seq_id": "482410778", "text": "\"\"\"Functions to query for statistics in the database.\"\"\"\n\nimport sqlite3\n\n__authors__ = [\n 'Hayden Phothong, Research Anslyst Intern @ ITD'\n]\n\n\nclass ImageRecognitionStats:\n \"\"\"Represent the statistics of the image recognition database.\"\"\"\n\n @classmethod\n def __get_cursor(cls):\n \"\"\"Return the cursor to the ImageRecognition database.\"\"\"\n cls.connection = sqlite3.connect('../../ImageRecognition.db')\n return cls.connection.cursor()\n\n @classmethod\n def completed_images_count(cls, field):\n \"\"\"Return the number of completed images.\"\"\"\n cursor = cls.__get_cursor()\n try:\n # Load assets\n cursor.execute('SELECT asset_name FROM Assets;')\n asset_names = []\n for row in cursor:\n asset_names.append(row[0].lower())\n\n # Query number of completed images\n for asset_name in asset_names:\n if field.lower() in asset_name:\n query = (\n \"\"\"\n SELECT COUNT(*) FROM CompletedImages\n WHERE {}_present = 1 OR {}_present = 0;\n \"\"\".format(field, field)\n )\n break\n cursor.execute(query)\n completed_images_count = cursor.fetchone()[0]\n\n cls.connection.close()\n\n return completed_images_count\n\n except Exception as e:\n return e\n\n @classmethod\n def percent_images_completed(cls, field):\n \"\"\"Return the percent of completed images from the total.\"\"\"\n try:\n cursor = cls.__get_cursor()\n\n # Load number of total images\n cursor.execute('SELECT COUNT(*) FROM Images;')\n total_images = cursor.fetchone()[0]\n\n # Get completed images count\n completed_images_count = cls.completed_images_count(field)\n\n percent = float(completed_images_count) / total_images\n percent *= 100\n percent = round(percent, 2)\n\n cls.connection.close()\n\n return percent\n\n except Exception as e:\n return e\n", "sub_path": "ImageRecognition/Code/classifier/applications/image_recognition_stats.py", "file_name": "image_recognition_stats.py", "file_ext": "py", "file_size_in_byte": 2172, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sqlite3.connect", "line_number": 16, "usage_type": "call"}]}
+{"seq_id": "579283089", "text": "#\n# Copyright 2020 - Swiss Data Science Center (SDSC)\n# A partnership between École Polytechnique Fédérale de Lausanne (EPFL) and\n# Eidgenössische Technische Hochschule Zürich (ETHZ).\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"Renku service templates view.\"\"\"\nfrom flask import request\n\nfrom renku.ui.service.config import SERVICE_PREFIX\nfrom renku.ui.service.controllers.templates_create_project import TemplatesCreateProjectCtrl\nfrom renku.ui.service.controllers.templates_read_manifest import TemplatesReadManifestCtrl\nfrom renku.ui.service.views.api_versions import ALL_VERSIONS, V2_0, V2_1, VersionedBlueprint\nfrom renku.ui.service.views.decorators import accepts_json, requires_cache, requires_identity\nfrom renku.ui.service.views.error_handlers import (\n handle_common_except,\n handle_templates_create_errors,\n handle_templates_read_errors,\n)\nfrom renku.ui.service.views.v1.templates import add_v1_specific_endpoints\n\nTEMPLATES_BLUEPRINT_TAG = \"templates\"\ntemplates_blueprint = VersionedBlueprint(TEMPLATES_BLUEPRINT_TAG, __name__, url_prefix=SERVICE_PREFIX)\n\n\n@templates_blueprint.route(\n \"/templates.read_manifest\", methods=[\"GET\"], provide_automatic_options=False, versions=[V2_0, V2_1]\n)\n@handle_common_except\n@handle_templates_read_errors\n@requires_cache\n@requires_identity\ndef read_manifest_from_template(user_data, cache):\n \"\"\"\n Read templates from the manifest file of a template repository.\n\n ---\n get:\n description: Read templates from the manifest file of a template repository.\n parameters:\n - in: query\n name: url\n required: true\n schema:\n type: string\n - in: query\n name: ref\n schema:\n type: string\n - in: query\n name: depth\n schema:\n type: string\n responses:\n 200:\n description: Listing of templates in the repository.\n content:\n application/json:\n schema: ManifestTemplatesResponseRPC\n tags:\n - templates\n \"\"\"\n return TemplatesReadManifestCtrl(cache, user_data, dict(request.args)).to_response()\n\n\n@templates_blueprint.route(\n \"/templates.create_project\", methods=[\"POST\"], provide_automatic_options=False, versions=ALL_VERSIONS\n)\n@handle_common_except\n@handle_templates_create_errors\n@accepts_json\n@requires_cache\n@requires_identity\ndef create_project_from_template(user_data, cache):\n \"\"\"\n Create a new project starting using a remote template.\n\n ---\n post:\n description: Create a new project using a remote template.\n requestBody:\n content:\n application/json:\n schema: ProjectTemplateRequest\n responses:\n 200:\n description: Details of the created project.\n content:\n application/json:\n schema: ProjectTemplateResponseRPC\n tags:\n - templates\n \"\"\"\n return TemplatesCreateProjectCtrl(cache, user_data, dict(request.json)).to_response() # type: ignore\n\n\ntemplates_blueprint = add_v1_specific_endpoints(templates_blueprint)\n", "sub_path": "renku/ui/service/views/templates.py", "file_name": "templates.py", "file_ext": "py", "file_size_in_byte": 3604, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "renku.ui.service.views.api_versions.VersionedBlueprint", "line_number": 33, "usage_type": "call"}, {"api_name": "renku.ui.service.config.SERVICE_PREFIX", "line_number": 33, "usage_type": "name"}, {"api_name": "renku.ui.service.controllers.templates_read_manifest.TemplatesReadManifestCtrl", "line_number": 73, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 73, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 73, "usage_type": "name"}, {"api_name": "renku.ui.service.views.api_versions.V2_0", "line_number": 37, "usage_type": "name"}, {"api_name": "renku.ui.service.views.api_versions.V2_1", "line_number": 37, "usage_type": "name"}, {"api_name": "renku.ui.service.views.error_handlers.handle_common_except", "line_number": 39, "usage_type": "name"}, {"api_name": "renku.ui.service.views.error_handlers.handle_templates_read_errors", "line_number": 40, "usage_type": "name"}, {"api_name": "renku.ui.service.views.decorators.requires_cache", "line_number": 41, "usage_type": "name"}, {"api_name": "renku.ui.service.views.decorators.requires_identity", "line_number": 42, "usage_type": "name"}, {"api_name": "renku.ui.service.controllers.templates_create_project.TemplatesCreateProjectCtrl", "line_number": 104, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 104, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 104, "usage_type": "name"}, {"api_name": "renku.ui.service.views.api_versions.ALL_VERSIONS", "line_number": 77, "usage_type": "name"}, {"api_name": "renku.ui.service.views.error_handlers.handle_common_except", "line_number": 79, "usage_type": "name"}, {"api_name": "renku.ui.service.views.error_handlers.handle_templates_create_errors", "line_number": 80, "usage_type": "name"}, {"api_name": "renku.ui.service.views.decorators.accepts_json", "line_number": 81, "usage_type": "name"}, {"api_name": "renku.ui.service.views.decorators.requires_cache", "line_number": 82, "usage_type": "name"}, {"api_name": "renku.ui.service.views.decorators.requires_identity", "line_number": 83, "usage_type": "name"}, {"api_name": "renku.ui.service.views.v1.templates.add_v1_specific_endpoints", "line_number": 107, "usage_type": "call"}]}
+{"seq_id": "634031460", "text": "from __future__ import division, print_function, absolute_import\n\nfrom nnmnkwii.util import trim_zeros_frames\nfrom nnmnkwii.baseline.gmm import MLPG\n\nfrom fastdtw import fastdtw\n\nimport numpy as np\nfrom numpy.linalg import norm\n\nfrom sklearn.mixture import GaussianMixture\n\n\nclass DTWAligner(object):\n \"\"\"Align feature matcies\n\n Attributes:\n dist (function): Distance function\n radius (int): Radius\n verbose (int): Default is 0\n \"\"\"\n\n def __init__(self, dist=lambda x, y: norm(x - y), radius=1, verbose=0):\n self.verbose = verbose\n self.dist = dist\n self.radius = radius\n\n def transform(self, XY):\n X, Y = XY\n assert X.ndim == 3 and Y.ndim == 3\n\n X_aligned = np.zeros_like(X)\n Y_aligned = np.zeros_like(Y)\n for idx, (x, y) in enumerate(zip(X, Y)):\n x, y = trim_zeros_frames(x), trim_zeros_frames(y)\n dist, path = fastdtw(x, y, radius=self.radius, dist=self.dist)\n dist /= (len(x) + len(y))\n pathx = list(map(lambda l: l[0], path))\n pathy = list(map(lambda l: l[1], path))\n x, y = x[pathx], y[pathy]\n X_aligned[idx][:len(x)] = x\n Y_aligned[idx][:len(y)] = y\n if self.verbose > 0:\n print(\"{}, distance: {}\".format(idx, dist))\n return X_aligned, Y_aligned\n\n\nclass IterativeDTWAligner(object):\n \"\"\"Align feature matcies iteratively using GMM-based feature conversion\n\n Attributes:\n n_iter (int): Number of iterations.\n dist (function): Distance function\n radius (int): Radius\n verbose (int): Default is 0\n \"\"\"\n\n def __init__(self, n_iter=3, dist=lambda x, y: norm(x - y), radius=1, verbose=0):\n self.n_iter = n_iter\n self.dist = dist\n self.radius = radius\n self.verbose = verbose\n\n def transform(self, XY):\n X, Y = XY\n assert X.ndim == 3 and Y.ndim == 3\n\n Xc = X.copy() # this will be updated iteratively\n X_aligned = np.zeros_like(X)\n Y_aligned = np.zeros_like(Y)\n refined_paths = np.empty(len(X), dtype=np.object)\n\n for idx in range(self.n_iter):\n for idx, (x, y) in enumerate(zip(Xc, Y)):\n x, y = trim_zeros_frames(x), trim_zeros_frames(y)\n dist, path = fastdtw(x, y, radius=self.radius, dist=self.dist)\n dist /= (len(x) + len(y))\n pathx = list(map(lambda l: l[0], path))\n pathy = list(map(lambda l: l[1], path))\n\n refined_paths[idx] = pathx\n x, y = x[pathx], y[pathy]\n X_aligned[idx][:len(x)] = x\n Y_aligned[idx][:len(y)] = y\n if self.verbose > 0:\n print(\"{}, distance: {}\".format(idx, dist))\n\n # Fit\n gmm = GaussianMixture(\n n_components=32, covariance_type=\"full\", max_iter=100)\n XY = np.concatenate((X_aligned, Y_aligned),\n axis=-1).reshape(-1, X.shape[-1] * 2)\n gmm.fit(XY)\n windows = [(0, 0, np.array([1.0]))] # no delta\n paramgen = MLPG(gmm, windows=windows)\n for idx in range(len(Xc)):\n x = trim_zeros_frames(Xc[idx])\n Xc[idx][:len(x)] = paramgen.transform(x)\n\n # Finally we can get aligned X\n for idx in range(len(X_aligned)):\n x = X[idx][refined_paths[idx]]\n X_aligned[idx][:len(x)] = x\n\n return X_aligned, Y_aligned\n", "sub_path": "nnmnkwii/preprocessing/alignment.py", "file_name": "alignment.py", "file_ext": "py", "file_size_in_byte": 3537, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "numpy.linalg.norm", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 33, "usage_type": "call"}, {"api_name": "nnmnkwii.util.trim_zeros_frames", "line_number": 35, "usage_type": "call"}, {"api_name": "fastdtw.fastdtw", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.object", "line_number": 71, "usage_type": "attribute"}, {"api_name": "nnmnkwii.util.trim_zeros_frames", "line_number": 75, "usage_type": "call"}, {"api_name": "fastdtw.fastdtw", "line_number": 76, "usage_type": "call"}, {"api_name": "sklearn.mixture.GaussianMixture", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 94, "usage_type": "call"}, {"api_name": "nnmnkwii.baseline.gmm.MLPG", "line_number": 95, "usage_type": "call"}, {"api_name": "nnmnkwii.util.trim_zeros_frames", "line_number": 97, "usage_type": "call"}]}
+{"seq_id": "234536827", "text": "import uuid\nimport datetime\nfrom app import sql_id_yaml\n\n# 有効期限 3時間\nEXPIRATION_TIME = datetime.timedelta(hours=3)\n\n\nclass AuthorizationUuid:\n def __init__(self, postgres_instance, auth):\n self.postgres_instance = postgres_instance\n self.auth = auth\n # 現在時刻を取得\n self.now_time = datetime.datetime.now()\n # users_loginのデータ取得\n self.__get_users_login()\n\n # users_login情報取得\n def __get_users_login(self):\n self.get_results = self.postgres_instance.select(\n sql_id_yaml['get_users_login_user_name'],\n self.auth.username()\n )\n\n # users_loginのtokenとlogin_timeを更新する\n def __put_token_login_time(self):\n self.postgres_instance.update(\n sql_id_yaml['put_users_login_token_login_time'],\n *(self.token, self.now_time),\n where_id=self.auth.username()\n )\n\n # users_loginのlogin_timeを更新する\n def __put_login_time(self):\n self.postgres_instance.update(\n sql_id_yaml['put_users_login_login_time'],\n self.now_time,\n where_id=self.auth.username()\n )\n\n # uuidのチェック\n def __create_check_token(self):\n # 既存のtokenとは別のtokenを生成する(入れ替え用)\n __token = str(uuid.uuid4())\n while True:\n # 生成したtokenと既存のtokenが不一致であれば新しいtokenを返却\n if not(__token == self.get_results['records'][0]['token']):\n self.token = __token\n break\n # 存在していた場合は再度uuidを作成、検索\n __token = str(uuid.uuid4())\n\n # uuidを生成して返却\n def users_login(self):\n # ログインユーザーの最終ログイン時刻を取得\n login_time = self.get_results['records'][0]['login_time']\n\n if EXPIRATION_TIME > (self.now_time - login_time):\n print(\"有効期限内\")\n # nowの方が小さい場合有効期限内 その時は既存のlogin_timeを更新(uuidはそのまま)\n self.__put_login_time()\n elif EXPIRATION_TIME <= (self.now_time - login_time):\n print(\"有効期限切れ\")\n # 更新用のtokenを生成する。\n self.__create_check_token()\n # nowの方が大きいので有効期限切れ tokenとlogin_timeを更新\n self.__put_token_login_time()\n\n # 更新後のデータを返却用に取得\n self.__get_users_login()\n\n return self.get_results\n", "sub_path": "app/AuthorizationUuid.py", "file_name": "AuthorizationUuid.py", "file_ext": "py", "file_size_in_byte": 2606, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "datetime.timedelta", "line_number": 6, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 14, "usage_type": "attribute"}, {"api_name": "app.sql_id_yaml", "line_number": 21, "usage_type": "name"}, {"api_name": "app.sql_id_yaml", "line_number": 28, "usage_type": "name"}, {"api_name": "app.sql_id_yaml", "line_number": 36, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 44, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 51, "usage_type": "call"}]}
+{"seq_id": "475982539", "text": "import os\nimport glob\n\nimport numpy as np\nimport cv2\nimport matplotlib.pyplot as plt\n\nfrom skimage.color import rgb2gray\nfrom skimage.filters import gaussian\nfrom skimage.segmentation import active_contour\nfrom skimage.io import imread, imsave\n\ndef show_lmks(im, landmarks, color=(0, 0, 255), show_txt=False, scale=10):\n for i in range(landmarks.shape[0]):\n x, y = landmarks[i, 0], landmarks[i, 1]\n if show_txt:\n cv2.putText(im, str(i), (int(x), int(y)),\n fontFace=cv2.FONT_HERSHEY_SCRIPT_SIMPLEX,\n fontScale=0.5, color=color)\n cv2.circle(im, (int(x), int(y)), scale, color, -1)\n return im\n\n\ndef show_img(img, wait=1):\n cv2.namedWindow('debugg', cv2.WINDOW_NORMAL)\n cv2.imshow('debugg', img)\n cv2.waitKey(wait)\n\n\ndef skimage2opencv(src):\n src *= 255\n src = src.astype(np.uint8)\n src = cv2.cvtColor(src, cv2.COLOR_RGB2BGR)\n return src\n\n\ndef opencv2skimage(src):\n src = cv2.cvtColor(src, cv2.COLOR_BGR2RGB)\n src = src.astype(np.float32)\n src /= 255\n return src\n\n\ndef get_bounding_rect(points):\n max_vals = np.max(points, axis=0)\n min_vals = np.min(points, axis=0)\n\n print('max_vals: {}'.format(max_vals))\n print('min_vals: {}'.format(min_vals))\n\n y_max = int(max_vals[0]+0.5)\n y_min = int(min_vals[0]-0.5)\n x_max = int(max_vals[1]+0.5)\n x_min = int(min_vals[1]-0.5)\n\n return (x_min, y_min, x_max, y_max)\n\n\ndef extend_rect(x_min, y_min, x_max, y_max, ratio=0.25):\n ht = y_max - y_min\n wd = x_max - x_min\n\n ext_y = ht*ratio\n ext_x = wd*ratio\n\n y_min -= ext_y\n y_max += ext_y\n x_min -= ext_x\n x_max += ext_x\n\n y_max = int(y_max+0.5)\n y_min = int(y_min-0.5)\n x_max = int(x_max+0.5)\n x_min = int(x_min-0.5)\n\n return (x_min, y_min, x_max, y_max)\n\n\nimg_path = 'material/00218.jpg'\nmouth_img_path = 'material/00218_mouth.jpg'\nlmks_path = 'material/00218_lmks.txt'\nmouth_lmks_path = 'material/00218_lmks_mouth.txt'\n# scale = 4.0\n\nimg = cv2.imread(img_path)\nprint('img.shape: {}'.format(img.shape))\n\nlmks = np.loadtxt(lmks_path).reshape(-1, 2)\nlmks = lmks[:, ::-1] # (x,y) to (r,c)\n\n# img = cv2.resize(img, (int(img.shape[1] / scale), int(img.shape[0] / scale)))\n# lmks = lmks / scale\n\n# #####crop face area\n# (x_min, y_min, x_max, y_max) = get_bounding_rect(lmks)\n# (x_min, y_min, x_max, y_max) = get_bounding_rect(lmks)\n# (x_min, y_min, x_max, y_max) = extend_rect(x_min, y_min, x_max, y_max, ratio=0.25)\n\n\n# img = img[y_min:y_max, x_min:x_max, :]\n\n# outer_mouth_lmks -= np.array([y_min, x_min])\n############\n\n#outer_mouth_index = [58, 118, 119, 59, 120, 121, 60, 122, 123, 61, 124, 125, 62, 126, 127, 63, 128, 129, 64, 130, 131, 65, 132, 133, 58]\nouter_mouth_index = [58, 118, 119, 59, 120, 121, 60, 122, 123, 61,\n 124, 125, 62, 126, 127, 63, 128, 129, 64, 130, 131, 65, 132, 133]\nouter_mouth_lmks = lmks[outer_mouth_index, :]\nprint('outer mouth landmarks: {}'.format(outer_mouth_lmks))\n\n#inner_mouth_index = [116, 134, 135, 66, 136, 137, 67, 138, 139, 68, 140, 141, 117, 142, 143, 69, 144, 145, 70, 146, 147, 71, 148, 149, 116]\ninner_mouth_index = [116, 134, 135, 66, 136, 137, 67, 138, 139, 68,\n 140, 141, 117, 142, 143, 69, 144, 145, 70, 146, 147, 71, 148, 149]\ninner_mouth_upper_index = [116, 134, 135, 66, 136, 137, 67, 138, 139, 68,\n 140, 141, 117]\ninner_mouth_lower_index = [142, 143, 69, 144, 145, 70, 146, 147, 71, 148, 149]\n\ninner_mouth_lmks = lmks[inner_mouth_index, :]\n# inner_mouth_lmks = lmks[inner_mouth_upper_index, :]\nprint('inner mouth landmarks: {}'.format(inner_mouth_lmks))\n\n# crop face area\n(x_min, y_min, x_max, y_max) = get_bounding_rect(outer_mouth_lmks)\nprint(\"bounding box: {}\".format((x_min, y_min, x_max, y_max)))\n(x_min, y_min, x_max, y_max) = extend_rect(\n x_min, y_min, x_max, y_max, ratio=0.25)\nprint(\"extended bounding box: {}\".format((x_min, y_min, x_max, y_max)))\n\nimg = img[y_min:y_max, x_min:x_max, ::-1]\nimsave(mouth_img_path, img)\n############\n\nouter_mouth_lmks -= np.array([y_min, x_min])\nprint('outer mouth landmarks (after crop): {}'.format(outer_mouth_lmks))\n\ninner_mouth_lmks -= np.array([y_min, x_min])\nprint('inner mouth landmarks (after crop): {}'.format(inner_mouth_lmks))\n\n# lmks_cropped = lmks - np.array([y_min, x_min])\n# print('landmarks (after crop): {}'.format(lmks_cropped))\n# lmks_cropped = lmks_cropped[:,:-1] # (r,c) to (x,y)\n# np.savetxt(mouth_lmks_path, lmks_cropped.flatten(), fmt='%.2f')\nmouth_lmks = np.vstack([outer_mouth_lmks, inner_mouth_lmks])\nmouth_lmks = mouth_lmks[:,::-1] # (r,c) to (x,y)\nnp.savetxt(mouth_lmks_path, mouth_lmks.flatten(), fmt='%.2f')\n\nimg_skimage = opencv2skimage(img)\nimg_skimage_gray = rgb2gray(img_skimage)\nprint('img_skimage_gray.shape: {}'.format(img_skimage_gray.shape))\n\ninit_contour = inner_mouth_lmks\n# init_contour = outer_mouth_lmks\nprint('init contour landmarks: {}'.format(init_contour))\n\n#snake = active_contour(gaussian(img_skimage_gray, 3), init_contour, alpha=0.015, beta=10, gamma=0.001, coordinates='rc')\n# snake = active_contour(gaussian(img_skimage_gray, 3), init_contour, coordinates='rc')\nsnake = active_contour(img_skimage_gray, init_contour,\n alpha=0.015, beta=10, gamma=0.001,\n w_line=-1, w_edge=1,\n max_px_move=5,\n boundary_condition='free', coordinates='rc')\n\nsnake2 = active_contour(img_skimage_gray, init_contour,\n alpha=0.015, beta=1, gamma=0.001,\n w_line=2, w_edge=1,\n max_px_move=5,\n boundary_condition='fixed', coordinates='rc')\nprint('active contour landmarks: {}'.format(snake))\nprint('active contour landmarks 2: {}'.format(snake2))\n\n# img = show_lmks(img, init_contour, scale=1, color=(0, 0, 255))\n# img = show_lmks(img, snake, scale=1, color=(0, 255, 0))\n\nfig, ax = plt.subplots(figsize=(7, 7))\nax.imshow(img_skimage_gray, cmap=plt.cm.gray)\n# ax.plot(init_contour[:, 1], init_contour[:, 0], '--r', lw=1)\nax.plot(init_contour[:, 1], init_contour[:, 0], 'or', ms=5)\nax.plot(snake[:, 1], snake[:, 0], 'ob', ms=5)\nax.plot(snake2[:, 1], snake2[:, 0], 'og', ms=5)\nax.set_xticks([]), ax.set_yticks([])\nax.axis([0, img_skimage_gray.shape[1], img_skimage_gray.shape[0], 0])\n\nplt.show()\n\n# show_img(img, 0)\n# cv2.destroyAllWindows()\n", "sub_path": "scikit_image_test/lip_active_contour.py", "file_name": "lip_active_contour.py", "file_ext": "py", "file_size_in_byte": 6379, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "cv2.putText", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SCRIPT_SIMPLEX", "line_number": 18, "usage_type": "attribute"}, {"api_name": "cv2.circle", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.namedWindow", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.WINDOW_NORMAL", "line_number": 25, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 32, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 33, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2BGR", "line_number": 33, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 38, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 39, "usage_type": "attribute"}, {"api_name": "numpy.max", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 46, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 88, "usage_type": "call"}, {"api_name": "skimage.io.imsave", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 145, "usage_type": "call"}, {"api_name": "skimage.color.rgb2gray", "line_number": 148, "usage_type": "call"}, {"api_name": "skimage.segmentation.active_contour", "line_number": 157, "usage_type": "call"}, {"api_name": "skimage.segmentation.active_contour", "line_number": 163, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 174, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 174, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 175, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 175, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}]}
+{"seq_id": "335391952", "text": "\"\"\"DAQ Module.\"\"\"\n\nimport logging\nimport typing as t\nfrom collections import namedtuple\n\nimport numpy as np\nfrom zhinst.core import DataAcquisitionModule as ZIDAQModule\n\nfrom zhinst.toolkit.driver.modules.base_module import BaseModule\nfrom zhinst.toolkit.nodetree.helper import NodeDict\n\nif t.TYPE_CHECKING: # pragma: no cover\n from zhinst.toolkit.session import Session\n\nlogger = logging.getLogger(__name__)\nDAQResult = namedtuple(\"DAQResult\", [\"header\", \"value\", \"time\", \"frequency\", \"shape\"])\n\n\nclass DAQModule(BaseModule):\n \"\"\"Data Acquisition Module.\n\n The Data Acquisition Module corresponds to the Data Acquisition tab of the\n LabOne User Interface. It enables the user to record and align time and\n frequency domain data from multiple instrument signal sources at a defined\n data rate. The data may be recorded either continuously or in bursts based\n upon trigger criteria analogous to the functionality provided by laboratory\n oscilloscopes.\n\n For a complete documentation see the LabOne user manual\n https://docs.zhinst.com/labone_programming_manual/data_acquisition_module.html\n\n Args:\n daq_module: Instance of the core DAQ module.\n session: Session to the Data Server.\n \"\"\"\n\n def __init__(self, daq_module: ZIDAQModule, session: \"Session\"):\n super().__init__(daq_module, session)\n self.root.update_nodes(\n {\n \"/triggernode\": {\n \"GetParser\": self._get_node,\n \"SetParser\": self._set_node,\n }\n },\n raise_for_invalid_node=False,\n )\n\n @staticmethod\n def _process_burst(\n node: str, burst: t.Dict[str, t.Any], clk_rate: float\n ) -> DAQResult:\n \"\"\"Process a single burst into a formatted DAQResult object.\n\n Args:\n node: Name of the node of the burst.\n burst: raw burst data.\n clk_rate: Clock rate [Hz] for converting the timestamps. Only\n applies if the raw flag is reset.\n\n Returns:\n Processed and formatted burst data.\n \"\"\"\n if \"fft\" in node:\n bin_count = len(burst[\"value\"][0])\n bin_resolution = burst[\"header\"][\"gridcoldelta\"]\n frequency = np.arange(bin_count)\n bandwidth = bin_resolution * len(frequency)\n frequency = frequency * bin_resolution\n if \"xiy\" in node:\n frequency = frequency - bandwidth / 2.0 + bin_resolution / 2.0\n return DAQResult(\n burst.get(\"header\", {}),\n burst[\"value\"],\n None,\n frequency,\n burst[\"value\"].shape,\n )\n timestamp = burst[\"timestamp\"]\n return DAQResult(\n burst.get(\"header\", {}),\n burst[\"value\"],\n (timestamp[0] - timestamp[0][0]) / clk_rate,\n None,\n burst[\"value\"].shape,\n )\n\n @staticmethod\n def _process_node_data(\n node: str, data: t.List[t.Dict[str, t.Any]], clk_rate: float\n ) -> t.List[t.Union[t.Dict[str, t.Any], DAQResult]]:\n \"\"\"Process the data of a node.\n\n Only subscribed sample nodes are processed. Other nodes (module native nodes)\n are returned in the original format.\n\n Args:\n node: Name of the node of the burst.\n data: raw data for the node.\n clk_rate: Clock rate [Hz] for converting the timestamps. Only\n applies if the raw flag is reset.\n\n Returns:\n Processed and formatted node data.\n \"\"\"\n if isinstance(data[0], dict):\n return [DAQModule._process_burst(node, burst, clk_rate) for burst in data]\n return data\n\n def finish(self) -> None:\n \"\"\"Stop the module.\n\n .. versionadded:: 0.5.0\n \"\"\"\n self._raw_module.finish()\n\n def finished(self) -> bool:\n \"\"\"Check if the acquisition has finished.\n\n Returns:\n Flag if the acquisition has finished.\n\n .. versionadded:: 0.5.0\n \"\"\"\n return self._raw_module.finished()\n\n def trigger(self) -> None:\n \"\"\"Execute a manual trigger.\n\n .. versionadded:: 0.5.0\n \"\"\"\n self._raw_module.trigger()\n\n def read(self, *, raw: bool = False, clk_rate: float = 60e6) -> NodeDict:\n \"\"\"Read the acquired data from the module.\n\n The data is split into bursts.\n\n Args:\n raw: Flag if the acquired data from the subscribed device\n device nodes should be converted into the DAQResult format\n (raw = False) or not. (default = False)\n clk_rate: Clock rate [Hz] for converting the timestamps. Only\n applies if the raw flag is reset.\n\n Returns:\n Result of the burst grouped by the signals.\n \"\"\"\n raw_result = self._raw_module.read(flat=True)\n if raw:\n return NodeDict(raw_result)\n return NodeDict(\n {\n node: self._process_node_data(node, data, clk_rate)\n for node, data in raw_result.items()\n }\n )\n", "sub_path": "src/zhinst/toolkit/driver/modules/daq_module.py", "file_name": "daq_module.py", "file_ext": "py", "file_size_in_byte": 5178, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "typing.TYPE_CHECKING", "line_number": 13, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 16, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 17, "usage_type": "call"}, {"api_name": "zhinst.toolkit.driver.modules.base_module.BaseModule", "line_number": 20, "usage_type": "name"}, {"api_name": "zhinst.core.DataAcquisitionModule", "line_number": 38, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 52, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 52, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 68, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 91, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 91, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 91, "usage_type": "attribute"}, {"api_name": "typing.List", "line_number": 92, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 92, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 92, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 92, "usage_type": "attribute"}, {"api_name": "zhinst.toolkit.nodetree.helper.NodeDict", "line_number": 152, "usage_type": "call"}, {"api_name": "zhinst.toolkit.nodetree.helper.NodeDict", "line_number": 153, "usage_type": "call"}, {"api_name": "zhinst.toolkit.nodetree.helper.NodeDict", "line_number": 135, "usage_type": "name"}]}
+{"seq_id": "572799846", "text": "from django.db import models\n\n\n# Create your models here.\nclass TMDB(models.Model):\n popularity = models.FloatField(null=True)\n vote_count = models.IntegerField(null=True)\n movie_id = models.IntegerField(null=True)\n title = models.CharField(max_length=100)\n vote_average = models.FloatField(null=True)\n description = models.TextField()\n release_date = models.DateTimeField()\n\n def __str__(self):\n return self.title\n\n\n", "sub_path": "webviz/viz/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 448, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.db.models.Model", "line_number": 5, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 5, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 6, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 6, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 7, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 8, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 9, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 9, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 10, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 10, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 11, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 11, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 12, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 12, "usage_type": "name"}]}
+{"seq_id": "409558305", "text": "import pyaudio\nimport wave\nimport string\nimport random\n\nrand_str = lambda n: ''.join([random.choice(string.lowercase) for i in xrange(n)])\n\nCHUNK = 1024\nFORMAT = pyaudio.paInt16\nCHANNELS = 2\nRATE = 44100\nRECORD_SECONDS = 10\n\nwhile True:\n\tp = pyaudio.PyAudio()\n\n\tstream = p.open(format=FORMAT,\n\t\t\tchannels=CHANNELS,\n\t\t\trate=RATE,\n\t\t\tinput=True,\n\t\t\tframes_per_buffer=CHUNK)\n\n\tprint(\"* recording\")\n\n\tframes = []\n\n\tfor i in range(0, int(RATE / CHUNK * RECORD_SECONDS)):\n\t data = stream.read(CHUNK)\n\t frames.append(data)\n\n\tprint(\"* done recording\")\n\n\tstream.stop_stream()\n\tstream.close()\n\tp.terminate()\n\n\twf = wave.open('../data/speech/'+rand_str(10)+'.wav', 'wb')\n\twf.setnchannels(CHANNELS)\n\twf.setsampwidth(p.get_sample_size(FORMAT))\n\twf.setframerate(RATE)\n\twf.writeframes(b''.join(frames))\n\twf.close()\n", "sub_path": "examples/e.py", "file_name": "e.py", "file_ext": "py", "file_size_in_byte": 806, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "random.choice", "line_number": 6, "usage_type": "call"}, {"api_name": "string.lowercase", "line_number": 6, "usage_type": "attribute"}, {"api_name": "pyaudio.paInt16", "line_number": 9, "usage_type": "attribute"}, {"api_name": "pyaudio.PyAudio", "line_number": 15, "usage_type": "call"}, {"api_name": "wave.open", "line_number": 37, "usage_type": "call"}]}
+{"seq_id": "225072150", "text": "import psycopg2\nimport logging\nimport config\nimport datetime\n\n## Logging handler for PostgreSQL\n#\nclass psqlHandler(logging.Handler):\n\n def connect(self):\n try:\n self._connect = psycopg2.connect(\n database = self._database,\n host = self._host,\n user = self._user,\n password = self._password,\n sslmode = \"disable\")\n return True\n except:\n return False\n\n def __init__(self, params):\n\n if not params:\n raise Exception (\"No database where to log...\")\n\n self._database = params['database']\n self._host = params['host']\n self._user = params['user']\n self._password = params['password']\n\n self._connect = None\n\n if not self.connect():\n raise Exception (\"Database connection error\")\n\n logging.Handler.__init__(self)\n \n self._connect.cursor().execute(config.LOGGER_INITIAL)\n self._connect.commit()\n self._connect.cursor().close()\n\n def emit(self, record):\n\n # Use default formatting:\n self.format(record)\n\n if record.exc_info:\n record.exc_text = logging._defaultFormatter.formatException(record.exc_info)\n else:\n record.exc_text = \"\"\n\n # Insert log record:\n try:\n cur = self._connect.cursor()\n except:\n self.connect()\n cur = self._connect.cursor()\n\n record.__dict__['created'] = datetime.datetime.utcfromtimestamp(record.__dict__['created']) \\\n - datetime.timedelta(hours=4)\n \n \n cur.execute(config.LOGGER_INSERTION, record.__dict__)\n\n self._connect.commit()\n self._connect.cursor().close()\n", "sub_path": "reminders/psqlhandler.py", "file_name": "psqlhandler.py", "file_ext": "py", "file_size_in_byte": 1829, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.Handler", "line_number": 8, "usage_type": "attribute"}, {"api_name": "psycopg2.connect", "line_number": 12, "usage_type": "call"}, {"api_name": "logging.Handler.__init__", "line_number": 37, "usage_type": "call"}, {"api_name": "logging.Handler", "line_number": 37, "usage_type": "attribute"}, {"api_name": "config.LOGGER_INITIAL", "line_number": 39, "usage_type": "attribute"}, {"api_name": "logging._defaultFormatter.formatException", "line_number": 49, "usage_type": "call"}, {"api_name": "logging._defaultFormatter", "line_number": 49, "usage_type": "attribute"}, {"api_name": "datetime.datetime.utcfromtimestamp", "line_number": 60, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 60, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 61, "usage_type": "call"}, {"api_name": "config.LOGGER_INSERTION", "line_number": 64, "usage_type": "attribute"}]}
+{"seq_id": "497776615", "text": "import multiprocessing as mp\nfrom multiprocessing import Process\nfrom more_itertools import chunked\nfrom tqdm import tqdm\n\n__all__ = [\"batch_multiprocess\"]\n\ndef batch_multiprocess(function_list, n_cores=mp.cpu_count(), show_progress=True):\n \"\"\"\n Run a list of functions on `n_cores` (default: all CPU cores),\n with the option to show a progress bar using tqdm (default: shown).\n \"\"\"\n iterator = [*chunked(function_list, n_cores)]\n if show_progress:\n iterator = tqdm(iterator)\n for func_batch in iterator:\n procs = []\n for f in func_batch:\n procs.append(Process(target=f))\n for p in procs:\n p.start()\n for p in procs:\n p.join()\n\ndef _batch_multiprocess(function_list, n_cores=mp.cpu_count(), show_progress=True):\n \"\"\"\n Dummy sequential version of `batch_multiprocess` to be swapped in\n when there's an error to debug in that\n \"\"\"\n for func in function_list:\n func()\n", "sub_path": "src/dx/share/multiprocessing.py", "file_name": "multiprocessing.py", "file_ext": "py", "file_size_in_byte": 976, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "multiprocessing.cpu_count", "line_number": 8, "usage_type": "call"}, {"api_name": "more_itertools.chunked", "line_number": 13, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 15, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 19, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 25, "usage_type": "call"}]}
+{"seq_id": "458684113", "text": "import csv\nimport operator\n\nimport math\nfrom deap import base\nfrom deap import creator\nfrom deap import gp\nfrom deap import tools\n\nminim = 10 ** 8\nmaxim = 10000\nbestindivid = -1\n\npath = \"C:\\\\Users\\\\stanc\\\\OneDrive\\\\Desktop\\\\MyDataTest3.csv\"\n\nfile = open(path, newline='')\nreader = csv.reader(file)\nheader = next(reader)\n\nvTp = []\nvCl = []\nvpH = []\nvRedox = []\nvLeit = []\nvTrueb = []\nvCl_2 = []\nvFm = []\nvFm2 = []\ndata = []\n\nv = [1, 2, 3, 4]\n\n\ndef getmed(v):\n '''minim = v[0]\n maxim = v[0]\n for i in v:\n if i>maxim:\n maxim = i\n if i 50:\n x = 50\n elif func(data[i][0], data[i][1], data[i][2], data[i][3], data[i][4], data[i][5], data[i][6], data[i][7],\n data[i][8], data[i][9], data[i][10], data[i][11], data[i][12], data[i][13], data[i][14],\n data[i][15], data[i][16], data[i][17]) < -50:\n x = -50\n else:\n x = func(data[i][0], data[i][1], data[i][2], data[i][3], data[i][4], data[i][5], data[i][6], data[i][7],\n data[i][8], data[i][9], data[i][10], data[i][11], data[i][12], data[i][13], data[i][14],\n data[i][15], data[i][16], data[i][17])\n predictie = 1 / (1 + math.exp(x))\n v.append(predictie)\n return v\n\n\ndef getPPVsingular(p):\n classifiedpositives = 0\n truepositives = 0\n allpositives = 0\n for i in range(1, len(data)):\n suma = 0\n if data[i][18] == 1:\n allpositives += 1\n if p[i - 1] > 0.5:\n classifiedpositives += 1\n if data[i][18] == 1:\n truepositives += 1\n return truepositives, allpositives, classifiedpositives\n\n\n# p1 = getprobabilities(\"safeDiv(cos(add(neg(add(pHmedie, safeDiv(cos(cos(pH)), neg(Cl_2)))), safeDiv(min(Cl_2, mul(Clmedie, Clmedie)), neg(Cl_2)))), mul(mul(pH, safeDiv(Cl_2, Fmmedie)), max(sin(neg(cos(add(neg(add(pHmedie, safeDiv(cos(cos(pH)), neg(Cl_2)))), safeDiv(min(Cl_2, Truebmedie), neg(Cl_2)))))), add(min(cos(cos(pH)), mul(neg(safeDiv(Leit, Redoxmedie)), add(add(Tp, add(add(safeDiv(safeDiv(cos(cos(pH)), neg(Cl_2)), min(min(neg(Cl_2), Cl_2medie), safeDiv(Fm, Fmmedie))), Fmmedie), max(neg(Cl_2), Fmmedie))), max(Fmmedie, Fm_2)))), Fmmedie))))\")\n# p1 = getprobabilities(\"safeDiv(cos(neg(add(cos(sub(pHmedie, cos(Cl_2))), safeDiv(add(Clmedie, safeDiv(add(Clmedie, safeDiv(add(Clmedie, safeDiv(Truebmedie, add(safeDiv(mul(pHmedie, Cl), cos(Clmedie)), sub(Cl_2, cos(Cl_2))))), neg(Cl_2medie))), Fm_2medie)), neg(Cl_2))))), neg(add(cos(safeDiv(add(sin(Trueb), neg(add(cos(cos(Clmedie)), safeDiv(add(Clmedie, Trueb), Clmedie)))), neg(Cl_2))), safeDiv(add(Clmedie, neg(add(cos(sin(add(Clmedie, neg(add(cos(Cl_2), safeDiv(add(Clmedie, Truebmedie), neg(Cl_2))))))), safeDiv(add(Clmedie, safeDiv(Truebmedie, add(safeDiv(mul(pHmedie, Cl), cos(Cl_2)), sub(neg(sub(safeDiv(Trueb, Trueb), min(safeDiv(add(Clmedie, Trueb), Clmedie), Cl_2))), cos(Cl_2))))), neg(Cl_2))))), neg(Cl_2)))))\")\n# p1 = getprobabilities(\"safeDiv(cos(neg(add(cos(sub(pHmedie, cos(Cl_2))), safeDiv(add(Clmedie, safeDiv(add(Clmedie, safeDiv(add(Clmedie, safeDiv(Truebmedie, add(safeDiv(mul(pHmedie, Cl), cos(Clmedie)), sub(sin(Cl_2), cos(Cl_2))))), neg(Cl_2medie))), Fm_2medie)), neg(Cl_2))))), neg(add(cos(safeDiv(add(sin(Trueb), neg(add(cos(cos(Clmedie)), safeDiv(add(Clmedie, Trueb), Clmedie)))), neg(Cl_2))), safeDiv(add(Clmedie, neg(add(cos(sin(add(Clmedie, neg(add(cos(Cl_2), safeDiv(add(Clmedie, Truebmedie), neg(Cl_2))))))), safeDiv(add(Clmedie, safeDiv(Truebmedie, add(safeDiv(mul(pHmedie, Cl), cos(Cl_2)), sub(neg(sub(safeDiv(Trueb, Trueb), min(safeDiv(add(Clmedie, Trueb), Clmedie), Cl_2))), cos(Cl_2))))), neg(Cl_2))))), neg(Cl_2)))))\")\np1 = getprobabilities(\n \"safeDiv(cos(add(neg(add(pHmedie, safeDiv(cos(cos(pH)), neg(Cl_2)))), safeDiv(min(Cl_2, mul(Clmedie, Clmedie)), neg(Cl_2)))), mul(mul(pH, safeDiv(safeDiv(safeDiv(neg(Cl_2), Tp), min(min(neg(Cl_2), Cl_2medie), safeDiv(Fm, Fmmedie))), Fmmedie)), max(sin(neg(cos(add(neg(add(pHmedie, safeDiv(cos(cos(pH)), neg(Cl_2)))), safeDiv(min(Cl_2, mul(Clmedie, Clmedie)), neg(Cl_2)))))), add(min(cos(safeDiv(min(Cl_2, mul(Clmedie, Clmedie)), neg(Cl_2))), mul(neg(safeDiv(Leit, Redoxmedie)), add(add(sub(add(mul(Fmmedie, Cl_2), mul(Fmmedie, Cl_2)), neg(Cl_2)), add(add(safeDiv(safeDiv(cos(cos(pH)), neg(Cl_2)), min(min(neg(Cl_2), Cl_2medie), safeDiv(Fm, Fmmedie))), Fmmedie), max(neg(Cl_2), Fmmedie))), max(Fmmedie, Redoxmedie)))), Fmmedie))))\")\nprint(getPPVsingular(p1))\n", "sub_path": "Licenta2019StanciuCSebastian/Cod/ProgramareGenetica/verificaritimeseries.py", "file_name": "verificaritimeseries.py", "file_ext": "py", "file_size_in_byte": 7811, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "csv.reader", "line_number": 17, "usage_type": "call"}, {"api_name": "deap.gp.PrimitiveSet", "line_number": 102, "usage_type": "call"}, {"api_name": "deap.gp", "line_number": 102, "usage_type": "name"}, {"api_name": "operator.add", "line_number": 105, "usage_type": "attribute"}, {"api_name": "operator.sub", "line_number": 106, "usage_type": "attribute"}, {"api_name": "operator.mul", "line_number": 107, "usage_type": "attribute"}, {"api_name": "operator.neg", "line_number": 109, "usage_type": "attribute"}, {"api_name": "math.cos", "line_number": 110, "usage_type": "attribute"}, {"api_name": "math.sin", "line_number": 111, "usage_type": "attribute"}, {"api_name": "deap.creator.create", "line_number": 135, "usage_type": "call"}, {"api_name": "deap.creator", "line_number": 135, "usage_type": "name"}, {"api_name": "deap.base.Fitness", "line_number": 135, "usage_type": "attribute"}, {"api_name": "deap.base", "line_number": 135, "usage_type": "name"}, {"api_name": "deap.creator.create", "line_number": 136, "usage_type": "call"}, {"api_name": "deap.creator", "line_number": 136, "usage_type": "name"}, {"api_name": "deap.gp.PrimitiveTree", "line_number": 136, "usage_type": "attribute"}, {"api_name": "deap.gp", "line_number": 136, "usage_type": "name"}, {"api_name": "deap.creator.FitnessMin", "line_number": 136, "usage_type": "attribute"}, {"api_name": "deap.base.Toolbox", "line_number": 138, "usage_type": "call"}, {"api_name": "deap.base", "line_number": 138, "usage_type": "name"}, {"api_name": "deap.gp.genFull", "line_number": 139, "usage_type": "attribute"}, {"api_name": "deap.gp", "line_number": 139, "usage_type": "name"}, {"api_name": "deap.tools.initIterate", "line_number": 140, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 140, "usage_type": "name"}, {"api_name": "deap.creator.Individual", "line_number": 140, "usage_type": "attribute"}, {"api_name": "deap.creator", "line_number": 140, "usage_type": "name"}, {"api_name": "deap.tools.initRepeat", "line_number": 141, "usage_type": "attribute"}, {"api_name": "deap.tools", "line_number": 141, "usage_type": "name"}, {"api_name": "deap.gp.compile", "line_number": 142, "usage_type": "attribute"}, {"api_name": "deap.gp", "line_number": 142, "usage_type": "name"}, {"api_name": "math.exp", "line_number": 162, "usage_type": "call"}]}
+{"seq_id": "122835458", "text": "# -*- coding: utf-8 -*-\n# @Time : 2021/1/18 0018 10:38\n# @Author : Owen\n# @Email : zhangwx0794@gmail.com\n# @File : ClsTaobao.py\n# @Project : Taobao\n\nimport xlwt, xlrd\nimport pymysql\nimport os\nimport re\nimport platform\nfrom openpyxl import *\n\n\n# import win32com.client as win32\n\nclass Taobao():\n\n # 1. 数据库操作returnCnt默认返回结果集条数,默认返回所有\n def sql_operation(self, sql, returnCnt=-1):\n conn = pymysql.Connect(\n host='47.111.113.238',\n port=3318,\n db='taobao',\n user='taobao',\n passwd='Boss123456..taobao',\n charset='utf8'\n )\n # 获取游标\n cursor = conn.cursor()\n try:\n cursor.execute(sql)\n except Exception as e:\n conn.rollback()\n print('异常sql: ', sql)\n print('事务处理失败!异常信息:', e)\n else:\n conn.commit()\n sqlRes = cursor.fetchall()\n # 关闭连接\n cursor.close()\n conn.close()\n if returnCnt == -1:\n return sqlRes\n elif returnCnt == 1:\n return sqlRes[0][0]\n else:\n return 0\n\n # 2. 查找当前工作目录下所有的xls文件\n def get_path_xls(self, absPath):\n allXls = os.listdir(absPath)\n zz = re.compile('(\\.xls)$')\n xlsList = []\n for xls in allXls:\n zzRes = zz.findall(xls)\n if 'xls' in xls and len(zzRes) > 0:\n xlsList.append(xls)\n return xlsList\n\n # 3. 查找当前工作目录下所有的xlsx文件\n def get_path_xlsx(self, absPath):\n allXlsx = os.listdir(absPath)\n if self.getSystemPlatform() == 'Windows':\n zz = re.compile('(\\.xlsx)$')\n else:\n zz = re.compile('(\\.xlsx)$')\n xlsxList = []\n for xlsx in allXlsx:\n zzRes = zz.findall(xlsx)\n if 'xlsx' in xlsx and len(zzRes) > 0:\n xlsxList.append(xlsx)\n return xlsxList\n\n # 4. 数据规范校验\n\n # 5. xls转xlsx\n # def xls_to_xlsx(self,xlsPath):\n # try:\n # excel = win32.gencache.EnsureDispatch('Excel.Application')\n # wb = excel.Workbooks.Open(xlsPath)\n # wb.SaveAs(xlsPath + \"x\", FileFormat=51) # FileFormat = 51 is for .xlsx extension\n # wb.Close() # FileFormat = 56 is for .xls extension\n # excel.Application.Quit()\n # except Exception as e:\n # print('xls转换xlsx异常',e)\n # else:\n # os.remove(xlsPath)\n # print(xlsPath,'转换成功,源文件已删除')\n\n # 6. excel文件重命名\n def format_xls_name(self, xlsPath):\n # 1.拼接变量\n # * xls名称\n if self.getSystemPlatform() == 'Windows':\n xlsName = str(xlsPath).split('\\\\')[-1]\n else:\n xlsName = str(xlsPath).split('/')[-1]\n # * xls所在路径\n xlsPwd = str(xlsPath).split(xlsName)[0]\n # * 订单日期\n length1 = len(re.compile('^\\d+\\.\\d+').findall(xlsName))\n if length1 > 0:\n rq = re.compile('^\\d+\\.\\d+').findall(xlsName)[0]\n month = str(rq).split('.')[0]\n day = str(rq).split('.')[1]\n if int(month) >= 10:\n year = 2020\n else:\n year = 2021\n date = str(year) + '-' + str(month).rjust(2, '0') + '-' + str(day).rjust(2, '0')\n print(xlsName, date)\n else:\n length2 = len(re.compile('^\\d{4}-\\d{2}-\\d{2}').findall(xlsName))\n if length2 > 0:\n date = re.compile('^\\d{4}-\\d{2}-\\d{2}').findall(xlsName)[0]\n else:\n # 文件中不含有日期关键词就会重命名为2099-12-31+原有中文字符+xlsx\n date = '2099-12-31'\n\n # * 店铺名称\n shopNameTemp = re.compile('[\\u4e00-\\u9fff]+').findall(xlsName)[0]\n shopName = str(shopNameTemp).replace('汇总', '').replace('订单', '').replace('总汇', '').replace('副本', '')\n # 2.拼接新的文件名称\n newXlsName = date + shopName + '.xlsx'\n newXlsPath = xlsPwd + newXlsName\n # 如果新旧文件名称不一样,则重命名\n if xlsName != newXlsName:\n # * 判断新文件是否已经存在\n for i in range(2, 100):\n if os.path.exists(newXlsPath):\n newXlsPath = xlsPwd + date + shopName + '_' + str(i) + '.xlsx'\n else:\n os.rename(xlsPath, newXlsPath)\n print('重命名成功!', xlsName, ' => ', newXlsName)\n break\n return None\n\n # 7. 删除excel含有关键字的列\n def del_col_from_key(self, xlsPath, key_word, col=1):\n # xlsPath必须得是绝对路径\n wbb = load_workbook(xlsPath)\n wss = wbb.active\n # 删除第一列【时间】数据\n kw = str(wss.cell(1, col).value).strip()\n if str(key_word).strip() == kw:\n wss.delete_cols(col)\n wbb.save(xlsPath)\n print(xlsPath + '第一列关键字' + key_word + '删除成功!')\n wbb.close()\n return None\n\n # 8. 校验指定范围列数据有效性\n def list_none_check(self, row, startCol, endCol):\n for i in range(startCol, endCol):\n if str(row[i]) == '':\n # print('none',end=' ')\n return 0\n return 1\n\n # 9. 订单号唯一检测\n # 查询数据库检查订单号是否已存在\n def chk_data_is_exist(self, order_id):\n sql = 'select count(0) from orderInfo where orderId = %s' % order_id\n cnt = self.sql_operation(sql, 1)\n # 返回订单号重复查询结果 存在返回1 不存在返回0\n return cnt\n\n # 9.1 批量检测订单号是否唯一\n def chkXlsOrderUniq(self, xlsPath):\n # 将数据库中所有的订单号存入数组orderIdLst中\n sql = 'select orderId from orderInfo where isDel = 0'\n orderIdRes = self.sql_operation(sql, -1)\n orderIdLst = []\n for orderId in orderIdRes:\n orderIdLst.append(str(orderId[0]).strip())\n # 获取xls中所有订单号\n wb = xlrd.open_workbook(xlsPath)\n # * 打开第一个sheet\n ws = wb.sheet_by_index(0)\n cnt = 0\n for line in range(1, ws.nrows):\n # 获取订单号\n colValue = str(ws.cell_value(rowx=line, colx=5)).strip()\n if colValue in orderIdLst:\n print(xlsPath, '数据库存在相同订单号', colValue, '第', line + 1, '行')\n cnt += 1\n else:\n if cnt > 0:\n print(xlsPath, '在订单号重复的表格中竟然发现订单号[{0}]竟然没有插入数据里...'.format(colValue))\n orderList = ws.row_values(line)\n self.insertOrder(orderList, xlsPath)\n print('呜呜,既然被发现了,只能老老实实的插进数据库里~~~')\n return cnt\n\n # 10. 数据导入\n def importData(self, xlsPath):\n if self.getSystemPlatform() == 'Windows':\n xlsName = str(xlsPath).split('\\\\')[-1]\n else:\n xlsName = str(xlsPath).split('/')[-1]\n wb = xlrd.open_workbook(xlsPath)\n # * 打开第一个sheet\n ws = wb.sheet_by_index(0)\n # * 从第二行开始导入数据\n dataImportNum = 0\n if self.data_format_check(xlsPath) == 0 and self.chkXlsOrderUniq(xlsPath) == 0:\n for line in range(1, ws.nrows):\n # * 获取当前行数据\n rowList = ws.row_values(line)\n # * 校验当前行指定范围列数据是否完整\n # 日期、经手人、店铺名称、宝贝名称、关键词、旺旺ID、订单号、客单价、佣金\n try:\n sqlFormat = 'insert into orderInfo(shopName,goodsName,goodsKey,wangwangId,orderId,goodsPrice,goodsYj,redPackets,ssyj,handlerName,opWechatId,custName,date,note) values({0},{1},{2},{3},{4},{5},{6},{7},{8},{9},{10},{11},{12},{13})'\n shopName = '\\'' + str(rowList[1]) + '\\''\n goodsName = '\\'' + str(rowList[2]) + '\\''\n goodsKey = '\\'' + str(rowList[3]) + '\\''\n wangwangId = '\\'' + str(rowList[4]).strip() + '\\''\n orderId = '\\'' + str(rowList[5]).strip() + '\\''\n goodsPrice = rowList[6]\n goodsYj = rowList[7]\n redPackets = rowList[8]\n ssyj = rowList[9]\n handlerName = '\\'' + str(rowList[10]) + '\\''\n opWechatId = '\\'' + str(rowList[11]) + '\\''\n custName = '\\'' + str(rowList[12]) + '\\''\n date = '\\'' + str(re.compile('^\\d{4}-\\d{2}-\\d{2}').findall(xlsName)[0]) + '\\''\n try:\n note = '\\'' + str(rowList[14]) + '\\''\n except Exception as e:\n print(xlsPath, '第{0}行获取备注列数据异常'.format(line + 1),e)\n note = '\\'\\''\n except Exception as e:\n print(xlsPath,'第{0}行数据异常'.format(line + 1),e)\n return 0\n sql = sqlFormat.format(shopName, goodsName, goodsKey, wangwangId, orderId, goodsPrice, goodsYj,\n redPackets, ssyj, handlerName, opWechatId, custName, date,note)\n # 根据订单号检查数据,如果不重复,则将表格中的数据插入数据库;0不存在 1存在\n try:\n # print('插入数据库……', sql)\n self.sql_operation(sql)\n dataImportNum += 1\n except Exception as e:\n print(xlsName, '有毛病,插入数据库异常')\n return 0\n print(xlsName, '成功导入{0}条数据'.format(dataImportNum))\n return dataImportNum\n\n # 11. 店铺名&旺旺ID唯一检测\n\n # 12. 更新店铺名\n\n # 13. 删除重复的xls文件\n def delRepeName(self, xlsPath, xlsxList):\n if self.getSystemPlatform() == 'Windows':\n xlsName = str(xlsPath).split('\\\\')[-1]\n else:\n xlsName = str(xlsPath).split('/')[-1]\n if xlsName + 'x' in xlsxList:\n os.remove(xlsPath)\n print('已删除', xlsPath)\n\n # 14. 取出文件中第一列的值\n def getColValues(self, xlsPath, col=1):\n wbb = load_workbook(xlsPath)\n wss = wbb.active\n # 删除第一列【时间】数据\n kw = str(wss.cell(1, col).value).strip()\n wbb.close()\n return kw\n\n # 15. 写入数据到excel\n def writeData2Xls(self, xlsPath, value, row=1, col=1):\n try:\n wbb = load_workbook(xlsPath)\n wss = wbb.active\n # 写入数据内容到单元格中\n wss.cell(row, col).value = value\n wbb.save(xlsPath)\n wbb.close()\n except Exception as e:\n print('写入数据异常', e)\n return 0\n else:\n return 1\n\n # 16. 插入新列\n def insertColum(self, xlsPath, col):\n try:\n wbb = load_workbook(xlsPath)\n wss = wbb.active\n # 插入新列\n wss.insert_cols(idx=col)\n wbb.save(xlsPath)\n wbb.close()\n except Exception as e:\n print('插入列异常', e)\n return 0\n else:\n return 1\n\n # 17. 删除空行\n def delteBlankRow(self, xlsPath):\n try:\n wbb = load_workbook(xlsPath)\n wss = wbb.active\n # 写入数据内容到单元格中\n for row in range(1, wss.max_row + 1):\n rowList = []\n for col in range(1, wss.max_column + 1):\n if wss.cell(row=row, column=col).value != None:\n rowList.append(wss.cell(row=row, column=col).value)\n if len(rowList) <= 1 or 'SUM' in str(rowList):\n print(xlsPath, '第', row, '行为空,已删除', rowList)\n wss.delete_rows(idx=row)\n wbb.save(xlsPath)\n wbb.close()\n return -1\n else:\n pass\n # print(xlsPath,rowList)\n wbb.save(xlsPath)\n wbb.close()\n except Exception as e:\n print('删除列异常', e)\n return -1\n else:\n return 0\n\n # 18. 根据规则表格完善\n def completeForm(self, xlsPath, col, value):\n try:\n wbb = load_workbook(xlsPath)\n wss = wbb.active\n # 写入数据内容到单元格中,从第2行开始填充数据\n for row in range(2, wss.max_row + 1):\n # 红包及其他\n if wss.cell(row=row, column=col).value == None and col == 9:\n wss.cell(row=row, column=col).value = 0\n # 店铺名称\n elif wss.cell(row=row, column=col).value == None and col == 2:\n wss.cell(row=row, column=col).value = value\n # 刷手佣金\n elif wss.cell(row=row, column=col).value == None and col == 10:\n wss.cell(row=row, column=col).value = 0\n # 经手人\n elif wss.cell(row=row, column=col).value == None and col == 11:\n wss.cell(row=row, column=col).value = value\n # 操作人微信\n elif wss.cell(row=row, column=col).value == None and col == 12:\n wss.cell(row=row, column=col).value = value\n # 客户名称\n elif col == 13:\n wss.cell(row=row, column=col).value = value\n # 日期\n elif col == 14:\n wss.cell(row=row, column=col).value = value\n wbb.save(xlsPath)\n wbb.close()\n except Exception as e:\n print('插入列异常', e)\n return 0\n else:\n return 1\n\n # 19. 删除订单号为空或订单号重复的行\n def delBlankOrderRow(self, xlsPath):\n while (True):\n # 接收错误行号\n line = self.chkRepeOrderInXls(xlsPath)\n if line > 0:\n try:\n wbb = load_workbook(xlsPath)\n wss = wbb.active\n # 删除错误行\n wss.delete_rows(idx=line)\n print('delete', xlsPath, '第', line, '行')\n wbb.save(xlsPath)\n except Exception as e:\n print('删除列异常', e)\n wbb.close()\n finally:\n wbb.close()\n else:\n break\n\n # 20. 数据规范检查\n def data_format_check(self, xlsPath):\n wb = xlrd.open_workbook(xlsPath)\n # * 打开第一个sheet\n ws = wb.sheet_by_index(0)\n for line in range(1, ws.nrows):\n # * 获取当前行数据\n rowList = list(ws.row_values(line))\n if rowList[5] == '':\n print(xlsPath, '第', line + 1, '行', '第', 6, '列数据为空!')\n return -1\n for i in range(5, 10):\n if not str(rowList[i]).replace('.', '').isdigit():\n print('错误: 第{0}行第{1}列数据不规范,含有非数字或小数点字符,或者为空!'.format(line + 1, i + 1))\n return -2\n return 0\n\n # 21. 检测单张表是否有重复订单号,正常返回0,异常返回订单号所在行号\n def chkRepeOrderInXls(self, xlsPath):\n try:\n wb = xlrd.open_workbook(xlsPath)\n ws = wb.sheet_by_index(0)\n lstTmp = []\n for line in range(1, ws.nrows):\n if ws.cell(line, 5).value not in lstTmp and ws.cell(line, 5).value != '':\n lstTmp.append(ws.cell(line, 5).value)\n elif ws.cell(line, 5).value == '':\n print(xlsPath, '订单号为空,第', line + 1, '行')\n return line + 1\n else:\n print(xlsPath, '有订单号重复,第', line + 1, '行')\n return line + 1\n except Exception as e:\n print(xlsPath, '捕捉到打开表异常', e)\n return -1\n return 0\n\n # 22. 插入单条订单数据\n def insertOrder(self, orderList, xlsPath):\n if self.getSystemPlatform() == 'Windows':\n xlsName = str(xlsPath).split('\\\\')[-1]\n else:\n xlsName = str(xlsPath).split('/')[-1]\n rowList = orderList\n # * 校验当前行指定范围列数据是否完整\n # 日期、经手人、店铺名称、宝贝名称、关键词、旺旺ID、订单号、客单价、佣金\n sqlFormat = 'insert into orderInfo(shopName,goodsName,goodsKey,wangwangId,orderId,goodsPrice,goodsYj,redPackets,ssyj,handlerName,opWechatId,custName,date) values({0},{1},{2},{3},{4},{5},{6},{7},{8},{9},{10},{11},{12})'\n shopName = '\\'' + str(rowList[1]) + '\\''\n goodsName = '\\'' + str(rowList[2]) + '\\''\n goodsKey = '\\'' + str(rowList[3]) + '\\''\n wangwangId = '\\'' + str(rowList[4]).strip() + '\\''\n orderId = '\\'' + str(rowList[5]).strip() + '\\''\n goodsPrice = rowList[6]\n goodsYj = rowList[7]\n redPackets = rowList[8]\n ssyj = rowList[9]\n handlerName = '\\'' + str(rowList[10]) + '\\''\n opWechatId = '\\'' + str(rowList[11]) + '\\''\n custName = '\\'' + str(rowList[12]) + '\\''\n date = '\\'' + str(re.compile('^\\d{4}-\\d{2}-\\d{2}').findall(xlsName)[0]) + '\\''\n sql = sqlFormat.format(shopName, goodsName, goodsKey, wangwangId, orderId, goodsPrice, goodsYj,\n redPackets, ssyj, handlerName, opWechatId, custName, date)\n # 根据订单号检查数据,如果不重复,则将表格中的数据插入数据库;0不存在 1存在\n try:\n # print('插入数据库……', sql)\n self.sql_operation(sql)\n print('插入单条订单信息到数据库成功,订单号[{0}]'.format(orderId))\n except Exception as e:\n print('订单号[{0}]插入数据库异常'.format(orderId))\n return 0\n\n # 23. 获取系统类型\n def getSystemPlatform(self):\n plat_tuple = platform.architecture()\n system = platform.system()\n return system\n", "sub_path": "TaobaoSDMS/models/ClsTaobao.py", "file_name": "ClsTaobao.py", "file_ext": "py", "file_size_in_byte": 18751, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pymysql.Connect", "line_number": 22, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 53, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 54, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 64, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 66, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 68, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 103, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 105, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 115, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 117, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 132, "usage_type": "call"}, {"api_name": "os.path", "line_number": 132, "usage_type": "attribute"}, {"api_name": "os.rename", "line_number": 135, "usage_type": "call"}, {"api_name": "xlrd.open_workbook", "line_number": 179, "usage_type": "call"}, {"api_name": "xlrd.open_workbook", "line_number": 203, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 228, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 261, "usage_type": "call"}, {"api_name": "xlrd.open_workbook", "line_number": 390, "usage_type": "call"}, {"api_name": "xlrd.open_workbook", "line_number": 408, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 447, "usage_type": "call"}, {"api_name": "platform.architecture", "line_number": 461, "usage_type": "call"}, {"api_name": "platform.system", "line_number": 462, "usage_type": "call"}]}
+{"seq_id": "161524228", "text": "# -*- coding: utf-8 -*-\r\n\r\n\r\nimport sys\r\nimport strat\r\nimport time\r\nimport rest\r\nimport pprint \r\nimport Var\r\nimport threading\r\nfrom datetime import datetime\r\nimport numpy as np\r\nimport matplotlib.dates as pld\r\nimport matplotlib.pyplot as plt\r\nimport matplotlib.animation as animation\r\n\r\n\r\nepic = \"CS.D.EURUSD.MINI.IP\"\r\ndata = rest.ig().Auth(Var.login,Var.password)\r\nThread1 = None\r\nThread2 = None\r\ndates = []\t\r\nPosition = {\"Bull\":[],\"Bear\":[]}\r\nfig = plt.figure()\r\nax1 = fig.add_subplot(1,1,1)\r\n\r\nclass cafe(object):\r\n\t\"\"\"docstring for cafe\"\"\"\r\n\tdef __init__(self):\r\n\t\tThread1 = threading.Thread(name=\"Thread1\", target=follow)\r\n\t\tThread1.setDaemon(True)\r\n\t\trunning = True\r\n\tdef follow():\r\n\t\t#THREAD\r\n\t\twhile self.running:\r\n\t\t\tBullish, Bearish = strat.strat().ClientSentiment(epic)\r\n\t\t\tPosition[\"Bull\"].append(Bullish)\r\n\t\t\tPosition[\"Bear\"].append(Bearish)\r\n\t\t\tdates.append(datetime.now())\r\n\t\t\t#print(\"Long: \"+str(Bullish),\"Short: \"+str(Bearish))\r\n\t\t\ttime.sleep(10)\r\n\r\n\tdef stop():\r\n\t\tself.running = False\r\n\r\ndef animate(i):\r\n\tax1.ylabel(\"Sentiment\")\r\n\tax1.xlabel(\"Temps\")\r\n\t\r\n\tax1.clear()\r\n\tax1.plot(dates, Position[\"Bull\"], xdate=True, ydate=False)\r\n\r\ndef gui():\r\n\tpass\r\n\t\r\n\r\n\r\n\r\ndef main():\r\n\t\r\n\tcf = cafe()\r\n\ttime.sleep(1)\r\n\t\r\n\tani = animation.FuncAnimation(fig, animate,interval=1000)\r\n\tplt.show()\r\n\r\n\tinput(\"Hit to Stop3\")\r\n\tprint(\"Fin\")\r\n\tThread1.join()\r\n\t\t\t\r\nif __name__ == '__main__':\r\n\tmain()\r\n", "sub_path": "sentiment.py", "file_name": "sentiment.py", "file_ext": "py", "file_size_in_byte": 1402, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "rest.ig", "line_number": 19, "usage_type": "call"}, {"api_name": "Var.login", "line_number": 19, "usage_type": "attribute"}, {"api_name": "Var.password", "line_number": 19, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "threading.Thread", "line_number": 30, "usage_type": "call"}, {"api_name": "strat.strat", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 39, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 39, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 41, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.animation.FuncAnimation", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.animation", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}]}
+{"seq_id": "641699383", "text": "import datetime\nimport string\nfrom random import randrange\n\n\ndef now():\n return str(datetime.datetime.now().strftime('%Y%m%d%H%M%S'))\n \n\ndef mkdir(directory,sudo=False,warn=True):\n command = \"mkdir %s\" % directory\n with settings(warn_only=warn):\n if sudo:\n sudo(command)\n else:\n run(command)\n\ndef rand(n):\n alphabets = string.digits + string.letters\n return ''.join(alphabets[randrange(len(alphabets))] for i in xrange(n))\n", "sub_path": "util.py", "file_name": "util.py", "file_ext": "py", "file_size_in_byte": 481, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "datetime.datetime.now", "line_number": 7, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 7, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 19, "usage_type": "attribute"}, {"api_name": "string.letters", "line_number": 19, "usage_type": "attribute"}, {"api_name": "random.randrange", "line_number": 20, "usage_type": "call"}]}
+{"seq_id": "52230202", "text": "from prompt_toolkit.shortcuts import prompt\nfrom prompt_toolkit.styles import style_from_dict\nfrom prompt_toolkit.token import Token\nfrom prompt_toolkit.validation import Validator, ValidationError\n\nclass _OptionValidator(Validator):\n def __init__(self, options, default):\n super().__init__()\n self.options = [o.lower() for o in options]\n self.defaultAllowed = default is not None\n\n def validate(self, document):\n text = document.text.lower().strip()\n if self.defaultAllowed and not text:\n return\n elif text not in self.options:\n raise ValidationError(message='Invalid response', cursor_position=len(document.text))\n\nclass _StringValidator(Validator):\n def validate(self, document):\n text = document.text.strip()\n if not text:\n raise ValidationError(message='Invalid response', cursor_position=len(document.text))\n\n_prompt_styles = style_from_dict({\n Token: '#dddddd',\n Token.Sigil: '#00ff00',\n Token.Prompt: '#dddddd',\n Token.Symbol: '#777777',\n Token.Option: '#00ffff',\n Token.Default: '#ff77ff',\n})\n\ndef promptOptions(msg, options, default=None):\n tokens = [(Token.Sigil, \"* \"),\n (Token.Prompt, msg),\n (Token.Symbol, \" [\"),]\n\n first = True\n for option in options:\n if first:\n first = False\n else:\n tokens.append((Token.Symbol, \",\"))\n if option == default:\n tokens.append((Token.Default, option.upper()))\n else:\n tokens.append((Token.Option, option))\n\n tokens.append((Token.Symbol, \"] : \"))\n val = prompt(get_prompt_tokens=lambda x: tokens, style=_prompt_styles, validator=_OptionValidator(options, default))\n if val:\n return val.lower().strip()\n return default\n\ndef promptString(msg):\n tokens = [(Token.Sigil, \"* \"),\n (Token.Prompt, msg),\n (Token.Symbol, \" : \")]\n val = prompt(get_prompt_tokens=lambda x: tokens, style=_prompt_styles, validator=_StringValidator())\n if val:\n return val.strip()\n return None\n", "sub_path": "pdfs/Prompt.py", "file_name": "Prompt.py", "file_ext": "py", "file_size_in_byte": 2116, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "prompt_toolkit.validation.Validator", "line_number": 6, "usage_type": "name"}, {"api_name": "prompt_toolkit.validation.ValidationError", "line_number": 17, "usage_type": "call"}, {"api_name": "prompt_toolkit.validation.Validator", "line_number": 19, "usage_type": "name"}, {"api_name": "prompt_toolkit.validation.ValidationError", "line_number": 23, "usage_type": "call"}, {"api_name": "prompt_toolkit.styles.style_from_dict", "line_number": 25, "usage_type": "call"}, {"api_name": "prompt_toolkit.token.Token", "line_number": 26, "usage_type": "name"}, {"api_name": "prompt_toolkit.token.Token.Sigil", "line_number": 27, "usage_type": "attribute"}, {"api_name": "prompt_toolkit.token.Token", "line_number": 27, "usage_type": "name"}, {"api_name": "prompt_toolkit.token.Token.Prompt", "line_number": 28, "usage_type": "attribute"}, {"api_name": "prompt_toolkit.token.Token", "line_number": 28, "usage_type": "name"}, {"api_name": "prompt_toolkit.token.Token.Symbol", "line_number": 29, "usage_type": "attribute"}, {"api_name": "prompt_toolkit.token.Token", "line_number": 29, "usage_type": "name"}, {"api_name": "prompt_toolkit.token.Token.Option", "line_number": 30, "usage_type": "attribute"}, {"api_name": "prompt_toolkit.token.Token", "line_number": 30, "usage_type": "name"}, {"api_name": "prompt_toolkit.token.Token.Default", "line_number": 31, "usage_type": "attribute"}, {"api_name": "prompt_toolkit.token.Token", "line_number": 31, "usage_type": "name"}, {"api_name": "prompt_toolkit.token.Token.Sigil", "line_number": 35, "usage_type": "attribute"}, {"api_name": "prompt_toolkit.token.Token", "line_number": 35, "usage_type": "name"}, {"api_name": "prompt_toolkit.token.Token.Prompt", "line_number": 36, "usage_type": "attribute"}, {"api_name": "prompt_toolkit.token.Token", "line_number": 36, "usage_type": "name"}, {"api_name": "prompt_toolkit.token.Token.Symbol", "line_number": 37, "usage_type": "attribute"}, {"api_name": "prompt_toolkit.token.Token", "line_number": 37, "usage_type": "name"}, {"api_name": "prompt_toolkit.token.Token.Symbol", "line_number": 44, "usage_type": "attribute"}, {"api_name": "prompt_toolkit.token.Token", "line_number": 44, "usage_type": "name"}, {"api_name": "prompt_toolkit.token.Token.Default", "line_number": 46, "usage_type": "attribute"}, {"api_name": "prompt_toolkit.token.Token", "line_number": 46, "usage_type": "name"}, {"api_name": "prompt_toolkit.token.Token.Option", "line_number": 48, "usage_type": "attribute"}, {"api_name": "prompt_toolkit.token.Token", "line_number": 48, "usage_type": "name"}, {"api_name": "prompt_toolkit.token.Token.Symbol", "line_number": 50, "usage_type": "attribute"}, {"api_name": "prompt_toolkit.token.Token", "line_number": 50, "usage_type": "name"}, {"api_name": "prompt_toolkit.shortcuts.prompt", "line_number": 51, "usage_type": "call"}, {"api_name": "prompt_toolkit.token.Token.Sigil", "line_number": 57, "usage_type": "attribute"}, {"api_name": "prompt_toolkit.token.Token", "line_number": 57, "usage_type": "name"}, {"api_name": "prompt_toolkit.token.Token.Prompt", "line_number": 58, "usage_type": "attribute"}, {"api_name": "prompt_toolkit.token.Token", "line_number": 58, "usage_type": "name"}, {"api_name": "prompt_toolkit.token.Token.Symbol", "line_number": 59, "usage_type": "attribute"}, {"api_name": "prompt_toolkit.token.Token", "line_number": 59, "usage_type": "name"}, {"api_name": "prompt_toolkit.shortcuts.prompt", "line_number": 60, "usage_type": "call"}]}
+{"seq_id": "492918777", "text": "from scipy.fft import dct\nimport cv2\nimport numpy as np\nfrom numba import jit\nfrom os import listdir, remove\nfrom joblib import Parallel, delayed\nimport sqlite3\nimport io\nconn = sqlite3.connect('phashes.db')\nIMAGE_PATH = \"./../../../public/images\"\n\n\ndef create_table():\n cursor = conn.cursor()\n query = '''\n\t CREATE TABLE IF NOT EXISTS phashes(\n\t \tid INTEGER NOT NULL UNIQUE PRIMARY KEY, \n\t \tphash BLOB NOT NULL\n\t )\n\t'''\n cursor.execute(query)\n conn.commit()\n\n\ndef check_if_exists_by_id(id):\n cursor = conn.cursor()\n query = '''SELECT EXISTS(SELECT 1 FROM phashes WHERE id=(?))'''\n cursor.execute(query, (id,))\n all_rows = cursor.fetchone()\n return all_rows[0] == 1\n\n\ndef delete_descriptor_by_id(id):\n cursor = conn.cursor()\n query = '''DELETE FROM phashes WHERE id=(?)'''\n cursor.execute(query, (id,))\n conn.commit()\n\n\ndef get_all_ids():\n cursor = conn.cursor()\n query = '''SELECT id FROM phashes'''\n cursor.execute(query)\n all_rows = cursor.fetchall()\n return list(map(lambda el: el[0], all_rows))\n\n\ndef adapt_array(arr):\n out = io.BytesIO()\n np.save(out, arr)\n out.seek(0)\n return sqlite3.Binary(out.read())\n\n\ndef add_descriptor(id, phash):\n cursor = conn.cursor()\n query = '''INSERT INTO phashes(id, phash) VALUES (?,?)'''\n cursor.execute(query, (id, phash))\n conn.commit()\n\n\ndef sync_db():\n file_names = listdir(IMAGE_PATH)\n ids_in_db = set(get_all_ids())\n\n for file_name in file_names:\n file_id = int(file_name[:file_name.index('.')])\n if file_id in ids_in_db:\n ids_in_db.remove(file_id)\n for id in ids_in_db:\n delete_descriptor_by_id(id) # Fix this\n print(f\"deleting {id}\")\n\n\n@jit(cache=True, nopython=True)\ndef diff(dct, hash_size):\n dctlowfreq = dct[:hash_size, :hash_size]\n med = np.median(dctlowfreq)\n diff = dctlowfreq > med\n return diff.flatten()\n\n\ndef fast_phash(image, hash_size=16, highfreq_factor=4):\n img_size = hash_size * highfreq_factor\n image = cv2.resize(image, (img_size, img_size), interpolation=cv2.INTER_LINEAR) # cv2.INTER_AREA\n dct_data = dct(dct(image, axis=0), axis=1)\n return diff(dct_data, hash_size)\n\n\n@jit(cache=True, nopython=True)\ndef bit_list_to_32_uint8(bit_list_256):\n uint8_arr = []\n for i in range(32):\n bit_list = []\n for j in range(8):\n if(bit_list_256[i*8+j] == True):\n bit_list.append(1)\n else:\n bit_list.append(0)\n uint8_arr.append(bit_list_to_int(bit_list))\n return np.array(uint8_arr, dtype=np.uint8)\n\n\n@jit(cache=True, nopython=True)\ndef bit_list_to_int(bitlist):\n out = 0\n for bit in bitlist:\n out = (out << 1) | bit\n return out\n\n\ndef get_phash(query_image):\n bit_list_256 = fast_phash(query_image)\n phash = bit_list_to_32_uint8(bit_list_256)\n return phash\n\n\ndef calc_phash(file_name):\n file_id = int(file_name[:file_name.index('.')])\n img_path = IMAGE_PATH+\"/\"+file_name\n query_image = cv2.imread(img_path, cv2.IMREAD_GRAYSCALE)\n if query_image is None:\n print(f'error reading {img_path}')\n remove(img_path)\n return None\n phash = get_phash(query_image)\n phash_bin = adapt_array(phash)\n print(file_name)\n return (file_id, phash_bin)\n\n\nfile_names = listdir(IMAGE_PATH)\ncreate_table()\nsync_db()\nnew_images = []\n\nfor file_name in file_names:\n file_id = int(file_name[:file_name.index('.')])\n if check_if_exists_by_id(file_id):\n continue\n new_images.append(file_name)\n\nnew_images = [new_images[i:i + 5000] for i in range(0, len(new_images), 5000)]\nfor batch in new_images:\n phashes = Parallel(n_jobs=-1, verbose=1)(delayed(calc_phash)(file_name) for file_name in batch)\n phashes = [i for i in phashes if i] # remove None's\n print(\"pushing data to db\")\n conn.executemany('''INSERT INTO phashes(id, phash) VALUES (?,?)''', phashes)\n conn.commit()", "sub_path": "generate_public_phashes.py", "file_name": "generate_public_phashes.py", "file_ext": "py", "file_size_in_byte": 3942, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sqlite3.connect", "line_number": 9, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 50, "usage_type": "call"}, {"api_name": "sqlite3.Binary", "line_number": 52, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 63, "usage_type": "call"}, {"api_name": "scipy.fft.dct", "line_number": 77, "usage_type": "name"}, {"api_name": "numpy.median", "line_number": 78, "usage_type": "call"}, {"api_name": "numba.jit", "line_number": 75, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 85, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 85, "usage_type": "attribute"}, {"api_name": "scipy.fft.dct", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 101, "usage_type": "attribute"}, {"api_name": "numba.jit", "line_number": 90, "usage_type": "call"}, {"api_name": "numba.jit", "line_number": 104, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 121, "usage_type": "call"}, {"api_name": "cv2.IMREAD_GRAYSCALE", "line_number": 121, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 124, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 132, "usage_type": "call"}, {"api_name": "joblib.Parallel", "line_number": 145, "usage_type": "call"}, {"api_name": "joblib.delayed", "line_number": 145, "usage_type": "call"}]}
+{"seq_id": "221394407", "text": "import matplotlib.pyplot as plt\n\ndef plot():\n '''Plot output from sensitivity randomizer.'''\n\n x_vals = []\n y_vals = []\n\n try:\n with open('sens_list.txt') as fp:\n lims = [float(e) for e in fp.readline().strip().split(\",\")]\n\n for line in fp:\n vals = [float(e) for e in line.strip().split(\",\")]\n #print(vals)\n x_vals.append(vals[0])\n y_vals.append(vals[1])\n\n if lims[3] == 0:\n new_x = [x_vals[0]]\n new_y = [y_vals[0]]\n for i in range(1, len(x_vals)):\n new_x.append(x_vals[i]-0.00001)\n new_x.append(x_vals[i])\n new_y.append(y_vals[i-1])\n new_y.append(y_vals[i])\n\n x_vals = new_x\n y_vals = new_y\n\n plt.title(\"Sensitivity Randomizer\")\n plt.xlabel(\"Time (seconds)\")\n plt.ylabel(\"Sensitivity Multiplier\")\n plt.ylim(lims[1]-0.1, lims[2]+0.1)\n plt.plot(x_vals, y_vals)\n plt.show()\n\n except FileNotFoundError:\n print(\"Sensitivity list not found. Please first run SensitivityRandomizer.exe to generate output.\")\n\nif __name__ == \"__main__\":\n plot()\n", "sub_path": "Source/SensitivityRandomizer/Win32/Release/visualize.py", "file_name": "visualize.py", "file_ext": "py", "file_size_in_byte": 1215, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "matplotlib.pyplot.title", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}]}
+{"seq_id": "294870878", "text": "import pandas as pd\nimport matplotlib.pyplot as plt\nimport seaborn as sb\nfrom scipy.optimize import curve_fit\nimport numpy as np\n\nc1_df = pd.read_csv('/Users/felipeantoniomendezsalcido/Desktop/V_M_Project/C1_21_2_17/C1_21_2_17_data.csv')\n\npos_c1 = c1_df[(c1_df['Latency']>0.5) & (c1_df['RiseTime']>0) & (c1_df['Latency']<3.5) & (c1_df['RiseTime']<5) & (c1_df['Peaks'] <0)]\npos_c1['Peaks'] = pos_c1['Peaks'].abs()\n\ngrouped_df = pos_c1.copy()\ngrouped_df['ID'] = grouped_df['ID'].apply(lambda x: x.rsplit('_', 4)[0])\n\ngrouped_df.describe()\n\ndef get_fq(ID):\n if '06hz' in ID:\n return '0.06Hz'\n elif '1hz' in ID:\n return '1Hz'\n elif '4hz' in ID:\n return '4Hz'\ngrouped_df['Fq']= grouped_df['ID'].apply(get_fq)\n\ndef get_cond(ID):\n if 'L' in ID:\n return 'Low'\n elif 'N' in ID:\n return 'Normal'\n elif 'H' in ID:\n return 'High'\n\ndef get_pair(ID):\n if 'PP1' in ID:\n return 'P1'\n elif 'PP2' in ID:\n return 'P2'\n\ngrouped_df['Cond'] = grouped_df['ID'].apply(get_cond)\ngrouped_df['Pair'] = grouped_df['ID'].apply(get_pair)\nod = ['PP1_L06hz', 'PP2_L06hz', 'PP1_L1hz', 'PP2_L1hz', 'PP1_L4hz',\n 'PP2_L4hz', 'PP1_Nl06hz', 'PP2_Nl06hz', 'PP1_Nl1hz', 'PP2_Nl1hz',\n 'PP1_Nl4hz', 'PP2_Nl4hz', 'PP1_H06hz', 'PP2_H06hz', 'PP1_H1hz','PP2_H1hz', 'PP1_H4hz', 'PP2_H4hz', 'DCG']\n\n#plotting all points\nfig, all_plot = plt.subplots()\nsb.set_context('poster')\nsb.set_style('ticks')\nsb.stripplot(x='ID', y='Peaks', data= grouped_df, jitter=True, palette=(sb.dark_palette('blue', reverse= True)), order = od)\nsb.despine()\nplt.xticks(rotation=45)\nall_plot.set(xlabel='Conditions', ylabel= 'Peaks (pA)')\nplt.tight_layout()\nplt.show()\nfig.savefig('All_plot(2).svg')\n#ends plotting\n\n#V-M part\n#gettinh only the peaks for a V-M\nanother = grouped_df.copy()\njust_peaks = another.drop(['Latency', 'RiseTime', 'ID'], axis=1)\n\n#grouped by ionic condition as first level\nan_grouped = just_peaks.groupby(['Cond', 'Fq', 'Pair']).agg(['mean', 'var'])\n\n#grouped by Fq as first level\nfq_grouped = just_peaks.groupby(['Fq', 'Cond', 'Pair']).agg(['mean', 'var'])\n\n#function to fit\ndef curv(x, a, b):\n return ((a*x)-((x**2)/b))*(1+0.45)\n\n\n#plotting points by frequency and adjusting the function\nv_m1, ax = plt.subplots()\nsb.set_context('notebook')\nsb.set_style('ticks')\nfor ii in fq_grouped.index.levels[0]:\n xs = fq_grouped.loc[ii, ('Peaks', 'mean')]\n ys = fq_grouped.loc[ii, ('Peaks', 'var')]\n popt, pcov = curve_fit(curv, xs, ys)\n ax.scatter(xs, ys, label= ii)\n allx = np.linspace(0, 350)\n ax.plot(allx, curv(allx, *popt))\nax.legend()\nax.set_xlabel('Mean')\nax.set_ylabel('Var')\nsb.despine()\nplt.show()\n", "sub_path": "Python/Exploratory.py", "file_name": "Exploratory.py", "file_ext": "py", "file_size_in_byte": 2691, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pandas.read_csv", "line_number": 7, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "seaborn.set_context", "line_number": 48, "usage_type": "call"}, {"api_name": "seaborn.set_style", "line_number": 49, "usage_type": "call"}, {"api_name": "seaborn.stripplot", "line_number": 50, "usage_type": "call"}, {"api_name": "seaborn.dark_palette", "line_number": 50, "usage_type": "call"}, {"api_name": "seaborn.despine", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "seaborn.set_context", "line_number": 77, "usage_type": "call"}, {"api_name": "seaborn.set_style", "line_number": 78, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 84, "usage_type": "call"}, {"api_name": "seaborn.despine", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}]}
+{"seq_id": "586522750", "text": "\"\"\"\nBase classes for JetNet datasets.\n\"\"\"\n\nfrom typing import Any, Callable, List, Optional, Tuple\n\nimport torch\nfrom torch import Tensor\n\nfrom .normalisations import NormaliseABC\nfrom .utils import checkListNotEmpty, checkStrToList, firstNotNoneElement\n\n\nclass JetDataset(torch.utils.data.Dataset):\n \"\"\"\n Base class for jet datasets.\n Inspired by https://pytorch.org/vision/main/generated/torchvision.datasets.VisionDataset.html\n\n Args:\n data_dir (str): directory where dataset is or will be stored.\n particle_features (List[str], optional): list of particle features to retrieve. If empty\n or None, gets no particle features. Should default to all.\n jet_features (List[str], optional): list of jet features to retrieve. If empty or None,\n gets no particle features. Should default to all.\n particle_normalisation (Optional[NormaliseABC], optional): optional normalisation for\n particle-level features. Defaults to None.\n jet_normalisation (Optional[NormaliseABC], optional): optional normalisation for jet-level\n features. Defaults to None.\n particle_transform (callable, optional): A function/transform that takes in the particle\n data tensor and transforms it. Defaults to None.\n jet_transform (callable, optional): A function/transform that takes in the jet\n data tensor and transforms it. Defaults to None.\n num_particles (int, optional): max number of particles to retain per jet. Defaults to None.\n \"\"\"\n\n _repr_indent = 4\n\n particle_data = None\n jet_data = None\n max_num_particles = None\n\n def __init__(\n self,\n data_dir: str = \"./\",\n particle_features: Optional[List[str]] = None,\n jet_features: Optional[List[str]] = None,\n particle_normalisation: Optional[NormaliseABC] = None,\n jet_normalisation: Optional[NormaliseABC] = None,\n particle_transform: Optional[Callable] = None,\n jet_transform: Optional[Callable] = None,\n num_particles: Optional[int] = None,\n ):\n self.data_dir = data_dir\n\n self.particle_features, self.jet_features = checkStrToList(particle_features, jet_features)\n self.use_particle_features, self.use_jet_features = checkListNotEmpty(\n particle_features, jet_features\n )\n\n self.particle_normalisation = particle_normalisation\n self.jet_normalisation = jet_normalisation\n\n if self.use_particle_features:\n if self.particle_normalisation is not None:\n if self.particle_normalisation.features_need_deriving():\n self.particle_normalisation.derive_dataset_features(self.particle_data)\n self.particle_data = self.particle_normalisation(self.particle_data)\n\n if self.use_jet_features:\n if self.jet_normalisation is not None:\n if self.jet_normalisation.features_need_deriving():\n self.jet_normalisation.derive_dataset_features(self.jet_data)\n self.jet_data = self.jet_normalisation(self.jet_data)\n\n self.particle_transform = particle_transform\n self.jet_transform = jet_transform\n\n self.num_particles = num_particles\n\n @classmethod\n def getData(**opts) -> Any:\n \"\"\"Class method to download and return numpy arrays of the data\"\"\"\n raise NotImplementedError\n\n def __getitem__(self, index) -> Tuple[Optional[Tensor], Optional[Tensor]]:\n \"\"\"\n Gets data and if needed transforms it.\n\n Args:\n index (int): Index\n\n Returns:\n (Tuple[Optional[Tensor], Optional[Tensor]]): particle, jet data\n \"\"\"\n\n if self.use_particle_features:\n particle_data = self.particle_data[index]\n\n if self.particle_transform is not None:\n particle_data = self.particle_transform(particle_data)\n\n particle_data = Tensor(particle_data)\n else:\n particle_data = []\n\n if self.use_jet_features:\n jet_data = self.jet_data[index]\n\n if self.jet_transform is not None:\n jet_data = self.jet_transform(jet_data)\n\n jet_data = Tensor(jet_data)\n else:\n jet_data = []\n\n return particle_data, jet_data\n\n def __len__(self) -> int:\n return len(firstNotNoneElement(self.particle_data, self.jet_data))\n\n def __repr__(self) -> str:\n head = \"Dataset \" + self.__class__.__name__\n body = [f\"Number of datapoints: {self.__len__()}\"]\n\n if self.data_dir is not None:\n body.append(f\"Data location: {self.data_dir}\")\n\n body += self.extra_repr().splitlines()\n\n if self.particle_features is not None:\n bstr = f\"Particle features: {self.particle_features}\"\n if self.num_particles is not None:\n bstr += f\", max {self.num_particles} particles per jet\"\n\n body += [bstr]\n\n if self.jet_features is not None:\n body += [f\"Jet features: {self.jet_features}\"]\n\n if self.particle_normalisation is not None:\n body += [f\"Particle normalisation: {self.particle_normalisation}\"]\n\n if self.jet_normalisation is not None:\n body += [f\"Jet normalisation: {self.jet_normalisation}\"]\n\n if self.particle_transform is not None:\n body += [f\"Particle transform: {self.particle_transform}\"]\n\n if self.jet_transform is not None:\n body += [f\"Jet transform: {self.jet_transform}\"]\n\n lines = [head] + [\" \" * self._repr_indent + line for line in body]\n\n return \"\\n\".join(lines)\n\n def extra_repr(self) -> str:\n return \"\"\n", "sub_path": "jetnet/datasets/dataset.py", "file_name": "dataset.py", "file_ext": "py", "file_size_in_byte": 5726, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "torch.utils", "line_number": 14, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 45, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 45, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 46, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 46, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 47, "usage_type": "name"}, {"api_name": "normalisations.NormaliseABC", "line_number": 47, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 48, "usage_type": "name"}, {"api_name": "normalisations.NormaliseABC", "line_number": 48, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 49, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 49, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 50, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 50, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 51, "usage_type": "name"}, {"api_name": "utils.checkStrToList", "line_number": 55, "usage_type": "call"}, {"api_name": "utils.checkListNotEmpty", "line_number": 56, "usage_type": "call"}, {"api_name": "typing.Any", "line_number": 81, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 112, "usage_type": "call"}, {"api_name": "typing.Tuple", "line_number": 85, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 85, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 85, "usage_type": "name"}, {"api_name": "utils.firstNotNoneElement", "line_number": 119, "usage_type": "call"}]}
+{"seq_id": "501568324", "text": "########## Credits\n# Sprites idea by KidsCanCode on Youtube\n# https://www.youtube.com/watch?v=VO8rTszcW4s&list=PLsk-HSGFjnaH5yghzu7PcOzm9NhsW0Urw\n\n# Music Guinea Pig Hero by Trevor Lentz\n# https://opengameart.org/content/guinea-pig-hero\n# Art by Kenney, www.kenney.nl\n# https://opengameart.org/content/platformer-art-xeno-diversity\n# Explosion https://opengameart.org/content/bubble-explosion\n\n# Programmed by Ojars\n\nimport pygame\n\n\nfrom game import *\n\n_WDT = 700\n_HGT = 400\n_FPS = 30\n\npygame.init()\npygame.mixer.init() #required to play sounds\nscreen = pygame.display.set_mode((_WDT,_HGT))\nclock = pygame.time.Clock()\n\ngame = Game(screen, _WDT, _HGT)\n\n# Game loop\nrunning = True\nwhile running:\n # keep right speed\n clock.tick(_FPS)\n # process input\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n running = False\n if event.type == pygame.KEYDOWN:\n if event.key == pygame.K_F4 and (event.mod & pygame.KMOD_ALT):\n running = False\n else:# event.key == pygame.K_SPACE:\n game.keydown(event.key)\n\n #update\n game.update()\n\n # draw/render\n screen.fill(GRAY)\n game.draw()\n\n # after drawing everything, flip display\n pygame.display.flip()\n\npygame.quit()\n\n\n## # mob hits player\n## hits = pygame.sprite.spritecollide(player, mobs, False)\n## if hits:\n## running = False\n##\n## #bullet hits mob\n## hits = pygame.sprite.groupcollide( mobs, bullets, True, True )\n## for hit in hits:\n## score += hit.speedx\n## random.choice(expl_sounds).play()\n## expl = Explosion( hit.rect.center, hit.speedx )\n## all_sprites.add(expl)\n## m = Mob()\n## all_sprites.add(m)\n## mobs.add(m)\n##\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1758, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pygame.init", "line_number": 22, "usage_type": "call"}, {"api_name": "pygame.mixer.init", "line_number": 23, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 23, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 24, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 25, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 25, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 35, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 35, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 38, "usage_type": "attribute"}, {"api_name": "pygame.K_F4", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pygame.KMOD_ALT", "line_number": 39, "usage_type": "attribute"}, {"api_name": "game.keydown", "line_number": 42, "usage_type": "call"}, {"api_name": "game.update", "line_number": 45, "usage_type": "call"}, {"api_name": "game.draw", "line_number": 49, "usage_type": "call"}, {"api_name": "pygame.display.flip", "line_number": 52, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 52, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 54, "usage_type": "call"}]}
+{"seq_id": "357284656", "text": "from clarifai.rest import ClarifaiApp\nfrom clarifai.rest import Image as ClImage\n\napp = ClarifaiApp(api_key='c60682cfc90e43478682e309e247ba92')\n\ndef get_food(filename):\n\tmodel = app.models.get('food-items-v1.0')\n\timage = ClImage(file_obj=open(filename, 'rb'))\n\tresponse = model.predict([image])\n\tconcepts = response['outputs'][0]['data']['concepts']\n\tnew = []\n\tfor concept in concepts:\n\t new.append(concept['name'])\n\treturn new", "sub_path": "food.py", "file_name": "food.py", "file_ext": "py", "file_size_in_byte": 430, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "clarifai.rest.ClarifaiApp", "line_number": 4, "usage_type": "call"}, {"api_name": "clarifai.rest.Image", "line_number": 8, "usage_type": "call"}]}
+{"seq_id": "473055574", "text": "import discord\nfrom redbot.core import Config, checks, commands\n\n\ndef guild_only_check():\n async def pred(ctx: commands.Context):\n if ctx.guild is not None:\n return True\n else:\n return False\n\n return commands.check(pred)\n\n\nclass SetParser:\n def __init__(self, argument):\n allowed = (\"+\", \"-\")\n self.sum = int(argument)\n if argument and argument[0] in allowed:\n if self.sum < 0:\n self.operation = \"sub\"\n elif self.sum > 0:\n self.operation = \"add\"\n else:\n raise RuntimeError\n self.sum = abs(self.sum)\n elif argument.isdigit():\n self.operation = \"set\"\n else:\n raise RuntimeError\n\n\nclass Tokens(commands.Cog):\n \"\"\"Token system\"\"\"\n\n def __init__(self, bot):\n self.bot = bot\n self.config = Config.get_conf(self, identifier=192153481165930496, force_registration=True)\n default_member = {\"balance\": 0}\n self.config.register_member(**default_member)\n\n default_guild = {\"channel\": 0}\n self.config.register_guild(**default_guild)\n\n @guild_only_check()\n @commands.command()\n async def tokens(self, ctx, member: discord.Member = None):\n \"\"\"Checks your Token balance\n Pass a Member to check their balance (Bot Admin only)\"\"\"\n\n if member == None:\n token = await self.config.member(ctx.author).balance()\n await ctx.send(f\"Your current Token Balance is {token}\")\n else:\n auth_checks = (\n await ctx.bot.is_owner(ctx.author),\n await ctx.bot.is_admin(ctx.author),\n ctx.author == ctx.guild.owner,\n )\n if any(auth_checks):\n token = await self.config.member(member).balance()\n await ctx.send(f\"{member.display_name}'s Token Balance is {token}\")\n\n @guild_only_check()\n @checks.admin()\n @commands.command()\n async def tokenset(self, ctx, member: discord.Member, tokens: SetParser):\n \"\"\"Sets, Adds or Subtracts a Member's Token Balance\"\"\"\n\n token = await self.config.member(member).balance()\n channel = await self.config.guild(ctx.guild).channel()\n if tokens.operation == \"add\":\n await self.config.member(member).balance.set(token + tokens.sum)\n await ctx.send(f\"Added {tokens.sum} Tokens to {member.display_name}\")\n if channel != 0:\n await ctx.guild.get_channel(channel).send(\n embed=discord.Embed(\n colour=discord.Colour(0x2ECC71),\n description=f\"{ctx.author.mention} added {tokens.sum} Tokens to {member.mention}\",\n )\n )\n elif tokens.operation == \"sub\":\n if tokens.sum > token:\n await ctx.send(f\"{member.display_name} does not have that many tokens\")\n else:\n await self.config.member(member).balance.set(token - tokens.sum)\n await ctx.send(f\"Removed {tokens.sum} Tokens to {member.display_name}\")\n if channel != 0:\n await ctx.guild.get_channel(channel).send(\n embed=discord.Embed(\n colour=discord.Colour(0xE74C3C),\n description=f\"{ctx.author.mention} removed {tokens.sum} Tokens to {member.mention}\",\n )\n )\n else:\n await self.config.member(member).balance.set(tokens.sum)\n await ctx.send(f\"Set {tokens.sum} Tokens on {member.display_name}\")\n if channel != 0:\n await ctx.guild.get_channel(channel).send(\n embed=discord.Embed(\n colour=discord.Colour(0xF1C40F),\n description=f\"{ctx.author.mention} set {tokens.sum} Tokens on {member.mention}\",\n )\n )\n\n @guild_only_check()\n @checks.admin()\n @commands.command()\n async def tokenlog(self, ctx, channel: discord.TextChannel = None):\n \"\"\"Sets the channel to log the usage of tokenset to.\n Use the command in the channel you want the logging posted in or pass a channel\"\"\"\n\n if channel is None:\n channel = ctx.channel\n\n await self.config.guild(ctx.guild).channel.set(channel.id)\n await ctx.send(f\"Token Logs set to {channel.mention}\")\n \n @guild_only_check()\n @commands.command()\n async def gifttoken(self, ctx, member: discord.Member, amount: int):\n \"\"\"Send Tokens to another Member\"\"\"\n \n balance1 = await self.config.member(ctx.author).balance()\n balance2 = await self.config.member(member).balance()\n channel = await self.config.guild(ctx.guild).channel()\n\n if amount > balance1:\n await ctx.send(f\"Sorry, you cannot gift more tokens than you have\")\n else:\n await self.config.member(ctx.author).balance.set(balance1 - amount)\n await self.config.member(member).balance.set(balance2 + amount)\n await ctx.send(f\"You gifted {amount} Tokens to {member.display_name}\")\n if channel != 0:\n await ctx.guild.get_channel(channel).send(\n embed=discord.Embed(\n colour=discord.Colour(0x3498DB),\n description=f\"{ctx.author.mention} gifted {amount} Tokens to {member.mention}\",\n )\n )\n", "sub_path": "tokens/tokens.py", "file_name": "tokens.py", "file_ext": "py", "file_size_in_byte": 5550, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "redbot.core.commands.Context", "line_number": 6, "usage_type": "attribute"}, {"api_name": "redbot.core.commands", "line_number": 6, "usage_type": "name"}, {"api_name": "redbot.core.commands.check", "line_number": 12, "usage_type": "call"}, {"api_name": "redbot.core.commands", "line_number": 12, "usage_type": "name"}, {"api_name": "redbot.core.commands.Cog", "line_number": 33, "usage_type": "attribute"}, {"api_name": "redbot.core.commands", "line_number": 33, "usage_type": "name"}, {"api_name": "redbot.core.Config.get_conf", "line_number": 38, "usage_type": "call"}, {"api_name": "redbot.core.Config", "line_number": 38, "usage_type": "name"}, {"api_name": "discord.Member", "line_number": 47, "usage_type": "attribute"}, {"api_name": "redbot.core.commands.command", "line_number": 46, "usage_type": "call"}, {"api_name": "redbot.core.commands", "line_number": 46, "usage_type": "name"}, {"api_name": "discord.Member", "line_number": 67, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 77, "usage_type": "call"}, {"api_name": "discord.Colour", "line_number": 78, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 90, "usage_type": "call"}, {"api_name": "discord.Colour", "line_number": 91, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 100, "usage_type": "call"}, {"api_name": "discord.Colour", "line_number": 101, "usage_type": "call"}, {"api_name": "redbot.core.checks.admin", "line_number": 65, "usage_type": "call"}, {"api_name": "redbot.core.checks", "line_number": 65, "usage_type": "name"}, {"api_name": "redbot.core.commands.command", "line_number": 66, "usage_type": "call"}, {"api_name": "redbot.core.commands", "line_number": 66, "usage_type": "name"}, {"api_name": "discord.TextChannel", "line_number": 109, "usage_type": "attribute"}, {"api_name": "redbot.core.checks.admin", "line_number": 107, "usage_type": "call"}, {"api_name": "redbot.core.checks", "line_number": 107, "usage_type": "name"}, {"api_name": "redbot.core.commands.command", "line_number": 108, "usage_type": "call"}, {"api_name": "redbot.core.commands", "line_number": 108, "usage_type": "name"}, {"api_name": "discord.Member", "line_number": 121, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 136, "usage_type": "call"}, {"api_name": "discord.Colour", "line_number": 137, "usage_type": "call"}, {"api_name": "redbot.core.commands.command", "line_number": 120, "usage_type": "call"}, {"api_name": "redbot.core.commands", "line_number": 120, "usage_type": "name"}]}
+{"seq_id": "387209707", "text": "from django.shortcuts import render,get_object_or_404,redirect\nfrom .models import Blog,Comment\nfrom django.utils import timezone\n# Create your views here.\ndef blog(request):\n blogs = Blog.objects.all()\n return render(request,'blog.html', { 'blogs' : blogs })\n\n# R\ndef detail(request, blog_id):\n detail = get_object_or_404(Blog, pk=blog_id)\n comments = Comment.objects.all().filter(post = detail)\n\n\n# likes -> 순서쌍 ('현재 blog.id', '현재 user.id' )\n# 이 순서쌍이 like 모델에 있다면 좋아요를 누른 것! -> 좋아요 취소 message 전달\n# 이 순서쌍이 like 모델에 없다면 좋아요를 누르지 않은 것 -> 좋아요 message 전달 \n\n if detail.likes.filter(id=request.user.id):\n message=\"좋아요 취소\"\n else:\n message=\"좋아요\"\n\n return render(request ,'detail.html', { 'detail' : detail, 'comments' : comments, \"message\" : message } )\n\n\ndef new(request):\n return render(request, 'new.html')\n\ndef create(request):\n blog = Blog() # 객체 틀 하나 가져오기\n blog.title = \"NoTitle\" # 내용 채우기\n if request.GET['title']:\n blog.title=request.GET['title']\n blog.body = request.GET['body'] # 내용 채우기\n blog.pub_date = timezone.datetime.now() # 내용 채우기\n blog.writer = request.user\n blog.save() # 객체 저장하기\n\n # 새로운 글 url 주소로 이동\n return redirect('/blog/' + str(blog.id))\n\n#삭제\ndef delete(request, blog_id):\n blog = get_object_or_404(Blog, pk=blog_id)\n blog.delete()\n return redirect('/blog/')\n#update\n\ndef update(request, blog_id):\n blog = get_object_or_404(Blog, pk =blog_id)\n\n if request.method == \"POST\":\n if request.POST['title']:\n blog.title=request.POST['title']\n blog.body = request.POST['body']\n blog.pub_date = timezone.datetime.now()\n blog.save()\n return redirect('/blog/' +str(blog.id))\n else:\n return render(request,'update.html')\n\n\ndef comment(request,blog_id):\n if request.method == \"POST\" :\n comment = Comment()\n comment.body = request.POST['body']\n comment.pub_date = timezone.datetime.now()\n comment.writer = request.user\n comment.post = get_object_or_404(Blog, pk=blog_id)\n comment.save()\n\n return redirect('/blog/'+str(blog_id))\n else:\n return redirect('/blog/'+str(blog_id))\n\n\ndef comment_delete(request, comment_id):\n comment = get_object_or_404(Comment, pk=comment_id)\n blog_id = comment.post.id\n comment.delete()\n\n return redirect('/blog/'+str(blog_id))\n\n# like 관련 함수\ndef post_like(request, blog_id):\n blog = get_object_or_404(Blog, pk=blog_id)\n user = request.user\n\n if blog.likes.filter(id=user.id):\n blog.likes.remove(user)\n else: \n blog.likes.add(user)\n\n return redirect('/blog/'+str(blog_id))", "sub_path": "skeleton_code/bp/blog/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2870, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "models.Blog.objects.all", "line_number": 6, "usage_type": "call"}, {"api_name": "models.Blog.objects", "line_number": 6, "usage_type": "attribute"}, {"api_name": "models.Blog", "line_number": 6, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 7, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 11, "usage_type": "call"}, {"api_name": "models.Blog", "line_number": 11, "usage_type": "argument"}, {"api_name": "models.Comment.objects.all", "line_number": 12, "usage_type": "call"}, {"api_name": "models.Comment.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "models.Comment", "line_number": 12, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 24, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 28, "usage_type": "call"}, {"api_name": "models.Blog", "line_number": 31, "usage_type": "call"}, {"api_name": "django.utils.timezone.datetime.now", "line_number": 36, "usage_type": "call"}, {"api_name": "django.utils.timezone.datetime", "line_number": 36, "usage_type": "attribute"}, {"api_name": "django.utils.timezone", "line_number": 36, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 41, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 45, "usage_type": "call"}, {"api_name": "models.Blog", "line_number": 45, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 47, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 51, "usage_type": "call"}, {"api_name": "models.Blog", "line_number": 51, "usage_type": "argument"}, {"api_name": "django.utils.timezone.datetime.now", "line_number": 57, "usage_type": "call"}, {"api_name": "django.utils.timezone.datetime", "line_number": 57, "usage_type": "attribute"}, {"api_name": "django.utils.timezone", "line_number": 57, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 59, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 61, "usage_type": "call"}, {"api_name": "models.Comment", "line_number": 66, "usage_type": "call"}, {"api_name": "django.utils.timezone.datetime.now", "line_number": 68, "usage_type": "call"}, {"api_name": "django.utils.timezone.datetime", "line_number": 68, "usage_type": "attribute"}, {"api_name": "django.utils.timezone", "line_number": 68, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 70, "usage_type": "call"}, {"api_name": "models.Blog", "line_number": 70, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 73, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 75, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 79, "usage_type": "call"}, {"api_name": "models.Comment", "line_number": 79, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 83, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 87, "usage_type": "call"}, {"api_name": "models.Blog", "line_number": 87, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 95, "usage_type": "call"}]}
+{"seq_id": "326266032", "text": "from feature_extract import *\nfrom sklearn.decomposition import PCA\nimport random\nimport time\nimport math\nfrom sklearn.model_selection import train_test_split \nfrom sklearn import metrics \nfrom xgboost.sklearn import XGBClassifier \nimport pandas as pd\nimport numpy as np \nfrom scipy.stats import pearsonr\nfrom sklearn.model_selection import cross_val_score\nfrom sklearn.model_selection import cross_val_predict\nimport pylab as plt\n\npca=PCA(n_components=10)\n\n#**********Read the file and output the list in [exoRbase id,circR2disease id] format ************\ncircRNA_id = np.loadtxt(r'.\\data\\exoRBase-circR2disease id conversion.txt',dtype=bytes).astype(str)\n\na=np.array(circRNA_id)\nexoRBase_id_all=a[:,0]\ncircBase_id_all=a[:,1]\n\nall_id=[]\nfor i in range(1000):\n exo_id=exoRBase_id_all[i]\n circ_id=circBase_id_all[i]\n all_id.append([exo_id,circ_id])\n\n\n# **************Delete useless columns, only keep those used for calculation **************\nexoRBase_all= np.loadtxt(r'.\\data\\Normal_circRNA_RPM.txt',dtype=bytes).astype(str)\nexoRBase_for_calculate=np.delete(exoRBase_all,[0,1],1)\nexoRBase_id=exoRBase_all[:,0] \n\nexoR_id_in_Normal_circRNA_RPM=[]\nexoR_id_to_circRNA_id=[]\nfor i in exoRBase_id:\n for j in range(1000):\n if i==all_id[j][0]:\n exoR_id_in_Normal_circRNA_RPM.append(all_id[j])\nnew_all_id=[]\na=0\nfor i in all_id:\n if all_id[a][1]!='NA':\n new_all_id.append(all_id[a])\n a=a+1\n# print(new_all_id)\n \n#*********************Screen out the circRNA that exists in the circRNA number ID.txt file \ncircRNA_number_ID= np.loadtxt(r'.\\data\\circRNA number ID.txt',dtype=bytes).astype(str)\ncircRNA_number_ID=np.delete(circRNA_number_ID,0,1)\n# print(circRNA_number_ID)\ncirc_location=[]\nnew_all_id_justCircID=np.delete(new_all_id,0,1)\nnew_all_id_in_circRNA_number_ID=[]\nfor i in range(604):\n for j in range(824):\n if circRNA_number_ID[i][0]==new_all_id[j][1]:\n circ_location.append(i)\n new_all_id_in_circRNA_number_ID.append(new_all_id[j])\n\n\n\nexoR_id_in_circRNA_number_ID=[] \nfor i in exoRBase_id:\n for j in range(49):\n if i==new_all_id_in_circRNA_number_ID[j][0]:\n exoR_id_in_circRNA_number_ID.append(i)\n\n\nnew_exoR_id_in_circRNA_number_ID=[]\nfor i in exoR_id_in_circRNA_number_ID:\n for j in range(40645):\n if i==exoRBase_all[j][0]:\n new_exoR_id_in_circRNA_number_ID.append(exoRBase_all[j])\nnew_exoR_id_in_circRNA_number_ID=np.delete(new_exoR_id_in_circRNA_number_ID,[0,1],1)\n\nFS_integration_1=np.zeros((604,604))\nfor i in range(49):\n for j in range(49): \n x=list(map(float,new_exoR_id_in_circRNA_number_ID[i]))\n y=list(map(float,new_exoR_id_in_circRNA_number_ID[j]))\n p=pearsonr(x, y)\n a=circ_location[i]\n b=circ_location[j]\n FS_integration_1[a,b] = p[1]\n\nfor i in range(604):\n for j in range(604):\n if FS_integration_1[i,j] >=0.4:\n FS_integration_1[i,j] = FS_integration_1[i,j]\n else:\n FS_integration_1[i,j] = 0\n\ndef mergeToOne(X,X2): \n X3=[] \n for i in range(X.shape[0]): \n tmp=np.array([list(X[i]),list(X2[i])])\n X3.append(list(np.hstack(tmp))) \n X3=np.array(X3) \n return X3 \n\nna = 604 \nnd = 88 \nna = 659 \nr = 0.5 \nnn = 604*88-659 \n\ncircRNAnumbercode = np.loadtxt(r'.\\data\\circRNA number ID.txt',dtype=bytes).astype(str)\ndiseasenumbercode = np.genfromtxt(r'.\\data\\disease number ID.txt',dtype=str,delimiter='\\t')\n\n\ndef Getgauss_circRNA(adjacentmatrix,nc):\n KC = np.zeros((nc,nc))\n gamaa=1\n sumnormm=0\n for i in range(nc):\n normm = np.linalg.norm(adjacentmatrix[i])**2\n sumnormm = sumnormm + normm \n gamam = gamaa/(sumnormm/nc)\n for i in range(nc):\n for j in range(nc):\n KC[i,j]= math.exp (-gamam*(np.linalg.norm(adjacentmatrix[i]-adjacentmatrix[j])**2))\n return KC\n \ndef Getgauss_disease(adjacentmatrix,nd):\n KD = np.zeros((nd,nd))\n gamaa=1\n sumnormd=0\n for i in range(nd):\n normd = np.linalg.norm(adjacentmatrix[:,i])**2\n sumnormd = sumnormd + normd\n gamad=gamaa/(sumnormd/nd)\n for i in range(nd):\n for j in range(nd):\n KD[i,j]= math.exp(-(gamad*(np.linalg.norm(adjacentmatrix[:,i]-adjacentmatrix[:,j])**2)))\n return KD\n\n\nA = np.zeros((nc,nd),dtype=float)\nConnectDate = np.loadtxt(r'.\\data\\known disease-circRNA association number ID.txt',dtype=int)-1 \nfor i in range(na):\n A[ConnectDate[i,0], ConnectDate[i,1]] = 1 \n\ndataset_n = np.argwhere(A == 0)\nTrainset_p = np.argwhere(A == 1)\ndisease_sm = np.loadtxt(r'.\\data\\disease_sm.txt',dtype=int)\nFS_integration = Getgauss_circRNA(A,nc)\nFS_integration = 0.5*FS_integration+0.5*FS_integration_1\nDS_integration = 0.5*Getgauss_disease(A,nd)+0.5*disease_sm \n\n\ncircRNAFeature,DiseaseFeature,numberOfDiseaseNeighborAssociations,\\\nnumberOfcircRNANeighborAssociations = threetypes_features(nc,nd,A,FS_integration,DS_integration)\npredict_0 =np.zeros((dataset_n.shape[0]+Trainset_p.shape[0]))\nTrainset_n = dataset_n[random.sample(list(range(nn)),na)]\nTrainset= np.vstack((Trainset_n,Trainset_p)) \n \n \nTraincircRNAFeature = circRNAFeature[Trainset[:,0]]\nTrainDiseaseFeature = DiseaseFeature[Trainset[:,1]]\n\ncircRNANumberNeighborTrain = numberOfcircRNANeighborAssociations[Trainset[:,0],Trainset[:,1]]\nDiseaseNumberNeighborTrain = numberOfDiseaseNeighborAssociations[Trainset[:,0],Trainset[:,1]]\n \nTraincircRNAFeatureOfPair = np.hstack((TraincircRNAFeature, DiseaseNumberNeighborTrain.reshape(DiseaseNumberNeighborTrain.shape[0],1)))\nPCA_TraincircRNAFeatureOfPair = pca.fit_transform(TraincircRNAFeatureOfPair)\nTrainDiseaseFeatureOfPair = np.hstack((TrainDiseaseFeature, circRNANumberNeighborTrain.reshape(circRNANumberNeighborTrain.shape[0],1)))\nPCA_TrainDiseaseFeatureOfPair = pca.transform(TrainDiseaseFeatureOfPair)\n\nX_train = np.hstack((PCA_TraincircRNAFeatureOfPair,PCA_TrainDiseaseFeatureOfPair))\nY_value=[]\nfor i in range(Trainset_n.shape[0]):\n Y_value.append(0.0)\nfor i in range(Trainset_n.shape[0],Trainset.shape[0]):\n Y_value.append(1.0)\n\nX1_train, X1_test, y1_train, y1_test = train_test_split(X_train, Y_value, test_size=0.3, random_state=0)\n\nclf = XGBClassifier(\n learning_rate =0.2, \n n_estimators=200, \n max_depth=8, \n min_child_weight=10, \n gamma=0.5, \n subsample=0.75, \n colsample_bytree=0.75, \n objective= 'binary:logistic', \n nthread=8, \n scale_pos_weight=1, \n reg_alpha=1e-05, \n reg_lambda=10, \n seed=1024) \n \nclf.fit(X1_train, y1_train) \n\n\nnew_feature= clf.apply(X1_train)\nnew_feature2=clf.apply(X1_test)\nnew_feature_all=clf.apply(X_train)\nX_train_new=mergeToOne(X1_train,new_feature)\nX_train_new2=mergeToOne(X1_test,new_feature2)\nX_train_all=mergeToOne(X_train,new_feature_all)\nmodel = XGBClassifier( \n learning_rate =0.3, \n n_estimators=200, \n max_depth=5, \n min_child_weight=1, \n gamma=0.5, \n subsample=0.8, \n colsample_bytree=0.8, \n objective= 'binary:logistic', \n nthread=8, \n scale_pos_weight=1, \n reg_alpha=1e-05, \n reg_lambda=1, \n seed=1024) \nmodel.fit(X_train_new, y1_train)\n\npredict_0 = model.predict_proba(X_test_new)[:,1]\npredict_0scoreranknumber =np.argsort(-predict_0)\npredict_0scorerank = predict_0[predict_0scoreranknumber]\ndiseaserankname_pos = Trainset[predict_0scoreranknumber,1]\ndiseaserankname = diseasenumbercode[diseaserankname_pos,1]\ncircRNArankname_pos = Trainset[predict_0scoreranknumber,0]\ncircRNArankname = circRNAnumbercode[circRNArankname_pos,1]\npredict_0scorerank_pd=pd.Series(predict_0scorerank)\ndiseaserankname_pd=pd.Series(diseaserankname)\ncircRNArankname_pd=pd.Series(circRNArankname)\nprediction_0_out = pd.concat([diseaserankname_pd,circRNArankname_pd,predict_0scorerank_pd],axis=1)\nprediction_0_out.columns=['Disease','circRNA','Score']\nprediction_0_out.to_excel(r'prediction results for all unknown samples.xlsx', sheet_name='Sheet1',index=False)\n\n\n \n\n\n\n\n\n\n \n \n\n \n \n \n \n \n \n\n\n", "sub_path": "XGBCDA.py", "file_name": "XGBCDA.py", "file_ext": "py", "file_size_in_byte": 8066, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sklearn.decomposition.PCA", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 80, "usage_type": "call"}, {"api_name": "scipy.stats.pearsonr", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.genfromtxt", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 120, "usage_type": "attribute"}, {"api_name": "math.exp", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 125, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 133, "usage_type": "attribute"}, {"api_name": "math.exp", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 138, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 142, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 157, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 170, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 173, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 180, "usage_type": "call"}, {"api_name": "xgboost.sklearn.XGBClassifier", "line_number": 182, "usage_type": "call"}, {"api_name": "xgboost.sklearn.XGBClassifier", "line_number": 206, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 223, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 229, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 230, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 231, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 232, "usage_type": "call"}]}
+{"seq_id": "190276482", "text": "#!/usr/local/bin/python3\n\nimport numpy as np\nimport pandas as pd\nimport healpy as hp\n\npart = 'I'\nr_sc = 8.2 # 8.2 kpc\nsize = 0.1\nrad = 45 # 45 for I, 300 for II, 600 for III\nnumber = int(2*rad/size)\ndenGrid = pd.read_csv('IDenGrid.csv')\nprint('Step 1')\nxGrid = np.round(np.linspace(-rad + size/2, rad - size/2, number), decimals = 2)\nyGrid = np.round(np.linspace(-rad + size/2, rad - size/2, number), decimals = 2)\nzGrid = np.round(np.linspace(-rad + size/2, rad - size/2, number), decimals = 2)\nGridy, Gridx, Gridz = np.meshgrid(yGrid, xGrid, zGrid)\n# Set an empty meshgrid for density \nrho = Gridx*0\nflag = 0\nfor i in range(number):\n for j in range(number):\n for k in range(number):\n # rho[i][j][k] means density in different point\n # The iteration will extract density information from the top to the bottom\n # and the program technically works\n rho[i][j][k] = denGrid.loc[flag, 'rho']\n flag +=1\n if flag%10000 == 0:\n print(flag)\n\nfrom scipy.interpolate import RegularGridInterpolator\nprofile = RegularGridInterpolator((xGrid, yGrid, zGrid), rho)\nprint('Step 2')\n\ndftmp = pd.read_csv('hpLumin.csv')\ndftmp.drop(['r2', 'r3'], axis = 1, inplace = True)\nerror = []\nlum = {}\nlum2 = {}\nPsi = np.array([3*np.pi/4, np.pi, 5*np.pi/4, 3*np.pi/2, 7*np.pi/4])\nfor psi in Psi:\n lum2[int(360*psi/(2*np.pi))] = []\n lum[int(360*psi/(2*np.pi))] = []\n\ndef rot(xx, yy, zz, theta):\n x = xx*np.cos(theta) + yy*np.sin(theta)\n y = -xx*np.sin(theta) + yy*np.cos(theta)\n z = zz\n return [x,y,z]\n\nprint('Step 3')\nfor i in range(len(dftmp)):\n ll = dftmp.loc[i, 'long']\n bb = dftmp.loc[i, 'lat']\n ll = 2*np.pi*ll/360\n bb = 2*np.pi*bb/360\n r0 = dftmp.loc[i,'r1']\n\n r1 = (r0//0.001)/1000 - 0.7*size # round down to 3 decimals\n rnum = int(r0//size) + 2\n r = np.delete(np.linspace(0, r1, rnum), 0)\n deltar = r[0]\n for psi in Psi:\n rr = pd.DataFrame(r, columns=['r'])\n lsr = np.array([-r_sc*np.cos(psi), -r_sc*np.sin(psi)])\n ## 1. Galactic coordinate to cartesian\n rr['x'] = r * np.cos(bb) * np.cos(ll)\n rr['y'] = r * np.cos(bb) * np.sin(ll)\n rr['z'] = r * np.sin(bb)\n ## 2. Rotate counter clockwise psi\n rr['x'], rr['y'], rr['z'] = rot(rr['x'], rr['y'], rr['z'], -psi)\n ## 3. Transfer to Galactic center coordinate\n rr['x'] = rr['x'] + lsr[0]\n rr['y'] = rr['y'] + lsr[1]\n try:\n rr['lum2'] = profile(rr.loc[:,['x','y','z']])**2 * deltar\n rr['lum'] = profile(rr.loc[:,['x','y','z']]) * deltar\n lum2[int(360*psi/(2*np.pi))].append(np.log10(rr['lum2'].sum()))\n lum[int(360*psi/(2*np.pi))].append(np.log10(rr['lum'].sum()))\n except:\n error.append('Error '+str(i) + ' ' + str(int(360*psi/(2*np.pi))))\n lum2[int(360*psi/(2*np.pi))].append(np.nan)\n lum[int(360*psi/(2*np.pi))].append(np.nan)\n del rr\n if i%10000 == 0:\n print(i)\n\ndftmp.drop(['r1'], axis = 1, inplace = True)\nfor psi in Psi:\n dftmp['lum2 '+str(int(360*psi/(2*np.pi)))] = lum2[int(360*psi/(2*np.pi))]\n dftmp['lum '+str(int(360*psi/(2*np.pi)))] = lum[int(360*psi/(2*np.pi))]\n\ndftmp.to_csv(r'./Icur2.csv', index=False)\n", "sub_path": "S_to_N_curve/hp256/Icur2.py", "file_name": "Icur2.py", "file_ext": "py", "file_size_in_byte": 3273, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pandas.read_csv", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 17, "usage_type": "call"}, {"api_name": "scipy.interpolate.RegularGridInterpolator", "line_number": 33, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 43, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 44, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 56, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 57, "usage_type": "attribute"}, {"api_name": "numpy.delete", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 62, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 79, "usage_type": "attribute"}, {"api_name": "numpy.log10", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 80, "usage_type": "attribute"}, {"api_name": "numpy.log10", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 82, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 83, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 83, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 84, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 84, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 91, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 92, "usage_type": "attribute"}]}
+{"seq_id": "87303172", "text": "# %load q03_skewness_log/build.py\nfrom scipy.stats import skew\nimport pandas as pd\nimport numpy as np\n\ndata = pd.read_csv('data/train.csv')\n\n\n# Write code here:\ndef skewness_log(df):\n logsaleprice = np.log(df['SalePrice'])\n loggrlivarea = np.log(df['GrLivArea'])\n \n skewsaleprice = skew(logsaleprice)\n skewgrlivarea = skew(loggrlivarea)\n \n return skewgrlivarea, skewsaleprice\n\nskewness_log(data)\n\n", "sub_path": "q03_skewness_log/build.py", "file_name": "build.py", "file_ext": "py", "file_size_in_byte": 418, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pandas.read_csv", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 12, "usage_type": "call"}, {"api_name": "scipy.stats.skew", "line_number": 14, "usage_type": "call"}, {"api_name": "scipy.stats.skew", "line_number": 15, "usage_type": "call"}]}
+{"seq_id": "644695865", "text": "\"\"\"This script simultaneously plot the mesh grid and the selected tracks,\nat either single time point in rotating 3D axes, or single view in time series,\nor time series with rotating 3D axes, save a movie for presentation.\n\"\"\"\nARROW_LENGTH = 2.0 # magnification fold of plotted arrow length\nVOXEL_DEPTH = 2.0 # in micron\nTIME_INTERVAL = 5.0 # in minutes\nW, H = 20, 20 # fig size in inches\n\nimport matplotlib as mpl\nmpl.use(\"Agg\")\n# mpl.use(\"Qt5Agg\")\n# AXES_COLOR = '#888888'# grey\nAXES_COLOR = '#FFFFFF'# white\nmpl.rc('figure', facecolor='k', edgecolor=AXES_COLOR)\nmpl.rc('axes', facecolor='k', edgecolor=AXES_COLOR, labelcolor=AXES_COLOR)\nmpl.rc('xtick', color=AXES_COLOR)\nmpl.rc('ytick', color=AXES_COLOR)\n# mpl.rc('grid', color='k')\nimport matplotlib.pyplot as plt\nimport matplotlib.animation as animation\nfrom matplotlib import cm\nfrom mpl_toolkits.mplot3d import Axes3D\nimport shutil, os\nimport numpy as np\nfrom shaohe_tracking_lib import getDisFoot, decomposeVector\nimport pandas as pd\n\ntMeshData = 'combinedMeshDataSorted.csv'\ntMeshDF = pd.read_csv(tMeshData)\nprint('timeLapseMeshData', tMeshDF.columns.values)\n# calculate the range, shifting values for putting the plot at (0,0,0) and\n# scaling factors to maintain the aspect ratios\ntMin, tMax = tMeshDF.t.min(), tMeshDF.t.max()\nxMin, xMax = tMeshDF.x.min(), tMeshDF.x.max()\nyMin, yMax = tMeshDF.y.min(), tMeshDF.y.max()\nzMin, zMax = tMeshDF.z.min(), tMeshDF.z.max()\nscale_x, scale_y, scale_z = xMax - xMin, yMax - yMin, zMax - zMin\nscale_range = max(scale_x, scale_y, scale_z)\nscale_x, scale_y, scale_z = scale_x/scale_range, scale_y/scale_range, scale_z/scale_range\nprint('x, y, z scales: ', scale_x, scale_y, scale_z)\nlenCorr = 1 / np.sum(i**2 for i in [scale_x, scale_y, scale_z])**.5\nprint('Length correction factor: ', lenCorr)\nshift_x, shift_y, shift_z = np.mean([xMax, xMin]), np.mean([yMax, yMin]), np.mean([zMax, zMin])\n\n# # read and format the track data from Imaris exported csv\n# trackData = '180325-mTmGHisG-SMG2-combined-ROI1_t001_z001_c001_Position.csv'\n# trackDF = pd.read_csv(trackData, skiprows=3)\n# # ['Position X', 'Position Y', 'Position Z', 'Unit', 'Category',\n# # 'Collection', 'Time', 'TrackID', 'ID', 'Unnamed: 9']\n# # remove the useless columns\n# trackDF.drop(['Unit', 'Category', 'Collection', 'Unnamed: 9'], axis=1, inplace=True)\n# trackDF.columns = ['x', 'y', 'z', 't', 'trackID', 'spotID']\ntrackData = 'selected-171030 03_SMG1bud3-1 cleft initiation_Position.csv'\ntrackDF = pd.read_csv(trackData)\nprint('Track data columns: ', trackDF.columns.values)\n\n# plotting\nfig = plt.figure(figsize=(W, H), dpi=72)\nax = fig.add_axes([0, 0, 1.0, 1.0], projection='3d')\n# Set the background color of the panes\nPANECOLOR = (0.1, 0.1, 0.1, 1.0)\nax.w_xaxis.set_pane_color(PANECOLOR)\nax.w_yaxis.set_pane_color(PANECOLOR)\nax.w_zaxis.set_pane_color(PANECOLOR)\n# Set the line colors of x,y,z axes\nAXISCOLOR = (1.0, 1.0, 1.0, 1.0)\nax.w_xaxis.line.set_color(AXISCOLOR)\nax.w_yaxis.line.set_color(AXISCOLOR)\nax.w_zaxis.line.set_color(AXISCOLOR)\n# plt.show()\nax.get_proj = lambda: np.dot(Axes3D.get_proj(ax), np.diag([scale_x, scale_y, scale_z, lenCorr]))\n\n# create one folder inside the temp folder for each desired view angle\nelevation = [0, 15, 30, 60, 90]\nazimuth = [30, -150]\n\nif not os.path.exists('./temp/'):\n os.mkdir('./temp/')\nfor azi in azimuth:\n for elev in elevation:\n tempPath = './temp/azimuth_'+str(azi)+'-elevation_'+str(elev)+'/'\n if not os.path.exists(tempPath):\n os.mkdir(tempPath)\n\n\niii = 0 # counter for testing runs\nfor imgIndex, t in enumerate(np.linspace(tMin, tMax, tMax-tMin+1)):\n # iii += 1 # counter for testing runs\n # if iii > 5: # for testing so that only this many loops are executed\n # break\n trackDF_t2 = trackDF.loc[trackDF.t == t + 1]\n# # to reverse time, also remember to reverse the sign in trackDF_t2!\n# for imgIndex, t in enumerate(np.linspace(tMax, tMin, tMax-tMin+1)): # for reversing time\n# trackDF_t2 = trackDF.loc[trackDF.t == t - 1] # for reversed time!\n trackDF_t = trackDF.loc[trackDF.t == t]\n tMeshDF_t = tMeshDF.loc[tMeshDF.t == t]\n # plot spots and arrows\n for i in trackDF_t.trackID.unique():\n temp = trackDF_t.loc[trackDF_t.trackID == i]\n x, y, z = temp.x.values, temp.y.values, temp.z.values\n spotCoor = np.c_[x, y, z].flatten()\n _, footCoor = getDisFoot(spotCoor, tMeshDF_t)\n ax.plot(x-shift_x, y-shift_y, z-shift_z, 'o', color='#00FFFF', alpha=1.0)\n temp2 = trackDF_t2.loc[trackDF_t2.trackID == i]\n if len(temp2) == 0:\n u, v, w = 0, 0, 0\n u1, v1, w1 = 0, 0, 0\n u2, v2, w2 = 0, 0, 0\n else:\n vArrow = [temp2.x.values-x, temp2.y.values-y, temp2.z.values-z]\n vPara, vPerpend = decomposeVector(footCoor - spotCoor, vArrow)\n u, v, w = vArrow[0], vArrow[1], vArrow[2]\n u1, v1, w1 = vPara[0], vPara[1], vPara[2]\n u2, v2, w2 = vPerpend[0], vPerpend[1], vPerpend[2]\n ax.quiver(x-shift_x, y-shift_y, z-shift_z, u, v, w, color='#FFFF00', alpha=.5,\\\n length=ARROW_LENGTH, normalize=False)\n ax.quiver(x-shift_x, y-shift_y, z-shift_z, u1, v1, w1, color='#00FF00', alpha=.5,\\\n length=ARROW_LENGTH, normalize=False)\n ax.quiver(x-shift_x, y-shift_y, z-shift_z, u2, v2, w2, color='#FF0000', alpha=1.0,\\\n length=ARROW_LENGTH, normalize=False)\n # ax.quiver(x, y, z, u, v, w, color='#FF00FF')\n # plot the vertical lines of mesh\n for i in tMeshDF_t.pathPos.unique():\n temp = tMeshDF_t.loc[tMeshDF_t.pathPos == i]\n x, y, z = temp.x.values, temp.y.values, temp.z.values\n ax.plot(x-shift_x, y-shift_y, z-shift_z, '-', color='#00FF00', alpha=.1)\n # plot the horizontal lines pf mesh\n for i in tMeshDF_t.z.unique():\n temp = tMeshDF_t.loc[tMeshDF_t.z == i]\n x, y, z = temp.x.values, temp.y.values, temp.z.values\n ax.plot(x-shift_x, y-shift_y, z-shift_z, '-', color='#00FF00', alpha=.1)\n\n # plot tracks\n # backward dragon tail\n trackToPlotDF = trackDF.loc[trackDF.t > t - 10]\n trackToPlotDF = trackToPlotDF.loc[trackToPlotDF.t <= t]\n for i in trackToPlotDF.trackID.unique():\n temp = trackToPlotDF.loc[trackToPlotDF.trackID == i]\n temp = temp.sort_values('t', ascending=True)\n x, y, z = temp.x.values, temp.y.values, temp.z.values\n # ax.plot(x-shift_x, y-shift_y, z-shift_z, '-', color='#00FF00', alpha=.8)\n ax.plot(x-shift_x, y-shift_y, z-shift_z, '-', color='#FFFFFF', alpha=.3)\n\n # adjust axis limits and turn on/off grids -- has to be here\n ax.set_xlim(xMin - shift_x, xMax - shift_x)\n ax.set_ylim(yMin - shift_y, yMax - shift_y)\n ax.set_zlim(zMin - shift_z, zMax - shift_z)\n ax.grid(False)\n\n for azi in azimuth:\n for elev in elevation:\n tempPath = './temp/azimuth_'+str(azi)+'-elevation_'+str(elev)+'/'\n ax.view_init(elev, azi)\n tempFile = tempPath + \"%04d\"%imgIndex + '.png'\n plt.savefig(tempFile)\n\n ax.clear()\n\n# plt.show()\n\n# the following code use ffmpeg to generate a few movies\nfps = 12\n# fps = 24\nfor azi in azimuth:\n for elev in elevation:\n tempPath = './temp/azimuth_'+str(azi)+'-elevation_'+str(elev)+'/'\n outMovie = './azimuth_'+str(azi)+'-elevation_'+str(elev)+'.mp4'\n os.system('ffmpeg -r '+str(fps)+' -i '+tempPath+'%04d.png -b:v 800k\\\n -vcodec mpeg4 -q:v 0 -y '+outMovie)\n\n# # the following code use ffmpeg to generate a movie\n# outMovie = './tempMovie-800k.mp4'\n# os.system('ffmpeg -r 6 -i '+tempPath+'%04d.png -b:v 800k\\\n# -vcodec mpeg4 -q:v 0 -y '+outMovie)\n# # shutil.rmtree(tempPath)\n\n# # the following code use fiji to generate a movie -- not as good as the ffmpeg method!\n# outMovie = './tempMovie.avi'\n# os.system('ImageJ-macosx --ij2 --headless -macro makeMovieFromImgSeq.ijm '+outMovie)\n# shutil.rmtree(tempPath)\n", "sub_path": "trackingAnalysis/plot3DmeshAndTracks.py", "file_name": "plot3DmeshAndTracks.py", "file_ext": "py", "file_size_in_byte": 7950, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "matplotlib.use", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 18, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 44, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "numpy.dot", "line_number": 72, "usage_type": "call"}, {"api_name": "mpl_toolkits.mplot3d.Axes3D.get_proj", "line_number": 72, "usage_type": "call"}, {"api_name": "mpl_toolkits.mplot3d.Axes3D", "line_number": 72, "usage_type": "name"}, {"api_name": "numpy.diag", "line_number": 72, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path", "line_number": 78, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.c_", "line_number": 102, "usage_type": "attribute"}, {"api_name": "shaohe_tracking_lib.getDisFoot", "line_number": 103, "usage_type": "call"}, {"api_name": "shaohe_tracking_lib.decomposeVector", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 156, "usage_type": "name"}, {"api_name": "os.system", "line_number": 169, "usage_type": "call"}]}
+{"seq_id": "396951593", "text": "# -*- coding: UTF-8 -*-\n\nimport os\nimport json\n\nsrc_dir = './src'\ndist_dir = './dist'\n\ndef load_json(file_name, file_path):\n\t# 读取json文件内容,返回字典格式\n\twith open(file_name, 'r', encoding='utf8') as f:\n\t\tjson_data = json.load(f)\n\t\tjson_data = dict(json_data)\n\t\twrite_info(file_path, json_data)\n\ndef write_info(file_name, file_info):\n\t# 写入文件\n\twith open(file_name, 'w') as fp:\n\t\tjson.dump(file_info, fp, indent=4, sort_keys=True)\n\ndef main():\n\t# 判断文件夹是否存在\n\tif not os.path.exists(dist_dir):\n\t\tos.makedirs(dist_dir)\n\n\t# 需要格式化数据\n\tpages = os.listdir(src_dir)\n\tfor page_name in pages:\n\t\t# 源路径\n\t\tsrc_path = os.path.join(src_dir, page_name)\n\t\t# 写入路径\n\t\tfile_path = os.path.join(dist_dir, page_name)\n\t\tload_json(src_path, file_path)# 源路径\n\t\tprint(\"format JSON Sort => {}\".format(file_path))\n\n\tprint(\"format complete !\")\n\nif __name__ == \"__main__\":\n\tmain()", "sub_path": "json/format.py", "file_name": "format.py", "file_ext": "py", "file_size_in_byte": 920, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "json.load", "line_number": 12, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 24, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}]}
+{"seq_id": "276031423", "text": "\"\"\"\nFactories for models used in Partaj tests.\n\"\"\"\nfrom datetime import timedelta\nfrom io import BytesIO\nfrom random import randrange\n\nfrom django.contrib.auth import get_user_model\nfrom django.core.files.base import File\n\nimport factory\n\nfrom . import models\n\n\nclass UserFactory(factory.django.DjangoModelFactory):\n \"\"\"Create users for test purposes.\"\"\"\n\n class Meta:\n model = get_user_model()\n\n email = factory.Faker(\"email\")\n first_name = factory.Faker(\"first_name\")\n last_name = factory.Faker(\"last_name\")\n phone_number = factory.Faker(\"phone_number\")\n title = factory.Faker(\"prefix\")\n unit_name = factory.Faker(\"company\")\n username = factory.Faker(\"email\")\n\n\nclass UnitFactory(factory.django.DjangoModelFactory):\n \"\"\"Create units for test purposes.\"\"\"\n\n class Meta:\n model = models.Unit\n\n name = factory.Faker(\"company\")\n\n\nclass UnitMembershipFactory(factory.django.DjangoModelFactory):\n \"\"\"Create unit memberships for test purposes.\"\"\"\n\n class Meta:\n model = models.UnitMembership\n\n role = factory.Faker(\"word\", ext_word_list=models.UnitMembershipRole.values)\n user = factory.SubFactory(UserFactory)\n unit = factory.SubFactory(UnitFactory)\n\n\nclass UnitMemberFactory(UserFactory):\n \"\"\"Create unit members for test purposes.\"\"\"\n\n class Meta:\n model = get_user_model()\n\n membership = factory.RelatedFactory(UnitMembershipFactory, \"user\")\n\n\nclass TopicFactory(factory.django.DjangoModelFactory):\n \"\"\"Create topics for test purposes.\"\"\"\n\n class Meta:\n model = models.Topic\n\n # pylint: disable=no-member\n name = factory.Faker(\"text\", max_nb_chars=models.Topic.name.field.max_length)\n unit = factory.SubFactory(UnitFactory)\n\n\nclass ReferralUrgencyFactory(factory.django.DjangoModelFactory):\n \"\"\"Create referral urgencies for test purposes.\"\"\"\n\n class Meta:\n model = models.ReferralUrgency\n\n # pylint: disable=no-member\n name = factory.Faker(\n \"text\", max_nb_chars=models.ReferralUrgency.name.field.max_length\n )\n is_default = factory.Faker(\"boolean\")\n requires_justification = factory.Faker(\"boolean\")\n\n @factory.lazy_attribute\n def duration(self):\n \"\"\"\n Generate a random duration for the urgency level.\n \"\"\"\n return timedelta(days=randrange(2, 30))\n\n\nclass ReferralFactory(factory.django.DjangoModelFactory):\n \"\"\"Create referrals for test purposes.\"\"\"\n\n class Meta:\n model = models.Referral\n\n context = factory.Faker(\"text\", max_nb_chars=500)\n prior_work = factory.Faker(\"text\", max_nb_chars=500)\n question = factory.Faker(\"text\", max_nb_chars=500)\n requester = factory.Faker(\"name\")\n topic = factory.SubFactory(TopicFactory)\n urgency_level = factory.SubFactory(ReferralUrgencyFactory)\n user = factory.SubFactory(UserFactory)\n\n @factory.lazy_attribute\n def urgency_explanation(self):\n \"\"\"\n Only generate an explanation if the urgency level requires it.\n \"\"\"\n return (\n factory.Faker(\"text\", max_nb_chars=500).generate()\n if self.urgency_level.requires_justification\n else \"\"\n )\n\n @factory.post_generation\n def post(referral, create, extracted, **kwargs):\n \"\"\"\n Add the topic's linked unit to the units linked to the referral.\n \"\"\"\n referral.units.add(referral.topic.unit)\n referral.save()\n\n\nclass ReferralActivityFactory(factory.django.DjangoModelFactory):\n \"\"\"Create referral activities for test purposes.\"\"\"\n\n class Meta:\n model = models.ReferralActivity\n\n actor = factory.SubFactory(UserFactory)\n referral = factory.SubFactory(ReferralFactory)\n verb = factory.Faker(\"word\", ext_word_list=models.ReferralActivityVerb.values)\n\n @factory.post_generation\n def post(activity, create, extracted, **kwargs):\n \"\"\"\n Generate a content object matching the verb on the referral activity.\n \"\"\"\n if activity.verb in [\n models.ReferralActivityVerb.ASSIGNED,\n models.ReferralActivityVerb.UNASSIGNED,\n ]:\n # pylint: disable=attribute-defined-outside-init\n activity.item_content_object = UserFactory()\n elif activity.verb in [\n models.ReferralActivityVerb.DRAFT_ANSWERED,\n models.ReferralActivityVerb.ANSWERED,\n ]:\n # pylint: disable=attribute-defined-outside-init\n activity.item_content_object = ReferralAnswerFactory(\n referral=activity.referral\n )\n elif activity.verb in [\n models.ReferralActivityVerb.VALIDATED,\n models.ReferralActivityVerb.VALIDATION_DENIED,\n ]:\n # pylint: disable=attribute-defined-outside-init\n activity.item_content_object = ReferralAnswerValidationResponseFactory()\n elif activity.verb == models.ReferralActivityVerb.VALIDATION_REQUESTED:\n # pylint: disable=attribute-defined-outside-init\n activity.item_content_object = ReferralAnswerValidationRequestFactory()\n elif activity.verb == models.ReferralActivityVerb.CREATED:\n pass\n else:\n raise Exception(f\"Incorrect activity verb {activity.verb}\")\n\n\nclass ReferralAnswerFactory(factory.django.DjangoModelFactory):\n \"\"\"Create referral answers for test purposes.\"\"\"\n\n class Meta:\n model = models.ReferralAnswer\n\n content = factory.Faker(\"text\", max_nb_chars=500)\n created_by = factory.SubFactory(UserFactory)\n referral = factory.SubFactory(ReferralFactory)\n state = factory.Faker(\"word\", ext_word_list=models.ReferralAnswerState.values)\n\n\nclass ReferralAnswerAttachmentFactory(factory.django.DjangoModelFactory):\n \"\"\"Create referral answer attachments for test purposes.\"\"\"\n\n class Meta:\n model = models.ReferralAnswerAttachment\n\n name = factory.Faker(\"text\", max_nb_chars=200)\n\n @factory.lazy_attribute\n def file(self):\n \"\"\"\n Create a bogus file field on the answer.\n \"\"\"\n file = BytesIO(b\"the_file\")\n file.name = \"the file name\"\n return File(file)\n\n @factory.post_generation\n def post(answer, create, extracted, **kwargs):\n \"\"\"\n Make sure the size on the answer field matches the actual size of the file.\n \"\"\"\n # pylint: disable=attribute-defined-outside-init\n answer.size = answer.file.size\n\n\nclass ReferralMessageFactory(factory.django.DjangoModelFactory):\n \"\"\"Create referral messages for test purposes.\"\"\"\n\n class Meta:\n model = models.ReferralMessage\n\n content = factory.Faker(\"text\", max_nb_chars=500)\n referral = factory.SubFactory(ReferralFactory)\n user = factory.SubFactory(UserFactory)\n\n\nclass ReferralMessageAttachmentFactory(factory.django.DjangoModelFactory):\n \"\"\"Create referral message attachments for test purposes.\"\"\"\n\n class Meta:\n model = models.ReferralMessageAttachment\n\n name = factory.Faker(\"file_name\")\n\n @factory.lazy_attribute\n def file(self):\n \"\"\"\n Create a bogus file field on the message.\n \"\"\"\n file = BytesIO(b\"the_file\")\n file.name = self.name\n return File(file)\n\n @factory.post_generation\n def post(referral_message, create, extracted, **kwargs):\n \"\"\"\n Make sure the size on the message field matches the actual size of the file.\n \"\"\"\n # pylint: disable=attribute-defined-outside-init\n referral_message.size = referral_message.file.size\n\n\nclass ReferralAnswerValidationRequestFactory(factory.django.DjangoModelFactory):\n \"\"\"Create referral answer validation requests for test purposes.\"\"\"\n\n class Meta:\n model = models.ReferralAnswerValidationRequest\n\n answer = factory.SubFactory(ReferralAnswerFactory)\n validator = factory.SubFactory(UserFactory)\n\n\nclass ReferralAnswerValidationResponseFactory(factory.django.DjangoModelFactory):\n \"\"\"Create referral answer validation responses for test purposes.\"\"\"\n\n class Meta:\n model = models.ReferralAnswerValidationResponse\n\n comment = factory.Faker(\"text\", max_nb_chars=500)\n state = factory.Faker(\n \"word\", ext_word_list=models.ReferralAnswerValidationResponseState.values\n )\n validation_request = factory.SubFactory(ReferralAnswerValidationRequestFactory)\n\n\nclass ReferralAssignmentFactory(factory.django.DjangoModelFactory):\n \"\"\"Create referral assignments for test purposes.\"\"\"\n\n class Meta:\n model = models.ReferralAssignment\n\n referral = factory.SubFactory(ReferralFactory)\n unit = factory.SubFactory(UnitFactory)\n\n @factory.lazy_attribute\n def assignee(self):\n \"\"\"\n Generate a membership to the unit with a brand new user and make this new user\n the assignee.\n \"\"\"\n membership = UnitMembershipFactory(unit=self.unit)\n return membership.user\n\n @factory.lazy_attribute\n def created_by(self):\n \"\"\"\n Generate a membership to the unit with a brand new user and have this news user\n be the the assignment creator.\n \"\"\"\n membership = UnitMembershipFactory(\n unit=self.unit, role=models.UnitMembershipRole.OWNER\n )\n return membership.user\n", "sub_path": "src/backend/partaj/core/factories.py", "file_name": "factories.py", "file_ext": "py", "file_size_in_byte": 9257, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "factory.django", "line_number": 16, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.get_user_model", "line_number": 20, "usage_type": "call"}, {"api_name": "factory.Faker", "line_number": 22, "usage_type": "call"}, {"api_name": "factory.Faker", "line_number": 23, "usage_type": "call"}, {"api_name": "factory.Faker", "line_number": 24, "usage_type": "call"}, {"api_name": "factory.Faker", "line_number": 25, "usage_type": "call"}, {"api_name": "factory.Faker", "line_number": 26, "usage_type": "call"}, {"api_name": "factory.Faker", "line_number": 27, "usage_type": "call"}, {"api_name": "factory.Faker", "line_number": 28, "usage_type": "call"}, {"api_name": "factory.django", "line_number": 31, "usage_type": "attribute"}, {"api_name": "factory.Faker", "line_number": 37, "usage_type": "call"}, {"api_name": "factory.django", "line_number": 40, "usage_type": "attribute"}, {"api_name": "factory.Faker", "line_number": 46, "usage_type": "call"}, {"api_name": "factory.SubFactory", "line_number": 47, "usage_type": "call"}, {"api_name": "factory.SubFactory", "line_number": 48, "usage_type": "call"}, {"api_name": "django.contrib.auth.get_user_model", "line_number": 55, "usage_type": "call"}, {"api_name": "factory.RelatedFactory", "line_number": 57, "usage_type": "call"}, {"api_name": "factory.django", "line_number": 60, "usage_type": "attribute"}, {"api_name": "factory.Faker", "line_number": 67, "usage_type": "call"}, {"api_name": "factory.SubFactory", "line_number": 68, "usage_type": "call"}, {"api_name": "factory.django", "line_number": 71, "usage_type": "attribute"}, {"api_name": "factory.Faker", "line_number": 78, "usage_type": "call"}, {"api_name": "factory.Faker", "line_number": 81, "usage_type": "call"}, {"api_name": "factory.Faker", "line_number": 82, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 89, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 89, "usage_type": "call"}, {"api_name": "factory.lazy_attribute", "line_number": 84, "usage_type": "attribute"}, {"api_name": "factory.django", "line_number": 92, "usage_type": "attribute"}, {"api_name": "factory.Faker", "line_number": 98, "usage_type": "call"}, {"api_name": "factory.Faker", "line_number": 99, "usage_type": "call"}, {"api_name": "factory.Faker", "line_number": 100, "usage_type": "call"}, {"api_name": "factory.Faker", "line_number": 101, "usage_type": "call"}, {"api_name": "factory.SubFactory", "line_number": 102, "usage_type": "call"}, {"api_name": "factory.SubFactory", "line_number": 103, "usage_type": "call"}, {"api_name": "factory.SubFactory", "line_number": 104, "usage_type": "call"}, {"api_name": "factory.Faker", "line_number": 112, "usage_type": "call"}, {"api_name": "factory.lazy_attribute", "line_number": 106, "usage_type": "attribute"}, {"api_name": "factory.post_generation", "line_number": 117, "usage_type": "attribute"}, {"api_name": "factory.django", "line_number": 126, "usage_type": "attribute"}, {"api_name": "factory.SubFactory", "line_number": 132, "usage_type": "call"}, {"api_name": "factory.SubFactory", "line_number": 133, "usage_type": "call"}, {"api_name": "factory.Faker", "line_number": 134, "usage_type": "call"}, {"api_name": "factory.post_generation", "line_number": 136, "usage_type": "attribute"}, {"api_name": "factory.django", "line_number": 170, "usage_type": "attribute"}, {"api_name": "factory.Faker", "line_number": 176, "usage_type": "call"}, {"api_name": "factory.SubFactory", "line_number": 177, "usage_type": "call"}, {"api_name": "factory.SubFactory", "line_number": 178, "usage_type": "call"}, {"api_name": "factory.Faker", "line_number": 179, "usage_type": "call"}, {"api_name": "factory.django", "line_number": 182, "usage_type": "attribute"}, {"api_name": "factory.Faker", "line_number": 188, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 195, "usage_type": "call"}, {"api_name": "django.core.files.base.File", "line_number": 197, "usage_type": "call"}, {"api_name": "factory.lazy_attribute", "line_number": 190, "usage_type": "attribute"}, {"api_name": "factory.post_generation", "line_number": 199, "usage_type": "attribute"}, {"api_name": "factory.django", "line_number": 208, "usage_type": "attribute"}, {"api_name": "factory.Faker", "line_number": 214, "usage_type": "call"}, {"api_name": "factory.SubFactory", "line_number": 215, "usage_type": "call"}, {"api_name": "factory.SubFactory", "line_number": 216, "usage_type": "call"}, {"api_name": "factory.django", "line_number": 219, "usage_type": "attribute"}, {"api_name": "factory.Faker", "line_number": 225, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 232, "usage_type": "call"}, {"api_name": "django.core.files.base.File", "line_number": 234, "usage_type": "call"}, {"api_name": "factory.lazy_attribute", "line_number": 227, "usage_type": "attribute"}, {"api_name": "factory.post_generation", "line_number": 236, "usage_type": "attribute"}, {"api_name": "factory.django", "line_number": 245, "usage_type": "attribute"}, {"api_name": "factory.SubFactory", "line_number": 251, "usage_type": "call"}, {"api_name": "factory.SubFactory", "line_number": 252, "usage_type": "call"}, {"api_name": "factory.django", "line_number": 255, "usage_type": "attribute"}, {"api_name": "factory.Faker", "line_number": 261, "usage_type": "call"}, {"api_name": "factory.Faker", "line_number": 262, "usage_type": "call"}, {"api_name": "factory.SubFactory", "line_number": 265, "usage_type": "call"}, {"api_name": "factory.django", "line_number": 268, "usage_type": "attribute"}, {"api_name": "factory.SubFactory", "line_number": 274, "usage_type": "call"}, {"api_name": "factory.SubFactory", "line_number": 275, "usage_type": "call"}, {"api_name": "factory.lazy_attribute", "line_number": 277, "usage_type": "attribute"}, {"api_name": "factory.lazy_attribute", "line_number": 286, "usage_type": "attribute"}]}
+{"seq_id": "193773827", "text": "# based on ../zhymir_scripts/train_model.py\nimport os\nimport keras\nimport numpy as np\n\nfrom utils.file import dump_to_json\nimport matplotlib.pyplot as plt\n\ndef train_model(data, labels, model_p, save, filename, save_history, h_filename):\n model_history = model_p.fit(data, labels, batch_size=10)\n if save:\n model_p.save(filename)\n if save_history:\n dump_to_json(model_history.history, h_filename)\nif __name__ == '__main__':\n train_data = np.load('../zhymir_scripts/train_test/train_data.npy')\n train_labels = []\n with np.load('../zhymir_scripts/train_test/train_labels.npz') as data:\n train_labels = data['arr_0']\n model_root = '../../../Task2/models'\n history_root = '../../../Task2/data'\n filepath = os.path.join(model_root, 'cody_model.h5')\n history_filename = os.path.join(history_root, 'cody_model_history')\n batch_size = 10\n num_classifiers = 16\n model = keras.models.Sequential([\n keras.layers.Dense(units=100, input_shape=(num_classifiers, 10), activation='relu', name='D1'),\n keras.layers.Flatten(),\n keras.layers.Dense(10, name='output_layer', activation='softmax')\n ])\n metrics = ['accuracy']\n model.compile('adam', 'categorical_crossentropy', metrics=metrics)\n\n history = model.fit(train_data, train_labels, epochs=20, batch_size=batch_size, validation_split=0.1, verbose=0)\n print(history.history.keys())\n\n # summarize history for accuracy\n plt.plot(history.history['accuracy'])\n plt.plot(history.history['val_accuracy'])\n plt.title('model accuracy')\n plt.ylabel('accuracy')\n plt.xlabel('epoch')\n plt.legend(['train', 'test'], loc='upper left')\n plt.show()\n\n # summarize history for loss\n plt.plot(history.history['loss'])\n plt.plot(history.history['val_loss'])\n plt.title('model loss')\n plt.ylabel('loss')\n plt.xlabel('epoch')\n plt.legend(['train', 'test'], loc='upper left')\n plt.show()\n\n model.save(filepath)\n dump_to_json(history.history, history_filename)\n", "sub_path": "src/scripts/cody_scripts/train_ensemble_model.py", "file_name": "train_ensemble_model.py", "file_ext": "py", "file_size_in_byte": 2032, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "utils.file.dump_to_json", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "keras.models.Sequential", "line_number": 26, "usage_type": "call"}, {"api_name": "keras.models", "line_number": 26, "usage_type": "attribute"}, {"api_name": "keras.layers.Dense", "line_number": 27, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 27, "usage_type": "attribute"}, {"api_name": "keras.layers.Flatten", "line_number": 28, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 28, "usage_type": "attribute"}, {"api_name": "keras.layers.Dense", "line_number": 29, "usage_type": "call"}, {"api_name": "keras.layers", "line_number": 29, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "utils.file.dump_to_json", "line_number": 56, "usage_type": "call"}]}
+{"seq_id": "507008596", "text": "import csv\nimport json\n\n\ndef radarGen(infile, outfile):\n\n # open outfile in write mode\n f = open(outfile, 'w')\n\n # build rowsArr and populate with data from csv\n # so we can access with indices\n with open(infile,'r') as csvFile:\n rows = csv.reader(csvFile)\n c = 0\n rowsArr = []\n for row in rows:\n rowsArr.append(row)\n\n # create dictionary to store output data\n radarData = {}\n\n # loop through rowsArray\n for i in range(1, len(rowsArr)):\n countryArray = []\n for j in range(3,len(rowsArr[i])):\n subdict = {}\n subdict['axis'] = rowsArr[0][j]\n subdict['value'] = float(rowsArr[i][j])\n countryArray.append(subdict)\n dct[rowsArr[i][1]] = countryArray\n # write dictionary, converted to json string, to outfile\n f.write( json.dumps(dct) )\n\n\nif __name__ == '__main__':\n main()\n", "sub_path": "csv2json-radar-data.py", "file_name": "csv2json-radar-data.py", "file_ext": "py", "file_size_in_byte": 908, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "csv.reader", "line_number": 13, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 32, "usage_type": "call"}]}
+{"seq_id": "248015275", "text": "#!/usr/bin/env python3\n# -*-coding=utf-8-*-\n\nimport os\nimport re\nimport shutil\nfrom datetime import datetime as dt\nfrom threading import Thread\nfrom diffandpatch.kandilliobservatory import GetFile as GF\nfrom diffandpatch.kandilliobservatory import ObservatoryManager,GetDataCSV\n\nfrom dbHelper.dbhelper import PostgreSqlCrud\n\n\n\n\n\nclass DiffAndPatch():\n\n run = os.system\n ishere = os.path.exists\n time = str(dt.now().hour)\n time_min = str(dt.now().minute).zfill(2)\n time_sc = str(dt.now().second).zfill(2)\n real_time = dt.now().strftime(\"%Y%m%d\")\n file_name = real_time + '_' + time + '.csv'\n output_name = 'Out_'+real_time + '_' + time + time_min+ time_sc + '.csv'\n compare_name = 'Comp_'+real_time +'_'+ time + time_min+ time_sc +'.csv'\n database_name = 'Db_'+real_time +'_'+ time + '.csv'\n isload = False\n __error = True\n dir_name = '/home/helis/kandilli'\n dir_compare = 'just_compare'\n control = [\"Tarih,Saat,Enlem(N),Boylam(E),Derinlik(km),MD,ML,Mw,Yer,Çözüm-Niteliği\"]\n\n before_mylist = []\n new_mylist = []\n after_mylist = []\n\n\n @classmethod\n def error_message(cls, err):\n with open('/tmp/diffandpatch_err.txt', 'a') as f:\n f.write('\\n' + str(err) + '' + cls.real_time)\n print( str(err) + '' + str(cls.real_time))\n f.close()\n cls.__error = False\n return cls.__error\n\n @classmethod\n def callabel_dir(cls):\n # cls.inside_dir = os.getcwd()\n cls.inside_dir = cls.dir_name\n for i in cls.inside_dir.split('/'):\n if i == cls.dir_name:\n return True\n\n @classmethod\n def control(cls):\n cls.isload = os.chdir(cls.dir_name)\n if not os.path.exists(cls.dir_compare):\n os.mkdir(cls.dir_compare)\n cls.my_list = sorted(os.listdir(cls.isload))\n\n if len(cls.my_list) == 1:\n # print('Veri çek')\n ObservatoryManager(GetDataCSV()).create()\n cls.contol_before_compare()\n elif len(cls.my_list) ==2:\n cls.chckOne = [ i for i in cls.my_list if re.match(\"[Out]+\", i)]\n if cls.chckOne:\n FileManager(DbStg(cls.chckOne)).run()\n else:\n ObservatoryManager(GetDataCSV()).create()\n cls.contol_before_compare()\n elif len(cls.my_list) >= 3:\n cls.new_mylist = [ k for k in cls.my_list if re.match(\"[0-9]+\", k)]\n FileManager(OutPutStg(cls.new_mylist)).run()\n else:\n print('Nothing')\n\n @classmethod\n def contol_before_compare(cls):\n cls.before_comp_list= os.listdir(os.chdir(cls.dir_compare))\n if len(cls.before_comp_list) == 0:\n return None\n if len(cls.before_comp_list) == 1:\n print(cls.before_comp_list)\n return None\n elif len(cls.before_comp_list) == 2:\n FileManager(CompareTwoFiles(cls.before_comp_list)).run()\n\n\n\n @classmethod\n def recursive_operation(cls):\n # for unrepaired files like txt file..\n for num, i in enumerate(cls.my_list):\n if re.search(\"^.*csv$\", i):\n cls.diff_and_patch(i)\n\n cls.after_mylist = (list(dict.fromkeys(cls.before_mylist)))\n\n cls.data_text = ''\n cls.data_text = [i for i in cls.after_mylist[2:]]\n\n with open(cls.file_name, 'a') as write_csv:\n for data in cls.data_text:\n write_csv.write(\"%s\\n\" % data)\n write_csv.close()\n try:\n # if cls.callabel_dir():\n cls.new_mylist = sorted(os.listdir(cls.dir_name))\n cls. go_to_module = GF(cls.new_mylist[0][0:])\n\n\n except Exception as err:\n cls.error_message(err)\n finally:\n if cls.__error:\n os.remove(cls.new_mylist[0])\n\n\n\n @classmethod\n def diff_and_patch(cls, file):\n cls.file = file\n\n try:\n with open(cls.file, 'r') as f:\n print(cls.file)\n cls.read_data = f.read()\n f.close()\n for j in cls.read_data.splitlines():\n if not j.strip():continue\n cls.before_mylist.append(j)\n\n\n except (Exception) as err:\n cls.error_message(err)\n\n finally:\n if cls.__error:\n os.remove(cls.file)\n\n\n @classmethod\n def diff_and_patch_with_twoFiles(cls):\n try:\n for i in range(len(cls.my_list)):\n cls.file_for_patch = cls.my_list[i]\n with open(cls.file_for_patch,'r') as readOne:\n line = readOne.read()\n for k in line.splitlines():\n cls.before_mylist.append(k)\n readOne.close()\n cls.after_mylist = (list(dict.fromkeys(cls.before_mylist)))\n with open(cls.output_name,'a') as writeFile:\n for j in cls.after_mylist:\n writeFile.write(\"%s\\n\" % j)\n writeFile.close()\n except Exception as err:\n print(err)\n cls.__error = False\n finally:\n if cls.__error:\n for p in cls.my_list:\n print(p)\n os.remove(p)\n return cls.control()\n\n\nclass CompareTwoFiles(DiffAndPatch):\n def __init__(self,before_comp_list):\n self.file_comp = before_comp_list\n pass\n\n def run(self):\n with open(self.file_comp[0], 'r') as file1:\n with open(self.file_comp[1], 'r') as file2:\n self.same = set(file1).intersection(file2)\n\n self.same.discard('\\n')\n\n\n with open(self.database_name, 'w') as file_out:\n for line in self.same:\n file_out.write(line)\n file1.close()\n file2.close()\n file_out.close()\n\n self.my_list = os.listdir(os.chdir(self.dir_compare))\n for del_it in self.my_list:\n if re.match(\"[Db]+\",del_it):continue\n print('ready for db')\n os.remove(del_it)\n\n\n\n\n\n\n\n\nclass OutPutStg(DiffAndPatch):\n def __init__(self,new_file):\n self.new_file = new_file\n # self.output ='Out_'+ dt.now().strftime(\"%Y%m%d\")\n super(OutPutStg, self).__init__()\n\n def run(self):\n try:\n for for_read in self.new_file:\n print(for_read)\n with open(for_read,'r') as readOne:\n line = readOne.read()\n for k in line.splitlines():\n self.before_mylist.append(k)\n readOne.close()\n self.after_mylist = (list(dict.fromkeys(self.before_mylist)))\n with open(self.output_name,'w') as writeFile:\n for j in self.after_mylist:\n writeFile.write(\"%s\\n\" % j)\n writeFile.close()\n\n except Exception as err:\n print(err)\n finally:\n for p in self.new_file:\n os.remove(p)\n\n\nclass DbStg(DiffAndPatch):\n def __init__(self,file_name):\n super(DbStg, self).__init__()\n self.file_name = file_name\n\n def run(self):\n self.move_file(self.file_name[0])\n self.contol_before_compare()\n # self.post =PostgreSqlCrud(self.file_name[0]).insert_csf_file()\n # self.move_file(self.file_name[0])\n # FileManager(CompareTwoFiles()).run()\n\n\n def move_file(self, name):\n # os.rename(os.path.join(self.dir_name, name),\n # os.path.join(self.dir_name, self.dir_compare, self.compare_name))\n # print(os.path.join(self.dir_name, name),os.path.join(self.dir_name, self.dir_compare, self.compare_name))\n shutil.move(os.path.join(self.dir_name, name),\n os.path.join(self.dir_name, self.dir_compare, self.compare_name))\n\n\nclass FileManager():\n def __init__(self,filestg):\n self.filestg = filestg\n\n def run(self):\n self.filestg.run()\n\n\n\n\n\n\n\n\n\n\n\nif __name__ == '__main__':\n app = DiffAndPatch.control()\n", "sub_path": "kandilli_diffandpatch.py", "file_name": "kandilli_diffandpatch.py", "file_ext": "py", "file_size_in_byte": 8133, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.system", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 22, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 23, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 24, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 24, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 25, "usage_type": "name"}, {"api_name": "os.chdir", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path", "line_number": 61, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 62, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 63, "usage_type": "call"}, {"api_name": "diffandpatch.kandilliobservatory.ObservatoryManager", "line_number": 67, "usage_type": "call"}, {"api_name": "diffandpatch.kandilliobservatory.GetDataCSV", "line_number": 67, "usage_type": "call"}, {"api_name": "re.match", "line_number": 70, "usage_type": "call"}, {"api_name": "diffandpatch.kandilliobservatory.ObservatoryManager", "line_number": 74, "usage_type": "call"}, {"api_name": "diffandpatch.kandilliobservatory.GetDataCSV", "line_number": 74, "usage_type": "call"}, {"api_name": "re.match", "line_number": 77, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 84, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 84, "usage_type": "call"}, {"api_name": "re.search", "line_number": 99, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 113, "usage_type": "call"}, {"api_name": "diffandpatch.kandilliobservatory.GetFile", "line_number": 114, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 121, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 144, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 169, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 193, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 193, "usage_type": "call"}, {"api_name": "re.match", "line_number": 195, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 197, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 231, "usage_type": "call"}, {"api_name": "shutil.move", "line_number": 251, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 251, "usage_type": "call"}, {"api_name": "os.path", "line_number": 251, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 252, "usage_type": "call"}, {"api_name": "os.path", "line_number": 252, "usage_type": "attribute"}]}
+{"seq_id": "119024230", "text": "import sys\n\nfrom sqlalchemy import *\n\nfrom sqlalchemy.ext.declarative import declarative_base\n\nfrom sqlalchemy.orm import relationship\n\nfrom sqlalchemy import create_engine\n\nfrom flask.ext.login import UserMixin\n\n\nBase = declarative_base()\n\n\nmetadata = MetaData()\n\nclass Path(Base):\n\t__tablename__ = \"path\"\n\n\tname = Column(String(100), nullable=False)\n\tdescription = Column(Text, nullable=False)\n\tid = Column(Integer, primary_key = True)\n\t\n\nclass User(UserMixin, Base):\n\t__tablename__ = \"user\"\n\n\tid = Column(Integer, primary_key=True)\n\tusername = Column(String(100))\n\tfull_name = Column(String(100), nullable=False)\n\temail = Column(String(200), unique=True, nullable=False)\n\tpassword = Column(String(100), nullable=False)\n\txp = Column(Integer)\n\tpath_id = Column(Integer, ForeignKey('path.id'))\n\tpath = relationship(Path)\n\n\nclass Step(Base):\n\t__tablename__ = \"step\"\n\n\tid = Column(Integer, primary_key = True)\n\ttitle = Column(String(200), nullable=False)\n\tdescription = Column(Text, nullable=False)\n\tis_challenge = Column(Boolean)\n\tfilename = Column(String(500))\n\tpath_id = Column(Integer, ForeignKey('path.id'), nullable=False)\n\tpath = relationship(Path)\n\tposition = Column(Integer, nullable=False)\n\nclass Submission(Base):\n\t__tablename__ = \"submission\"\n\tid = Column(Integer, primary_key = True)\n\tfilename = Column(String(500))\n\tstep_id = Column(Integer, ForeignKey('step.id'), nullable=False)\n\tstep = relationship(Step)\n\tuser_id = Column(Integer, ForeignKey('user.id'), nullable=False)\n\tuser = relationship(User)\n\n\n\n\nengine = create_engine(\"sqlite:///l2code.db\")\n\nBase.metadata.create_all(engine)\n", "sub_path": "db_setup.py", "file_name": "db_setup.py", "file_ext": "py", "file_size_in_byte": 1597, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sqlalchemy.ext.declarative.declarative_base", "line_number": 14, "usage_type": "call"}, {"api_name": "flask.ext.login.UserMixin", "line_number": 27, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.relationship", "line_number": 37, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.relationship", "line_number": 49, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.relationship", "line_number": 57, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.relationship", "line_number": 59, "usage_type": "call"}, {"api_name": "sqlalchemy.create_engine", "line_number": 64, "usage_type": "call"}]}
+{"seq_id": "627436571", "text": "#!/usr/bin/env python\nimport json\nimport sys\nfrom src.write_MDcmds import write_MDscript\nfrom src.write_PBS import write_PBSheader\nimport argparse\n\nparser = argparse.ArgumentParser(usage=\"{} input_file.json\".format(sys.argv[0]),\n epilog=\"\"\"Generates the necessary input scripts\n for PBS submission in the HPC of a long\n MD run.\\n\"\"\")\nparser.add_argument(\"InputFile\", help=\"A JSON file with the different options\")\n\nargs = parser.parse_args()\n\ndef read_jsonfile(file):\n \"\"\"\n Parse the input JSON file and return a dictionary with the info\n \"\"\"\n with open(file) as data:\n json_data = json.load(data)\n return(json_data)\n\n\ndef get_NumberOfJobs(json):\n total_time = json['simulation_details']['final_time'] - json['simulation_details']['start_time']\n job_length = json['simulation_details']['job_length']\n if (total_time % job_length) != 0:\n sys.exit(\"Job lenght must be a divisor of total simulation time.\")\n else:\n return(int(total_time/job_length) + 1)\n\n\ndef get_Times(number_of_jobs, job_length, start_time):\n \"\"\"\n Returns a dictionary with the corresponding time window for each\n job\n \"\"\"\n timeList = {}\n for job in range(1, number_of_jobs):\n if job == 1:\n time_at_start = start_time\n time_at_finish = time_at_start + job_length\n else:\n time_at_finish = start_time + (job * job_length)\n time_at_start = time_at_finish - job_length\n seq = (str(time_at_start).zfill(4), str(time_at_finish).zfill(4))\n timeList[job] = '-'.join(seq)\n return(timeList)\n\n\ndef main():\n if args:\n input_file = read_jsonfile(args.InputFile)\n dictionary = get_Times(get_NumberOfJobs(input_file),\n input_file['simulation_details']['job_length'],\n input_file['simulation_details']['start_time'])\n print(dictionary)\n for i in range(1, get_NumberOfJobs(input_file)):\n write_PBSheader(i, input_file)\n write_MDscript(i, input_file, dictionary)\n\n\nif __name__ == \"__main__\":\n main()\n", "sub_path": "generate_scripts.py", "file_name": "generate_scripts.py", "file_ext": "py", "file_size_in_byte": 2221, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 8, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 8, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 21, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 29, "usage_type": "call"}, {"api_name": "src.write_PBS.write_PBSheader", "line_number": 60, "usage_type": "call"}, {"api_name": "src.write_MDcmds.write_MDscript", "line_number": 61, "usage_type": "call"}]}
+{"seq_id": "147155166", "text": "from array import array\nimport colorama\ncolorama.init(strip=False)\n\na = [0 for x in range(15)]\ncount_null = 0\ncount_otr = 0\nproiz = 1\nfor i in range(0, 15):\n a[i] = int(input('Введите {}/15 элемент массива: '.format(i)))\n#Обработаем первый элемент\nif (a[0] == 0):\n count_null += 1\nelif (a[0] < 0):\n count_otr += 1\n#конец обработки\nfor i in range(1,15):\n if (a[i] > 0):\n temp = a[i-1] * a[i]\n proiz = proiz + temp\n elif (a[i] == 0):\n count_null += 1\n elif (a[i] < 0):\n count_otr += 1\nprint('{endcolor}Количество отрицательных: {color}{out}{endcolor}'.format(color = colorama.Fore.GREEN, out = count_otr, endcolor = colorama.Fore.WHITE))\nprint('{endcolor}Произведение положительных элементов: {color}{out}{endcolor}'.format(color = colorama.Fore.GREEN, out = proiz, endcolor = colorama.Fore.WHITE))\nprint('{endcolor}Нулевых элементов: {color}{out}{endcolor}'.format(color = colorama.Fore.GREEN, out = count_null, endcolor = colorama.Fore.WHITE))\n\n", "sub_path": "5_1.py", "file_name": "5_1.py", "file_ext": "py", "file_size_in_byte": 1126, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "colorama.init", "line_number": 3, "usage_type": "call"}, {"api_name": "colorama.Fore", "line_number": 25, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 26, "usage_type": "attribute"}, {"api_name": "colorama.Fore", "line_number": 27, "usage_type": "attribute"}]}
+{"seq_id": "421732316", "text": "from django.contrib.auth.models import User\nfrom rest_framework.permissions import IsAuthenticated, IsAdminUser\nfrom performancereview.models import PerformanceReview, ReviewFeedback\nfrom rest_framework import viewsets, mixins\nfrom rest_framework.reverse import reverse\nfrom django.core.exceptions import PermissionDenied\nfrom django.shortcuts import get_object_or_404\nfrom performancereview.serializers import UserSerializer, PerformanceReviewSerializer, ReviewFeedbackSerializer, UserSerializerPut, PerformanceReviewSerializerGet, PerformanceReviewSerializerPut, ReviewFeedbackSerializerGet, ReviewFeedbackSerializerPut\n\nclass NoUpdateViewSet(mixins.RetrieveModelMixin,\n mixins.CreateModelMixin,\n mixins.ListModelMixin,\n mixins.DestroyModelMixin,\n viewsets.GenericViewSet):\n pass\n\nclass UserViewSet(viewsets.ModelViewSet):\n \"\"\"\n API endpoint that allows users to be viewed or edited.\n \"\"\"\n queryset = User.objects.all().order_by('-date_joined')\n serializer_class = UserSerializer\n permission_classes = (IsAuthenticated, IsAdminUser)\n\n def get_serializer_class(self):\n serializer_class = self.serializer_class\n\n if self.request.method == 'PUT':\n serializer_class = UserSerializerPut\n\n return serializer_class\n\n\nclass PerformanceReviewViewSet(NoUpdateViewSet):\n \"\"\"\n API endpoint that allows performance reviews to be viewed or edited.\n \"\"\"\n queryset = PerformanceReview.objects.all().order_by('reviewee__pk')\n serializer_class = PerformanceReviewSerializer\n permission_classes = (IsAuthenticated, IsAdminUser)\n\n def get_serializer_class(self):\n serializer_class = self.serializer_class\n\n if self.request.method == 'PUT':\n serializer_class = PerformanceReviewSerializerPut\n if self.request.method == 'GET':\n serializer_class = PerformanceReviewSerializerGet\n\n return serializer_class\n\n # custom method so we can create via username instead of resource url\n def create(self, request):\n reviewee = get_object_or_404(User, username=request.data['reviewee'])\n request.data['reviewee'] = UserSerializer(reviewee, context={ 'request': request }).data.get('url')\n return super().create(request)\n\n# only admin to delete\n# only allow reviewer to update\nclass ReviewFeedbackViewSet(viewsets.ModelViewSet):\n \"\"\"\n API endpoint that allows individual review feedback to be viewed or edited.\n \"\"\"\n queryset = ReviewFeedback.objects.all().order_by('reviewer__pk')\n serializer_class = ReviewFeedbackSerializer\n permission_classes = (IsAuthenticated,)\n\n def get_queryset(self):\n # if user is admin, return all\n if self.request.user.is_staff:\n return self.queryset\n # if user is not admin, fetch all where user = reviewer\n return ReviewFeedback.objects.filter(reviewer=self.request.user)\n\n\n def get_object(self):\n obj = get_object_or_404(self.get_queryset(), pk=self.kwargs[\"pk\"])\n self.check_object_permissions(self.request, obj)\n return obj\n\n def perform_destroy(self, instance):\n print(self.request.user)\n # allow admin to delete object\n if self.request.user.is_staff:\n super().perform_destroy(instance)\n else:\n # otherwise raise exception\n raise PermissionDenied\n\n def perform_create(self, serializer):\n # allow admin to create object\n print(self.request.user.is_staff)\n if self.request.user.is_staff:\n super().perform_create(serializer)\n else:\n # otherwise raise exception\n raise PermissionDenied\n\n def perform_update(self, serializer):\n # allow admin or reviewer to update object\n if self.request.user.is_staff or self.get_object():\n super().perform_update(serializer)\n else:\n # otherwise raise exception\n raise PermissionDenied\n\n\n def get_serializer_class(self):\n serializer_class = self.serializer_class\n\n if self.request.method == 'PUT':\n serializer_class = ReviewFeedbackSerializerPut\n if self.request.method == 'GET':\n serializer_class = ReviewFeedbackSerializerGet\n\n return serializer_class\n\n # custom method so we can create via username instead of resource url\n def create(self, request):\n reviewer = get_object_or_404(User, username=request.data['reviewer'])\n performanceReview = get_object_or_404(PerformanceReview, pk=request.data['performanceReviewId'])\n\n request.data['reviewer'] = UserSerializer(reviewer, context={ 'request': request }).data.get('url')\n request.data['performanceReview'] = PerformanceReviewSerializer(performanceReview, context={ 'request': request }).data.get('url')\n\n return super().create(request)\n", "sub_path": "backend/api/performancereview/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4914, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "rest_framework.mixins.RetrieveModelMixin", "line_number": 10, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 10, "usage_type": "name"}, {"api_name": "rest_framework.mixins.CreateModelMixin", "line_number": 11, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 11, "usage_type": "name"}, {"api_name": "rest_framework.mixins.ListModelMixin", "line_number": 12, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 12, "usage_type": "name"}, {"api_name": "rest_framework.mixins.DestroyModelMixin", "line_number": 13, "usage_type": "attribute"}, {"api_name": "rest_framework.mixins", "line_number": 13, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.GenericViewSet", "line_number": 14, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 14, "usage_type": "name"}, {"api_name": "rest_framework.viewsets.ModelViewSet", "line_number": 17, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 17, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.all", "line_number": 21, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 21, "usage_type": "name"}, {"api_name": "performancereview.serializers.UserSerializer", "line_number": 22, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 23, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAdminUser", "line_number": 23, "usage_type": "name"}, {"api_name": "performancereview.serializers.UserSerializerPut", "line_number": 29, "usage_type": "name"}, {"api_name": "performancereview.models.PerformanceReview.objects.all", "line_number": 38, "usage_type": "call"}, {"api_name": "performancereview.models.PerformanceReview.objects", "line_number": 38, "usage_type": "attribute"}, {"api_name": "performancereview.models.PerformanceReview", "line_number": 38, "usage_type": "name"}, {"api_name": "performancereview.serializers.PerformanceReviewSerializer", "line_number": 39, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 40, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAdminUser", "line_number": 40, "usage_type": "name"}, {"api_name": "performancereview.serializers.PerformanceReviewSerializerPut", "line_number": 46, "usage_type": "name"}, {"api_name": "performancereview.serializers.PerformanceReviewSerializerGet", "line_number": 48, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 54, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 54, "usage_type": "argument"}, {"api_name": "performancereview.serializers.UserSerializer", "line_number": 55, "usage_type": "call"}, {"api_name": "rest_framework.viewsets.ModelViewSet", "line_number": 60, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 60, "usage_type": "name"}, {"api_name": "performancereview.models.ReviewFeedback.objects.all", "line_number": 64, "usage_type": "call"}, {"api_name": "performancereview.models.ReviewFeedback.objects", "line_number": 64, "usage_type": "attribute"}, {"api_name": "performancereview.models.ReviewFeedback", "line_number": 64, "usage_type": "name"}, {"api_name": "performancereview.serializers.ReviewFeedbackSerializer", "line_number": 65, "usage_type": "name"}, {"api_name": "rest_framework.permissions.IsAuthenticated", "line_number": 66, "usage_type": "name"}, {"api_name": "performancereview.models.ReviewFeedback.objects.filter", "line_number": 73, "usage_type": "call"}, {"api_name": "performancereview.models.ReviewFeedback.objects", "line_number": 73, "usage_type": "attribute"}, {"api_name": "performancereview.models.ReviewFeedback", "line_number": 73, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 77, "usage_type": "call"}, {"api_name": "django.core.exceptions.PermissionDenied", "line_number": 88, "usage_type": "name"}, {"api_name": "django.core.exceptions.PermissionDenied", "line_number": 97, "usage_type": "name"}, {"api_name": "django.core.exceptions.PermissionDenied", "line_number": 105, "usage_type": "name"}, {"api_name": "performancereview.serializers.ReviewFeedbackSerializerPut", "line_number": 112, "usage_type": "name"}, {"api_name": "performancereview.serializers.ReviewFeedbackSerializerGet", "line_number": 114, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 120, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 120, "usage_type": "argument"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 121, "usage_type": "call"}, {"api_name": "performancereview.models.PerformanceReview", "line_number": 121, "usage_type": "argument"}, {"api_name": "performancereview.serializers.UserSerializer", "line_number": 123, "usage_type": "call"}, {"api_name": "performancereview.serializers.PerformanceReviewSerializer", "line_number": 124, "usage_type": "call"}]}
+{"seq_id": "135781602", "text": "import logging\nlogging.getLogger('googleapiclient.discovery_cache').setLevel(logging.ERROR)\n\nimport gevent.monkey\ngevent.monkey.patch_all()\n\nimport argparse\nimport gevent.pywsgi\nimport sys\n\nimport backend\n\ndef ycdl_flask_launch(port, refresh_rate):\n if port == 443:\n http = gevent.pywsgi.WSGIServer(\n listener=('', port),\n application=backend.site,\n keyfile='https\\\\flasksite.key',\n certfile='https\\\\flasksite.crt',\n )\n else:\n http = gevent.pywsgi.WSGIServer(\n listener=('0.0.0.0', port),\n application=backend.site,\n )\n\n if refresh_rate is not None:\n backend.common.start_refresher_thread(refresh_rate)\n\n print(f'Starting server on port {port}')\n http.serve_forever()\n\ndef ycdl_flask_launch_argparse(args):\n if args.do_refresh:\n refresh_rate = args.refresh_rate\n else:\n refresh_rate = None\n\n return ycdl_flask_launch(\n port=args.port,\n refresh_rate=refresh_rate,\n )\n\ndef main(argv):\n parser = argparse.ArgumentParser(description=__doc__)\n\n parser.add_argument('port', nargs='?', type=int, default=5000)\n parser.add_argument('--no_refresh', dest='do_refresh', action='store_false', default=True)\n parser.add_argument('--refresh_rate', dest='refresh_rate', type=int, default=60 * 60 * 6)\n parser.set_defaults(func=ycdl_flask_launch_argparse)\n\n args = parser.parse_args(argv)\n return args.func(args)\n\nif __name__ == '__main__':\n raise SystemExit(main(sys.argv[1:]))\n", "sub_path": "frontends/ycdl_flask/ycdl_flask_launch.py", "file_name": "ycdl_flask_launch.py", "file_ext": "py", "file_size_in_byte": 1543, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.getLogger", "line_number": 2, "usage_type": "call"}, {"api_name": "logging.ERROR", "line_number": 2, "usage_type": "attribute"}, {"api_name": "gevent.monkey.monkey.patch_all", "line_number": 5, "usage_type": "call"}, {"api_name": "gevent.monkey.monkey", "line_number": 5, "usage_type": "attribute"}, {"api_name": "gevent.monkey", "line_number": 5, "usage_type": "name"}, {"api_name": "gevent.monkey.pywsgi.WSGIServer", "line_number": 15, "usage_type": "call"}, {"api_name": "gevent.monkey.pywsgi", "line_number": 15, "usage_type": "attribute"}, {"api_name": "gevent.monkey", "line_number": 15, "usage_type": "name"}, {"api_name": "backend.site", "line_number": 17, "usage_type": "attribute"}, {"api_name": "gevent.monkey.pywsgi.WSGIServer", "line_number": 22, "usage_type": "call"}, {"api_name": "gevent.monkey.pywsgi", "line_number": 22, "usage_type": "attribute"}, {"api_name": "gevent.monkey", "line_number": 22, "usage_type": "name"}, {"api_name": "backend.site", "line_number": 24, "usage_type": "attribute"}, {"api_name": "backend.common.start_refresher_thread", "line_number": 28, "usage_type": "call"}, {"api_name": "backend.common", "line_number": 28, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 45, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 56, "usage_type": "attribute"}]}
+{"seq_id": "325960415", "text": "#GlobalData.py\n#p1.py\nimport time\nimport pickle as pk\nimport glob\nfrom threading import Thread\nimport os.path as PATH\nimport os\nimport dill as d\n\n'''\nfrom Globals import *\ngd = GlobalData()\ndef pp(data):\n\tprint(\"data\",data)\n\treturn setBackgroundColor(data)\ngd.Run(\"rotr\",(100,200,233),pp)\n'''\n\n''' > > > > >\n# gd.SaveFuncX(\"foo\",(lambda:print(\"yo!!!\")))\n# x = lambda: print(\"XXXXXXX\")\n# gd.SaveFuncX(\"foo\",x)\n# gd.SaveFuncX(\"foo\",lambda:os._exit(1))\n# gd.SaveFuncX(\"foo\",x)\n# >>> gd.SaveFuncX(\"foo\",(lambda:print(tt.ctime())))\n > > > > > Search funcExample1 '''\n\nprintX = False\n\ndef PubExit(gd):\n\tif gd is None:\n\t\tgd = GlobalData()\n\n\tgd.pub(\"exitAll\",[True, time.time()])\n\n\ndef SubExit(gd, id = None):\n\tif gd is None:\n\t\tgd = GlobalData()\n\tgd.sub(\"exitAll\",Exit,block = False)\n\tif id is not None:\n\t\tgd.sub(\"exit\"+id,Exit,block = False)\n\n# def SubExit(gd, id = None):\n# \tif gd is None:\n# \t\tgd = GlobalData()\n# \twaitSubExitAll(gd)\n# \tif id is not None:\n# \t\twaitSubExitID(gd,id)\n#\n# def waitSubExitID(gd, id):\n# \tt = Thread(target = waitSubExitIDT, args = [[gd, id],])\n# \tt.start()\n#\n# def waitSubExitAll(gd):\n# \tt = Thread(target = waitSubExitAllT, args = [[gd, ],])\n# \tt.start()\n#\n# def waitSubExitIDT(data):\n# \tgd, id = data\n# \tgd.sub(\"exit\"+id,Exit,block = True)\n#\n# def waitSubExitAllT(data):\n# \tgd = data[0]\n# \tgd.sub(\"exitAll\",Exit,block = True)\n# \t# gd.sub(\"exitAll\",Exit,block = True)\n\n\n###### from GlobalData import *\n###### r = lambda c : c is 100 and True or (lambda a: print(a) is None and r(a))(c+1)\n\ndef exxit():\n\tprint(\"SAYONARA!\")\n\tos._exit(1)\n\ndef Exit(d=None):\n\tprint(\"!!!!!!!!!!!!!!!!!\")\n\tprint(\"!!!!!!!!!!!!!!!!!\")\n\tprint(\"!!!!!!!!!!!!!!!!!\")\n\tprint(\"!!!!!!!!!!!!!!!!!\")\n\tif d is None:\n\t\texxit()\n\telse:\n\t\tprint(d)\n\tdata, channel = d\n\tif data[0]:\n\t\t# print(\"AAAAA\")\n\t\t# printT(channel)\n\t\t# print(\"BBBB\")\n\t\texxit()\n\telse:\n\t\tprint(\"Not Exiting...\")\n\ndef waitExit(x = 3):\n\tt = Thread(target = waitExitT, args = [[x],])\n\tt.start()\ndef waitExitT(data):\n\tx = data[0]\n\tfor i in range(x):\n\t\tprint((\"EXITING IN \",x-i))\n\t\ttime.sleep(1)\n\n\tprint(\"BYE!!!!!!!!\")\n\tos._exit(1)\n\tprint(\"BYE!!!!!!!!2\")\n\n\n\n\ndef funcExample1():\n\tc = \"foo\"\n\tprint(\"test channel = \",c)\n\t# fff = [(lambda: print(\"FFFFFFFF\"))]\n\tgd = GlobalData()\n\tfff = gd.bindFunc(c)\n\tq = fff[0]\n\twhile(True):\n\t\ttime.sleep(0.01)\n\t\tprint(fff, fff[0](), q())\n\ndef funcExample2():\n\tc = \"foo\"\n\tprint(\"test channel = \",c)\n\t# fff = [(lambda: print(\"FFFFFFFF\"))]\n\tgd = GlobalData()\n\tfff = gd.bindFunc(c)\n\tq = fff[0]\n\twhile(True):\n\t\ttime.sleep(0.01)\n\t\tprint(fff, fff[0](), q())\n\n\ndef funcExample0():\n\tfff = [(lambda: print(\"FFFFFFFF\"))]\n\tgd = GlobalData()\n\tgd.bind(\"fff\",fff)\n\twhile(True):\n\t\td.loads(fff[0])()\n\n\ndef non():\n\tprint(\"-----non-----\")\n\treturn None\n\nclass Wrapper:\n\tdef __init__(self, data):\n\t\tself.data = [data]\n\n\tdef __call__(self):\n\t\treturn self.data[0]()\n\nclass GlobalData:\n\t#def __init__(self,dir = \"/media/magic/EV1/GlobalData/\", max = 100000000, wait=0.01):\n\tdef __init__(self,dir = \"X/GlobalData/\", max = 10000000, wait=0.01, lowerCase = False, printPub = False, debug = False):\n\t\t# print(\"AAAAAAAAAAAAAAAAAAAAa\")\n\t\tself.Threads = []\n\t\tself.Debug = [debug]\n\t\tself.dir = dir\n\t\tself.extChannel = \".gdc\"\n\t\tself.extData = \".gdd\"\n\t\tself.__Channels = list()\n\t\tself.max = max\n\t\tself.wait = wait\n\t\tself.lowerCase = lowerCase\n\t\tself.printPub = printPub\n\t\tdbexist = PATH.exists(dir)\n\t\tif not dbexist:\n\t\t\tos.makedirs(dir)\n\t\tuT = Thread(target = self.__updateChannels, args = [self])\n\t\tuT.start()\n\t\tself.Threads.append(uT)\n\n\tdef GetChannels(self):\n\t\tchn = list()\n\t\tfor k in self.__Channels:\n\t\t\ta = k.split(\".\")[0]\n\t\t\tif a not in chn:\n\t\t\t\tchn.append(a)\n\t\tchn.sort()\n\t\treturn chn\n\n\tdef __updateChannels(self, data):\n\t\t#channels = sorted(glob.glob(dirPath+'*'),key=os.path.getmtime)\n\t\tttt = str(time.time())\n\t\t#print(\"updating channels\",\"OOOOOOOOOOOOOOOOOOO\")\n\t\twhile(True):\n\t\t\t# print(\"x Checking Channels...\", ttt)\n\t\t\tchannels = glob.glob(self.dir+'*')\n\t\t\tif len(channels) == 0:\n\t\t\t\tself.printF(\"0 Channels\")\n\t\t\t\tself.__Channels = list()\n\t\t\tfor c in channels:\n\t\t\t\tcn = c.split(self.dir)\n\t\t\t\tif len(cn)>0 and cn[-1] not in self.__Channels:\n\t\t\t\t\t\tself.__Channels.append(cn[-1])\n\t\t\t\t\t\tself.printF(\"added channel\",cn[-1])\n\t\t\ttime.sleep(0.1)\n\n\tdef channelUpdate(self, channel, untilTrue = True):\n\t\tif self.lowerCase:\n\t\t\tchannel = channel.lower()\n\t\tself.addChannel(channel)\n\n\t\tdata = self.load(self.dir+channel+self.extChannel, untilTrue = untilTrue)\n\t\tif data is None:\n\t\t\tdata = -1\n\t\tif (data > self.max):\n\t\t\tdata = 0\n\t\tdata += 1\n\t\tself.savePath(data,self.dir+channel+self.extChannel,untilTrue = untilTrue)\n\n\tdef awaitChannelUpdate(self, channel):\n\t\tif self.lowerCase:\n\t\t\tchannel = channel.lower()\n\t\twait=self.wait\n\t\td = self.load(self.dir+channel+self.extChannel)\n\t\t#print(\"ddddddddddddddddd\",d)\n\t\tif d is None:\n\t\t\td = self.load(self.dir+channel+self.extChannel)\n\t\t\tif d is None:\n\t\t\t\treturn None\n\t\ttempD = d+0\n\t\t#print(d,tempD,str(d) == str(tempD))\n\n\t\twhile(str(d) == str(tempD)):\n\t\t\t# print(\".\",d)\n\t\t\td = self.load(self.dir+channel+self.extChannel)\n\t\t\ttime.sleep(wait)\n\t\tself.printF(\"!!!!!!!!!!!!!!!!!\")\n\t\treturn self.load(self.dir+channel+self.extData)\n\n\tdef pub(self, channel, data, untilTrue = True):\n\t\tif self.lowerCase:\n\t\t\tchannel = channel.lower()\n\t\tif self.printPub:\n\t\t\tprint(\"PUB: \", channel,\"\\n:::\", data)\n\t\treturn self.save(data, channel, untilTrue = untilTrue)\n\n\t# def pubZZ(self, channel, data):\n\t# \t#if channel+self.extChannel not in self.__Channels:\n\t# \t#\tself.addChannel(channel)\n\t# \t#else:\n\t# \t#\tself.channelUpdate(channel+self.extChannel)\n\t# \t#\tself.save(data,self.dir+channel+self.extData)\n\t#\n\t# \t#self.printF(\"########\")\n\t# \tself.addChannel(channel)\n\t# \t#self.printF(\"########\")\n\t# \tself.channelUpdate(channel+self.extChannel)\n\t# \tself.savePath(data,self.dir+channel+self.extData)\n\n\tdef sub(self, channel, func=None, autoPub = None, block = True, once= False, echo=True, debug = False):\n\t\tif func is None:\n\t\t\tfunc = self.defFunc\n\t\tif self.lowerCase:\n\t\t\tchannel = channel.lower()\n\t\t#if channel+self.extChannel not in self.__Channels:\n\t\t#self.printF(\"########\")\n\t\tself.addChannel(channel)\n\t\t#self.printF(\"########\")\n\n\t\tif echo:\n\t\t\tprint(\"Subscribing to\", channel)\n\t\tif block:\n\t\t\tself.waitAndExe(channel, func, once, autoPub = autoPub, debug = debug)\n\t\telse:\n\t\t\tuT = Thread(target = self.waitAndExe, args = [channel, func, once, autoPub, debug])\n\t\t\tself.Threads.append(uT)\n\t\t\tuT.start()\n\n\tdef defFunc(self, data):\n\n\t\tdata, channel = data\n\t\tprint()\n\t\tprint(\":::::::: \"+channel+ \"- incoming data ::::::::::::::::::::::::::::: \"+time.ctime())\n\t\tprint(data)\n\t\tprint(\":::::::: \"+channel+ \"- incoming data :::::::::::::::::::::::::::::\")\n\t\tprint()\n\t\treturn data\n\n\t# def subUDP(self, channel, func, autoPub = None, block = True, once= False, echo=True):\n\t# \tif self.lowerCase:\n\t# \t\tchannel = channel.lower()\n\t# \t#if channel+self.extChannel not in self.__Channels:\n\t# \t#self.printF(\"########\")\n\t# \tself.addChannel(channel)\n\t# \t#self.printF(\"########\")\n\t# \tautoPub = [channel]\n\t#\n\t# \tif echo:\n\t# \t\tprint(\"Subscribing to UDP\", channel, \"IP:\", ip ,\"PORT\",port)\n\t# \tif block:\n\t# \t\tself.waitAndExe(channel, func, once, autoPub = autoPub)\n\t# \telse:\n\t# \t\tuT = Thread(target = self.waitAndExe, args = [channel, func, once, autoPub])\n\t# \t\tuT.start()\n\tdef bind(self, channel, default = None, var = None):\n\t\tif var is None or len(var)<1:\n\t\t\tvar = [None]\n\n\t\tcurr = self.GetX(channel)\n\n\t\tif curr is None:\n\t\t\tself.channelUpdate(channel)\n\t\telse:\n\t\t\t# self.SaveX(channel,curr)\n\t\t\tvar = [curr]\n\t\tif default is not None and var[0] is None:\n\t\t\tvar[0] = default\n\t\t\tself.SaveX(channel,default)\n\t\t# var[0] = curr\n\n\t\tt = Thread(target = self.__bindT, args = [[channel,var],])\n\t\tself.Threads.append(t)\n\t\tt.start()\n\t\treturn var\n\n\n\tdef __bindT(self, data):\n\t\tchannel, var = data\n\t\twhile(True):\n\t\t\ttime.sleep(0.01)\n\t\t\tvar[0] = self.awaitChannelUpdate(channel)\n\t\t\tprint(\"@@@@@@@@@@@@@@@@@\", \"updated bind for \",channel, \":::\",var)\n\n\tdef ReRun(self, channel, params = None):\n\t\treturn self.SaveFuncX(channel, self.GetFuncX(channel), params = params)\n\n\tdef Run(self, channel, params = None, newFunc = None, untilRes = False):\n\t\tfunc = self.GetFuncX(channel)\n\t\tif func is None and newFunc is None:\n\t\t\tprint(\"No Function! for\",channel,\"- First SaveFuncX !!!!!!!\")\n\t\t\treturn None\n\t\telse:\n\t\t\tself.SaveX(channel+\"_done\", None)\n\t\t\tif newFunc is None:\n\t\t\t\tself.ReRun(channel, params = params)\n\t\t\telse:\n\t\t\t\tself.SaveFuncX(channel,newFunc,params = params)\n\t\t\tres = self.GetX(channel+\"_done\")\n\t\t\tstop = False\n\t\t\twhile res is None and not stop:\n\t\t\t\tif not untilRes:\n\t\t\t\t\tstop = True\n\t\t\t\tres = self.GetX(channel+\"_done\")\n\n\t\t\t# print(\"!!!!!!!!!!!!!!!\",\"res\")\n\t\t\t# print(res)\n\t\t\t# print(\"!!!!!!!!!!!!!!!\")\n\t\t\treturn res\n\n\n\tdef bindFunc(self, channel, params = None, autoRun = False, preRun = None, ofunc = None):\n\t\tvar = [None]\n\t\tif ofunc is None:\n\t\t\t# if var is None or len(var)<1:\n\t\t\t\t# var = [(lambda:print(\"___\"))]\n\t\t\tcurr = self.GetX(channel)\n\t\t\tif curr is None:\n\t\t\t\tself.channelUpdate(channel)\n\t\t\t# var[0] = d.loads(curr)\n\t\t\ttry:\n\t\t\t\tvar[0] = d.loads(curr)\n\t\t\texcept Exception as e:\n\t\t\t\tprint(\"Exception:\",e)\n\t\t\t\tvar[0] = None\n\n\t\t\tif var[0] is None:\n\t\t\t\tvar[0] = non\n\t\telse:\n\t\t\tvar = [ofunc]\n\n\t\tif preRun is None:\n\t\t\t# print(\"TTTTTTTTTTTTT\")\n\t\t\t# preRun = self.defPreRun\n\t\t\tpass\n\n\t\tt = Thread(target = self.__bindFuncT, args = [[channel,var, autoRun, preRun],])\n\t\tself.Threads.append(t)\n\t\tt.start()\n\t\tprint(\"binding func to\",channel)\n\t\treturn var\n\n\tdef defPreRun(self):\n\t\tprint(\"preRun\")\n\t\treturn None\n\n\tdef timeAsTT(self):\n\t\t# print(\"\")\n\t\treturn None\n\t\timport time as tt\n\n\n\tdef __bindFuncT(self, data):\n\t\tchannel, var, autoRun, preRun = data\n\t\tstop = False\n\t\twhile(not stop):\n\t\t\ttime.sleep(0.01)\n\t\t\tif not self.Debug[0]:\n\t\t\t\ttry:\n\t\t\t\t\tdata = self.__bindLogic(data)\n\t\t\t\t\tchannel, var, autoRun, preRun, res, params = data\n\n\t\t\t\texcept Exception as e:\n\t\t\t\t\tprint()\n\t\t\t\t\tprint()\n\t\t\t\t\tprint(\"EEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEe\")\n\t\t\t\t\tprint(\"Exception: \",e)\n\t\t\t\t\tprint()\n\t\t\t\t\tprint(\"EEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEe\")\n\t\t\t\t\tprint()\n\n\t\t\t\t\tself.SaveX(channel+\"_done\", False)\n\t\t\t\t\t# print(e.message)\n\t\t\t\t\t# print(e.args)\n\t\t\t\t\tif \"keyboard\" in str(e).lower():\n\t\t\t\t\t\tprint(\"0000000000000000000000000000\")\n\t\t\t\t\t\tstop = True\n\t\t\telse:\n\t\t\t\tdata = self.__bindLogic(data)\n\t\t\t\tchannel, var, autoRun, preRun, res, params = data\n\n\tdef __bindLogic(self, data):\n\t\t# print()\n\t\t# print(\"..............\")\n\t\tres, params = None, None\n\t\tif len(data) == 6:\n\t\t\t# print(len(data))\n\t\t\tres, params = data[4], data[5]\n\n\n\t\t# print(1111111)\n\t\tchannel, var, autoRun, preRun = data[:4]\n\t\tprint(channel, var, autoRun, preRun)\n\t\tvar[0] = d.loads(self.awaitChannelUpdate(channel))\n\t\t# print(\"fffffffff\")\n\t\tif var[0] is None:\n\t\t\tvar[0] = non\n\t\tprint(\"\\n@@@@@@@@@@@@@@@@@\", \"updated bind for \",channel)\n\t\t# print(1111111)\n\t\tif autoRun:\n\t\t\t# print(222222)\n\t\t\tparams = [None]\n\t\t\ttparams = self.GetX(channel+\"_params\")\n\t\t\tif tparams is not None:\n\t\t\t\tparams[0] = tparams\n\n\t\t\t# print(\"func:\",var, \"params\",params)\n\t\t\tif preRun is not None:\n\t\t\t\tpreRun()\n\n\t\t\t# print(\"PPPPPPPPPPPPPPPPPPPPPP\")\n\t\t\t# print(\"Params :::\",params)\n\t\t\tif params is not None:\n\t\t\t\tif \"list\" in str(type(params)) and len(params) > 0:\n\t\t\t\t\tres = var[0](params[0])\n\t\t\t\telse:\n\t\t\t\t\tres = var[0](params)\n\t\t\telse:\n\t\t\t\tres = var[0](None)\n\t\t\t# print(222222)\n\t\t\t# print(\"source res\",res)\n\t\t\tself.SaveX(channel+\"_done\", [True,res])\n\t\t\t# print(\"xx\",\"...f\")\n\t\t# print(\"..............f\")\n\t\treturn channel, var, autoRun, preRun, res, params\n\n\tdef SubFunc(self, channel):\n\t\treturn self.bindFunc(channel, autoRun=True)\n\n\tdef Func(self, channel, func):\n\t\tself.SaveFuncX(channel,func)\n\t\treturn self.bindFunc(channel, autoRun=True, ofunc = func)\n\n\tdef waitAndExe(self, channel, func, once= False, autoPub = None, debug = False):\n\t\tif self.lowerCase:\n\t\t\tchannel = channel.lower()\n\t\trun = True\n\t\tnewdata = None\n\t\twhile run:\n\t\t\trun = not once\n\t\t\tdata = self.awaitChannelUpdate(channel)\n\t\t\tself.printF(\"@@@@@@@@@@@@\")\n\t\t\tif not debug:\n\t\t\t\ttry:\n\t\t\t\t\tnewdata = func([data,channel])\n\t\t\t\texcept:\n\t\t\t\t\tpass\n\t\t\telse:\n\t\t\t\tnewdata = func([data,channel])\n\t\t\tself.printF(\"############\")\n\t\t\tif autoPub is not None:\n\t\t\t\tif type(autoPub) is list:\n\t\t\t\t\tfor pb in autoPub:\n\t\t\t\t\t\tprint(\"autoPub is \",pb)\n\t\t\t\t\t\tself.pub(pb,newdata)\n\t\t\t\telse:\n\t\t\t\t\tprint(\"autoPub is \",autoPub)\n\t\t\t\t\tself.pub(autoPub,newdata)\n\t\t\telse:\n\t\t\t\tpass\n\t\t\t\t#print(\"autoPub is None\")\n\n\n\n\n\tdef addChannel(self, channel):\n\t\tif self.lowerCase:\n\t\t\tchannel = channel.lower()\n\t\t#for l in self.__Channels:\n\t\t\t#self.printF(\"CHANNEL:\",l)\n\t\t#\tpass\n\t\tif channel not in self.__Channels and channel+self.extChannel not in self.__Channels:\n\t\t\t#self.__Channels.append(channel+self.extData)\n\t\t\t#self.__Channels.append(channel+self.extChannel)\n\t\t\tself.savePath(0, self.dir+channel+self.extChannel)\n\t\t\tself.savePath(-1, self.dir+channel+self.extData)\n\t\telse:\n\t\t\tself.printF(\"Already Have Channel\",channel)\n\n\tdef save(self, data, channel, untilTrue = True):\n\t\tif self.lowerCase:\n\t\t\tchannel = channel.lower()\n\t\t### ???\n\t\tself.channelUpdate(channel)\n\t\treturn self.savePath(data,self.dir+channel+self.extData,untilTrue=untilTrue)\n\n\tdef savePath(self, data, filename, untilTrue = True):\n\t# with open(filename,\"wb\") as file:\n\t# pk.dump(data,file)\n\t# self.printF(\"saved\",data)\n\t\tself.printF(\"data to be saved:\",data)\n\t\tself.printF(\"path:\",filename)\n\t\trun = True\n\t\twhile(run):\n\t\t\trun = untilTrue\n\t\t\ttry:\n\t\t\t\twith open(filename,\"wb\") as file:\n\t\t\t\t\tpk.dump(data,file)\n\t\t\t\t\tself.printF(\"saved\",str(data), \"@\",filename)\n\t\t\t\t\trun = False\n\t\t\t\t# del(data)\n\t\t\t\t### ???\n\t\t\t\t# print(\"//////////////////////////\")\n\t\t\t\treturn True\n\t\t\texcept:\n\t\t\t\tself.printF(\"XXXXXXX failed saving\")\n\t\t\t\ttime.sleep(0.01)\n\t\treturn False\n\n\tdef get(self, filename, untilTrue = True):\n\t\treturn self.getData(filename,untilTrue)\n\n\tdef getData(self, filename, untilTrue = True):\n\t\tfilename = self.dir+str(filename)+self.extData\n\t\treturn self.load(filename,untilTrue)\n\n\tdef getChannelData(self, channel):\n\t\treturn self.get(str(channel),untilTrue=False )\n\n\tdef load(self, filename, untilTrue = True):\n\t\trun = True\n\t\tuntilTrue = False\n\t\ttry:\n\t\t\tdata = None\n\t\t\twith open(filename,\"rb\") as file:\n\t\t\t\treturn pk.load(file)\n\t\t\t\t#self.printF(\"loaded\",data, \"from\",filename)\n\t\t\treturn data\n\t\texcept:\n\t\t\tif untilTrue:\n\t\t\t\tself.printF(\"ERR - Try again, loading\",filename)\n\t\t\t\ttime.sleep(0.1)\n\t\t\t\tself.load(filename)\n\t\treturn None\n\n\tdef SaveX(self, channel, data, echo = False):\n\t\ttime.sleep(0.01)\n\t\tif echo:\n\t\t\tprint(\"Saving \",data,\"to channel\", channel)\n\t\t# gd = GlobalData()\n\t\treturn self.pub(str(channel), data, untilTrue = True)\n\n\tdef GetX(self, channel, echo = False):\n\t\ttime.sleep(0.01)\n\t\tif echo:\n\t\t\t# print(\"Loading channel\", channel)\n\t\t\tpass\n\t\t# gd = GlobalData()\n\t\treturn self.get(str(channel),untilTrue=False )\n\n\tdef LoadX(self, channel, echo = False):\n\t\treturn self.GetX(channel, echo = echo)\n\n\tdef SaveFuncX(self, channel, func, params = None):\n\t\tself.SaveX(channel+\"_params\", params)\n\t\timport dill as d\n\t\treturn self.pub(channel,d.dumps(func))\n\t\t# return SaveX(channel, d.dumps(func))\n\n\tdef GetFuncX(self, channel):\n\t\timport dill as d\n\t\ttry:\n\t\t\treturn d.loads(self.GetX(channel))\n\t\texcept:\n\t\t\treturn None\n\n\tdef printF(self, data=None, a=None, b=None, c=None, d=None,e=None,f=None,g=None,h=None,i=None,j=None):\n\t\tglobal printX\n\t\ts = \"\"\n\t\tls = [data,a,b,c,d,e,f,g,h,i,j]\n\t\tfor x in ls:\n\t\t\tif x is not None:\n\t\t\t\ts = s + str(x)\n\n\t\tif printX:\n\t\t\tprint(s)\n\n\tdef printZ(text, char, thick = 5, length = 70, newline=True, minimal = False):\n\n\t\t#x = list()\n\t\t#x.append(text)\n\t\t#return x\n\t\tTextList = list()\n\t\ttext = str(text)\n\t\thead = char\n\t\tfor l in range(0,int(length/(len(char)))):\n\t\t\thead += char\n\t\tif minimal:\n\t\t\tchar = \" \"\n\t\tif newline:\n\t\t\tprint()\n\t\t\tTextList.append(\"\")\n\t\textra = length - len(text)\n\t\tntext= \"\"\n\t\tfor l in range(0, int(extra/2)+1):\n\t\t\tntext += char\n\t\tntext+=text\n\t\tfor l in range(0, int(extra/2)):\n\t\t\tntext += char\n\t\twhile (len(ntext)MIN_ANGLE)[0]+1\n\n # 可视化\n # fig = plt.figure(figsize=(10,8))\n # ax =fig.add_subplot(111)\n # ax.plot(x, y, 'b-', label='original path')\n # ax.plot(sx, sy, 'g--', label='simplified path')\n # ax.plot(sx[idx], sy[idx], 'ro', markersize = 10, label='turning points')\n # plt.legend(loc='best')\n # plt.show()\n\n # 转折点\n res = list(zip(list(sx[idx]),list(sy[idx])))\n\n # 保存数据库\n sql = \" create TABLE 'route_brief' ('id' int(11) NOT NULL AUTO_INCREMENT, 'longitude' varchar(45) DEFAULT NULL, 'latitude' varchar(45) DEFAULT NULL, PRIMARY KEY ('id'))\"\n try:\n sql = \"DROP TABLE IF EXISTS route_brief;\"\n cursor.execute(sql)\n sql = \"\"\"CREATE TABLE route_brief (\n id int(11) NOT NULL AUTO_INCREMENT,\n longitude varchar(45) DEFAULT NULL,\n latitude varchar(45) DEFAULT NULL, \n PRIMARY KEY (id))\"\"\"\n cursor.execute(sql)\n for longitude, latitude in res:\n sql = \"insert into route_brief(longitude, latitude) values (%s, %s);\" % (str(longitude), str(latitude))\n cursor.execute(sql)\n except Exception as e:\n connect.rollback() # 事务回滚\n print('事务处理失败', e)\n else:\n connect.commit() # 事务提交\n\n # 关闭连接\n cursor.close()\n connect.close()\n\n # with open('./turningPoints.data', 'w') as f:\n # f.write(str(res))\n\n", "sub_path": "convoy/py/tracing.py", "file_name": "tracing.py", "file_ext": "py", "file_size_in_byte": 3012, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "numpy.pi", "line_number": 23, "usage_type": "attribute"}, {"api_name": "numpy.arccos", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 29, "usage_type": "call"}, {"api_name": "pymysql.cursors.Connect", "line_number": 34, "usage_type": "call"}, {"api_name": "pymysql.cursors", "line_number": 34, "usage_type": "name"}, {"api_name": "numpy.float64", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 55, "usage_type": "call"}, {"api_name": "rdp.rdp", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.diff", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 59, "usage_type": "call"}]}
+{"seq_id": "641564036", "text": "#\n# Copyright (c) 2014, Prometheus Research, LLC\n#\n\n\nfrom rex.core import Extension, cached, Error, guard\nfrom .arm import (TableArm, TrunkArm, BranchArm, FacetArm, JoinArm, ColumnArm,\n LinkArm, SyntaxArm)\nfrom .constraint import ConstraintSet, reserved_parameters\nfrom htsql.core.util import to_name\nfrom htsql.core.error import Error as HTSQLError\nfrom htsql.core.domain import (UntypedDomain, BooleanDomain, NumberDomain,\n IntegerDomain, TextDomain, EnumDomain, DateDomain, TimeDomain,\n DateTimeDomain, IdentityDomain)\nfrom htsql.core.cmd.embed import Embed\nfrom htsql.core.tr.lookup import prescribe, identify, lookup_reference\nfrom htsql.core.tr.binding import (LiteralBinding, IdentityBinding,\n ImplicitCastBinding, SieveBinding, FormulaBinding, DirectionBinding,\n SortBinding, DefineReferenceBinding, ClipBinding, LiteralRecipe)\nfrom htsql.core.tr.signature import (IsEqualSig, IsInSig, NotSig, IsNullSig,\n CompareSig, AndSig, OrSig)\nfrom htsql.core.tr.fn.signature import ContainsSig\nfrom htsql.tweak.etl.cmd.insert import Clarify\n\n\ndef embed(values):\n # Converts raw values to `Value` instances.\n try:\n return [Embed.__invoke__(value) for value in values]\n except TypeError as exc:\n raise Error(\"Cannot recognize value:\", str(exc)) from None\n\n\ndef clarify(domains, values):\n # Casts input values to the respective domains.\n if isinstance(values, list):\n if isinstance(domains, list):\n # Cast:\n # ([domain1, ...], [value1, ...]) -> [domain1(value1), ...]\n if len(domains) != len(values):\n raise Error(\"Got unexpected number of values:\",\n \"expected %s; got %s\" % (len(domains), len(values)))\n return [clarify(domain, value)\n for domain, value in zip(domains, values)]\n else:\n # Cast:\n # (domain, [value1, ...]) -> [domain(value1), ...]\n domain = domains\n return [clarify(domain, value) for value in values]\n else:\n # Cast:\n # (domain, value) -> domain(value)\n domain = domains\n value = values\n convert = Clarify.__invoke__(value.domain, domain)\n if convert is None:\n raise Error(\"Cannot convert value of type %s to %s:\"\n % (value.domain, domain),\n value)\n try:\n return convert(value.data)\n except ValueError as exc:\n raise Error(\"Failed to convert value to %s:\" % domain, str(exc)) from None\n\n\ndef locate(state, scope, value):\n # Generates:\n # `scope.id()=value`\n\n # Prepare the left operand.\n recipe = identify(scope)\n assert recipe is not None\n identity = state.use(recipe, scope.syntax, scope=scope)\n domain = identity.domain\n lops = identity.elements\n # Prepare the right operand.\n if value.domain == domain:\n rops = value.data\n elif isinstance(value.domain, UntypedDomain):\n try:\n rops = domain.parse(value.data)\n except ValueError:\n raise Error(\"Failed to convert value to %s:\" % domain, value) from None\n else:\n raise Error(\"Failed to convert value of type %s to %s:\"\n % (value.domain, domain), value)\n\n # If value is not provided, always return `False`.\n if rops is None:\n return LiteralBinding(state.scope, False, BooleanDomain(),\n scope.syntax)\n\n # Generate `lops=rops`, where `lops` is an `IdentityBinding.fields`\n # and `rops` is an `IdentityDomain` value.\n def match(lops, rops):\n images = []\n for lop, rop in zip(lops, rops):\n if isinstance(lop.domain, IdentityDomain):\n images.extend(match(lop.elements, rop))\n else:\n rop = LiteralBinding(state.scope, rop, lop.domain,\n scope.syntax)\n image = FormulaBinding(state.scope, IsEqualSig(+1),\n BooleanDomain(), state.scope.syntax,\n lop=lop, rop=rop)\n images.append(image)\n return images\n images = match(lops, rops)\n if len(images) == 1:\n return images[0]\n else:\n return FormulaBinding(state.scope, AndSig(), BooleanDomain(),\n state.scope.syntax, ops=images)\n\n\ndef ensure(arm, ArmClass):\n # Verifies if `arm` is an instance of one of the Arm classes.\n if not isinstance(arm, ArmClass):\n if isinstance(ArmClass, tuple):\n expected = \", \".join(cls.kind for cls in ArmClass[:-1]) \\\n + \" or \" + ArmClass[-1].kind\n else:\n expected = ArmClass.kind\n raise Error(\"Got unexpected arm type:\",\n \"expected %s; got %s\" % (expected, arm.kind))\n\n\nclass Condition(Extension):\n \"\"\"\n Implements a constraint operator.\n \"\"\"\n\n # Operator name.\n operator = None\n # The operator to use when the operator is not specified.\n default_operator = 'eq'\n # Expected arm types.\n kind = None\n\n @classmethod\n def enabled(cls):\n return (cls.operator is not None)\n\n @classmethod\n @cached\n def map_all(cls):\n # Generates a mapping:\n # operator -> `Constraint` subclass.\n mapping = {}\n for extension in cls.all():\n assert extension.operator not in mapping, \\\n \"duplicate condition: %r\" % extension.operator\n mapping[extension.operator] = extension\n return mapping\n\n @classmethod\n def apply(cls, arm, state, constraint, binding, scope):\n # Applies the given constraint to the binding.\n condition_map = cls.map_all()\n with guard(\"While applying constraint:\", constraint):\n operator = constraint.operator or cls.default_operator\n if operator not in condition_map:\n raise Error(\"Got unknown constraint operator:\", operator)\n arguments = embed(constraint.arguments)\n condition = condition_map[operator](arm, state, arguments)\n if condition.kind:\n ensure(arm, condition.kind)\n return condition(binding, scope)\n\n def __init__(self, arm, state, arguments):\n self.arm = arm\n self.state = state\n self.arguments = arguments\n\n def __call__(self, binding, scope):\n # `binding` is a `Binding` object to wrap; `scope` is the arm scope;\n # `self.state.scope` is the scope of the parent arm.\n raise NotImplementedError()\n\n\nclass ReferenceCondition:\n # Implements a reference definition.\n\n @classmethod\n def apply(cls, arm, state, constraint, binding, scope):\n with guard(\"While applying constraint:\", constraint):\n operator = constraint.operator\n if not (to_name(operator) in arm.parameters or\n operator in reserved_parameters):\n raise Error(\"Got unknown parameter:\", operator)\n arguments = embed(constraint.arguments)\n condition = cls(arm, state, operator, arguments)\n return condition(binding, scope)\n\n @classmethod\n def apply_missing(cls, arm, state, binding, scope):\n for name in sorted(arm.parameters)+reserved_parameters:\n name = to_name(name)\n if lookup_reference(binding, name) is not None:\n continue\n arguments = embed([arm.parameters.get(name)])\n condition = cls(arm, state, name, arguments)\n binding = condition(binding, scope)\n return binding\n\n def __init__(self, arm, state, name, arguments):\n self.arm = arm\n self.state = state\n self.name = name\n self.arguments = arguments\n\n def __call__(self, binding, scope):\n if len(self.arguments) != 1:\n raise Error(\"Got unexpected number of arguments:\",\n \"expected 1; got %s\" % len(self.arguments))\n [argument] = self.arguments\n recipe = LiteralRecipe(argument.data, argument.domain)\n binding = DefineReferenceBinding(binding, self.name, recipe,\n binding.syntax)\n return binding\n\n\nclass FilterCondition:\n # Implements a conditional filter expression on an entity arm.\n\n @classmethod\n def apply(cls, arm, state, constraint, binding, scope):\n with guard(\"While applying constraint:\", constraint):\n arguments = embed(constraint.arguments)\n condition = cls(arm, state, constraint.operator, arguments)\n return condition(binding, scope)\n\n def __init__(self, arm, state, name, arguments):\n self.arm = arm\n self.state = state\n self.name = name\n self.arguments = arguments\n self.filter = arm.filters[name]\n self.syntax = self.filter.syntax\n self.parameters = self.filter.parameters\n\n def __call__(self, binding, scope):\n # Check the number of arguments.\n if len(self.parameters) != len(self.arguments):\n raise Error(\"Got unexpected number of arguments:\",\n \"expected %s; got %s\"\n % (len(self.parameters), len(self.arguments)))\n # Add the formal parameters to the current scope.\n for parameter, argument in zip(self.parameters, self.arguments):\n recipe = LiteralRecipe(argument.data, argument.domain)\n scope = DefineReferenceBinding(scope, parameter,\n recipe, scope.syntax)\n # Try the compile the filter expression.\n try:\n condition = self.state.bind(self.syntax, scope=scope)\n except HTSQLError:\n raise Error(\"Failed to compile filter:\", self.name)\n # Wrap the given `binding`.\n condition = ImplicitCastBinding(condition,\n BooleanDomain(), condition.syntax)\n binding = SieveBinding(binding, condition,\n binding.syntax)\n return binding\n\n\nclass TopSkipCondition:\n # Takes a slice of a plural entity.\n\n @classmethod\n def apply(cls, arm, state, top_constraint, skip_constraint, binding, scope):\n constraints = []\n top_arguments = skip_arguments = None\n if top_constraint is not None:\n constraints.append(top_constraint)\n top_arguments = top_constraint.arguments\n if skip_constraint is not None:\n constraints.append(skip_constraint)\n skip_arguments = skip_constraint.arguments\n constraints = ConstraintSet(0, constraints)\n with guard(\"While applying constraint:\", constraints):\n if top_arguments is not None:\n top_arguments = embed(top_arguments)\n if skip_arguments is not None:\n skip_arguments = embed(skip_arguments)\n condition = cls(top_arguments, skip_arguments, arm, state)\n ensure(arm, (TrunkArm, BranchArm))\n return condition(binding, scope)\n\n def __init__(self, top_arguments, skip_arguments, arm, state):\n self.top_arguments = top_arguments\n self.skip_arguments = skip_arguments\n self.arm = arm\n self.state = state\n\n def __call__(self, binding, scope):\n limit = offset = None\n if self.top_arguments is not None:\n [limit] = clarify([IntegerDomain()], self.top_arguments)\n if limit < 0:\n raise Error(\"Expected non-negative integer; got:\", limit)\n if self.skip_arguments is not None:\n [offset] = clarify([IntegerDomain()], self.skip_arguments)\n if offset < 0:\n raise Error(\"Expected non-negative integer; got:\", offset)\n binding = ClipBinding(self.state.scope, binding, [], limit, offset,\n binding.syntax)\n return binding\n\n\nclass EqualCondition(Condition):\n # Implements `=` condition.\n\n operator = 'eq'\n kind = (TableArm, ColumnArm, LinkArm, SyntaxArm)\n\n def __call__(self, binding, scope):\n # With no arguments, it is unconditional `False` expression.\n if not self.arguments:\n condition = LiteralBinding(self.state.scope, False, BooleanDomain(),\n self.state.scope.syntax)\n\n # On tables, links, generate `scope.id()=arguments`.\n elif isinstance(self.arm, (TableArm, LinkArm)):\n conditions = [locate(self.state, scope, argument)\n for argument in self.arguments]\n if len(conditions) == 1:\n [condition] = conditions\n else:\n condition = FormulaBinding(self.state.scope, OrSig(),\n BooleanDomain(), self.state.scope.syntax,\n ops=conditions)\n\n # On columns and calculated fields, generate `scope=arguments`.\n else:\n arguments = clarify(self.arm.domain, self.arguments)\n literals = [LiteralBinding(self.state.scope, argument,\n self.arm.domain, self.state.scope.syntax)\n for argument in arguments]\n if len(literals) == 1:\n [literal] = literals\n condition = FormulaBinding(self.state.scope, IsEqualSig(+1),\n BooleanDomain(), self.state.scope.syntax,\n lop=scope, rop=literal)\n else:\n condition = FormulaBinding(self.state.scope, IsInSig(+1),\n BooleanDomain(), self.state.scope.syntax,\n lop=scope, rops=literals)\n\n return SieveBinding(binding, condition, binding.syntax)\n\n\nclass NullCondition(Condition):\n # Implements `is_null`.\n\n operator = 'null'\n kind = (FacetArm, JoinArm, ColumnArm, LinkArm, SyntaxArm)\n\n def __call__(self, binding, scope):\n if not self.arguments:\n argument = True\n else:\n [argument] = clarify([BooleanDomain()], self.arguments)\n if isinstance(self.arm, (TableArm, LinkArm)):\n condition = ImplicitCastBinding(scope,\n BooleanDomain(), scope.syntax)\n if argument:\n condition = FormulaBinding(self.state.scope, NotSig(),\n BooleanDomain(), scope.syntax, op=condition)\n else:\n polarity = 1 if argument is True else -1\n condition = FormulaBinding(self.state.scope, IsNullSig(polarity),\n BooleanDomain(), self.state.scope.syntax, op=scope)\n return SieveBinding(binding, condition, binding.syntax)\n\n\nclass CompareCondition(Condition):\n # Implements `<`, `<=`, `>`, `>=`.\n\n operator = None\n relation = None\n kind = (ColumnArm, SyntaxArm)\n\n def __call__(self, binding, scope):\n domain = self.arm.domain\n if not isinstance(domain, (TextDomain, NumberDomain,\n DateDomain, TimeDomain, DateTimeDomain)):\n raise Error(\"Got unsupported %s type:\" % self.arm.kind,\n \"expected text, number, date, time or datetime; \"\n \"got %s\" % domain)\n [argument] = clarify([domain], self.arguments)\n literal = LiteralBinding(self.state.scope, argument, domain,\n self.state.scope.syntax)\n condition = FormulaBinding(self.state.scope, CompareSig(self.relation),\n BooleanDomain(), self.state.scope.syntax,\n lop=scope, rop=literal)\n return SieveBinding(binding, condition, binding.syntax)\n\n\nclass LessThanCondition(CompareCondition):\n\n operator = 'lt'\n relation = '<'\n\n\nclass LessOrEqualToCondition(CompareCondition):\n\n operator = 'le'\n relation = '<='\n\n\nclass GreaterThanCondition(CompareCondition):\n\n operator = 'gt'\n relation = '>'\n\n\nclass GreaterOrEqualToCondition(CompareCondition):\n\n operator = 'ge'\n relation = '>='\n\n\nclass ContainsCondition(Condition):\n # Implements `~`.\n\n operator = 'contains'\n kind = (ColumnArm, SyntaxArm)\n\n def __call__(self, binding, scope):\n domain = self.arm.domain\n if not isinstance(domain, TextDomain):\n raise Error(\"Got unsupported %s type:\" % self.arm.kind,\n \"expected text; got %s\" % domain)\n [argument] = clarify([domain], self.arguments)\n literal = LiteralBinding(self.state.scope, argument, domain,\n self.state.scope.syntax)\n condition = FormulaBinding(self.state.scope, ContainsSig(+1),\n BooleanDomain(), self.state.scope.syntax,\n lop=scope, rop=literal)\n return SieveBinding(binding, condition, binding.syntax)\n\n\nclass SortCondition(Condition):\n # Implements `.sort()`.\n\n operator = 'sort'\n kind = (ColumnArm, SyntaxArm)\n\n def __call__(self, binding, scope):\n [argument] = clarify([EnumDomain(['asc', 'desc'])], self.arguments)\n direction = +1 if argument == 'asc' else -1\n condition = DirectionBinding(scope, direction, scope.syntax)\n return SortBinding(binding, [condition], None, None, binding.syntax)\n\n\n", "sub_path": "src/rex.port/src/rex/port/condition.py", "file_name": "condition.py", "file_ext": "py", "file_size_in_byte": 17044, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "htsql.core.cmd.embed.Embed.__invoke__", "line_number": 29, "usage_type": "call"}, {"api_name": "htsql.core.cmd.embed.Embed", "line_number": 29, "usage_type": "name"}, {"api_name": "rex.core.Error", "line_number": 31, "usage_type": "call"}, {"api_name": "rex.core.Error", "line_number": 41, "usage_type": "call"}, {"api_name": "htsql.tweak.etl.cmd.insert.Clarify.__invoke__", "line_number": 55, "usage_type": "call"}, {"api_name": "htsql.tweak.etl.cmd.insert.Clarify", "line_number": 55, "usage_type": "name"}, {"api_name": "rex.core.Error", "line_number": 57, "usage_type": "call"}, {"api_name": "rex.core.Error", "line_number": 63, "usage_type": "call"}, {"api_name": "htsql.core.tr.lookup.identify", "line_number": 71, "usage_type": "call"}, {"api_name": "htsql.core.domain.UntypedDomain", "line_number": 79, "usage_type": "argument"}, {"api_name": "rex.core.Error", "line_number": 83, "usage_type": "call"}, {"api_name": "rex.core.Error", "line_number": 85, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.LiteralBinding", "line_number": 90, "usage_type": "call"}, {"api_name": "htsql.core.domain.BooleanDomain", "line_number": 90, "usage_type": "call"}, {"api_name": "htsql.core.domain.IdentityDomain", "line_number": 98, "usage_type": "argument"}, {"api_name": "htsql.core.tr.binding.LiteralBinding", "line_number": 101, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.FormulaBinding", "line_number": 103, "usage_type": "call"}, {"api_name": "htsql.core.tr.signature.IsEqualSig", "line_number": 103, "usage_type": "call"}, {"api_name": "htsql.core.domain.BooleanDomain", "line_number": 104, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.FormulaBinding", "line_number": 112, "usage_type": "call"}, {"api_name": "htsql.core.tr.signature.AndSig", "line_number": 112, "usage_type": "call"}, {"api_name": "htsql.core.domain.BooleanDomain", "line_number": 112, "usage_type": "call"}, {"api_name": "rex.core.Error", "line_number": 124, "usage_type": "call"}, {"api_name": "arm.kind", "line_number": 125, "usage_type": "attribute"}, {"api_name": "rex.core.Extension", "line_number": 128, "usage_type": "name"}, {"api_name": "rex.core.cached", "line_number": 145, "usage_type": "name"}, {"api_name": "rex.core.guard", "line_number": 160, "usage_type": "call"}, {"api_name": "constraint.operator", "line_number": 161, "usage_type": "attribute"}, {"api_name": "rex.core.Error", "line_number": 163, "usage_type": "call"}, {"api_name": "constraint.arguments", "line_number": 164, "usage_type": "attribute"}, {"api_name": "rex.core.guard", "line_number": 186, "usage_type": "call"}, {"api_name": "constraint.operator", "line_number": 187, "usage_type": "attribute"}, {"api_name": "htsql.core.util.to_name", "line_number": 188, "usage_type": "call"}, {"api_name": "arm.parameters", "line_number": 188, "usage_type": "attribute"}, {"api_name": "constraint.reserved_parameters", "line_number": 189, "usage_type": "name"}, {"api_name": "rex.core.Error", "line_number": 190, "usage_type": "call"}, {"api_name": "constraint.arguments", "line_number": 191, "usage_type": "attribute"}, {"api_name": "arm.parameters", "line_number": 197, "usage_type": "attribute"}, {"api_name": "constraint.reserved_parameters", "line_number": 197, "usage_type": "name"}, {"api_name": "htsql.core.util.to_name", "line_number": 198, "usage_type": "call"}, {"api_name": "htsql.core.tr.lookup.lookup_reference", "line_number": 199, "usage_type": "call"}, {"api_name": "arm.parameters.get", "line_number": 201, "usage_type": "call"}, {"api_name": "arm.parameters", "line_number": 201, "usage_type": "attribute"}, {"api_name": "rex.core.Error", "line_number": 214, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.LiteralRecipe", "line_number": 217, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.DefineReferenceBinding", "line_number": 218, "usage_type": "call"}, {"api_name": "rex.core.guard", "line_number": 228, "usage_type": "call"}, {"api_name": "constraint.arguments", "line_number": 229, "usage_type": "attribute"}, {"api_name": "constraint.operator", "line_number": 230, "usage_type": "attribute"}, {"api_name": "arm.filters", "line_number": 238, "usage_type": "attribute"}, {"api_name": "rex.core.Error", "line_number": 245, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.LiteralRecipe", "line_number": 250, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.DefineReferenceBinding", "line_number": 251, "usage_type": "call"}, {"api_name": "htsql.core.error.Error", "line_number": 256, "usage_type": "name"}, {"api_name": "rex.core.Error", "line_number": 257, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.ImplicitCastBinding", "line_number": 259, "usage_type": "call"}, {"api_name": "htsql.core.domain.BooleanDomain", "line_number": 260, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.SieveBinding", "line_number": 261, "usage_type": "call"}, {"api_name": "constraint.ConstraintSet", "line_number": 279, "usage_type": "call"}, {"api_name": "rex.core.guard", "line_number": 280, "usage_type": "call"}, {"api_name": "arm.TrunkArm", "line_number": 286, "usage_type": "name"}, {"api_name": "arm.BranchArm", "line_number": 286, "usage_type": "name"}, {"api_name": "htsql.core.domain.IntegerDomain", "line_number": 298, "usage_type": "call"}, {"api_name": "rex.core.Error", "line_number": 300, "usage_type": "call"}, {"api_name": "htsql.core.domain.IntegerDomain", "line_number": 302, "usage_type": "call"}, {"api_name": "rex.core.Error", "line_number": 304, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.ClipBinding", "line_number": 305, "usage_type": "call"}, {"api_name": "arm.TableArm", "line_number": 314, "usage_type": "name"}, {"api_name": "arm.ColumnArm", "line_number": 314, "usage_type": "name"}, {"api_name": "arm.LinkArm", "line_number": 314, "usage_type": "name"}, {"api_name": "arm.SyntaxArm", "line_number": 314, "usage_type": "name"}, {"api_name": "htsql.core.tr.binding.LiteralBinding", "line_number": 319, "usage_type": "call"}, {"api_name": "htsql.core.domain.BooleanDomain", "line_number": 319, "usage_type": "call"}, {"api_name": "arm.TableArm", "line_number": 323, "usage_type": "name"}, {"api_name": "arm.LinkArm", "line_number": 323, "usage_type": "name"}, {"api_name": "htsql.core.tr.binding.FormulaBinding", "line_number": 329, "usage_type": "call"}, {"api_name": "htsql.core.tr.signature.OrSig", "line_number": 329, "usage_type": "call"}, {"api_name": "htsql.core.domain.BooleanDomain", "line_number": 330, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.LiteralBinding", "line_number": 336, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.FormulaBinding", "line_number": 341, "usage_type": "call"}, {"api_name": "htsql.core.tr.signature.IsEqualSig", "line_number": 341, "usage_type": "call"}, {"api_name": "htsql.core.domain.BooleanDomain", "line_number": 342, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.FormulaBinding", "line_number": 345, "usage_type": "call"}, {"api_name": "htsql.core.tr.signature.IsInSig", "line_number": 345, "usage_type": "call"}, {"api_name": "htsql.core.domain.BooleanDomain", "line_number": 346, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.SieveBinding", "line_number": 349, "usage_type": "call"}, {"api_name": "arm.FacetArm", "line_number": 356, "usage_type": "name"}, {"api_name": "arm.JoinArm", "line_number": 356, "usage_type": "name"}, {"api_name": "arm.ColumnArm", "line_number": 356, "usage_type": "name"}, {"api_name": "arm.LinkArm", "line_number": 356, "usage_type": "name"}, {"api_name": "arm.SyntaxArm", "line_number": 356, "usage_type": "name"}, {"api_name": "htsql.core.domain.BooleanDomain", "line_number": 362, "usage_type": "call"}, {"api_name": "arm.TableArm", "line_number": 363, "usage_type": "name"}, {"api_name": "arm.LinkArm", "line_number": 363, "usage_type": "name"}, {"api_name": "htsql.core.tr.binding.ImplicitCastBinding", "line_number": 364, "usage_type": "call"}, {"api_name": "htsql.core.domain.BooleanDomain", "line_number": 365, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.FormulaBinding", "line_number": 367, "usage_type": "call"}, {"api_name": "htsql.core.tr.signature.NotSig", "line_number": 367, "usage_type": "call"}, {"api_name": "htsql.core.domain.BooleanDomain", "line_number": 368, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.FormulaBinding", "line_number": 371, "usage_type": "call"}, {"api_name": "htsql.core.tr.signature.IsNullSig", "line_number": 371, "usage_type": "call"}, {"api_name": "htsql.core.domain.BooleanDomain", "line_number": 372, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.SieveBinding", "line_number": 373, "usage_type": "call"}, {"api_name": "arm.ColumnArm", "line_number": 381, "usage_type": "name"}, {"api_name": "arm.SyntaxArm", "line_number": 381, "usage_type": "name"}, {"api_name": "htsql.core.domain.TextDomain", "line_number": 385, "usage_type": "name"}, {"api_name": "htsql.core.domain.NumberDomain", "line_number": 385, "usage_type": "name"}, {"api_name": "htsql.core.domain.DateDomain", "line_number": 386, "usage_type": "name"}, {"api_name": "htsql.core.domain.TimeDomain", "line_number": 386, "usage_type": "name"}, {"api_name": "htsql.core.domain.DateTimeDomain", "line_number": 386, "usage_type": "name"}, {"api_name": "rex.core.Error", "line_number": 387, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.LiteralBinding", "line_number": 391, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.FormulaBinding", "line_number": 393, "usage_type": "call"}, {"api_name": "htsql.core.tr.signature.CompareSig", "line_number": 393, "usage_type": "call"}, {"api_name": "htsql.core.domain.BooleanDomain", "line_number": 394, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.SieveBinding", "line_number": 396, "usage_type": "call"}, {"api_name": "arm.ColumnArm", "line_number": 427, "usage_type": "name"}, {"api_name": "arm.SyntaxArm", "line_number": 427, "usage_type": "name"}, {"api_name": "htsql.core.domain.TextDomain", "line_number": 431, "usage_type": "argument"}, {"api_name": "rex.core.Error", "line_number": 432, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.LiteralBinding", "line_number": 435, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.FormulaBinding", "line_number": 437, "usage_type": "call"}, {"api_name": "htsql.core.tr.fn.signature.ContainsSig", "line_number": 437, "usage_type": "call"}, {"api_name": "htsql.core.domain.BooleanDomain", "line_number": 438, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.SieveBinding", "line_number": 440, "usage_type": "call"}, {"api_name": "arm.ColumnArm", "line_number": 447, "usage_type": "name"}, {"api_name": "arm.SyntaxArm", "line_number": 447, "usage_type": "name"}, {"api_name": "htsql.core.domain.EnumDomain", "line_number": 450, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.DirectionBinding", "line_number": 452, "usage_type": "call"}, {"api_name": "htsql.core.tr.binding.SortBinding", "line_number": 453, "usage_type": "call"}]}
+{"seq_id": "198634207", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Sep 27 09:48:53 2017\n\n@author: c1625914\n\"\"\"\n\nfrom __future__ import absolute_import, print_function, division\n\nimport montage_wrapper as montage\nimport matplotlib\nmatplotlib.rcParams['font.family'] = 'Latin Modern Roman'\nimport os\nimport time\nfrom astropy.io import fits\nfrom astropy.wcs import WCS\nfrom scipy.optimize import curve_fit\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom matplotlib.pyplot import cm\nfrom mpl_toolkits.axes_grid1 import make_axes_locatable\nfrom lmfit import minimize, Parameters, Model\nfrom scipy.odr import ODR, RealData\nfrom scipy.odr import Model as odrmodel\nimport sys\nimport statsmodels.api as sm\nfrom scipy import linalg, stats\nsys.path.append('/home/daedalusdata/c1625914/UsefulCode/Functions/')\nimport aplpy_edit as aplpy\nimport tgfunctions as tg\nimport warnings\n\nwarnings.simplefilter('ignore')\n\ndef standard_aplpy(figure,tick_format=None,sig_digits=0,vmin=None,vmax=None):\n figure.set_tick_labels_format(xformat='hh:mm',yformat='dd:mm')\n figure.set_tick_xspacing(0.75)\n figure.set_tick_labels_font(size='10')\n figure.set_axis_labels_font(size='12')\n figure.set_frame_color('black')\n figure.set_tick_color('black')\n if not vmin:\n figure.show_colorscale(cmap='gist_heat',pmin=1,pmax=99,interpolation='none')\n else:\n figure.show_colorscale(cmap='gist_heat',vmin=vmin,vmax=vmax,interpolation='none')\n figure.show_colorbar(tick_format=tick_format,sig_digits=sig_digits)\n figure.colorbar.set_width(0.1)\n figure.colorbar.set_pad(0)\n\ndef straight_line_fit(x,m,c):\n \n return m*x+c\n \ndef straight_line_fit_odr(beta,x):\n \n return beta[0]*x+beta[1]\n\ndef schmidt_fit(x,c):\n \n return 1.4*x+c\n \ndef linear_fit(x,a,b,n):\n \n return a*x**n+b\n \ndef minimise_straight_line(p0,x,y,errors):\n \n m = p0['m']\n c = p0['c']\n \n line_values = straight_line_fit(x,m,c)\n \n residuals = (((line_values-y)/errors))\n \n return np.array(residuals,dtype=float)\n \ndef gaussian(bin_positions,mean,st_dev,height):\n \n gaussian_function = height * np.exp( -(bin_positions-mean)**2 / (2*st_dev**2) ) \n \n return gaussian_function\n \ndef ks_annuli(data,x_centre,y_centre,a_annuli,b_annuli,annulus_number,theta=23):\n theta *= np.pi/180\n \n data_total = []\n \n for i in range(data.shape[0]):\n for j in range(data.shape[1]):\n x = x_centre-j\n y = y_centre-i\n if ((x*np.cos(theta)+y*np.sin(theta))**2/a_annuli[annulus_number]**2) + ((x*np.sin(theta)-y*np.cos(theta))**2/b_annuli[annulus_number]**2) <= 1 and \\\n ((x*np.cos(theta)+y*np.sin(theta))**2/a_annuli[annulus_number-1]**2) + ((x*np.sin(theta)-y*np.cos(theta))**2/b_annuli[annulus_number-1]**2) > 1:\n data_total.append(data[i,j])\n \n# mean = np.nanmean(ks_total)\n# st_dev = np.nanstd(ks_total)\n \n return data_total\n \ndef ks_annuli_plot(data,x_centre,y_centre,a_annuli,b_annuli,annulus_number,colour,theta=23):\n theta *= np.pi/180\n \n plt.figure(98)\n \n j_collected = []\n i_collected = []\n \n for i in range(data.shape[0]):\n for j in range(data.shape[1]):\n x = x_centre-j\n y = y_centre-i\n if ((x*np.cos(theta)+y*np.sin(theta))**2/a_annuli[annulus_number]**2) + ((x*np.sin(theta)-y*np.cos(theta))**2/b_annuli[annulus_number]**2) <= 1 and \\\n ((x*np.cos(theta)+y*np.sin(theta))**2/a_annuli[annulus_number-1]**2) + ((x*np.sin(theta)-y*np.cos(theta))**2/b_annuli[annulus_number-1]**2) > 1:\n j_collected.append(j)\n i_collected.append(i)\n \n plt.scatter(j_collected,i_collected,c=colour,lw=0,alpha=0.4) \n \ndef ks_annuli_test(data,x_centre,y_centre,a_annuli,b_annuli,annulus_number,theta=23):\n theta *= np.pi/180\n \n data_total = []\n \n for i in range(data.shape[0]):\n for j in range(data.shape[1]):\n x = x_centre-j\n y = y_centre-i\n if ((x*np.cos(theta)+y*np.sin(theta))**2/a_annuli[annulus_number]**2) + ((x*np.sin(theta)-y*np.cos(theta))**2/b_annuli[annulus_number]**2) < 1:\n data_total.append(data[i,j])\n data[i,j] = np.nan\n \n mean = np.nanmean(data_total)\n st_dev = np.nanstd(data_total)\n \n return data,mean,st_dev\n \ndef minimise_straight_line_fit(p0,x,data,errors):\n \n m = p0['m'].value \n c = p0['c'].value\n \n straight_line = straight_line_fit(x,m,c)\n residuals = ((straight_line-data)/errors)\n \n return residuals\n \ndef reject_outliers(x,y, percentile = 40):\n \n #Remove the lowest 40% of data\n\n idx = np.where( (x>np.nanpercentile(x,percentile)) & (y>np.nanpercentile(y,percentile)) )\n x = x[idx]\n y = y[idx] \n \n return x,y\n\nstart = time.clock()\n \nsteradian = 42545170296.1522 #arcsec^2\nm33_distance = 840 #kpc\nm33_ra = \t23.462051\nm33_dec = 30.660184\ninclination = 56\ninclination *= np.pi/180\n\nalpha_co = 3.1 #Msun/pc^2 (K km/s)^-1\nco_ratio = 0.7 #from 2-1 to 1-0\nco_conversion = 2.1e20 # From the HerM33es paper (https://arxiv.org/pdf/1005.3422.pdf)\nsandstrom_conversion = 1.42e20 #Sandstrom et al (2013)\ninner_co_conversion = 1.54e20\nouter_co_conversion = 2.87e20\nhi_conversion = 1.8e18\nhelium_correction = 1.36\n\nrad_conv = np.pi/(3600*180)\nmsun = 1.989e30\nmproton = 1.67e-27\npc_to_cm = 3.0857e18\n\nr_25_amount = 1\n\nplt.close('all')\n\nfig_number = 1 \nsubplot = 1\n\ntotal_sigma = 5\nsigma = 5\n\nuse_dust = 0\nuse_total_gas = 0\nuse_atomic_gas = 0\nuse_mol_gas = 1\n\nuse_braine_radial = 1\nuse_braine_total = 0\nuse_sandstrom = 0\n\n#filenames = ['magphys_sfr','M33CO','mbb_total_gas','M33HI','magphys_total_gas']\n#filenames = ['SCUBA2_850','SCUBA2_450']\n\ngmod = Model(gaussian)\n\nlinemod = odrmodel(straight_line_fit_odr)\n\n##############################MASK HI-RES DATA#################################\n\n#regrid_dir = '/home/daedalusdata/c1625914/M33/regrids/25_originals/'\n#\n#for filename in filenames:\n# \n# data,header = fits.getdata(str(regrid_dir)+str(filename)+'.fits', header=True)\n#\n# #Background subtract\n# \n# if filename in ['M33HI']:\n# data_background_level = tg.calculate_background_median(data)\n# data -= data_background_level\n# \n# if filename in ['magphys_sfr']:\n# data[data < 1e-7] = np.nan\n# \n## if filename in ['M33CO']:\n## w = WCS(str(regrid_dir)+'M33CO.fits')\n## pixscale = tg.get_pixscale(header)\n## #For rms background calculation\n## a_background = 20*60/pixscale\n## b_background = 15*60/pixscale\n## x_centre_co,y_centre_co = w.all_world2pix(m33_ra,m33_dec,1)\n## data_masked = tg.ellipse(data,x_centre_co,y_centre_co,b_background,a_background,theta=22.5)\n## co_data_min = sigma*tg.find_background_rms(data_masked)\n## data[data=0 and y>=0:\n sfr_data[i,j] = magphys_data[y,x] \n sfr_error[i,j] = magphys_error[y,x] \n except IndexError:\n pass\n \nsfr_error *= sfr_data\n \n#Match up CO\n \ndata, header = fits.getdata('/export/daedalusdata/c1625914/M33/regrids/25/M33CO.fits',header=True)\nw = WCS('/export/daedalusdata/c1625914/M33/regrids/25/M33CO.fits')\n\nco_data = np.zeros([magphys_data.shape[0],magphys_data.shape[1]])\nco_data[co_data == 0] = np.nan\n\nfor i in range(magphys_data.shape[0]):\n for j in range(magphys_data.shape[1]):\n try:\n x,y = w.all_world2pix(world_coords[i,j,0],world_coords[i,j,1],1)\n x = int(round(x))\n y = int(round(y)) \n if x>=0 and y>=0:\n co_data[i,j] = data[y,x]\n except IndexError:\n pass\n \n#Match up HI\n \ndata, header = fits.getdata('/export/daedalusdata/c1625914/M33/regrids/25/M33HI.fits',header=True)\nw = WCS('/export/daedalusdata/c1625914/M33/regrids/25/M33HI.fits')\n\nhi_data = np.zeros([magphys_data.shape[0],magphys_data.shape[1]])\nhi_data[hi_data == 0] = np.nan\n\nfor i in range(magphys_data.shape[0]):\n for j in range(magphys_data.shape[1]):\n try:\n x,y = w.all_world2pix(world_coords[i,j,0],world_coords[i,j,1],1)\n x = int(round(x))\n y = int(round(y)) \n if x>=0 and y>=0:\n hi_data[i,j] = data[y,x]\n except IndexError:\n pass\n \n##Match up dust\n# \n#data, header = fits.getdata('/export/daedalusdata/c1625914/M33/regrids/25/magphys_total_gas.fits',header=True)\n#w = WCS('/export/daedalusdata/c1625914/M33/regrids/25/magphys_total_gas.fits')\n#\n#dust_data = np.zeros([magphys_data.shape[0],magphys_data.shape[1]])\n#dust_data[dust_data == 0] = np.nan\n#\n#for i in range(magphys_data.shape[0]):\n# for j in range(magphys_data.shape[1]):\n# try:\n# x,y = w.all_world2pix(world_coords[i,j,0],world_coords[i,j,1],1)\n# x = int(round(x))\n# y = int(round(y)) \n# if x>=0 and y>=0:\n# dust_data[i,j] = data[y,x]\n# except IndexError:\n# pass\n\nschmidt_to_save = []\n\nfor pixscale in pixscales:\n \n pixscale_ratio = 25/pixscale\n \n #Pixel scale in pc^2\n \n pixel_size_pc = (pixscale*rad_conv)**2 * (m33_distance*1000)**2\n pixel_size_cm = pixel_size_pc * pc_to_cm**2\n \n #Scale the SFR data\n \n sfr_data_scaled = np.zeros([int(sfr_data.shape[0]*pixscale_ratio),int(sfr_data.shape[1]*pixscale_ratio)])\n sfr_data_scaled[sfr_data_scaled == 0] = np.nan \n sfr_error_scaled = np.zeros([int(sfr_data.shape[0]*pixscale_ratio),int(sfr_data.shape[1]*pixscale_ratio)])\n sfr_error_scaled[sfr_error_scaled == 0] = np.nan \n sfr_data_scaled_nans = np.zeros([int(sfr_data.shape[0]*pixscale_ratio),int(sfr_data.shape[1]*pixscale_ratio)])\n \n coords = np.zeros([int(sfr_data.shape[0]*pixscale_ratio),int(sfr_data.shape[1]*pixscale_ratio),2])\n\n for i in range(sfr_data_scaled.shape[0]):\n for j in range(sfr_data_scaled.shape[1]):\n \n cropped_array = sfr_data[int(i/pixscale_ratio):int(i/pixscale_ratio)+int(1/pixscale_ratio),\n int(j/pixscale_ratio):int(j/pixscale_ratio)+int(1/pixscale_ratio)] \n \n cropped_err_array = sfr_error[int(i/pixscale_ratio):int(i/pixscale_ratio)+int(1/pixscale_ratio),\n int(j/pixscale_ratio):int(j/pixscale_ratio)+int(1/pixscale_ratio)] \n \n sfr_x = np.mean([int(i/pixscale_ratio),int(i/pixscale_ratio)+int(1/pixscale_ratio)]) \n sfr_y = np.mean([int(j/pixscale_ratio),int(j/pixscale_ratio)+int(1/pixscale_ratio)]) \n \n sfr_data_scaled[i,j] = np.nansum(cropped_array) \n sfr_error_scaled[i,j] = np.sqrt(np.nansum(np.multiply(cropped_err_array,cropped_err_array))) \n \n if np.sqrt(np.nansum(np.multiply(cropped_err_array,cropped_err_array))) == 0:\n sfr_data_scaled[i,j] = np.nan \n \n sfr_data_scaled_nans[i,j] = np.sum(np.isfinite(cropped_array))\n \n coords[i,j,0], coords[i,j,1] = w_sfr.all_pix2world(sfr_y,sfr_x,1) \n \n sfr_data_scaled[sfr_data_scaled == 0] = np.nan\n sfr_error_scaled /= sfr_data_scaled\n sfr_data_scaled_nans[sfr_data_scaled_nans == 0] = np.nan\n sfr_data_scaled_nans *= pixscale_ratio**2\n \n# plt.figure(2)\n# plt.imshow(sfr_error_scaled,origin='lower',interpolation='none')\n \n #############################CO############################################\n \n co_data_scaled = np.zeros([int(co_data.shape[0]*pixscale_ratio),int(co_data.shape[1]*pixscale_ratio)])\n co_data_scaled[co_data_scaled == 0] = np.nan \n\n for i in range(co_data_scaled.shape[0]):\n for j in range(co_data_scaled.shape[1]):\n \n cropped_array = co_data[int(i/pixscale_ratio):int(i/pixscale_ratio)+int(1/pixscale_ratio),\n int(j/pixscale_ratio):int(j/pixscale_ratio)+int(1/pixscale_ratio)] \n \n co_data_scaled[i,j] = np.nanmean(cropped_array) \n \n co_data_scaled[co_data_scaled == 0] = np.nan\n \n if use_braine_radial == 1:\n \n #Convert CO to surface density of H2 gas. Take inclination into account. Also\n #take into account these two populations\n \n column_density_h2 = np.zeros([co_data_scaled.shape[0],co_data_scaled.shape[1]])\n column_density_h2[column_density_h2 == 0] = np.nan\n \n theta = 22.5*np.pi/180 \n \n inner_ring_b = 2/(m33_distance*rad_conv*pixscale)\n inner_ring_a = 0.55*2/(m33_distance*rad_conv*pixscale)\n \n for i in range(co_data_scaled.shape[0]):\n for j in range(co_data_scaled.shape[1]):\n x = x_centre-j\n y = y_centre-i\n if ((x*np.cos(theta)+y*np.sin(theta))**2/inner_ring_a**2) + ((x*np.sin(theta)-y*np.cos(theta))**2/inner_ring_b**2) <= 1:\n column_density_h2[i,j] = co_data_scaled[i,j] * inner_co_conversion/co_ratio\n else:\n column_density_h2[i,j] = co_data_scaled[i,j] * outer_co_conversion/co_ratio\n \n elif use_braine_total == 1: \n column_density_h2 = co_data_scaled * co_conversion/co_ratio\n \n elif use_sandstrom == 1:\n column_density_h2 = co_data_scaled * sandstrom_conversion/co_ratio\n \n number_density_h2 = column_density_h2 * pixel_size_cm \n mass_h2 = number_density_h2*2*mproton*helium_correction/ msun \n surface_density_h2 = mass_h2/pixel_size_pc\n \n ################################HI#########################################\n\n hi_data_scaled = np.zeros([int(hi_data.shape[0]*pixscale_ratio),int(hi_data.shape[1]*pixscale_ratio)])\n hi_data_scaled[hi_data_scaled == 0] = np.nan \n\n for i in range(hi_data_scaled.shape[0]):\n for j in range(hi_data_scaled.shape[1]):\n hi_data_scaled[i,j] = np.nanmean(hi_data[int(i/pixscale_ratio):int(i/pixscale_ratio)+int(1/pixscale_ratio),\n int(j/pixscale_ratio):int(j/pixscale_ratio)+int(1/pixscale_ratio)]) \n \n hi_data_scaled[hi_data_scaled == 0] = np.nan\n \n #Calculate surface density of HI gas.\n \n column_density_hi = hi_data_scaled * hi_conversion \n number_density_hi = column_density_hi * pixel_size_cm \n mass_hi = number_density_hi*mproton*helium_correction / msun\n \n surface_density_hi = mass_hi/pixel_size_pc\n \n ############################GAS FROM DUST################################## \n \n try:\n dust_data_magphys,dust_data_header = fits.getdata('/home/daedalusdata/c1625914/M33/regrids/'+str(pixscale)+'/m_dust.fits',header=True)\n dust_wcs = WCS('/home/daedalusdata/c1625914/M33/regrids/'+str(pixscale)+'/m_dust.fits')\n dust_data_err,dust_data_err_header = fits.getdata('/home/daedalusdata/c1625914/M33/regrids/'+str(pixscale)+'/m_dust_err.fits',header=True)\n \n dust_data_scaled = np.zeros([int(sfr_data.shape[0]*pixscale_ratio),int(sfr_data.shape[1]*pixscale_ratio)])\n dust_data_scaled[dust_data_scaled == 0] = np.nan \n \n dust_err_scaled = np.zeros([int(sfr_data.shape[0]*pixscale_ratio),int(sfr_data.shape[1]*pixscale_ratio)])\n dust_err_scaled[dust_err_scaled == 0] = np.nan \n \n for i in range(dust_data_scaled.shape[0]):\n for j in range(dust_data_scaled.shape[1]):\n try:\n x,y = dust_wcs.all_world2pix(coords[i,j,0],coords[i,j,1],1)\n x = int(round(x))\n y = int(round(y))\n if x>=0 and y>=0:\n dust_data_scaled[i,j] = dust_data_magphys[y,x]\n dust_err_scaled[i,j] = dust_data_err[y,x]\n except IndexError:\n pass\n \n #Convert dust to gas mass\n \n x_centre = dust_data_scaled.shape[0]/2\n y_centre = dust_data_scaled.shape[1]/2\n \n for i in range(dust_data_scaled.shape[0]):\n for j in range(dust_data_scaled.shape[1]):\n r = np.sqrt( (i-x_centre)**2+(j-y_centre)**2 )\n r *= pixscale\n \n #convert this r from arcsec to kpc\n r *= rad_conv\n r *= m33_distance\n \n log_dgr = -0.015*r + -1.902\n \n dust_data_scaled[i,j] /= 10**log_dgr\n \n \n surface_density_gas = dust_data_scaled/(pixel_size_pc)\n \n except IOError:\n\n print('File not found eh',pixscale) \n dust_data_scaled = np.zeros([int(dust_data.shape[0]*pixscale_ratio),int(dust_data.shape[1]*pixscale_ratio)])\n dust_data_scaled[dust_data_scaled == 0] = np.nan \n dust_data_scaled_nans = np.zeros([int(dust_data.shape[0]*pixscale_ratio),int(dust_data.shape[1]*pixscale_ratio)])\n \n for i in range(dust_data_scaled.shape[0]):\n for j in range(dust_data_scaled.shape[1]):\n \n cropped_array = dust_data[int(i/pixscale_ratio):int(i/pixscale_ratio)+int(1/pixscale_ratio),\n int(j/pixscale_ratio):int(j/pixscale_ratio)+int(1/pixscale_ratio)] \n \n dust_data_scaled[i,j] = np.nansum(cropped_array) \n dust_data_scaled_nans[i,j] = np.sum(np.isfinite(cropped_array))\n \n dust_data_scaled[dust_data_scaled == 0] = np.nan\n dust_data_scaled_nans[dust_data_scaled_nans == 0] = np.nan\n dust_err_scaled[dust_err_scaled == 0] = np.nan\n dust_data_scaled_nans *= pixscale_ratio**2\n \n surface_density_gas = dust_data_scaled/(pixel_size_pc*dust_data_scaled_nans)\n \n #SFR from MAGPHYS\n \n sigma_sfr = ( (sfr_data_scaled)*1e6 )/ (pixel_size_pc)#*sfr_data_scaled_nans)\n \n sigma_sfr_flatten = sigma_sfr.flatten()\n \n #############################KS LAW########################################\n \n ##############################SFR##########################################\n\n sfr_density = sigma_sfr.copy() \n log_sfr_density = np.log10(sfr_density)\n \n log_sfr_density_flatten = log_sfr_density.flatten()\n sfr_err_flatten = sfr_error_scaled.flatten()\n \n ############################GAS############################################\n \n if use_dust == 1:\n sigma_gas_density = surface_density_gas.copy()\n sigma_gas_err = dust_err_scaled.copy()\n sigma_gas_err = dust_err_scaled.flatten()\n elif use_total_gas == 1:\n sigma_gas_density = (surface_density_h2+surface_density_hi).copy()\n elif use_atomic_gas == 1:\n sigma_gas_density = surface_density_hi.copy()\n elif use_mol_gas == 1:\n sigma_gas_density = surface_density_h2.copy()\n \n log_total_gas_density = np.log10(sigma_gas_density)\n \n log_total_gas_density_flatten = log_total_gas_density.flatten()\n \n total_gas_index = np.isfinite(log_total_gas_density_flatten) & np.isfinite(log_sfr_density_flatten)\n \n x_to_plot = log_total_gas_density_flatten[total_gas_index]\n y_to_plot = log_sfr_density_flatten[total_gas_index] \n y_err = sfr_err_flatten[total_gas_index]\n \n if use_dust == 1:\n x_err = sigma_gas_err[total_gas_index]\n \n xmin = x_to_plot.min()\n xmax = x_to_plot.max()\n ymin = y_to_plot.min()\n ymax = y_to_plot.max() \n \n ##############################ERRORS#######################################\n \n sfr_calculation_uncertainty = 0.32\n log_sfr_uncertainty = 0.434*sfr_calculation_uncertainty\n \n log_sfr_error_flatten = log_sfr_density_flatten*log_sfr_uncertainty\n \n co_calibration_uncertainty = 0.15\n hi_calibration_uncertainty = 0.02\n \n if use_dust == 1:\n log_gas_uncertainty = log_total_gas_density_flatten*log_sfr_uncertainty\n else:\n log_gas_uncertainty = 0.434*co_calibration_uncertainty\n \n total_uncertainty = np.sqrt(log_sfr_uncertainty**2+log_gas_uncertainty**2)\n \n ##########################SFR/GAS PLOT#####################################\n\n x_to_fit = x_to_plot.copy()\n y_to_fit = y_to_plot.copy()\n \n x_to_clip = x_to_plot.copy()\n y_to_clip = y_to_plot.copy()\n \n #Spearman's rank and Pearson correlation\n \n spearman_correlation = stats.spearmanr(x_to_plot,y_to_plot)\n pearson_correlation = stats.pearsonr(x_to_plot,y_to_plot)[0]\n \n print(pixscale,spearman_correlation,pearson_correlation)\n spearman_correlation = spearman_correlation.correlation\n \n #Save the average x and y to put on the global KS plot\n \n if pixscale == 25:\n global_ks = []\n global_ks.append(np.nanmean(x_to_fit))\n global_ks.append(np.nanmean(y_to_fit))\n if use_dust == 1:\n np.savetxt('/export/daedalusdata/c1625914/M33/arrays_for_plots/dust_gas_global_ks.txt', global_ks)\n elif use_total_gas == 1:\n np.savetxt('/export/daedalusdata/c1625914/M33/arrays_for_plots/tot_gas_global_ks.txt', global_ks)\n elif use_atomic_gas == 1:\n pass\n elif use_mol_gas == 1:\n np.savetxt('/export/daedalusdata/c1625914/M33/arrays_for_plots/mol_gas_global_ks.txt', global_ks)\n \n\n popt,pcov = curve_fit(straight_line_fit,x_to_fit,y_to_fit)\n schmidt_index,offset = popt\n schmidt_index_error = np.sqrt(np.diag(pcov))[0]*schmidt_index\n \n schmidt_to_save.append(schmidt_index)\n \n if pixscale == 25:\n schmidt_to_correct = schmidt_index\n \n chi_square_scatter = ( np.divide(y_to_fit-(np.multiply(x_to_fit,schmidt_index)+offset),log_sfr_error_flatten[total_gas_index]) )**2\n \n #Best fit with classic Schmidt index 1.4\n \n classic_offset = curve_fit(schmidt_fit, x_to_fit, y_to_fit)[0] \n x = np.linspace(np.min(x_to_plot),np.max(x_to_plot)) \n y = schmidt_index*x+offset \n y_classic = 1.4*x+classic_offset\n \n chi_square_scatter = ( np.divide(y_to_fit-(np.multiply(x_to_fit,1.4)+classic_offset),log_sfr_error_flatten[total_gas_index]) )**2\n \n #Plot all this up\n \n plt.figure(1)\n \n if len(pixscales) == 1:\n ax = plt.subplot(1,np.ceil(len(pixscales)/2),subplot)\n else: \n ax = plt.subplot(2,np.ceil(len(pixscales)/2),subplot)\n plt.scatter(x_to_plot,y_to_plot)\n \n plt.plot(x,y,'r',linewidth=3.0)\n plt.plot(x,y_classic,'g',linewidth=3.0)\n \n plt.xlim([xmin,xmax])\n plt.ylim([ymin,ymax])\n \n if subplot == 1 or subplot == 4: \n plt.ylabel(r'log[$\\Sigma_{SFR}(M_{\\odot}\\mathregular{yr}^{-1}\\mathregular{kpc}^{-2})$]')\n plt.xlabel(r'log[$\\Sigma_{gas}(M_{\\odot}\\mathregular{pc}^{-2})$]')\n plt.title(str(pixscale)+r'\", $N$ = '+str('{0:.2f}'.format(schmidt_index))+'$\\pm$'+str('{0:.2f}'.format(schmidt_index_error)))\n \n tx = plt.text(0.05,0.95,r'$\\rho_{\\mathregular{sp}} = '+str('{0:.2f}'.format(spearman_correlation))+'$',\n verticalalignment='top', horizontalalignment='left',\n transform = ax.transAxes) \n tx.set_bbox(dict(color='white', alpha=0.8))\n \n if use_dust == 1:\n np.savetxt('/export/daedalusdata/c1625914/M33/arrays_for_plots/dust_gas_x_to_plot_'+str(pixscale)+'.txt', x_to_plot)\n np.savetxt('/export/daedalusdata/c1625914/M33/arrays_for_plots/dust_gas_x_err_to_plot_'+str(pixscale)+'.txt', x_err)\n np.savetxt('/export/daedalusdata/c1625914/M33/arrays_for_plots/dust_gas_y_to_plot_'+str(pixscale)+'.txt', y_to_plot)\n np.savetxt('/export/daedalusdata/c1625914/M33/arrays_for_plots/dust_gas_y_err_to_plot_'+str(pixscale)+'.txt', y_err)\n elif use_total_gas == 1:\n np.savetxt('/export/daedalusdata/c1625914/M33/arrays_for_plots/tot_gas_x_to_plot_'+str(pixscale)+'.txt', x_to_plot)\n np.savetxt('/export/daedalusdata/c1625914/M33/arrays_for_plots/tot_gas_y_to_plot_'+str(pixscale)+'.txt', y_to_plot) \n np.savetxt('/export/daedalusdata/c1625914/M33/arrays_for_plots/tot_gas_y_err_to_plot_'+str(pixscale)+'.txt', y_err)\n elif use_atomic_gas == 1:\n pass\n elif use_mol_gas == 1:\n np.savetxt('/export/daedalusdata/c1625914/M33/arrays_for_plots/mol_gas_x_to_plot_'+str(pixscale)+'.txt', x_to_plot)\n np.savetxt('/export/daedalusdata/c1625914/M33/arrays_for_plots/mol_gas_y_to_plot_'+str(pixscale)+'.txt', y_to_plot)\n np.savetxt('/export/daedalusdata/c1625914/M33/arrays_for_plots/mol_gas_y_err_to_plot_'+str(pixscale)+'.txt', y_err)\n \n subplot += 1\n \n##############################ANNULI############################################\n \n#def ks_annuli(data,x_centre,y_centre,a_annuli,b_annuli,annulus_number,theta=23):\n# theta *= np.pi/180\n# \n# data_total = []\n# \n# for i in range(data.shape[0]):\n# for j in range(data.shape[1]):\n# x = x_centre-j\n# y = y_centre-i\n# if ((x*np.cos(theta)+y*np.sin(theta))**2/a_annuli[annulus_number]**2) + ((x*np.sin(theta)-y*np.cos(theta))**2/b_annuli[annulus_number]**2) <= 1 and \\\n# ((x*np.cos(theta)+y*np.sin(theta))**2/a_annuli[annulus_number-1]**2) + ((x*np.sin(theta)-y*np.cos(theta))**2/b_annuli[annulus_number-1]**2) > 1:\n# data_total.append(data[i,j])\n# \n# return data_total\n \n number_of_annuli = 6\n a_annuli = np.linspace(0,a,number_of_annuli)\n b_annuli = np.linspace(0,b,number_of_annuli)\n radius_fraction = np.linspace(0,1,number_of_annuli)\n schmidt_annuli = [np.nan]\n schmidt_annuli_err = [np.nan] \n \n colour=iter(cm.rainbow(np.linspace(0,1,number_of_annuli)))\n \n if pixscale == 25:\n \n for annulus_number in range(1,number_of_annuli):\n \n sfr_annuli = ks_annuli(log_sfr_density,x_centre,y_centre,a_annuli,b_annuli,annulus_number,theta=22.5)\n gas_annuli = ks_annuli(log_total_gas_density,x_centre,y_centre,a_annuli,b_annuli,annulus_number,theta=22.5) \n y_err = ks_annuli(sfr_error_scaled,x_centre,y_centre,a_annuli,b_annuli,annulus_number,theta=22.5) \n x_err = ks_annuli(dust_err_scaled,x_centre,y_centre,a_annuli,b_annuli,annulus_number,theta=22.5) \n \n idx = np.isfinite(sfr_annuli) & np.isfinite(gas_annuli)\n \n if use_dust == 1:\n np.savetxt('/export/daedalusdata/c1625914/M33/arrays_for_plots/dust_'+str(annulus_number)+'_annulus.txt',np.c_[np.asarray(gas_annuli)[idx],np.asarray(sfr_annuli)[idx],np.asarray(y_err)[idx],np.asarray(x_err)[idx]]) \n elif use_total_gas == 1:\n np.savetxt('/export/daedalusdata/c1625914/M33/arrays_for_plots/tot_'+str(annulus_number)+'_annulus.txt',np.c_[np.asarray(gas_annuli)[idx],np.asarray(sfr_annuli)[idx],np.asarray(y_err)[idx]])\n elif use_atomic_gas == 1:\n pass\n elif use_mol_gas == 1:\n np.savetxt('/export/daedalusdata/c1625914/M33/arrays_for_plots/mol_'+str(annulus_number)+'_annulus.txt',np.c_[np.asarray(gas_annuli)[idx],np.asarray(sfr_annuli)[idx],np.asarray(y_err)[idx]])", "sub_path": "sf_law/ks_law_final.py", "file_name": "ks_law_final.py", "file_ext": "py", "file_size_in_byte": 30387, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "matplotlib.rcParams", "line_number": 12, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 28, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "warnings.simplefilter", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 84, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 102, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "numpy.cos", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 118, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 118, "usage_type": "name"}, {"api_name": "numpy.pi", "line_number": 121, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 131, "usage_type": "attribute"}, {"api_name": "numpy.nanmean", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.nanstd", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.nanpercentile", "line_number": 152, "usage_type": "call"}, {"api_name": "time.clock", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 165, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 176, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.close", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}, {"api_name": "lmfit.Model", "line_number": 203, "usage_type": "call"}, {"api_name": "scipy.odr.Model", "line_number": 205, "usage_type": "call"}, {"api_name": "astropy.io.fits.getdata", "line_number": 269, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 269, "usage_type": "name"}, {"api_name": "astropy.io.fits.getdata", "line_number": 270, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 270, "usage_type": "name"}, {"api_name": "astropy.wcs.WCS", "line_number": 271, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 272, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 274, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 275, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 278, "usage_type": "attribute"}, {"api_name": "tgfunctions.ellipse", "line_number": 286, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 290, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 294, "usage_type": "call"}, {"api_name": "astropy.io.fits.getdata", "line_number": 310, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 310, "usage_type": "name"}, {"api_name": "astropy.wcs.WCS", "line_number": 311, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 313, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 314, "usage_type": "attribute"}, {"api_name": "astropy.io.fits.getdata", "line_number": 329, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 329, "usage_type": "name"}, {"api_name": "astropy.wcs.WCS", "line_number": 330, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 332, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 333, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 378, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 379, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 380, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 381, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 382, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 384, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 395, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 396, "usage_type": "call"}, {"api_name": "numpy.nansum", "line_number": 398, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 399, "usage_type": "call"}, {"api_name": "numpy.nansum", "line_number": 399, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 399, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 401, "usage_type": "call"}, {"api_name": "numpy.nansum", "line_number": 401, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 401, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 402, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 404, "usage_type": "call"}, {"api_name": "numpy.isfinite", "line_number": 404, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 408, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 410, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 418, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 419, "usage_type": "attribute"}, {"api_name": "numpy.nanmean", "line_number": 427, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 429, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 436, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 437, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 439, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 448, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 448, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 465, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 466, "usage_type": "attribute"}, {"api_name": "numpy.nanmean", "line_number": 470, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 473, "usage_type": "attribute"}, {"api_name": "astropy.io.fits.getdata", "line_number": 486, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 486, "usage_type": "name"}, {"api_name": "astropy.wcs.WCS", "line_number": 487, "usage_type": "call"}, {"api_name": "astropy.io.fits.getdata", "line_number": 488, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 488, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 490, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 491, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 493, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 494, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 515, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 532, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 533, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 534, "usage_type": "call"}, {"api_name": "numpy.nansum", "line_number": 542, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 543, "usage_type": "call"}, {"api_name": "numpy.isfinite", "line_number": 543, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 545, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 546, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 547, "usage_type": "attribute"}, {"api_name": "numpy.log10", "line_number": 563, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 581, "usage_type": "call"}, {"api_name": "numpy.isfinite", "line_number": 585, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 614, "usage_type": "call"}, {"api_name": "scipy.stats.spearmanr", "line_number": 626, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 626, "usage_type": "name"}, {"api_name": "scipy.stats.pearsonr", "line_number": 627, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 627, "usage_type": "name"}, {"api_name": "numpy.nanmean", "line_number": 636, "usage_type": "call"}, {"api_name": "numpy.nanmean", "line_number": 637, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 639, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 641, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 645, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 648, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 650, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 650, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 657, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 657, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 661, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 662, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 662, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 662, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 666, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 666, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 670, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 670, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 673, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 673, "usage_type": "name"}, {"api_name": "numpy.ceil", "line_number": 673, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 675, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 675, "usage_type": "name"}, {"api_name": "numpy.ceil", "line_number": 675, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 676, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 676, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 678, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 678, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 679, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 679, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 681, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 681, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 682, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 682, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 685, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 685, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 686, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 686, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 687, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 687, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 689, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 689, "usage_type": "name"}, {"api_name": "numpy.savetxt", "line_number": 695, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 696, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 697, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 698, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 700, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 701, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 702, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 706, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 707, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 708, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 730, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 731, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 732, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 733, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 734, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.cm.rainbow", "line_number": 736, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 736, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 736, "usage_type": "call"}, {"api_name": "numpy.isfinite", "line_number": 747, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 750, "usage_type": "call"}, {"api_name": "numpy.c_", "line_number": 750, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 750, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 752, "usage_type": "call"}, {"api_name": "numpy.c_", "line_number": 752, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 752, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 756, "usage_type": "call"}, {"api_name": "numpy.c_", "line_number": 756, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 756, "usage_type": "call"}]}
+{"seq_id": "6469629", "text": "\"\"\"\nScript to start a production server on Kubernetes. This script can serve as the mainApplicationFile for the SparkApplication custom resource of the spark-operator\n\"\"\"\n\nimport logging\nimport os\n\nfrom openeo_driver.server import run_gunicorn\nfrom openeo_driver.util.logging import get_logging_config, setup_logging\nfrom openeo_driver.views import build_app\nfrom openeogeotrellis import deploy\nfrom openeogeotrellis.configparams import ConfigParams\nfrom openeogeotrellis.deploy import get_socket\nfrom openeogeotrellis.job_registry import ZkJobRegistry\n\nlog = logging.getLogger(__name__)\n\n\ndef main():\n setup_logging(get_logging_config(\n root_handlers=[\"stderr_json\"],\n loggers={\n \"openeo\": {\"level\": \"DEBUG\"},\n \"openeo_driver\": {\"level\": \"DEBUG\"},\n 'openeogeotrellis': {'level': 'DEBUG'},\n \"flask\": {\"level\": \"DEBUG\"},\n \"werkzeug\": {\"level\": \"DEBUG\"},\n \"gunicorn\": {\"level\": \"INFO\"},\n 'kazoo': {'level': 'WARN'},\n },\n ))\n\n from pyspark import SparkContext\n log.info(\"starting spark context\")\n SparkContext.getOrCreate()\n\n def setup_batch_jobs():\n if not ConfigParams().is_ci_context:\n with ZkJobRegistry() as job_registry:\n job_registry.ensure_paths()\n\n def on_started():\n app.logger.setLevel('DEBUG')\n deploy.load_custom_processes()\n setup_batch_jobs()\n\n from openeogeotrellis.backend import GeoPySparkBackendImplementation\n\n app = build_app(backend_implementation=GeoPySparkBackendImplementation())\n\n host = os.environ.get('SPARK_LOCAL_IP', None)\n if host is None:\n host, _ = get_socket()\n port = os.environ.get('KUBE_OPENEO_API_PORT', 50001)\n\n run_gunicorn(\n app,\n threads=10,\n host=host,\n port=port,\n on_started=on_started\n )\n\n\nif __name__ == '__main__':\n main()\n", "sub_path": "openeogeotrellis/deploy/kube.py", "file_name": "kube.py", "file_ext": "py", "file_size_in_byte": 1907, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.getLogger", "line_number": 16, "usage_type": "call"}, {"api_name": "openeo_driver.util.logging.setup_logging", "line_number": 20, "usage_type": "call"}, {"api_name": "openeo_driver.util.logging.get_logging_config", "line_number": 20, "usage_type": "call"}, {"api_name": "pyspark.SparkContext.getOrCreate", "line_number": 35, "usage_type": "call"}, {"api_name": "pyspark.SparkContext", "line_number": 35, "usage_type": "name"}, {"api_name": "openeogeotrellis.configparams.ConfigParams", "line_number": 38, "usage_type": "call"}, {"api_name": "openeogeotrellis.job_registry.ZkJobRegistry", "line_number": 39, "usage_type": "call"}, {"api_name": "openeogeotrellis.deploy.load_custom_processes", "line_number": 44, "usage_type": "call"}, {"api_name": "openeogeotrellis.deploy", "line_number": 44, "usage_type": "name"}, {"api_name": "openeo_driver.views.build_app", "line_number": 49, "usage_type": "call"}, {"api_name": "openeogeotrellis.backend.GeoPySparkBackendImplementation", "line_number": 49, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 51, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 51, "usage_type": "attribute"}, {"api_name": "openeogeotrellis.deploy.get_socket", "line_number": 53, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 54, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 54, "usage_type": "attribute"}, {"api_name": "openeo_driver.server.run_gunicorn", "line_number": 56, "usage_type": "call"}]}
+{"seq_id": "568709257", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nfrom django.utils.timezone import utc\nimport datetime\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('pdt', '0006_auto_20150712_1921'),\n ]\n\n operations = [\n migrations.CreateModel(\n name='Operation',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('operation_no', models.IntegerField(verbose_name='Nr. rejsu', null=True, blank=True)),\n ('operation_type', models.CharField(verbose_name='Rodzaj lotu', max_length=3)),\n ('pax', models.IntegerField(verbose_name='Liczba pasa?erow', null=True, blank=True)),\n ('bags', models.IntegerField(verbose_name='Ci??ar bagazu [kg]', null=True, blank=True)),\n ('fuel_refill', models.IntegerField(verbose_name='Uzupe?nione paliwo', default=0)),\n ('fuel_available', models.IntegerField(verbose_name='Stan paliwa do lotu')),\n ('oil_lh_refill', models.DecimalField(verbose_name='Uzupe?niony olej', default=0, decimal_places=1, max_digits=3)),\n ('oil_rh_refill', models.DecimalField(verbose_name='Uzupe?niony olej RH', default=0, decimal_places=1, max_digits=3)),\n ('oil_state', models.CharField(verbose_name='Olej w normie', null=True, choices=[('ok', 'OK')], blank=True, max_length=3)),\n ('trans_oil_refill', models.DecimalField(verbose_name='Uzupe?niony olej przek?.', null=True, max_digits=3, decimal_places=1, blank=True, default=0)),\n ('trans_oil_state', models.CharField(verbose_name='Olej przek?. w normie', null=True, choices=[('ok', 'OK')], blank=True, max_length=3)),\n ('hydr_oil_refill', models.DecimalField(verbose_name='Uzupe?niony olej hydr.', null=True, max_digits=3, decimal_places=1, blank=True, default=0)),\n ('hydr_oil_state', models.CharField(verbose_name='Olej hydr. w normie', null=True, choices=[('ok', 'OK')], blank=True, max_length=3)),\n ('loc_start', models.CharField(verbose_name='Miejsce startu', max_length=20)),\n ('time_start', models.TimeField(verbose_name='Czas off-block')),\n ('tth_start', models.DecimalField(verbose_name='Licznik pocz?tkowy', max_digits=6, decimal_places=1)),\n ('loc_end', models.CharField(verbose_name='Miejsce l?dowania', max_length=20)),\n ('time_end', models.TimeField(verbose_name='Czas on-block')),\n ('tth_end', models.DecimalField(verbose_name='Licznik ko?cowy', max_digits=6, decimal_places=1)),\n ('landings', models.IntegerField(verbose_name='Liczba l?dowa?')),\n ('pdt', models.ForeignKey(to='pdt.PDT')),\n ],\n options={\n },\n bases=(models.Model,),\n ),\n migrations.AddField(\n model_name='pdt',\n name='close_time',\n field=models.DateTimeField(verbose_name='Czas zamkni?cia', null=True, blank=True),\n preserve_default=True,\n ),\n migrations.AddField(\n model_name='pdt',\n name='n1_cycles',\n field=models.IntegerField(verbose_name='Liczba cykli N1', null=True, blank=True),\n preserve_default=True,\n ),\n migrations.AddField(\n model_name='pdt',\n name='n2_cycles',\n field=models.IntegerField(verbose_name='Liczba cykli N2', null=True, blank=True),\n preserve_default=True,\n ),\n migrations.AddField(\n model_name='pdt',\n name='open_time',\n field=models.DateTimeField(verbose_name='Czas otwarcia', default=datetime.datetime(2015, 7, 13, 20, 24, 38, 5412, tzinfo=utc)),\n preserve_default=False,\n ),\n migrations.AlterField(\n model_name='pdt',\n name='pic',\n field=models.ForeignKey(verbose_name='Pierwszy pilot', to='pdt.Pilot'),\n preserve_default=True,\n ),\n migrations.AlterField(\n model_name='pdt',\n name='sic',\n field=models.ForeignKey(verbose_name='Drugi pilot', to='pdt.Pilot', null=True, related_name='pdt_sic_set', blank=True),\n preserve_default=True,\n ),\n ]\n", "sub_path": "pdt/migrations/0007_auto_20150713_2224.py", "file_name": "0007_auto_20150713_2224.py", "file_ext": "py", "file_size_in_byte": 4402, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 9, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 9, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 24, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.models.DecimalField", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.models.DecimalField", "line_number": 27, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 27, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 28, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 28, "usage_type": "name"}, {"api_name": "django.db.models.DecimalField", "line_number": 29, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 29, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.models.DecimalField", "line_number": 31, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 31, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 32, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 32, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 33, "usage_type": "name"}, {"api_name": "django.db.models.TimeField", "line_number": 34, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 34, "usage_type": "name"}, {"api_name": "django.db.models.DecimalField", "line_number": 35, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 35, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 36, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 36, "usage_type": "name"}, {"api_name": "django.db.models.TimeField", "line_number": 37, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 37, "usage_type": "name"}, {"api_name": "django.db.models.DecimalField", "line_number": 38, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 38, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 39, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 39, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 40, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 40, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 44, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 44, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 46, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 46, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 49, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 49, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 52, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 52, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 55, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 55, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 58, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 58, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 61, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 61, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 64, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 64, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 67, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 67, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 67, "usage_type": "call"}, {"api_name": "django.utils.timezone.utc", "line_number": 67, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 70, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 70, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 73, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 73, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 76, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 76, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 79, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 79, "usage_type": "name"}]}
+{"seq_id": "653156577", "text": "import random\nimport requests\nimport json\nimport pprint\nimport time\n\n\ndef mkdir(path):\n import os\n path = path.strip()\n path = path.rstrip(\"\\\\\")\n is_exist = os.path.exists(path)\n if not is_exist:\n os.makedirs(path)\n print(path + ' create path successfully')\n return True\n else:\n print(path + ' path already exist')\n return False\n\n\ndef get_dict(approved_date):\n b_id = random.randint(100000, 999999)\n url = r'https://osu.ppy.sh/beatmapsets/search?cursor%5Bapproved_date%5D=' + str(\n approved_date) + '&cursor%5B_id%5D=' + str(b_id)\n # url = r'https://osu.ppy.sh/beatmapsets/search?cursor%5Bapproved_date%5D=1512104463000&cursor%5B_id%5D=685822'\n response = requests.get(url)\n if response.status_code == 200:\n with open('archive/archive_' + str(approved_date) + r'.txt', 'w') as f:\n f.write(response.text)\n f.close()\n d = json.loads(response.text)\n return d\n else:\n print(\"{} failed\".format(url))\n return None\n\n\ndef judge_is_std(beat_maps):\n for beat_map in beat_maps:\n if beat_map['mode'] != 'osu':\n return False\n return True\n\n\ndef download_osz(d):\n count_success = 0\n count_fail = 0\n for i in range(49):\n oid = str(d['beatmapsets'][i]['id'])\n if d['beatmapsets'][i]['ranked'] == 1 and judge_is_std(d['beatmapsets'][i]['beatmaps']):\n url = r'https://bloodcat.com/osu/s/' + oid\n r = requests.get(url)\n if len(r.content) > 100:\n path = r\"oszs/\" + oid + \".osz\"\n with open(path, \"wb\") as f:\n f.write(r.content)\n print(\"download {}.osz successfully!\".format(oid))\n count_success += 1\n f.close()\n else:\n count_fail += 1\n print(\"{} is not accessible!\".format(oid))\n time.sleep(5)\n print(\"success {} in 50\".format(count_success))\n print(\"fail {} in 50\".format(count_fail))\n\n\ndef main():\n approved_date = 1572104463000\n mkdir('archive')\n mkdir('oszs')\n for i in range(1, 7):\n print(\"start approved_date:{}\".format(approved_date))\n d = get_dict(approved_date)\n # pprint.pprint(d)\n if d is not None:\n download_osz(d)\n print(\"end approved_date:{}\".format(approved_date))\n approved_date -= 500000000\n\n\nmain()\n", "sub_path": "beat_map_sets_json_crawler.py", "file_name": "beat_map_sets_json_crawler.py", "file_ext": "py", "file_size_in_byte": 2439, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.path.exists", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 14, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 23, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 27, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 32, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 53, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 64, "usage_type": "call"}]}
+{"seq_id": "339277526", "text": "from typing import Union, Optional, Callable\nimport torch\nimport torch.nn.functional as F\nfrom torch.nn import Parameter, ReLU, LeakyReLU\nfrom torch_sparse import spspmm\nfrom torch_scatter import scatter_add, scatter_max\nfrom torch_geometric.nn import GCNConv, GATConv, Sequential\nfrom torch_geometric.nn.conv import FastRGCNConv, RGCNConv\nfrom torch_geometric.utils import (add_self_loops, sort_edge_index,\n remove_self_loops, softmax)\nfrom torch_geometric.utils.repeat import repeat\nimport random\n\nimport sys\nimport os\nsys.path.append(os.environ[\"GRAPH_SUM\"])\nsys.path.append(\"/u/helbling/GraphSummarization/src/external_code/allRank/\")\nimport src.external_code.allRank.allrank.models.losses.neuralNDCG as neuralNDCG\n\ndef topk(score, indices, ratio, batch, min_score=None, tol=1e-7, num_output_sentences=None):\n if indices is None:\n indices = torch.arange(score.shape[0])\n assert score.shape == indices.shape\n\n sentence_ratio = ratio\n word_ratio = ratio\n num_nodes = scatter_add(batch.new_ones(score.size(0)), batch, dim=0)\n batch_size, max_num_nodes = num_nodes.size(0), num_nodes.max().item()\n\n cum_num_nodes = torch.cat(\n [num_nodes.new_zeros(1),\n num_nodes.cumsum(dim=0)[:-1]], dim=0)\n #print(\"batch\")\n #print(batch)\n #print(\"num nodes\")\n #print(num_nodes)\n #print(\"score shape\")\n #print(score.shape)\n #print(\"indices shape\")\n #print(indices.shape)\n #print(\"cum num nodes\")\n #print(cum_num_nodes)\n assert cum_num_nodes.shape == num_nodes.shape\n \n index = torch.arange(batch.size(0), dtype=torch.long, device=score.device)\n index = (index - cum_num_nodes[batch]) + (batch * max_num_nodes)\n\n dense_x = score.new_full((batch_size * max_num_nodes, ),\n torch.finfo(score.dtype).min)\n #print(\"index\")\n #print(index)\n #print(\"dense x shape\")\n #print(dense_x.shape)\n dense_x[index] = score\n dense_x = dense_x.view(batch_size, max_num_nodes)\n _, perm = dense_x.sort(dim=-1, descending=True)\n #print(_)\n #print(perm)\n perm = perm + cum_num_nodes.view(-1, 1)\n perm = perm.view(-1)\n if not num_output_sentences is None:\n k = torch.ones(batch_size) * num_output_sentences\n k = k.to(score.device)\n print(k.device)\n print(num_nodes.device)\n k = torch.min(k, num_nodes)\n elif isinstance(ratio, int):\n k = num_nodes.new_full((num_nodes.size(0), ), ratio)\n k = torch.min(k, num_nodes)\n else:\n k = (ratio * num_nodes.to(torch.float)).ceil().to(torch.long)\n\n mask = [\n torch.arange(k[i], dtype=torch.long, device=score.device) +\n i * max_num_nodes for i in range(batch_size)\n ]\n\n mask = torch.cat(mask, dim=0)\n perm = perm[mask]\n # filter indices to be in range\n #in_range = torch.logical_and(perm >= 0, perm < indices.shape[0])\n #perm = perm[torch.nonzero(perm >= 0)]\n #perm = perm[torch.nonzero(perm < indices.shape[0])]\n indices_perm = indices[perm]\n\n return indices_perm\n\ndef filter_adj(edge_index, edge_attr, perm, num_nodes=None):\n mask = perm.new_full((num_nodes, ), -1)\n i = torch.arange(perm.size(0), dtype=torch.long, device=perm.device)\n mask[perm] = i\n row, col = edge_index\n row, col = mask[row], mask[col]\n mask = (row >= 0) & (col >= 0)\n row, col = row[mask], col[mask]\n\n if edge_attr is not None:\n edge_attr = edge_attr[mask]\n\n return torch.stack([row, col], dim=0), edge_attr\n\ndef convert_y_to_onehot(y, num_sentences):\n \"\"\"\n Converts the y vector to a one hot encoded vector\n \"\"\"\n batch_size = num_sentences.shape[0]\n onehot_vectors = []\n for index in range(batch_size):\n current_num_sentences = num_sentences[index].int()\n current_y = y[index]\n current_y = current_y[torch.nonzero(current_y > 0)].squeeze().long()\n base_vector = torch.zeros(current_num_sentences).to(y.device)\n base_vector[current_y] = 1\n onehot_vectors.append(base_vector)\n\n return onehot_vectors\n\nclass TopKPooling(torch.nn.Module):\n r\"\"\":math:`\\mathrm{top}_k` pooling operator from the `\"Graph U-Nets\"\n `_, `\"Towards Sparse\n Hierarchical Graph Classifiers\" `_\n and `\"Understanding Attention and Generalization in Graph Neural\n Networks\" `_ papers\n\n Args:\n in_channels (int): Size of each input sample.\n ratio (float or int): Graph pooling ratio, which is used to compute\n :math:`k = \\lceil \\mathrm{ratio} \\cdot N \\rceil`, or the value\n of :math:`k` itself, depending on whether the type of :obj:`ratio`\n is :obj:`float` or :obj:`int`.\n This value is ignored if :obj:`min_score` is not :obj:`None`.\n (default: :obj:`0.5`)\n min_score (float, optional): Minimal node score :math:`\\tilde{\\alpha}`\n which is used to compute indices of pooled nodes\n :math:`\\mathbf{i} = \\mathbf{y}_i > \\tilde{\\alpha}`.\n When this value is not :obj:`None`, the :obj:`ratio` argument is\n ignored. (default: :obj:`None`)\n multiplier (float, optional): Coefficient by which features gets\n multiplied after pooling. This can be useful for large graphs and\n when :obj:`min_score` is used. (default: :obj:`1`)\n nonlinearity (torch.nn.functional, optional): The nonlinearity to use.\n (default: :obj:`torch.tanh`)\n \"\"\"\n def __init__(self, in_channels: int, ratio: Union[int, float] = 0.5,\n min_score: Optional[float] = None, multiplier: float = 1.,\n nonlinearity: Callable = torch.tanh, hidden=1024, gat=False):\n super(TopKPooling, self).__init__()\n self.in_channels = in_channels\n self.ratio = ratio\n self.min_score = min_score\n self.gat = gat\n self.multiplier = multiplier\n self.nonlinearity = nonlinearity\n self.device = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n self.hidden = hidden \n if not gat:\n self.model = Sequential(\"x, edge_index, edge_weight\", [\n (GCNConv(self.in_channels, self.hidden), \"x, edge_index, edge_weight -> x\"),\n LeakyReLU(inplace=True),\n (GCNConv(self.hidden, self.hidden // 2), \"x, edge_index, edge_weight -> x\"),\n LeakyReLU(inplace=True),\n (GCNConv(self.hidden // 2, 1), \"x, edge_index, edge_weight -> x\"),\n #LeakyReLU(inplace=True),\n #(GCNConv(self.hidden // 4, 1), \"x, edge_index, edge_weight -> x\"),\n LeakyReLU(inplace=True)\n ])\n else:\n self.model = Sequential(\"x, edge_index\", [\n (GATConv(self.in_channels, self.hidden), \"x, edge_index -> x\"),\n LeakyReLU(inplace=True),\n (GATConv(self.hidden, self.hidden // 2), \"x, edge_index -> x\"),\n LeakyReLU(inplace=True),\n (GATConv(self.hidden // 2, 1), \"x, edge_index -> x\"),\n #LeakyReLU(inplace=True),\n #(GCNConv(self.hidden // 4, 1), \"x, edge_index, edge_weight -> x\"),\n LeakyReLU(inplace=True)\n ])\n\n \n\n self.reset_parameters()\n\n def reset_parameters(self):\n pass\n\n def __repr__(self):\n return '{}({}, {}={}, multiplier={})'.format(\n self.__class__.__name__, self.in_channels,\n 'ratio' if self.min_score is None else 'min_score',\n self.ratio if self.min_score is None else self.min_score,\n self.multiplier)\n\n def forward(self, x, edge_index, edge_attr=None, batch=None, attn=None, num_sentences=None, num_output_sentences=3, is_last_layer=True):\n \"\"\"\"\"\"\n if batch is None:\n batch = edge_index.new_zeros(x.size(0))\n\n #attn = x if attn is None else attn\n #attn = attn.unsqueeze(-1) if attn.dim() == 1 else attn\n #score = (attn * self.weight).sum(dim=-1)\n edge_weight = edge_attr.squeeze()\n if not self.gat:\n score = self.model(x, edge_index, edge_weight).squeeze()\n else:\n score = self.model(x, edge_index).squeeze()\n score = torch.nan_to_num(score)\n #score = softmax(score, batch)\n # get seperate permutation for sentences and the rest (word and document nodes)\n num_nodes = scatter_add(batch.new_ones(x.size(0)), batch, dim=0).to(x.device)\n num_sentence_nodes = num_sentences.to(x.device) # the number of sentence nodes in each batch graph \n cumulative_num_nodes = torch.cat(\n [num_nodes.new_zeros(1),\n num_nodes.cumsum(dim=0)[:-1]]\n , dim=0).to(x.device)\n batch_size = num_nodes.size(0)\n # get the sentence perm\n sentence_indices = [torch.arange(num_sentence_nodes[i], device=x.device) + cumulative_num_nodes[i] for i in range(batch_size)]\n sentence_indices = torch.cat(sentence_indices).long()\n sentence_scores = score[sentence_indices]\n sentence_batch = [torch.ones(num_sentence_nodes[i]) * i for i in range(len(num_sentence_nodes))]\n sentence_batch = torch.cat(sentence_batch).long().to(x.device)\n sentence_perm = topk(sentence_scores, sentence_indices, self.ratio, sentence_batch, self.min_score, num_output_sentences=num_output_sentences)\n # get the other perm\n num_other_nodes = num_nodes - num_sentence_nodes\n if num_other_nodes.sum() > 0:\n other_indices = [torch.arange(num_other_nodes[i], device=x.device) + cumulative_num_nodes[i] + num_sentence_nodes[i] for i in range(batch_size)]\n other_indices = torch.cat(other_indices).long().to(x.device)\n other_scores = score[other_indices]\n other_batch = [torch.ones(num_other_nodes[i]) * i for i in range(len(num_other_nodes))]\n other_batch = torch.cat(other_batch).long().to(x.device)\n other_perm = topk(other_scores, other_indices, self.ratio, other_batch, self.min_score)\n # combine the other and sentence perms\n perm = torch.cat((sentence_perm, other_perm)) \n else:\n perm = sentence_perm\n # process the batch\n x = x[perm] * score[perm].view(-1, 1)\n x = self.multiplier * x if self.multiplier != 1 else x\n\n batch = batch[perm]\n edge_index, edge_attr = filter_adj(edge_index, edge_attr, perm,\n num_nodes=score.size(0))\n\n return x, edge_index, edge_attr, batch, perm, score[perm], sentence_scores, sentence_batch\n\nclass GraphUNetCoarsening(torch.nn.Module):\n r\"\"\"The Graph U-Net model from the `\"Graph U-Nets\"\n `_ paper which implements a U-Net like\n architecture with graph pooling and unpooling operations.\n\n Args:\n in_channels (int): Size of each input sample.\n hidden_channels (int): Size of each hidden sample.\n out_channels (int): Size of each output sample.\n depth (int): The depth of the U-Net architecture.\n pool_ratios (float or [float], optional): Graph pooling ratio for each\n depth. (default: :obj:`0.5`)\n sum_res (bool, optional): If set to :obj:`False`, will use\n concatenation for integration of skip connections instead\n summation. (default: :obj:`True`)\n activation (torch.nn.functional, optional): The nonlinearity to use.\n (default: :obj:`torch.nn.functional.relu`)\n \"\"\"\n def __init__(self, in_channels, hidden_channels, out_channels, depth,\n pool_ratios=0.5, sum_res=True, activation=F.relu, gat=False, embedding_mapping=False):\n super(GraphUNetCoarsening, self).__init__()\n assert depth >= 1\n self.in_channels = in_channels\n self.hidden_channels = hidden_channels\n self.out_channels = out_channels\n self.depth = depth\n self.pool_ratios = repeat(pool_ratios, depth)\n self.activation = activation\n self.embedding_mapping = embedding_mapping\n self.sum_res = sum_res\n self.loss = torch.nn.MSELoss()\n self.supervised_loss = torch.nn.BCELoss()\n self.triplet_loss = torch.nn.TripletMarginLoss()\n channels = hidden_channels\n embedding_channels = 1024\n # embedding function \n if self.embedding_mapping:\n self.embedding_function = torch.nn.Linear(self.in_channels, embedding_channels) \n self.inverse_embedding_function = torch.nn.Linear(embedding_channels, out_channels)\n else:\n self.embedding_function = torch.nn.Identity()\n self.inverse_embedding_function = torch.nn.Identity()\n embedding_channels = in_channels\n\n self.down_convs = torch.nn.ModuleList()\n self.pools = torch.nn.ModuleList()\n if gat:\n base_module = GATConv\n else:\n base_module = GCNConv\n self.down_convs.append(base_module(embedding_channels, channels))\n for i in range(depth):\n self.pools.append(TopKPooling(channels, self.pool_ratios[i], gat=gat))\n self.down_convs.append(base_module(channels, channels))\n\n in_channels = channels if sum_res else 2 * channels\n\n self.up_convs = torch.nn.ModuleList()\n for i in range(depth - 1):\n self.up_convs.append(base_module(in_channels, channels))\n self.up_convs.append(base_module(in_channels, embedding_channels))\n #self.output_conv = GCNConv(out_channels, out_channels)\n self.reset_parameters()\n\n def reset_parameters(self):\n for conv in self.down_convs:\n conv.reset_parameters()\n for pool in self.pools:\n pool.reset_parameters()\n for conv in self.up_convs:\n conv.reset_parameters()\n\n def augment_adj(self, edge_index, edge_weight, num_nodes):\n edge_index, edge_weight = remove_self_loops(edge_index, edge_weight)\n edge_index, edge_weight = add_self_loops(edge_index, edge_weight,\n num_nodes=num_nodes)\n edge_index, edge_weight = sort_edge_index(edge_index, edge_weight,\n num_nodes)\n edge_index, edge_weight = spspmm(edge_index, edge_weight, edge_index,\n edge_weight, num_nodes, num_nodes,\n num_nodes)\n edge_index, edge_weight = remove_self_loops(edge_index, edge_weight)\n return edge_index, edge_weight\n\n def __repr__(self):\n return '{}({}, {}, {}, depth={}, pool_ratios={})'.format(\n self.__class__.__name__, self.in_channels, self.hidden_channels,\n self.out_channels, self.depth, self.pool_ratios)\n\n \"\"\"\n Computes the supervised ranking loss\n \"\"\"\n def compute_supervised_ranking_loss(self, data, sentence_scores, sentence_batch):\n max_label_length = 10\n num_sentences = data.num_sentences\n y_all = data.y\n y_all = torch.reshape(y_all, (-1, max_label_length))\n batch_size = sentence_batch.max() + 1\n # go through each element\n loss = 0.0\n for index in range(batch_size):\n # compute output onehot vector\n current_indices = torch.nonzero(sentence_batch == index)\n scores = sentence_scores[current_indices].squeeze()\n output_onehot = torch.sigmoid(scores)\n # convert label to a onehot vector\n y = y_all[index]\n label_indices = y[torch.nonzero(y > 0)].long().squeeze()\n if len(label_indices.shape) == 0:\n label_indices = label_indices[None]\n current_num_sentences = num_sentences[index]\n label_onehot = torch.zeros(current_num_sentences).to(data.x.device).float()\n label_onehot[label_indices] = 1.0\n # compute the loss\n loss += self.supervised_loss(output_onehot, label_onehot)\n \n mean_loss = loss / batch_size\n return mean_loss\n \n \"\"\"\n Computes the supervised loss\n \"\"\"\n def compute_supervised_loss(self, data, sentence_scores, sentence_batch):\n max_label_length = 10\n num_sentences = data.num_sentences\n y_all = data.y\n y_all = torch.reshape(y_all, (-1, max_label_length))\n batch_size = sentence_batch.max() + 1\n # go through each element\n loss = 0.0\n for index in range(batch_size):\n # compute output onehot vector\n current_indices = torch.nonzero(sentence_batch == index)\n scores = sentence_scores[current_indices].squeeze()\n output_onehot = torch.sigmoid(scores)\n if len(output_onehot.shape) < 1:\n output_onehot = output_onehot.unsqueeze(0)\n # convert label to a onehot vector\n y = y_all[index]\n label_indices = y[torch.nonzero(y > 0)].long().squeeze()\n if len(label_indices.shape) == 0:\n label_indices = label_indices[None]\n current_num_sentences = num_sentences[index]\n label_onehot = torch.zeros(current_num_sentences).to(data.x.device).float()\n label_onehot[label_indices] = 1.0\n # compute the loss\n loss += self.supervised_loss(output_onehot, label_onehot)\n \n mean_loss = loss / batch_size\n return mean_loss\n\n \"\"\"\n Computes a basic MSE between an input graph and a \n reconstructed graph. \n \"\"\"\n def compute_unsupervised_loss(self, original_x, reconstructed_x, num_sentences=None, sentence_only=False, batch=None):\n if not num_sentences is None and sentence_only:\n x = original_x\n # computes the loss solely based on the sentence nodes\n num_nodes = scatter_add(batch.new_ones(x.size(0)), batch, dim=0).to(x.device)\n num_sentence_nodes = num_sentences.to(x.device) # the number of sentence nodes in each batch graph \n cumulative_num_nodes = torch.cat(\n [num_nodes.new_zeros(1),\n num_nodes.cumsum(dim=0)[:-1]]\n , dim=0).to(x.device)\n batch_size = num_nodes.size(0)\n # get the sentence perm\n sentence_indices = [torch.arange(num_sentence_nodes[i], device=x.device) + cumulative_num_nodes[i] for i in range(batch_size)]\n sentence_indices = torch.cat(sentence_indices).long()\n # select the correct nodes\n original_sentence_x = original_x[sentence_indices]\n reconstructed_sentence_x = reconstructed_x[sentence_indices]\n loss = self.loss(original_sentence_x, reconstructed_sentence_x)\n else:\n loss = self.loss(original_x, reconstructed_x)\n return loss\n\n \"\"\"\n Computes the basic triplet loss of a given set of output scores on\n a set of triplets constructed from the specific examples triplets\n \"\"\"\n def compute_triplet_loss(self, sentence_scores, rankings, sentence_batch, batch, num_sentences, num_triplets=5, ndcg=True):\n if ndcg:\n return self.compute_neural_ndcg_loss(sentence_scores, rankings, sentence_batch, batch, num_sentences)\n # generate a list of triplets from the rankings\n triplets = []\n batch_size = sentence_batch.max() + 1\n total_triplet_loss = 0.0\n for index in range(batch_size):\n current_scores = sentence_scores[torch.nonzero(sentence_batch == index)]\n current_rankings = rankings[torch.nonzero(sentence_batch == index)]\n #current_rankings = current_rankings[0: current_rankings.shape[0]]\n assert current_scores.shape == current_rankings.shape\n num_scores = current_scores.shape[0]\n if num_scores <= 3:\n continue\n current_triplet_loss = 0.0\n for triplet_num in range(num_triplets):\n random_indices = random.sample(range(0, num_scores), 3)\n # sort the numbers as anchor, positive, negative\n anchor_rank = current_rankings[random_indices[0]]\n first_rank = current_rankings[random_indices[1]]\n second_rank = current_rankings[random_indices[2]]\n # is first closer than second\n first_closer_than_second = abs(anchor_rank - first_rank) < abs(anchor_rank - second_rank)\n if not first_closer_than_second:\n # swap indices\n temp = random_indices[1]\n random_indices[1] = random_indices[2]\n random_indices[2] = temp\n # evaluate the triplet\n anchor_score = current_scores[random_indices[0]].unsqueeze(0)\n left_score = current_scores[random_indices[1]].unsqueeze(0)\n right_score = current_scores[random_indices[2]].unsqueeze(0)\n # calculate the triplet margin loss of the scores\n triplet_loss = self.triplet_loss(anchor_score, left_score, right_score)\n current_triplet_loss += triplet_loss\n\n total_triplet_loss += current_triplet_loss / num_triplets\n\n return total_triplet_loss\n\n def compute_neural_ndcg_loss(self, sentence_scores, rankings, sentence_batch, batch, num_sentences):\n batch_size = sentence_batch.max() + 1\n slate_size = 200 # max num sentences\n # reshape the inputs\n stacked_scores = []\n stacked_rankings = []\n for index in range(batch_size):\n # sentence_scores should be [batch_size, slate_size]\n # rankings should be [batch_size, slate_size] \n current_scores = sentence_scores[torch.nonzero(sentence_batch == index)]\n current_rankings = rankings[torch.nonzero(sentence_batch == index)]\n # pad the scores and rankings with -1 to have the desired size\n if current_scores.shape[0] == 1:\n continue\n current_scores = current_scores.squeeze()\n current_scores_shape = current_scores.shape[0]\n current_scores = F.pad(input=current_scores, pad=(0, slate_size - current_scores_shape), mode='constant', value=-1)\n stacked_scores.append(current_scores)\n current_rankings = current_rankings.squeeze()\n current_rankings_shape = current_rankings.shape[0]\n current_rankings = F.pad(input=current_rankings, pad=(0, slate_size - current_rankings_shape), mode='constant', value=-1)\n stacked_rankings.append(current_rankings.squeeze())\n stacked_scores = torch.stack(stacked_scores)\n stacked_rankings = torch.stack(stacked_rankings)\n assert stacked_scores.shape == stacked_rankings.shape\n # calculate loss\n loss = -1 * neuralNDCG(stacked_scores, stacked_rankings, padded_value_indicator=-1)\n return loss \n\n \"\"\"\n def forward(self, input_graph, num_output_sentences=3):\n # unpack input\n data_list = input_graph.to_data_list()\n loss = 0.0\n coarse_indices = []\n for data in data_list:\n x, edge_index, edge_attr = data.x, data.edge_index, data.edge_attr\n edge_weight = edge_attr.squeeze().float()\n num_sentences = data.num_sentences\n y = data.y\n # run the layers\n x = self.layer_one(x, edge_index, edge_weight)\n output_attention = self.layer_two(x, edge_index, edge_weight).squeeze()\n assert output_attention.shape[0] == x.shape[0]\n # convert the attention to a onehot vector\n sentence_attention = output_attention[0:num_sentences]\n # get topk\n num_output_nodes = torch.nonzero(y >= 0).shape[0]\n if num_output_nodes == 0:\n num_output_nodes = 1\n topk_values, topk_indices = torch.topk(sentence_attention, num_output_nodes)\n cutoff = topk_values[-1]\n # get all coarse indices\n all_topk_values, all_topk_indices = torch.topk(output_attention, x.shape[0])\n coarse_inds = all_topk_indices[torch.nonzero(all_topk_values >= cutoff).squeeze()]\n coarse_indices.append(coarse_inds)\n # compute output \n output_onehot = torch.sigmoid(sentence_attention)\n # convert label to a onehot vector\n label_indices = y[torch.nonzero(y > 0)].long().squeeze()\n if len(label_indices.shape) == 0:\n label_indices = label_indices[None]\n label_onehot = torch.zeros(num_sentences).to(device).float()\n label_onehot[label_indices] = 1.0\n # compute the loss\n loss += self.supervised_loss(output_onehot, label_onehot)\n \"\"\" \n \n \"\"\"\n Performs forward pass \n \"\"\"\n def forward(self, data, num_output_sentences=3, batch=None, mode=\"supervised\"):\n if batch is None:\n batch = data.batch\n original_batch = batch.clone()\n # unpack the data\n x, y, edge_index, edge_attr, num_sentences = data.x, data.y, data.edge_index, data.edge_attr, data.num_sentences\n rankings = data.rankings\n # convert edge_attr to edge_weight\n edge_weight = edge_attr.squeeze().float()\n edge_index = edge_index.long()\n x = x.float()\n # save a copy of the input x\n original_x = x.clone()\n # do input embedding function\n x = self.embedding_function(x)\n # perform first layer down convolution \n x = self.down_convs[0](x, edge_index)\n x = self.activation(x)\n xs = [x]\n edge_indices = [edge_index]\n edge_weights = [edge_weight]\n perms = []\n last_batch = None\n # perform downward pooling and convolutions\n \n for i in range(1, self.depth + 1):\n edge_index, edge_weight = self.augment_adj(edge_index, edge_weight,\n x.size(0))\n is_last_layer = not i < self.depth\n x, edge_index, edge_weight, batch, perm, perm_scores, sentence_scores, sentence_batch = self.pools[i - 1](\n x,\n edge_index,\n edge_attr=edge_weight, \n num_sentences=num_sentences, \n num_output_sentences=num_output_sentences, \n batch=batch,\n is_last_layer=is_last_layer)\n\n x = self.down_convs[i](x, edge_index)\n x = self.activation(x)\n if i < self.depth:\n xs += [x]\n edge_indices += [edge_index]\n edge_weights += [edge_weight]\n perms += [perm]\n last_batch = batch\n coarsened_indices = perms[-1]\n # perform upward pooling and convolutions\n for i in range(self.depth):\n j = self.depth - 1 - i\n\n res = xs[j]\n edge_index = edge_indices[j]\n edge_weight = edge_weights[j]\n perm = perms[j]\n\n up = torch.zeros_like(res)\n up[perm] = x\n x = res + up if self.sum_res else torch.cat((res, up), dim=-1)\n\n x = self.up_convs[i](x, edge_index)\n x = self.activation(x) if i < self.depth - 1 else x\n # output embedding\n x = self.inverse_embedding_function(x)\n # compute the loss\n reconstructed_x = x\n supervised_loss = self.compute_supervised_loss(data, sentence_scores, sentence_batch)\n unsupervised_loss = self.compute_unsupervised_loss(original_x, reconstructed_x, num_sentences=num_sentences, batch=original_batch)\n triplet_loss = self.compute_triplet_loss(sentence_scores, rankings, sentence_batch, batch, num_sentences)\n if mode == \"supervised\":\n loss = supervised_loss\n elif mode == \"triplet\":\n loss = triplet_loss\n elif mode == \"unsupervised\":\n loss = unsupervised_loss\n # make a loss dictionary\n loss_dict = {\n \"supervised_loss\": supervised_loss,\n \"unsupervised_loss\": unsupervised_loss,\n \"triplet_loss\": triplet_loss, \n }\n\n return loss_dict, loss, coarsened_indices\n", "sub_path": "src/ot_coarsening/u_net.py", "file_name": "u_net.py", "file_ext": "py", "file_size_in_byte": 28849, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sys.path.append", "line_number": 16, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 16, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 17, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "torch.arange", "line_number": 22, "usage_type": "call"}, {"api_name": "torch_scatter.scatter_add", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.arange", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.long", "line_number": 45, "usage_type": "attribute"}, {"api_name": "torch.finfo", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.min", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.min", "line_number": 69, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 71, "usage_type": "attribute"}, {"api_name": "torch.long", "line_number": 71, "usage_type": "attribute"}, {"api_name": "torch.arange", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.long", "line_number": 74, "usage_type": "attribute"}, {"api_name": "torch.cat", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.arange", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.long", "line_number": 90, "usage_type": "attribute"}, {"api_name": "torch.stack", "line_number": 100, "usage_type": "call"}, {"api_name": "torch.nonzero", "line_number": 111, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 112, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 118, "usage_type": "attribute"}, {"api_name": "typing.Union", "line_number": 144, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 145, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 146, "usage_type": "name"}, {"api_name": "torch.tanh", "line_number": 146, "usage_type": "attribute"}, {"api_name": "torch.device", "line_number": 154, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 154, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 154, "usage_type": "attribute"}, {"api_name": "torch_geometric.nn.Sequential", "line_number": 157, "usage_type": "call"}, {"api_name": "torch_geometric.nn.GCNConv", "line_number": 158, "usage_type": "call"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 159, "usage_type": "call"}, {"api_name": "torch_geometric.nn.GCNConv", "line_number": 160, "usage_type": "call"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 161, "usage_type": "call"}, {"api_name": "torch_geometric.nn.GCNConv", "line_number": 162, "usage_type": "call"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 165, "usage_type": "call"}, {"api_name": "torch_geometric.nn.Sequential", "line_number": 168, "usage_type": "call"}, {"api_name": "torch_geometric.nn.GATConv", "line_number": 169, "usage_type": "call"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 170, "usage_type": "call"}, {"api_name": "torch_geometric.nn.GATConv", "line_number": 171, "usage_type": "call"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 172, "usage_type": "call"}, {"api_name": "torch_geometric.nn.GATConv", "line_number": 173, "usage_type": "call"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 176, "usage_type": "call"}, {"api_name": "torch.nan_to_num", "line_number": 206, "usage_type": "call"}, {"api_name": "torch_scatter.scatter_add", "line_number": 209, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 211, "usage_type": "call"}, {"api_name": "torch.arange", "line_number": 217, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 218, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 220, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 221, "usage_type": "call"}, {"api_name": "torch.arange", "line_number": 226, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 227, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 229, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 230, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 233, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 246, "usage_type": "attribute"}, {"api_name": "torch.nn.functional.relu", "line_number": 265, "usage_type": "attribute"}, {"api_name": "torch.nn.functional", "line_number": 265, "usage_type": "name"}, {"api_name": "torch_geometric.utils.repeat.repeat", "line_number": 272, "usage_type": "call"}, {"api_name": "torch.nn.MSELoss", "line_number": 276, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 276, "usage_type": "attribute"}, {"api_name": "torch.nn.BCELoss", "line_number": 277, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 277, "usage_type": "attribute"}, {"api_name": "torch.nn.TripletMarginLoss", "line_number": 278, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 278, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 283, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 283, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 284, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 284, "usage_type": "attribute"}, {"api_name": "torch.nn.Identity", "line_number": 286, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 286, "usage_type": "attribute"}, {"api_name": "torch.nn.Identity", "line_number": 287, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 287, "usage_type": "attribute"}, {"api_name": "torch.nn.ModuleList", "line_number": 290, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 290, "usage_type": "attribute"}, {"api_name": "torch.nn.ModuleList", "line_number": 291, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 291, "usage_type": "attribute"}, {"api_name": "torch_geometric.nn.GATConv", "line_number": 293, "usage_type": "name"}, {"api_name": "torch_geometric.nn.GCNConv", "line_number": 295, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 303, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 303, "usage_type": "attribute"}, {"api_name": "torch_geometric.utils.remove_self_loops", "line_number": 319, "usage_type": "call"}, {"api_name": "torch_geometric.utils.add_self_loops", "line_number": 320, "usage_type": "call"}, {"api_name": "torch_geometric.utils.sort_edge_index", "line_number": 322, "usage_type": "call"}, {"api_name": "torch_sparse.spspmm", "line_number": 324, "usage_type": "call"}, {"api_name": "torch_geometric.utils.remove_self_loops", "line_number": 327, "usage_type": "call"}, {"api_name": "torch.reshape", "line_number": 342, "usage_type": "call"}, {"api_name": "torch.nonzero", "line_number": 348, "usage_type": "call"}, {"api_name": "torch.sigmoid", "line_number": 350, "usage_type": "call"}, {"api_name": "torch.nonzero", "line_number": 353, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 357, "usage_type": "call"}, {"api_name": "torch.reshape", "line_number": 372, "usage_type": "call"}, {"api_name": "torch.nonzero", "line_number": 378, "usage_type": "call"}, {"api_name": "torch.sigmoid", "line_number": 380, "usage_type": "call"}, {"api_name": "torch.nonzero", "line_number": 385, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 389, "usage_type": "call"}, {"api_name": "torch_scatter.scatter_add", "line_number": 405, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 407, "usage_type": "call"}, {"api_name": "torch.arange", "line_number": 413, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 414, "usage_type": "call"}, {"api_name": "torch.nonzero", "line_number": 435, "usage_type": "call"}, {"api_name": "torch.nonzero", "line_number": 436, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 444, "usage_type": "call"}, {"api_name": "torch.nonzero", "line_number": 477, "usage_type": "call"}, {"api_name": "torch.nonzero", "line_number": 478, "usage_type": "call"}, {"api_name": "torch.nn.functional.pad", "line_number": 484, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 484, "usage_type": "name"}, {"api_name": "torch.nn.functional.pad", "line_number": 488, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 488, "usage_type": "name"}, {"api_name": "torch.stack", "line_number": 490, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 491, "usage_type": "call"}, {"api_name": "src.external_code.allRank.allrank.models.losses.neuralNDCG", "line_number": 494, "usage_type": "call"}, {"api_name": "torch.zeros_like", "line_number": 595, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 597, "usage_type": "call"}]}
+{"seq_id": "194815564", "text": "\"\"\"\nFilename: calc_heat_budget_change.py\nAuthor: Damien Irving, irving.damien@gmail.com\nDescription: Calculate the heat budget change between two time periods\n\n\"\"\"\n\n# Import general Python modules\n\nimport sys, os, pdb\nimport collections\nimport argparse\nimport numpy\nimport pandas\nimport iris\nimport iris.coord_categorisation\n\n\n# Import my modules\n\ncwd = os.getcwd()\nrepo_dir = '/'\nfor directory in cwd.split('/')[1:]:\n repo_dir = os.path.join(repo_dir, directory)\n if directory == 'ocean-analysis':\n break\n\nmodules_dir = os.path.join(repo_dir, 'modules')\nsys.path.append(modules_dir)\ntry:\n import general_io as gio\n import convenient_universal as uconv\n import spatial_weights\n import timeseries\nexcept ImportError:\n raise ImportError('Must run this script from anywhere within the ocean-analysis git repo')\n\n\n# Define functions\n\nhistory = []\n\ndef save_history(cube, field, filename):\n \"\"\"Save the history attribute when reading the data.\n (This is required because the history attribute differs between input files \n and is therefore deleted upon equilising attributes) \n \"\"\" \n\n history.append(cube.attributes['history'])\n\n\ndef get_attributes(cube):\n \"\"\"Get the model, experiment and rip information.\"\"\"\n\n model = cube.attributes['model_id']\n experiment = cube.attributes['experiment_id']\n \n realization = cube.attributes['realization']\n initialization = cube.attributes['initialization_method']\n physics = cube.attributes['physics_version']\n \n rip = 'r%si%sp%s' %(realization, initialization, physics)\n\n return model, experiment, rip\n\n\ndef generate_results(data_dict, cube_list, time_constraint, time_bounds):\n \"\"\"Generate results.\"\"\"\n\n model, experiment, rip = get_attributes(cube_list[0])\n period = '%s-%s' %(time_bounds[0].split('-')[0], time_bounds[1].split('-')[0])\n\n data_dict['model'].append(model)\n data_dict['experiment'].append(experiment)\n data_dict['rip'].append(rip)\n data_dict['period'].append(period)\n\n for cube in cube_list:\n cube = cube.copy()\n temporal_subset = cube.extract(time_constraint)\n if 'ohc' in cube.var_name:\n agg_method = iris.analysis.MEAN\n elif 'hfds' in cube.var_name:\n agg_method = iris.analysis.SUM\n clim = temporal_subset.collapsed('time', agg_method)\n data_dict[clim.var_name].append(float(clim.data))\n\n return data_dict\n\n \ndef main(inargs):\n \"\"\"Run the program.\"\"\"\n\n hist_cube_list = iris.load(inargs.historical_file)\n control_cube_list = iris.load(inargs.control_file)\n\n total_time = (inargs.start_time[0], inargs.end_time[-1])\n\n hist_start_constraint = gio.get_time_constraint(inargs.start_time)\n hist_end_constraint = gio.get_time_constraint(inargs.end_time)\n hist_total_constraint = gio.get_time_constraint(total_time)\n\n control_start_constraint = timeseries.get_control_time_constraint(control_cube_list[0], hist_cube_list[0], inargs.start_time)\n control_end_constraint = timeseries.get_control_time_constraint(control_cube_list[0], hist_cube_list[0], inargs.end_time)\n control_total_constraint = timeseries.get_control_time_constraint(control_cube_list[0], hist_cube_list[0], total_time)\n \n column_headers = ['model', 'experiment', 'rip', 'period',\n 'hfds-globe-sum', 'hfds-nh-sum', 'hfds-sh-sum', 'hfds-nhext-sum', 'hfds-tropics-sum', 'hfds-shext-sum',\n 'ohc-globe-sum', 'ohc-nh-sum', 'ohc-sh-sum', 'ohc-nhext-sum', 'ohc-tropics-sum', 'ohc-shext-sum']\n\n data_dict = collections.OrderedDict()\n for column in column_headers:\n data_dict[column] = []\n\n data_dict = generate_results(data_dict, hist_cube_list, hist_start_constraint, inargs.start_time)\n data_dict = generate_results(data_dict, hist_cube_list, hist_end_constraint, inargs.end_time)\n data_dict = generate_results(data_dict, hist_cube_list, hist_total_constraint, total_time)\n\n data_dict = generate_results(data_dict, control_cube_list, control_start_constraint, inargs.start_time)\n data_dict = generate_results(data_dict, control_cube_list, control_end_constraint, inargs.end_time)\n data_dict = generate_results(data_dict, control_cube_list, control_total_constraint, total_time)\n\n data_df = pandas.DataFrame.from_dict(data_dict)\n data_df.to_csv(inargs.outfile)\n\n metadata_dict = {inargs.historical_file: hist_cube_list[0].attributes['history'],\n inargs.control_file: control_cube_list[0].attributes['history']}\n gio.write_metadata(inargs.outfile, file_info=metadata_dict)\n\n\nif __name__ == '__main__':\n\n extra_info =\"\"\" \n\nauthor:\n Damien Irving, irving.damien@gmail.com\n \n\"\"\"\n\n description = 'Calculate the heat budget change between two time periods'\n parser = argparse.ArgumentParser(description=description,\n epilog=extra_info, \n argument_default=argparse.SUPPRESS,\n formatter_class=argparse.RawDescriptionHelpFormatter)\n \n parser.add_argument(\"historical_file\", type=str,\n help=\"historical experiment heat budget file from calc_heat_budget_timeseries.py\") \n parser.add_argument(\"control_file\", type=str,\n help=\"control experiment heat budget file from calc_heat_budget_timeseries.py\")\n parser.add_argument(\"outfile\", type=str, help=\"Output .csv file\") \n \n parser.add_argument(\"--start_time\", type=str, nargs=2, metavar=('START_DATE', 'END_DATE'),\n default=('1861-01-01', '1880-12-31'), help=\"Start time period\")\n parser.add_argument(\"--end_time\", type=str, nargs=2, metavar=('START_DATE', 'END_DATE'),\n default=('1986-01-01', '2005-12-31'), help=\"End time period\")\n\n args = parser.parse_args() \n main(args)\n", "sub_path": "data_processing/calc_heat_budget_change.py", "file_name": "calc_heat_budget_change.py", "file_ext": "py", "file_size_in_byte": 5905, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.getcwd", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 29, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "iris.analysis", "line_number": 82, "usage_type": "attribute"}, {"api_name": "iris.analysis", "line_number": 84, "usage_type": "attribute"}, {"api_name": "iris.load", "line_number": 94, "usage_type": "call"}, {"api_name": "iris.load", "line_number": 95, "usage_type": "call"}, {"api_name": "general_io.get_time_constraint", "line_number": 99, "usage_type": "call"}, {"api_name": "general_io.get_time_constraint", "line_number": 100, "usage_type": "call"}, {"api_name": "general_io.get_time_constraint", "line_number": 101, "usage_type": "call"}, {"api_name": "timeseries.get_control_time_constraint", "line_number": 103, "usage_type": "call"}, {"api_name": "timeseries.get_control_time_constraint", "line_number": 104, "usage_type": "call"}, {"api_name": "timeseries.get_control_time_constraint", "line_number": 105, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 111, "usage_type": "call"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 123, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 123, "usage_type": "attribute"}, {"api_name": "general_io.write_metadata", "line_number": 128, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 141, "usage_type": "call"}, {"api_name": "argparse.SUPPRESS", "line_number": 143, "usage_type": "attribute"}, {"api_name": "argparse.RawDescriptionHelpFormatter", "line_number": 144, "usage_type": "attribute"}]}
+{"seq_id": "642432368", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[1]:\n\n\n\nimport os\nfrom pycocotools.coco import COCO\nfrom torchvision import transforms\nimport torch\nimport numpy as np\n#import matplotlib.pyplot as plt\n#matplotlib inline\n\nfrom data_loader import get_loader\nfrom model import EncoderCNN, DecoderRNN\nfrom utils import clean_sentence, get_prediction\n\n\n# In[2]:\n\n\n# Define a transform to pre-process the testing images\ntransform_test = transforms.Compose([ \n transforms.Resize(256), # smaller edge of image resized to 256\n transforms.CenterCrop(224), # get 224x224 crop from the center\n transforms.ToTensor(), # convert the PIL Image to a tensor\n transforms.Normalize((0.485, 0.456, 0.406), # normalize image for pre-trained model\n (0.229, 0.224, 0.225))])\n\n# Create the data loader\ndata_loader = get_loader(transform=transform_test, \n mode='test')\n\n\n# In[3]:\n\n\n# Load the most recent checkpoint\ncheckpoint = torch.load(os.path.join('/home/osboxes/image_captioning/example', 'train-model-112000.pkl'))\n\n# Specify values for embed_size and hidden_size - we use the same values as in training step\nembed_size = 256\nhidden_size = 512\n\n# Get the vocabulary and its size\nvocab = data_loader.dataset.vocab\nvocab_size = len(vocab)\n\n# Initialize the encoder and decoder, and set each to inference mode\nencoder = EncoderCNN(embed_size)\nencoder.eval()\ndecoder = DecoderRNN(embed_size, hidden_size, vocab_size)\ndecoder.eval()\n\n# Load the pre-trained weights\nencoder.load_state_dict(checkpoint['encoder'])\ndecoder.load_state_dict(checkpoint['decoder'])\n\n# Move models to GPU if CUDA is available.\nif torch.cuda.is_available():\n encoder.cuda()\n decoder.cuda()\n\n\n# In[5]:\n\n\nx=get_prediction(data_loader, encoder, decoder, vocab)\n\n\n# In[6]:\n\n\nprint(x)\n\n\n\n\n\n\n\n", "sub_path": "predcition.py", "file_name": "predcition.py", "file_ext": "py", "file_size_in_byte": 1879, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "torchvision.transforms.Compose", "line_number": 25, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 25, "usage_type": "name"}, {"api_name": "torchvision.transforms.Resize", "line_number": 26, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 26, "usage_type": "name"}, {"api_name": "torchvision.transforms.CenterCrop", "line_number": 27, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 27, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 28, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 28, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 29, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 29, "usage_type": "name"}, {"api_name": "data_loader.get_loader", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "data_loader.dataset", "line_number": 48, "usage_type": "attribute"}, {"api_name": "model.EncoderCNN", "line_number": 52, "usage_type": "call"}, {"api_name": "model.DecoderRNN", "line_number": 54, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 62, "usage_type": "attribute"}, {"api_name": "utils.get_prediction", "line_number": 70, "usage_type": "call"}]}
+{"seq_id": "175776124", "text": "import json\nimport requests\nimport os\nfrom dotenv import load_dotenv\nfrom pathlib import Path\n\n\ndef load_env():\n env_path=Path('.')/'.env'\n load_dotenv(dotenv_path=env_path)\n print(API_ACCESS_KEY)\n \n\ndef get_help():\n with open('currency_code.txt','r') as file:\n print(file.read())\n exit()\n\n\ndef start():\n print(\" if you want to get some help about currency code type help \")\n from_currency=input(\" enter currency code that you want to convert from \")\n if from_currency.lower()=='help':\n get_help()\n to_currency=input(\" enter currency code that you want to convert from \")\n\n if to_currency.lower()=='help':\n get_help()\n\n return from_currency.upper(),to_currency.upper()\n\n\ndef api_request(from_currency,to_currency):\n \n try:\n API_ACCESS_KEY=os.environ.get(\"API_ACCESS_KEY\")\n res=requests.get(f\"https://v6.exchangerate-api.com/v6/{API_ACCESS_KEY}/latest/{from_currency}\")\n if res.status_code!=200:\n raise Exception(\"invalid-key\")\n json_obj=json.loads(res.text)\n if(json_obj['result']!='success'):\n raise Exception(json_obj['error-type'])\n except Exception as exc:\n print(exc)\n exit()\n else:\n try:\n result=json_obj['conversion_rates'][to_currency]\n time_last_update=json_obj['time_last_update_utc']\n except Exception:\n print(\"unsupported-code\")\n exit()\n\n return result,time_last_update\n \n\ndef print_result(result,time_last_update,from_currency,to_currency):\n print(f'{from_currency} is equal to {result} {to_currency}')\n print(f'time since last update: {time_last_update}')\n\n\nif __name__==\"__main__\":\n load_dotenv()\n from_currency,to_currency=start()\n result,time_last_update=api_request(from_currency,to_currency)\n print_result(result,time_last_update,from_currency,to_currency)\n\n", "sub_path": "currency_converter.py", "file_name": "currency_converter.py", "file_ext": "py", "file_size_in_byte": 1915, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pathlib.Path", "line_number": 9, "usage_type": "call"}, {"api_name": "dotenv.load_dotenv", "line_number": 10, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 36, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 36, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 37, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 40, "usage_type": "call"}, {"api_name": "dotenv.load_dotenv", "line_number": 63, "usage_type": "call"}]}
+{"seq_id": "435568490", "text": "import seaborn as sns\nimport matplotlib.pyplot as plt\n\niris = sns.load_dataset('iris')\nprint(iris.head())\n\n# Pair Plot automatically gives you a grid of multiple plots\n# It is basically all scatter plots except the diagonal, which is a bar plot.\nsns.pairplot(iris)\nplt.show()\n\n# Pair Grid allows you to customise the pair plot grid.\n# You can use different plot functions to the upper, lower and diagonal sections.\ngrid = sns.PairGrid(iris)\ngrid.map_diag(sns.distplot) # This will use distributed plots on the diagonal\ngrid.map_upper(sns.stripplot) # This will use strip plots above the diagonal line.\ngrid.map_lower(sns.kdeplot) # This will use kde plots below the diagonal line.\nplt.show()\n\ntips = sns.load_dataset('tips')\n\n# Facet Grid allows you to create multiple graphs based on rows and columns of variables.\n# It also allows you to customize the type of plots that you use.\ngrid = sns.FacetGrid(tips, col='time', row='smoker')\ngrid.map(sns.distplot, 'total_bill')\nplt.show()\n\n# When you create a FacetGrid with a plot that requires multiple columns, such as a scatter\n# You just need to pass the second column name, in this case 'tips'.\ngrid = sns.FacetGrid(tips, col='time', row='smoker')\ngrid.map(plt.scatter, 'total_bill', 'tip')\nplt.show()\n", "sub_path": "09-Seaborn/51-Grids.py", "file_name": "51-Grids.py", "file_ext": "py", "file_size_in_byte": 1259, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "seaborn.load_dataset", "line_number": 4, "usage_type": "call"}, {"api_name": "seaborn.pairplot", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 10, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 10, "usage_type": "name"}, {"api_name": "seaborn.PairGrid", "line_number": 14, "usage_type": "call"}, {"api_name": "seaborn.distplot", "line_number": 15, "usage_type": "attribute"}, {"api_name": "seaborn.stripplot", "line_number": 16, "usage_type": "attribute"}, {"api_name": "seaborn.kdeplot", "line_number": 17, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "seaborn.load_dataset", "line_number": 20, "usage_type": "call"}, {"api_name": "seaborn.FacetGrid", "line_number": 24, "usage_type": "call"}, {"api_name": "seaborn.distplot", "line_number": 25, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "seaborn.FacetGrid", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 31, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}]}
+{"seq_id": "239690151", "text": "#!/usr/bin/python\n# coding:utf8\n\"\"\"\n@author: Cong Yu\n@time: 2019-08-23 16:06\n\"\"\"\nimport unicodedata\nimport six\nimport sentencepiece as spm\nimport tensorflow as tf\nimport collections\nimport pandas as pd\nfrom sklearn.utils import shuffle\n\nSEG_ID_A = 0\nSEG_ID_B = 1\nSEG_ID_CLS = 2\nSEG_ID_SEP = 3\nSEG_ID_PAD = 4\n\nspecial_symbols = {\n \"\": 0,\n \"\": 1,\n \"\": 2,\n \"\": 3,\n \"\": 4,\n \"\": 5,\n \"\": 6,\n \"\": 7,\n \"\": 8,\n}\n\nVOCAB_SIZE = 32000\nUNK_ID = special_symbols[\"\"]\nCLS_ID = special_symbols[\"\"]\nSEP_ID = special_symbols[\"\"]\nMASK_ID = special_symbols[\"\"]\nEOD_ID = special_symbols[\"\"]\n\n\ndef preprocess_text(inputs, lower=False, remove_space=True, keep_accents=False):\n if remove_space:\n outputs = ' '.join(inputs.strip().split())\n else:\n outputs = inputs\n outputs = outputs.replace(\"``\", '\"').replace(\"''\", '\"')\n\n if six.PY2 and isinstance(outputs, str):\n outputs = outputs.decode('utf-8')\n\n if not keep_accents:\n outputs = unicodedata.normalize('NFKD', outputs)\n outputs = ''.join([c for c in outputs if not unicodedata.combining(c)])\n if lower:\n outputs = outputs.lower()\n\n return outputs\n\n\ndef encode_pieces(sp_model, text, return_unicode=True, sample=False):\n # return_unicode is used only for py2\n # note(zhiliny): in some systems, sentencepiece only accepts str for py2\n if six.PY2 and isinstance(text, unicode):\n text = text.encode('utf-8')\n if not sample:\n pieces = sp_model.EncodeAsPieces(text)\n else:\n pieces = sp_model.SampleEncodeAsPieces(text, 64, 0.1)\n new_pieces = []\n for piece in pieces:\n if len(piece) > 1 and piece[-1] == ',' and piece[-2].isdigit():\n cur_pieces = sp_model.EncodeAsPieces(\n piece[:-1].replace(SPIECE_UNDERLINE, ''))\n if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE:\n if len(cur_pieces[0]) == 1:\n cur_pieces = cur_pieces[1:]\n else:\n cur_pieces[0] = cur_pieces[0][1:]\n cur_pieces.append(piece[-1])\n new_pieces.extend(cur_pieces)\n else:\n new_pieces.append(piece)\n\n # note(zhiliny): convert back to unicode for py2\n if six.PY2 and return_unicode:\n ret_pieces = []\n for piece in new_pieces:\n if isinstance(piece, str):\n piece = piece.decode('utf-8')\n ret_pieces.append(piece)\n new_pieces = ret_pieces\n\n return new_pieces\n\n\ndef encode_ids(sp_model, text, sample=False):\n pieces = encode_pieces(sp_model, text, return_unicode=False, sample=sample)\n ids = [sp_model.PieceToId(piece) for piece in pieces]\n return ids\n\n\ndef _truncate_seq_pair(tokens_a, tokens_b, max_length):\n \"\"\"Truncates a sequence pair in place to the maximum length.\"\"\"\n\n # This is a simple heuristic which will always truncate the longer sequence\n # one token at a time. This makes more sense than truncating an equal percent\n # of tokens from each, since if one sequence is very short then each token\n # that's truncated likely contains more information than a longer sequence.\n while True:\n total_length = len(tokens_a) + len(tokens_b)\n if total_length <= max_length:\n break\n if len(tokens_a) > len(tokens_b):\n tokens_a.pop()\n else:\n tokens_b.pop()\n\n\ndef convert_single_example(text_a, text_b, max_seq_length, tokenize_fn):\n \"\"\"Converts a single `InputExample` into a single `InputFeatures`.\"\"\"\n tokens_a = tokenize_fn(text_a)\n tokens_b = None\n if text_b:\n tokens_b = tokenize_fn(text_b)\n if tokens_b:\n # Modifies `tokens_a` and `tokens_b` in place so that the total\n # length is less than the specified length.\n # Account for two [SEP] & one [CLS] with \"- 3\"\n _truncate_seq_pair(tokens_a, tokens_b, max_seq_length - 3)\n else:\n # Account for one [SEP] & one [CLS] with \"- 2\"\n if len(tokens_a) > max_seq_length - 2:\n tokens_a = tokens_a[:max_seq_length - 2]\n\n tokens = []\n segment_ids = []\n for token in tokens_a:\n tokens.append(token)\n segment_ids.append(SEG_ID_A)\n tokens.append(SEP_ID)\n segment_ids.append(SEG_ID_A)\n\n if tokens_b:\n for token in tokens_b:\n tokens.append(token)\n segment_ids.append(SEG_ID_B)\n tokens.append(SEP_ID)\n segment_ids.append(SEG_ID_B)\n\n tokens.append(CLS_ID)\n segment_ids.append(SEG_ID_CLS)\n\n input_ids = tokens\n\n # The mask has 0 for real tokens and 1 for padding tokens. Only real\n # tokens are attended to.\n input_mask = [0] * len(input_ids)\n\n # Zero-pad up to the sequence length.\n if len(input_ids) < max_seq_length:\n delta_len = max_seq_length - len(input_ids)\n input_ids = [0] * delta_len + input_ids\n input_mask = [1] * delta_len + input_mask\n segment_ids = [SEG_ID_PAD] * delta_len + segment_ids\n\n assert len(input_ids) == max_seq_length\n assert len(input_mask) == max_seq_length\n assert len(segment_ids) == max_seq_length\n return (input_ids, input_mask, segment_ids)\n\n\ndef file_based_convert_examples_to_features(path, label2id, max_seq_length, tokenize_fn, output_file):\n \"\"\"Convert a set of `InputExample`s to a TFRecord file.\"\"\"\n tf.logging.info(\"Create new tfrecord {}.\".format(output_file))\n writer = tf.python_io.TFRecordWriter(output_file)\n df = pd.read_csv(path, index_col=0)\n df = shuffle(df)\n count = 0\n for index, row in df.iterrows():\n # label = label2id[row[\"topic\"].strip()]\n feature = convert_single_example(row[config[\"column_name_x1\"]],\n row[config[\"column_name_x2\"]] if config[\"column_name_x2\"] != \"\" else None,\n max_seq_length, tokenize_fn)\n\n def create_int_feature(values):\n f = tf.train.Feature(int64_list=tf.train.Int64List(value=list(values)))\n return f\n\n def create_float_feature(values):\n f = tf.train.Feature(float_list=tf.train.FloatList(value=list(values)))\n return f\n\n label = label2id.get(str(row[config[\"column_name_y\"]]))\n features = collections.OrderedDict()\n features[\"input_ids\"] = create_int_feature(feature[0])\n features[\"input_mask\"] = create_float_feature(feature[1])\n features[\"segment_ids\"] = create_int_feature(feature[2])\n features[\"label_ids\"] = create_int_feature([label])\n count += 1\n if count < 5:\n print(\"*** Example ***\")\n print(\"input_ids: %s\" % \" \".join([str(x) for x in feature[0]]))\n print(\"input_mask: %s\" % \" \".join([str(x) for x in feature[1]]))\n print(\"segment_ids: %s\" % \" \".join([str(x) for x in feature[2]]))\n\n print(\"label: %s (id = %s)\" % (row[config[\"column_name_y\"]], str(label)))\n\n tf_example = tf.train.Example(features=tf.train.Features(feature=features))\n writer.write(tf_example.SerializeToString())\n if count % 1000 == 0:\n print(count)\n writer.close()\n print(\"example count:\", count)\n\n\nlabel2id = {'劳动纠纷': 0, '婚姻家庭': 1, '公司法': 2, '交通事故': 3, '合同纠纷': 4, '刑事辩护': 5, '房产纠纷': 6, '债权债务': 7}\nconfig = {\n \"spiece_model_file\": \"./chinese_xlnet_mid_L-24_H-768_A-12/spiece.model\",\n \"csv_file\": \"./data/dev.csv\",\n \"tf_record_file\": \"./data/dev.tf_record\",\n \"column_name_x1\": \"question\",\n \"column_name_x2\": \"\",\n \"column_name_y\": \"label\",\n \"max_seq_len\": 128,\n}\n\nSPIECE_UNDERLINE = '▁'\nsp = spm.SentencePieceProcessor()\nsp.Load(config[\"spiece_model_file\"])\n\n\ndef tokenize_fn(text):\n text = preprocess_text(text, lower=False)\n return encode_ids(sp, text)\n\n\nfile_based_convert_examples_to_features(config[\"csv_file\"], label2id, config[\"max_seq_len\"], tokenize_fn,\n config[\"tf_record_file\"])\n", "sub_path": "component_xlnet_data_processor.py", "file_name": "component_xlnet_data_processor.py", "file_ext": "py", "file_size_in_byte": 8068, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "six.PY2", "line_number": 48, "usage_type": "attribute"}, {"api_name": "unicodedata.normalize", "line_number": 52, "usage_type": "call"}, {"api_name": "unicodedata.combining", "line_number": 53, "usage_type": "call"}, {"api_name": "six.PY2", "line_number": 63, "usage_type": "attribute"}, {"api_name": "six.PY2", "line_number": 85, "usage_type": "attribute"}, {"api_name": "tensorflow.logging.info", "line_number": 174, "usage_type": "call"}, {"api_name": "tensorflow.logging", "line_number": 174, "usage_type": "attribute"}, {"api_name": "tensorflow.python_io.TFRecordWriter", "line_number": 175, "usage_type": "call"}, {"api_name": "tensorflow.python_io", "line_number": 175, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 176, "usage_type": "call"}, {"api_name": "sklearn.utils.shuffle", "line_number": 177, "usage_type": "call"}, {"api_name": "tensorflow.train.Feature", "line_number": 186, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 186, "usage_type": "attribute"}, {"api_name": "tensorflow.train.Int64List", "line_number": 186, "usage_type": "call"}, {"api_name": "tensorflow.train.Feature", "line_number": 190, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 190, "usage_type": "attribute"}, {"api_name": "tensorflow.train.FloatList", "line_number": 190, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 194, "usage_type": "call"}, {"api_name": "tensorflow.train.Example", "line_number": 208, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 208, "usage_type": "attribute"}, {"api_name": "tensorflow.train.Features", "line_number": 208, "usage_type": "call"}, {"api_name": "sentencepiece.SentencePieceProcessor", "line_number": 228, "usage_type": "call"}]}
+{"seq_id": "291742086", "text": "from tornado.ioloop import IOLoop\nfrom tornado.web import Application, url\n\nfrom handlers import (\n MainHandler,\n LoginHandler,\n LogoutHandler,\n AddAddressHandler,\n DeleteAddressHandler,\n ViewAddressHandler\n)\n\n\ndef make_app():\n return Application([\n url(r\"/\", MainHandler),\n url(r\"/login\", LoginHandler),\n url(r\"/logout\", LogoutHandler),\n url(r\"/add\", AddAddressHandler),\n url(r\"/del/(\\d)\", DeleteAddressHandler),\n url(r\"/all\", ViewAddressHandler),\n ], login_url=\"/login\",\n debug=True,\n compiled_template_cache=False,\n cookie_secret=\"__TODO:_GENERATE_YOUR_OWN_RANDOM_VALUE_HERE__\")\n\n\ndef main():\n app = make_app()\n app.listen(8888)\n IOLoop.current().start()\n\n\nif __name__ == \"__main__\":\n main()\n", "sub_path": "examples/addressbook/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 795, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "tornado.web.Application", "line_number": 15, "usage_type": "call"}, {"api_name": "tornado.web.url", "line_number": 16, "usage_type": "call"}, {"api_name": "handlers.MainHandler", "line_number": 16, "usage_type": "argument"}, {"api_name": "tornado.web.url", "line_number": 17, "usage_type": "call"}, {"api_name": "handlers.LoginHandler", "line_number": 17, "usage_type": "argument"}, {"api_name": "tornado.web.url", "line_number": 18, "usage_type": "call"}, {"api_name": "handlers.LogoutHandler", "line_number": 18, "usage_type": "argument"}, {"api_name": "tornado.web.url", "line_number": 19, "usage_type": "call"}, {"api_name": "handlers.AddAddressHandler", "line_number": 19, "usage_type": "argument"}, {"api_name": "tornado.web.url", "line_number": 20, "usage_type": "call"}, {"api_name": "handlers.DeleteAddressHandler", "line_number": 20, "usage_type": "argument"}, {"api_name": "tornado.web.url", "line_number": 21, "usage_type": "call"}, {"api_name": "handlers.ViewAddressHandler", "line_number": 21, "usage_type": "argument"}, {"api_name": "tornado.ioloop.IOLoop.current", "line_number": 31, "usage_type": "call"}, {"api_name": "tornado.ioloop.IOLoop", "line_number": 31, "usage_type": "name"}]}
+{"seq_id": "252289725", "text": "import numpy as np\nimport cv2\nimport video\nimport time\n \ncap = cv2.VideoCapture(0)\nret, background = cap.read()\ntime.sleep(1)\nret, background = cap.read()\nbackground = cv2.cvtColor(background, cv2.COLOR_BGR2GRAY)\n\nwhile(True):\n # Capture frame-by-frame\n ret,frame = cap.read()\n\n # Our operations on the frame come here\n gray = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)\n\n diff = cv2.absdiff(background,gray)\n\n ret,thresh = cv2.threshold(diff, 80, 255, 0)\n\n im2, contours, hierarchy = cv2.findContours(thresh,cv2.RETR_TREE,cv2.CHAIN_APPROX_SIMPLE)\n\n if len(contours) > 1:\n\n areaArray = []\n \n for i, c in enumerate(contours):\n area = cv2.contourArea(c)\n areaArray.append(area)\n\n #first sort the array by area\n sorteddata = sorted(zip(areaArray, contours), key=lambda x: x[0], reverse=True)\n\n\n #find the nth largest contour [n-1][1], in this case 2\n firstcontour = sorteddata[0][1]\n #secondcontour = sorteddata[1][1]\n\n #draw it\n x, y, w, h = cv2.boundingRect(firstcontour)\n cv2.drawContours(thresh, firstcontour, -1, (255, 255, 0), 2)\n cv2.rectangle(thresh, (x, y), (x+w, y+h), (127,255,0), 2)\n \n #cv2.drawContours(thresh, contours, len(contours)-1, (128,255,0), 3)\n # Display the resulting frame\n cv2.imshow('frame',thresh)\n if cv2.waitKey(1) & 0xFF == ord('q'):\n break\n\n# When everything done, release the capture\ncap.release()\ncv2.destroyAllWindows()\n", "sub_path": "python/motion_detect/bg_sub.py", "file_name": "bg_sub.py", "file_ext": "py", "file_size_in_byte": 1500, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "cv2.VideoCapture", "line_number": 6, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 8, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 10, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 17, "usage_type": "attribute"}, {"api_name": "cv2.absdiff", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.threshold", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.findContours", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.RETR_TREE", "line_number": 23, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_SIMPLE", "line_number": 23, "usage_type": "attribute"}, {"api_name": "cv2.contourArea", "line_number": 30, "usage_type": "call"}, {"api_name": "cv2.boundingRect", "line_number": 42, "usage_type": "call"}, {"api_name": "cv2.drawContours", "line_number": 43, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 44, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 48, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 49, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 54, "usage_type": "call"}]}
+{"seq_id": "317214798", "text": "import numpy as np\nimport pandas as pd\nimport os, sys\n\nos.environ['KERAS_BACKEND']='tensorflow'\nos.environ['CUDA_VISIBLE_DEVICES']=sys.argv[1]\n\nfrom keras.utils.np_utils import to_categorical\nfrom keras.layers import Embedding, Dropout, Reshape, GRU, Bidirectional, TimeDistributed, Dot, Activation, Dense, Input\nfrom keras.models import Model\n\nimport preprocessor\nimport nltk\nimport itertools\n\n# Hyper parameters\nMAX_SENT_LENGTH = 300\nMAX_NUM_SENTS = 20\nVOCABULARY_SIZE = 50000\nEMBEDDING_DIM = 200\nVALIDATION_RATIO = 0.2\nBATCH_SIZE = 64\nNUM_EPOCHS = 20\n\n\n# Data pre-processing\nword_to_index = {}\ndef build_vocab(texts):\n global word_to_index\n\n # Tokenize the data into sentences\n sentences = itertools.chain(*[text.split('') for text in texts])\n # Tokenize the sentences into words\n tokenized_sentences = [nltk.word_tokenize(sent) for sent in sentences]\n\n # Count the word frequencies\n word_freq = nltk.FreqDist(itertools.chain(*tokenized_sentences))\n print(\"Found %d unique words tokens.\" % len(word_freq.items()))\n # Get the most common words and build index_to_word and word_to_index vectors\n vocab = word_freq.most_common(VOCABULARY_SIZE)\n # word index starts from 1, 0-index is reserved for padding\n word_to_index = dict([(w[0], i+1) for i, w in enumerate(vocab)])\n\n print(\"Using vocabulary size %d.\" % VOCABULARY_SIZE)\n print(\"The least frequent word in our vocabulary is '%s' and appeared %d times.\" % (vocab[-1][0], vocab[-1][1]))\n\n\ndef texts_to_tensor(texts):\n global word_to_index\n documents = np.zeros((len(texts), MAX_NUM_SENTS, MAX_SENT_LENGTH), dtype='int32')\n reviews = [text.split('') for text in texts]\n for i, review in enumerate(reviews):\n for j, sent in enumerate(review):\n if (j >= MAX_NUM_SENTS): continue\n review[j] = [w for w in sent if w in word_to_index]\n for k in range(MAX_SENT_LENGTH):\n if k < len(review[j]):\n documents[i,j,k] = word_to_index[review[j][k]]\n return documents\n\n\n# Load data (Yelp)\nlabel_map = {1:0, 2:1, 3:2, 4:3, 5:4}\n\ndata_train = pd.read_csv('data/emnlp-2015/yelp-2013-train.txt.ss', sep='\\t', header=None, usecols=[4, 6])\nprint(data_train.shape)\nbuild_vocab(data_train[6])\nx_train = texts_to_tensor(data_train[6])\ny_train = to_categorical(data_train[4].map(label_map))\n\ndata_val = pd.read_csv('data/emnlp-2015/yelp-2013-test.txt.ss', sep='\\t', header=None, usecols=[4, 6])\nprint(data_val.shape)\nx_val = texts_to_tensor(data_val[6])\ny_val = to_categorical(data_val[4].map(label_map))\n\nprint('Number of reviews per class in training and validation set ')\nprint(y_train.sum(axis=0))\nprint(y_val.sum(axis=0))\n\nembedding_weights = preprocessor.load_embedding(EMBEDDING_DIM)\nembedding_matrix = np.zeros((VOCABULARY_SIZE + 1, EMBEDDING_DIM))\nfor word, i in word_to_index.items():\n embedding_vector = embedding_weights.get(word)\n if embedding_vector is not None:\n embedding_matrix[i] = embedding_vector\n\nembedding_layer = Embedding(VOCABULARY_SIZE + 1,\n EMBEDDING_DIM,\n weights=[embedding_matrix],\n input_length=MAX_SENT_LENGTH,\n trainable=True)\n\n\n# Hyper parameters\nhidden_dim = 50\ndropout = 0.5\n\n# Word level\nsentence_input = Input(shape=(MAX_SENT_LENGTH,), dtype='int32')\nembedded_sequences = (embedding_layer(sentence_input))\nl_dropout1 = Dropout(dropout)(embedded_sequences)\n\nh_word = Bidirectional(GRU(hidden_dim, return_sequences=True), name='h_word')(l_dropout1)\nu_word = TimeDistributed(Dense(2 * hidden_dim, activation='tanh'), name='u_word')(h_word)\n\nalpha_word = TimeDistributed(Dense(1, use_bias=False))(u_word)\nalpha_word = Reshape((MAX_SENT_LENGTH,))(alpha_word)\nalpha_word = Activation('softmax')(alpha_word)\n\nh_word_combined = Dot(axes=[1, 1], name='h_word_combined')([h_word, alpha_word])\n\nsent_encoder = Model(sentence_input, h_word_combined)\nsent_encoder.summary()\n\n# Sentence level\nreview_input = Input(shape=(MAX_NUM_SENTS, MAX_SENT_LENGTH), dtype='int32')\nreview_encoder = TimeDistributed(sent_encoder, name='sent_encoder')(review_input)\nl_dropout2 = Dropout(dropout)(review_encoder)\n\nh_sent = Bidirectional(GRU(hidden_dim, return_sequences=True), name='h_sent')(l_dropout2)\nu_sent = TimeDistributed(Dense(2 * hidden_dim, activation='tanh'), name='u_sent')(h_sent)\n\nalpha_sent = TimeDistributed(Dense(1, use_bias=False))(u_sent)\nalpha_sent = Reshape((MAX_NUM_SENTS,))(alpha_sent)\nalpha_sent = Activation('softmax')(alpha_sent)\n\nh_sent_combined = Dot(axes=[1, 1], name='h_sent_combined')([h_sent, alpha_sent])\n\n# Classifier layer\nl_classifier = Dense(5, activation='softmax')(h_sent_combined)\n\n# Build model\nmodel = Model(review_input, l_classifier)\nmodel.compile(loss='categorical_crossentropy',\n optimizer='adam',\n metrics=['accuracy'])\n\n# Train model\nmodel.summary()\nmodel.fit(x_train, y_train, validation_data=(x_val, y_val), epochs=NUM_EPOCHS, batch_size=BATCH_SIZE)", "sub_path": "hatt_classifier.py", "file_name": "hatt_classifier.py", "file_ext": "py", "file_size_in_byte": 5027, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.environ", "line_number": 5, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 6, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 6, "usage_type": "attribute"}, {"api_name": "itertools.chain", "line_number": 32, "usage_type": "call"}, {"api_name": "nltk.word_tokenize", "line_number": 34, "usage_type": "call"}, {"api_name": "nltk.FreqDist", "line_number": 37, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 50, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 65, "usage_type": "call"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 69, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 71, "usage_type": "call"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 74, "usage_type": "call"}, {"api_name": "preprocessor.load_embedding", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 81, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 87, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 99, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 101, "usage_type": "call"}, {"api_name": "keras.layers.Bidirectional", "line_number": 103, "usage_type": "call"}, {"api_name": "keras.layers.GRU", "line_number": 103, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 104, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 104, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 106, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 106, "usage_type": "call"}, {"api_name": "keras.layers.Reshape", "line_number": 107, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 108, "usage_type": "call"}, {"api_name": "keras.layers.Dot", "line_number": 110, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 112, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 116, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 117, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 118, "usage_type": "call"}, {"api_name": "keras.layers.Bidirectional", "line_number": 120, "usage_type": "call"}, {"api_name": "keras.layers.GRU", "line_number": 120, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 121, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 121, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 123, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 123, "usage_type": "call"}, {"api_name": "keras.layers.Reshape", "line_number": 124, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 125, "usage_type": "call"}, {"api_name": "keras.layers.Dot", "line_number": 127, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 130, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 133, "usage_type": "call"}]}
+{"seq_id": "368658586", "text": "from envs.rover_lander_1 import rover_lander_1\nfrom envs.rover_lander_2 import rover_lander_2\n\nimport argparse\nimport time\nimport tensorflow as tf\nimport numpy as np\nimport random\nimport os\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"--model\", help=\"Path to model to be used by the agent\")\nparser.add_argument(\"--fps\", help=\"Frames per second\", type=int, default=20)\nparser.add_argument(\"--env\", help=\"Env name\")\nparser.add_argument(\"--save-gif\", help=\"Save gif\", action='store_true', default=False)\n\nargs = parser.parse_args()\n\nmodel_path = args.model\nfps = args.fps\n\n\nclass Agent:\n def __init__(self, model_path=None):\n self.testing = (model_path == None)\n if not self.testing:\n self.model = self.load_model(model_path)\n \n def load_model(self, model_path):\n return tf.keras.models.load_model(model_path)\n \n def qs(self, state):\n if self.testing:\n return (random.randint(0, 4) - 1)\n state = state.reshape(1, np.prod(state.shape[:]))/255\n return np.argmax(self.model.predict(state))\n \n \nif __name__ == '__main__':\n agent = Agent(model_path)\n if args.env == 'rover_lander_1':\n env = rover_lander_1(save_gif=args.save_gif, filename=os.path.basename(model_path).replace(\".h5\", \"\"))\n elif args.env == 'rover_lander_2':\n env = rover_lander_2(save_gif=args.save_gif, filename=os.path.basename(model_path).replace(\".h5\", \"\"))\n \n for i in range(5):\n state = env.reset()\n while True:\n time.sleep(1/fps)\n env.render()\n action = agent.qs(state)\n state, reward, done = env.step(action)\n print(action, reward, done)\n if done:\n break\n env.export_gif()", "sub_path": "play.py", "file_name": "play.py", "file_ext": "py", "file_size_in_byte": 1766, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 11, "usage_type": "call"}, {"api_name": "tensorflow.keras.models.load_model", "line_number": 30, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 30, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.prod", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 36, "usage_type": "call"}, {"api_name": "envs.rover_lander_1.rover_lander_1", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path", "line_number": 42, "usage_type": "attribute"}, {"api_name": "envs.rover_lander_2.rover_lander_2", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 49, "usage_type": "call"}]}
+{"seq_id": "401112482", "text": "import logging\nimport os\nimport sys\nimport traceback\n\nfrom importlib.machinery import SourceFileLoader\nfrom jinja2 import Environment, select_autoescape, FileSystemLoader, StrictUndefined\n\nfrom sceptre.exceptions import UnsupportedTemplateFileTypeError, TemplateSceptreHandlerError\nfrom sceptre.template_handlers import TemplateHandler\n\n\nclass File(TemplateHandler):\n \"\"\"\n Template handler that can load files from disk. Supports JSON, YAML, Jinja2 and Python.\n \"\"\"\n\n def __init__(self, *args, **kwargs):\n self.logger = logging.getLogger(__name__)\n super(File, self).__init__(*args, **kwargs)\n\n def schema(self):\n return {\n \"type\": \"object\",\n \"properties\": {\n \"path\": {\"type\": \"string\"},\n },\n \"required\": [\"path\"]\n }\n\n def handle(self):\n file_extension = os.path.splitext(self.arguments[\"path\"])[1]\n path = self.arguments[\"path\"]\n try:\n if file_extension in {\".json\", \".yaml\", \".template\"}:\n with open(path) as template_file:\n return template_file.read()\n elif file_extension == \".j2\":\n return self._render_jinja_template(\n os.path.dirname(path),\n os.path.basename(path),\n {\"sceptre_user_data\": self.sceptre_user_data}\n )\n elif file_extension == \".py\":\n return self._call_sceptre_handler()\n else:\n raise UnsupportedTemplateFileTypeError(\n \"Template has file extension %s. Only .py, .yaml, \"\n \".template, .json and .j2 are supported.\",\n os.path.splitext(path)[1]\n )\n except Exception as e:\n self._print_template_traceback()\n raise e\n\n def _call_sceptre_handler(self):\n \"\"\"\n Calls the function `sceptre_handler` within templates that are python\n scripts.\n\n :returns: The string returned from sceptre_handler in the template.\n :rtype: str\n :raises: IOError\n :raises: TemplateSceptreHandlerError\n \"\"\"\n # Get relative path as list between current working directory and where\n # the template is\n # NB: this is a horrible hack...\n path = self.arguments[\"path\"]\n\n relpath = os.path.relpath(path, os.getcwd()).split(os.path.sep)\n relpaths_to_add = [\n os.path.sep.join(relpath[:i + 1])\n for i in range(len(relpath[:-1]))\n ]\n # Add any directory between the current working directory and where\n # the template is to the python path\n for directory in relpaths_to_add:\n sys.path.append(os.path.join(os.getcwd(), directory))\n self.logger.debug(\n \"%s - Getting CloudFormation from %s\", self.name, path\n )\n\n if not os.path.isfile(path):\n raise IOError(\"No such file or directory: '%s'\", path)\n\n module = SourceFileLoader(path, path).load_module()\n\n try:\n body = module.sceptre_handler(self.sceptre_user_data)\n except AttributeError as e:\n if 'sceptre_handler' in str(e):\n raise TemplateSceptreHandlerError(\n \"The template does not have the required \"\n \"'sceptre_handler(sceptre_user_data)' function.\"\n )\n else:\n raise e\n for directory in relpaths_to_add:\n sys.path.remove(os.path.join(os.getcwd(), directory))\n return body\n\n def _print_template_traceback(self):\n \"\"\"\n Prints a stack trace, including only files which are inside a\n 'templates' directory. The function is intended to give the operator\n instant feedback about why their templates are failing to compile.\n\n :rtype: None\n \"\"\"\n\n def _print_frame(filename, line, fcn, line_text):\n self.logger.error(\"{}:{}: Template error in '{}'\\n=> `{}`\".format(\n filename, line, fcn, line_text))\n\n try:\n _, _, tb = sys.exc_info()\n stack_trace = traceback.extract_tb(tb)\n search_string = os.path.join('', 'templates', '')\n if search_string in self.arguments[\"path\"]:\n template_path = self.arguments[\"path\"].split(search_string)[0] + search_string\n else:\n return\n for frame in stack_trace:\n if isinstance(frame, tuple):\n # Python 2 / Old style stack frame\n if template_path in frame[0]:\n _print_frame(frame[0], frame[1], frame[2], frame[3])\n else:\n if template_path in frame.filename:\n _print_frame(frame.filename, frame.lineno, frame.name, frame.line)\n except Exception as tb_exception:\n self.logger.error(\n 'A template error occured. ' +\n 'Additionally, a traceback exception occured. Exception: %s',\n tb_exception\n )\n\n @staticmethod\n def _render_jinja_template(template_dir, filename, jinja_vars):\n \"\"\"\n Renders a jinja template.\n\n Sceptre supports passing sceptre_user_data to JSON and YAML\n CloudFormation templates using Jinja2 templating.\n\n :param template_dir: The directory containing the template.\n :type template_dir: str\n :param filename: The name of the template file.\n :type filename: str\n :param jinja_vars: Dict of variables to render into the template.\n :type jinja_vars: dict\n :returns: The body of the CloudFormation template.\n :rtype: str\n \"\"\"\n logger = logging.getLogger(__name__)\n logger.debug(\"%s Rendering CloudFormation template\", filename)\n env = Environment(\n autoescape=select_autoescape(\n disabled_extensions=('j2',),\n default=True,\n ),\n loader=FileSystemLoader(template_dir),\n undefined=StrictUndefined\n )\n template = env.get_template(filename)\n body = template.render(**jinja_vars)\n return body\n", "sub_path": "sceptre/template_handlers/file.py", "file_name": "file.py", "file_ext": "py", "file_size_in_byte": 6252, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sceptre.template_handlers.TemplateHandler", "line_number": 13, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "sceptre.exceptions.UnsupportedTemplateFileTypeError", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.relpath", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path", "line_number": 71, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path.sep.join", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 79, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 79, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path", "line_number": 79, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "importlib.machinery.SourceFileLoader", "line_number": 87, "usage_type": "call"}, {"api_name": "sceptre.exceptions.TemplateSceptreHandlerError", "line_number": 93, "usage_type": "call"}, {"api_name": "sys.path.remove", "line_number": 100, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 100, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path", "line_number": 100, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 100, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 117, "usage_type": "call"}, {"api_name": "traceback.extract_tb", "line_number": 118, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 119, "usage_type": "call"}, {"api_name": "os.path", "line_number": 119, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 156, "usage_type": "call"}, {"api_name": "jinja2.Environment", "line_number": 158, "usage_type": "call"}, {"api_name": "jinja2.select_autoescape", "line_number": 159, "usage_type": "call"}, {"api_name": "jinja2.FileSystemLoader", "line_number": 163, "usage_type": "call"}, {"api_name": "jinja2.StrictUndefined", "line_number": 164, "usage_type": "name"}]}
+{"seq_id": "321372299", "text": "import matplotlib.pyplot as plt\nimport seaborn as sns\nimport click\nimport os\n\nfrom src.data_preparation.anomaly_data_generator import DataGenerator\nfrom src.data_reading.anomaly_data_reader import AnomalyDataReader\n\n\n@click.command()\n@click.option('--data_path', type=click.Path(exists=True), required=True)\n@click.option('--output_dir', type=click.Path(), required=True)\n@click.option('--data_type', type=str, required=True)\ndef main(data_path, output_dir, data_type):\n channels = DataGenerator.wanted_electrodes['EEG'] + DataGenerator.wanted_electrodes['EKG']\n\n train_info_dicts = AnomalyDataReader.load_info_dicts(data_path, data_type)\n\n stdv_list = []\n mean_list = []\n for info_dict in train_info_dicts:\n stdv_list.append(info_dict['std'])\n mean_list.append(info_dict['mean'])\n\n os.makedirs(output_dir, exist_ok=True)\n for channel_i, channel in enumerate(channels):\n stdv = [float(s[channel_i]) for s in stdv_list]\n mean = [float(m[channel_i]) for m in mean_list]\n\n plt.title(channel)\n sns.distplot(stdv, rug=True)\n plt.savefig(os.path.join(output_dir, channel + ' std'))\n plt.clf()\n\n plt.title(channel)\n sns.distplot(mean, rug=True)\n plt.savefig(os.path.join(output_dir, channel + ' mean'))\n plt.clf()\n\n\nif __name__ == '__main__':\n main()\n", "sub_path": "tools/channel_scale_histogram.py", "file_name": "channel_scale_histogram.py", "file_ext": "py", "file_size_in_byte": 1352, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "src.data_preparation.anomaly_data_generator.DataGenerator.wanted_electrodes", "line_number": 15, "usage_type": "attribute"}, {"api_name": "src.data_preparation.anomaly_data_generator.DataGenerator", "line_number": 15, "usage_type": "name"}, {"api_name": "src.data_reading.anomaly_data_reader.AnomalyDataReader.load_info_dicts", "line_number": 17, "usage_type": "call"}, {"api_name": "src.data_reading.anomaly_data_reader.AnomalyDataReader", "line_number": 17, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "seaborn.distplot", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "seaborn.distplot", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "click.command", "line_number": 10, "usage_type": "call"}, {"api_name": "click.option", "line_number": 11, "usage_type": "call"}, {"api_name": "click.Path", "line_number": 11, "usage_type": "call"}, {"api_name": "click.option", "line_number": 12, "usage_type": "call"}, {"api_name": "click.Path", "line_number": 12, "usage_type": "call"}, {"api_name": "click.option", "line_number": 13, "usage_type": "call"}]}
+{"seq_id": "484154042", "text": "import abc\nfrom enum import Enum\nfrom collections import namedtuple\n\nPrimitiveType = Enum('PrimitiveType', 'NUMBR NUMBAR LETTR TROOF')\nArrayType = namedtuple('ArrayType', ['subtype'])\nclass CompilerTypeError(Exception): pass\nclass ArgumentError(Exception): pass\nclass NestedError(Exception): pass\n\nclass ASTNode(abc.ABC):\n def __init__(self, children=None):\n self.children = children if children else []\n \n def __repr__(self):\n result = type(self).__name__ # class name\n if self.children:\n children_reprs = [repr(child) for child in self.children]\n children_lines = '\\n'.join(children_reprs)\n children_lines_tabbed = map(lambda x: '\\t' + x, children_lines.splitlines())\n result += '\\n' + '\\n'.join(children_lines_tabbed)\n return result\n\n @abc.abstractmethod\n def compile(self, symbol_table, compiled_code):\n for child in self.children:\n child.compile(symbol_table, compiled_code)\n\n\nclass CodeBlock(ASTNode):\n \"\"\"\n Represents a block of statements. \n For instance, the main program or part of a \n flow control statement. Its children are a list\n of statements.\n \"\"\"\n def __init__(self, children):\n super().__init__(children=children)\n\n def compile(self, symbol_table, compiled_code):\n symbol_table.increment_scope()\n super().compile(symbol_table, compiled_code)\n symbol_table.decrement_scope()\n\nclass MainProgram(CodeBlock):\n \"\"\"\n Represents the entire program, has a CodeBlock as\n its only child, and a version\n \"\"\"\n def __init__(self, children, version):\n super().__init__(children=children)\n assert version.value == '1.450', version\n\n def compile(self, symbol_table, compiled_code):\n self.children[0].compile(symbol_table, compiled_code)\n\nclass PrimitiveLiteral(ASTNode):\n \"\"\"\n An abstract base class that represents primitive literals\n The string of the value is stored as its only child.\n \"\"\"\n def __init__(self, data, expr_type):\n super().__init__(children=[data])\n self.expr_type = expr_type\n\n def compile(self, symbol_table, compiled_code):\n entry = symbol_table.get_entry(expr_type=self.expr_type)\n compiled_code.append(['VAL_COPY', self.children[0], entry])\n return entry\n\nclass NumbrLiteral(PrimitiveLiteral):\n \"\"\"\n An expression that represents a Numbr (like 5).\n The string of the value is stored as its only child.\n \"\"\"\n def __init__(self, data):\n PrimitiveLiteral.__init__(self, data=data, expr_type=PrimitiveType.NUMBR)\n\nclass TroofLiteral(PrimitiveLiteral):\n \"\"\"\n An expression that represents a Troof (like WIN).\n The string of the value is stored as its only child.\n Note the enclosing quotes are included in the string.\n \"\"\"\n def __init__(self, data):\n PrimitiveLiteral.__init__(self, data=data, expr_type=PrimitiveType.TROOF)\n\n def compile(self, symbol_table, compiled_code):\n entry = symbol_table.get_entry(expr_type=self.expr_type)\n value = 1 if self.children[0] == 'WIN' else 0\n compiled_code.append(['VAL_COPY', value, entry])\n return entry\n\nclass LettrLiteral(PrimitiveLiteral):\n \"\"\"\n An expression that represents a Lettr (like 'a').\n The string of the value is stored as its only child.\n Note the enclosing quotes are included in the string.\n \"\"\"\n def __init__(self, data):\n PrimitiveLiteral.__init__(self, data=data, expr_type=PrimitiveType.LETTR)\n \n def compile(self, symbol_table, compiled_code):\n entry = symbol_table.get_entry(expr_type=self.expr_type)\n lettr = self.children[0] # like ':)'\n mapping_to_lmao_char = {\n \"':)'\": r\"'\\n'\",\n \"':>'\": r\"'\\t'\",\n \"':''\": r\"'\\''\",\n \"'::'\": r\"':'\", \n r\"'\\'\": r\"'\\\\'\", \n } \n lmao_char = mapping_to_lmao_char.get(lettr, lettr)\n compiled_code.append(['VAL_COPY', lmao_char, entry])\n return entry\n\nclass VisibleStatement(ASTNode):\n \"\"\"\n A statement generated from \"VISIBLE , , \".\n The expr node is stored as its only child.\n \"\"\"\n def __init__(self, children, output_newline=True):\n super().__init__(children=children)\n self.output_newline = output_newline\n\n def compile(self, symbol_table, compiled_code):\n def print_entry(entry, compiled_code):\n if entry.expr_type in {PrimitiveType.NUMBAR, PrimitiveType.NUMBR, PrimitiveType.TROOF}:\n compiled_code.append(['OUT_NUM', entry])\n elif entry.expr_type == PrimitiveType.LETTR:\n compiled_code.append(['OUT_CHAR', entry])\n\n for child in self.children:\n child_entry = child.compile(symbol_table, compiled_code)\n if isinstance(child_entry.expr_type, PrimitiveType):\n print_entry(child_entry, compiled_code)\n elif isinstance(child_entry.expr_type, ArrayType):\n compiled_code.append(['# Printing Array', child_entry])\n size_entry = symbol_table.get_entry(expr_type=PrimitiveType.NUMBR)\n compiled_code.append(['AR_GET_SIZE', child_entry, size_entry])\n index_entry = symbol_table.get_entry(expr_type=PrimitiveType.NUMBR)\n compiled_code.append(['VAL_COPY', 0, index_entry])\n \n loop_start = symbol_table.get_unique_label('visible_array_loop_start')\n compiled_code.append([loop_start + ':'])\n test_entry = symbol_table.get_entry(expr_type=PrimitiveType.TROOF)\n compiled_code.append(['TEST_GTE', index_entry, size_entry, test_entry])\n\n loop_end = symbol_table.get_unique_label('visible_array_loop_end')\n compiled_code.append(['JUMP_IF_N0', test_entry, loop_end])\n\n subtype = child_entry.expr_type.subtype\n value_entry = symbol_table.get_entry(expr_type=subtype)\n compiled_code.append(['AR_GET_IDX', child_entry, index_entry, value_entry])\n\n print_entry(value_entry, compiled_code)\n\n compiled_code.append(['ADD', 1, index_entry, index_entry])\n compiled_code.append(['JUMP', loop_start])\n compiled_code.append([loop_end + ':'])\n\n compiled_code.append(['# Done Printing Array', child_entry])\n else:\n raise CompilerTypeError(f'Unable to print type {child_entry.expr_type}')\n if self.output_newline:\n compiled_code.append(['OUT_CHAR', r\"'\\n'\"])\n\n\nclass VariableDeclaration(ASTNode):\n \"\"\"\n An expression that represents a varible identifier (like x).\n The string of the variable's name and its type are its children.\n \"\"\"\n def compile(self, symbol_table, compiled_code):\n name, declaration_type = self.children\n return symbol_table.declare_variable(name, declaration_type)\n\nclass VariableUse(ASTNode):\n \"\"\"\n An expression that represents a varible identifier (like x).\n The string of the variable's name is stored as its only child.\n \"\"\"\n def compile(self, symbol_table, compiled_code):\n name = self.children[0]\n return symbol_table.get_entry_for_variable(name)\n\nclass MathBinaryExpression(ASTNode):\n \"\"\"\n An expression that represents a math binary operation \n (like 'SUM OF josh AN 6'). The children consist of\n the operator as a string (like 'SUM'), the first operand,\n and the second operand.\n \"\"\"\n def compile(self, symbol_table, compiled_code):\n operator, expr_1, expr_2 = self.children\n entry_1 = expr_1.compile(symbol_table, compiled_code)\n entry_2 = expr_2.compile(symbol_table, compiled_code)\n\n numeric_types = {PrimitiveType.NUMBR, PrimitiveType.NUMBAR}\n if entry_1.expr_type not in numeric_types:\n raise CompilerTypeError(f'{expr_1} is not a numeric type.')\n if entry_2.expr_type not in numeric_types:\n raise CompilerTypeError(f'{expr_1} is not a numeric type.')\n if entry_1.expr_type != entry_2.expr_type:\n raise CompilerTypeError(f'{expr_1} and {expr_2} do not match types.')\n\n result_entry = symbol_table.get_entry(expr_type=entry_1.expr_type)\n \n math_lol_to_lmao = {\n 'SUM': 'ADD',\n 'DIFF': 'SUB',\n 'PRODUKT': 'MULT',\n 'QUOSHUNT': 'DIV',\n }\n lmao_command = math_lol_to_lmao[operator]\n compiled_code.append([lmao_command, entry_1, entry_2, result_entry])\n\n\n return result_entry\n\nclass MathUnaryExpression(ASTNode):\n \"\"\"\n An expression that represents a math unary operation \n (like 'FLIP OF 6'). The children consist of\n the operator as a string (like 'FLIP') and the operand.\n \"\"\"\n def compile(self, symbol_table, compiled_code):\n operator, expr = self.children\n entry = expr.compile(symbol_table, compiled_code)\n\n numeric_types = {PrimitiveType.NUMBR, PrimitiveType.NUMBAR}\n if entry.expr_type not in numeric_types:\n raise CompilerTypeError(f'{entry} is not a numeric type.')\n\n result_entry = symbol_table.get_entry(expr_type=entry.expr_type)\n \n if operator == 'FLIP':\n compiled_code.append(['DIV', 1, entry, result_entry])\n else: # operator == 'SQUAR':\n compiled_code.append(['MULT', entry, entry, result_entry])\n return result_entry\n\n\nclass AssignmentExpression(ASTNode):\n \"\"\"\n An expression that represents an assignment (like 'toyz R \"us\"')\n or intializations (like 'I HAS A x ITZ A NUMBR AN ITZ 5').\n Its expr_type is the type of the right side of the assignment\n (YARN and NUMBR in the above examples).\n The left side (the variable expression) and the right side (the value)\n being assigned compose its two children\n \"\"\"\n def __init__(self, left_side, right_side):\n super().__init__(children=[left_side, right_side])\n \n \n def compile(self, symbol_table, compiled_code):\n left_side, right_side = self.children\n right_entry = right_side.compile(symbol_table, compiled_code)\n left_entry = left_side.compile(symbol_table, compiled_code)\n if left_entry.expr_type != right_entry.expr_type:\n raise CompilerTypeError(f'{left_entry.expr_type} != {right_entry.expr_type}')\n if left_entry.array_entry:\n compiled_code.append(['AR_SET_IDX', left_entry.array_entry, \n left_entry.index_entry, right_entry])\n compiled_code.append(['VAL_COPY', right_entry, left_entry]) \n elif isinstance(left_entry.expr_type, PrimitiveType):\n compiled_code.append(['VAL_COPY', right_entry, left_entry])\n else:\n compiled_code.append(['AR_COPY', right_entry, left_entry])\n return left_entry\n\nclass LogicalExpressionLazy(ASTNode):\n \"\"\"\n An expression that represents a logical expression \n (like 'BOTH OF WIN AN FAIL').\n The first child is the operator, and the rest of the children\n are the TROOF expressions to be evaluated.\n Only evaluates as many operands as needed to determine result.\n \"\"\"\n def compile(self, symbol_table, compiled_code):\n def check_is_troof_and_get_entry(expr):\n entry = expr.compile(symbol_table, compiled_code)\n if entry.expr_type != PrimitiveType.TROOF:\n raise CompilerTypeError(\n f'Using non-TROOF type {entry.expr_type} in logical expression')\n return entry\n\n operator = self.children[0]\n \n result_entry = symbol_table.get_entry(expr_type=PrimitiveType.TROOF)\n child_exprs = self.children[1:]\n\n compiled_code.append([f'# Logical Expression (result in {result_entry})'])\n \n if operator == 'NOT':\n entry = check_is_troof_and_get_entry(child_exprs[0])\n compiled_code.append(['TEST_EQU', entry, 0, result_entry])\n elif operator in {'BOTH', 'ALL', 'EITHER', 'ANY'}:\n lazy_jump_label = symbol_table.get_unique_label(root='logical_lazy_jump')\n for expr in child_exprs:\n entry = check_is_troof_and_get_entry(expr)\n command = 'JUMP_IF_0' if operator in {'BOTH', 'ALL'} else 'JUMP_IF_N0'\n compiled_code.append([command, entry, lazy_jump_label])\n compiled_code.append(['VAL_COPY', entry, result_entry])\n\n end_label = symbol_table.get_unique_label(root='logical_end')\n compiled_code.append(['JUMP', end_label])\n compiled_code.append([lazy_jump_label + ':'])\n value = 0 if operator in {'BOTH', 'ALL'} else 1\n compiled_code.append(['VAL_COPY', value, result_entry])\n compiled_code.append([end_label + ':'])\n elif operator in {}:\n pass\n else: # operator == 'WON'\n entry_1 = check_is_troof_and_get_entry(child_exprs[0])\n entry_2 = check_is_troof_and_get_entry(child_exprs[1])\n compiled_code.append(['TEST_NEQU', entry_1, entry_2, result_entry])\n\n compiled_code.append([f'# Logical Expression (result in {result_entry}) Done'])\n return result_entry\n\nclass LogicalExpression(ASTNode):\n \"\"\"\n An expression that represents a logical expression \n (like 'BOTH OF WIN AN FAIL').\n The first child is the operator, and the rest of the children\n are the TROOF expressions to be evaluated.\n \"\"\"\n def compile(self, symbol_table, compiled_code):\n operator = self.children[0]\n entries = [expr.compile(symbol_table, compiled_code) \n for expr in self.children[1:]]\n result_entry = symbol_table.get_entry(expr_type=PrimitiveType.TROOF)\n if operator == 'NOT':\n compiled_code.append(['TEST_EQU', entries[0], 0, result_entry])\n elif operator in {'BOTH', 'EITHER', 'WON'}:\n compiled_code.append(['ADD', entries[0], entries[1], result_entry])\n if operator == 'BOTH': \n compiled_code.append(['TEST_EQU', result_entry, 2, result_entry])\n elif operator == 'EITHER': \n compiled_code.append(['TEST_GTE', result_entry, 1, result_entry])\n else: # operator == 'WON': \n compiled_code.append(['TEST_EQU', result_entry, 1, result_entry])\n else: # operator in {'ALL', 'ANY'}:\n compiled_code.append(['VAL_COPY', 0, result_entry])\n for entry in entries:\n compiled_code.append(['ADD', entry, result_entry, result_entry])\n if operator == 'ALL':\n compiled_code.append(['TEST_EQU', len(entries), result_entry, result_entry])\n else: # operator == 'ANY'\n compiled_code.append(['TEST_GTE', result_entry, 1, result_entry])\n return result_entry\n\n\n\nclass ComparisonExpression(ASTNode):\n \"\"\"\n An expression that represents a comparison expression \n (like 'BOTH SAEM 5 AN 7').\n The first child is the operator, and the rest of the children\n are the two operands.\n \"\"\"\n def compile(self, symbol_table, compiled_code):\n operator, expr_1, expr_2 = self.children\n entry_1 = expr_1.compile(symbol_table, compiled_code)\n entry_2 = expr_2.compile(symbol_table, compiled_code)\n result_entry = symbol_table.get_entry(expr_type=PrimitiveType.TROOF)\n \n if entry_1.expr_type != entry_2.expr_type:\n compiled_code.append(['VAL_COPY', 0, result_entry])\n return result_entry\n\n lol_to_lmao = {\n 'SAEM': 'TEST_EQU',\n 'DIFFRINT': 'TEST_NEQU',\n 'FURSTSMALLR': 'TEST_LESS',\n 'FURSTBIGGR': 'TEST_GTR',\n }\n lmao_command = lol_to_lmao[operator]\n compiled_code.append([lmao_command, entry_1, entry_2, result_entry])\n return result_entry\n\n\nclass WhatevrExpression(ASTNode):\n \"\"\"\n A node representing a random NUMBR.\n \"\"\"\n def __init__(self):\n super().__init__()\n \n def compile(self, symbol_table, compiled_code):\n result_entry = symbol_table.get_entry(expr_type=PrimitiveType.NUMBR)\n compiled_code.append(['RANDOM', result_entry])\n return result_entry\n\nclass GimmehExpression(ASTNode):\n \"\"\"\n A node representing a request of a LETTR from standard input.\n \"\"\"\n def __init__(self):\n super().__init__()\n \n def compile(self, symbol_table, compiled_code):\n result_entry = symbol_table.get_entry(expr_type=PrimitiveType.LETTR)\n compiled_code.append(['IN_CHAR', result_entry])\n return result_entry\n\nclass ORLYStatement(ASTNode):\n \"\"\"\n A node representing a O RLY? statement.\n Its children are (in the following order):\n a conditional expression,\n a code block (YA RLY),\n a code block (possibly None) of NO WAI\n\n \"\"\"\n def __init__(self, children):\n super().__init__(children=children)\n \n def compile(self, symbol_table, compiled_code):\n def compile_and_check_troof(expr):\n entry = expr.compile(symbol_table, compiled_code)\n if entry.expr_type != PrimitiveType.TROOF:\n raise CompilerTypeError(\n f'{cond_entry.expr_type} is not an acceptable conditional expression')\n return entry\n\n compiled_code.append(['# Compiling O RLY Statement'])\n expr, if_true_block, otherwise_block = self.children\n symbol_table.push_if_stack('if')\n \n oic_label = symbol_table.get_unique_label(root='oic')\n \n expr_entry = compile_and_check_troof(expr)\n after_label = symbol_table.get_unique_label(root='after_if_true_block')\n compiled_code.append(['JUMP_IF_0', expr_entry, after_label])\n if_true_block.compile(symbol_table, compiled_code)\n compiled_code.append(['JUMP', oic_label])\n compiled_code.append([after_label + ':'])\n\n\n if otherwise_block:\n otherwise_block.compile(symbol_table, compiled_code)\n\n compiled_code.append([oic_label + ':'])\n symbol_table.pop_if_stack()\n compiled_code.append(['# Done with O RLY Statement'])\n\nclass LoopStatement(ASTNode):\n \"\"\"\n A node representing a loop statement.\n Its children are (in the following order):\n a code block representing the body of the loop\n \"\"\"\n def compile(self, symbol_table, compiled_code):\n def compile_and_check_troof(expr):\n entry = expr.compile(symbol_table, compiled_code)\n if entry.expr_type != PrimitiveType.TROOF:\n raise CompilerTypeError(\n f'{cond_entry.expr_type} is not an acceptable conditional expression')\n return entry\n\n compiled_code.append(['# Compiling Loop Statement'])\n assign_expression, til_expression, body = self.children\n\n start_label = symbol_table.get_unique_label(root='loop_start')\n compiled_code.append([start_label + ':'])\n\n end_label = symbol_table.get_unique_label(root='loop_end')\n\n if til_expression is not None:\n til_entry = compile_and_check_troof(til_expression)\n compiled_code.append(['JUMP_IF_N0', til_entry, end_label])\n\n symbol_table.push_GTFO_stack(end_label)\n\n body.compile(symbol_table, compiled_code)\n\n if assign_expression is not None:\n assign_expression.compile(symbol_table, compiled_code)\n\n compiled_code.append(['JUMP', start_label])\n compiled_code.append([end_label + ':'])\n symbol_table.pop_GTFO_stack()\n compiled_code.append(['# Done with Loop Statement'])\n\nclass GTFOStatement(ASTNode):\n \"\"\"\n A node representing a GTFO (break) statement.\n It has no children. It relies on the Symbol Table to determine\n jump destination.\n \"\"\"\n def compile(self, symbol_table, compiled_code):\n destination = symbol_table.read_GTFO_stack()\n compiled_code.append(['JUMP', destination])\n\nclass YarnLiteral(ASTNode):\n \"\"\"\n An expression that represents a YARN (LOTZ OF LETTRs).\n It's only child is a string (with enclosing double quotes).\n \"\"\"\n def compile(self, symbol_table, compiled_code):\n value = self.children[0]\n expr_type = ArrayType(PrimitiveType.LETTR)\n array_entry = symbol_table.get_entry(expr_type=expr_type)\n compiled_code.append(['# Compiling YARN', value, array_entry])\n\n letters = []\n value = value[1:-1] # remove enclosing double quotes\n while value:\n first_letter = value[0]\n if first_letter == ':':\n escaped_letter = value[1]\n mapping_to_lmao_char = {\n \")\": r\"'\\n'\",\n \">\": r\"'\\t'\",\n '\"': '\\'\"\\'',\n \":\": r\"':'\", \n } \n value = value[2:]\n letters.append(mapping_to_lmao_char[escaped_letter])\n else: \n if first_letter == \"'\":\n letter = r'\\''\n else:\n letter = first_letter\n letters.append(\"'\" + letter + \"'\")\n value = value[1:]\n compiled_code.append(['# LMAO chars', letters])\n compiled_code.append(['AR_SET_SIZE', array_entry, len(letters)])\n for i, letter in enumerate(letters):\n compiled_code.append(['AR_SET_IDX', array_entry, i, letter])\n compiled_code.append(['# Done compiling YARN'])\n return array_entry\n\nclass ArrayDeclaration(ASTNode):\n \"\"\"\n A node representing an array declaration. Its children are its name,\n the type of the array, and its size (as an expression).\n \"\"\"\n def compile(self, symbol_table, compiled_code):\n name, array_type, size_expr = self.children\n array_entry = symbol_table.declare_variable(name, array_type)\n size_entry = size_expr.compile(symbol_table, compiled_code)\n if size_entry.expr_type != PrimitiveType.NUMBR:\n raise CompilerTypeError(f'{size_entry.expr_type} is not accceptable size for array')\n compiled_code.append(['AR_SET_SIZE', array_entry, size_entry])\n\nclass ArrayIndex(ASTNode):\n \"\"\"\n A node representing an array index (like \"x'Z 0\"). It has two children,\n an VariableUse node and an index expression.\n \"\"\"\n def compile(self, symbol_table, compiled_code):\n array_node, index_expr = self.children\n array_entry = array_node.compile(symbol_table, compiled_code)\n index_entry = index_expr.compile(symbol_table, compiled_code)\n if index_entry.expr_type != PrimitiveType.NUMBR:\n raise CompilerTypeError(f'{index_entry.expr_type} is not accceptable size for array')\n \n array_subtype = array_entry.expr_type.subtype\n result_entry = symbol_table.get_array_index_entry(\n expr_type=array_subtype, array_entry=array_entry, index_entry=index_entry)\n \n compiled_code.append(['AR_GET_IDX', array_entry, index_entry, result_entry])\n return result_entry\n\nclass LengthzExpression(ASTNode):\n \"\"\"\n A node representing an lengthz expression, its only child is the array expression.\n \"\"\"\n def compile(self, symbol_table, compiled_code):\n expr = self.children[0]\n array_entry = expr.compile(symbol_table, compiled_code)\n if not isinstance(array_entry.expr_type, ArrayType):\n raise CompilerTypeError(f'Can not get length of non-array type {array_entry.expr_type}')\n result_entry = symbol_table.get_entry(expr_type=PrimitiveType.NUMBR)\n compiled_code.append(['AR_GET_SIZE', array_entry, result_entry])\n return result_entry\n\nclass CaseStatement(ASTNode):\n \"\"\"\n A node representing a case (WTF?) statement.\n Its children are (in the following order):\n an expression to be matched with the cases,\n a (possibly empty list of cases).\n Cases with two elements consist of a literal and a block.\n A last case with a single element is the default block.\n\n \"\"\"\n def compile(self, symbol_table, compiled_code):\n compiled_code.append(['# Case Statement'])\n expr, cases = self.children\n\n expr_entry = expr.compile(symbol_table, compiled_code)\n\n literals = []\n blocks = []\n block_labels = []\n for case in cases:\n if len(case) == 2:\n literal, block = case\n literals.append(literal)\n else:\n block = case[0]\n blocks.append(block)\n block_labels.append(symbol_table.get_unique_label(root='block_start'))\n \n for literal, block_label in zip(literals, block_labels):\n literal_entry = literal.compile(symbol_table, compiled_code)\n\n if literal_entry.expr_type != expr_entry.expr_type:\n raise CompilerTypeError(f'Type mismatch in case statement')\n\n test_entry = symbol_table.get_entry(expr_type=PrimitiveType.TROOF)\n\n if isinstance(literal_entry.expr_type, PrimitiveType):\n compiled_code.append(['TEST_EQU', expr_entry, literal_entry, test_entry])\n else:\n index = symbol_table.get_entry(expr_type=PrimitiveType.NUMBR)\n literal_size = symbol_table.get_entry(expr_type=PrimitiveType.NUMBR)\n expr_size = symbol_table.get_entry(expr_type=PrimitiveType.NUMBR)\n\n literal_element = symbol_table.get_entry(expr_type=literal_entry.expr_type.subtype)\n expr_element = symbol_table.get_entry(expr_type=literal_entry.expr_type.subtype)\n\n equal_array_start = symbol_table.get_unique_label(root='equal_array_start')\n equal_array_end = symbol_table.get_unique_label(root='equal_array_end')\n compiled_code += [\n ['# Doing array equality test for case statement'],\n ['AR_GET_SIZE', literal_entry, literal_size],\n ['AR_GET_SIZE', expr_entry, expr_size],\n ['TEST_EQU', literal_size, expr_size, test_entry],\n ['JUMP_IF_0', test_entry, equal_array_end, '# Jump if different sizes'],\n\n ['# Check all indices'],\n ['VAL_COPY', 0, index],\n [equal_array_start + ':'],\n\n ['TEST_GTE', index, literal_size, test_entry],\n ['JUMP_IF_N0', test_entry, equal_array_end, '# Reached end of array with no mismatches'],\n\n ['AR_GET_IDX', literal_entry, index, literal_element],\n ['AR_GET_IDX', expr_entry, index, expr_element],\n ['TEST_EQU', literal_element, expr_element, test_entry],\n ['JUMP_IF_0', test_entry, equal_array_end],\n ['ADD', 1, index, index],\n ['JUMP', equal_array_start],\n [equal_array_end + ':']\n ]\n compiled_code.append(['JUMP_IF_N0', test_entry, block_label])\n \n if len(block_labels) > len(literals): # default provided\n compiled_code.append(['JUMP', block_labels[-1]])\n \n end_label = symbol_table.get_unique_label(root='case_end')\n compiled_code.append(['JUMP', end_label])\n \n symbol_table.push_GTFO_stack(end_label)\n\n for block_label, block in zip(block_labels, blocks):\n compiled_code.append([block_label + ':'])\n block.compile(symbol_table, compiled_code)\n\n\n symbol_table.pop_GTFO_stack()\n compiled_code.append([end_label + ':'])\n \nclass FunctionStatement(ASTNode):\n \"\"\"\n A node for functions\n \n func[0] = function name (string)\n func[1] = args (list of identifiers and type declarations)\n func[2] = code block\n func[3] = return type (type declaration)\n \"\"\"\n def compile(self, symbol_table, compiled_code):\n func_name, args, block, returnType = self.children\n if len(symbol_table.function_stack) != 0 or len(symbol_table.gtfo_stack) != 0 or len(symbol_table.if_stack) != 0:\n raise NestedError('Function is nested')\n symbol_table.push_func_stack(func_name)\n function_label_start = symbol_table.get_unique_label(root=func_name)\n function_label_end = symbol_table.get_unique_label(root='end_of_' + func_name)\n arg_entries = symbol_table.get_arg_entries(func_name, function_label_start, args, returnType)\n \n compiled_code += [\n ['JUMP', function_label_end],\n [function_label_start + ':']\n ]\n buffer_code = []\n for entry in arg_entries:\n buffer_code.append(['POP', entry])\n buffer_code.reverse()\n for line in buffer_code:\n compiled_code.append(line)\n block.compile(symbol_table, compiled_code)\n compiled_code.append([function_label_end + ':'])\n symbol_table.decrement_scope()\n symbol_table.pop_func_stack()\n \nclass ReturnStatement(ASTNode):\n \"\"\"\n A node to return out of a function\n \"\"\"\n def compile(self, symbol_table, compiled_code):\n f_r_vars = symbol_table.f_recur_variables.copy()\n compiled_code.append(['# Found compiled'])\n expr = self.children[0]\n f_name = symbol_table.read_func_stack()\n func = symbol_table.declared_function_variables[f_name]\n expr_entry = expr.compile(symbol_table, compiled_code)\n if expr_entry.expr_type != func.returnType:\n raise TypeError('Function return type declaration is not the same')\n func_return = symbol_table.get_entry(PrimitiveType.LETTR)\n compiled_code += [\n ['POP', func_return],\n ['PUSH', expr_entry],\n ['JUMP', func_return]\n ]\n symbol_table.f_recur_variables = f_r_vars\n\nclass FunctionCallExpression(ASTNode):\n \"\"\"\n A node for functions \n function[0] = function name\n function[1] = args passed in\n \"\"\"\n def compile(self, symbol_table, compiled_code):\n f_r_vars = symbol_table.f_recur_variables.copy()\n func_name, args = self.children\n f_vars = symbol_table.declared_function_variables[func_name]\n \n return_str = 'return_from_' + func_name\n return_label = symbol_table.get_unique_label(root=return_str)\n func_obj = symbol_table.declared_function_variables[func_name]\n function_label_start = func_obj.funcLabel\n func_entry = symbol_table.get_entry(func_obj.returnType)\n if args:\n if len(f_vars.args) == 0:\n raise ArgumentError('Function does not take arguments')\n if symbol_table.function_stack:\n for entry in f_r_vars:\n compiled_code.append(['PUSH', entry])\n compiled_code.append(['PUSH', return_label])\n arg_pos = 0\n for elm in args:\n elm_entry = elm.compile(symbol_table, compiled_code)\n if elm_entry.expr_type != f_vars.args[arg_pos].expr_type:\n raise ArgumentError('Function typing is incorrect')\n arg_pos += 1\n compiled_code.append(['PUSH', elm_entry])\n compiled_code += [\n ['JUMP', function_label_start],\n [return_label + ':'],\n ['POP', func_entry]\n ]\n if symbol_table.function_stack:\n for entry in reversed(f_r_vars):\n compiled_code.append(['POP', entry])\n else:\n compiled_code += [\n ['PUSH', return_label],\n ['JUMP', function_label_start],\n [return_label + ':'],\n ['POP', func_entry]\n ]\n return func_entry\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n ", "sub_path": "Project7/ast_nodes.py", "file_name": "ast_nodes.py", "file_ext": "py", "file_size_in_byte": 31851, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "enum.Enum", "line_number": 5, "usage_type": "call"}, {"api_name": "collections.namedtuple", "line_number": 6, "usage_type": "call"}, {"api_name": "abc.ABC", "line_number": 11, "usage_type": "attribute"}, {"api_name": "abc.abstractmethod", "line_number": 24, "usage_type": "attribute"}]}
+{"seq_id": "86809456", "text": "# -*- coding: utf-8 -*- \nimport wx\n\nfrom src.Package.package import ReqElecTrend,FrameHeader,FrameTail,Time\nfrom src.CommonUse.staticVar import staticVar\nimport time\nclass ReqElecTrendDialog(wx.Dialog):\n def __init__(self,parent):\n wx.Dialog.__init__(self,parent,-1,u\"电磁分布态势数据请求\",size=(400,500))\n\n self.SetFont(wx.Font(10, wx.ROMAN, wx.NORMAL, wx.LIGHT, underline=False, faceName=u\"微软雅黑\",\n encoding=wx.FONTENCODING_DEFAULT))\n self.parent=parent\n \n \n ###############################\n self.tail=FrameTail(0,0,0xAA)\n \n self.id=staticVar.getid()\n self.lowid=self.id&0x00FF\n self.highid=self.id>>8\n \n self.List=staticVar.getCentreFreq()\n self.ListFreq=staticVar.getFreq()\n ##############################\n panel=wx.Panel(self,-1)\n \n self.FreqSection=wx.ComboBox(panel,-1,u\"FM调频广播频段\",choices=self.List)\n self.FreqSection.SetSelection(0)\n self.radioChoose=wx.RadioButton(panel,-1,u\"选择频率\")\n self.radioHand=wx.RadioButton(panel,-1,u\"手动频率\")\n self.CentreFreq=wx.TextCtrl(panel,-1,size=(80,25))\n self.BandWidth=wx.TextCtrl(panel,-1,size=(80,25))\n self.Radius=wx.TextCtrl(panel,-1,size=(80,25))\n self.FenBianLv=wx.TextCtrl(panel,-1,size=(80,25))\n self.RefreshIntv=wx.TextCtrl(panel,-1,size=(80,25))\n\n #############################################\n curTime = time.strftime('%Y%m%d%H%M%S', time.localtime(time.time()))\n Year = int(curTime[0:4])\n Month = int(curTime[4:6])\n Day = int(curTime[6:8])\n Hour = int(curTime[8:10])\n Min = int(curTime[10:12]) + 2\n\n ###############################################\n\n self.StartTimeYear = wx.ComboBox(panel, -1, str(Year), choices=[\"2015\", \"2016\", \"2017\", \"2018\"])\n self.StartTimeMonth = wx.ComboBox(panel, -1, str(Month),\n choices=[\"1\", \"2\", \"3\", \"4\", \"5\", \"6\", \"7\", \"8\", \"9\", \"10\", \"11\", \"12\"])\n self.StartTimeDay = wx.TextCtrl(panel, -1, str(Day), size=(60, 25))\n self.StartTimeHour = wx.TextCtrl(panel, -1, str(Hour), size=(60, 25))\n self.StartTimeMinute = wx.TextCtrl(panel, -1, str(Min), size=(60, 25))\n\n self.EndTimeYear = wx.ComboBox(panel, -1, str(Year), choices=[\"2015\", \"2016\", \"2017\", \"2018\"])\n self.EndTimeMonth = wx.ComboBox(panel, -1, str(Month),\n choices=[\"1\", \"2\", \"3\", \"4\", \"5\", \"6\", \"7\", \"8\", \"9\", \"10\", \"11\", \"12\"])\n self.EndTimeDay = wx.TextCtrl(panel, -1, str(Day), size=(60, 25))\n self.EndTimeHour = wx.TextCtrl(panel, -1, str(Hour), size=(60, 25))\n self.EndTimeMinute = wx.TextCtrl(panel, -1, str(Min), size=(60, 25))\n\n\n sizer=wx.BoxSizer(wx.VERTICAL)\n hBox=wx.BoxSizer(wx.HORIZONTAL)\n sizer.Add((30,30))\n hBox.Add(self.radioChoose,0,wx.LEFT,20)\n hBox.Add(self.radioHand,0,wx.LEFT,20)\n sizer.Add(hBox)\n sizer.Add(self.FreqSection,0,wx.LEFT|wx.TOP,20)\n\n sizer.Add((10,10))\n hBox1=wx.BoxSizer(wx.HORIZONTAL)\n hBox1.Add(wx.StaticText(panel,-1,u\"中心频率(MHz)\",size=(100,25)),0,wx.LEFT,20)\n hBox1.Add(self.CentreFreq,0,wx.LEFT,20)\n sizer.Add(hBox1)\n \n sizer.Add((10,10))\n hBox1=wx.BoxSizer(wx.HORIZONTAL)\n hBox1.Add(wx.StaticText(panel,-1,u\"带宽(MHz)\",size=(100,25)),0,wx.LEFT,20)\n hBox1.Add(self.BandWidth,0,wx.LEFT,20)\n sizer.Add(hBox1)\n\n sizer.Add((10,10))\n hBox1=wx.BoxSizer(wx.HORIZONTAL)\n hBox1.Add(wx.StaticText(panel,-1,u\"地理半径(km)\",size=(100,25)),0,wx.LEFT,20)\n hBox1.Add(self.Radius,0,wx.LEFT,20)\n sizer.Add(hBox1)\n \n sizer.Add((10,10))\n hBox1=wx.BoxSizer(wx.HORIZONTAL)\n hBox1.Add(wx.StaticText(panel,-1,u\"经纬度分辨率\",size=(100,25)),0,wx.LEFT,20)\n hBox1.Add(self.FenBianLv,0,wx.LEFT,20)\n sizer.Add(hBox1)\n\n sizer.Add((10,10))\n hBox1=wx.BoxSizer(wx.HORIZONTAL)\n hBox1.Add(wx.StaticText(panel,-1,u\"动态刷新间隔(Min)\",size=(130,25)),0,wx.LEFT,20)\n hBox1.Add(self.RefreshIntv,0,wx.LEFT,20)\n sizer.Add(hBox1)\n\n sizer.Add(wx.StaticText(panel,-1,u\"起始时间(年-月-日-时-分):\",size=(160,25)),0,wx.LEFT,20)\n sizer.Add((10,10))\n hBox1=wx.BoxSizer(wx.HORIZONTAL)\n hBox1.Add(self.StartTimeYear,0,wx.LEFT,20)\n hBox1.Add(wx.StaticText(panel,-1,\"-\"),0,wx.LEFT|wx.RIGHT|wx.ALIGN_BOTTOM,5)\n hBox1.Add(self.StartTimeMonth,0)\n hBox1.Add(wx.StaticText(panel,-1,\"-\"),0,wx.LEFT|wx.RIGHT|wx.ALIGN_BOTTOM,5)\n hBox1.Add(self.StartTimeDay,0)\n hBox1.Add(wx.StaticText(panel,-1,\"-\"),0,wx.LEFT|wx.RIGHT|wx.ALIGN_BOTTOM,5)\n hBox1.Add(self.StartTimeHour,0)\n hBox1.Add(wx.StaticText(panel,-1,\"-\"),0,wx.LEFT|wx.RIGHT|wx.ALIGN_BOTTOM,5)\n hBox1.Add(self.StartTimeMinute,0)\n sizer.Add(hBox1)\n\n sizer.Add(wx.StaticText(panel,-1,u\"终止时间(年-月-日-时-分):\",size=(160,25)),0,wx.LEFT,20)\n sizer.Add((10,10))\n hBox1=wx.BoxSizer(wx.HORIZONTAL)\n hBox1.Add(self.EndTimeYear,0,wx.LEFT,20)\n hBox1.Add(wx.StaticText(panel,-1,\"-\"),0,wx.LEFT|wx.RIGHT|wx.ALIGN_BOTTOM,5)\n hBox1.Add(self.EndTimeMonth,0)\n hBox1.Add(wx.StaticText(panel,-1,\"-\"),0,wx.LEFT|wx.RIGHT|wx.ALIGN_BOTTOM,5)\n hBox1.Add(self.EndTimeDay,0)\n hBox1.Add(wx.StaticText(panel,-1,\"-\"),0,wx.LEFT|wx.RIGHT|wx.ALIGN_BOTTOM,5)\n hBox1.Add(self.EndTimeHour,0)\n hBox1.Add(wx.StaticText(panel,-1,\"-\"),0,wx.LEFT|wx.RIGHT|wx.ALIGN_BOTTOM,5)\n hBox1.Add(self.EndTimeMinute,0)\n sizer.Add(hBox1)\n sizer.Add((30,30))\n hBox1=wx.BoxSizer(wx.HORIZONTAL)\n self.btn_ok=wx.Button(panel,-1,\"OK\",size=(60,25))\n \n hBox1.Add(self.btn_ok,0,wx.LEFT,20)\n hBox1.Add(wx.Button(panel,wx.ID_CANCEL,\"CANCEL\",size=(60,25)),0,wx.LEFT,20)\n sizer.Add(hBox1)\n panel.SetSizer(sizer)\n \n self.CentreFreq.Enable(True)\n self.BandWidth.Enable(True)\n self.radioHand.SetValue(True)\n \n self.Layout()\n self.Centre( wx.BOTH )\n \n ##Events\n self.btn_ok.Bind(wx.EVT_BUTTON,self.OnbtnOK)\n self.Bind(wx.EVT_RADIOBUTTON, self.OnRadio)\n \n def OnRadio(self,event):\n if(self.radioChoose.GetValue()):\n self.FreqSection.Enable(True)\n self.BandWidth.Enable(False)\n self.CentreFreq.Enable(False)\n\n elif(self.radioHand.GetValue()):\n self.FreqSection.Enable(False)\n self.CentreFreq.Enable(True)\n self.BandWidth.Enable(True)\n \n \n def ByteToTime(self,time):\n Obj=Time()\n Obj.HighYear=time[0]>>4\n Obj.LowYear=time[0]&0x00F\n Obj.Month=time[1]\n Obj.Day=time[2]\n Obj.HighHour=time[3]>>2\n Obj.LowHour=time[3]&0x03\n Obj.Minute=time[4]\n \n return Obj \n \n def OnbtnOK(self,event):\n reqElec=ReqElecTrend()\n reqElec.CommonHeader=FrameHeader(0x55,0xA2,self.lowid,self.highid)\n reqElec.CommonTail=self.tail\n if(self.radioChoose.GetValue()):\n centreFreq=int(self.ListFreq[self.FreqSection.GetSelection()][0])\n bandWidth=int(self.ListFreq[self.FreqSection.GetSelection()][1])\n\n else:\n centreFreq=int(self.CentreFreq.GetValue())\n bandWidth=int(self.BandWidth.GetValue())\n \n reqElec.HighCentreFreq=centreFreq>>8\n reqElec.LowCentreFreq=centreFreq&0x00FF\n reqElec.BandWidth=bandWidth\n reqElec.Radius=int(self.Radius.GetValue())\n fenBianLv=float(self.FenBianLv.GetValue())\n startTime=(int(self.StartTimeYear.GetValue()),int(self.StartTimeMonth.GetValue()), \\\n int(self.StartTimeDay.GetValue()),int(self.StartTimeHour.GetValue()), \\\n int(self.StartTimeMinute.GetValue())\n )\n endTime=(int(self.EndTimeYear.GetValue()),int(self.EndTimeMonth.GetValue()), \\\n int(self.EndTimeDay.GetValue()),int(self.EndTimeHour.GetValue()), \\\n int(self.EndTimeMinute.GetValue())\n )\n \n reqElec.FenBianLvInteger=int(fenBianLv)\n reqElec.FenBianLvFraction=int((fenBianLv-int(fenBianLv))*8)\n \n reqElec.RefreshIntv=int(self.RefreshIntv.GetValue())\n reqElec.StartTime=self.ByteToTime(startTime)\n reqElec.EndTime=self.ByteToTime(endTime)\n\n\n # if(staticVar.getSock()):\n # staticVar.getSock().sendall(bytearray(reqElec))\n # for i in bytearray(reqElec):\n # print i,\n\n self.parent.queueRequest.put(reqElec)\n\n self.Close()\n \n \n \n \n", "sub_path": "src/MapDialog/req_elec_trend.py", "file_name": "req_elec_trend.py", "file_ext": "py", "file_size_in_byte": 8928, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "wx.Dialog", "line_number": 7, "usage_type": "attribute"}, {"api_name": "wx.Dialog.__init__", "line_number": 9, "usage_type": "call"}, {"api_name": "wx.Dialog", "line_number": 9, "usage_type": "attribute"}, {"api_name": "wx.Font", "line_number": 11, "usage_type": "call"}, {"api_name": "wx.ROMAN", "line_number": 11, "usage_type": "attribute"}, {"api_name": "wx.NORMAL", "line_number": 11, "usage_type": "attribute"}, {"api_name": "wx.LIGHT", "line_number": 11, "usage_type": "attribute"}, {"api_name": "wx.FONTENCODING_DEFAULT", "line_number": 12, "usage_type": "attribute"}, {"api_name": "src.Package.package.FrameTail", "line_number": 17, "usage_type": "call"}, {"api_name": "src.CommonUse.staticVar.staticVar.getid", "line_number": 19, "usage_type": "call"}, {"api_name": "src.CommonUse.staticVar.staticVar", "line_number": 19, "usage_type": "name"}, {"api_name": "src.CommonUse.staticVar.staticVar.getCentreFreq", "line_number": 23, "usage_type": "call"}, {"api_name": "src.CommonUse.staticVar.staticVar", "line_number": 23, "usage_type": "name"}, {"api_name": "src.CommonUse.staticVar.staticVar.getFreq", "line_number": 24, "usage_type": "call"}, {"api_name": "src.CommonUse.staticVar.staticVar", "line_number": 24, "usage_type": "name"}, {"api_name": "wx.Panel", "line_number": 26, "usage_type": "call"}, {"api_name": "wx.ComboBox", "line_number": 28, "usage_type": "call"}, {"api_name": "wx.RadioButton", "line_number": 30, "usage_type": "call"}, {"api_name": "wx.RadioButton", "line_number": 31, "usage_type": "call"}, {"api_name": "wx.TextCtrl", "line_number": 32, "usage_type": "call"}, {"api_name": "wx.TextCtrl", "line_number": 33, "usage_type": "call"}, {"api_name": "wx.TextCtrl", "line_number": 34, "usage_type": "call"}, {"api_name": "wx.TextCtrl", "line_number": 35, "usage_type": "call"}, {"api_name": "wx.TextCtrl", "line_number": 36, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 39, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 39, "usage_type": "call"}, {"api_name": "time.time", "line_number": 39, "usage_type": "call"}, {"api_name": "wx.ComboBox", "line_number": 48, "usage_type": "call"}, {"api_name": "wx.ComboBox", "line_number": 49, "usage_type": "call"}, {"api_name": "wx.TextCtrl", "line_number": 51, "usage_type": "call"}, {"api_name": "wx.TextCtrl", "line_number": 52, "usage_type": "call"}, {"api_name": "wx.TextCtrl", "line_number": 53, "usage_type": "call"}, {"api_name": "wx.ComboBox", "line_number": 55, "usage_type": "call"}, {"api_name": "wx.ComboBox", "line_number": 56, "usage_type": "call"}, {"api_name": "wx.TextCtrl", "line_number": 58, "usage_type": "call"}, {"api_name": "wx.TextCtrl", "line_number": 59, "usage_type": "call"}, {"api_name": "wx.TextCtrl", "line_number": 60, "usage_type": "call"}, {"api_name": "wx.BoxSizer", "line_number": 63, "usage_type": "call"}, {"api_name": "wx.VERTICAL", "line_number": 63, "usage_type": "attribute"}, {"api_name": "wx.BoxSizer", "line_number": 64, "usage_type": "call"}, {"api_name": "wx.HORIZONTAL", "line_number": 64, "usage_type": "attribute"}, {"api_name": "wx.LEFT", "line_number": 66, "usage_type": "attribute"}, {"api_name": "wx.LEFT", "line_number": 67, "usage_type": "attribute"}, {"api_name": "wx.LEFT", "line_number": 69, "usage_type": "attribute"}, {"api_name": "wx.TOP", "line_number": 69, "usage_type": "attribute"}, {"api_name": "wx.BoxSizer", "line_number": 72, "usage_type": "call"}, {"api_name": "wx.HORIZONTAL", "line_number": 72, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 73, "usage_type": "call"}, {"api_name": "wx.LEFT", "line_number": 73, "usage_type": "attribute"}, {"api_name": "wx.LEFT", "line_number": 74, "usage_type": "attribute"}, {"api_name": "wx.BoxSizer", "line_number": 78, "usage_type": "call"}, {"api_name": "wx.HORIZONTAL", "line_number": 78, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 79, "usage_type": "call"}, {"api_name": "wx.LEFT", "line_number": 79, "usage_type": "attribute"}, {"api_name": "wx.LEFT", "line_number": 80, "usage_type": "attribute"}, {"api_name": "wx.BoxSizer", "line_number": 84, "usage_type": "call"}, {"api_name": "wx.HORIZONTAL", "line_number": 84, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 85, "usage_type": "call"}, {"api_name": "wx.LEFT", "line_number": 85, "usage_type": "attribute"}, {"api_name": "wx.LEFT", "line_number": 86, "usage_type": "attribute"}, {"api_name": "wx.BoxSizer", "line_number": 90, "usage_type": "call"}, {"api_name": "wx.HORIZONTAL", "line_number": 90, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 91, "usage_type": "call"}, {"api_name": "wx.LEFT", "line_number": 91, "usage_type": "attribute"}, {"api_name": "wx.LEFT", "line_number": 92, "usage_type": "attribute"}, {"api_name": "wx.BoxSizer", "line_number": 96, "usage_type": "call"}, {"api_name": "wx.HORIZONTAL", "line_number": 96, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 97, "usage_type": "call"}, {"api_name": "wx.LEFT", "line_number": 97, "usage_type": "attribute"}, {"api_name": "wx.LEFT", "line_number": 98, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 101, "usage_type": "call"}, {"api_name": "wx.LEFT", "line_number": 101, "usage_type": "attribute"}, {"api_name": "wx.BoxSizer", "line_number": 103, "usage_type": "call"}, {"api_name": "wx.HORIZONTAL", "line_number": 103, "usage_type": "attribute"}, {"api_name": "wx.LEFT", "line_number": 104, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 105, "usage_type": "call"}, {"api_name": "wx.LEFT", "line_number": 105, "usage_type": "attribute"}, {"api_name": "wx.RIGHT", "line_number": 105, "usage_type": "attribute"}, {"api_name": "wx.ALIGN_BOTTOM", "line_number": 105, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 107, "usage_type": "call"}, {"api_name": "wx.LEFT", "line_number": 107, "usage_type": "attribute"}, {"api_name": "wx.RIGHT", "line_number": 107, "usage_type": "attribute"}, {"api_name": "wx.ALIGN_BOTTOM", "line_number": 107, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 109, "usage_type": "call"}, {"api_name": "wx.LEFT", "line_number": 109, "usage_type": "attribute"}, {"api_name": "wx.RIGHT", "line_number": 109, "usage_type": "attribute"}, {"api_name": "wx.ALIGN_BOTTOM", "line_number": 109, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 111, "usage_type": "call"}, {"api_name": "wx.LEFT", "line_number": 111, "usage_type": "attribute"}, {"api_name": "wx.RIGHT", "line_number": 111, "usage_type": "attribute"}, {"api_name": "wx.ALIGN_BOTTOM", "line_number": 111, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 115, "usage_type": "call"}, {"api_name": "wx.LEFT", "line_number": 115, "usage_type": "attribute"}, {"api_name": "wx.BoxSizer", "line_number": 117, "usage_type": "call"}, {"api_name": "wx.HORIZONTAL", "line_number": 117, "usage_type": "attribute"}, {"api_name": "wx.LEFT", "line_number": 118, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 119, "usage_type": "call"}, {"api_name": "wx.LEFT", "line_number": 119, "usage_type": "attribute"}, {"api_name": "wx.RIGHT", "line_number": 119, "usage_type": "attribute"}, {"api_name": "wx.ALIGN_BOTTOM", "line_number": 119, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 121, "usage_type": "call"}, {"api_name": "wx.LEFT", "line_number": 121, "usage_type": "attribute"}, {"api_name": "wx.RIGHT", "line_number": 121, "usage_type": "attribute"}, {"api_name": "wx.ALIGN_BOTTOM", "line_number": 121, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 123, "usage_type": "call"}, {"api_name": "wx.LEFT", "line_number": 123, "usage_type": "attribute"}, {"api_name": "wx.RIGHT", "line_number": 123, "usage_type": "attribute"}, {"api_name": "wx.ALIGN_BOTTOM", "line_number": 123, "usage_type": "attribute"}, {"api_name": "wx.StaticText", "line_number": 125, "usage_type": "call"}, {"api_name": "wx.LEFT", "line_number": 125, "usage_type": "attribute"}, {"api_name": "wx.RIGHT", "line_number": 125, "usage_type": "attribute"}, {"api_name": "wx.ALIGN_BOTTOM", "line_number": 125, "usage_type": "attribute"}, {"api_name": "wx.BoxSizer", "line_number": 129, "usage_type": "call"}, {"api_name": "wx.HORIZONTAL", "line_number": 129, "usage_type": "attribute"}, {"api_name": "wx.Button", "line_number": 130, "usage_type": "call"}, {"api_name": "wx.LEFT", "line_number": 132, "usage_type": "attribute"}, {"api_name": "wx.Button", "line_number": 133, "usage_type": "call"}, {"api_name": "wx.ID_CANCEL", "line_number": 133, "usage_type": "attribute"}, {"api_name": "wx.LEFT", "line_number": 133, "usage_type": "attribute"}, {"api_name": "wx.BOTH", "line_number": 142, "usage_type": "attribute"}, {"api_name": "wx.EVT_BUTTON", "line_number": 145, "usage_type": "attribute"}, {"api_name": "wx.EVT_RADIOBUTTON", "line_number": 146, "usage_type": "attribute"}, {"api_name": "src.Package.package.Time", "line_number": 161, "usage_type": "call"}, {"api_name": "src.Package.package.ReqElecTrend", "line_number": 173, "usage_type": "call"}, {"api_name": "src.Package.package.FrameHeader", "line_number": 174, "usage_type": "call"}]}
+{"seq_id": "86654210", "text": "\"\"\"\nPlot posterior distributions for a binomial parameter.\n\nThis implementation uses the object oriented programming (OOP) paradigm.\n\nCreated Feb 27, 2015 by Tom Loredo\n2018-03-08 Modified for Py-3, updated for BDA18\n2020-02-26 Updated for BDA20\n\"\"\"\n\nimport numpy as np\n# import numpy.testing as npt\nimport scipy\nimport matplotlib as mpl\nfrom matplotlib.pyplot import *\nfrom numpy import *\nfrom scipy import stats, special, integrate\n\n\nion()\n\n\nclass BinomialInference:\n \"\"\"\n Bayesian inference for the probability of a Bernoulli outcome, based\n on binomial data.\n \"\"\"\n\n def __init__(self, n, n_trials, prior=1., na=200, arange=(0., 1.)):\n \"\"\"\n Define a posterior PDF for the probability of a Bernoulli outcome,\n alpha, based on binomial data.\n\n Parameters\n ----------\n\n n : int\n Number of successes\n\n n_trials : int\n Number of trials (>= n)\n\n prior : const or function\n Prior PDF for alpha, as a constant for flat prior, or\n a function that can evaluate the PDF on an array\n \"\"\"\n self.n, self.n_trials = n, n_trials\n self.na = na\n self.alphas = linspace(arange[0], arange[1], na)\n\n # Evaluate the prior on the grid.\n self.prior = prior # save for possible future reference\n if callable(prior):\n self.prior_pdf = prior(self.alphas)\n else:\n self.prior_pdf = prior * ones_like(self.alphas)\n\n # Evaluate the binomial likelihood function; ignore the\n # combinatorial factor (indep. of alpha).\n self.like = (self.alphas)**n * (1. - self.alphas)**(n_trials - n)\n\n # Bayes's theorem:\n numer = self.prior_pdf * self.like\n self.da = self.alphas[1] - self.alphas[0]\n self.mlike = np.trapz(numer, dx=self.da)\n self.post_pdf = numer / self.mlike\n\n def plot(self, ls='b-', lw=3, **kwds):\n \"\"\"\n Plot the posterior PDF in the current axes.\n \"\"\"\n plot(self.alphas, self.post_pdf, ls, lw=lw, **kwds)\n\n\n#-------------------------------------------------------------------------------\n# 1st 2 cases: const & beta(.5,.5) priors, (n, n_trials) = (8, 12)\n\n# Define the data.\nn, n_trials = 8, 12\n\nbi1 = BinomialInference(n, n_trials)\nbi1.plot(alpha=.5)\n\nbeta_half = stats.beta(a=.5, b=.5)\nbi2 = BinomialInference(n, n_trials, beta_half.pdf, arange=(1.e-4, 1 - 1.e-4))\nbi2.plot(ls='g--')\n\nxlabel(r'$\\alpha$')\nylabel('Posterior PDF')\n\n\n#-------------------------------------------------------------------------------\n# 2nd 2 cases: const & beta(.5,.5) priors, (n, n_trials) = (4, 12)\n\nn, n_trials = 4, 12\n\nbi3 = BinomialInference(n, n_trials)\nbi3.plot(alpha=.5)\n\nbi4 = BinomialInference(n, n_trials, beta_half.pdf, arange=(1.e-4, 1 - 1.e-4))\nbi4.plot(ls='g--')\n\n\n#-------------------------------------------------------------------------------\n# 3rd 2 cases: const & beta(.5,.5) priors, (n, n_trials) = 4*(8, 12)\n\nn, n_trials = 4 * 8, 4 * 12\n\nbi5 = BinomialInference(n, n_trials)\nbi5.plot(alpha=.5)\n\nbi6 = BinomialInference(n, n_trials, beta_half.pdf, arange=(1.e-4, 1 - 1.e-4))\nbi6.plot(ls='g--')\n", "sub_path": "Lab06-Assignment04/3-BinomialAlpha-OOP.py", "file_name": "3-BinomialAlpha-OOP.py", "file_ext": "py", "file_size_in_byte": 3156, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "numpy.trapz", "line_number": 65, "usage_type": "call"}, {"api_name": "scipy.stats.beta", "line_number": 84, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 84, "usage_type": "name"}]}
+{"seq_id": "192549958", "text": "import pandas as pd\nimport nmslib\nfrom codesearch.query.encoder import encode_query_string\nfrom codesearch.code.summarizer import summarize_code\n\n\nclass SemanticCodeSearchEngine:\n def __init__(self, search_index_dir):\n self.codebase_df = pd.read_feather(search_index_dir + '/codebase.df.feather')\n self.search_index = nmslib.init(method='hnsw', space='cosinesimil')\n self.search_index.loadIndex(search_index_dir + '/searchindex.nmslib')\n\n def search(self, query, num_result=10, summarize=True):\n embedding = encode_query_string(query)\n idxs, dists = self.search_index.knnQuery(embedding, k=num_result)\n\n results = []\n rank = 1\n\n for idx, dist in zip(idxs, dists):\n row = self.codebase_df[self.codebase_df.code_vec_hash == idx]\n if row.empty:\n row = self.codebase_df[self.codebase_df.comment_vec_hash == idx]\n\n row = row.iloc[0]\n if not row.empty:\n results.append({\n 'rank': rank,\n 'score': float(str(dist)),\n 'file': row.file_name,\n 'line': int(str(row.line)),\n 'code': row.function_body,\n 'summary': summarize_code(row.tokenized_body) if summarize else None,\n 'docstring': row.docstring\n })\n rank += 1\n\n return results\n\n\nif __name__ == '__main__':\n import json\n import time\n import sys\n from codesearch.utils.redis_queue import RedisQueueExchange\n\n search_engine = SemanticCodeSearchEngine(sys.argv[1])\n queue_exchange = RedisQueueExchange()\n\n while True:\n try:\n while not queue_exchange.empty('input'):\n # pop query from the input queue\n query = queue_exchange.fetch('input').decode()\n print('Found query:', query)\n # perform search and write result to output queue\n result = json.dumps(search_engine.search(query), indent=4)\n print('Writing results:', result)\n queue_exchange.write('output', query, result)\n queue_exchange.delete('input', query)\n pass\n\n # goto sleep for 100 milliseconds\n time.sleep(0.1)\n except Exception as exc:\n print(exc)\n print('Restarting searcher.py ...')\n search_engine = SemanticCodeSearchEngine(sys.argv[1])\n queue_exchange = RedisQueueExchange()\n", "sub_path": "search-engine/codesearch/query/searcher.py", "file_name": "searcher.py", "file_ext": "py", "file_size_in_byte": 2530, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pandas.read_feather", "line_number": 9, "usage_type": "call"}, {"api_name": "nmslib.init", "line_number": 10, "usage_type": "call"}, {"api_name": "codesearch.query.encoder.encode_query_string", "line_number": 14, "usage_type": "call"}, {"api_name": "codesearch.code.summarizer.summarize_code", "line_number": 33, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 47, "usage_type": "attribute"}, {"api_name": "codesearch.utils.redis_queue.RedisQueueExchange", "line_number": 48, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 57, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 64, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 68, "usage_type": "attribute"}, {"api_name": "codesearch.utils.redis_queue.RedisQueueExchange", "line_number": 69, "usage_type": "call"}]}
+{"seq_id": "65238679", "text": "import urllib.request\r\nimport re\r\nimport matplotlib.pyplot as plt\r\n\r\n\r\n\"\"\"\r\nКакие годы были самыми популярными с точки зрения выхода игр?\r\nКакие жанры были популярны в различные периоды времени?\r\n\"\"\"\r\n\r\nurl = 'https://raw.githubusercontent.com/Newbilius/Old-Games_DOS_Game_Gauntlet/master/GAMES.csv'\r\nresp = urllib.request.urlopen(url)\r\nrespData = resp.read()\r\ndata_str = str(respData.decode())\r\ndata_str = re.split(';|\\n', data_str)\r\ndates = data_str[3::4]\r\ngenres = data_str[1::4]\r\nset_dates = sorted(set(dates))\r\nset_genres = sorted(set(genres))\r\ndDates = dict.fromkeys(set_dates, 0)\r\ndGenres = dict.fromkeys(set_genres, 0)\r\nfor i in range(len(dates)):\r\n dDates[dates[i]] += 1\r\nfor i in range(len(genres)):\r\n dGenres[genres[i]] += 1\r\n\r\nfig, axs = plt.subplots(2, figsize=(25, 10))\r\naxs[0].bar(set_dates, list(dDates.values()))\r\naxs[1].bar(set_genres, list(dGenres.values()))\r\nplt.show()\r\n", "sub_path": "P3/Additional tasks/2_7.py", "file_name": "2_7.py", "file_ext": "py", "file_size_in_byte": 995, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "urllib.request.request.urlopen", "line_number": 12, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 12, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 12, "usage_type": "name"}, {"api_name": "re.split", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}]}
+{"seq_id": "282349446", "text": "import os\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport math\n\n\nfrom getData import Trainset\n\n\n\ndef trafficData():\n\t#获取traffic的文件列表,文件名列表\n\tfile_list = os.listdir('traffic-fix/raw')\n\tfile_name = []\n\tfor file in file_list:\n\t\tfile_name.append(file.split('.')[0])\n\n\tfile_path = os.path.join('traffic-fix/traffic', file_list[0])\n\twith open(file_path,'r') as file:\n \t#将所有行读取\n\t\tids = []\n\t\tlabels = []\n\t\tcurrent_t = []\n\t\ttime_diff = []\n\t\tprediction_t = []\n\n\t\tlines = file.readlines()\n\t\t\n\t\tfor line in lines:\n\t\t\tsplitBySpace = line.split(' ')\n\t\t\tids.append(splitBySpace[0])\n\t\t\tlabels.append(splitBySpace[1])\n\t\t\tcurrent_t.append(splitBySpace[2])\n\t\t\tprediction_t.append(splitBySpace[3].split(';')[0])\n\n\t\t\ttime_diff.append(int(splitBySpace[3].split(';')[0]) - int(splitBySpace[2])) \n\n\treturn max(time_diff)\n\ndef getRspeed(speed, index):\n\n\tfile_name = str(index) + '.txt'\n\n\twith open(os.path.join('attr',file_name),'r') as f:\n\n\t\tlines = f.readlines()\n\t\n\tline = lines[0].split('\t')\n\n\tvk = float(line[6])\n\n\t\n\n\treturn speed/vk\n\ndef getLevel(index):\n\n\twith open('attr/'+str(index)+'.txt','r') as f:\n\n\t\tlines = f.readlines()\n\n\t\tline = lines[0].split('\\t')\n\n\t\tlevel = int(line[1])\n\n\treturn level\n\ndef stateVspeed():\n\n\ttrainset = Trainset('test/27.777778')\n\tlabels = []\n\tspeeds = []\n\tep = 0.001\n\n\tfor i,data in enumerate(trainset,0):\n\t\t\n\t\tif i + 1 < 0:\n\t\t\tcontinue\n\n\t\tif i + 1 == 2037:\n\t\t\tbreak\n\n\t\tif data[2][0][2] != 3:\n\t\t\tcontinue\n\n\t\tlabels.append(data[5] - 4)\n\n\t\tspeeds.append(data[2][0][0])\n\n\t\t\n\n\tplt.scatter(speeds, labels)\n\tplt.show()\n\ndef speedDvL():\n\n\ttrainset = Trainset()\n\tdiffs = []\n\tlns = []\n\tep = 0.01\n\n\tfor data in trainset:\n\n\t\tif len(lns) > 1000:\n\t\t\tbreak\n\n\t\tif data[2][0][2] != 4:\n\t\t\tcontinue\n\n\t\tindex = data[4]\n\t\t# for seq in data[2]:\n\t\t# \tlabels.append(seq[2])\n\t\t# \tdensity = getDensity(seq[3], index)\n\t\t# \tdensitys.append(density)\n\n\t\t\n\n\t\tlns.append(data[2][0][0])\n\n\t\trspeed = getRspeed(data[2][0][0], index)\n\t\tdiffs.append(rspeed)\n\n\n\n\tplt.scatter(lns, diffs)\n\tplt.show()\n\ndef getMDn():\n\n\t\n\n\tdataset = Trainset('20190724')\n\n\tnum = len(dataset)\n\n\tprint(num)\n\n\tM = 0.0\n\n\tfor i,data in enumerate(dataset,0):\n\n\t\tM = data[2][0][3] + data[2][1][3] + data[2][2][3] + data[2][3][3] + data[2][4][3] + M\n\n\t\tif (i + 1) == 2000:\n\n\t\t\tbreak\n\n\tM = M / (2000 * 5)\n\n\tprint(M) \n\n\tD = 0.0\n\n\tfor i,data in enumerate(dataset,0):\n\n\t\tD = pow((data[2][0][3] - M),2) + pow((data[2][1][3] - M),2) + pow((data[2][2][3] - M),2) + pow((data[2][3][3] - M),2) + pow((data[2][4][3] - M),2) + D\n\n\t\tif (i + 1) == 2000:\n\n\t\t\tbreak\n\n\tD = D / (2000 * 5)\n\n\tD = math.sqrt(D)\n\n\tprint(D) \n\n\treturn num, M, D\n\ndef getMD(vk):\n\n\troot = os.path.join('test',str(vk))\n\n\tdataset = Trainset(root)\n\n\tnum = len(dataset)\n\n\tprint(num)\n\n\tM = 0.0\n\n\tfor i,data in enumerate(dataset,0):\n\n\t\tM = data[2][0][0] + data[2][1][0] + data[2][2][0] + data[2][3][0] + data[2][4][0] + M\n\n\t\tif (i + 1) == len(dataset):\n\n\t\t\tbreak\n\n\tM = M / (num * 5)\n\n\tprint(M) \n\n\tD = 0.0\n\n\tfor i,data in enumerate(dataset,0):\n\n\t\tD = pow((data[2][0][0] - M),2) + pow((data[2][1][0] - M),2) + pow((data[2][2][0] - M),2) + pow((data[2][3][0] - M),2) + pow((data[2][4][0] - M),2) + D\n\n\t\tif (i + 1) == len(dataset):\n\n\t\t\tbreak\n\n\tD = D / (num * 5)\n\n\tD = math.sqrt(D)\n\n\tprint(D) \n\n\treturn num, M, D\n\n\n\n\nif __name__ == '__main__':\n\n\t# 计算预测时间片与当前时间片的最大插值\n\n\t# max_time_diff = trafficData()\n\n\t# print(max_time_diff)\n\n\t\n\t# n,m,d = getMD(19.444444)\n\n\tstateVspeed()\n\n\t# getMDn()\n\n\t", "sub_path": "LSTM/getFeature.py", "file_name": "getFeature.py", "file_ext": "py", "file_size_in_byte": 3466, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.listdir", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "getData.Trainset", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 94, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 94, "usage_type": "name"}, {"api_name": "getData.Trainset", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "getData.Trainset", "line_number": 133, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 165, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 173, "usage_type": "call"}, {"api_name": "os.path", "line_number": 173, "usage_type": "attribute"}, {"api_name": "getData.Trainset", "line_number": 175, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 207, "usage_type": "call"}]}
+{"seq_id": "305191859", "text": "import os\nfrom flask import Flask\n\nfrom flashapp.debug import trace_print\nfrom flashapp.routes import add_routes\n\ntrace_print(f\"Importing app/__init__\")\n\n\ndef create_app(test_config=None):\n # create and configure the flashapp\n _app = Flask(__name__, instance_relative_config=True)\n _app.config.from_mapping(\n SECRET_KEY=\"dev\", DATABASE=os.path.join(_app.instance_path, \"flask.sqlite\"),\n )\n\n if test_config is None:\n # load the instance config, if it exists, when not testing\n _app.config.from_pyfile(\"config.py\", silent=True)\n else:\n # load the test config if passed in\n _app.config.from_mapping(test_config)\n\n # ensure the instance folder exists\n try:\n os.makedirs(_app.instance_path)\n except OSError:\n pass\n\n return _app\n\n\napp = create_app()\n\nadd_routes(app)\n", "sub_path": "flashapp/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 840, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "flashapp.debug.trace_print", "line_number": 7, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 26, "usage_type": "call"}, {"api_name": "flashapp.routes.add_routes", "line_number": 35, "usage_type": "call"}]}
+{"seq_id": "620732626", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nfrom django.conf import settings\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n migrations.swappable_dependency(settings.AUTH_USER_MODEL),\n ('coltrane', '0003_auto_20141229_1157'),\n ]\n\n operations = [\n migrations.AddField(\n model_name='entry',\n name='author',\n field=models.ForeignKey(to=settings.AUTH_USER_MODEL, default=1),\n preserve_default=False,\n ),\n migrations.AddField(\n model_name='entry',\n name='enable_comments',\n field=models.BooleanField(default=True),\n preserve_default=True,\n ),\n migrations.AddField(\n model_name='entry',\n name='featured',\n field=models.BooleanField(default=False),\n preserve_default=True,\n ),\n migrations.AddField(\n model_name='entry',\n name='slug',\n field=models.SlugField(help_text='Suggested value automatically generated from title. Must be unique.', default='', unique_for_date='pub_date'),\n preserve_default=False,\n ),\n migrations.AddField(\n model_name='entry',\n name='status',\n field=models.IntegerField(help_text='Only entries with live status will be publicly displayed.', default=1, choices=[(1, 'Live'), (2, 'Draft'), (3, 'Hidden')]),\n preserve_default=True,\n ),\n ]\n", "sub_path": "coltrane/migrations/0004_auto_20141229_1202.py", "file_name": "0004_auto_20141229_1202.py", "file_ext": "py", "file_size_in_byte": 1545, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.migrations.swappable_dependency", "line_number": 11, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 11, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 11, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 11, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 19, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 28, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 28, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 31, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 31, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 34, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 34, "usage_type": "name"}, {"api_name": "django.db.models.SlugField", "line_number": 37, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 37, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 40, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 40, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 43, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 43, "usage_type": "name"}]}
+{"seq_id": "613876641", "text": "from bot import DiscordWrapper\nfrom webWrapper import WebWrapper\nimport globalSettings\n\nimport asyncio\nimport logging\nfrom logging import handlers\nimport signal\nimport functools\nimport sys\nimport os\n\nhandler = handlers.TimedRotatingFileHandler(\"logs/log_ottobot.log\", when=\"midnight\", interval=1)\nlogging.basicConfig(format='%(asctime)s,%(msecs)d %(levelname)-8s [%(filename)s:%(lineno)d] %(message)s',\n filename=os.devnull,\n level=logging.INFO)\n_logger = logging.getLogger()\n_logger.addHandler(handler)\n\nensure_future = asyncio.ensure_future\n\nclass OttoBot:\n def __init__(self, token, prefix, connectionString, spamLimit, spamTimeout, display_response_id, broker_id, super_user_role, tip_verifier, exchange_rate, tip_command, broker_api_key):\n self.loop = asyncio.get_event_loop()\n self.web = WebWrapper(self.loop)\n self.discord = DiscordWrapper(token, self.web, prefix, connectionString, spamLimit, spamTimeout, display_response_id, broker_id, super_user_role, tip_verifier, exchange_rate, tip_command, broker_api_key)\n self.discord_task = None\n self.web_task = None\n self.response_checker_task = None\n self.status_updater_task = None\n self.shutdown_error = False\n self.do_shutdown = False\n \n def start(self):\n _logger.info(\"Starting OttoBot\")\n \n def begin_shutdown(signame):\n is_error = True if signame == \"SIGTERM\" else False\n msg = \"Shutting down bot due to signal: {}\".format(signame)\n if is_error:\n _logger.error(msg)\n else:\n _logger.info(msg)\n self.stop(is_error)\n \n # make sure the primary loop is interruptable\n for signame in (\"SIGINT\", \"SIGTERM\"):\n # windows doesn't have signals, so this will error if running on windows\n try:\n self.loop.add_signal_handler(getattr(signal, signame),functools.partial(begin_shutdown, signame))\n except NotImplementedError:\n _logger.info('Couldn\\'t set up signal handler for {}'.format(signame))\n pass\n\n self.discord_task = ensure_future(self.discord.start())\n self.web_task = ensure_future(self.web.run())\n self.response_checker_task = ensure_future(self.discord.check_pending_responses())\n if (globalSettings.config.get('DEFAULT', 'btc_status') == 'True'):\n self.status_updater_task = ensure_future(self.discord.start_status_updater())\n \n try:\n self.loop.run_until_complete(self.process())\n except asyncio.CancelledError:\n pass\n \n self.loop.close()\n sys.exit(self.shutdown_error)\n \n def stop(self, is_error=False):\n _logger.info(\"Stopping OttoBot\")\n self.shutdown_error = is_error\n self.do_shutdown = True\n \n if self.discord_task and not self.discord_task.done():\n ensure_future(self.discord.disconnect())\n \n async def process(self):\n task_list = [self.web_task, self.discord_task, self.response_checker_task]\n if self.status_updater_task:\n task_list.append(self.status_updater_task)\n while True:\n await asyncio.wait(task_list, return_when=asyncio.ALL_COMPLETED)\n if self.do_shutdown:\n break\n\n\ndef main():\n globalSettings.init()\n bot = OttoBot(globalSettings.config.get('DEFAULT', 'token'),\n globalSettings.config.get('DEFAULT', 'prefix'),\n globalSettings.config.get('DEFAULT', 'connectionString'),\n int(globalSettings.config.get('DEFAULT', 'spam_limit')),\n int(globalSettings.config.get('DEFAULT', 'spam_timeout')),\n globalSettings.config.get('DEFAULT', 'display_response_id') == 'True',\n globalSettings.config.get('DEFAULT', 'broker_id'),\n globalSettings.config.get('DEFAULT', 'super_user_role'),\n globalSettings.config.get('DEFAULT', 'tip_verifier_id'),\n globalSettings.config.get('DEFAULT', 'exchange_rate'),\n globalSettings.config.get('DEFAULT', 'tip_command'),\n globalSettings.config.get('DEFAULT', 'broker_api_key'))\n bot.start()\n\nmain()\n", "sub_path": "bot/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 4232, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.handlers.TimedRotatingFileHandler", "line_number": 13, "usage_type": "call"}, {"api_name": "logging.handlers", "line_number": 13, "usage_type": "name"}, {"api_name": "logging.basicConfig", "line_number": 14, "usage_type": "call"}, {"api_name": "os.devnull", "line_number": 15, "usage_type": "attribute"}, {"api_name": "logging.INFO", "line_number": 16, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 17, "usage_type": "call"}, {"api_name": "asyncio.ensure_future", "line_number": 20, "usage_type": "attribute"}, {"api_name": "asyncio.get_event_loop", "line_number": 24, "usage_type": "call"}, {"api_name": "webWrapper.WebWrapper", "line_number": 25, "usage_type": "call"}, {"api_name": "bot.DiscordWrapper", "line_number": 26, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 50, "usage_type": "call"}, {"api_name": "globalSettings.config.get", "line_number": 58, "usage_type": "call"}, {"api_name": "globalSettings.config", "line_number": 58, "usage_type": "attribute"}, {"api_name": "asyncio.CancelledError", "line_number": 63, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 67, "usage_type": "call"}, {"api_name": "asyncio.wait", "line_number": 82, "usage_type": "call"}, {"api_name": "asyncio.ALL_COMPLETED", "line_number": 82, "usage_type": "attribute"}, {"api_name": "globalSettings.init", "line_number": 88, "usage_type": "call"}, {"api_name": "globalSettings.config.get", "line_number": 89, "usage_type": "call"}, {"api_name": "globalSettings.config", "line_number": 89, "usage_type": "attribute"}, {"api_name": "globalSettings.config.get", "line_number": 90, "usage_type": "call"}, {"api_name": "globalSettings.config", "line_number": 90, "usage_type": "attribute"}, {"api_name": "globalSettings.config.get", "line_number": 91, "usage_type": "call"}, {"api_name": "globalSettings.config", "line_number": 91, "usage_type": "attribute"}, {"api_name": "globalSettings.config.get", "line_number": 92, "usage_type": "call"}, {"api_name": "globalSettings.config", "line_number": 92, "usage_type": "attribute"}, {"api_name": "globalSettings.config.get", "line_number": 93, "usage_type": "call"}, {"api_name": "globalSettings.config", "line_number": 93, "usage_type": "attribute"}, {"api_name": "globalSettings.config.get", "line_number": 94, "usage_type": "call"}, {"api_name": "globalSettings.config", "line_number": 94, "usage_type": "attribute"}, {"api_name": "globalSettings.config.get", "line_number": 95, "usage_type": "call"}, {"api_name": "globalSettings.config", "line_number": 95, "usage_type": "attribute"}, {"api_name": "globalSettings.config.get", "line_number": 96, "usage_type": "call"}, {"api_name": "globalSettings.config", "line_number": 96, "usage_type": "attribute"}, {"api_name": "globalSettings.config.get", "line_number": 97, "usage_type": "call"}, {"api_name": "globalSettings.config", "line_number": 97, "usage_type": "attribute"}, {"api_name": "globalSettings.config.get", "line_number": 98, "usage_type": "call"}, {"api_name": "globalSettings.config", "line_number": 98, "usage_type": "attribute"}, {"api_name": "globalSettings.config.get", "line_number": 99, "usage_type": "call"}, {"api_name": "globalSettings.config", "line_number": 99, "usage_type": "attribute"}, {"api_name": "globalSettings.config.get", "line_number": 100, "usage_type": "call"}, {"api_name": "globalSettings.config", "line_number": 100, "usage_type": "attribute"}, {"api_name": "bot.start", "line_number": 101, "usage_type": "call"}]}
+{"seq_id": "273644419", "text": "from django.shortcuts import render, redirect, get_object_or_404\nfrom django.utils import timezone\nfrom .models import Course, Comment, Like_course, Like, Profile, Tag\nfrom .forms import CourseForm, CommentForm, SignupForm, SigninForm, ProfileForm, TagForm\nfrom django.contrib.auth.models import User\nfrom django.contrib import auth\nfrom django.http.response import HttpResponseRedirect\nfrom django.urls.base import reverse\nfrom django.contrib.auth import login, authenticate\nfrom django.contrib.auth.decorators import login_required\nfrom django.contrib.auth.hashers import check_password # seohyun added\n# Create your views here.\n\ndef home(request):\n crs = Course.objects.all().order_by('-date')\n return render(request, 'sgapp/home.html', {'crs':crs})\n\ndef new(request):\n if request.method == 'POST':\n form = CourseForm(request.POST, request.FILES)\n tform = TagForm(request.POST)\n if form.is_valid():\n crs = form.save(commit=False)\n crs.author = request.user\n crs.save()\n if tform.is_valid():\n tag = tform.save()\n crs.tag_set.add(tag)\n return redirect('/'+str(crs.id))\n else:\n form = CourseForm()\n tag = TagForm()\n return render(request, 'sgapp/new.html',{'crs':form, 'tag':tag})\n\ndef detail(request, crs_id):\n crs = get_object_or_404(Course, pk=crs_id)\n form = CommentForm()\n cmt = Comment.objects.filter(crs=crs).order_by('-date')\n lk = Like_course.objects.filter(course=crs, author=request.user)\n alk = Like_course.objects.filter(course=crs)\n t = TagForm()\n return render(request, 'sgapp/detail.html', {'cmt':cmt,'crs':crs, 'cform':form, 'lk':lk, 'alk':alk, 't':t})\n\ndef edit(request, crs_id):\n crs = get_object_or_404(Course, pk=crs_id)\n if request.method == \"POST\":\n form = CourseForm(request.POST, instance=crs)\n if form.is_valid():\n c = form.save(commit=False)\n c.save()\n return redirect('detail', crs_id=c.id)\n else:\n form = CourseForm(instance=crs)\n return render(request, 'sgapp/edit.html', {'crs':form})\n\ndef delete(request, crs_id):\n crs = get_object_or_404(Course, pk=crs_id)\n crs.delete()\n return redirect('home')\n\ndef c_create(request, crs_id):\n crs = get_object_or_404(Course, pk=crs_id)\n cform = CommentForm(request.POST)\n score = request.POST['score']\n if cform.is_valid():\n cmt = cform.save(commit=False)\n cmt.score = score\n cmt.crs = crs\n cmt.author = request.user\n cmt.save()\n return redirect('detail', crs_id=crs.id)\n\ndef search(request):\n s = request.GET['search']\n c = request.GET['cate']\n sc = {f\"{c}__contains\":s}\n if s:\n crs = Course.objects.filter(**sc)\n return render(request, 'sgapp/search.html', {'crs':crs,'s':s})\n\ndef like(request, crs_id):\n crs = get_object_or_404(Course, pk=crs_id)\n like = Like(like=True)\n like.save()\n lc = Like_course(\n like = like,\n course = crs,\n author = request.user\n )\n lc.save()\n return redirect('detail', crs_id=crs.id)\n\ndef unlike(request, crs_id):\n crs = get_object_or_404(Course, pk=crs_id)\n like = get_object_or_404(Like_course, course=crs, author=request.user)\n like.like.delete()\n return redirect('detail', crs_id=crs.id)\n\ndef signup(request):#역시 GET/POST 방식을 사용하여 구현한다.\n if request.method == \"GET\":\n return render(request, 'sgapp/signup.html', {'f':SignupForm()} )\n \n elif request.method == \"POST\":\n form = SignupForm(request.POST)\n if form.is_valid():\n if form.cleaned_data['password'] == form.cleaned_data['password_check']:\n new_user = User.objects.create_user(form.cleaned_data['username'],form.cleaned_data['email'],form.cleaned_data['password'])\n new_user.last_name = form.cleaned_data['last_name']\n new_user.first_name = form.cleaned_data['first_name']\n new_user.save()\n auth.login(request, new_user)\n return HttpResponseRedirect(reverse('home'))\n else:\n return render(request, 'sgapp/signup.html',{'f':form, 'error':'비밀번호와 비밀번호 확인이 다릅니다.'})#password와 password_check가 다를 것을 대비하여 error를 지정해준다.\n return render(request, 'sgapp/signup.html',{'f':form})\n \ndef signin(request):#로그인 기능\n if request.method == \"GET\":\n return render(request, 'sgapp/signin.html', {'f':SigninForm()} )\n\n elif request.method == \"POST\":\n form = SigninForm(request.POST)\n id = request.POST.get('username')\n pw = request.POST.get('password')\n u = authenticate(username=id, password=pw)\n if u: #u에 특정 값이 있다면\n login(request, user=u) #u 객체로 로그인해라\n return HttpResponseRedirect(reverse('home'))\n else:\n return render(request, 'sgapp/signin.html',{'f':form, 'error':'아이디나 비밀번호가 일치하지 않습니다.'})\n\nfrom django.contrib.auth import logout #logout을 처리하기 위해 선언\n\ndef signout(request): #logout 기능\n logout(request) #logout을 수행한다.\n return HttpResponseRedirect(reverse('signin'))\n\ndef mypage(request):\n pf = Profile.objects.all()\n crs = Course.objects.filter(author=request.user).order_by('-date')\n if request.method == \"POST\":\n form = ProfileForm(request.POST, request.FILES)\n if form.is_valid():\n form.save()\n pform = ProfileForm()\n return render(request, 'sgapp/mypage.html', {'pf':pf, 'pform':pform, 'crs':crs})\n\n else:\n form = ProfileForm()\n return render(request, 'sgapp/mypage.html', {'pf':pf, 'form':form, 'crs':crs})\n\ndef change_pw(request): #비밀번호 변경 기능\n context= {}\n if request.method == \"POST\":\n current_pw = request.POST.get(\"current_pw\")\n user = request.user\n if check_password(current_pw,user.password):\n new_pw = request.POST.get(\"new_pw\")\n new_pw_check = request.POST.get(\"new_pw_check\")\n if new_pw == new_pw_check:\n user.set_password(new_pw)\n user.save()\n auth.login(request,user)\n return redirect('mypage')\n else:\n context.update({'error':\"새로운 비밀번호를 다시 확인해주세요.\"})\n else:\n context.update({'error':\"현재 비밀번호가 일치하지 않습니다.\"})\n\n return render(request, 'sgapp/change_pw.html', context)\n#def create(request):\n # 생략\n #profile.photo = request.FILES['photo']", "sub_path": "sgapp/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 6739, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "models.Course.objects.all", "line_number": 15, "usage_type": "call"}, {"api_name": "models.Course.objects", "line_number": 15, "usage_type": "attribute"}, {"api_name": "models.Course", "line_number": 15, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 16, "usage_type": "call"}, {"api_name": "forms.CourseForm", "line_number": 20, "usage_type": "call"}, {"api_name": "forms.TagForm", "line_number": 21, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 29, "usage_type": "call"}, {"api_name": "forms.CourseForm", "line_number": 31, "usage_type": "call"}, {"api_name": "forms.TagForm", "line_number": 32, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 33, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 36, "usage_type": "call"}, {"api_name": "models.Course", "line_number": 36, "usage_type": "argument"}, {"api_name": "forms.CommentForm", "line_number": 37, "usage_type": "call"}, {"api_name": "models.Comment.objects.filter", "line_number": 38, "usage_type": "call"}, {"api_name": "models.Comment.objects", "line_number": 38, "usage_type": "attribute"}, {"api_name": "models.Comment", "line_number": 38, "usage_type": "name"}, {"api_name": "models.Like_course.objects.filter", "line_number": 39, "usage_type": "call"}, {"api_name": "models.Like_course.objects", "line_number": 39, "usage_type": "attribute"}, {"api_name": "models.Like_course", "line_number": 39, "usage_type": "name"}, {"api_name": "models.Like_course.objects.filter", "line_number": 40, "usage_type": "call"}, {"api_name": "models.Like_course.objects", "line_number": 40, "usage_type": "attribute"}, {"api_name": "models.Like_course", "line_number": 40, "usage_type": "name"}, {"api_name": "forms.TagForm", "line_number": 41, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 42, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 45, "usage_type": "call"}, {"api_name": "models.Course", "line_number": 45, "usage_type": "argument"}, {"api_name": "forms.CourseForm", "line_number": 47, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 51, "usage_type": "call"}, {"api_name": "forms.CourseForm", "line_number": 53, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 54, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 57, "usage_type": "call"}, {"api_name": "models.Course", "line_number": 57, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 59, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 62, "usage_type": "call"}, {"api_name": "models.Course", "line_number": 62, "usage_type": "argument"}, {"api_name": "forms.CommentForm", "line_number": 63, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 71, "usage_type": "call"}, {"api_name": "models.Course.objects.filter", "line_number": 78, "usage_type": "call"}, {"api_name": "models.Course.objects", "line_number": 78, "usage_type": "attribute"}, {"api_name": "models.Course", "line_number": 78, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 79, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 82, "usage_type": "call"}, {"api_name": "models.Course", "line_number": 82, "usage_type": "argument"}, {"api_name": "models.Like", "line_number": 83, "usage_type": "call"}, {"api_name": "models.Like_course", "line_number": 85, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 91, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 94, "usage_type": "call"}, {"api_name": "models.Course", "line_number": 94, "usage_type": "argument"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 95, "usage_type": "call"}, {"api_name": "models.Like_course", "line_number": 95, "usage_type": "argument"}, {"api_name": "django.shortcuts.redirect", "line_number": 97, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 101, "usage_type": "call"}, {"api_name": "forms.SignupForm", "line_number": 101, "usage_type": "call"}, {"api_name": "forms.SignupForm", "line_number": 104, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.create_user", "line_number": 107, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 107, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 107, "usage_type": "name"}, {"api_name": "django.contrib.auth.login", "line_number": 111, "usage_type": "call"}, {"api_name": "django.contrib.auth", "line_number": 111, "usage_type": "name"}, {"api_name": "django.http.response.HttpResponseRedirect", "line_number": 112, "usage_type": "call"}, {"api_name": "django.urls.base.reverse", "line_number": 112, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 114, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 115, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 119, "usage_type": "call"}, {"api_name": "forms.SigninForm", "line_number": 119, "usage_type": "call"}, {"api_name": "forms.SigninForm", "line_number": 122, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 125, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 127, "usage_type": "call"}, {"api_name": "django.http.response.HttpResponseRedirect", "line_number": 128, "usage_type": "call"}, {"api_name": "django.urls.base.reverse", "line_number": 128, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 130, "usage_type": "call"}, {"api_name": "django.contrib.auth.logout", "line_number": 135, "usage_type": "call"}, {"api_name": "django.http.response.HttpResponseRedirect", "line_number": 136, "usage_type": "call"}, {"api_name": "django.urls.base.reverse", "line_number": 136, "usage_type": "call"}, {"api_name": "models.Profile.objects.all", "line_number": 139, "usage_type": "call"}, {"api_name": "models.Profile.objects", "line_number": 139, "usage_type": "attribute"}, {"api_name": "models.Profile", "line_number": 139, "usage_type": "name"}, {"api_name": "models.Course.objects.filter", "line_number": 140, "usage_type": "call"}, {"api_name": "models.Course.objects", "line_number": 140, "usage_type": "attribute"}, {"api_name": "models.Course", "line_number": 140, "usage_type": "name"}, {"api_name": "forms.ProfileForm", "line_number": 142, "usage_type": "call"}, {"api_name": "forms.ProfileForm", "line_number": 145, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 146, "usage_type": "call"}, {"api_name": "forms.ProfileForm", "line_number": 149, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 150, "usage_type": "call"}, {"api_name": "django.contrib.auth.hashers.check_password", "line_number": 157, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 163, "usage_type": "call"}, {"api_name": "django.contrib.auth", "line_number": 163, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 164, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 170, "usage_type": "call"}]}
+{"seq_id": "71558221", "text": "# -*- coding: utf-8 -*-\n\"\"\"\n@author: imalenica\n\"\"\"\nfrom collections import Counter\nfrom collections import defaultdict\nimport numpy as np\nimport pandas as pd\nfrom sklearn.base import TransformerMixin\nfrom pprint import pprint\nfrom sklearn.feature_extraction import DictVectorizer\nfrom sklearn.utils import shuffle\n\nclass DataFrameImputer(TransformerMixin):\n\n def __init__(self):\n \"\"\" \n Impute missing values.\n \"\"\"\n def fit(self, X, y=None):\n\n self.fill = pd.Series([X[c].value_counts().index[0]\n if X[c].dtype == np.dtype('O') else X[c].mean() for c in X],\n index=X.columns)\n\n return self\n\n def transform(self, X, y=None):\n return X.fillna(self.fill)\n \n#Preprocess data:\n#1)Deal with categorical variables (map categories to binary variables)\n#2)Missing values\n\n#################################\n# Process the test data \n#################################\n\nNumColumns = []\nCatColumns = []\nfor col in test_datacsv.columns:\n if (test_datacsv[col].dtype == np.object):\n CatColumns.append(col)\n else:\n NumColumns.append(col)\n \n#Processes numeric columns\ntest_dataNum=test_datacsv.loc[:,NumColumns]\n\n#Processes categorical columns\ntest_dataCat=test_datacsv.loc[:,CatColumns]\ntest_dataCat2=test_dataCat.replace('?',np.NaN)\ntest_dataCat= DataFrameImputer().fit_transform(test_dataCat2)\n\ntest_as_dicts = [dict(r.iteritems()) for _, r in test_dataCat.iterrows()]\nv = DictVectorizer(sparse=False)\ntest_dataCat_2=v.fit_transform(test_as_dicts)\n\nnames=defaultdict(list)\n\nfor i in range(0,len(test_dataCat.columns)):\n\n name=sorted(pd.unique(test_dataCat.ix[:,i].ravel()))\n \n names[i].append(name)\n\n#Rename the dictionary key; it will automatically sort the values\nnames['workclass'] = names.pop(0)\nnames['education'] = names.pop(1)\nnames['marital-status'] = names.pop(2)\nnames['occupation'] = names.pop(3)\nnames['relationship'] = names.pop(4)\nnames['race'] = names.pop(5)\nnames['sex'] = names.pop(6)\nnames['native-country'] = names.pop(7)\n\nheader=names['education'] + names['marital-status'] + names['native-country'] + names['occupation'] + names['race'] + names['relationship'] + names['sex'] + names['workclass']\n \nimport itertools\nmerged_header = list(itertools.chain.from_iterable(header))\n\ntest_dataCat=pd.DataFrame(data=test_dataCat_2[0:,0:],index=test_dataNum.index.values,columns=merged_header)\n \nn='not'\nh=list(test_dataCat.columns.values)[0]\n \ntest_dataCat2=test_dataCat.ix[:,0].replace(1,h) \ntest_dataCat2=test_dataCat2.replace(0,n+h)\ntest_dataCat_fin=test_dataCat2.to_frame()\n\nfor i in range(1,len(test_dataCat.columns)): \n n='not'\n h=list(test_dataCat.columns.values)[i]\n \n test_dataCat2=test_dataCat.ix[:,i].replace(1,h) \n test_dataCat2=test_dataCat2.replace(0,n+h)\n test_dataCat2=test_dataCat2.to_frame()\n \n test_dataCat_fin=test_dataCat_fin.join(test_dataCat2)\n\n#Combine Numerical and Categorical data\ntest_data_fin=test_dataNum.join(test_dataCat_fin)\n\nheader_test=list(test_data_fin.columns.values)\n\ntest_data_fin2=test_data_fin.values.tolist()\n\ntemp = np.zeros((1, len(test_data_fin2)))\nd = pd.DataFrame(temp)\nd=d.T\nd.columns=['Holand-Netherlands']\nd=d.replace(0,'notHoland-Netherlands')\n\ntest_data_final=test_data_fin.join(d)\n\ntest_data_fin=test_data_final.values.tolist()\n\n#################################\n# Process the training data \n#################################\nimport random\nrandom.seed(100)\n\nNumColumns = []\nCatColumns = []\nfor col in train_datacsv.columns:\n if (train_datacsv[col].dtype == np.object):\n CatColumns.append(col)\n else:\n NumColumns.append(col)\n \n#Processes numeric columns\ntrain_dataNum=train_datacsv.loc[:,NumColumns]\n\n#Processes categorical columns\ntrain_dataCat=train_datacsv.loc[:,CatColumns]\ntrain_dataCat2=train_dataCat.replace('?',np.NaN)\ntrain_dataCat= DataFrameImputer().fit_transform(train_dataCat2)\n\ntrain_as_dicts = [dict(r.iteritems()) for _, r in train_dataCat.iterrows()]\nv = DictVectorizer(sparse=False)\ntrain_dataCat_2=v.fit_transform(train_as_dicts)\n\nnames=defaultdict(list)\n\nfor i in range(0,len(train_dataCat.columns)):\n\n name=sorted(pd.unique(train_dataCat.ix[:,i].ravel()))\n \n names[i].append(name)\n\n#Rename the dictionary key; it will automatically sort the values\nnames['workclass'] = names.pop(0)\nnames['education'] = names.pop(1)\nnames['marital-status'] = names.pop(2)\nnames['occupation'] = names.pop(3)\nnames['relationship'] = names.pop(4)\nnames['race'] = names.pop(5)\nnames['sex'] = names.pop(6)\nnames['native-country'] = names.pop(7)\n\nheader=names['education'] + names['marital-status'] + names['native-country'] + names['occupation'] + names['race'] + names['relationship'] + names['sex'] + names['workclass']\n\nimport itertools\nmerged_header = list(itertools.chain.from_iterable(header))\n\ntrain_dataCat=pd.DataFrame(data=train_dataCat_2[0:,0:],index=train_dataNum.index.values,columns=merged_header)\n \nn='not'\nh=list(train_dataCat.columns.values)[0]\n \ntrain_dataCat2=train_dataCat.ix[:,0].replace(1,h) \ntrain_dataCat2=train_dataCat2.replace(0,n+h)\ntrain_dataCat_fin=train_dataCat2.to_frame()\n\nfor i in range(1,len(train_dataCat.columns)): \n n='not'\n h=list(train_dataCat.columns.values)[i]\n \n train_dataCat2=train_dataCat.ix[:,i].replace(1,h) \n train_dataCat2=train_dataCat2.replace(0,n+h)\n train_dataCat2=train_dataCat2.to_frame()\n \n train_dataCat_fin=train_dataCat_fin.join(train_dataCat2)\n\n#Combine Numerical and Categorical data\ntrain_data_fin=train_dataNum.join(train_dataCat_fin)\nheader_train=list(train_data_fin.columns.values)\n\ntrain_data_fin2=train_data_fin.values.tolist()\n\ntrain_data_fin2 = shuffle(train_data_fin2, random_state=0)\n\n#Functions:\n\nfrom math import log\n\ndef counts(samples, outcome_index=0):\n res={}\n for sample in samples:\n i=sample[outcome_index]\n if i not in res:\n res[i]=0\n res[i]+=1\n return res \n \ndef entropy(samples, outcome_index=0):\n res=counts(samples, outcome_index)\n # Now calculate the entropy\n entr=0.0\n for i in res.keys():\n # current probability of class\n prob=res[i]/len(samples) \n entr=entr-prob*log(prob,2)\n return entr\n\ndef divide2(samples,pred,value):\n \n split=None\n \n if isinstance(value,int) or isinstance(value,float):\n def split(sample):\n return sample[pred]>value\n\n else:\n def split(sample):\n return sample[pred]==value\n \n left=[sample for sample in samples if split(sample)]\n right=[sample for sample in samples if not split(sample)]\n return (left,right) \n \ndef info2(samples, outcome_index=0):\n\n current_score = entropy(samples,outcome_index)\n\n best_gain = 0.0\n best_pred = None\n best_val = None\n best_sets = None\n \n column_count = len(samples[0])\n \n for col in range(1, column_count):\n # find different values in this column\n column_values = set([sample[col] for sample in samples])\n\n # for each possible value, try to divide on that value\n for value in column_values:\n \n set1, set2 = divide2(samples, col, value)\n\n # Information gain\n gain = current_score-(len(set1)*entropy(set1)+len(set2)*entropy(set2))/len(samples)\n \n if gain > best_gain and len(set1) > 0 and len(set2) > 0:\n best_gain = gain\n best_pred = col\n best_val = value\n best_sets = (set1, set2) \n\n return best_gain,best_pred,best_val\n\ndef majority(data, outcome_index=0):\n counts = Counter([i[outcome_index] for i in data])\n return counts.most_common(1)[0][0]\n \ndef divide(samples,pred,value):\n \n split=None\n \n if isinstance(value,int) or isinstance(value,float):\n \n def split(sample):\n return sample[pred]>value\n \n right=[sample for sample in samples if split(sample)]\n left=[sample for sample in samples if not split(sample)]\n \n left_dict=defaultdict(list)\n right_dict=defaultdict(list)\n \n for i in right:\n right_dict[value+1].append(i)\n\n for i in left:\n left_dict[value-1].append(i)\n \n splits = {**left_dict, **right_dict} \n \n return splits\n \n else:\n def split(sample):\n return sample[pred]==value\n \n left=[sample for sample in samples if split(sample)]\n right=[sample for sample in samples if not split(sample)]\n \n left_dict=defaultdict(list)\n right_dict=defaultdict(list)\n \n other=right[0][pred]\n \n for i in right:\n right_dict[other].append(i)\n\n for i in left:\n left_dict[value].append(i)\n \n splits = {**left_dict, **right_dict} \n \n return splits \n\ndef create_tree(data,pred_inx=None,out_inx=0,default=None):\n \n class_labels_and_counts = Counter([instance[out_inx] for instance in data])\n \n column_count = len(data[0])\n classes=[]\n \n for col in range(1, column_count):\n num=Counter([instance[col] for instance in data])\n n=[len(num)]\n classes.extend(n)\n \n if sum(classes) == (column_count-1) or len(class_labels_and_counts) == 1:\n class_label = class_labels_and_counts.most_common(1)[0][0]\n return class_label\n \n else:\n if info2(data)[0]== 0.0:\n class_label = class_labels_and_counts.most_common(1)[0][0]\n return class_label\n \n else:\n best_index = info2(data)[1] \n best_value = info2(data)[2]\n tree = {best_index:{}}\n partitions = divide(data, best_index, best_value)\n \n #Tree pruning\n for attribute_value in partitions:\n if len(partitions[attribute_value])<11:\n tree[best_index][attribute_value]=majority(partitions[attribute_value], out_inx)\n \n else:\n #Note: when calling create_tree again, now we are working with a subset of data corresponding to a leaf...\n subtree = create_tree(partitions[attribute_value],pred_inx=None,out_inx=0,default=None)\n tree[best_index][attribute_value] = subtree\n \n return tree\n\ndef predict(tree, instance, data):\n \n if not isinstance(tree, dict):\n return tree\n \n attribute_index = list(tree.keys())[0] \n attribute_values = list(tree.values())[0]\n \n instance_attribute_value = instance[attribute_index]\n \n if isinstance(list(attribute_values.keys())[0],str):\n if instance_attribute_value not in attribute_values:\n return majority(data, attribute_index)\n else: \n return predict(attribute_values[instance_attribute_value],instance,data)\n \n else:\n num=list(attribute_values.keys())[0]+1\n \n less=list(attribute_values.keys())[0]\n more=list(attribute_values.keys())[1]\n \n if instance_attribute_value>=num:\n return predict(attribute_values[more],instance,data)\n else:\n return predict(attribute_values[less],instance,data) \n \nclass DecisionTree:\n\n _tree = {} \n\n def __init__(self):\n pass\n \n def fit(self, data, pred_inx=None,out_inx=0,default=None):\n if not pred_inx:\n pred_inx = [i for i in range(len(data[0])) if i != out_inx]\n self._tree = self._create_tree(data,pred_inx,out_inx,default)\n \n def _create_tree(self,data,pred_inx=None,out_inx=0,default=None):\n \n class_labels_and_counts = Counter([instance[out_inx] for instance in data])\n \n column_count = len(data[0])\n classes=[]\n \n for col in range(1, column_count):\n num=Counter([instance[col] for instance in data])\n n=[len(num)]\n classes.extend(n)\n \n if sum(classes) == (column_count-1) or len(class_labels_and_counts) == 1:\n class_label = class_labels_and_counts.most_common(1)[0][0]\n return class_label\n \n else:\n if info2(data)[0]== 0.0:\n class_label = class_labels_and_counts.most_common(1)[0][0]\n return class_label\n \n else:\n best_index = info2(data)[1] \n best_value = info2(data)[2]\n tree = {best_index:{}}\n partitions = divide(data, best_index, best_value)\n \n #Tree pruning\n for attribute_value in partitions:\n \n if len(partitions[attribute_value])<11:\n tree[best_index][attribute_value]=majority(partitions[attribute_value], out_inx)\n \n else:\n #Note: when calling create_tree again, now we are working with a subset of data corresponding to a leaf...\n subtree = self._create_tree(partitions[attribute_value],pred_inx=None,out_inx=0,default=None)\n tree[best_index][attribute_value] = subtree\n \n return tree \n \n #Parse samples \n def predict(self, test, data):\n \n return [self._predict(self._tree, instance, data) for instance in test]\n \n #Predict samples one at a time \n def _predict(self, tree, instance, data):\n \n if not isinstance(tree, dict):\n return tree\n \n attribute_index = list(tree.keys())[0] \n attribute_values = list(tree.values())[0]\n \n instance_attribute_value = instance[attribute_index]\n \n if isinstance(list(attribute_values.keys())[0],str):\n if instance_attribute_value not in attribute_values:\n return majority(data, attribute_index)\n else: \n return self._predict(attribute_values[instance_attribute_value],instance,data)\n \n else:\n num=list(attribute_values.keys())[0]+1\n \n less=list(attribute_values.keys())[0]\n more=list(attribute_values.keys())[1]\n \n if instance_attribute_value>=num:\n return self._predict(attribute_values[more],instance,data)\n else:\n return self._predict(attribute_values[less],instance,data)\n \n def accuracy(self, data, default=None):\n predicted_labels = self.predict(data, default)\n actual_labels = [x[0] for x in data]\n counts = Counter([x == y for x, y in zip(predicted_labels, actual_labels)])\n return counts[True]/len(data), counts[True], counts[False]\n\n def graph(self):\n pprint(self._tree)\n \n\n#Create test and training datasets\ntrain_data_testing=train_data_fin2[1000:2000]\ntrain_data_test=train_data_fin2[10000:11000]\n \ndecision_tree = DecisionTree()\ndecision_tree.fit(train_data_testing)\ndecision_tree.graph()\ndecision_tree.predict(train_data_test,train_data_testing)\ndecision_tree.accuracy(train_data_test)\n\n###################\n#Random forests\n###################\n\ntrain=train_dataNum.join(train_dataCat)\ntrain_data=train.values.tolist()\n\ntest=test_dataNum.join(test_dataCat)\ntest_data=test.values.tolist()\n\nimport random\nnum=len(train_data_fin)\npred=len(train_data_fin.columns)\n\n#Create master file for all the samples tested\nall_samples=list(range(0, num))\ncolumns = ['Prediction']\n\nAll_predictions=pd.DataFrame(index=all_samples, columns=columns) \nAll_predictions = All_predictions.fillna(0) \n\n#Create master file counter\nall_counter=list(range(0, num))\ncolumns = ['Counter']\n\nAll_counter=pd.DataFrame(index=all_counter, columns=columns) \nAll_counter = All_counter.fillna(0) \n\n#Create master predictor counter\nall_predictors=list(range(0, pred))\ncolumns = ['Predictor']\n\nAll_predictors=pd.DataFrame(index=all_predictors, columns=columns) \nAll_predictors = All_predictors.fillna(0) \n\n#data is a list with all the samples and all the predictors. \ndef RandomForest(data,train,test,pred):\n \n x=np.random.uniform(0,len(data)-train,1) \n x=[int(i) for i in x]\n \n lowLim=x[0]\n highLim=x[0]+train\n\n train_data=data[lowLim:highLim]\n \n x1=np.random.uniform(0,len(data)-test,1) \n x1=[int(i) for i in x1]\n \n lowLim_test=x1[0]\n highLim_test=x1[0]+test\n\n test_data=data[lowLim_test:highLim_test]\n \n train_data=np.matrix(train_data)\n test_data=np.matrix(test_data)\n \n x3=random.sample(range(1,106),pred-1) \n \n lab=[0]\n lab.extend(x3)\n \n data_pred_subset=train_data[:,lab]\n\n data_both_subset=pd.DataFrame(data_pred_subset, columns=lab)\n data_both_subset_test=pd.DataFrame(test_data)\n\n data_both_subset=data_both_subset.convert_objects(convert_numeric=True)\n data_both_subset_test=data_both_subset_test.convert_objects(convert_numeric=True)\n\n data_subset_fin=data_both_subset.values.tolist()\n data_subset_fin_test=data_both_subset_test.values.tolist()\n \n ref=data_both_subset.columns.values.tolist()\n \n tree=create_tree(data_subset_fin,pred_inx=None,out_inx=0,default=None)\n predictions=[predict(tree, instance, data_subset_fin) for instance in data_subset_fin_test]\n \n if isinstance(tree,int):\n rootNode=0\n else:\n rootNode=list(tree.keys())[0]\n \n rootNode_actually=ref[rootNode]\n \n samples=[]\n nonUsedSamples1=[]\n nonUsedSamples2=[]\n nonUsedPreds1=[]\n nonUsedPreds2=[]\n\n samples=list(range(lowLim_test, highLim_test))\n nonUsedSamples1=list(range(0, lowLim_test))\n nonUsedSamples2=list(range(highLim_test, len(data)))\n nonUsedPreds1=list(range(0,rootNode_actually))\n nonUsedPreds2=list(range(rootNode_actually+1,106))\n \n tree_predictors=[]\n \n listofzeros1 = [0] * len(nonUsedPreds1)\n listofzeros2 = [0] * len(nonUsedPreds2)\n listofones = [1] * len([rootNode_actually])\n \n tree_predictors.extend(listofzeros1)\n tree_predictors.extend(listofones)\n tree_predictors.extend(listofzeros2)\n \n tree_samples=[]\n\n tree_samples.extend(nonUsedSamples1)\n tree_samples.extend(samples)\n tree_samples.extend(nonUsedSamples2)\n\n listofzeros1 = [0] * len(nonUsedSamples1)\n listofzeros2 = [0] * len(nonUsedSamples2)\n listofones = [1] * len(predictions)\n \n tree_predictions=[]\n\n tree_predictions.extend(listofzeros1)\n tree_predictions.extend(predictions)\n tree_predictions.extend(listofzeros2)\n\n tree_counts=[]\n\n tree_counts.extend(listofzeros1)\n tree_counts.extend(listofones)\n tree_counts.extend(listofzeros2)\n \n columns = ['Counter']\n Tree_count=pd.DataFrame(tree_counts,index=tree_samples, columns=columns)\n\n columns = ['Prediction']\n Tree_pred=pd.DataFrame(tree_predictions,index=tree_samples, columns=columns)\n \n columns = ['Predictor']\n Tree_predictors=pd.DataFrame(tree_predictors, columns=columns)\n\n return Tree_count, Tree_pred, Tree_predictors\n\nfor i in range(1000):\n \n Tree_count,Tree_pred, Tree_predictors = RandomForest(train_data,1000,10000,11)\n \n All_counter=All_counter.add(Tree_count)\n All_predictions=All_predictions.add(Tree_pred)\n All_predictors=All_predictors.add(Tree_predictors)\n\nresult=All_predictions['Prediction']/All_counter['Counter']\nresult2=result.tolist()\n\nresult3=[\"%.0f\" % a for a in result2]\n\npredicted_labels=[int(i) for i in result3]\nactual_labels = [x[0] for x in train_data_fin2]\ncounts = Counter([x == y for x, y in zip(predicted_labels, actual_labels)])\ncounts[True]/len(data)\n\ncounts[True]\ncounts[False]\n", "sub_path": "RandomForests.py", "file_name": "RandomForests.py", "file_ext": "py", "file_size_in_byte": 19876, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sklearn.base.TransformerMixin", "line_number": 14, "usage_type": "name"}, {"api_name": "pandas.Series", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.dtype", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.object", "line_number": 42, "usage_type": "attribute"}, {"api_name": "numpy.NaN", "line_number": 52, "usage_type": "attribute"}, {"api_name": "sklearn.feature_extraction.DictVectorizer", "line_number": 56, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 59, "usage_type": "call"}, {"api_name": "pandas.unique", "line_number": 63, "usage_type": "call"}, {"api_name": "itertools.chain.from_iterable", "line_number": 80, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 80, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 108, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 109, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.object", "line_number": 127, "usage_type": "attribute"}, {"api_name": "numpy.NaN", "line_number": 137, "usage_type": "attribute"}, {"api_name": "sklearn.feature_extraction.DictVectorizer", "line_number": 141, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 144, "usage_type": "call"}, {"api_name": "pandas.unique", "line_number": 148, "usage_type": "call"}, {"api_name": "itertools.chain.from_iterable", "line_number": 165, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 165, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 167, "usage_type": "call"}, {"api_name": "sklearn.utils.shuffle", "line_number": 192, "usage_type": "call"}, {"api_name": "math.log", "line_number": 214, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 265, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 280, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 281, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 300, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 301, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 317, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 323, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 395, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 401, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 469, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 473, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 504, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 511, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 518, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 524, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 524, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 532, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 532, "usage_type": "attribute"}, {"api_name": "numpy.matrix", "line_number": 540, "usage_type": "call"}, {"api_name": "numpy.matrix", "line_number": 541, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 543, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 550, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 551, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 616, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 619, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 622, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 641, "usage_type": "call"}]}
+{"seq_id": "565155899", "text": "from django.shortcuts import render, HttpResponse, redirect\nfrom django.contrib import messages\nfrom .models import User\nfrom time import gmtime, strftime\n\ndef index(request):\n # get all users from the database\n users = User.objects.values().all()\n return render(request, 'crud_users/users.html', {'users': users})\n\ndef new(request):\n return render(request, 'crud_users/new.html')\n\ndef edit(request, id):\n # get the user info from the database\n u = User.objects.get(id = id)\n user = {\n 'id': id,\n 'first_name': u.first_name,\n 'last_name': u.last_name,\n 'email': u.email,\n 'created_at': u.created_at\n }\n return render(request, 'crud_users/edit.html', user)\n\ndef show(request, id):\n # get the user info from the database\n u = User.objects.get(id = id)\n user = {\n 'id': id,\n 'first_name': u.first_name,\n 'last_name': u.last_name,\n 'email': u.email,\n 'created_at': u.created_at\n }\n return render(request, 'crud_users/show.html', user)\n\n\ndef destroy(request, id):\n # get the user info from the database\n u = User.objects.get(id = id)\n u.delete()\n messages.success(request, \"A user was delete from database\")\n return redirect('/users')\n\ndef update(request):\n errors = User.objects.basic_validator(request.POST)\n\n # check if any error exist\n if len(errors):\n for key, value in errors.items():\n messages.error(request, value)\n\n # redirect to the form to display the message\n return redirect('/users/'+request.POST['id']+'/edit')\n else:\n # No erros so send the data to the database\n\n #first the the user info that needed to be update\n user = User.objects.get(id = request.POST['id'])\n user.first_name = request.POST['first_name']\n user.last_name = request.POST['last_name']\n user.email = request.POST['email']\n user.save()\n\n # Redirect witht the new user id\n messages.success(request, 'This user info was was updated!')\n return redirect('/users/'+str(user.id))\n\ndef create(request):\n user = User.objects\n user.create(first_name = request.POST['first_name'], last_name = request.POST['last_name'], email = request.POST['email'])\n\n # Redirect witht the new user id\n return render(request, 'crud_users/all.html', {'users': User.objects.values().all() } )", "sub_path": "AJAX/users_ajax/apps/crud_users/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2395, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "models.User.objects.values", "line_number": 8, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 8, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 8, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 9, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 12, "usage_type": "call"}, {"api_name": "models.User.objects.get", "line_number": 16, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 16, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 16, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 24, "usage_type": "call"}, {"api_name": "models.User.objects.get", "line_number": 28, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 28, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 28, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 36, "usage_type": "call"}, {"api_name": "models.User.objects.get", "line_number": 41, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 41, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 41, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 43, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 43, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 44, "usage_type": "call"}, {"api_name": "models.User.objects.basic_validator", "line_number": 47, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 47, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 47, "usage_type": "name"}, {"api_name": "django.contrib.messages.error", "line_number": 52, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 52, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 55, "usage_type": "call"}, {"api_name": "models.User.objects.get", "line_number": 60, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 60, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 60, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 67, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 67, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 68, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 71, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 71, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 75, "usage_type": "call"}, {"api_name": "models.User.objects.values", "line_number": 75, "usage_type": "call"}, {"api_name": "models.User.objects", "line_number": 75, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 75, "usage_type": "name"}]}
+{"seq_id": "397733148", "text": "# -*- coding: utf-8 -*-\n# Copyright (c) 2016-present, CloudZero, Inc. All rights reserved.\n# Licensed under the BSD-style license. See LICENSE file in the project root for full license information.\nimport os\n\nimport json\nimport time\n\nfrom datetime import datetime\nfrom dateutil.tz import tzlocal # tzlocal vs datetime.timezone for py 2.7 compatibility\nimport boto3\nimport botocore.credentials\nimport botocore.session\nimport botocore.exceptions\nimport pkg_resources\n\nfrom cloudzerocli.constants import AwsProviderConnectionType, AWS_RESOURCE_SERVICES\nfrom cloudzerocli.settings import Settings\nfrom cloudzerocli.utils.formatters import create_random_string, render_template\n\n# Load settings\n# TODO: pass this in somehow rather than instantiating it at module load time\nsettings = Settings()\n\nCLOUDTRAIL_BUCKET_LIFECYCLE_POLICY = {\n 'Rules': [\n {\n 'Expiration': {\n 'Days': settings.defaults.aws.cloudtrail_bucket.max_age\n },\n 'ID': 'CloudZeroLifeCycleRules20170109',\n 'Status': 'Enabled',\n 'NoncurrentVersionExpiration': {\n 'NoncurrentDays': settings.defaults.aws.cloudtrail_bucket.max_age\n },\n 'Transitions': [\n {\n 'Days': settings.defaults.aws.cloudtrail_bucket.transition_to_infrequent_access,\n 'StorageClass': 'STANDARD_IA'\n },\n ],\n 'Prefix': '',\n 'AbortIncompleteMultipartUpload': {\n 'DaysAfterInitiation': 7\n }\n },\n ]\n}\n\n\ndef get_reactor_api_keys(aws_session):\n try:\n apigateway = aws_session.connection('apigateway')\n response = apigateway.client.get_api_keys(nameQuery='cloudzero-reactor',\n includeValues=True)\n return response['items']\n except botocore.exceptions.ClientError:\n return []\n\n\ndef get_reactor_encryption_keys(aws_session):\n try:\n kms = aws_session.connection('kms')\n keys = kms.client.list_keys()['Keys']\n aliases = kms.client.list_aliases()['Aliases']\n\n # One day Python 3 will rule them all\n # return [{**keys[x], **aliases[x]} for x in range(len(keys))]\n # until then....\n for x in range(len(keys)):\n keys[x].update(aliases[x])\n return keys\n\n except botocore.exceptions.ClientError:\n return []\n\n\ndef get_name_from_arn(arn):\n return arn.split(':')[-1]\n\n\ndef get_account_id(arn):\n \"\"\"\n Return the AWS account id from a given arn\n\n Args:\n arn (str):\n\n Returns:\n str : AWS Account ID\n\n \"\"\"\n try:\n return str(arn.split(\":\")[4])\n except (IndexError, AttributeError, ValueError):\n return None\n\n\ndef generate_assume_role_policy(account_id_to_trust, external_id):\n policy_template = load_policy_document('assume-role-policy-template.json')\n new_policy = render_template(policy_template, {'external_id': external_id,\n 'account_id_to_trust': account_id_to_trust})\n return new_policy\n\n\ndef generate_external_id():\n return create_random_string(32)\n\n\ndef get_role(aws_session, role_name):\n \"\"\"\n Get an aws IAM access role and its external id\n\n Args:\n aws_session (awsProvider): awsProvider Object\n role_name (str): Name of role to get\n\n Returns:\n str, str: The found role ARN and external_id\n \"\"\"\n iam = aws_session.connection('iam')\n try:\n role = iam.client.get_role(RoleName=role_name)\n external_id = get_external_id_from_role_object(role)\n return role[\"Role\"][\"Arn\"], external_id\n except botocore.exceptions.ClientError:\n return None, None\n\n\ndef get_all_roles(aws_session, name_filters=None):\n iam = aws_session.connection('iam')\n found_roles = []\n try:\n roles = iam.client.list_roles()\n if name_filters:\n for role in roles.get('Roles', []):\n for name_filter in name_filters:\n if name_filter in role['RoleName']:\n found_roles.append(role)\n else:\n found_roles = roles.get('Roles', [])\n except botocore.exceptions.ClientError:\n pass\n\n return found_roles\n\n\ndef get_external_id_from_role_object(role):\n try:\n return role['Role']['AssumeRolePolicyDocument']['Statement'][0]['Condition']['StringEquals']['sts:ExternalId']\n except (AttributeError, KeyError):\n return None\n\n\ndef create_cross_account_role(aws_session, account_id_to_trust, role_name):\n \"\"\"\n Create the IAM role that the Reactor will use to access the target AWS Account\n\n Args:\n aws_session (awsProvider): awsProvider Object\n account_id_to_trust(int): Name of AWS account we are going to allow access from\n role_name (str): Name of role to create\n\n Returns:\n str, str: The found or created role ARN and external_id\n \"\"\"\n iam = aws_session.connection('iam')\n external_id = generate_external_id()\n trusted_policy_doc = generate_assume_role_policy(account_id_to_trust, external_id)\n iam.client.create_role(RoleName=role_name,\n AssumeRolePolicyDocument=trusted_policy_doc)\n role = iam.client.get_role(RoleName=role_name)\n time.sleep(5) # A (wishful) attempt to wait for the role to settle in on the AWS side\n role_arn = role[\"Role\"][\"Arn\"]\n\n return role_arn, external_id\n\n\ndef create_or_update_sqs_queue(sqs_client, queue_name=None, sns_topic_arn=None, reactor_account_id=None):\n policy_template = load_policy_document(\"sqs-queue-policy.json\")\n sqs_policy = render_template(policy_template, {'reactor_account_id': reactor_account_id,\n 'sns_topic_arn': sns_topic_arn})\n\n try:\n response = sqs_client.create_queue(QueueName=queue_name, Attributes={\"Policy\": sqs_policy})\n sqs_url = response['QueueUrl']\n except botocore.exceptions.ClientError:\n sqs_url = sqs_client.get_queue_url(QueueName=queue_name)['QueueUrl']\n sqs_client.set_queue_attributes(QueueUrl=sqs_url, Attributes={\"Policy\": sqs_policy})\n\n sqs_arn = sqs_client.get_queue_attributes(QueueUrl=sqs_url,\n AttributeNames=['QueueArn'])['Attributes']['QueueArn']\n\n return sqs_arn\n\n\ndef create_sns_topic(sns_client, topic_name, connected_account_id, reactor_account_id):\n response = sns_client.create_topic(Name=topic_name)\n sns_arn = response['TopicArn']\n update_sns_policy(sns_client=sns_client, topic_name=topic_name, topic_arn=sns_arn,\n connected_account_id=connected_account_id,\n reactor_account_id=reactor_account_id)\n return sns_arn\n\n\ndef verify_arn(aws_session, service_arn):\n \"\"\"\n Verify an ARN points to a valid and accessable amazon resource\n\n note: Currently only supports SNS ARNs\n\n Args:\n aws_session (awsProvider): awsProvider Object\n service_arn (str): ARN to verify\n\n Returns:\n str\n \"\"\"\n try:\n service_name = service_arn.split(':')[2]\n service_object = aws_session.connection(service_name)\n if service_name == 'sns':\n service_object.client.get_topic_attributes(TopicArn=service_arn)\n return service_arn\n else:\n raise Exception('Unsupported service')\n except (botocore.exceptions.ClientError, AttributeError):\n return None\n\n\ndef get_all_subscriptions_to_sns_topic(aws_session, sns_arn):\n sns = aws_session.connection('sns')\n subscriptions = sns.client.list_subscriptions_by_topic(TopicArn=sns_arn)['Subscriptions']\n return subscriptions\n\n\ndef get_sqs_subscriptions_to_sns_topic(aws_session, sns_arn):\n subscriptions = get_all_subscriptions_to_sns_topic(aws_session, sns_arn)\n return [x for x in subscriptions if x['Protocol'] == 'sqs']\n\n\ndef print_iam_user_details(aws_session):\n \"\"\"\n Displays IAM user details to the screen\n\n Args:\n aws_session (awsProvider):\n \"\"\"\n try:\n iam = aws_session.connection('iam')\n user = iam.resource.CurrentUser()\n print('CURRENT USER: {}'.format(user.arn))\n user_policies = iam.client.list_user_policies(UserName=user.user_name)['PolicyNames']\n print('--- Policies ---')\n for policy in user_policies:\n print(policy)\n print(json.dumps(iam.client.get_user_policy(UserName=user.user_name, PolicyName=policy), indent=4))\n user_groups = iam.client.list_groups_for_user(UserName=user.user_name)['Groups']\n print('--- groups ---')\n for group in user_groups:\n print(group['GroupName'])\n group_policy_names = iam.client.list_group_policies(GroupName=group['GroupName'])['PolicyNames']\n print('---group policies---')\n for group_policy_name in group_policy_names:\n print(group_policy_name)\n group_policy = iam.client.get_group_policy(GroupName=group['GroupName'],\n PolicyName=group_policy_name)\n print(json.dumps(group_policy, indent=4))\n print('---managed group policies---')\n managed_group_policy_names = iam.client.list_attached_group_policies(GroupName=group['GroupName'])[\n 'AttachedPolicies']\n for group_policy in managed_group_policy_names:\n print(group_policy['PolicyArn'])\n\n except Exception as e:\n print(e)\n\n\ndef combine_iam_policies(left_json, right_json):\n \"\"\"\n Combine two IAM policy documents into one. The id information from the\n `left` is used in the final document, and the statements\n from `right` are merged into `left`, overwriting duplicate Sid's\n\n Args:\n left_json (str): Primary JSON document\n right_json (str): Updating JSON document\n\n Returns:\n str : combined JSON policy document\n\n \"\"\"\n # if the left policy was empty there is nothing to combine\n if not left_json:\n return right_json\n\n left = json.loads(left_json)\n right = json.loads(right_json)\n\n statements = {x['Sid']: x for x in left['Statement']}\n statements.update({x['Sid']: x for x in right['Statement']})\n\n left['Statement'] = list(statements.values())\n return json.dumps(left)\n\n\ndef update_sns_policy(sns_client, topic_name, topic_arn, connected_account_id, reactor_account_id):\n policy_template = load_policy_document(\"cloudtrail-sns-policy-template.json\")\n sns_policy = render_template(policy_template, {'reactor_account_id': reactor_account_id,\n 'sns_topic_arn': topic_arn,\n 'sns_topic_name': topic_name,\n 'region': sns_client.meta.region_name,\n 'connected_account_id': connected_account_id})\n existing_policy = sns_client.get_topic_attributes(TopicArn=topic_arn)['Attributes']['Policy']\n combined_policy = combine_iam_policies(existing_policy, sns_policy)\n response = sns_client.set_topic_attributes(TopicArn=topic_arn,\n AttributeName='Policy',\n AttributeValue=combined_policy)\n return response\n\n\ndef apply_bucket_lifecycle_policy(aws_session, bucket_name, lifecycle_policy, force=False):\n \"\"\"\n Safely set the bucket lifecycle policy. Safely means if there is a policy already present,\n do nothing unless forced\n\n Args:\n aws_session (awsProvider): aws provider object\n bucket_name (str): Name of s3 bucket\n lifecycle_policy (dict): a AWS S3 lifecycle policy\n force (bool): Force policy apply\n\n Returns:\n object: A response object if a lifecycle policy was applied\n \"\"\"\n try:\n s3 = aws_session.connection('s3')\n existing_lifecycle_policy = s3.client.get_bucket_lifecycle_configuration(Bucket=bucket_name)\n except botocore.exceptions.ClientError:\n existing_lifecycle_policy = None\n\n if not existing_lifecycle_policy or force:\n response = s3.client.put_bucket_lifecycle_configuration(\n Bucket=bucket_name,\n LifecycleConfiguration=lifecycle_policy\n )\n return response\n else:\n return None\n\n\ndef apply_bucket_policy(aws_session, bucket_name, new_policy):\n \"\"\"\n\n Args:\n aws_session (awsProvider):\n bucket_name:\n new_policy:\n \"\"\"\n s3 = aws_session.connection('s3')\n\n try:\n existing_policy = s3.client.get_bucket_policy(Bucket=bucket_name)['Policy']\n except botocore.exceptions.ClientError:\n existing_policy = \"\" # bucket has no pre-existing policy, which is fine\n\n combined_policy = combine_iam_policies(existing_policy, new_policy)\n s3.client.put_bucket_policy(\n Bucket=bucket_name,\n Policy=combined_policy\n )\n\n\ndef apply_cloudtrail_bucket_policy(aws_session, bucket_name, reactor_account_id):\n policy_template = load_policy_document(\"cloudtrail-bucket-policy-template.json\")\n new_policy = render_template(policy_template, {'reactor_account_id': reactor_account_id,\n 'cloudtrail_bucket': bucket_name})\n apply_bucket_policy(aws_session, bucket_name, new_policy)\n tag_bucket(aws_session, bucket_name)\n response = apply_bucket_lifecycle_policy(aws_session, bucket_name, CLOUDTRAIL_BUCKET_LIFECYCLE_POLICY)\n if response:\n print(' - Bucket lifecycle rules have been set')\n else:\n print(' - Bucket lifecycle rules found and left alone')\n\n\ndef load_policy_document(policy_file_name, packaged=True):\n \"\"\"\n Loads an Amazon IAM policy from a specified file or uses a recommended default\n\n Args:\n policy_file_name (str): Policy file name\n packaged (bool): Look for file in package location or provided path. Default packaged (True)\n\n Returns:\n str: The JSON policy document, as a string\n \"\"\"\n if packaged:\n policy_document = pkg_resources.resource_string('cloudzerocli', \"data/{}\".format(policy_file_name))\n else:\n with open(policy_file_name, 'r') as input_file:\n policy_document = input_file.read()\n\n return policy_document.decode(\"utf-8\")\n\n\ndef extract_trail_tags(input):\n extracted_tags = {}\n\n try:\n for taglist in input['ResourceTagList']:\n for tags in taglist['TagsList']:\n extracted_tags[tags['Key']] = tags['Value']\n except TypeError:\n return None\n except KeyError:\n raise KeyError('Malformed ResourceTagList input dictionary')\n\n return extracted_tags\n\n\ndef tag_cloudtrail_configuration(cloudtrail_client, trail_arn):\n response = cloudtrail_client.add_tags(\n ResourceId=trail_arn,\n TagsList=[\n {\n 'Key': 'cloudzero-reactor',\n 'Value': datetime.now(tzlocal()).isoformat()\n },\n ]\n )\n return response\n\n\ndef tag_bucket(aws_session, bucket_name):\n \"\"\"\n\n Args:\n aws_session (awsProvider):\n bucket_name:\n\n Returns:\n\n \"\"\"\n s3 = aws_session.connection('s3')\n\n response = s3.client.put_bucket_tagging(\n Bucket=bucket_name,\n Tagging={\n 'TagSet': [\n {\n 'Key': 'cloudzero-reactor',\n 'Value': datetime.now(tzlocal()).isoformat()\n },\n ]\n }\n )\n return response\n\n\ndef verify_bucket_read(aws_session, bucket, prefix=None):\n \"\"\"\n Verify that we can read files from a given S3 bucket. Since the only way to truly ensure we have read rights\n is to actually read a file, that's what we attempt to do. However, in some cases we might be testing an empty\n bucket, in which case we can't be 100% sure we have read rights, so in those cases we return ResourceAccess.MAYBE_OK\n Args:\n aws_session: Boto3 wrapper class that provides account connections and client access\n bucket (str): The bucket\n prefix (str): A key path\n\n Returns:\n str\n\n \"\"\"\n prefix_string = '/' + prefix if prefix else ''\n result = \"no\"\n\n if not bucket:\n return result\n\n try:\n s3 = aws_session.connection('s3')\n s3.client.head_bucket(Bucket=bucket)\n if prefix:\n response = s3.client.list_objects_v2(Bucket=bucket, Prefix=prefix_string)\n else:\n response = s3.client.list_objects_v2(Bucket=bucket)\n\n objects = response.get('Contents', {})\n for obj in objects:\n # Iterate over objects, use size to quickly determine if object is a file (not a folder)\n if obj.get('Size'):\n try:\n # get object will fail if storage class is Glacier\n response = s3.client.get_object(Bucket=bucket, Key=obj.get('Key'))\n if response.get('ContentLength'):\n result = \"yes\"\n break\n except botocore.exceptions.ClientError as err:\n if \"The operation is not valid for the object's storage class\" in str(err):\n # The bucket has mixed storage classes, but permissions look ok\n result = \"maybe\"\n else:\n # No permissions\n result = \"no\"\n break\n else:\n result = \"maybe\"\n except botocore.exceptions.ClientError:\n # unknown error, would be nice to have some logging in here in the future, for now ¯\\_(ツ)_/¯!\n result = \"no\"\n\n return result\n\n\ndef verify_bucket_read_write(aws_session, bucket, prefix=None):\n s3 = aws_session.connection('s3')\n\n if prefix:\n verify_key = '{}/VERIFY-{}'.format(prefix, create_random_string())\n else:\n verify_key = 'VERIFY-{}'.format(create_random_string())\n\n content = 'CloudZero Reactor read/write verification'.encode()\n\n # try:\n s3.client.put_object(ACL='private',\n Body=content,\n Bucket=bucket,\n Key=verify_key)\n\n response = s3.client.get_object(Bucket=bucket,\n Key=verify_key)\n\n received_content = response['Body'].read()\n\n assert received_content == content\n\n s3.client.delete_object(Bucket=bucket,\n Key=verify_key)\n\n return True\n\n\nclass awsProvider(object):\n class awsConnection(object):\n def __init__(self, service_name, client, resource):\n self.__client = client\n self.__resource = resource\n self.__service_name = service_name\n\n @property\n def service_name(self):\n return self.__service_name\n\n @property\n def client(self):\n return self.__client\n\n @property\n def resource(self):\n return self.__resource\n\n def __init__(self, role_arn=None, external_id=None, profile_name='default',\n region_name='us-east-1', endpoint_url=None, aws_access_key_id=None, aws_secret_access_key=None):\n super(awsProvider, self).__init__()\n\n self.region_name = region_name\n self.profile_name = profile_name\n self.mfa_serial = None\n\n self.__session = None\n self.__endpoint_url = endpoint_url\n self.__aws_access_key_id = aws_access_key_id\n self.__aws_secret_access_key = aws_secret_access_key\n self.__connection_type = AwsProviderConnectionType.UNKNOWN\n self.__role_session_name = 'cloudzero-reactor'\n self.__account_id = None\n self.__account_name = None\n\n if role_arn and external_id:\n self.role_arn = role_arn\n self.external_id = external_id\n self.__connection_type = AwsProviderConnectionType.ROLE\n else:\n self.__connection_type = AwsProviderConnectionType.LOCAL\n\n self.configure_session()\n\n def account_id(self):\n if not self.__account_id:\n sts_client = self.__session.client('sts')\n response = sts_client.get_caller_identity()\n self.__account_id = response.get('Account')\n\n return self.__account_id\n\n def account_name(self):\n \"\"\"\n Returns:\n str: Account alias\n \"\"\"\n if not self.__account_name:\n iam_client = self.__session.client('iam')\n response = iam_client.list_account_aliases()\n try:\n self.__account_name = response['AccountAliases'][0]\n except (AttributeError, IndexError):\n self.__account_name = None\n\n return self.__account_name\n\n @property\n def session(self):\n return self.__session\n\n def configure_session(self):\n if self.__connection_type is AwsProviderConnectionType.ROLE:\n # Create a initial session using user provided profile from which we will then assume the role from\n initial_session = boto3.Session(profile_name=self.profile_name, region_name=self.region_name)\n sts_client = initial_session.client('sts')\n credentials = sts_client.assume_role(RoleArn=self.role_arn,\n RoleSessionName=self.__role_session_name,\n ExternalId=self.external_id)['Credentials']\n\n # Create final boto3 session from the assumed role\n self.__session = boto3.Session(profile_name=self.profile_name,\n aws_access_key_id=credentials['AccessKeyId'],\n aws_secret_access_key=credentials['SecretAccessKey'],\n aws_session_token=credentials['SessionToken'],\n region_name=self.region_name)\n\n elif self.__connection_type is AwsProviderConnectionType.LOCAL:\n # LOCAL connections support MFA and use the credentials cache\n\n # Configure system credentials cache which by default is ~/.aws/boto/cache\n cli_cache = os.path.join(os.path.expanduser('~'), '.aws/cli/cache')\n\n # Construct low level botocore session with cache, which allows MFA session reuse\n session = botocore.session.Session(profile=self.profile_name)\n session.get_component('credential_provider').get_provider('assume-role').cache = \\\n botocore.credentials.JSONFileCache(cli_cache)\n\n # this mfa_serial code is _only_ here to deal with boto profiles that _do not_ assume a role\n # which is required because this PR is still open: https://github.com/boto/botocore/pull/1399\n # otherwise MFA is nicely handled automatically by boto\n self.mfa_serial = session.full_config['profiles'][self.profile_name].get('mfa_serial')\n if self.mfa_serial and not session.full_config['profiles'][self.profile_name].get('role_arn'):\n sts = session.create_client('sts')\n mfa = self.mfa_serial\n mfa_code = input(\"Enter MFA code for {}: \".format(self.mfa_serial))\n response = sts.get_session_token(DurationSeconds=3600, SerialNumber=mfa, TokenCode=mfa_code)\n credentials = response['Credentials']\n self.__session = boto3.Session(botocore_session=session,\n aws_access_key_id=credentials['AccessKeyId'],\n aws_secret_access_key=credentials['SecretAccessKey'],\n aws_session_token=credentials['SessionToken'],\n region_name=self.region_name)\n else:\n # create final boto3 session for system use\n self.__session = boto3.Session(botocore_session=session,\n profile_name=self.profile_name,\n region_name=self.region_name)\n else:\n raise Exception('UNKNOWN Connection type {}'.format(self.__connection_type))\n\n def connection(self, service_name):\n client = self.__session.client(service_name=service_name,\n aws_access_key_id=self.__aws_access_key_id,\n aws_secret_access_key=self.__aws_secret_access_key,\n endpoint_url=self.__endpoint_url)\n\n if service_name in AWS_RESOURCE_SERVICES:\n resource = self.__session.resource(service_name=service_name)\n else:\n resource = None\n\n return self.awsConnection(service_name, client, resource)\n\n\ndef is_org_account(aws_session):\n \"\"\"\n\n Args:\n aws_session (awsProvider):\n\n Returns:\n\n \"\"\"\n organizations = aws_session.connection('organizations')\n try:\n response = organizations.client.describe_organization()\n org_details = response.get('Organization')\n except organizations.client.exceptions.AWSOrganizationsNotInUseException: # No Org for this account\n org_details = None\n\n return org_details\n\n\ndef delete_sqs_queue(sqs_arn, sqs_client):\n try:\n queue_name = sqs_arn.split(':')[-1]\n queue_url = sqs_client.get_queue_url(QueueName=queue_name).get('QueueUrl')\n sqs_client.delete_queue(QueueUrl=queue_url)\n return True\n except (botocore.exceptions.ClientError, AttributeError):\n return False\n\n\ndef verify_sqs_queue_exists(aws_session, queue_name):\n sqs = aws_session.connection('sqs')\n try:\n return bool(sqs.client.get_queue_url(QueueName=queue_name))\n except botocore.exceptions.ClientError as err:\n if err.response['Error']['Code'] == \"AWS.SimpleQueueService.NonExistentQueue\":\n return False\n else:\n raise\n\n\ndef delete_sns_subscription(sns_arn, sns_client):\n try:\n sns_client.unsubscribe(SubscriptionArn=sns_arn)\n return True\n except botocore.exceptions.ClientError:\n return False\n\n\ndef delete_iam_role(role_arn, iam_client):\n\n # find all policies attached to role, and remove them\n resource_type_and_name = role_arn.split(':')[-1]\n role_name = resource_type_and_name.split('/')[-1]\n inline_policy_names = iam_client.list_role_policies(RoleName=role_name).get('PolicyNames', [])\n attached_policy_arns = [p['PolicyArn'] for p\n in iam_client.list_attached_role_policies(RoleName=role_name).get('AttachedPolicies', [])]\n\n try:\n for n in inline_policy_names:\n iam_client.delete_role_policy(RoleName=role_name, PolicyName=n)\n for a in attached_policy_arns:\n iam_client.detach_role_policy(RoleName=role_name, PolicyArn=a)\n iam_client.delete_role(RoleName=role_name)\n except botocore.exceptions.ClientError:\n return False\n return True\n", "sub_path": "cloudzerocli/utils/aws.py", "file_name": "aws.py", "file_ext": "py", "file_size_in_byte": 27139, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "cloudzerocli.settings.Settings", "line_number": 23, "usage_type": "call"}, {"api_name": "botocore.credentials.exceptions", "line_number": 57, "usage_type": "attribute"}, {"api_name": "botocore.credentials", "line_number": 57, "usage_type": "name"}, {"api_name": "botocore.credentials.exceptions", "line_number": 74, "usage_type": "attribute"}, {"api_name": "botocore.credentials", "line_number": 74, "usage_type": "name"}, {"api_name": "cloudzerocli.utils.formatters.render_template", "line_number": 101, "usage_type": "call"}, {"api_name": "cloudzerocli.utils.formatters.create_random_string", "line_number": 107, "usage_type": "call"}, {"api_name": "botocore.credentials.exceptions", "line_number": 126, "usage_type": "attribute"}, {"api_name": "botocore.credentials", "line_number": 126, "usage_type": "name"}, {"api_name": "botocore.credentials.exceptions", "line_number": 142, "usage_type": "attribute"}, {"api_name": "botocore.credentials", "line_number": 142, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 173, "usage_type": "call"}, {"api_name": "cloudzerocli.utils.formatters.render_template", "line_number": 181, "usage_type": "call"}, {"api_name": "botocore.credentials.exceptions", "line_number": 187, "usage_type": "attribute"}, {"api_name": "botocore.credentials", "line_number": 187, "usage_type": "name"}, {"api_name": "botocore.credentials.exceptions", "line_number": 227, "usage_type": "attribute"}, {"api_name": "botocore.credentials", "line_number": 227, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 257, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 268, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 297, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 298, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 304, "usage_type": "call"}, {"api_name": "cloudzerocli.utils.formatters.render_template", "line_number": 309, "usage_type": "call"}, {"api_name": "botocore.credentials.exceptions", "line_number": 339, "usage_type": "attribute"}, {"api_name": "botocore.credentials", "line_number": 339, "usage_type": "name"}, {"api_name": "botocore.credentials.exceptions", "line_number": 364, "usage_type": "attribute"}, {"api_name": "botocore.credentials", "line_number": 364, "usage_type": "name"}, {"api_name": "cloudzerocli.utils.formatters.render_template", "line_number": 376, "usage_type": "call"}, {"api_name": "pkg_resources.resource_string", "line_number": 399, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 428, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 428, "usage_type": "name"}, {"api_name": "dateutil.tz.tzlocal", "line_number": 428, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 453, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 453, "usage_type": "name"}, {"api_name": "dateutil.tz.tzlocal", "line_number": 453, "usage_type": "call"}, {"api_name": "botocore.credentials.exceptions", "line_number": 499, "usage_type": "attribute"}, {"api_name": "botocore.credentials", "line_number": 499, "usage_type": "name"}, {"api_name": "botocore.credentials.exceptions", "line_number": 509, "usage_type": "attribute"}, {"api_name": "botocore.credentials", "line_number": 509, "usage_type": "name"}, {"api_name": "cloudzerocli.utils.formatters.create_random_string", "line_number": 520, "usage_type": "call"}, {"api_name": "cloudzerocli.utils.formatters.create_random_string", "line_number": 522, "usage_type": "call"}, {"api_name": "cloudzerocli.constants.AwsProviderConnectionType.UNKNOWN", "line_number": 576, "usage_type": "attribute"}, {"api_name": "cloudzerocli.constants.AwsProviderConnectionType", "line_number": 576, "usage_type": "name"}, {"api_name": "cloudzerocli.constants.AwsProviderConnectionType.ROLE", "line_number": 584, "usage_type": "attribute"}, {"api_name": "cloudzerocli.constants.AwsProviderConnectionType", "line_number": 584, "usage_type": "name"}, {"api_name": "cloudzerocli.constants.AwsProviderConnectionType.LOCAL", "line_number": 586, "usage_type": "attribute"}, {"api_name": "cloudzerocli.constants.AwsProviderConnectionType", "line_number": 586, "usage_type": "name"}, {"api_name": "cloudzerocli.constants.AwsProviderConnectionType.ROLE", "line_number": 618, "usage_type": "attribute"}, {"api_name": "cloudzerocli.constants.AwsProviderConnectionType", "line_number": 618, "usage_type": "name"}, {"api_name": "boto3.Session", "line_number": 620, "usage_type": "call"}, {"api_name": "boto3.Session", "line_number": 627, "usage_type": "call"}, {"api_name": "cloudzerocli.constants.AwsProviderConnectionType.LOCAL", "line_number": 633, "usage_type": "attribute"}, {"api_name": "cloudzerocli.constants.AwsProviderConnectionType", "line_number": 633, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 637, "usage_type": "call"}, {"api_name": "os.path", "line_number": 637, "usage_type": "attribute"}, {"api_name": "os.path.expanduser", "line_number": 637, "usage_type": "call"}, {"api_name": "botocore.credentials.session.Session", "line_number": 640, "usage_type": "call"}, {"api_name": "botocore.credentials.session", "line_number": 640, "usage_type": "attribute"}, {"api_name": "botocore.credentials", "line_number": 640, "usage_type": "name"}, {"api_name": "botocore.credentials.credentials.JSONFileCache", "line_number": 642, "usage_type": "call"}, {"api_name": "botocore.credentials.credentials", "line_number": 642, "usage_type": "attribute"}, {"api_name": "botocore.credentials", "line_number": 642, "usage_type": "name"}, {"api_name": "boto3.Session", "line_number": 654, "usage_type": "call"}, {"api_name": "boto3.Session", "line_number": 661, "usage_type": "call"}, {"api_name": "cloudzerocli.constants.AWS_RESOURCE_SERVICES", "line_number": 673, "usage_type": "name"}, {"api_name": "botocore.credentials.exceptions", "line_number": 706, "usage_type": "attribute"}, {"api_name": "botocore.credentials", "line_number": 706, "usage_type": "name"}, {"api_name": "botocore.credentials.exceptions", "line_number": 714, "usage_type": "attribute"}, {"api_name": "botocore.credentials", "line_number": 714, "usage_type": "name"}, {"api_name": "botocore.credentials.exceptions", "line_number": 725, "usage_type": "attribute"}, {"api_name": "botocore.credentials", "line_number": 725, "usage_type": "name"}, {"api_name": "botocore.credentials.exceptions", "line_number": 744, "usage_type": "attribute"}, {"api_name": "botocore.credentials", "line_number": 744, "usage_type": "name"}]}
+{"seq_id": "627952222", "text": "import functools\n\n\ndef is_module_definition(definition):\n return (type(definition) is dict and 'name' in definition and\n 'type' in definition)\n\n\ndef config_module_partial(module_name, module_type, get_module_builders,\n **params):\n module_builders = get_module_builders(module_type)\n try:\n module_builder: dict = module_builders[module_name]\n except KeyError:\n raise ValueError(f'Wrong module name \"{module_name}\" provided\\n'\n f'Available names: {module_builders.keys()}\\n')\n return functools.partial(module_builder, **params)\n\n\ndef config_module(module_name, module_type, get_module_builders, **params):\n try:\n return config_module_partial(module_name, module_type,\n get_module_builders, **params)()\n except TypeError as e:\n raise TypeError('Error, trying to initialize module:\\n'\n f'module type: {type}\\n'\n f'module name: {module_name}\\n'\n f'params:\\n{params}') from e\n\n\nclass ResourceManager:\n def __init__(self, config, get_module_builders):\n self.config = config\n self.get_module_builders = get_module_builders\n self.resources = {}\n\n def __getitem__(self, item):\n return self._get_resource(item)\n\n def __setitem__(self, key, value):\n assert key not in self.resources, f'Tried to overwrite resource {key}'\n self.resources[key] = value\n\n def get(self, item, default=None):\n return self.resources.get(item, default)\n\n def _get_resource(self, name):\n if name not in self.resources:\n try:\n definition = self.config[name]\n except KeyError as e:\n raise TypeError(\n f\"Couldn't find definition for resource {name}\\n\"\n f\"Config: {self.config}\") from e\n\n self.resources[name] = self._define_resource(definition)\n return self.resources[name]\n\n def _define_resource(self, definition):\n if is_module_definition(definition):\n return self._define_module(definition)\n else:\n # Consider definition to be a simple python type\n return definition\n\n def _define_module(self, definition: dict):\n module_name = definition['name']\n module_type = definition['type']\n inputs = self._get_inputs(definition.get('inputs', {}))\n params = definition.get('params', {})\n if definition.get('init', True):\n initialize = config_module\n else:\n initialize = config_module_partial\n return initialize(module_name, module_type, self.get_module_builders,\n **params, **inputs)\n\n def _get_inputs(self, inputs_definition):\n inputs = {}\n for name, value in inputs_definition.items():\n if type(value) is str:\n inputs[name] = self._get_resource(value)\n elif is_module_definition(value):\n inputs[name] = self._define_module(value)\n else:\n raise ValueError(\"Couldn't define module:\\n\"\n f'{inputs_definition}\\n'\n 'Wrong type for input:\\n'\n f'{name}: {value}\\n'\n 'Possible types: \"str\" or \"module\"')\n return inputs\n", "sub_path": "dpipe/externals/resource_manager/resource_manager/base.py", "file_name": "base.py", "file_ext": "py", "file_size_in_byte": 3438, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "functools.partial", "line_number": 17, "usage_type": "call"}]}
+{"seq_id": "444628848", "text": "from kaggle_environments import evaluate, make, utils\nimport numpy as np\nimport gym\nimport matplotlib.pyplot as plt\nfrom tqdm.notebook import tqdm\nimport torch.optim as optim\nimport network\nimport torch.nn as nn\nimport torch\nimport game_environment\nimport agents\n\ndef plot_grad_flow(named_parameters):\n ave_grads = []\n layers = []\n for n, p in named_parameters:\n if(p.requires_grad) and (\"bias\" not in n):\n layers.append(n)\n ave_grads.append(p.grad.abs().mean())\n plt.plot(ave_grads, alpha=0.3, color=\"b\")\n plt.hlines(0, 0, len(ave_grads)+1, linewidth=1, color=\"k\" )\n plt.xticks(range(0,len(ave_grads), 1), layers, rotation=\"vertical\")\n plt.xlim(xmin=0, xmax=len(ave_grads))\n plt.xlabel(\"Layers\")\n plt.ylabel(\"Average Gradient\")\n plt.title(\"Gradient Flow\")\n plt.grid(True)\n return plt\n\nclass OpponentDQN:\n def __init__(self, num_states, num_actions):\n self.num_actions = num_actions\n self.model = network.ConnectXNetwork2(num_states, num_actions)\n self.mark = 6\n self.name = 0\n self.EVALenv = game_environment.ConnectXEnvironment(7, 6, 4)\n \n def predict(self, inputs):\n return self.model(torch.from_numpy(inputs).float())\n \n def preprocess(self, state):\n result = state[:]\n return result\n\n def lookahead(self, state, action, mark, depth = 2):\n self.EVALenv.copy_board(state)\n new_state, valid, done, reward = self.EVALenv.step(action, mark)\n mark = mark % 2 +1\n if done:\n #print(\"DONE\")\n value = reward[mark-1]\n #print(\"value\", value)\n return value \n elif depth == 0:\n prediction = self.predict(np.atleast_2d(self.preprocess(new_state)))[0].detach().numpy()\n value = max(prediction)\n if value > 1:\n value = 0.999\n elif value < -1:\n value = -0.999\n return value\n else:\n value = -1e7\n possible_actions = [i for i in range(self.num_actions) if new_state[i] == 0]\n for action in possible_actions:\n value = max(value, -self.lookahead(new_state, action, mark, depth -1))\n return value\n\n def get_action(self, state, epsilon):\n mark = 1\n if np.random.random() < epsilon:\n return(int(np.random.choice([c for c in range(self.num_actions) if state[c] == 0])))\n else:\n best_value = -1e7 \n best_action = 20 #want the shit to crash if an action isnt selected through negamax\n possible_actions = [i for i in range(self.num_actions) if state[i] == 0]\n for action in possible_actions:\n value = -self.lookahead(state, action, mark)\n \n if value > best_value:\n best_action = action\n best_value = value\n return best_action\n \n def load_weights(self, path):\n self.model.load_state_dict(torch.load(path))\n self.name = path\n\n\n\nclass DQN:\n def __init__(self, num_states, num_actions, gamma, max_exp, min_exp, batch_size, learning_rate):\n self.num_actions = num_actions\n self.batch_size = batch_size\n self.gamma = gamma\n self.model = network.ConnectXNetwork2(num_states, num_actions)\n self.optimizer = optim.Adam(self.model.parameters() ,lr = learning_rate)\n self.criterion = nn.MSELoss()\n self.mark = 1 #placeholder for verticalbot funticonality\n self.name = 0 \n self.EVALenv = game_environment.ConnectXEnvironment(7, 6, 4)\n \n self.experience = {'prev_obs' : [], 'a' : [], 'r': [], 'obs' : [], 'done': [] } \n\n self.max_exp = max_exp\n self.min_exp = min_exp\n\n def predict(self, inputs):\n return self.model(torch.from_numpy(inputs).float())\n \n \n def preprocess(self, state):\n result = state[:]\n return result\n\n def get_action_no_lookahead(self, state, epsilon):\n if np.random.random() < epsilon:\n return(int(np.random.choice([c for c in range(self.num_actions) if state[c] == 0])))\n else:\n prediction = self.predict(np.atleast_2d(self.preprocess(state)))[0].detach().numpy()\n for i in range(self.num_actions):\n if state[i] != 0:\n prediction[i] = -1e7\n return int(np.argmax(prediction)) \n\n \n def lookahead(self, state, action, mark, alpha, beta, depth = 2):\n self.EVALenv.copy_board(state)\n new_state, valid, done, reward = self.EVALenv.step(action, mark)\n mark = mark % 2 +1\n if done:\n value = reward[mark-1]\n return value \n elif depth == 0:\n prediction = self.predict(np.atleast_2d(self.preprocess(new_state)))[0].detach().numpy()\n value = max(prediction)\n \n if value > 20:\n value = 19.999\n elif value < -20:\n value = -19.999\n return value\n else:\n value = -1e7\n possible_actions = [i for i in range(self.num_actions) if new_state[i] == 0]\n for action in possible_actions:\n value = max(value, -self.lookahead(new_state, action, mark, -beta, -alpha, depth -1))\n alpha = max(alpha, value)\n if alpha >= beta:\n break\n return value\n\n def get_action(self, state, epsilon):\n mark = 1\n if np.random.random() < epsilon:\n #print(\"Random\")\n return(int(np.random.choice([c for c in range(self.num_actions) if state[c] == 0])))\n else:\n alpha = float(\"-inf\")\n beta = float(\"inf\")\n best_value = -1e7\n best_action = 20 #want the shit to crash if an action isnt selected through negamax\n possible_actions = [i for i in range(self.num_actions) if state[i] == 0]\n for action in possible_actions:\n value = -self.lookahead(state, action, mark, -beta, -alpha)\n \n if value > best_value:\n best_action = action\n best_value = value\n return best_action\n \n def get_values(self, state, TargetNet):\n alpha = float(\"-inf\")\n beta = float(\"inf\")\n mark = 1\n best_value = -1e7\n best_action = 20 #want the shit to crash if an action isnt selected through negamax\n possible_actions = [i for i in range(self.num_actions) if state[i] == 0]\n value = TargetNet.predict(state).detach().numpy()\n for i in range(len(value)):\n if value[i] > 20:\n value[i] = 19.999\n elif value[i] < -20:\n value[i] = -19.999\n for action in possible_actions:\n value[action] = -self.lookahead_values(state, action, mark, -beta, -alpha, TargetNet )\n return value\n\n def lookahead_values(self, state, action, mark, alpha, beta, TargetNet, depth = 1):\n self.EVALenv.copy_board(state)\n new_state, valid, done, reward = self.EVALenv.step(action, mark)\n mark = mark % 2 +1\n if done:\n value = reward[mark-1]\n return value \n elif depth == 0:\n prediction = TargetNet.predict(np.atleast_2d(self.preprocess(new_state)))[0].detach().numpy()\n value = max(prediction)\n if value > 20:\n value = 19.999\n elif value < -20:\n value = -19.999\n return value\n else:\n value = -1e7\n possible_actions = [i for i in range(self.num_actions) if new_state[i] == 0]\n for action in possible_actions:\n value = max(value, -self.lookahead_values(new_state, action, mark, -beta, -alpha, TargetNet, depth -1))\n alpha = max(alpha, value)\n if alpha >= beta:\n \n break\n return value\n\n \n def add_experience(self, exp):\n if len(self.experience['prev_obs']) >= self.max_exp:\n for key in self.experience.keys():\n self.experience[key].pop(0)\n for key, value in exp.items():\n self.experience[key].append(value)\n\n def load_weights(self, path):\n self.model.load_state_dict(torch.load(path))\n self.name = path\n\n def save_weights(self, path):\n torch.save(self.model.state_dict(), path)\n \n def copy_weights(self, TrainNet):\n self.model.load_state_dict(TrainNet.model.state_dict())\n\n def train(self, TargetNet):\n if len(self.experience['prev_obs']) < self.min_exp:\n return 0\n \n ids = np.random.randint(low = 0, high = len(self.experience['prev_obs']), size = self.batch_size)\n states = np.asarray([self.preprocess(self.experience['prev_obs'][i]) for i in ids])\n actions = np.asarray([self.experience['a'][i] for i in ids])\n rewards = np.asarray([self.experience['r'][i] for i in ids])\n next_states = np.asarray([self.preprocess(self.experience['obs'][i]) for i in ids])\n dones = np.asarray([self.experience['done'][i] for i in ids]) \n next_value = np.max(TargetNet.predict(next_states).detach().numpy(), axis=1)\n \"\"\"\n next_value = np.zeros(self.batch_size)\n k = 0\n\n for next_state in next_states:\n next_value[k] = np.max(self.get_values(next_state, TargetNet))\n k+=1\n \"\"\"\n\n ''' Q - learning aspect '''\n actual_values = np.where(dones, rewards, rewards+self.gamma*next_value)\n \n \n\n \n ''' !!! '''\n actions = np.expand_dims(actions, axis = 1)\n \n actions_one_hot = torch.FloatTensor(self.batch_size, self.num_actions).zero_()\n actions_one_hot = actions_one_hot.scatter_(1, torch.LongTensor(actions), 1)\n selected_action_values = torch.sum(self.predict(states) * actions_one_hot, dim = 1)\n \n actual_values = torch.FloatTensor(actual_values)\n \n \n self.optimizer.zero_grad()\n loss = self.criterion(selected_action_values, actual_values)\n loss.backward()\n self.optimizer.step()\n return loss\n\n\n\nclass ConnectXGym2(gym.Env):\n def __init__(self):\n self.env = game_environment.ConnectXEnvironment(7, 6, 4)\n \n self.trainer = 0\n \n \n self.columns = self.env.num_columns\n self.rows = self.env.num_rows\n self.actions = gym.spaces.Discrete(self.columns)\n self.positions = gym.spaces.Discrete(self.columns * self.rows)\n self.list_of_trainers = [\"new2\", \"new\", \"new1\", \"new3\", \"new4\"]\n self.score_list = {i : 0 for i in self.list_of_trainers}\n self.games_list = {i : 0 for i in self.list_of_trainers}\n self.change_trainer_at_random()\n \n def reset_scores(self):\n self.score_list = {i : 0 for i in self.list_of_trainers}\n self.games_list = {i : 0 for i in self.list_of_trainers}\n def print_scores(self):\n for i in self.score_list:\n print(i, self.score_list[i], end = \"\")\n print(\"/\", self.games_list[i])\n \n def change_trainer(self, trainer):\n self.trainer = trainer\n\n def change_trainer_at_random(self):\n choice = np.random.choice(self.list_of_trainers)\n if choice == \"verticalbot\":\n trainer = agents.verticalBot(7,6)\n self.change_trainer(trainer)\n else:\n trainer = OpponentDQN(42,7)\n trainer.load_weights(choice)\n self.change_trainer(trainer)\n \n\n def step(self, action, mark):\n return self.env.step(action, mark)\n\n def render(self, **kwargs):\n return self.env.render(**kwargs)\n\n def generate_data(self, TrainNet, TargetNet, epsilon, copy_step):\n rewards = 0\n opp_action = 0\n done = False\n\n env = self.env\n \n trainee_mark, observations = env.reset()\n \n if trainee_mark == 1:\n TrainNet.mark = 1\n self.trainer.mark = 2\n opp_mark = 2\n \n while not done:\n '''trainee makes a move '''\n action = TrainNet.get_action(observations, epsilon)\n\n prev_observations = np.array(observations)\n \n observations, valid, done, reward = env.step(action, trainee_mark)\n\n reward = reward[trainee_mark-1]\n if not done:\n ''' opponent makes a move '''\n ''' flip_the board '''\n observations = env.flip()\n opp_action = self.trainer.get_action(observations, 0.03)\n\n observations, valid, done, reward = env.step(opp_action, opp_mark)\n reward = reward[trainee_mark-1]\n\n if done:\n rewards += reward\n\n exp = {'prev_obs': prev_observations, 'a' : action, 'r': reward, 'obs': observations, 'done' : done }\n TrainNet.add_experience(exp)\n\n loss = TrainNet.train(TargetNet)\n\n return rewards, loss\n else:\n TrainNet.mark = 2\n self.trainer.mark = 1\n opp_mark = 1 \n ''' opponent makes a move '''\n opp_action = self.trainer.get_action(observations, 0.03)\n\n observations, valid, done, reward = env.step(opp_action, opp_mark)\n reward = reward[trainee_mark-1]\n\n while not done:\n observations = env.flip()\n action = TrainNet.get_action(observations, epsilon)\n\n prev_observations = np.array(observations)\n \n observations, valid, done, reward = env.step(action, trainee_mark)\n reward = reward[trainee_mark-1]\n if not done:\n ''' opponent makes a move '''\n opp_action = self.trainer.get_action(observations, 0.03)\n\n observations, valid, done, reward = env.step(opp_action, opp_mark)\n reward = reward[trainee_mark-1]\n\n if done:\n rewards += reward\n\n exp = {'prev_obs': prev_observations, 'a' : action, 'r': reward, 'obs': env.flip(), 'done' : done }\n TrainNet.add_experience(exp)\n\n loss = TrainNet.train(TargetNet)\n return rewards, loss\n\ndef dojo(games, gym, TrainNet, TargetNet, min_epsilon, epsilon, copy_step):\n total_loss = 0\n even_match = 0\n test_match = game_environment.ConnectXEnvironment(7, 6, 4)\n _, test_state = test_match.reset()\n print(TrainNet.predict(test_state)) \n decay = 0.9995\n for i in range(games):\n rewards, loss = gym.generate_data(TrainNet, TargetNet, epsilon, copy_step)\n if rewards == 0:\n even_match += 1\n print(\"motstander\", gym.trainer.mark)\n print(\"SCORE:\", rewards)\n gym.score_list[gym.trainer.name] += rewards\n gym.games_list[gym.trainer.name] += 1\n total_loss += loss\n print(i)\n if i%10 == 0 and i != 0:\n gym.change_trainer_at_random()\n print(TrainNet.predict(test_state)) \n if i%2 == 0 and i != 0:\n epsilon = max(min_epsilon, epsilon*decay)\n if i%100 == 0 and i != 0:\n print('Total Loss:', total_loss)\n print('Even matches:', even_match)\n gym.print_scores()\n gym.reset_scores()\n even_match = 0\n total_loss = 0\n print(\"games\", i)\n print(\"epsilon\", epsilon)\n if i % copy_step == 0:\n TargetNet.copy_weights(TrainNet)\n if i%50000 == 0 and i != 0:\n plt = plot_grad_flow(TrainNet.model.named_parameters())\n path = \"plot\" + str(i)+ \".png\"\n plt.savefig(path)\n\n", "sub_path": "DQN2.py", "file_name": "DQN2.py", "file_ext": "py", "file_size_in_byte": 15952, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "matplotlib.pyplot.plot", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hlines", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "network.ConnectXNetwork2", "line_number": 33, "usage_type": "call"}, {"api_name": "game_environment.ConnectXEnvironment", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.atleast_2d", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.random.random", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 71, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 72, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 86, "usage_type": "call"}, {"api_name": "network.ConnectXNetwork2", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 97, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 97, "usage_type": "name"}, {"api_name": "torch.nn.MSELoss", "line_number": 98, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 98, "usage_type": "name"}, {"api_name": "game_environment.ConnectXEnvironment", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.random.random", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 117, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 118, "usage_type": "attribute"}, {"api_name": "numpy.atleast_2d", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.atleast_2d", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.random.random", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 155, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 157, "usage_type": "attribute"}, {"api_name": "numpy.atleast_2d", "line_number": 197, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 224, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 228, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 237, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 237, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 238, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 239, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 240, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 241, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 242, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 243, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 254, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 260, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 262, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 263, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 264, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 266, "usage_type": "call"}, {"api_name": "gym.Env", "line_number": 277, "usage_type": "attribute"}, {"api_name": "game_environment.ConnectXEnvironment", "line_number": 279, "usage_type": "call"}, {"api_name": "gym.spaces.Discrete", "line_number": 286, "usage_type": "call"}, {"api_name": "gym.spaces", "line_number": 286, "usage_type": "attribute"}, {"api_name": "gym.spaces.Discrete", "line_number": 287, "usage_type": "call"}, {"api_name": "gym.spaces", "line_number": 287, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 305, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 305, "usage_type": "attribute"}, {"api_name": "agents.verticalBot", "line_number": 307, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 339, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 376, "usage_type": "call"}, {"api_name": "game_environment.ConnectXEnvironment", "line_number": 399, "usage_type": "call"}, {"api_name": "gym.generate_data", "line_number": 404, "usage_type": "call"}, {"api_name": "gym.trainer", "line_number": 407, "usage_type": "attribute"}, {"api_name": "gym.score_list", "line_number": 409, "usage_type": "attribute"}, {"api_name": "gym.trainer", "line_number": 409, "usage_type": "attribute"}, {"api_name": "gym.games_list", "line_number": 410, "usage_type": "attribute"}, {"api_name": "gym.trainer", "line_number": 410, "usage_type": "attribute"}, {"api_name": "gym.change_trainer_at_random", "line_number": 414, "usage_type": "call"}, {"api_name": "gym.print_scores", "line_number": 421, "usage_type": "call"}, {"api_name": "gym.reset_scores", "line_number": 422, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 430, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 432, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 432, "usage_type": "name"}]}
+{"seq_id": "374731632", "text": "import locale\nimport socket\n\nimport psutil\n\nfrom health_check.backends import BaseHealthCheckBackend\n\n\nhost = socket.gethostname()\n\n\nclass DiskUsageHealthCheckBackend(BaseHealthCheckBackend):\n\n def __init__(self, disk_usage_max):\n # type: (int) -> None\n super(DiskUsageHealthCheckBackend, self).__init__()\n self.disk_usage_max = disk_usage_max # type: int\n\n def check_status(self):\n try:\n du = psutil.disk_usage('/')\n if du.percent >= self.disk_usage_max:\n msg = \"{0} {1}% disk usage exceeds {2}%\".format(host, du.percent, self.disk_usage_max) # type: str\n self.add_service_warning(msg, exc_info=False)\n except ValueError as e:\n self.add_service_warning(e)\n\n @property\n def component_name(self):\n # type: () -> str\n return 'disk'\n\n\nclass MemoryUsageHealthCheckBackend(BaseHealthCheckBackend):\n\n def __init__(self, memory_min):\n # type: (int) -> None\n super(MemoryUsageHealthCheckBackend, self).__init__()\n self.memory_min = memory_min # type: int\n\n def check_status(self):\n try:\n memory = psutil.virtual_memory()\n if memory.available < (self.memory_min * 1024 * 1024):\n locale.setlocale(locale.LC_ALL, '')\n avail = int(memory.available / 1024 / 1024) # type: int\n msg = \"{0} {1} MB available RAM below {2} MB\".format(host, avail, self.memory_min) # type: str\n self.add_service_warning(msg, exc_info=False)\n except ValueError as e:\n self.add_service_warning(e)\n\n @property\n def component_name(self):\n # type: () -> str\n return 'memory'\n", "sub_path": "health_check/contrib/psutil/backends.py", "file_name": "backends.py", "file_ext": "py", "file_size_in_byte": 1718, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "socket.gethostname", "line_number": 9, "usage_type": "call"}, {"api_name": "health_check.backends.BaseHealthCheckBackend", "line_number": 12, "usage_type": "name"}, {"api_name": "psutil.disk_usage", "line_number": 21, "usage_type": "call"}, {"api_name": "health_check.backends.BaseHealthCheckBackend", "line_number": 34, "usage_type": "name"}, {"api_name": "psutil.virtual_memory", "line_number": 43, "usage_type": "call"}, {"api_name": "locale.setlocale", "line_number": 45, "usage_type": "call"}, {"api_name": "locale.LC_ALL", "line_number": 45, "usage_type": "attribute"}]}
+{"seq_id": "647900564", "text": "import pytest\nfrom unittest.mock import MagicMock\n\nfrom connections.sim.hw.xbee_module_sim import XBeeModuleSim, FRAME_PARSED_EVENT, SENT_TO_ROCKET_EVENT\nfrom util.event_stats import get_event_stats_snapshot\n\nTEST_GS_ADDR = bytes.fromhex('0013A200400A0127')\nTEST_GS_ADDR_ESCAPED = bytes.fromhex('007D33A200400A0127')\n\n\n@pytest.fixture()\ndef xbee():\n xbee = XBeeModuleSim(TEST_GS_ADDR)\n xbee.rocket_callback = MagicMock()\n xbee.ground_callback = MagicMock()\n yield xbee\n xbee.shutdown()\n\n\ndef test_rocket_rx(xbee):\n test_data = b\"TxData0A\"\n tx_example = bytearray(\n b\"\\x7E\\x00\\x16\\x10\\x01\" + TEST_GS_ADDR_ESCAPED +\n b\"\\xFF\\xFE\\x00\\x00\" + test_data + b\"\\x7D\\x33\"\n )\n\n snapshot = get_event_stats_snapshot()\n xbee.recieved_from_rocket(tx_example)\n\n assert FRAME_PARSED_EVENT.wait(snapshot) == 1\n xbee.ground_callback.assert_called_with(test_data)\n\n\ndef test_ground_rx(xbee):\n snapshot = get_event_stats_snapshot()\n xbee.send_to_rocket(b\"HelloRocket\") # 11 bytes\n\n assert SENT_TO_ROCKET_EVENT.wait(snapshot) == 1\n\n msg = xbee.rocket_callback.call_args[0][0]\n\n # Skip any escape characters\n assert len(msg) - msg.count(b'\\x7D') == 27\n\n assert msg[-12:-1] == b\"HelloRocket\"\n\n\ndef test_rocket_rx_pieces(xbee):\n test_data = b\"TxData0A\"\n tx_1 = bytearray(b\"\\x7E\\x00\\x16\\x10\\x01\") + TEST_GS_ADDR_ESCAPED[:7]\n tx_2 = bytearray(TEST_GS_ADDR_ESCAPED[7:] + b\"\\xFF\\xFE\\x00\\x00\" + test_data + b\"\\x7D\\x33\")\n\n snapshot = get_event_stats_snapshot()\n xbee.recieved_from_rocket(tx_1)\n xbee.recieved_from_rocket(tx_2)\n\n assert FRAME_PARSED_EVENT.wait(snapshot) == 1\n xbee.ground_callback.assert_called_with(test_data)\n\ndef test_rocket_rx_bad_addr(xbee):\n test_data = b\"TxData0A\"\n tx_example = bytearray(\n b\"\\x7E\\x00\\x16\\x10\\x01\" + bytes(8) +\n b\"\\xFF\\xFE\\x00\\x00\" + test_data + b\"\\x3A\"\n )\n\n snapshot = get_event_stats_snapshot()\n xbee.recieved_from_rocket(tx_example)\n\n assert FRAME_PARSED_EVENT.wait(snapshot) == 1\n assert xbee.ground_callback.call_count == 0\n", "sub_path": "tests/test_xbee_module_sim.py", "file_name": "test_xbee_module_sim.py", "file_ext": "py", "file_size_in_byte": 2068, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "connections.sim.hw.xbee_module_sim.XBeeModuleSim", "line_number": 13, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 14, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 15, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 11, "usage_type": "call"}, {"api_name": "util.event_stats.get_event_stats_snapshot", "line_number": 27, "usage_type": "call"}, {"api_name": "connections.sim.hw.xbee_module_sim.FRAME_PARSED_EVENT.wait", "line_number": 30, "usage_type": "call"}, {"api_name": "connections.sim.hw.xbee_module_sim.FRAME_PARSED_EVENT", "line_number": 30, "usage_type": "name"}, {"api_name": "util.event_stats.get_event_stats_snapshot", "line_number": 35, "usage_type": "call"}, {"api_name": "connections.sim.hw.xbee_module_sim.SENT_TO_ROCKET_EVENT.wait", "line_number": 38, "usage_type": "call"}, {"api_name": "connections.sim.hw.xbee_module_sim.SENT_TO_ROCKET_EVENT", "line_number": 38, "usage_type": "name"}, {"api_name": "util.event_stats.get_event_stats_snapshot", "line_number": 53, "usage_type": "call"}, {"api_name": "connections.sim.hw.xbee_module_sim.FRAME_PARSED_EVENT.wait", "line_number": 57, "usage_type": "call"}, {"api_name": "connections.sim.hw.xbee_module_sim.FRAME_PARSED_EVENT", "line_number": 57, "usage_type": "name"}, {"api_name": "util.event_stats.get_event_stats_snapshot", "line_number": 67, "usage_type": "call"}, {"api_name": "connections.sim.hw.xbee_module_sim.FRAME_PARSED_EVENT.wait", "line_number": 70, "usage_type": "call"}, {"api_name": "connections.sim.hw.xbee_module_sim.FRAME_PARSED_EVENT", "line_number": 70, "usage_type": "name"}]}
+{"seq_id": "452856493", "text": "# coding: utf-8\n\nfrom __future__ import absolute_import\nfrom datetime import date, datetime # noqa: F401\n\nfrom typing import List, Dict # noqa: F401\n\nfrom capif_routing_info.models.base_model_ import Model\nfrom capif_routing_info.models.custom_operation import CustomOperation\nfrom capif_routing_info.models.resource import Resource\nfrom capif_routing_info import util\n\nfrom capif_routing_info.models.custom_operation import CustomOperation # noqa: E501\nfrom capif_routing_info.models.resource import Resource # noqa: E501\n\nclass Version(Model):\n \"\"\"NOTE: This class is auto generated by OpenAPI Generator (https://openapi-generator.tech).\n\n Do not edit the class manually.\n \"\"\"\n\n def __init__(self, api_version=None, expiry=None, resources=None, cust_operations=None): # noqa: E501\n \"\"\"Version - a model defined in OpenAPI\n\n :param api_version: The api_version of this Version. # noqa: E501\n :type api_version: str\n :param expiry: The expiry of this Version. # noqa: E501\n :type expiry: datetime\n :param resources: The resources of this Version. # noqa: E501\n :type resources: List[Resource]\n :param cust_operations: The cust_operations of this Version. # noqa: E501\n :type cust_operations: List[CustomOperation]\n \"\"\"\n self.openapi_types = {\n 'api_version': str,\n 'expiry': datetime,\n 'resources': List[Resource],\n 'cust_operations': List[CustomOperation]\n }\n\n self.attribute_map = {\n 'api_version': 'apiVersion',\n 'expiry': 'expiry',\n 'resources': 'resources',\n 'cust_operations': 'custOperations'\n }\n\n self._api_version = api_version\n self._expiry = expiry\n self._resources = resources\n self._cust_operations = cust_operations\n\n @classmethod\n def from_dict(cls, dikt) -> 'Version':\n \"\"\"Returns the dict as a model\n\n :param dikt: A dict.\n :type: dict\n :return: The Version of this Version. # noqa: E501\n :rtype: Version\n \"\"\"\n return util.deserialize_model(dikt, cls)\n\n @property\n def api_version(self):\n \"\"\"Gets the api_version of this Version.\n\n API major version in URI (e.g. v1) # noqa: E501\n\n :return: The api_version of this Version.\n :rtype: str\n \"\"\"\n return self._api_version\n\n @api_version.setter\n def api_version(self, api_version):\n \"\"\"Sets the api_version of this Version.\n\n API major version in URI (e.g. v1) # noqa: E501\n\n :param api_version: The api_version of this Version.\n :type api_version: str\n \"\"\"\n if api_version is None:\n raise ValueError(\"Invalid value for `api_version`, must not be `None`\") # noqa: E501\n\n self._api_version = api_version\n\n @property\n def expiry(self):\n \"\"\"Gets the expiry of this Version.\n\n string with format \\\"date-time\\\" as defined in OpenAPI. # noqa: E501\n\n :return: The expiry of this Version.\n :rtype: datetime\n \"\"\"\n return self._expiry\n\n @expiry.setter\n def expiry(self, expiry):\n \"\"\"Sets the expiry of this Version.\n\n string with format \\\"date-time\\\" as defined in OpenAPI. # noqa: E501\n\n :param expiry: The expiry of this Version.\n :type expiry: datetime\n \"\"\"\n\n self._expiry = expiry\n\n @property\n def resources(self):\n \"\"\"Gets the resources of this Version.\n\n Resources supported by the API. # noqa: E501\n\n :return: The resources of this Version.\n :rtype: List[Resource]\n \"\"\"\n return self._resources\n\n @resources.setter\n def resources(self, resources):\n \"\"\"Sets the resources of this Version.\n\n Resources supported by the API. # noqa: E501\n\n :param resources: The resources of this Version.\n :type resources: List[Resource]\n \"\"\"\n if resources is not None and len(resources) < 1:\n raise ValueError(\"Invalid value for `resources`, number of items must be greater than or equal to `1`\") # noqa: E501\n\n self._resources = resources\n\n @property\n def cust_operations(self):\n \"\"\"Gets the cust_operations of this Version.\n\n Custom operations without resource association. # noqa: E501\n\n :return: The cust_operations of this Version.\n :rtype: List[CustomOperation]\n \"\"\"\n return self._cust_operations\n\n @cust_operations.setter\n def cust_operations(self, cust_operations):\n \"\"\"Sets the cust_operations of this Version.\n\n Custom operations without resource association. # noqa: E501\n\n :param cust_operations: The cust_operations of this Version.\n :type cust_operations: List[CustomOperation]\n \"\"\"\n if cust_operations is not None and len(cust_operations) < 1:\n raise ValueError(\"Invalid value for `cust_operations`, number of items must be greater than or equal to `1`\") # noqa: E501\n\n self._cust_operations = cust_operations\n", "sub_path": "services/TS29222_CAPIF_Routing_Info_API/capif_routing_info/models/version.py", "file_name": "version.py", "file_ext": "py", "file_size_in_byte": 5105, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "capif_routing_info.models.base_model_.Model", "line_number": 16, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 36, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 37, "usage_type": "name"}, {"api_name": "capif_routing_info.models.resource.Resource", "line_number": 37, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 38, "usage_type": "name"}, {"api_name": "capif_routing_info.models.custom_operation.CustomOperation", "line_number": 38, "usage_type": "name"}, {"api_name": "capif_routing_info.util.deserialize_model", "line_number": 62, "usage_type": "call"}, {"api_name": "capif_routing_info.util", "line_number": 62, "usage_type": "name"}]}
+{"seq_id": "56966628", "text": "import argparse\nimport yaml\nimport torch\nimport wandb\n\nfrom dr_spaam.dr_spaam.data_handle.drow_dataset import get_dataloader\nfrom dr_spaam.dr_spaam.pipeline.pipeline import Pipeline\nfrom dr_spaam.dr_spaam.model.get_model import get_model\n\n\n# Run benchmark to select fastest implementation of ops.\ntorch.backends.cudnn.benchmark = True\n\nparser = argparse.ArgumentParser(description=\"arg parser\")\nparser.add_argument(\n \"--cfg\", type=str, required=True, help=\"configuration of the experiment\"\n)\nparser.add_argument(\"--ckpt\", type=str, required=False, default=None)\nparser.add_argument(\"--cont\", default=False, action=\"store_true\")\nparser.add_argument(\"--tmp\", default=False, action=\"store_true\")\nparser.add_argument(\"--evaluation\", default=False, action=\"store_true\")\nargs = parser.parse_args()\n\nwith open(args.cfg, \"r\") as f:\n cfg = yaml.safe_load(f)\n cfg[\"pipeline\"][\"Logger\"][\"backup_list\"].append(args.cfg)\n if args.tmp:\n cfg[\"pipeline\"][\"Logger\"][\"tag\"] += \"_TMP\"\n\nproject = \"dr_spaam\" if not args.tmp else \"tmp\"\nwandb.init(\n project=project, name=cfg[\"pipeline\"][\"Logger\"][\"tag\"], sync_tensorboard=True\n)\nwandb.config.update(cfg)\n\nmodel = get_model(cfg[\"model\"])\nmodel.cuda()\nwandb.watch(model, log=\"all\")\n\npipeline = Pipeline(model, cfg[\"pipeline\"])\n\nif args.ckpt:\n pipeline.load_ckpt(model, args.ckpt)\nelif args.cont and pipeline.sigterm_ckpt_exists():\n pipeline.load_sigterm_ckpt(model)\n\n# training\nif not args.evaluation:\n # main train loop\n train_loader = get_dataloader(\n split=\"train\", cfg=cfg[\"dataset\"], **cfg[\"dataloader\"]\n )\n val_loader = get_dataloader(split=\"val\", cfg=cfg[\"dataset\"], **cfg[\"dataloader\"])\n status = pipeline.train(model, train_loader, val_loader)\n\n # test after training\n if not status:\n test_loader = get_dataloader(\n split=\"test\", batch_size=1, num_workers=1, cfg=cfg[\"dataset\"]\n )\n pipeline.evaluate(model, test_loader, tb_prefix=\"TEST\")\n\n# evaluation\nelse:\n val_loader = get_dataloader(\n split=\"val\", batch_size=1, num_workers=1, cfg=cfg[\"dataset\"]\n )\n pipeline.evaluate(model, val_loader, tb_prefix=\"VAL\")\n\n test_loader = get_dataloader(\n split=\"test\", batch_size=1, num_workers=1, cfg=cfg[\"dataset\"]\n )\n pipeline.evaluate(model, test_loader, tb_prefix=\"TEST\")\n\npipeline.close()\n\n# force wandb to push logs to server\n# https://github.com/wandb/client/issues/554\nwandb.log({\"dummy\": 1.0}, commit=False)\n", "sub_path": "depracted_scripts/train.py", "file_name": "train.py", "file_ext": "py", "file_size_in_byte": 2460, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "torch.backends", "line_number": 12, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 14, "usage_type": "call"}, {"api_name": "yaml.safe_load", "line_number": 25, "usage_type": "call"}, {"api_name": "wandb.init", "line_number": 31, "usage_type": "call"}, {"api_name": "wandb.config.update", "line_number": 34, "usage_type": "call"}, {"api_name": "wandb.config", "line_number": 34, "usage_type": "attribute"}, {"api_name": "dr_spaam.dr_spaam.model.get_model.get_model", "line_number": 36, "usage_type": "call"}, {"api_name": "wandb.watch", "line_number": 38, "usage_type": "call"}, {"api_name": "dr_spaam.dr_spaam.pipeline.pipeline.Pipeline", "line_number": 40, "usage_type": "call"}, {"api_name": "dr_spaam.dr_spaam.data_handle.drow_dataset.get_dataloader", "line_number": 50, "usage_type": "call"}, {"api_name": "dr_spaam.dr_spaam.data_handle.drow_dataset.get_dataloader", "line_number": 53, "usage_type": "call"}, {"api_name": "dr_spaam.dr_spaam.data_handle.drow_dataset.get_dataloader", "line_number": 58, "usage_type": "call"}, {"api_name": "dr_spaam.dr_spaam.data_handle.drow_dataset.get_dataloader", "line_number": 65, "usage_type": "call"}, {"api_name": "dr_spaam.dr_spaam.data_handle.drow_dataset.get_dataloader", "line_number": 70, "usage_type": "call"}, {"api_name": "wandb.log", "line_number": 79, "usage_type": "call"}]}
+{"seq_id": "2749200", "text": "import sys, os\n\nos.environ.setdefault(\"DJANGO_SETTINGS_MODULE\", \"dictionarysite.settings\")\n\nimport django\ndjango.setup()\n\nfrom dictionary.models import Translate\n\n\ndef save_word_pair_from_row(row, id):\n if row[1]:\n t = Translate()\n t.id = id\n t.languagefrom= row[0]\n t.texttranslatefrom = row[1]\n t.languageto = row[2]\n t.texttranslateto = row[3]\n t.save()\n\n\ndef delete(fromid):\n rows = Translate.objects.filter(id__gte=fromid)\n for r in rows:\n if fromid % 1000 == 0:\n print(fromid)\n r.delete()\n print(\"DONE\")\n\n\nif __name__ == \"__main__\":\n id = 1\n # delete(id)\n with open('wikt2dict_samplev2.tsv', 'r', encoding=\"utf8\") as tsvfile:\n # with open('test2.tsv', 'r', encoding=\"utf8\") as tsvfile:\n for line in tsvfile:\n # line = line.replace('\\r\\n', '\\n')\n line = line.split('\\t')\n save_word_pair_from_row(line, id)\n if id%1000 == 0:\n print(id)\n id += 1\n tsvfile.close()\n", "sub_path": "loadTranslate.py", "file_name": "loadTranslate.py", "file_ext": "py", "file_size_in_byte": 1044, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.environ.setdefault", "line_number": 3, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 3, "usage_type": "attribute"}, {"api_name": "django.setup", "line_number": 6, "usage_type": "call"}, {"api_name": "dictionary.models.Translate", "line_number": 13, "usage_type": "call"}, {"api_name": "dictionary.models.Translate.objects.filter", "line_number": 23, "usage_type": "call"}, {"api_name": "dictionary.models.Translate.objects", "line_number": 23, "usage_type": "attribute"}, {"api_name": "dictionary.models.Translate", "line_number": 23, "usage_type": "name"}]}
+{"seq_id": "168035511", "text": "import os.path\nimport sys\nimport matplotlib.pylab as plt\nimport numpy as np\nimport math\nimport nibabel as nib\nfrom sklearn.decomposition.pca import PCA\n# from skimage import filter\nimport string\nfrom sklearn.preprocessing import StandardScaler\nfrom scipy.signal import medfilt\nfrom skimage import exposure\nimport skimage.transform\nimport scipy.io as sio\nfrom configparser import ConfigParser\n#from visualizer import multi_slice_viewer\nfrom sklearn.cross_decomposition import PLSRegression\nfrom skimage.measure import structural_similarity as ssim\n\ncp = ConfigParser()\ncp.read(\"config.ini\")\ncfg = cp['DEFAULT']\nfor d in cfg:\n exec(\"%s = %s\"%(d, cfg[d]))\nsource_dvf += 'iter%d/' % nitern\n\n\n# read deformation vector fields and store in DVFs\nfiles = os.listdir(source_dvf)\nfpaths = []\nfor ifile in files:\n if (ifile.find('_dvf_pc_iter%d.nii' % nitern) > 0):\n fpaths.append(('%s%s' % (source_dvf, ifile)))\nfpaths = np.sort(fpaths)\nDVFs = []\nfor ifile in fpaths:\n DVFs.append(nib.load(ifile).get_data())\nDVFs = np.array(DVFs)\n# multi_slice_viewer(DVFs[:,:,:,0,0,0])\n[Nt, Nx, Ny, n1, n2, ndim] = np.shape(DVFs)\nprint(np.shape(DVFs))\n# collapse to 1D vector form\ndvf_clpsd = np.reshape(DVFs, [Nt, Nx * Ny * n1 * n2 * ndim])\n# standardize and calculate principal components\nscaler = StandardScaler()\ndvf_pr = scaler.fit_transform(dvf_clpsd)\npca = PCA(n_components=ncomp)\npca.fit(dvf_pr)\ny_training = pca.transform(dvf_pr)\n\n# if you want to recover data\nrecovered = pca.inverse_transform(y_training)\n\nprint(\"PCA COMPONENTS: \" + str(np.array(pca.components_).shape))\n\n\nrecovered = scaler.inverse_transform(recovered)\n\n### SURFACE DATA\n# surf_data = sio.loadmat('C:\\ps_lj/files/new_data\\WS_fx3_3comps/surf_training.mat')\nsurf_data = sio.loadmat(matpath)\nsurfaces = surf_data['surfaces_stored'][0]\nX_train = surfaces[1]\n# multi_slice_viewer(X_train)\n# print(np.array(X_train).shape)\nX_pred = surfaces[2]\ntime_fl = surf_data['time_fl'][0][1]\ntime_vrt = surf_data['vrt_time'][0][1]\n\n####\n# INTERPOLATE THE DVF COMPONENTS TO SURFACE TIME SPACE. CAN SWITCH TO SPLINE IF NEEDED.\ny_training_int = []\nfor ipc in range(np.shape(y_training)[1]):\n y_training_int.append(np.interp(time_vrt.ravel(), time_fl.ravel()[1::], y_training[:, ipc].ravel()))\ny_training_int = np.array(y_training_int)\n\n# CROP DATA!\ny_old = [0]\ny_cropped = []\nx_cropped = []\nfor idx, Y_ in enumerate(np.transpose(y_training_int)):\n if y_old[0] != Y_[0]:\n y_cropped.append(Y_)\n x_cropped.append(X_train[idx])\n #print(Y_, X_train[idx,10,10])\n y_old = Y_\n #print(idx)\n #print(idx, Y_)\n #print(np.array(X_train).shape)\n#plt.plot(np.array(y_cropped))\n#plt.show()\ny_training_int = np.transpose(np.array(y_cropped))\nX_train = np.array(x_cropped)\n#print('y_cropped shape ' + str(np.array(y_cropped).shape))\n#print('Y_training_int shape ' + str(y_training_int.shape))\n#print(\"x_cropped shape\" + str(np.array(x_cropped).shape))\n#print(\"xtrain shape: \" + str(np.array(X_train).shape))\n\n\"\"\"print(y_training_int.shape)\nplt.plot(y_training)\nplt.show()\nplt.plot(np.transpose(y_training_int))\nplt.show()\"\"\"\n\n\n# NORMALIZATION\ny_training_int = np.transpose(y_training_int)\nX_norm = (X_train - np.average(X_train)) / np.std(X_train)\nX_norm_pred = (X_pred - np.average(X_train)) / np.std(X_train)\n\nY_norm = (y_training_int - np.average(y_training_int)) / np.std(y_training_int)\n(examples, x__, y__) = X_norm.shape\nval = int(examples * 0.1)\n#plt.plot(y_training_int)\n#plt.show()\nprint(\"exaples: %s \\nval: %s \\ntrain :%s\" % (examples, val, examples - val - 1))\n\n\n\"\"\"def getBatch():\n X = [] #\n Y = []\n for i in iter(np.random.randint(0, int(examples - val - 1), size=(10))):\n X.append(X_norm[i])\n Y.append(Y_norm[i])\n return np.array(X) + np.random.normal(0, scale=0.1, size=np.array(X).shape), np.array(Y)\"\"\"\n\n\ntemp_resolution = 3\ntemp_steps = 3\nmin_ = temp_resolution*temp_steps\n\nprint(temp_steps, temp_resolution, min_)\n\ndef getBatch():\n X = [] #\n Y = []\n\n for i in iter(np.random.randint(min_, int(examples - val - 1), size=(10))):\n Xo = []\n for t in range(temp_steps):\n Xo.append(X_norm[i-t*temp_resolution])\n #print(\"sample: \" + str(i-t*temp_resolution))\n Y.append(Y_norm[i])\n X.append(np.transpose(np.array(Xo)))\n #print(np.array(X).shape)\n return np.array(X), np.array(Y)\n\nx_, y_ = getBatch()\nprint(np.array(x_).shape)\n\ndef getValBatch():\n X = []\n Y = []\n for i in iter(np.random.randint(examples - val - 1 + min_, examples - 1, size=(10))):\n Xo = []\n for t in range(temp_steps):\n Xo.append(X_norm[i - t * temp_resolution])\n # print(\"sample: \" + str(i-t*temp_resolution))\n Y.append(Y_norm[i])\n X.append(np.transpose(np.array(Xo)))\n \"\"\"X = []\n Y = []\n for i in iter(np.random.randint(examples - val - 1 + min_, examples - 1, size=(10))):\n X.append(X_norm[i])\n Y.append(Y_norm[i])\"\"\"\n\n return np.array(X), np.array(Y)\na,b = getValBatch()\nprint()\n\nxp = iter(X_norm_pred)\ndef getXpred():\n global xp\n Xpred = []\n try:\n to_append = xp.__next__()\n except:\n print(\"reset getXpred counter\")\n xp = iter(X_norm_pred)\n to_append = xp.__next__()\n for i in range(batchsize):\n Xo = []\n for d in range(temp_steps):\n Xo.append(to_append)\n Xpred.append(to_append)\n Xpred = np.expand_dims(Xpred,-1)\n return np.array(Xpred)\nprint(\"xpredshape:\" + str(getXpred().shape))\n\ndef getNoPred():\n (need, sth, sthelse) = np.array(X_pred).shape\n return need\n\nplt.plot(Y_norm)\nplt.show()\n\nprint(\"X_pred size: \" + str(getNoPred())) # 542 14 32\n#print(np.average(y_training_int, axis=0))\ndef getYmean():\n return np.average(y_training_int)\ndef getYstd():\n return np.std(y_training_int)\n\n\n\"\"\"plt.hist(y_training_int) # hist are to check the distribution\nplt.show()\n\nplt.hist(infer_Y)\nplt.show()\nprint(X_norm_pred.shape)\nprint(X_pred.shape)\nfor i in range(1000):\n print(getXpred().shape)\"\"\"\n\n####USE VARIABLES X_train and y_training_int to train your network. Then predict using X_pred as input\n# inputs:\n# X_train : 481 observations of 10x35 surfaces. <<<<< SIZE OF THE SURFACES WILL CHANGE FOR EACH DATASET !!!!!\n# y_training_int: first 3 components of DVF\n\n\n# assume we have Y_pred as output. Need to interpolate back to fluoro time\n\n\"\"\" BACK TO FLUORO INTERPOLATION - ONLY USE FOR POSTPROCESSING\"\"\"\n\"\"\"\ntime_fl_pred = surf_data['time_fl'][0][2]\ntime_vrt_pred = surf_data['vrt_time'][0][2]\nprint(np.array(time_fl).shape)\nprint(np.array(time_vrt).shape)\nprint(np.array(time_fl_pred).shape)\nprint(np.array(time_vrt_pred).shape)\n\ninfer_Y = np.load(\"C:/ps_lj/components_avg_5.99051_1535643741.1559029.npy\")\nprint(infer_Y.shape)\n\ny_pred = []\nfor ipf in range(np.shape(infer_Y)[1]):\n y_pred.append(np.interp(time_fl_pred.ravel(),time_vrt_pred.ravel(), infer_Y[:, ipf].ravel()))\ny_pred = np.transpose(np.array(y_pred))\nnp.save('interpolatedComponents.npy', y_pred)\n#plt.plot(y_pred[:,0])\n#plt.plot(y_pred[:,1])\n#plt.plot(y_pred[:,2])\n##plt.show()\nnew = pca.inverse_transform(y_pred)\nnew = scaler.inverse_transform(new)\nrecon = []\nfor dvf in new:\n recon.append(np.reshape(dvf, newshape=(Nx, Ny, n1, n2, ndim)))\nrecon = np.array(recon)\nnp.save(\"DVFs.npy\", recon)\nprint(\"reconshape \" + str(recon.shape))\n#multi_slice_viewer(recon[:,:,:,0,0,0])\nY0 = []\nY1 = []\nfor f in recon:\n Y0.append(f[:,:,0,0,0] - np.transpose([np.arange(256) for i in range(342)]))\n Y1.append(f[:,:,0,0, 1] - [np.arange(342) for i in range(256)])\nmulti_slice_viewer(np.array(Y0))\nmulti_slice_viewer(np.array(Y1))\n\"\"\"\n\n\n#########################SOME FUNCTIONS FOR NIFTIREG THAT I WROTE, THERE IS A PYTHON PACKAGE PYNIFTI BUT I COULDN'T GET IT TO WORK ON THE SERVER\n\ndef reg_f3d(fixed, moving, res, cpp, be, sx, dvf, key=''):\n if (os.path.isfile(dvf)):\n os.remove(dvf)\n if (os.path.isfile(cpp)):\n os.remove(cpp)\n cmd = ('reg_f3d -ref %s -flo %s -be %f -sx %d -res %s -cpp %s %s ' % (fixed, moving, be, sx, res, cpp, key))\n print(cmd)\n os.system(cmd)\n cmd = ('reg_transform -ref %s -cpp2def %s %s' % (moving, cpp, dvf))\n print(cmd)\n os.system(cmd)\n if (os.path.isfile(res)):\n os.remove(res)\n cmd = ('reg_resample -ref %s -flo %s -def %s -res %s' % (moving, moving, dvf, res))\n print(cmd)\n os.system(cmd)\n\n # read dvf and output\n DVF = nib.load(dvf)\n RES = nib.load(res)\n return [RES, DVF]\n\n\ndef compose_dvf(fixed, dvf1, dvf2, dvf_final, res):\n if (os.path.isfile(dvf_final)):\n os.remove(dvf_final)\n if (os.path.isfile(res)):\n os.remove(res)\n cmd = ('reg_transform -ref %s -comp3 %s %s %s' % (fixed, dvf1, dvf2, dvf_final))\n print(cmd)\n os.system(cmd)\n cmd = ('reg_resample -ref %s -flo %s -def %s -res %s' % (fixed, fixed, dvf_final, res))\n os.system(cmd)\n DVF = nib.load(dvf_final)\n RES = nib.load(res)\n return [RES, DVF]\n\n\ndef apply_dvf(moving, dvf, res_file):\n if (os.path.isfile(res_file)):\n os.remove(res_file)\n cmd = ('reg_resample -ref %s -flo %s -def %s -res %s' % (moving, moving, dvf, res_file))\n os.system(cmd)\n\n RES = nib.load(res_file)\n return RES\n\n\ndef def_2_disp(moving, dvf, dsp):\n if (os.path.isfile(dsp)):\n os.remove(dsp)\n cmd = ('reg_transform -ref %s -def2disp %s %s' % (moving, dvf, dsp))\n os.system(cmd)\n RES = nib.load(dsp)\n return RES\n\n\ndef disp_2_def(moving, dsp, dvf):\n if (os.path.isfile(dvf)):\n os.remove(dvf)\n cmd = ('reg_transform -ref %s -disp2def %s %s' % (moving, dsp, dvf))\n os.system(cmd)\n RES = nib.load(dvf)\n return RES\n\n", "sub_path": "preprocessing_temporal.py", "file_name": "preprocessing_temporal.py", "file_ext": "py", "file_size_in_byte": 9662, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "configparser.ConfigParser", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.listdir", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "name"}, {"api_name": "numpy.sort", "line_number": 34, "usage_type": "call"}, {"api_name": "nibabel.load", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 43, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 45, "usage_type": "call"}, {"api_name": "sklearn.decomposition.pca.PCA", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 54, "usage_type": "call"}, {"api_name": "scipy.io.loadmat", "line_number": 61, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 61, "usage_type": "name"}, {"api_name": "numpy.shape", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 138, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 154, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 191, "usage_type": "call"}, {"api_name": "matplotlib.pylab.plot", "line_number": 194, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 194, "usage_type": "name"}, {"api_name": "matplotlib.pylab.show", "line_number": 195, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 195, "usage_type": "name"}, {"api_name": "numpy.average", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 202, "usage_type": "call"}, {"api_name": "os.path.path.isfile", "line_number": 266, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 266, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 266, "usage_type": "name"}, {"api_name": "os.path.remove", "line_number": 267, "usage_type": "call"}, {"api_name": "os.path", "line_number": 267, "usage_type": "name"}, {"api_name": "os.path.path.isfile", "line_number": 268, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 268, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 268, "usage_type": "name"}, {"api_name": "os.path.remove", "line_number": 269, "usage_type": "call"}, {"api_name": "os.path", "line_number": 269, "usage_type": "name"}, {"api_name": "os.path.system", "line_number": 272, "usage_type": "call"}, {"api_name": "os.path", "line_number": 272, "usage_type": "name"}, {"api_name": "os.path.system", "line_number": 275, "usage_type": "call"}, {"api_name": "os.path", "line_number": 275, "usage_type": "name"}, {"api_name": "os.path.path.isfile", "line_number": 276, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 276, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 276, "usage_type": "name"}, {"api_name": "os.path.remove", "line_number": 277, "usage_type": "call"}, {"api_name": "os.path", "line_number": 277, "usage_type": "name"}, {"api_name": "os.path.system", "line_number": 280, "usage_type": "call"}, {"api_name": "os.path", "line_number": 280, "usage_type": "name"}, {"api_name": "nibabel.load", "line_number": 283, "usage_type": "call"}, {"api_name": "nibabel.load", "line_number": 284, "usage_type": "call"}, {"api_name": "os.path.path.isfile", "line_number": 289, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 289, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 289, "usage_type": "name"}, {"api_name": "os.path.remove", "line_number": 290, "usage_type": "call"}, {"api_name": "os.path", "line_number": 290, "usage_type": "name"}, {"api_name": "os.path.path.isfile", "line_number": 291, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 291, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 291, "usage_type": "name"}, {"api_name": "os.path.remove", "line_number": 292, "usage_type": "call"}, {"api_name": "os.path", "line_number": 292, "usage_type": "name"}, {"api_name": "os.path.system", "line_number": 295, "usage_type": "call"}, {"api_name": "os.path", "line_number": 295, "usage_type": "name"}, {"api_name": "os.path.system", "line_number": 297, "usage_type": "call"}, {"api_name": "os.path", "line_number": 297, "usage_type": "name"}, {"api_name": "nibabel.load", "line_number": 298, "usage_type": "call"}, {"api_name": "nibabel.load", "line_number": 299, "usage_type": "call"}, {"api_name": "os.path.path.isfile", "line_number": 304, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 304, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 304, "usage_type": "name"}, {"api_name": "os.path.remove", "line_number": 305, "usage_type": "call"}, {"api_name": "os.path", "line_number": 305, "usage_type": "name"}, {"api_name": "os.path.system", "line_number": 307, "usage_type": "call"}, {"api_name": "os.path", "line_number": 307, "usage_type": "name"}, {"api_name": "nibabel.load", "line_number": 309, "usage_type": "call"}, {"api_name": "os.path.path.isfile", "line_number": 314, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 314, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 314, "usage_type": "name"}, {"api_name": "os.path.remove", "line_number": 315, "usage_type": "call"}, {"api_name": "os.path", "line_number": 315, "usage_type": "name"}, {"api_name": "os.path.system", "line_number": 317, "usage_type": "call"}, {"api_name": "os.path", "line_number": 317, "usage_type": "name"}, {"api_name": "nibabel.load", "line_number": 318, "usage_type": "call"}, {"api_name": "os.path.path.isfile", "line_number": 323, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 323, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 323, "usage_type": "name"}, {"api_name": "os.path.remove", "line_number": 324, "usage_type": "call"}, {"api_name": "os.path", "line_number": 324, "usage_type": "name"}, {"api_name": "os.path.system", "line_number": 326, "usage_type": "call"}, {"api_name": "os.path", "line_number": 326, "usage_type": "name"}, {"api_name": "nibabel.load", "line_number": 327, "usage_type": "call"}]}
+{"seq_id": "513537715", "text": "\"\"\" Unit tests for Synergy Device. \"\"\"\nimport logging\nfrom mock import Mock, call, mock_open\nimport os\nimport sys\nimport copy\nimport time\nfrom tng_sl.device.endpoint.synergylite.synergylite_extended import (\n SynergyLiteExtended, media_available, verify_media_all_devices,\n wait_for_call_states, UnexpectedIceError, value_from_name_or_index)\nfrom tng_sl.plugins.synergylite_cli_common import (\n SynergyLiteCommandParser, SynergyLiteCommandHandler)\nfrom tng_sl.plugins.synergylite_cf_common import SynergyLiteCfNamespace\nfrom tng_sl.plugins.synergylite3pcc_ui_common import SynergyLite3pccUiNamespace\nfrom tng.error import TngError, TimeoutError\n\nfrom twisted.trial.unittest import TestCase\nfrom tng_sl.test.utilities import Clock\n\nfrom tng.frontend.engine import Engine\nfrom tng.device.endpoint.ata import ATA\nfrom tng_sl.plugins.ata_cli_common import ATACommandHandler\n\ntry:\n from unittest.mock import patch\nexcept ImportError:\n from mock import patch\n\nif sys.version_info.major == 2:\n import __builtin__ as builtins\nelse:\n import builtins\n\nlog = logging.getLogger('test_synergylite')\n\nGENERIC_MAC_ADDRESS = '01:02:03:04:05:06'\n\n\nclass MockATA(ATA, SynergyLiteExtended):\n def __init__(self, ip, *args, **kwargs):\n super(MockATA, self).__init__(ip, *args, **kwargs)\n\n\nclass TestLogFailureToDevice(TestCase):\n\n def test_basic(self):\n device = Mock(spec=SynergyLiteExtended)\n device.log = Mock(spec=logging.Logger)\n # should be copied over by spec?\n device._log_failure_to_device = (\n lambda test, error: SynergyLiteExtended._log_failure_to_device(\n device, test, error))\n\n device._log_failure_to_device('Some Random Test', 'Some Random Error')\n msg = 'Failed TNG test Some Random Test with error: Some Random Error'\n\n self.assertEqual(\n device.log_to_device.mock_calls,\n [call(msg)])\n self.assertEqual(\n device.log.warn.mock_calls,\n [call('_log_failure_to_device(%r)', msg)])\n self.assertEqual(device.log.debug.mock_calls, [])\n\n\nclass TestLogToDevice(TestCase):\n def setUp(self):\n # hardcode clock time so output is predictable\n self.clock = Clock(1458253930.25)\n patcher = patch('time.time', self.clock.time)\n patcher.start()\n self.addCleanup(patcher.stop)\n # standardize TZ\n orig_tz = os.environ.get('TZ', '')\n os.environ['TZ'] = 'US/Central'\n\n def restore_tz():\n os.environ['TZ'] = orig_tz\n self.addCleanup(restore_tz)\n\n self.device = Mock(spec=SynergyLiteExtended)\n self.device.cli = Mock(spec=SynergyLiteCommandHandler)\n self.device.log = Mock(spec=logging.Logger)\n self.device.debugsh = Mock()\n self.device.debugsh.plugin = Mock()\n self.device.debugsh.plugin.activated = True\n self.device.is_dev_phone = False\n\n # should be copied over by spec?\n self.device._log_max_chars_per_syslog = (\n SynergyLiteExtended._log_max_chars_per_syslog)\n self.device._log_max_msg_len = (\n SynergyLiteExtended._log_max_msg_len)\n self.device.log_to_device = (\n lambda msg: SynergyLiteExtended.log_to_device(self.device, msg))\n self.device._prefix = (\n SynergyLiteExtended._prefix)\n\n self.msg_prefix = '\"\\'!@$#^&*():;Test Message'\n self.cleaned_msg_prefix = (\n \"@2016-03-17 17:32:10,250 \"\n \"''!@$#^&*():;Test Message\")\n self.first_x_count = 35\n\n def test_basic(self):\n # make call we are testing\n msg = (self.msg_prefix + 'X' * (\n SynergyLiteExtended._log_max_chars_per_syslog))\n self.device.log_to_device(msg)\n # Validate:\n\n # * Ensure logging makes it to device log by changing double\n # quotes changed to single.\n # * Add Host TNG timestamp at start of log sent to device.\n # * Ensure SynergyLiteExtended._log_max_chars_per_syslog being honored\n # by chunking long logs into multiple syslog writes.\n\n self.assertEqual(\n self.device.cli.exec_command.mock_calls,\n [call(\n \"syslog write TNG \" + self.cleaned_msg_prefix +\n 'X' * self.first_x_count),\n call(\n 'syslog write TNG ' +\n 'X' * (\n SynergyLiteExtended._log_max_chars_per_syslog -\n self.first_x_count))])\n\n self.assertEqual(\n self.device.log.debug.mock_calls,\n [call('log_to_device: ' + repr(\n self.cleaned_msg_prefix +\n 'X' * SynergyLiteExtended._log_max_chars_per_syslog))])\n\n def test_too_long(self):\n # make call we are testing\n num_x = (\n SynergyLiteExtended._log_max_msg_len -\n len(self.cleaned_msg_prefix) + 1)\n msg = (self.msg_prefix + 'X' * num_x)\n self.device.log_to_device(msg)\n # Validate:\n\n # * Ensure logging makes it to device log by changing double\n # quotes changed to single.\n # * Add Host TNG timestamp at start of log sent to device.\n # * Ensure SynergyLiteExtended._log_max_chars_per_syslog being honored\n # by chunking long logs into multiple syslog writes.\n\n self.assertEqual(\n self.device.log.warn.mock_calls,\n [call('log_to_device():Truncating msg of len 426 to len 425')])\n\n # log_to_device debug should have thrown away last char\n self.assertEqual(\n self.device.log.debug.mock_calls,\n [call('log_to_device: ' + repr(\n self.cleaned_msg_prefix +\n 'X' * (num_x - 1)))])\n\n calls_list = [\n call(\n \"syslog write TNG \" + self.cleaned_msg_prefix +\n 'X' * self.first_x_count)]\n\n # -1 is because the intent is that the string is one too long\n remaining = num_x - self.first_x_count - 1\n while remaining > 0:\n this_chunk = SynergyLiteExtended._log_max_chars_per_syslog\n if remaining < this_chunk:\n this_chunk = remaining\n remaining -= this_chunk\n calls_list.append(call(\n 'syslog write TNG ' +\n 'X' * this_chunk))\n\n self.assertEqual(\n self.device.cli.exec_command.mock_calls,\n calls_list)\n\n def test_too_long_for_dev_phone(self):\n num_x = (\n SynergyLiteExtended._log_max_msg_len -\n len(self.cleaned_msg_prefix) + 1)\n msg = self.msg_prefix + 'X' * num_x\n\n self.device.is_dev_phone = True\n self.device.log_to_device(msg)\n\n # Validate:\n # Ensure entire command not be split into multiple parts.\n self.assertEqual(\n self.device.cli.exec_command.mock_calls,\n [call(\n 'syslog write TNG ' + self.cleaned_msg_prefix + 'X' * num_x)])\n\n def test_no_debugsh(self):\n self.device.debugsh._plugin.activated = False\n msg = (self.msg_prefix + 'X' * (\n SynergyLiteExtended._log_max_chars_per_syslog))\n self.device.log_to_device(msg)\n\n self.device.cli.exec_command.assert_not_called()\n self.device.log.debug.assert_called_with(\n 'No log_to_device() because debugsh plugin is not activated.')\n\n\nclass unmockedSLX(SynergyLiteExtended):\n '''need a full implementation to test the properties'''\n is_3pcc = None\n _kwargs = {}\n log_parser_server_ip = False\n log = Mock(spec=logging.Logger)\n\n def add_test_method_failure_callback(self, *args, **kwargs):\n pass\n\n\nclass SynergyLiteExtendedTestCase(TestCase):\n def build_generic_mock_synergy(self, method_name, target=None):\n target = target or Mock(spec=SynergyLiteExtended)\n target_method = getattr(SynergyLiteExtended, method_name)\n setattr(target, method_name, lambda *args, **kwargs: target_method(\n target, *args, **kwargs))\n target.log = Mock(spec=logging.Logger)\n target.ICE_OPTIMAL_PATHS = SynergyLiteExtended.ICE_OPTIMAL_PATHS\n target.ICE_TRAVERSAL_MODES = SynergyLiteExtended.ICE_TRAVERSAL_MODES\n target.ICE_STATE_NAMES = SynergyLiteExtended.ICE_STATE_NAMES\n target.wait_until_up = Mock()\n target.wait_until_down = Mock()\n\n return target\n\n\nclass TestSynergyLiteExtended(SynergyLiteExtendedTestCase):\n\n def setUp(self):\n self.target = Mock(spec=SynergyLiteExtended)\n self.target.cf = Mock(spec=SynergyLiteCfNamespace)\n self.target.cli = Mock(spec=SynergyLiteCommandHandler)\n self.target.cli.parser = SynergyLiteCommandParser()\n self.target.handle_startup = Mock()\n self.target.handle_shutdown = Mock()\n self.target._restart_action = Mock()\n self.target.http = Mock()\n self.target.dbus = Mock()\n self.target.log = Mock(spec=logging.Logger)\n self.target.credentials = Mock()\n self.target.ui = Mock(spec=SynergyLite3pccUiNamespace)\n\n self.jvmVal = {\n 'Bypass tftp addresses': 'false',\n 'Collabration Edge': '',\n 'Device Access Mode': 'Enterprise (deploy-mode:1)',\n 'Enable Password Persistent': 'false',\n 'Is Login window running': 'false',\n 'PRT URL': '',\n 'Registered': 'true',\n 'Saved Service Domain': '',\n 'State': 'idle',\n 'TFTP Available': 'true'\n }\n self.target._wait_patch = Mock()\n wait_patcher = patch(\n 'tng_sl.device.endpoint.synergylite.synergylite_extended.tng_wait',\n self.target._wait_patch)\n wait_patcher.start()\n self.addCleanup(wait_patcher.stop)\n\n self.clock = Clock()\n time_patcher = patch(\n 'tng.frontend.timing.time.time', self.clock.seconds)\n time_patcher.start()\n self.addCleanup(time_patcher.stop)\n\n sleep_patcher = patch(\n 'tng.frontend.timing.time.sleep', self.clock.sleep)\n sleep_patcher.start()\n self.addCleanup(sleep_patcher.stop)\n\n @patch(\n 'tng_sl.device.endpoint.synergylite.synergylite_extended.requests')\n def test__handle_test_failure(self, requests):\n target = self.build_generic_mock_synergy(\n '_handle_test_failure', self.target)\n target.gen_prt_log.return_value = False\n target.log_parser_server_ip = None\n target._handle_test_failure('Unused', 'Unused')\n target.gen_prt_log.assert_called_once_with(prt_type='FULL')\n requests.get.assert_not_called()\n\n target.gen_prt_log.reset_mock()\n target.gen_prt_log.return_value = True\n target._handle_test_failure('Unused', 'Unused')\n target.gen_prt_log.assert_called_once_with(prt_type='FULL')\n requests.get.assert_not_called()\n\n target.gen_prt_log.reset_mock()\n target.log_parser_server_ip = '1.1.1.1'\n target.get_ip = Mock()\n target.get_ip.return_value = '2.2.2.2'\n target._handle_test_failure('Unused', 'Unused')\n target.gen_prt_log.assert_called_once_with(prt_type='FULL')\n requests.get.assert_called_once_with(\n url='http://1.1.1.1/snapshot/log?ip=2.2.2.2')\n\n def test__get_stream_state(self):\n target = self.build_generic_mock_synergy(\n '_get_stream_state', self.target)\n target.cf.get_streams.return_value = [{'Direc': 'My Value'}]\n self.assertEqual('My Value', target._get_stream_state(1))\n\n target.cf.get_streams.return_value = []\n self.assertEqual(None, target._get_stream_state(2))\n\n def test_get_device_time(\n self, test_time='Tue May 10 20:12:20 UTC 2016', match=None,\n err=None):\n incoming_format = '%a %b %d %H:%M:%S %Z %Y'\n match = match or {\n 'year': '2016', 'month': '5', 'day': '10', 'hour': '20',\n 'min': '12', 'sec': '20', 'weekday': '1', 'day_of_year': '131',\n 'isdst': '0'}\n\n target = self.build_generic_mock_synergy(\n 'get_device_time', self.target)\n\n target.cli.exec_command.return_value = ['', test_time, '']\n if not err:\n ret = target.get_device_time()\n else:\n self.assertRaises(err, target.get_device_time)\n return\n\n t_struct = time.strptime(test_time.strip(), incoming_format)\n match.update({\n 'raw': test_time.strip(), 'struct_time': t_struct,\n 'time_zone': 'UTC', 'time': time.strftime('%H:%M:%S', t_struct)})\n self.assertDictEqual(match, ret)\n\n def test_get_device_time_2(self):\n match = {\n 'day': '24', 'day_of_year': '55', 'hour': '21', 'isdst': '0',\n 'min': '41', 'month': '2', 'sec': '56', 'weekday': '4',\n 'year': '2017'}\n return self.test_get_device_time(\n ' Fri Feb 24 21:41:56 UTC 2017 ', match)\n\n def test_get_device_time_raises(self):\n return self.test_get_device_time(\n \"long line with a Bad Date that can't be converted\", err=TngError)\n\n def test_check_is_available(self):\n target = self.build_generic_mock_synergy(\n 'check_is_available', self.target)\n self.assertFalse(target.check_is_available())\n target.check_ip_ping.return_value = True\n self.assertTrue(target.check_is_available())\n\n def test_get_huron_trackingid(self):\n self.target = self.build_generic_mock_synergy(\n 'get_huron_trackingid', self.target)\n\n self.target.cli.exec_command.return_value = [\n 'show huron TrackingID',\n '',\n '8851_e06d594d-0010-5000-a000-b07d47d313c8',\n '']\n self.assertEqual(\n self.target.get_huron_trackingid(),\n '8851_e06d594d-0010-5000-a000-b07d47d313c8')\n\n def test_get_huron_username(self):\n self.target = self.build_generic_mock_synergy(\n 'get_huron_username', self.target)\n\n self.target.cli.exec_command.return_value = [\n 'show huron username',\n '',\n 'tipbu_auto_63768@int1.huron-alpha.com',\n '']\n self.assertEqual(\n self.target.get_huron_username(),\n 'tipbu_auto_63768@int1.huron-alpha.com')\n\n def test_wait_until_close(self):\n target = self.build_generic_mock_synergy(\n 'wait_until_close', self.target)\n self.assertTrue(target.wait_until_close())\n target.handle_shutdown.assert_called_once_with()\n target.wait_until_down.assert_not_called()\n\n target.handle_shutdown.reset_mock()\n target.wait_until_down.reset_mock()\n\n self.assertTrue(target.wait_until_close(check=True))\n target.handle_shutdown.assert_called_once_with()\n target.wait_until_down.assert_called_once_with(30)\n\n target.handle_shutdown.reset_mock()\n target.wait_until_down.reset_mock()\n\n target.handle_shutdown.side_effect = [TngError]\n self.assertFalse(target.wait_until_close(check=True))\n\n def test_wait_until_restore(self):\n target = self.build_generic_mock_synergy(\n 'wait_until_restore', self.target)\n\n target.check_ip_ping.return_value = True\n target.handle_startup.return_value = True\n self.assertTrue(target.wait_until_restore(timeout=1, wait=0.1))\n\n def test_phone_set_idle(self):\n target = self.build_generic_mock_synergy('phone_set_idle', self.target)\n target.ui.close_all_apps.return_value = False\n self.assertFalse(target.phone_set_idle())\n target.cf.end_all_calls.assert_not_called()\n\n target.ui.close_all_apps.return_value = True\n target.cf.end_all_calls.return_value = False\n self.assertFalse(target.phone_set_idle())\n\n target.cf.end_all_calls.return_value = True\n self.assertTrue(target.phone_set_idle())\n\n @patch(\n 'tng_sl.device.endpoint.synergylite.synergylite_extended.get_system')\n def test_power(self, get_system):\n target = self.build_generic_mock_synergy('power', self.target)\n sw_obj = Mock()\n get_system.return_value = sw_obj\n\n self.assertFalse(target.power(reset_type='junk'))\n sw_obj.run_config_command.assert_not_called()\n self.assertEqual(sw_obj, target.sw_obj)\n\n target.get_port.return_value = \"MyPort\"\n\n self.assertTrue(target.power(reset_type='on', restore=False))\n sw_obj.run_config_command.assert_called_once_with(\n 'interface MyPort\\nno shut')\n target.wait_until_restore.assert_not_called()\n\n sw_obj.run_config_command.reset_mock()\n\n self.assertTrue(target.power(\n reset_type='off', timeout=0.1, wait=0.01, restore=True))\n\n sw_obj.run_config_command.assert_called_once_with(\n 'interface MyPort\\nshutdown')\n target.wait_until_restore.assert_called_once_with(\n timeout=0.1, wait=0.01)\n\n sw_obj.run_config_command.reset_mock()\n target.wait_until_restore.reset_mock()\n self.assertTrue(target.power(\n reset_type='toggle', timeout=0.2, wait=0.01, restore=True))\n\n sw_obj.run_config_command.has_calls([\n call('interface MyPort\\nshutdown'),\n call('interface MyPort\\nno shut')])\n target.wait_until_restore.assert_called_once_with(\n timeout=0.2, wait=0.01)\n\n def test_reset(self):\n target = self.build_generic_mock_synergy('reset', self.target)\n target.feedback = Mock()\n self.assertFalse(target.reset('junk'))\n target.cli.exec_command.assert_not_called()\n\n self.assertTrue(target.reset(timeout=1, wait=1))\n target.cli.exec_command.assert_called_once_with('reset hard\\n')\n target.wait_until_close.assert_called_once_with(check=True)\n target.wait_until_restore.assert_called_once_with(timeout=1, wait=1)\n\n target.cli.exec_command.reset_mock()\n target.wait_until_close.reset_mock()\n target.wait_until_restore.reset_mock()\n\n self.assertTrue(target.reset(reset_type='soft'))\n target.cli.exec_command.assert_called_once_with('reset soft\\n')\n target.wait_until_close.assert_called_once_with(check=False)\n target.wait_until_restore.assert_called_once_with(timeout=80, wait=20)\n\n target.cli.exec_command.reset_mock()\n target.wait_until_close.reset_mock()\n target.wait_until_restore.reset_mock()\n\n target.is_3pcc = False\n self.assertTrue(target.reset(reset_type='factory'))\n target.cli.exec_command.assert_called_once_with('reset factory\\n')\n target.cli.exec_command.reset_mock()\n target.wait_until_up.assert_not_called()\n target._set_3pcc_ssh.assert_not_called()\n target.cli.press_skip_key.assert_not_called()\n\n target.is_3pcc = 'True'\n target._set_3pcc_ssh.return_value = True\n target.cli.press_skip_key.return_value = True\n self.assertTrue(target.reset(reset_type='factory'))\n target.cli.exec_command.assert_called_once_with('reset factory\\n')\n target.wait_until_up.assert_called_once_with(80)\n target._set_3pcc_ssh.assert_called_once_with()\n target.cli.press_skip_key.assert_called_once_with()\n\n target._set_3pcc_ssh.return_value = False\n self.assertFalse(target.reset(reset_type='factory'))\n\n target._set_3pcc_ssh.return_value = True\n target.cli.press_skip_key.return_value = False\n self.assertFalse(target.reset(reset_type='factory'))\n\n target.cli.exec_command.reset_mock()\n\n self.assertTrue(target.reset(reset_type='service'))\n target.cli.exec_command.assert_called_once_with('reset servicemode\\n')\n target.cli.exec_command.reset_mock()\n\n self.assertTrue(target.reset(reset_type='servicemode'))\n target.cli.exec_command.assert_called_once_with('reset servicemode\\n')\n target.cli.exec_command.reset_mock()\n\n def test_reset_reboot(self):\n self.target = self.build_generic_mock_synergy('reset', self.target)\n self.target.feedback = Mock()\n self.target.debugsh = Mock()\n self.target.debugsh.exec_command.return_value = \"Hello\\nWorld\"\n\n # Invoke API under test\n self.assertTrue(\n self.target.reset(reset_type='reboot', timeout=120, wait=120))\n\n # Additional validations\n self.target.debugsh.exec_command.assert_called_once_with(\n 'reboot\\n', timeout=5, raw=True)\n self.target.wait_until_close.assert_called_once_with(check=True)\n self.target.wait_until_restore.assert_called_once_with(\n timeout=120, wait=120)\n\n self.target.debugsh.exec_command.reset_mock()\n self.target.wait_until_close.reset_mock()\n self.target.wait_until_restore.reset_mock()\n\n self.target.debugsh.exec_command.side_effect = [TimeoutError]\n # Invoke API under test\n self.assertTrue(\n self.target.reset(reset_type='reboot', timeout=120, wait=120))\n\n # Additional validations\n self.target.debugsh.exec_command.assert_called_once_with(\n 'reboot\\n', timeout=5, raw=True)\n self.target.wait_until_close.assert_called_once_with(check=True)\n self.target.wait_until_restore.assert_called_once_with(\n timeout=120, wait=120)\n\n def test__restart_action(self):\n target = self.build_generic_mock_synergy(\n '_restart_action', self.target)\n target.ssh = None\n self.assertEqual(target._restart_action(), None)\n target.log.info.assert_called_once_with('Did not reboot')\n\n target.ssh = Mock()\n target.reset.return_value = True\n self.assertTrue(target._restart_action())\n target.reset.assert_called_once_with(\n reset_type='reboot', timeout=120, wait=120)\n\n target.reset.reset_mock()\n target.reset.return_value = False\n self.assertRaises(TngError, target._restart_action)\n target.reset.assert_called_once_with(\n reset_type='reboot', timeout=120, wait=120)\n\n def test__boot_action(self):\n target = self.build_generic_mock_synergy('_boot_action', self.target)\n target._boot_action()\n target._restart_action.assert_called_once_with()\n\n def test_resume_all(self):\n target = self.build_generic_mock_synergy('resume_all', self.target)\n self.assertRaises(NotImplementedError, target.resume_all)\n\n def test_on_board_huron(self):\n self.target = self.build_generic_mock_synergy(\n 'on_board_huron', self.target)\n self.target.cli.exec_command.return_value = [\n 'login huron 8893501470247980 0'\n ]\n self.target.handle_shutdown.return_value = True\n self.target.handle_startup.return_value = True\n self.assertEqual(\n self.target.on_board_huron('8893501470247980', 0.001, 1),\n ['login huron 8893501470247980 0'])\n self.target.cli.exec_command.assert_called_once_with(\n 'login huron 8893501470247980')\n\n def test_on_board_ce(self):\n self.target = self.build_generic_mock_synergy(\n 'on_board_ce', self.target)\n self.target.cli.exec_command.return_value = (\n 'login ce cisco.com me pwd\\nSuccess'.splitlines())\n self.assertEqual(\n self.target.on_board_ce('cisco.com', 'me', 'pwd', 0.001, 1),\n ['login ce cisco.com me pwd', 'Success'])\n self.target.cli.exec_command.assert_called_once_with(\n 'login ce cisco.com me pwd')\n\n def test_reset_servicemode(self):\n target = self.build_generic_mock_synergy(\n 'reset_servicemode', self.target)\n target.reset_servicemode()\n target.reset.assert_called_once_with(\n 'service', timeout=80, wait=20, restore_conn=False)\n\n @patch(\n 'tng_sl.device.endpoint.synergylite.synergylite_extended.get_defines')\n def test_property_ICE_OPTIMAL_PATH(self, get_defines):\n target = unmockedSLX()\n # test values read from command line\n get_defines.return_value = {'ICE_OPTIMAL_PATH': 'SRFLX'}\n self.assertEqual(target.ICE_OPTIMAL_PATH, 'SRFLX')\n self.assertEqual(target._ice_optimal_path, 'SRFLX')\n self.assertEqual(\n target.ICE_OPTIMAL_PATHS.index(target.ICE_OPTIMAL_PATH), 2)\n\n get_defines.assert_called_once_with()\n get_defines.reset_mock()\n self.assertTrue(getattr(target, '_ice_optimal_path', False))\n del target.ICE_OPTIMAL_PATH\n self.assertFalse(getattr(target, '_ice_optimal_path', False))\n\n get_defines.return_value = {'Other_stuff': 'Nope'}\n self.assertEqual(target.ICE_OPTIMAL_PATH, 'NO_ICE')\n self.assertEqual(target._ice_optimal_path, 'NO_ICE')\n self.assertEqual(\n target.ICE_OPTIMAL_PATHS.index(target.ICE_OPTIMAL_PATH), 0)\n\n get_defines.assert_called_once_with()\n get_defines.reset_mock()\n\n # user sets value to a string-int\n target.ICE_OPTIMAL_PATH = '3'\n self.assertEqual(target.ICE_OPTIMAL_PATH, 'RELAY')\n self.assertEqual(target._ice_optimal_path, 'RELAY')\n self.assertEqual(\n target.ICE_OPTIMAL_PATHS.index(target.ICE_OPTIMAL_PATH), 3)\n get_defines.assert_not_called()\n\n # user sets value to a lower case\n target.ICE_OPTIMAL_PATH = 'prflx'\n self.assertEqual(target.ICE_OPTIMAL_PATH, 'PRFLX')\n self.assertEqual(target._ice_optimal_path, 'PRFLX')\n self.assertEqual(\n target.ICE_OPTIMAL_PATHS.index(target.ICE_OPTIMAL_PATH), 4)\n get_defines.assert_not_called()\n\n @patch(\n 'tng_sl.device.endpoint.synergylite.synergylite_extended.get_defines')\n def test_property_ICE_TRAVERSAL_MODE(self, get_defines):\n target = unmockedSLX()\n # test values read from command line\n get_defines.return_value = {'ICE_TRAVERSAL_MODE': 'ICE with TURN'}\n self.assertEqual(target.ICE_TRAVERSAL_MODE, 'ICE with TURN')\n self.assertEqual(target._ice_traversal_mode, 'ICE with TURN')\n self.assertEqual(\n target.ICE_TRAVERSAL_MODES.index(target.ICE_TRAVERSAL_MODE), 3)\n\n get_defines.assert_called_once_with()\n get_defines.reset_mock()\n self.assertTrue(getattr(target, '_ice_traversal_mode', False))\n\n # make sure deleting the property will delete the private variable\n del target.ICE_TRAVERSAL_MODE\n self.assertFalse(getattr(target, '_ice_traversal_mode', False))\n\n get_defines.return_value = {'Other_stuff': 'Nope'}\n self.assertEqual(target.ICE_TRAVERSAL_MODE, 'Media Latch')\n self.assertEqual(target._ice_traversal_mode, 'Media Latch')\n self.assertEqual(\n target.ICE_TRAVERSAL_MODES.index(target.ICE_TRAVERSAL_MODE), 1)\n\n get_defines.assert_called_once_with()\n get_defines.reset_mock()\n\n # user sets value to a string-int\n target.ICE_TRAVERSAL_MODE = '2'\n self.assertEqual(target.ICE_TRAVERSAL_MODE, 'TURN Relay Only')\n self.assertEqual(target._ice_traversal_mode, 'TURN Relay Only')\n self.assertEqual(\n target.ICE_TRAVERSAL_MODES.index(target.ICE_TRAVERSAL_MODE), 2)\n get_defines.assert_not_called()\n\n # user sets value to a lower case\n target.ICE_TRAVERSAL_MODE = 'media latch'\n self.assertEqual(target.ICE_TRAVERSAL_MODE, 'Media Latch')\n self.assertEqual(target._ice_traversal_mode, 'Media Latch')\n self.assertEqual(\n target.ICE_TRAVERSAL_MODES.index(target.ICE_TRAVERSAL_MODE), 1)\n get_defines.assert_not_called()\n\n def test_check_traversal_mode(self):\n target = self.build_generic_mock_synergy(\n 'check_traversal_mode', self.target)\n ret = {'Media Traversal Mode': 'TestValue'}\n target.cf.get_ice.return_value = ret\n\n resp = target.check_traversal_mode(\n expected_mode='ValueTest',\n show_ice={'Media Traversal Mode': 'ValueTest'})\n self.assertEqual(resp, {'Media Traversal Mode': 'ValueTest'})\n\n resp = target.check_traversal_mode('TestValue')\n self.assertEqual(resp, ret)\n target.cf.get_ice.assert_called_once_with()\n\n target.ICE_TRAVERSAL_MODE = 'TestValue'\n resp = target.check_traversal_mode()\n self.assertEqual(resp, ret)\n\n self.assertRaises(\n UnexpectedIceError, target.check_traversal_mode, 'unmatched')\n\n def test_check_optimal_path(self):\n target = self.build_generic_mock_synergy(\n 'check_optimal_path', self.target)\n\n ret = {\n 'Media Traversal Mode': 'TestValue',\n 'Audio Optimal media path': '4 (PRFLX)',\n 'Video Optimal media path': '1 (HOST)'}\n\n target.cf.get_ice.return_value = ret\n resp = target.check_optimal_path(expected_path='PRFLX')\n self.assertEqual(resp, ret)\n target.cf.get_ice.assert_called_once_with()\n target.wait_for_ice_completed.assert_called_once_with()\n\n target.ICE_OPTIMAL_PATH = 'TestValue'\n resp = target.check_optimal_path(expected_path='PRFLX')\n self.assertEqual(resp, ret)\n\n self.assertRaises(\n UnexpectedIceError, target.check_optimal_path,\n expected_path='unmatched')\n\n # test video only\n resp = target.check_optimal_path(\n expected_path='HOST', media={'video': 'junk'})\n\n ret = {\n 'Media Traversal Mode': 'TestValue',\n 'Audio Optimal media path': '3 (RELAY)',\n 'Video Optimal media path': '3 (RELAY)'}\n target.cf.get_ice.return_value = ret\n\n # test both video and audio on same path\n resp = target.check_optimal_path(\n expected_path='RELAY', media={'video': 'junk', 'audio': 'more'})\n\n # test both video and audio on different paths\n ret['Audio Optimal media path'] = '1 (HOST)'\n\n self.assertRaises(\n UnexpectedIceError, target.check_optimal_path,\n expected_path='RELAY', media={'video': 'junk', 'audio': 'more'})\n\n def test_check_ice_optimal_path(self):\n target = self.build_generic_mock_synergy(\n 'check_optimal_path', self.target)\n target.cf.get_ice.return_value = {\n \"Audio Optimal media path\": '3 (RELAY)'}\n target.check_optimal_path('RELAY')\n self.assertRaises(\n UnexpectedIceError, target.check_optimal_path, 'SRFLX')\n\n def test_check_ICE_mode_and_path(self):\n target = self.build_generic_mock_synergy(\n 'check_ICE_mode_and_path', self.target)\n target.check_traversal_mode.return_value = 'RET1'\n target.check_optimal_path.return_value = 'RET2'\n\n ret = target.check_ICE_mode_and_path(\n media={'a': 'b'}, expected_mode='MyMode', expected_path='MyPath')\n self.assertEqual(ret, \"RET2\")\n target.check_traversal_mode.assert_called_once_with(\n expected_mode='MyMode')\n target.check_optimal_path.assert_called_once_with(\n media={'a': 'b'}, expected_path='MyPath')\n\n def test_get_phone_traversal_mode(self):\n target = self.build_generic_mock_synergy(\n 'get_phone_traversal_mode', self.target)\n target.cli.get_config_info.return_value = '2'\n self.assertEqual('TURN Relay Only', target.get_phone_traversal_mode())\n\n target.cli.get_config_info.return_value = 'unmatched'\n self.assertEqual('Media Latch', target.get_phone_traversal_mode())\n\n @patch('tng_sl.device.endpoint.synergylite.synergylite_extended.until')\n def test_wait_for_media_traversal_mode(self, u):\n target = self.build_generic_mock_synergy(\n 'wait_for_media_traversal_mode', self.target)\n u.return_value = True\n target.get_phone_traversal_mode.return_value = 'TURN Relay Only'\n target.wait_for_media_traversal_mode(2)\n u.assert_called_once_with(\n target.get_phone_traversal_mode, desired_result='TURN Relay Only')\n u.reset_mock()\n u.side_effect = [TimeoutError('TestFailure')]\n self.assertRaises(\n TimeoutError, target.wait_for_media_traversal_mode)\n\n def test_get_ice_state(self):\n target = self.build_generic_mock_synergy('get_ice_state', self.target)\n target.cf.get_ice.return_value = {\"Ice state\": 'unmatched'}\n self.assertEqual('IDLE', target.get_ice_state())\n\n target.cf.get_ice.reset_mock()\n target.cf.get_ice.return_value = {\"Ice state\": '5 (ACTIVE)'}\n self.assertEqual('ACTIVE', target.get_ice_state())\n target.cf.get_ice.assert_called_once_with()\n target.cf.get_ice.reset_mock()\n self.assertEqual('NO_ICE', target.get_ice_state(\n {\"Ice state\": '7 (NO_ICE)'}))\n target.cf.get_ice.assert_not_called()\n # Ice not configured has no 'Ice state\"\n test_no_ice = {\n u'Audio Local media ip': u'10.4.10.23',\n u'Audio Local media port': u'23040',\n u'Audio Optimal media path': u'LATCHING',\n u'Audio Peer media ip': u'10.1.50.190',\n u'Audio Peer media port': u'47606',\n u'Media Traversal Mode': u'Media Latch',\n u'Video Local media ip': u'10.4.10.23',\n u'Video Local media port': u'26318',\n u'Video Optimal media path': u'LATCHING',\n u'Video Peer media ip': u'10.1.50.190',\n u'Video Peer media port': u'36150'}\n target.cf.get_ice.reset_mock()\n self.assertEqual('NO_ICE', target.get_ice_state(show_ice=test_no_ice))\n target.cf.get_ice.assert_not_called()\n\n def test_is_ice_state_started(self):\n target = self.build_generic_mock_synergy(\n 'is_ice_state_started', self.target)\n target.get_ice_state.return_value = 'CANDIDATES_DONE'\n self.assertTrue(target.is_ice_state_started())\n target.get_ice_state.assert_called_once_with(show_ice=None)\n\n target.get_ice_state.reset_mock()\n target.get_ice_state.return_value = 'ACTIVE'\n self.assertFalse(target.is_ice_state_started())\n target.get_ice_state.assert_called_once_with(show_ice=None)\n\n target.get_ice_state.reset_mock()\n target.get_ice_state.return_value = 'START_CONNECTIVITY_CHECK'\n self.assertTrue(target.is_ice_state_started(\n {'Ice state': '3 (START_CONNECTIVITY_CHECK)'}))\n target.get_ice_state.assert_called_once_with(\n show_ice={'Ice state': '3 (START_CONNECTIVITY_CHECK)'})\n\n @patch('tng_sl.device.endpoint.synergylite.synergylite_extended.until')\n def test_wait_for_ice_state_started(self, u):\n target = self.build_generic_mock_synergy(\n 'wait_for_ice_state_started', self.target)\n u.return_value = True\n target.wait_for_ice_state_started()\n u.assert_called_once_with(\n target.is_ice_state_started, interval=0.2, timeout=10.0,\n raise_msg=(\n \"{}: Not in expected ICE state 2 (CANDIDATES_DONE) or 3 \"\n \"(START_CONNECTIVITY_CHECK)\").format(target))\n\n def test_wait_for_ice_completed(self):\n target = self.build_generic_mock_synergy(\n 'wait_for_ice_completed', self.target)\n # negative cases first\n for s in range(6):\n ret_state_name = value_from_name_or_index(\n s, target.ICE_STATE_NAMES, 0)\n target.get_ice_state.return_value = ret_state_name\n # need to test both the Exception and the message in the exception\n try:\n target.wait_for_ice_completed()\n raise BaseException(\"expected TimeoutError\")\n except Exception as e:\n self.assertTrue(isinstance(\n e, TimeoutError), \"{} is not a TimeoutError\".format(e))\n self.assertIn(\n \"{} Not in completed ICE states\".format(ret_state_name),\n e.message)\n\n # success cases here\n for s in range(6, 9):\n translated_state = value_from_name_or_index(\n s, target.ICE_STATE_NAMES, 0)\n target.get_ice_state.reset_mock()\n target.get_ice_state.side_effect = [\n value_from_name_or_index(bad_val, target.ICE_STATE_NAMES, 0)\n for bad_val in range(5)] + [translated_state]\n ret = target.wait_for_ice_completed()\n self.assertEqual(ret, translated_state)\n\n @patch('tng_sl.device.endpoint.synergylite.synergylite_extended.until')\n def test_wait_for_ice_state_active(self, u):\n target = self.build_generic_mock_synergy(\n 'wait_for_ice_state_active', self.target)\n u.return_value = True\n target.wait_for_ice_state_active()\n u.assert_called_once_with(\n target.get_ice_state, desired_result='ACTIVE', timeout=20.0,\n raise_msg=\"{}: Not in expected ICE state ACTIVE\".format(target))\n\n def test_check_ice_remote_concluded_type(self):\n target = self.build_generic_mock_synergy(\n 'check_ice_remote_concluded_type', self.target)\n ret = {\n \"Audio Remote concluded candidate type\": '3 (RELAY)',\n 'Audio Peer media ip': '1.2.3.4'}\n target.cf.get_ice.return_value = ret.copy()\n self.assertEqual(ret, target.check_ice_remote_concluded_type(3))\n ret2 = {\n \"Audio Remote concluded candidate type\": '4 (PRFLX)',\n 'Audio Peer media ip': '4.3.2.1'}\n self.assertEqual(ret2, target.check_ice_remote_concluded_type(\n 'prflx', ret2.copy()))\n\n self.assertRaises(\n UnexpectedIceError, target.check_ice_remote_concluded_type,\n 'prflx')\n\n def test_is_registered(self):\n self.target = self.build_generic_mock_synergy(\n 'is_registered', self.target)\n self.target.cli.show_register_gsi.return_value = {\n 'Proxy Registration':\n 'ENABLED',\n 'line': '201',\n 'local uuid': '9313784100105000a000346f9017d6a6',\n 'remote uuid': '',\n 'state': 'REGISTERED'\n }\n self.target.cli.get_phone_info.return_value = {\n 'Active Load': 'sip88xx.11-5-1MN-7dev.loads',\n 'Active Serve Type': 'CMS',\n 'Active Server': 'cms-vip-a-01-internal.int-tx3.huron-int.com',\n 'IPv4 Address': '10.79.63.28',\n 'Last Upgrade': '11-04-15 10:00',\n 'MAC Address': '6C998984B90C',\n 'Model Number': 'CP-8861',\n 'Phone Information': '',\n 'Squared UC': 'huron-int.com',\n 'Stand-by Server': '',\n 'Stand-by Server Type': ''\n }\n self.assertTrue(self.target.is_registered())\n\n def test_wait_until_registered(self):\n target = self.build_generic_mock_synergy(\n 'wait_until_registered', self.target)\n target.is_registered.return_value = False\n self.assertRaises(\n TimeoutError, target.wait_until_registered, retry=1, period=0.1)\n target.is_registered.return_value = True\n self.assertTrue(target.wait_until_registered())\n\n def test_wait_until_login_page(self):\n target = self.build_generic_mock_synergy(\n 'wait_until_login_page', self.target)\n target.ui.check_ui.return_value = False\n self.assertRaises(\n TimeoutError, target.wait_until_login_page, retry=1, period=0.1)\n target.ui.check_ui.assert_called_with(\n expected='Welcome', name='in-focus', ui_type='ui_window_detail')\n\n target.ui.check_ui.reset_mock()\n target.ui.check_ui.return_value = True\n self.assertTrue(target.wait_until_login_page())\n target.ui.check_ui.assert_called_once_with(\n expected='Welcome', name='in-focus', ui_type='ui_window_detail')\n\n def test_is_cucm_mode(self):\n self.target = self.build_generic_mock_synergy(\n 'is_cucm_mode', self.target)\n self.target.get_service_mode.return_value = (\n 'Enterprise (deploy-mode:1)')\n self.assertTrue(self.target.is_cucm_mode())\n\n def test_is_ce_mode(self):\n self.target = self.build_generic_mock_synergy('is_ce_mode')\n self.target.get_service_mode.return_value = (\n 'Enterprise (deploy-mode:1)')\n self.assertFalse(self.target.is_ce_mode())\n\n def test_is_huron_mode(self):\n self.target = self.build_generic_mock_synergy('is_huron_mode')\n self.target.get_service_mode.return_value = (\n 'Enterprise (deploy-mode:1)')\n self.assertFalse(self.target.is_huron_mode())\n\n def test_get_huron_domain(self):\n self.target = self.build_generic_mock_synergy(\n 'get_huron_domain', self.target)\n self.target.cli.show_huron_info.return_value = 'hptx1.huron-dev.com'\n self.assertEqual(self.target.get_huron_domain(), 'hptx1.huron-dev.com')\n\n def test_get_nslookup_info(self):\n self.target = self.build_generic_mock_synergy(\n 'get_nslookup_info', self.target)\n self.target.cli.show_nslookup.return_value = {\n 'Address 1': '72.163.4.161 www1.cisco.com',\n 'Address 2': '2001:420:1101:1::a www1.cisco.com',\n 'Name': 'cisco.com',\n 'Server': '127.0.0.1',\n 'nslookup cisco.com': ''}\n expected_result = {\n 'Address 1': '72.163.4.161 www1.cisco.com',\n 'Address 2': '2001:420:1101:1::a www1.cisco.com',\n 'Name': 'cisco.com',\n 'Server': '127.0.0.1',\n 'nslookup cisco.com': ''}\n self.assertEqual(\n self.target.get_nslookup_info(\"cisco.com\"), expected_result)\n\n def test_get_wlan_auth_prompt_mode(self):\n self.target = self.build_generic_mock_synergy(\n 'get_wlan_auth_prompt_mode', self.target)\n self.target.cli.show_jvm_config_properties.return_value = {\n 'device.settings.config.vendorconfig.wlanauthpromptmode': {\n 'value': 'true',\n 'provisional.value': 'null',\n 'to.string': 'true'\n }\n }\n expected_result = {\n 'value': 'true',\n 'provisional.value': 'null',\n 'to.string': 'true'\n }\n self.assertEquals(\n self.target.get_wlan_auth_prompt_mode(), expected_result)\n\n def test_get_hedge_server_list(self):\n self.target = self.build_generic_mock_synergy(\n 'get_hedge_server_list', self.target)\n self.target.cli.show_hedge_info.return_value = {\n 'ActiveHedgeServer': '10.1.1.1',\n 'Hedge Infomation': '',\n 'HedgeServer1': '10.1.1.1',\n 'HedgeServer2': '10.1.1.2',\n 'HedgeServer3': '',\n 'HedgeServer4': ''\n }\n self.assertEqual(\n self.target.get_hedge_server_list(), ['10.1.1.1', '10.1.1.2'])\n\n def test_get_active_hedge_server(self):\n self.target = self.build_generic_mock_synergy(\n 'get_active_hedge_server', self.target)\n self.target.cli.show_hedge_info.return_value = {\n 'ActiveHedgeServer': '10.1.1.1',\n 'Hedge Infomation': '',\n 'HedgeServer1': '10.1.1.1',\n 'HedgeServer2': '10.1.1.2',\n 'HedgeServer3': '',\n 'HedgeServer4': ''\n }\n self.assertEqual(\n self.target.get_active_hedge_server(), '10.1.1.1')\n\n def test_get_park_number_toast(self):\n self.target = self.build_generic_mock_synergy(\n 'get_park_number', self.target)\n self.target.cli.show_dpark_info.return_value = {\n 'Title': 'null',\n 'Type': '0',\n 'Icon': 'N/A',\n 'Text': 'Call park at 7770'\n }\n self.assertEqual(self.target.get_park_number(), '7770')\n\n def test_get_park_number_toast_missing_text_value(self):\n self.target = self.build_generic_mock_synergy(\n 'get_park_number', self.target)\n self.target.cli.show_dpark_info.return_value = {\n 'Title': 'null',\n 'Type': '0',\n 'Icon': 'N/A',\n 'Text': ''\n }\n self.assertEqual(self.target.get_park_number(), None)\n\n def test_get_park_number_toast_variation_text_value(self):\n self.target = self.build_generic_mock_synergy(\n 'get_park_number', self.target)\n self.target.cli.show_dpark_info.return_value = {\n 'Title': 'null',\n 'Type': '0',\n 'Icon': 'N/A',\n 'Text': 'Call Parked at 7770'\n }\n self.assertEqual(self.target.get_park_number(), '7770')\n\n def test_get_park_number_toast_unknown_text_value(self):\n self.target = self.build_generic_mock_synergy(\n 'get_park_number', self.target)\n self.target.cli.show_dpark_info.return_value = {\n 'Title': 'null',\n 'Type': '0',\n 'Icon': 'N/A',\n 'Text': 'Dr Pais watc is now connected.'\n\n }\n self.assertEqual(self.target.get_park_number(), None)\n\n def test_get_park_number_toast_sp_variation(self):\n self.target = self.build_generic_mock_synergy(\n 'get_park_number', self.target)\n self.target.cli.show_dpark_info.return_value = {\n 'Text': 'Park at 7770',\n 'subwindow-0\\t0/0/330/105': '',\n 'Icon': 'N/A',\n 'Type': '5',\n 'Title': ''\n }\n self.assertEqual(self.target.get_park_number(), '7770')\n\n def test_get_park_number_no_toast(self):\n self.target = self.build_generic_mock_synergy(\n 'get_park_number', self.target)\n self.target.cli.show_dpark_info.return_value = {\n 'No Active Toast Displayed': ''\n }\n self.assertEqual(self.target.get_park_number(), None)\n\n def test_get_service_mode(self):\n self.target = self.build_generic_mock_synergy(\n 'get_service_mode', self.target)\n self.target.cli.show_jvm_ce.return_value = self.jvmVal\n self.assertEqual(\n self.target.get_service_mode(), 'Enterprise (deploy-mode:1)')\n\n def test_get_wlan_info(self):\n self.target = self.build_generic_mock_synergy(\n 'get_wlan_info', self.target)\n self.target.cli.show_wlan_profile.return_value = {\n 'OK': '',\n 'band': '5GHz',\n 'passphrase': '*',\n 'security_type': 'PSK',\n 'ssid': 'x930'\n }\n self.assertEqual(self.target.get_wlan_info('profile'), {\n 'OK': '',\n 'band': '5GHz',\n 'passphrase': '*',\n 'security_type': 'PSK',\n 'ssid': 'x930'})\n\n def test_press_hard_key(self):\n self.target = self.build_generic_mock_synergy(\n 'press_hard_key', self.target)\n self.target.cli.press_speaker.return_value = True\n self.assertTrue(self.target.press_hard_key('speaker'))\n self.target.cli.press_number_key.return_value = True\n self.assertTrue(self.target.press_hard_key(\n key_name=0, key_type='number'))\n\n def test_get_ssid_info_by_name(self):\n self.target = self.build_generic_mock_synergy(\n 'get_ssid_info_by_name', self.target)\n self.target.cli.show_wifi_ssid_list.side_effect = [None, [\n {\n 'BSSID': 'c8:f9:f9:d5:37:c2',\n 'Current': 'Yes',\n 'Frequency': '2462',\n 'SSID': 'automation_PSK',\n 'Security Mode': 'PSK',\n 'Signal': 't-57'\n },\n {\n 'BSSID': 'f0:b4:29:60:d1:30',\n 'Current': 'No',\n 'Frequency': '2457',\n 'SSID': 'Wenlan-XiaoMi-2.4G',\n 'Security Mode': 'PSK',\n 'Signal': '-27'\n }\n ]]\n expect_result = {\n 'BSSID': 'c8:f9:f9:d5:37:c2',\n 'Current': 'Yes',\n 'Frequency': '2462',\n 'SSID': 'automation_PSK',\n 'Security Mode': 'PSK',\n 'Signal': 't-57'\n }\n self.assertEqual(\n self.target.get_ssid_info_by_name('automation_PSK'), expect_result)\n self.target._wait_patch.assert_called_once_with(\n 1, 'wait 1s to get ssid info from cli again')\n\n def test_is_wlan_connected_by_ssid(self):\n self.target = self.build_generic_mock_synergy(\n 'is_wlan_connected_by_ssid', self.target)\n self.target.cli.show_wlan_supplicant_status.return_value = {\n 'bssid': 'cc:46:d6:a9:29:f6',\n 'ssid': 'test-peap-gtc',\n 'id': '0',\n 'mode': 'station',\n 'pairwise_cipher': 'CCMP',\n 'group_cipher': 'CCMP',\n 'key_mgmt': 'WPA2/IEEE 802.1X/EAP',\n 'wpa_state': 'COMPLETED',\n 'ip_address': '100.100.48.145',\n 'address': '28:34:a2:82:8f:66',\n 'Supplicant PAE state': 'AUTHENTICATED',\n 'suppPortStatus': 'Authorized',\n 'EAP state': 'SUCCESS',\n 'selectedMethod': '25 (EAP-PEAP)',\n 'EAP TLS cipher': 'ECDHE-RSA-AES256-SHA',\n 'EAP-PEAPv1 Phase2 method': 'GTC'\n }\n self.assertTrue(\n self.target.is_wlan_connected_by_ssid('test-peap-gtc'))\n\n def test_get_prt_status(self):\n self.target = self.build_generic_mock_synergy(\n 'get_prt_status', self.target)\n self.target.cli.show_prt_status.return_value = [{\n 'PRT': '6',\n 'State': 'uploaded',\n }]\n expect_result = {\n 'PRT': '6',\n 'STATE': 'uploaded',\n }\n self.assertEqual(\n self.target.get_prt_status(), expect_result)\n\n self.target.cli.show_prt_status.return_value = None\n self.assertRaises(TypeError, self.target.get_prt_status, 'junk')\n\n self.target.cli.show_prt_status.return_value = [\n {u'Command not found.': u\"subcommands available under 'prt':\"},\n {u'Command not found.': u'prt create'}]\n self.assertRaises(TngError, self.target.get_prt_status)\n\n def test_gen_prt_log(self):\n self.target = self.build_generic_mock_synergy(\n 'gen_prt_log', self.target)\n self.target._check_prt_generated.return_value = True\n self.target.cli.exec_command.return_value = [\n u'prt create FULL',\n u\"Full PRT Generation started.\"\n u\"Check phone's web page Console Logs for Full PRT logs.\",\n u'PRT Filename: prt-20161109-143351-00EBD5DA5AAC.tar.gz',\n u'Phone Webpage Link:'\n u'http://10.89.122.116/FS/prt-20161109-143351-00EBD5DA5AAC.tar.gz',\n u'',\n u'']\n self.assertEqual(\n self.target.gen_prt_log(),\n 'prt-20161109-143351-00EBD5DA5AAC.tar.gz')\n self.target.cli.exec_command.assert_called_once_with(\n 'prt create FULL noupload', timeout=120.0)\n\n self.target.cli.exec_command.return_value = []\n self.assertRaises(TngError, self.target.gen_prt_log)\n\n def test_check_webpage_for_prtname(self):\n self.target = self.build_generic_mock_synergy(\n '_check_webpage_for_prtname', self.target)\n self.target.http.request.return_value = 'Searching for prt'\n self.assertTrue(self.target._check_webpage_for_prtname('for prt'))\n self.assertFalse(self.target._check_webpage_for_prtname('abc xyz'))\n\n def test_check_prt_generated(self):\n self.target = self.build_generic_mock_synergy(\n '_check_prt_generated', self.target)\n self.target.get_prt_status.return_value = {\n u'PRT': u'5', u'State': u'uploading-failed'}\n self.assertTrue(self.target._check_prt_generated('FULL'))\n self.target.get_prt_status.return_value = {\n u'PRT': u'0', u'State': u'idle'}\n self.assertFalse(self.target._check_prt_generated('FULL'))\n\n self.target.get_prt_status.reset_mock()\n self.target.get_prt_status.side_effect = [TngError(\"test failure\")]\n self.assertTrue(self.target._check_prt_generated('FULL'))\n\n def test_get_logs_failed_prt(self):\n target = self.build_generic_mock_synergy('get_logs', self.target)\n target.trigger_prt_and_wait.return_value = 'mypath/myprt.test'\n target.is_3pcc = False\n ex = Mock()\n ex._exit_status = 'failed'\n m = mock_open()\n\n ge = (\n 'tng_sl.device.endpoint.synergylite.synergylite_extended.'\n 'get_engine')\n with patch(ge) as eng, patch.object(builtins, 'open', m), patch(\n 'time.sleep'):\n eng.return_value = ex\n\n ret = Mock()\n ret.code = 200\n ret.length = 27\n target.http.download.return_value = ret\n\n result = target.get_logs('mypath')\n self.assertEqual(result, 'mypath/myprt.test')\n\n target.trigger_prt_and_wait.assert_called_once_with(\n log_dir='mypath', timeout=120.0)\n eng.assert_called_once_with()\n\n def test_get_logs_failed(self):\n target = self.build_generic_mock_synergy('get_logs', self.target)\n target.trigger_prt_and_wait.return_value = 'myprt.test'\n target.is_3pcc = False\n ex = Mock()\n ex._exit_status = 'failed'\n m = mock_open()\n with patch('tng.api.get_engine') as eng, patch(\n 'tng.api'), patch.object(builtins, 'open', m):\n eng.return_value = ex\n ret = Mock()\n # continue on to collect messages when the prt is empty\n ret.code = 200\n ret.length = 0 # length of zero is an empty file\n target.http.download.return_value = ret\n target.trigger_prt_and_wait.side_effect = [TimeoutError]\n target.http.get.return_value = (\n '523 - Cisco Systems - # RELEASE Model=88xx Version=sip88xx.12'\n '-0-1MN-176dev\\n[TZ=CST+6:00CDT+5:00,70,308] Wed Feb 8 '\n '11:55:43 2017')\n target._get_message_file_list.return_value = []\n target.get_logs('mypath')\n target._get_message_file_list.assert_called_once()\n\n def test_get_logs(self):\n target = self.build_generic_mock_synergy('get_logs', self.target)\n target = self.build_generic_mock_synergy(\n 'get_time_from_logfile', target)\n target.is_3pcc = \"True\"\n self.assertFalse(target.get_logs('junk'))\n target.is_3pcc = False\n target.http.get.side_effect = [\n '523 - Cisco Systems - # RELEASE Model=88xx Version=sip88xx.12-0-1'\n 'MN-176dev\\n[TZ=CST+6:00CDT+5:00,70,308] Wed Feb 8 11:55:43 2017',\n '524 - Cisco Systems - # RELEASE Model=88xx Version=sip88xx.12-0-1'\n 'MN-176dev\\n[TZ=CST+6:00CDT+5:00,70,308] Wed Feb 8 12:00:41 2017',\n '522 - Cisco Systems - # RELEASE Model=88xx Version=sip88xx.12-0-1'\n 'MN-176dev\\n[TZ=CST+6:00CDT+5:00,70,308] Wed Feb 8 11:50:45 2017',\n '']\n\n test_files = [\n 'FX/messages', 'FX/messages.0', 'FX/messages.1', 'FX/messages.BAD']\n test_path = \"/path/to/nowhere\"\n test_paths = [\n os.path.join(test_path, os.path.basename(p))\n for p in test_files]\n test_calls = [call(p, decode_body=False) for p in test_files]\n\n # we know the last one is bad because the http body is ''\n bad_path = test_paths.pop()\n # we know the method will copy the newest body to messages.log\n test_paths.append(os.path.join(test_path, 'messages.log'))\n target._get_message_file_list.return_value = copy.copy(test_files)\n my_open = Mock()\n open_mock = mock_open(my_open)\n with patch.object(builtins, 'open', open_mock):\n target.get_logs(test_path)\n\n for config_file in test_paths:\n self.assertTrue(\n call(config_file, 'w') in my_open.call_args_list)\n\n self.assertTrue(\n call(bad_path, 'w') not in my_open.call_args_list)\n target.http.get.assert_has_calls(test_calls)\n\n def test_trigger_prt_and_wait(self):\n target = self.build_generic_mock_synergy(\n 'trigger_prt_and_wait', self.target)\n\n target.gen_prt_log.return_value = \"my_prt\"\n target.local_log_dir = None\n target._is_fetched_gz_valid.return_value = True\n self.assertEqual(target.trigger_prt_and_wait(), \"./my_prt\")\n target.gen_prt_log.assert_called_once_with(\n prt_type='FULL noupload', timeout=120.0)\n target._is_fetched_gz_valid.return_value = False\n self.assertRaises(TimeoutError, target.trigger_prt_and_wait)\n\n def test__is_fetched_gz_valid(self):\n target = self.build_generic_mock_synergy(\n '_is_fetched_gz_valid', self.target)\n # http response is not 200 ok\n self.assertFalse(target._is_fetched_gz_valid('1', '2'))\n\n resp = Mock()\n resp.code = 200\n resp.length = 27\n target.http.download.return_value = resp\n tfile = (\n 'tng_sl.device.endpoint.synergylite.synergylite_extended.'\n 'is_tarfile')\n with patch(tfile) as tf:\n tf.return_value = False\n self.assertFalse(target._is_fetched_gz_valid('1', '2'))\n tf.return_value = True\n self.assertTrue(target._is_fetched_gz_valid('1', '2'))\n resp.code = 201\n self.assertFalse(target._is_fetched_gz_valid('1', '2'))\n resp.code = 200\n resp.length = 0\n self.assertFalse(target._is_fetched_gz_valid('1', '2'))\n\n def test_get_cnf_xml(self):\n target = self.build_generic_mock_synergy('get_cnf_xml', self.target)\n target.trigger_prt_and_wait.return_value = 'prt_file'\n tf = 'tng_sl.device.endpoint.synergylite.synergylite_extended.TarFile'\n with patch(tf) as mytf:\n tfile = Mock()\n mytf.open.return_value.__enter__.return_value = tfile\n\n # check negative case with missing files\n tfile.getnames.return_value = []\n self.assertFalse(target.get_cnf_xml())\n\n tfile.getnames.return_value = ['nope', 'yep.cnf.xml']\n myread = Mock()\n myread.read.return_value = \"myxmlcontents\"\n tfile.extractfile.return_value = myread\n\n self.assertEqual(target.get_cnf_xml(), 'myxmlcontents')\n\n def test__get_message_file_list(self):\n target = self.build_generic_mock_synergy(\n '_get_message_file_list', self.target)\n target.http.get.return_value = (\n 'some http stuff\\nmessages\\n/fS/messages\\ndup of /fs/messages\\n'\n '/fs/MESSAGES.0')\n self.assertEqual(target._get_message_file_list(), [\n 'FS/messages', 'FS/messages.0', 'FS/messages.1'])\n target.http.get.assert_called_once_with(\n 'CGI/Java/Serviceability?adapter=device.statistics.consolelog')\n\n def test__set_3pcc_ssh(self):\n target = self.build_generic_mock_synergy('_set_3pcc_ssh', self.target)\n target.ui.set_web_parameter_curl = Mock()\n target.credentials.get('default').username = 'root'\n target.credentials.get('default').password = 'cisco'\n target.ui.set_web_parameter_curl.return_value = True\n target.ui.get_phone_parameter_from_web_http = Mock()\n target.ui.get_phone_parameter_from_web_http.side_effect = [\n 'true', 'root']\n self.assertTrue(target._set_3pcc_ssh(timeout=0), 'set ssh fail')\n target.ui.get_phone_parameter_from_web_http.side_effect = [\n 'true', 'root123', 'true', 'root123',\n 'true', 'root123', 'true', 'root123',\n 'true', 'root123']\n self.assertFalse(target._set_3pcc_ssh(timeout=0), 'set ssh succeed')\n\n def test_dbus_automation_enable(self):\n target = self.build_generic_mock_synergy(\n 'dbus_automation_enabled', self.target)\n self.target.cli.exec_command.return_value = [\n '', '', 'Dbus Automation Interface Enabled'\n ]\n delattr(target.dbus, 'bus_names')\n self.assertFalse(hasattr(target.dbus, 'bus_names'))\n with target.dbus_automation_enabled():\n pass\n target.cli.exec_command.assert_called_once_with(\n 'dbus-automation enable')\n target._restart_action.assert_called_once_with()\n\n def test_dbus_automation_enable_already_enabled(self):\n target = self.build_generic_mock_synergy(\n 'dbus_automation_enabled', self.target)\n self.target.cli.exec_command.return_value = [\n '', '', 'Dbus Automation Interface Enabled'\n ]\n self.assertTrue(hasattr(target.dbus, 'bus_names'))\n with target.dbus_automation_enabled():\n pass\n target.cli.exec_command.assert_not_called()\n\n def test_validate_media(self):\n target = self.build_generic_mock_synergy('validate_media', self.target)\n target = self.build_generic_mock_synergy(\"_is_media_ready\", target)\n target = self.build_generic_mock_synergy(\n \"_check_packet_increase\", target)\n target = self.build_generic_mock_synergy(\n \"_compare_packet_counts\", target)\n\n mock_return_1 = [{\n \"StrmID\": \"1\", \"SndPkts\": \"5\", \"RcvPkts\": \"50\", \"Status\": \"R\",\n \"Direc\": \"Both\"}, {\n \"StrmID\": \"2\", \"SndPkts\": \"10\", \"RcvPkts\": \"10\", \"Status\": \"R\",\n \"Direc\": \"Both\"}]\n\n mock_return_2 = [{\n \"StrmID\": \"1\", \"SndPkts\": \"20\", \"RcvPkts\": \"30\", \"Status\": \"R\",\n \"Direc\": \"Both\"}, {\n \"StrmID\": \"2\", \"SndPkts\": \"20\", \"RcvPkts\": \"30\", \"Status\": \"R\",\n \"Direc\": \"Both\"}]\n\n target.cf.get_streams.side_effect = [mock_return_1, mock_return_2] * 8\n\n self.assertTrue(target.validate_media(ice_check=False))\n target.wait_for_ice_completed.test_assert_not_called()\n target.check_ICE_mode_and_path.assert_not_called()\n self.assertEqual(target.cf.get_streams.call_count, 2)\n target.cf.get_streams.reset_mock()\n\n target.ICE_OPTIMAL_PATH = 'ACTIVE'\n self.assertTrue(target.validate_media())\n target.wait_for_ice_completed.assert_called_once_with(timeout=10)\n target.check_ICE_mode_and_path.assert_called_once_with(\n media={'audio': 'Both'})\n self.assertEqual(target.cf.get_streams.call_count, 2)\n\n self.assertTrue(target.validate_media(media={\"audio\": \"Both\"}))\n self.assertTrue(target.validate_media(media={\"video\": \"Both\"}))\n self.assertRaises(\n KeyError, target.validate_media, media={\"audio\": \"Rxxx\"})\n\n self.assertTrue(target.validate_media(media={\"audio\": \"Rx\"}))\n\n self.assertRaises(\n KeyError, target.validate_media, media={\"audiox\": \"Both\"})\n mock_return_3 = [[{\n \"StrmID\": \"1\", \"SndPkts\": \"5\", \"RcvPkts\": \"50\", \"Status\": \"R\",\n \"Direc\": \"Both\"}]]\n target.cf.get_streams.reset_mock()\n target.cf.get_streams.side_effect = mock_return_3 * 100\n self.assertRaises(\n TimeoutError, target.validate_media, media={\"video\": \"Both\"})\n\n # negative case where packet counts not expected to increase\n self.assertTrue(target.validate_media(\n media={\"audio\": \"Both\"}, media_flowing=False))\n\n def test_validate_media_MOH(self):\n target = self.build_generic_mock_synergy('validate_media', self.target)\n target = self.build_generic_mock_synergy(\"_is_media_ready\", target)\n target = self.build_generic_mock_synergy(\n \"_check_packet_increase\", target)\n target = self.build_generic_mock_synergy(\n \"_compare_packet_counts\", target)\n\n mock_MOH1 = [{\n \"StrmID\": \"1\", \"SndPkts\": \"5\", \"RcvPkts\": \"50\", \"Status\": \"NR\",\n \"Direc\": \"Both\"}, {\n \"StrmID\": \"2\", \"SndPkts\": \"10\", \"RcvPkts\": \"10\",\n \"Status\": \"NR\", \"Direc\": \"Both\"}]\n\n mock_MOH2 = [{\n \"StrmID\": \"1\", \"SndPkts\": \"20\", \"RcvPkts\": \"30\", \"Status\": \"NR\",\n \"Direc\": \"Both\"}, {\n \"StrmID\": \"2\", \"SndPkts\": \"20\", \"RcvPkts\": \"30\",\n \"Status\": \"NR\", \"Direc\": \"Both\"}]\n\n target.cf.get_streams.side_effect = [mock_MOH1, mock_MOH2] * 2\n target.ICE_OPTIMAL_PATH = 'DEFAULT'\n self.assertTrue(target.validate_media(ready_status='NR'))\n\n def test_on_board_onprem(self):\n self.target = self.build_generic_mock_synergy(\n 'on_board_onprem', self.target)\n output = 'login onprem 8893501470247981'\n self.target.cli.exec_command.return_value = output\n self.assertEqual(\n self.target.on_board_onprem('8893501470247981', 0.001),\n output)\n\n\nclass TestWaitForCallStates(TestCase):\n\n def setUp(self):\n concurrent_patcher = patch(\n 'tng_sl.device.endpoint.synergylite.synergylite_extended.'\n 'concurrent_map',\n lambda c, devices, lines, calls, data: [c(\n dev, l, cl, dat)\n for dev, l, cl, dat in zip(devices, lines, calls, data)])\n concurrent_patcher.start()\n self.addCleanup(concurrent_patcher.stop)\n\n self.tng_wait = Mock()\n wait_patcher = patch(\n 'tng_sl.device.endpoint.synergylite.synergylite_extended.tng_wait',\n self.tng_wait)\n wait_patcher.start()\n self.addCleanup(wait_patcher.stop)\n\n def build_mock_device(self, side_effect=['Unknown']):\n device = Mock(spec=SynergyLiteExtended)\n device.cf = Mock()\n device.cf.get_line_info = Mock()\n device.cf.get_line_info.side_effect = [\n {'State': s} for s in side_effect]\n device.log = Mock()\n return device\n\n def test_basic(self):\n d1 = self.build_mock_device(['IDLE'])\n d2 = self.build_mock_device(['IDLE'])\n d3 = self.build_mock_device(['IDLE'])\n wait_for_call_states((d1, d2, d3), ('IDLE', 'IDLE', 'IDLE'))\n\n d1.cf.get_line_info.assert_called_once_with(1)\n d2.cf.get_line_info.assert_called_once_with(1)\n d3.cf.get_line_info.assert_called_once_with(1)\n self.tng_wait.assert_called_once_with(\n 1.0, reason=('wait until the status of phone is stable'))\n\n def test_alt_lines(self):\n d1 = self.build_mock_device(['IDLE'])\n d2 = self.build_mock_device(['BUSY'])\n d3 = self.build_mock_device(['IDLE'])\n wait_for_call_states(\n (d1, d2, d3), ('IDLE', 'BUSY', 'IDLE'), (2, 3, 4))\n\n d1.cf.get_line_info.assert_called_once_with(2)\n d2.cf.get_line_info.assert_called_once_with(3)\n d3.cf.get_line_info.assert_called_once_with(4)\n\n def test_alt_calls_only(self):\n d1 = self.build_mock_device(['IDLE', 'PROCEEDING'])\n d2 = self.build_mock_device(['BUSY'])\n d3 = self.build_mock_device(['IDLE'])\n wait_for_call_states(\n (d1, d2, d3), ('PROCEEDING', 'BUSY', 'IDLE'), call_ids=(1, 1, 0),\n poll_interval=0.01)\n\n d1.cf.get_line_info.assert_called_with(1)\n self.assertEqual(d1.cf.get_line_info.call_count, 2)\n d2.cf.get_line_info.assert_called_once_with(1)\n d3.cf.get_line_info.assert_called_once_with(0)\n\n def test_error(self):\n d1 = self.build_mock_device(['IDLE'])\n d2 = self.build_mock_device(['NOT_IDLE', 'NOT_IDLE', 'NOT_IDLE'])\n d3 = self.build_mock_device(['IDLE'])\n self.assertRaises(TimeoutError, wait_for_call_states, (\n d1, d2, d3), ('IDLE', 'IDLE', 'IDLE'), timeout=.01,\n poll_interval=0.02)\n\n def test_set_pre_wait(self):\n d1 = self.build_mock_device(['IDLE'])\n d2 = self.build_mock_device(['BUSY'])\n d3 = self.build_mock_device(['IDLE'])\n wait_for_call_states(\n (d1, d2, d3), ('IDLE', 'BUSY', 'IDLE'), (2, 3, 4),\n pre_wait=123)\n d1.cf.get_line_info.assert_called_once_with(2)\n d2.cf.get_line_info.assert_called_once_with(3)\n d3.cf.get_line_info.assert_called_once_with(4)\n self.tng_wait.assert_called_once_with(\n 123, reason=('wait until the status of phone is stable'))\n\n def test_no_pre_wait(self):\n d1 = self.build_mock_device(['IDLE'])\n d2 = self.build_mock_device(['BUSY'])\n d3 = self.build_mock_device(['IDLE'])\n wait_for_call_states(\n (d1, d2, d3), ('IDLE', 'BUSY', 'IDLE'), (2, 3, 4),\n pre_wait=0)\n d1.cf.get_line_info.assert_called_once_with(2)\n d2.cf.get_line_info.assert_called_once_with(3)\n d3.cf.get_line_info.assert_called_once_with(4)\n self.tng_wait.assert_not_called()\n\n def test_no_state(self):\n d1 = self.build_mock_device(['IDLE'])\n d2 = self.build_mock_device(['BUSY'])\n d3 = self.build_mock_device(['IDLE'])\n # sometimes get_line_info is empty\n d3.cf.get_line_info.side_effect = [\n dict(), {'State': 'IDLE'}]\n wait_for_call_states(\n (d1, d2, d3), ('IDLE', 'BUSY', 'IDLE'), (2, 3, 4),\n pre_wait=0)\n d1.cf.get_line_info.assert_called_once_with(2)\n d2.cf.get_line_info.assert_called_once_with(3)\n d3.cf.get_line_info.assert_has_calls([call(4), call(4)])\n self.tng_wait.assert_not_called()\n\n\nclass TestMediaAvailable(TestCase):\n def build_mock_device(self, side_effect=[True]):\n device = Mock(spec=SynergyLiteExtended)\n device.is_video_device.side_effect = side_effect\n device.log = Mock()\n return device\n\n def test_media_available_video(self):\n d1 = self.build_mock_device()\n d2 = self.build_mock_device()\n d3 = self.build_mock_device()\n media = media_available((d1, d2, d3))\n self.assertEqual(media, {'audio': 'Both', 'video': 'Both'})\n\n d1.is_video_device.assert_called_once_with()\n d2.is_video_device.assert_called_once_with()\n d3.is_video_device.assert_called_once_with()\n\n def test_media_available_one_audio(self):\n d1 = self.build_mock_device()\n d2 = self.build_mock_device([False])\n d3 = self.build_mock_device()\n media = media_available((d1, d2, d3))\n self.assertEqual(media, {'audio': 'Both'})\n\n def test_media_available_all_audio(self):\n d1 = self.build_mock_device([False])\n d2 = self.build_mock_device([False])\n d3 = self.build_mock_device([False])\n media = media_available((d1, d2, d3))\n self.assertEqual(media, {'audio': 'Both'})\n\n\nclass TestVerifyMediaAllDevices(TestCase):\n\n def setUp(self):\n concurrent_patcher = patch(\n 'tng_sl.device.endpoint.synergylite.synergylite_extended.'\n 'concurrent',\n lambda funcs, **kwargs: [f(**kwargs) for f in funcs])\n concurrent_patcher.start()\n self.addCleanup(concurrent_patcher.stop)\n\n self.tng_wait = Mock()\n wait_patcher = patch(\n 'tng_sl.device.endpoint.synergylite.synergylite_extended.tng_wait',\n self.tng_wait)\n wait_patcher.start()\n self.addCleanup(wait_patcher.stop)\n\n def build_mock_device(self, vside_effect=[True], mside_effect=[True]):\n device = Mock(spec=SynergyLiteExtended)\n device.is_video_device.side_effect = vside_effect\n device.validate_media.side_effect = mside_effect\n device.log = Mock()\n return device\n\n def test_verify_media_all_devices_video(self):\n d1 = self.build_mock_device()\n d2 = self.build_mock_device()\n d3 = self.build_mock_device()\n self.assertTrue(verify_media_all_devices((d1, d2, d3)))\n\n d1.is_video_device.assert_called_once_with()\n d2.is_video_device.assert_called_once_with()\n d3.is_video_device.assert_called_once_with()\n\n kwargs = {\n 'media': {'audio': 'Both', 'video': 'Both'},\n 'media_flowing': True, 'timeout': 10}\n d1.validate_media.assert_called_once_with(**kwargs)\n d2.validate_media.assert_called_once_with(**kwargs)\n d3.validate_media.assert_called_once_with(**kwargs)\n\n def test_verify_media_all_devices_one_audio(self):\n d1 = self.build_mock_device()\n d2 = self.build_mock_device(vside_effect=[False])\n d3 = self.build_mock_device()\n self.assertTrue(verify_media_all_devices((d1, d2, d3)))\n\n d1.is_video_device.assert_called_once_with()\n d2.is_video_device.assert_called_once_with()\n d3.is_video_device.assert_called_once_with()\n\n kwargs = {\n 'media': {'audio': 'Both'}, 'media_flowing': True, 'timeout': 10}\n d1.validate_media.assert_called_once_with(**kwargs)\n d2.validate_media.assert_called_once_with(**kwargs)\n d3.validate_media.assert_called_once_with(**kwargs)\n\n def test_verify_media_all_devices_timeout(self):\n d1 = self.build_mock_device()\n d2 = self.build_mock_device(\n vside_effect=[False], mside_effect=[TimeoutError])\n d3 = self.build_mock_device()\n self.assertRaises(TimeoutError, verify_media_all_devices, (d1, d2, d3))\n\n def test_verify_media_all_devices_unexpected_media(self):\n d1 = self.build_mock_device(mside_effect=[False])\n d2 = self.build_mock_device(mside_effect=[True])\n d3 = self.build_mock_device(mside_effect=[False])\n self.assertRaises(\n TimeoutError, verify_media_all_devices, (d1, d2, d3),\n media_flowing=False)\n\n def test_verify_media_all_devices_expect_no_media(self):\n d1 = self.build_mock_device(mside_effect=[False])\n d2 = self.build_mock_device(mside_effect=[False])\n d3 = self.build_mock_device(mside_effect=[False])\n self.assertTrue(verify_media_all_devices(\n (d1, d2, d3), media_flowing=False))\n\n d1.is_video_device.assert_called_once_with()\n d2.is_video_device.assert_called_once_with()\n d3.is_video_device.assert_called_once_with()\n\n kwargs = {\n 'media': {'audio': 'Both', 'video': 'Both'},\n 'media_flowing': False, 'timeout': 2}\n d1.validate_media.assert_called_once_with(**kwargs)\n d2.validate_media.assert_called_once_with(**kwargs)\n d3.validate_media.assert_called_once_with(**kwargs)\n\n\nclass TestValueFromNameOrIndex(TestCase):\n def test_no_match(self):\n ldata = ['Zero', 'One', 'Two', 'Final']\n # exact match\n self.assertEqual('Two', value_from_name_or_index('Two', ldata))\n self.assertEqual('Final', value_from_name_or_index('3 (Final)', ldata))\n # lowercase match\n self.assertEqual('One', value_from_name_or_index('one', ldata))\n # incoming None -> default\n self.assertEqual('One', value_from_name_or_index(None, ldata, 1))\n # incoming Integer\n self.assertEqual('Final', value_from_name_or_index(3, ldata, 1))\n # incoming Integer too big -> default\n self.assertEqual('Final', value_from_name_or_index(9, ldata, 3))\n # incoming No Match -> Default\n self.assertEqual('Zero', value_from_name_or_index('Junk', ldata))\n\n def test_all_match_no_overlaps(self):\n def check_overlaps(ldata):\n for akey in ldata:\n self.assertEqual(akey, value_from_name_or_index(akey, ldata))\n self.assertEqual(akey, value_from_name_or_index(\n \"(0) {}\".format(akey), ldata))\n\n check_overlaps(SynergyLiteExtended.ICE_OPTIMAL_PATHS)\n check_overlaps(SynergyLiteExtended.ICE_TRAVERSAL_MODES)\n check_overlaps(SynergyLiteExtended.ICE_STATE_NAMES)\n\n\nclass AtaExtendedTestCase(TestCase):\n def build_generic_mock_ata(self, method_name, target=None):\n target = target or Mock(spec=MockATA)\n target_method = getattr(MockATA, method_name)\n setattr(target, method_name, lambda *args, **kwargs: target_method(\n target, *args, **kwargs))\n target.log = Mock(spec=logging.Logger)\n return target\n\n\nclass TestATAExtended(AtaExtendedTestCase):\n\n def setUp(self):\n engine = Engine(None)\n device_ip = '1.1.1.1'\n self.target = MockATA(device_ip, engine)\n self.target.log = Mock(spec=logging.Logger)\n self.target.handle_shutdown = Mock()\n self.target.handle_startup = Mock()\n self.target.cli = Mock(spec=ATACommandHandler)\n self.target.cli.parser = SynergyLiteCommandParser()\n\n def test_on_board_ce(self):\n self.target = self.build_generic_mock_ata(\n 'on_board_ce', self.target)\n self.target.cli.exec_command.return_value = [\n 'login ce my.domain me pwd 0']\n self.target.reset = Mock()\n self.target.reset.return_value = True\n self.target.handle_shutdown.return_value = True\n self.target.handle_startup.return_value = True\n self.assertEqual(\n self.target.on_board_ce('my.domain', 'me', 'pwd', 0.001, 1),\n 'login ce my.domain me pwd 0')\n\n self.target.cli.exec_command.assert_called_once_with(\n 'login ce my.domain me pwd 1', timeout=0.001)\n self.target.reset.assert_called_once_with(\n 'soft', wait=230, restore_conn=False)\n\n self.target.cli.exec_command.return_value = [\n 'login ce my.domain me pwd 1']\n\n self.assertRaises(\n TngError, self.target.on_board_ce, 'my.domain', 'me', 'pwd', 0.001,\n 1)\n\n def test_on_board_huron(self):\n self.target = self.build_generic_mock_ata(\n 'on_board_huron', self.target)\n self.target.cli.exec_command.return_value = [\n 'login huron 8893501470247980 0'\n ]\n self.target.reset = Mock()\n self.target.reset.return_value = True\n self.target.handle_shutdown.return_value = True\n self.target.handle_startup.return_value = True\n self.assertEqual(\n self.target.on_board_huron('8893501470247980', 0.001, 1),\n 'login huron 8893501470247980 0')\n\n def test_gen_prt_log(self):\n self.target = self.build_generic_mock_ata('gen_prt_log', self.target)\n self.target.cli.exec_command.return_value = [\n u'prt create^J/pcm/prt-20170323-010127-34DBFD19A149.tar.gz']\n self.assertEqual(\n self.target.gen_prt_log(),\n 'prt-20170323-010127-34DBFD19A149.tar.gz')\n self.target.cli.exec_command.return_value = []\n self.assertRaises(TngError, self.target.gen_prt_log)\n", "sub_path": "tng_sl/test/endpoint/device/test_synergylite.py", "file_name": "test_synergylite.py", "file_ext": "py", "file_size_in_byte": 78581, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sys.version_info", "line_number": 29, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 34, "usage_type": "call"}, {"api_name": "tng.device.endpoint.ata.ATA", "line_number": 39, "usage_type": "name"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 39, "usage_type": "name"}, {"api_name": "twisted.trial.unittest.TestCase", "line_number": 44, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 47, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 47, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 48, "usage_type": "call"}, {"api_name": "logging.Logger", "line_number": 48, "usage_type": "attribute"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended._log_failure_to_device", "line_number": 51, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 51, "usage_type": "name"}, {"api_name": "mock.call", "line_number": 59, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 62, "usage_type": "call"}, {"api_name": "twisted.trial.unittest.TestCase", "line_number": 66, "usage_type": "name"}, {"api_name": "tng_sl.test.utilities.Clock", "line_number": 69, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 70, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 74, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 74, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 75, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 78, "usage_type": "attribute"}, {"api_name": "mock.Mock", "line_number": 81, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 81, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 82, "usage_type": "call"}, {"api_name": "tng_sl.plugins.synergylite_cli_common.SynergyLiteCommandHandler", "line_number": 82, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 83, "usage_type": "call"}, {"api_name": "logging.Logger", "line_number": 83, "usage_type": "attribute"}, {"api_name": "mock.Mock", "line_number": 84, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 85, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended._log_max_chars_per_syslog", "line_number": 91, "usage_type": "attribute"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 91, "usage_type": "name"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended._log_max_msg_len", "line_number": 93, "usage_type": "attribute"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 93, "usage_type": "name"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended.log_to_device", "line_number": 95, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 95, "usage_type": "name"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended._prefix", "line_number": 97, "usage_type": "attribute"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 97, "usage_type": "name"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended._log_max_chars_per_syslog", "line_number": 108, "usage_type": "attribute"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 108, "usage_type": "name"}, {"api_name": "mock.call", "line_number": 120, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 123, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended._log_max_chars_per_syslog", "line_number": 126, "usage_type": "attribute"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 126, "usage_type": "name"}, {"api_name": "mock.call", "line_number": 131, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended._log_max_chars_per_syslog", "line_number": 133, "usage_type": "attribute"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 133, "usage_type": "name"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended._log_max_msg_len", "line_number": 138, "usage_type": "attribute"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 138, "usage_type": "name"}, {"api_name": "mock.call", "line_number": 152, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 157, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 162, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended._log_max_chars_per_syslog", "line_number": 169, "usage_type": "attribute"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 169, "usage_type": "name"}, {"api_name": "mock.call", "line_number": 173, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended._log_max_msg_len", "line_number": 183, "usage_type": "attribute"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 183, "usage_type": "name"}, {"api_name": "mock.call", "line_number": 194, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended._log_max_chars_per_syslog", "line_number": 200, "usage_type": "attribute"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 200, "usage_type": "name"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 208, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 213, "usage_type": "call"}, {"api_name": "logging.Logger", "line_number": 213, "usage_type": "attribute"}, {"api_name": "twisted.trial.unittest.TestCase", "line_number": 219, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 221, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 221, "usage_type": "name"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 222, "usage_type": "argument"}, {"api_name": "mock.Mock", "line_number": 225, "usage_type": "call"}, {"api_name": "logging.Logger", "line_number": 225, "usage_type": "attribute"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended.ICE_OPTIMAL_PATHS", "line_number": 226, "usage_type": "attribute"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 226, "usage_type": "name"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended.ICE_TRAVERSAL_MODES", "line_number": 227, "usage_type": "attribute"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 227, "usage_type": "name"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended.ICE_STATE_NAMES", "line_number": 228, "usage_type": "attribute"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 228, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 229, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 230, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 238, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 238, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 239, "usage_type": "call"}, {"api_name": "tng_sl.plugins.synergylite_cf_common.SynergyLiteCfNamespace", "line_number": 239, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 240, "usage_type": "call"}, {"api_name": "tng_sl.plugins.synergylite_cli_common.SynergyLiteCommandHandler", "line_number": 240, "usage_type": "name"}, {"api_name": "tng_sl.plugins.synergylite_cli_common.SynergyLiteCommandParser", "line_number": 241, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 242, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 243, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 244, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 245, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 246, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 247, "usage_type": "call"}, {"api_name": "logging.Logger", "line_number": 247, "usage_type": "attribute"}, {"api_name": "mock.Mock", "line_number": 248, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 249, "usage_type": "call"}, {"api_name": "tng_sl.plugins.synergylite3pcc_ui_common.SynergyLite3pccUiNamespace", "line_number": 249, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 263, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 264, "usage_type": "call"}, {"api_name": "tng_sl.test.utilities.Clock", "line_number": 270, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 271, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 276, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 300, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 281, "usage_type": "call"}, {"api_name": "time.strptime", "line_number": 335, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 338, "usage_type": "call"}, {"api_name": "tng.error.TngError", "line_number": 351, "usage_type": "name"}, {"api_name": "tng.error.TngError", "line_number": 403, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 431, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 461, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 462, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 427, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 468, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 526, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 527, "usage_type": "call"}, {"api_name": "tng.error.TimeoutError", "line_number": 545, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 564, "usage_type": "call"}, {"api_name": "tng.error.TngError", "line_number": 572, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 617, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 659, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.UnexpectedIceError", "line_number": 723, "usage_type": "argument"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.UnexpectedIceError", "line_number": 745, "usage_type": "argument"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.UnexpectedIceError", "line_number": 766, "usage_type": "argument"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.UnexpectedIceError", "line_number": 776, "usage_type": "argument"}, {"api_name": "tng.error.TimeoutError", "line_number": 811, "usage_type": "call"}, {"api_name": "tng.error.TimeoutError", "line_number": 813, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 801, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 864, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.value_from_name_or_index", "line_number": 881, "usage_type": "call"}, {"api_name": "tng.error.TimeoutError", "line_number": 890, "usage_type": "argument"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.value_from_name_or_index", "line_number": 897, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.value_from_name_or_index", "line_number": 901, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 906, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.UnexpectedIceError", "line_number": 931, "usage_type": "argument"}, {"api_name": "tng.error.TimeoutError", "line_number": 965, "usage_type": "argument"}, {"api_name": "tng.error.TimeoutError", "line_number": 974, "usage_type": "argument"}, {"api_name": "tng.error.TngError", "line_number": 1249, "usage_type": "argument"}, {"api_name": "tng.error.TngError", "line_number": 1271, "usage_type": "argument"}, {"api_name": "tng.error.TngError", "line_number": 1291, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1298, "usage_type": "call"}, {"api_name": "mock.mock_open", "line_number": 1300, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 1305, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 1305, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1309, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1325, "usage_type": "call"}, {"api_name": "mock.mock_open", "line_number": 1327, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 1328, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 1329, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 1329, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 1331, "usage_type": "call"}, {"api_name": "tng.error.TimeoutError", "line_number": 1336, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 1365, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1365, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 1365, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 1367, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1372, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1372, "usage_type": "attribute"}, {"api_name": "copy.copy", "line_number": 1373, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1374, "usage_type": "call"}, {"api_name": "mock.mock_open", "line_number": 1375, "usage_type": "call"}, {"api_name": "mock.patch.object", "line_number": 1376, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 1376, "usage_type": "name"}, {"api_name": "mock.call", "line_number": 1381, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 1384, "usage_type": "call"}, {"api_name": "tng.error.TimeoutError", "line_number": 1398, "usage_type": "argument"}, {"api_name": "mock.Mock", "line_number": 1406, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 1413, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 1428, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1429, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1437, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1456, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1460, "usage_type": "call"}, {"api_name": "tng.error.TimeoutError", "line_number": 1545, "usage_type": "argument"}, {"api_name": "twisted.trial.unittest.TestCase", "line_number": 1585, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 1588, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1597, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 1598, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1605, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 1605, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 1606, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1607, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1610, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.wait_for_call_states", "line_number": 1617, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.wait_for_call_states", "line_number": 1629, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.wait_for_call_states", "line_number": 1640, "usage_type": "call"}, {"api_name": "tng.error.TimeoutError", "line_number": 1653, "usage_type": "argument"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.wait_for_call_states", "line_number": 1653, "usage_type": "argument"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.wait_for_call_states", "line_number": 1661, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.wait_for_call_states", "line_number": 1674, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.wait_for_call_states", "line_number": 1689, "usage_type": "call"}, {"api_name": "mock.call", "line_number": 1694, "usage_type": "call"}, {"api_name": "twisted.trial.unittest.TestCase", "line_number": 1698, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 1700, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 1700, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 1702, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.media_available", "line_number": 1709, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.media_available", "line_number": 1720, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.media_available", "line_number": 1727, "usage_type": "call"}, {"api_name": "twisted.trial.unittest.TestCase", "line_number": 1731, "usage_type": "name"}, {"api_name": "mock.patch", "line_number": 1734, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1741, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 1742, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1749, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 1749, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 1752, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.verify_media_all_devices", "line_number": 1759, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.verify_media_all_devices", "line_number": 1776, "usage_type": "call"}, {"api_name": "tng.error.TimeoutError", "line_number": 1791, "usage_type": "name"}, {"api_name": "tng.error.TimeoutError", "line_number": 1793, "usage_type": "argument"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.verify_media_all_devices", "line_number": 1793, "usage_type": "argument"}, {"api_name": "tng.error.TimeoutError", "line_number": 1800, "usage_type": "argument"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.verify_media_all_devices", "line_number": 1800, "usage_type": "argument"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.verify_media_all_devices", "line_number": 1807, "usage_type": "call"}, {"api_name": "twisted.trial.unittest.TestCase", "line_number": 1822, "usage_type": "name"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.value_from_name_or_index", "line_number": 1826, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.value_from_name_or_index", "line_number": 1827, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.value_from_name_or_index", "line_number": 1829, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.value_from_name_or_index", "line_number": 1831, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.value_from_name_or_index", "line_number": 1833, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.value_from_name_or_index", "line_number": 1835, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.value_from_name_or_index", "line_number": 1837, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.value_from_name_or_index", "line_number": 1842, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.value_from_name_or_index", "line_number": 1843, "usage_type": "call"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended.ICE_OPTIMAL_PATHS", "line_number": 1846, "usage_type": "attribute"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 1846, "usage_type": "name"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended.ICE_TRAVERSAL_MODES", "line_number": 1847, "usage_type": "attribute"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 1847, "usage_type": "name"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended.ICE_STATE_NAMES", "line_number": 1848, "usage_type": "attribute"}, {"api_name": "tng_sl.device.endpoint.synergylite.synergylite_extended.SynergyLiteExtended", "line_number": 1848, "usage_type": "name"}, {"api_name": "twisted.trial.unittest.TestCase", "line_number": 1851, "usage_type": "name"}, {"api_name": "mock.Mock", "line_number": 1853, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1857, "usage_type": "call"}, {"api_name": "logging.Logger", "line_number": 1857, "usage_type": "attribute"}, {"api_name": "tng.frontend.engine.Engine", "line_number": 1864, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1867, "usage_type": "call"}, {"api_name": "logging.Logger", "line_number": 1867, "usage_type": "attribute"}, {"api_name": "mock.Mock", "line_number": 1868, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1869, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1870, "usage_type": "call"}, {"api_name": "tng_sl.plugins.ata_cli_common.ATACommandHandler", "line_number": 1870, "usage_type": "name"}, {"api_name": "tng_sl.plugins.synergylite_cli_common.SynergyLiteCommandParser", "line_number": 1871, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 1878, "usage_type": "call"}, {"api_name": "tng.error.TngError", "line_number": 1895, "usage_type": "argument"}, {"api_name": "mock.Mock", "line_number": 1904, "usage_type": "call"}, {"api_name": "tng.error.TngError", "line_number": 1920, "usage_type": "argument"}]}
+{"seq_id": "625036460", "text": "#!/usr/bin/env python3\nimport argparse\nimport asyncio\nimport gi\nimport logging\nimport os\nimport tempfile\nimport unittest\nimport xmlrunner\n\n# Add the current directory to the PLUGINDIR so we can use the plugin\n# file added here.\nos.environ[\"PLUGINDIR\"] += (\":\" + os.path.dirname(os.path.realpath(__file__)))\ngi.require_version(\"RwDts\", \"1.0\")\ngi.require_version(\"RwVnfrYang\", \"1.0\")\nfrom gi.repository import (\n RwDts,\n RwVnfrYang,\n)\n\nimport rift.tasklets\nimport rift.test.dts\n\ngi.require_version('RwKeyspec', '1.0')\nfrom gi.repository.RwKeyspec import quoted_key\n\nclass RwLogTestCase(rift.test.dts.AbstractDTSTest):\n # Disable the log_utest_mode so that log messages actually get logged\n # using the rwlog handler since that is what we are testing here.\n log_utest_mode = False\n\n @classmethod\n def configure_suite(cls, rwmain):\n pass\n\n @classmethod\n def start_test_tasklet(cls):\n cls.rwmain.add_tasklet(\n os.path.join(\n os.path.dirname(os.path.realpath(__file__)),\n 'reprotesttasklet-python'\n ),\n 'reprotesttasklet-python'\n )\n\n @classmethod\n def configure_schema(cls):\n return RwVnfrYang.get_schema()\n\n @classmethod\n def configure_timeout(cls):\n return 1000000\n\n def configure_test(self, loop, test_id):\n self.log.debug(\"STARTING - %s\", self.id())\n self.tinfo = self.new_tinfo(self.id())\n self.dts = rift.tasklets.DTS(self.tinfo, self.schema, self.loop)\n\n @rift.test.dts.async_test\n def test_tasklet_logging(self):\n self.start_test_tasklet()\n\n # The logtesttasklet signals being done, by moving into DTS Running state\n yield from self.wait_for_tasklet_running(\"reprotesttasklet-python\")\n @asyncio.coroutine\n def reader():\n while True:\n res_iter = yield from self.dts.query_read(\"D,/vnfr:vnfr-catalog/vnfr:vnfr[vnfr:id={}]/vnfr:vdur[vnfr:id={}]/rw-vnfr:nfvi-metrics\".format(\n quoted_key(\"a7f30def-0942-4425-8454-1ffe02b7db1e\"), quoted_key(\"a7f30def-0942-4425-8454-1ffe02b7db1e\"),\n ))\n for ent in res_iter:\n res = yield from ent\n metrics = res.result\n self.log.debug(\"got metrics result: %s\", metrics)\n\n for _ in range(20):\n self.loop.create_task(reader())\n\n while True:\n yield from asyncio.sleep(.001, loop=self.loop)\n\n\ndef main():\n runner = xmlrunner.XMLTestRunner(output=os.environ[\"RIFT_MODULE_TEST\"])\n\n parser = argparse.ArgumentParser()\n parser.add_argument('-v', '--verbose', action='store_true')\n args, _ = parser.parse_known_args()\n\n RwLogTestCase.log_level = logging.DEBUG if args.verbose else logging.WARN\n\n unittest.main(testRunner=runner)\n\nif __name__ == '__main__':\n main()\n", "sub_path": "osm/SO/rwlaunchpad/plugins/rwmonitor/test/repro_tasklet_test.py", "file_name": "repro_tasklet_test.py", "file_ext": "py", "file_size_in_byte": 2913, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.environ", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 13, "usage_type": "call"}, {"api_name": "gi.require_version", "line_number": 14, "usage_type": "call"}, {"api_name": "gi.require_version", "line_number": 15, "usage_type": "call"}, {"api_name": "gi.require_version", "line_number": 24, "usage_type": "call"}, {"api_name": "rift.tasklets.test", "line_number": 27, "usage_type": "attribute"}, {"api_name": "rift.tasklets", "line_number": 27, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 40, "usage_type": "call"}, {"api_name": "gi.repository.RwVnfrYang.get_schema", "line_number": 48, "usage_type": "call"}, {"api_name": "gi.repository.RwVnfrYang", "line_number": 48, "usage_type": "name"}, {"api_name": "rift.tasklets.tasklets.DTS", "line_number": 57, "usage_type": "call"}, {"api_name": "rift.tasklets.tasklets", "line_number": 57, "usage_type": "attribute"}, {"api_name": "rift.tasklets", "line_number": 57, "usage_type": "name"}, {"api_name": "gi.repository.RwKeyspec.quoted_key", "line_number": 69, "usage_type": "call"}, {"api_name": "asyncio.coroutine", "line_number": 65, "usage_type": "attribute"}, {"api_name": "asyncio.sleep", "line_number": 80, "usage_type": "call"}, {"api_name": "rift.tasklets.test", "line_number": 59, "usage_type": "attribute"}, {"api_name": "rift.tasklets", "line_number": 59, "usage_type": "name"}, {"api_name": "xmlrunner.XMLTestRunner", "line_number": 84, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 84, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 86, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 90, "usage_type": "attribute"}, {"api_name": "logging.WARN", "line_number": 90, "usage_type": "attribute"}, {"api_name": "unittest.main", "line_number": 92, "usage_type": "call"}]}
+{"seq_id": "199997056", "text": "import contextlib\nwith contextlib.redirect_stdout(None):\n import pygame, sys, random, math, time, ast, os\n from pygame.locals import *\n from multiprocessing import Pool\n\n\npool = Pool(os.cpu_count())\n\n\n\ndef events(segments):\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n pygame.quit()\n sys.exit()\n\n elif event.type == pygame.MOUSEBUTTONUP:\n Mouse = pygame.mouse.get_pos()\n segments.append([[Mouse[0]-25, Mouse[1]-25], [Mouse[0]+25, Mouse[1]-25]])\n segments.append([[Mouse[0]+25, Mouse[1]-25], [Mouse[0]+25, Mouse[1]+25]])\n segments.append([[Mouse[0]+25, Mouse[1]+25], [Mouse[0]-25, Mouse[1]+25]])\n segments.append([[Mouse[0]-25, Mouse[1]+25], [Mouse[0]-25, Mouse[1]-25]])\n return segments\n\n elif event.type == pygame.KEYDOWN and pygame.key == k_F5:\n print(\"F5\")\n\n return segments\n\n\ndef draw(screen, clock, segments):\n for seg in segments:\n pygame.draw.line(screen, [0,0,255], [seg[0][0], seg[0][1]],[seg[1][0], seg[1][1]],1)\n fps = font.render(str(int(clock.get_fps()))+\" FPS\", True, pygame.Color('Green'))\n screen.blit(fps, [0,0])\n\n\nif __name__ == \"__main__\":\n pygame.init()\n screen = pygame.display.set_mode([1280, 720])\n pygame.display.set_caption(\"Ray casting Engine\")\n font=pygame.font.SysFont(\"Courier\", 20, bold=True)\n clock = pygame.time.Clock()\n\n ##LINE SEGMENTS\n segments = [\n ##Border\n [[-1,-1], [1280,-1]],\n [[1280,-1], [1280,720]],\n [[1280,720], [-1,720]],\n [[-1,720], [-1,-1]]]\n\n while True:\n screen.fill([0,0,0])\n segments = events(segments)\n\n draw(screen, clock, segments)\n\n pygame.display.flip()\n clock.tick()\n", "sub_path": "Ray casting 10.1.1.1.py", "file_name": "Ray casting 10.1.1.1.py", "file_ext": "py", "file_size_in_byte": 1791, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "contextlib.redirect_stdout", "line_number": 2, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 8, "usage_type": "call"}, {"api_name": "os.cpu_count", "line_number": 8, "usage_type": "call"}, {"api_name": "pygame.event.get", "line_number": 13, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 14, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 15, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 16, "usage_type": "call"}, {"api_name": "pygame.MOUSEBUTTONUP", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pos", "line_number": 19, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pygame.key", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pygame.draw.line", "line_number": 34, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pygame.Color", "line_number": 35, "usage_type": "call"}, {"api_name": "pygame.init", "line_number": 40, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 41, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 41, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 42, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 42, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 43, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 43, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 44, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 44, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 60, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 60, "usage_type": "attribute"}]}
+{"seq_id": "510155065", "text": "__author__ = 'Jiggy'\n# Uses National Data on the relative frequency of given names\n# in the population of U.S. births,\n# stored in a directory \"names\", in files named \"yobxxxx.txt\n# with xxxx (the year of birth)\n# ranging from 1880 to2013.\n# Prompts the user for a first name, and finds out the first year\n# when this name was most popular in terms of frequency of names being given,\n# as a female name and as a male name.\n#\n# Written by *** and Eric Martin for COMP9021\n\nfrom datetime import datetime\nimport os\nfrom sys import exit\n\nstartTime = datetime.now()\n# print('Start time ', startTime)\n\nfirst_name = input('Enter a first name: ')\ndirectory = 'names'\nmin_male_frequency = 0\nmale_first_year = None\nmin_female_frequency = 0\nfemale_first_year = None\n\nmy_data_female = {}\nmy_data_male = {}\ntotal_stats = {}\ncurrent_year_stats = {}\n\ncurrent_year_male = {}\ncurrent_year_female = {}\n\n# Replace this comment with your code\nfor filename in os.listdir(directory):\n if not filename.endswith('.txt'):\n continue\n with open(directory + '/' + filename, 'r') as file:\n for line in file:\n L = []\n name, sex, tally = line.split(',')\n year = int(filename[3:7])\n if tally.endswith('\\n'):\n tally = int(tally[0:len(tally)-1])\n else:\n tally = int(tally)\n if sex == 'M' and year not in current_year_male:\n current_year_male[year] = tally\n elif sex == 'M' and year in current_year_male:\n current_year_male[year] += tally\n\n elif sex == 'F' and year not in current_year_female:\n current_year_female[year] = tally\n else:\n current_year_female[year] += tally\n\n if sex == 'F' and name not in my_data_female:\n L.append(year)\n L.append(tally)\n my_data_female[name] = L\n elif sex == 'F' and name in my_data_female:\n my_data_female[name].append(year)\n my_data_female[name].append(tally)\n elif sex == 'M' and name not in my_data_male:\n L.append(year)\n L.append(tally)\n my_data_male[name] = L\n else:\n my_data_male[name].append(year)\n my_data_male[name].append(tally)\n\n\n\n # break\n\n\n# print(my_data_male['Mary'])\n# print(my_data_female['Mary'])\n\ncurrent_high_fre = 0\ntotal_requested_male_count = 0\ntotal_requested_female_count = 0\nyear_to_be_displayed = 0\n\ntotal_male_count = 0\ntotal_female_count = 0\n\ncurrent_temp_female_count = 0\ncurrent_temp_male_count = 0\n\nfor key in current_year_male:\n total_male_count += current_year_male[key]\n\nfor key in current_year_female:\n total_female_count += current_year_female[key]\n\nmy_temp_female = my_data_female.get(first_name, None)\nif not my_temp_female:\n female_first_year = my_temp_female\nelse:\n my_requested_female_list = my_data_female[first_name]\n for index in range(0, len(my_requested_female_list), 2):\n year = my_requested_female_list[index]\n count = my_requested_female_list[index+1]\n fre = count / current_year_female[year]\n total_requested_female_count += count\n if current_high_fre < fre:\n current_high_fre = fre\n year_to_be_displayed = year\n current_temp_female_count = count\n\n female_first_year = year_to_be_displayed\n #sum_t = 0\n #for key in current_year_female:\n # if key <= female_first_year:\n # sum_t += current_year_female[key]\n\n #for index in range(0, len(my_requested_female_list), 2):\n # if my_requested_female_list[index] <= female_first_year:\n # current_temp_female_count += my_requested_female_list[index+1]\n\n min_female_frequency = current_temp_female_count * 100 / current_year_female[female_first_year]\n\ncurrent_high_fre = 0\nmy_temp_male = my_data_male.get(first_name, None)\nif not my_temp_male:\n male_first_year = my_temp_male\nelse:\n my_requested_male_list = my_data_male[first_name]\n # print(my_requested_male_list)\n for index in range(0, len(my_requested_male_list), 2):\n year = my_requested_male_list[index]\n count = my_requested_male_list[index+1]\n fre = count / current_year_male[year]\n total_requested_male_count += count\n if current_high_fre < fre:\n current_high_fre = fre\n year_to_be_displayed = year\n current_temp_male_count = count\n\n male_first_year = year_to_be_displayed\n #sum_t = 0\n #for key in current_year_male:\n # if key <= male_first_year:\n # sum_t += current_year_male[key]\n\n #for index in range(0, len(my_requested_male_list), 2):\n # if my_requested_male_list[index] <= male_first_year:\n # current_temp_male_count += my_requested_male_list[index+1]\n\n min_male_frequency = current_temp_male_count * 100 / current_year_male[male_first_year]\n\n\nif not female_first_year:\n print('In all years, {:} was never given as a female name.'.\n format(first_name))\nelse:\n print('In terms of frequency, {:} was the most popular '\n 'as a female name first in the year {:}.\\n'\n ' It then accounted for {:.2f}% of all female names'.\n format(first_name,\n female_first_year,\n min_female_frequency))\nif not male_first_year:\n print('In all years, {:} was never given as a male name.'\n .format(first_name))\nelse:\n print('In terms of frequency, {:} was the most popular '\n 'as a male name first in the year {:}.\\n'\n ' It then accounted for {:.2f}% of all male names'.\n format(first_name,\n male_first_year,\n min_male_frequency))\n\nendTime = datetime.now()\n\n# print('End time ', endTime)\n# print('Time Taken :: ', (endTime - startTime))\n", "sub_path": "quiz_4.py", "file_name": "quiz_4.py", "file_ext": "py", "file_size_in_byte": 5944, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "datetime.datetime.now", "line_number": 17, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 17, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 176, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 176, "usage_type": "name"}]}
+{"seq_id": "581811678", "text": "'''\nThis script is used to initialize menu.models\nplease run it with \"python manage.py shell\"\n'''\n\nfrom .models import Menu, Button\nfrom os import path\nimport json\n\ndef init():\n menu_file = open(\"%s/%s\"%(path.dirname(__file__), \"menu.json\"), \"r\")\n menu_data = json.load(menu_file)\n\n for i, menu_item in enumerate(menu_data[\"button\"]):\n # method get_or_create returns a tuple (object, created)\n # where created is a boolen\n new_menu = Menu.objects.get_or_create(position=i)[0]\n\n new_menu.name = menu_item[\"name\"] # change anyway\n new_menu.save()\n\n for j, button_item in enumerate(menu_item[\"sub_button\"]):\n new_button = new_menu.button_set.get_or_create(position=j)[0]\n\n for content in button_item.keys():\n if content == \"type\": # type cannot be used as a key while others can\n new_button.act_type = button_item[content]\n else:\n exec(\"new_button.%s = button_item['%s']\"%(content, content))\n\n new_button.up_menu = new_menu\n new_button.save()\n\n # below set a Menu and a Button named \"null\" for unsorted news\n new_menu = Menu.objects.get_or_create(position=9, name=\"null\")[0]\n new_button = new_menu.button_set.get_or_create(position=9, name=\"null\")[0]\n\n menu_file.close()\n\nif __name__==\"__main__\":\n init()\n", "sub_path": "menu/init_menu.py", "file_name": "init_menu.py", "file_ext": "py", "file_size_in_byte": 1380, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.path.dirname", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "name"}, {"api_name": "json.load", "line_number": 12, "usage_type": "call"}, {"api_name": "models.Menu.objects.get_or_create", "line_number": 17, "usage_type": "call"}, {"api_name": "models.Menu.objects", "line_number": 17, "usage_type": "attribute"}, {"api_name": "models.Menu", "line_number": 17, "usage_type": "name"}, {"api_name": "models.Menu.objects.get_or_create", "line_number": 35, "usage_type": "call"}, {"api_name": "models.Menu.objects", "line_number": 35, "usage_type": "attribute"}, {"api_name": "models.Menu", "line_number": 35, "usage_type": "name"}]}
+{"seq_id": "613593550", "text": "from flask import Flask\nfrom flask import request\nfrom pymongo import MongoClient\nimport json\nfrom bson import json_util\nfrom bson import objectid\nimport re\nimport os\n\napp = Flask(__name__)\n#add this so that flask doesn't swallow error messages\napp.config['PROPAGATE_EXCEPTIONS'] = True\n\n#a base urls that returns all the parks in the collection (of course in the future we would implement paging)\n@app.route(\"/ws/users\")\ndef parks():\n #setup the connection\n conn = MongoClient(os.environ['PATB_MONGODB_SERVICE_HOST'],\n int(os.environ['PATB_MONGODB_SERVICE_PORT']))\n db = conn[os.environ['MONGODB_DATABASE']]\n db.authenticate(os.environ['MONGODB_USER'],\n os.environ['MONGODB_PASSWORD'])\n\n #query the DB for all the parkpoints\n result = db.Users.find()\n\n #Now turn the results into valid JSON\n return str(json.dumps({'results':list(result)},default=json_util.default))\n\n\n@app.route('/')\ndef hello_world():\n return 'Hello World1!'\n\n@app.route('/DEBUG/env')\ndef dbg_env():\n env_list = ['%s: %s' % (key, value)\n for key, value in sorted(os.environ.items())]\n return \"env is\\n%s
\" % '\\n'.join(env_list)\n\n\nif __name__ == '__main__':\n app.run()\n", "sub_path": "patb.py", "file_name": "patb.py", "file_ext": "py", "file_size_in_byte": 1239, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "flask.Flask", "line_number": 10, "usage_type": "call"}, {"api_name": "pymongo.MongoClient", "line_number": 18, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 22, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 28, "usage_type": "call"}, {"api_name": "bson.json_util.default", "line_number": 28, "usage_type": "attribute"}, {"api_name": "bson.json_util", "line_number": 28, "usage_type": "name"}, {"api_name": "os.environ.items", "line_number": 38, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 38, "usage_type": "attribute"}]}
+{"seq_id": "187371132", "text": "\"\"\"\nhttp://tempo-db.com/api/read-series/#read-series-by-key\n\"\"\"\nfrom __future__ import print_function\n\nimport datetime\nfrom tempodb import Client\n\n# Modify these with your settings found at: http://tempo-db.com/manage/\nAPI_KEY = 'your-api-key'\nAPI_SECRET = 'your-api-secret'\nSERIES_KEY = 'your-custom-key'\n\nclient = Client(API_KEY, API_SECRET)\n\nstart = datetime.date(2012, 1, 1)\nend = start + datetime.timedelta(days=1)\n\ndata = client.read_key(SERIES_KEY, start, end)\n\nfor datapoint in data.data:\n print(datapoint)\n", "sub_path": "tempodb/demo/tempodb-read-demo.py", "file_name": "tempodb-read-demo.py", "file_ext": "py", "file_size_in_byte": 518, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "tempodb.Client", "line_number": 14, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 17, "usage_type": "call"}]}
+{"seq_id": "37392763", "text": "#!/usr/bin/python2.7\n## This Script is to read CSV File and process the records to be imported into db.\n##\n## Author: Ahmed Sammoud\n## Date: June, 2016\n##\n## Company: Red Hat, Red Hat University , Intern_Team_2016\n##\n## Description :\n## - This Script main purpose is to read in \".csv\" file and extract the needed information for data analysis.\n## - The Script runs in three phases: \n## -- 1- It uses the csv package and start reading Rows from the file in the Starter module.\n## -- 2- It uses the Extract_Info function to extract the needed info from columns in each row.\n## -- 3- The Process_Data function is used to :\n## 1- clean up the data from each columns\n## 2-populate the needed data structures.\n## -- \n## -- Starter --> Extract_Info --> Process_Data \n##\n\n\n\nimport csv\nimport re\nimport logging\nimport pprint\n\n\nclass CSV_Import:\n # A function that selects only the needed Columns and puts it in a new list.\n\n def __init__(self, filename=\"test_data.csv\", perm='rb'):\n\n # Open CSV File\n CSV_F = open(filename, perm)\n\n if perm == 'rb':\n # Starting the reader, One Row at a time. Puts it into a list.\n self.Reader = csv.reader(CSV_F)\n self.Reader.next() # skip header rows\n self.Reader.next() # skip header rows\n else:\n # Starting the Writer\n self.Writer = csv.writer(CSV_F)\n\n # Setting name for logger information\n self.logger = logging.getLogger('___CSV_IMPORT__')\n self.logger.debug(\"CSV_Import intiailized\")\n\n def __Extract_Data__(self, Row):\n Info = {} # list of values from csv\n\n ## Data Format in the Exam .csv file\n ## Entries in order (Col) :: Info (Notes)\n ## - Exam Name (A - 0) :: Strings no Exam code.\n ## - Duration (C - 2) :: In Minutes\n ## - Examinee name (D - 3) :: Whole String (First M Last)\n ## - Examinee email (E - 4) :: email (Tag if redhat email)\n ## - Status of Exam (F - 5) :: Status on how test go (completed,Error,No Show) (Not Scores)\n ## - Country (G - 6) :: Country Name\n ## - Site / Office (H - 7) :: Site Name (Include info such as: Employees only, Retired, City Name, On KoaLA )\n ## - Date of Test (K -10) :: Date (There are 3 Date columns, I choose K since its more consisted: Extract Month,Year)\n ## Data Dictionary Key : entries\n ## Exam : Exam Name\n ## Duration : time In Minutes\n ## Name : Examinee Name\n ## Email : Examinee Email\n ## RedHat : True or False\n ## Status : Completed, System_Error, No Show\n ## Country : Conuntry of Test Location\n ## City : City of Location -- Not Always Available\n ## Site : Office Name\n ## Site_Info : Employees Only/Retired\n ## KoaLa : True/False\n ## Date : {month,day,year}\n\n\n Info[\"Exam\"] = str(Row[0].strip())\n Info[\"Duration\"] = str(Row[2].strip())\n Info[\"Name\"] = str(Row[3].strip())\n Info[\"Email\"] = str(Row[4].strip())\n\n # This needs some review, since Only checks for RedHat emails.\n Info[\"RedHat\"] = False if re.search('[a-zA-Z0-9_.]*@redhat.com', Info[\"Email\"]) == None else True\n Info[\"Status\"] = str(Row[5].strip())\n Info[\"Country\"] = str(Row[6].strip())\n\n if \"employees only\" in str(Row[7].strip().lower()):\n Info[\"Site_Info\"] = \"Employees Only\"\n elif \"retired\" in str(Row[7].strip().lower()):\n Info[\"Site_Info\"] = \"Retired\"\n else:\n Info[\"Site_Info\"] = \"N/A\"\n\n ## There are a lot of cases with this field\n ## This is Might start with:\n ## Employees Only: Red Hat- City -KOALA, Empoyees Only: Red Hat -Office - City - KOALA,or Employees Only: City\n\n Site = str(Row[7].strip())\n Info[\"test\"] = Site\n if Site.startswith(\"Employees Only\".upper()):\n Info[\"Site\"] = \"Red Hat\"\n\n if \"KOALA\" in Site:\n Info[\"City\"] = re.search('(?<=-)[\\s\\w-]*(?=-KOALA)', Site).group(0) if re.search(\n \"(?<=-)[\\s\\w-]*(?=-KOALA)\", Site) is not None else \"N/A\"\n else:\n Info[\"City\"] = re.search('(?<=-)[\\s\\w-]*', Site).group(0)\n\n ## Else examples : Office - City or Office - City - KOALA or Office\n else:\n Info[\"Site\"] = re.search('[\\w|\\D]*(?=-)*', Site).group(0)\n\n if re.search('(?<=-)[\\s\\w]*(?=KOALA)*', Site) != None:\n Info[\"City\"] = re.search('(?<=-)[\\s\\w]*(?=KOALA)*', Site).group(0)\n else:\n Info[\"City\"] = \"N/A\"\n\n # Is This test on a KOALA ?\n Info[\"KoaLA\"] = False if re.search('[a-zA-Z0-9_.-]*KOALA', Site) is None else True\n\n # Extract Date .\n datetmp = re.search('\\d*/\\d*/\\d*', Row[10].strip()).group(0).split(\"/\")\n month = datetmp[0]\n day = datetmp[1]\n year = datetmp[2]\n Info[\"Date\"] = dict(month=month, day=day, year=year)\n\n return Info\n\n def getlist(self):\n\n list = []\n\n for Row in self.Reader:\n if len(Row) > 0:\n Info = self.__Extract_Data__(Row)\n list.append(Info)\n\n return list\n\n def store_Rows(self, Rows):\n for Row in Rows:\n self.Writer.writerow(Row)\n\n def store_Row(self, Row):\n self.Writer.writerow(Row)\n\n\n'''\nT = CSV_Import(\"test.csv\")\nlist = T.getlist()\nP = pprint.PrettyPrinter()\nP.pprint(list)\n'''\n", "sub_path": "csv_import.py", "file_name": "csv_import.py", "file_ext": "py", "file_size_in_byte": 5819, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "csv.reader", "line_number": 39, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 44, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 47, "usage_type": "call"}, {"api_name": "re.search", "line_number": 84, "usage_type": "call"}, {"api_name": "re.search", "line_number": 105, "usage_type": "call"}, {"api_name": "re.search", "line_number": 108, "usage_type": "call"}, {"api_name": "re.search", "line_number": 112, "usage_type": "call"}, {"api_name": "re.search", "line_number": 114, "usage_type": "call"}, {"api_name": "re.search", "line_number": 115, "usage_type": "call"}, {"api_name": "re.search", "line_number": 120, "usage_type": "call"}, {"api_name": "re.search", "line_number": 123, "usage_type": "call"}]}
+{"seq_id": "573602444", "text": "#!/usr/bin/env python3\n#coding: utf-8\n\nimport sys\n\nfrom PyQt5 import QtWidgets\nfrom classes import table_widget\nfrom libs import ui_utils\nfrom libs import string_utils\nfrom libs import nhi_utils\nfrom libs import personnel_utils\nfrom libs import case_utils\nfrom libs import system_utils\nfrom libs import cshis_utils\nfrom libs import number_utils\n\n\n# 病歷資料 2018.01.31\nclass MedicalRecordRegistration(QtWidgets.QMainWindow):\n # 初始化\n def __init__(self, parent=None, *args):\n super(MedicalRecordRegistration, self).__init__(parent)\n self.parent = parent\n self.database = args[0]\n self.system_settings = args[1]\n self.case_key = args[2]\n self.call_from = args[3]\n self.medical_record = None\n self.patient_data = None\n self.ui = None\n self.data_changed = False\n self.cshis_data_changed = False\n\n self._set_ui()\n self._read_case_registration()\n self._set_signal() # 先讀完資料才設定信號\n\n if self.call_from == '新增自費病歷':\n self._set_new_self_medical_record()\n self.case_key = -1\n\n self.user_name = self.system_settings.field('使用者')\n self._set_permission()\n\n # 解構\n def __del__(self):\n self.close_all()\n\n # 關閉\n def close_all(self):\n pass\n\n # 設定GUI\n def _set_ui(self):\n self.ui = ui_utils.load_ui_file(ui_utils.UI_MEDICAL_RECORD_REGISTRATION, self)\n system_utils.set_css(self, self.system_settings)\n system_utils.center_window(self)\n self.table_widget_prescript_sign = table_widget.TableWidget(\n self.ui.tableWidget_prescript_sign, self.database)\n self._set_combo_box()\n self._set_table_width()\n\n # 設定信號\n def _set_signal(self):\n self.ui.lineEdit_case_date.textChanged.connect(self._set_data_changed)\n self.ui.comboBox_period.currentTextChanged.connect(self._set_data_changed)\n self.ui.lineEdit_completion_time.textChanged.connect(self._set_data_changed)\n self.ui.lineEdit_charge_time.textChanged.connect(self._set_data_changed)\n self.ui.comboBox_charge_period.currentTextChanged.connect(self._set_data_changed)\n self.ui.comboBox_visit.currentTextChanged.connect(self._set_data_changed)\n self.ui.lineEdit_patient_key.textChanged.connect(self._set_data_changed)\n self.ui.lineEdit_name.textChanged.connect(self._set_data_changed)\n self.ui.comboBox_ins_type.currentTextChanged.connect(self._set_data_changed)\n self.ui.comboBox_reg_type.currentTextChanged.connect(self._set_data_changed)\n self.ui.comboBox_room.currentTextChanged.connect(self._set_data_changed)\n self.ui.lineEdit_regist_no.textChanged.connect(self._set_data_changed)\n\n self.ui.comboBox_registrar.currentTextChanged.connect(self._set_data_changed)\n self.ui.comboBox_cashier.currentTextChanged.connect(self._set_data_changed)\n self.ui.comboBox_doctor.currentTextChanged.connect(self._set_data_changed)\n self.ui.comboBox_pharmacist.currentTextChanged.connect(self._set_data_changed)\n self.ui.comboBox_massager.currentTextChanged.connect(self._set_data_changed)\n\n self.ui.comboBox_apply_type.currentTextChanged.connect(self._set_data_changed)\n self.ui.comboBox_pharmacy_type.currentTextChanged.connect(self._set_data_changed)\n self.ui.comboBox_share_type.currentTextChanged.connect(self._set_data_changed)\n self.ui.comboBox_treat_type.currentTextChanged.connect(self._set_data_changed)\n self.ui.comboBox_injury_type.currentTextChanged.connect(self._set_data_changed)\n self.ui.comboBox_xcard.currentTextChanged.connect(self._set_data_changed)\n self.ui.comboBox_card.currentTextChanged.connect(self._set_data_changed)\n self.ui.comboBox_course.currentTextChanged.connect(self._set_data_changed)\n self.ui.comboBox_tour_area.currentTextChanged.connect(self._set_data_changed)\n\n self.ui.comboBox_upload_type.currentTextChanged.connect(self._cshis_data_changed)\n self.ui.comboBox_treat_after_check.currentTextChanged.connect(self._cshis_data_changed)\n\n self.ui.lineEdit_special_code.textChanged.connect(self.set_special_code)\n self.ui.checkBox_designated_doctor.clicked.connect(self._set_data_changed)\n self.ui.checkBox_designated_massager.clicked.connect(self._set_data_changed)\n\n def _set_permission(self):\n if self.call_from == '醫師看診作業':\n return\n\n if self.user_name == '超級使用者':\n return\n\n if personnel_utils.get_permission(self.database, '病歷資料', '病歷修正', self.user_name) == 'Y':\n return\n\n self.ui.groupBox_registration.setEnabled(False)\n self.ui.groupBox_ins_apply.setEnabled(False)\n self.ui.groupBox_ic_card.setEnabled(False)\n\n def _set_table_width(self):\n width = [200, 90, 430]\n self.table_widget_prescript_sign.set_table_heading_width(width)\n\n # 檢查資料是否異動\n def _set_data_changed(self):\n self.data_changed = True\n sender_name = self.sender().objectName()\n\n if sender_name == 'comboBox_share_type':\n if self.ui.comboBox_share_type.currentText() == '職業傷害':\n if self.ui.comboBox_injury_type.currentText() not in ['職業傷害', '職業病']:\n self.ui.comboBox_injury_type.setCurrentText('職業傷害')\n card = string_utils.xstr(self.ui.comboBox_card.currentText()).split(' ')[0]\n if card != 'IC06':\n self.ui.comboBox_card.setCurrentText(nhi_utils.INJURY_CARD_DICT['IC06'])\n else:\n if self.ui.comboBox_injury_type.currentText() != '普通疾病':\n self.ui.comboBox_injury_type.setCurrentText('普通疾病')\n\n elif sender_name == 'comboBox_injury_type':\n if self.ui.comboBox_injury_type.currentText() in ['職業傷害', '職業病']:\n if self.ui.comboBox_share_type.currentText() != '職業傷害':\n self.ui.comboBox_share_type.setCurrentText('職業傷害')\n card = string_utils.xstr(self.ui.comboBox_card.currentText()).split(' ')[0]\n if card != 'IC06':\n self.ui.comboBox_card.setCurrentText(nhi_utils.INJURY_CARD_DICT['IC06'])\n elif sender_name == 'comboBox_card':\n card = string_utils.xstr(self.ui.comboBox_card.currentText()).split(' ')[0]\n if card == 'IC06':\n if self.ui.comboBox_injury_type.currentText() not in ['職業傷害', '職業病']:\n self.ui.comboBox_injury_type.setCurrentText('職業傷害')\n elif sender_name == 'comboBox_pharmacy_type':\n ins_prescript = self.parent.tab_list[0]\n if ins_prescript is None:\n return\n\n pharmacy_type = self.ui.comboBox_pharmacy_type.currentText()\n combo_box_pharmacy = ins_prescript.checkBox_pharmacy\n if pharmacy_type == '申報' and not combo_box_pharmacy.isChecked():\n combo_box_pharmacy.setChecked(True)\n elif pharmacy_type == '不申報' and combo_box_pharmacy.isChecked():\n combo_box_pharmacy.setChecked(False)\n\n self.parent.calculate_ins_fees()\n\n def _set_combo_box(self):\n ui_utils.set_combo_box(self.ui.comboBox_period, nhi_utils.PERIOD, None)\n ui_utils.set_combo_box(self.ui.comboBox_charge_period, nhi_utils.PERIOD, None)\n ui_utils.set_combo_box(self.ui.comboBox_visit, nhi_utils.VISIT, None)\n ui_utils.set_combo_box(self.ui.comboBox_ins_type, nhi_utils.INS_TYPE, None)\n ui_utils.set_combo_box(self.ui.comboBox_reg_type, nhi_utils.REG_TYPE, None)\n ui_utils.set_combo_box(self.ui.comboBox_room, nhi_utils.ROOM)\n ui_utils.set_combo_box(\n self.ui.comboBox_tour_area,\n list(nhi_utils.TOUR_AREA_DICT.keys()),\n None\n )\n ui_utils.set_combo_box(\n self.ui.comboBox_registrar,\n personnel_utils.get_personnel(self.database, '全部'), None,\n )\n ui_utils.set_combo_box(\n self.ui.comboBox_cashier,\n personnel_utils.get_personnel(self.database, '全部'), None,\n )\n\n ui_utils.set_combo_box(\n self.ui.comboBox_doctor,\n personnel_utils.get_personnel(self.database, '全部醫師'), None,\n )\n ui_utils.set_combo_box(\n self.ui.comboBox_pharmacist,\n personnel_utils.get_personnel(self.database, '藥師'), None,\n )\n ui_utils.set_combo_box(\n self.ui.comboBox_massager,\n personnel_utils.get_personnel(self.database, '推拿師父'), None,\n )\n ui_utils.set_combo_box(self.ui.comboBox_apply_type, nhi_utils.APPLY_TYPE, None)\n ui_utils.set_combo_box(self.ui.comboBox_pharmacy_type, nhi_utils.PHARMACY_APPLY_TYPE, None)\n ui_utils.set_combo_box(self.ui.comboBox_share_type, nhi_utils.SHARE_TYPE, None)\n ui_utils.set_combo_box(self.ui.comboBox_treat_type, nhi_utils.TREAT_TYPE, None)\n ui_utils.set_combo_box(self.ui.comboBox_injury_type, nhi_utils.INJURY_TYPE, None)\n ui_utils.set_combo_box(self.ui.comboBox_xcard, nhi_utils.ABNORMAL_CARD_WITH_HINT, None)\n ui_utils.set_combo_box(self.ui.comboBox_card, nhi_utils.CARD, None, '欠卡')\n ui_utils.set_combo_box(self.ui.comboBox_course, nhi_utils.COURSE, None)\n\n ui_utils.set_combo_box(self.ui.comboBox_upload_type, nhi_utils.UPLOAD_TYPE, None)\n ui_utils.set_combo_box(self.ui.comboBox_treat_after_check, nhi_utils.TREAT_AFTER_CHECK, None)\n\n def set_special_code(self):\n self.data_changed = True\n\n if self.ui.lineEdit_special_code.text() != '':\n self.parent.ui.lineEdit_disease_code1.setStyleSheet('color:red')\n self.parent.ui.lineEdit_disease_name1.setStyleSheet('color:red')\n else:\n self.parent.ui.lineEdit_disease_code1.setStyleSheet('color:black')\n self.parent.ui.lineEdit_disease_name1.setStyleSheet('color:black')\n\n def _read_case_registration(self):\n sql = '''\n SELECT * FROM cases WHERE \n CaseKey = {0}\n '''.format(self.case_key)\n row = self.database.select_record(sql)[0]\n self._set_registration_data(row)\n self._set_personnel(row)\n self._set_ic_card_data(row)\n self._set_ins_data(row)\n self._set_prescript_sign(row)\n\n def _set_registration_data(self, row):\n self.ui.lineEdit_case_date.setText(string_utils.xstr(row['CaseDate']))\n self.ui.comboBox_period.setCurrentText(string_utils.xstr(row['Period']))\n self.ui.lineEdit_completion_time.setText(string_utils.xstr(row['DoctorDate']))\n self.ui.lineEdit_charge_time.setText(string_utils.xstr(row['ChargeDate']))\n self.ui.comboBox_charge_period.setCurrentText(string_utils.xstr(row['ChargePeriod']))\n self.ui.comboBox_visit.setCurrentText(string_utils.xstr(row['Visit']))\n self.ui.lineEdit_patient_key.setText(string_utils.xstr(row['PatientKey']))\n self.ui.lineEdit_name.setText(string_utils.xstr(row['Name']))\n self.ui.comboBox_ins_type.setCurrentText(string_utils.xstr(row['InsType']))\n self.ui.comboBox_reg_type.setCurrentText(string_utils.xstr(row['RegistType']))\n self.ui.comboBox_room.setCurrentText(string_utils.xstr(row['Room']))\n self.ui.lineEdit_regist_no.setText(string_utils.xstr(row['RegistNo']))\n self.ui.comboBox_tour_area.setCurrentText(string_utils.xstr(row['TourArea']))\n\n def _set_personnel(self, row):\n system_utils.set_combo_box_item(self.ui.comboBox_registrar, string_utils.xstr(row['Register']))\n system_utils.set_combo_box_item(self.ui.comboBox_cashier, string_utils.xstr(row['Cashier']))\n system_utils.set_combo_box_item(self.ui.comboBox_pharmacist, string_utils.xstr(row['Pharmacist']))\n system_utils.set_combo_box_item(self.ui.comboBox_massager, string_utils.xstr(row['Massager']))\n system_utils.set_combo_box_item(self.ui.comboBox_doctor, string_utils.xstr(row['Doctor']))\n\n if row['DesignatedDoctor'] == 'True':\n self.ui.checkBox_designated_doctor.setChecked(True)\n if row['DesignatedMassager'] == 'True':\n self.ui.checkBox_designated_massager.setChecked(True)\n\n def _set_ic_card_data(self, row):\n card_datetime = case_utils.extract_security_xml(row['Security'], '寫卡時間')\n seq_number = case_utils.extract_security_xml(row['Security'], '健保卡序')\n clinic_id = case_utils.extract_security_xml(row['Security'], '院所代號')\n sam_id = case_utils.extract_security_xml(row['Security'], '安全模組')\n signature = case_utils.extract_security_xml(row['Security'], '安全簽章')\n upload_time = case_utils.extract_security_xml(row['Security'], '上傳時間')\n upload_type = case_utils.extract_security_xml(row['Security'], '資料格式')\n treat_after_check = case_utils.extract_security_xml(row['Security'], '補卡註記')\n prescript_sign_time = case_utils.extract_security_xml(row['Security'], '醫令時間')\n\n self.ui.lineEdit_ic_registration.setText(card_datetime)\n self.ui.lineEdit_seq_number.setText(seq_number)\n self.ui.lineEdit_clinic_id.setText(clinic_id)\n self.ui.lineEdit_sam_id.setText(sam_id)\n self.ui.lineEdit_upload_time.setText(upload_time)\n\n self.ui.comboBox_upload_type.setCurrentText(cshis_utils.UPLOAD_TYPE_DICT[upload_type])\n self.ui.comboBox_treat_after_check.setCurrentText(cshis_utils.TREAT_AFTER_CHECK_DICT[treat_after_check])\n\n self.ui.lineEdit_prescript_sign_time.setText(prescript_sign_time)\n self.ui.textEdit_signature.setPlainText(signature)\n\n def _set_ins_data(self, row):\n self.ui.comboBox_apply_type.setCurrentText(string_utils.xstr(row['ApplyType']))\n self.ui.comboBox_pharmacy_type.setCurrentText(string_utils.xstr(row['PharmacyType']))\n self.ui.comboBox_share_type.setCurrentText(string_utils.xstr(row['Share']))\n self.ui.comboBox_treat_type.setCurrentText(string_utils.xstr(row['TreatType']))\n self.ui.comboBox_injury_type.setCurrentText(string_utils.xstr(row['Injury']))\n\n xcard = string_utils.xstr(row['XCard'])\n if xcard in nhi_utils.ABNORMAL_CARD:\n xcard = nhi_utils.ABNORMAL_CARD_DICT[xcard]\n\n self.ui.comboBox_xcard.setCurrentText(xcard)\n\n card = string_utils.xstr(row['Card'])\n if card in nhi_utils.ABNORMAL_CARD:\n card = nhi_utils.ABNORMAL_CARD_DICT[card]\n\n if card not in nhi_utils.ABNORMAL_CARD_WITH_HINT + nhi_utils.CARD:\n self.ui.comboBox_card.insertItem(1, card)\n self.ui.comboBox_card.setCurrentText(card)\n\n self.ui.comboBox_course.setCurrentText(string_utils.xstr(row['Continuance']))\n self.ui.lineEdit_special_code.setText(string_utils.xstr(row['SpecialCode']))\n\n self.ui.lineEdit_ins_total_fee.setText(string_utils.xstr(number_utils.get_integer(row['InsTotalFee'])))\n # database.ui.lineEdit_share_fee.setText(\n # string_utils.xstr(\n # number_utils.get_integer(medical_row['DiagShareFee']) +\n # number_utils.get_integer(medical_row['DrugShareFee'])\n # )\n # )\n # database.ui.lineEdit_ins_apply_fee.setText(string_utils.xstr(number_utils.get_integer(medical_row['InsApplyFee'])))\n\n def _set_treat_sign(self):\n sql = '''\n SELECT * FROM presextend WHERE\n PrescriptKey = {0} AND ExtendType = \"處置簽章\"\n '''.format(self.case_key)\n self.table_widget_prescript_sign.set_db_data(sql, self._set_treat_sign_data)\n\n def _set_treat_sign_data(self, rec_no, rec):\n sql = '''\n SELECT Treatment FROM cases WHERE\n CaseKey = {0}\n '''.format(self.case_key)\n row = self.database.select_record(sql)[0]\n treatment = string_utils.xstr(row['Treatment'])\n ins_code = nhi_utils.get_treat_code(self.database, self.case_key)\n prescript_sign_rec = [\n treatment,\n ins_code,\n string_utils.xstr(rec['Content']),\n ]\n\n for column in range(len(prescript_sign_rec)):\n self.ui.tableWidget_prescript_sign.setItem(\n rec_no, column, QtWidgets.QTableWidgetItem(prescript_sign_rec[column])\n )\n\n def _set_prescript_sign(self, row):\n start_index = 0\n\n if string_utils.xstr(row['Treatment']) != '':\n self._set_treat_sign()\n start_index = 1\n\n sql = '''\n SELECT \n prescript.PrescriptKey, prescript.MedicineName, prescript.InsCode, \n presextend.Content FROM prescript\n LEFT JOIN presextend ON presextend.PrescriptKey = prescript.PrescriptKey \n WHERE\n prescript.CaseKey = {0} AND\n prescript.MedicineSet = 1 AND prescript.InsCode IS NOT NULL AND\n presextend.Content IS NOT NULL\n ORDER BY prescript.PrescriptNo, prescript.PrescriptKey\n '''.format(self.case_key)\n self.table_widget_prescript_sign.set_db_data(sql, self._set_prescript_sign_data, None, start_index)\n\n def _set_prescript_sign_data(self, row_no, row):\n prescript_sign_rec = [\n string_utils.xstr(row['MedicineName']),\n string_utils.xstr(row['InsCode']),\n string_utils.xstr(row['Content']),\n ]\n\n for column in range(len(prescript_sign_rec)):\n self.ui.tableWidget_prescript_sign.setItem(\n row_no, column,\n QtWidgets.QTableWidgetItem(prescript_sign_rec[column])\n )\n\n def save_record(self):\n if not self.data_changed:\n return\n\n fields = [\n 'CaseDate', 'Period', 'DoctorDate', 'ChargeDate', 'ChargePeriod',\n 'Visit', 'PatientKey', 'Name',\n 'InsType', 'RegistType', 'TourArea', 'Room', 'RegistNo',\n 'Register', 'Cashier', 'Doctor', 'Pharmacist', 'Massager',\n 'ApplyType', 'PharmacyType', 'Share', 'TreatType', 'Injury',\n 'XCard', 'Card', 'Continuance', 'SpecialCode',\n 'DesignatedDoctor', 'DesignatedMassager',\n ]\n xcard = string_utils.xstr(self.ui.comboBox_xcard.currentText()).split(' ')[0]\n card = string_utils.xstr(self.ui.comboBox_card.currentText()).split(' ')[0]\n\n massager = self.ui.comboBox_massager.currentText()\n designated_doctor = 'False'\n designated_massager = 'False'\n\n if self.ui.checkBox_designated_doctor.isChecked():\n designated_doctor = 'True'\n if massager != '' and self.ui.checkBox_designated_massager.isChecked():\n designated_massager = 'True'\n\n\n data = [\n self.ui.lineEdit_case_date.text(),\n self.ui.comboBox_period.currentText(),\n self.ui.lineEdit_completion_time.text(),\n self.ui.lineEdit_charge_time.text(),\n self.ui.comboBox_charge_period.currentText(),\n self.ui.comboBox_visit.currentText(),\n self.ui.lineEdit_patient_key.text(),\n self.ui.lineEdit_name.text(),\n self.ui.comboBox_ins_type.currentText(),\n self.ui.comboBox_reg_type.currentText(),\n self.ui.comboBox_tour_area.currentText(),\n self.ui.comboBox_room.currentText(),\n self.ui.lineEdit_regist_no.text(),\n\n self.ui.comboBox_registrar.currentText(),\n self.ui.comboBox_cashier.currentText(),\n self.ui.comboBox_doctor.currentText(),\n self.ui.comboBox_pharmacist.currentText(),\n massager,\n\n self.ui.comboBox_apply_type.currentText(),\n self.ui.comboBox_pharmacy_type.currentText(),\n self.ui.comboBox_share_type.currentText(),\n self.ui.comboBox_treat_type.currentText(),\n self.ui.comboBox_injury_type.currentText(),\n xcard,\n card,\n self.ui.comboBox_course.currentText(),\n self.ui.lineEdit_special_code.text(),\n designated_doctor,\n designated_massager,\n ]\n\n self.database.update_record('cases', fields, 'CaseKey', self.case_key, data)\n\n if self.cshis_data_changed:\n upload_type = self.ui.comboBox_upload_type.currentText().split('-')[0]\n treat_after_check = self.ui.comboBox_treat_after_check.currentText().split('-')[0]\n\n case_utils.update_xml(\n self.database, 'cases', 'Security', 'upload_type', upload_type, 'CaseKey', self.case_key,\n ) # 更新健保寫卡資料\n case_utils.update_xml(\n self.database, 'cases', 'Security', 'treat_after_check', treat_after_check, 'CaseKey', self.case_key,\n ) # 更新健保寫卡資料\n\n def _cshis_data_changed(self):\n self.cshis_data_changed = True\n\n def _set_new_self_medical_record(self):\n user_name = self.system_settings.field('使用者')\n\n self.ui.comboBox_ins_type.setCurrentText('自費')\n self.ui.lineEdit_completion_time.setText('')\n self.ui.lineEdit_charge_time.setText('')\n\n self.ui.comboBox_registrar.setCurrentText(user_name)\n self.ui.comboBox_doctor.setCurrentText(user_name)\n self.ui.comboBox_cashier.setCurrentText(user_name)\n self.ui.comboBox_charge_period.setCurrentText(None)\n\n self.ui.comboBox_upload_type.setCurrentText('1-正常上傳')\n self.ui.comboBox_treat_after_check.setCurrentText('1-正常')\n self.ui.lineEdit_clinic_id.setText('')\n self.ui.lineEdit_sam_id.setText('')\n self.ui.lineEdit_ic_registration.setText('')\n self.ui.lineEdit_seq_number.setText('')\n self.ui.lineEdit_prescript_sign_time.setText('')\n self.ui.lineEdit_upload_time.setText('')\n self.ui.textEdit_signature.setPlainText('')\n\n self.ui.comboBox_reg_type.setCurrentText(self.system_settings.field('掛號類別'))\n self.ui.comboBox_treat_type.setCurrentText('內科')\n self.ui.comboBox_card.setCurrentText('不需取得')\n self.ui.comboBox_course.setCurrentText(None)\n self.ui.comboBox_xcard.setCurrentText(None)\n self.ui.lineEdit_special_code.setText('')\n self.ui.lineEdit_ins_total_fee.setText('')\n\n self.ui.tableWidget_prescript_sign.setRowCount(0)\n", "sub_path": "medical_record_registration.py", "file_name": "medical_record_registration.py", "file_ext": "py", "file_size_in_byte": 22572, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 19, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 19, "usage_type": "name"}, {"api_name": "libs.ui_utils.load_ui_file", "line_number": 55, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 55, "usage_type": "name"}, {"api_name": "libs.ui_utils.UI_MEDICAL_RECORD_REGISTRATION", "line_number": 55, "usage_type": "attribute"}, {"api_name": "libs.system_utils.set_css", "line_number": 56, "usage_type": "call"}, {"api_name": "libs.system_utils", "line_number": 56, "usage_type": "name"}, {"api_name": "libs.system_utils.center_window", "line_number": 57, "usage_type": "call"}, {"api_name": "libs.system_utils", "line_number": 57, "usage_type": "name"}, {"api_name": "classes.table_widget.TableWidget", "line_number": 58, "usage_type": "call"}, {"api_name": "classes.table_widget", "line_number": 58, "usage_type": "name"}, {"api_name": "libs.personnel_utils.get_permission", "line_number": 108, "usage_type": "call"}, {"api_name": "libs.personnel_utils", "line_number": 108, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 128, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 128, "usage_type": "name"}, {"api_name": "libs.nhi_utils.INJURY_CARD_DICT", "line_number": 130, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 130, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 139, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 139, "usage_type": "name"}, {"api_name": "libs.nhi_utils.INJURY_CARD_DICT", "line_number": 141, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 141, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 143, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 143, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 162, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 162, "usage_type": "name"}, {"api_name": "libs.nhi_utils.PERIOD", "line_number": 162, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 162, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 163, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 163, "usage_type": "name"}, {"api_name": "libs.nhi_utils.PERIOD", "line_number": 163, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 163, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 164, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 164, "usage_type": "name"}, {"api_name": "libs.nhi_utils.VISIT", "line_number": 164, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 164, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 165, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 165, "usage_type": "name"}, {"api_name": "libs.nhi_utils.INS_TYPE", "line_number": 165, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 165, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 166, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 166, "usage_type": "name"}, {"api_name": "libs.nhi_utils.REG_TYPE", "line_number": 166, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 166, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 167, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 167, "usage_type": "name"}, {"api_name": "libs.nhi_utils.ROOM", "line_number": 167, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 167, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 168, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 168, "usage_type": "name"}, {"api_name": "libs.nhi_utils.TOUR_AREA_DICT.keys", "line_number": 170, "usage_type": "call"}, {"api_name": "libs.nhi_utils.TOUR_AREA_DICT", "line_number": 170, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 170, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 173, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 173, "usage_type": "name"}, {"api_name": "libs.personnel_utils.get_personnel", "line_number": 175, "usage_type": "call"}, {"api_name": "libs.personnel_utils", "line_number": 175, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 177, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 177, "usage_type": "name"}, {"api_name": "libs.personnel_utils.get_personnel", "line_number": 179, "usage_type": "call"}, {"api_name": "libs.personnel_utils", "line_number": 179, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 182, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 182, "usage_type": "name"}, {"api_name": "libs.personnel_utils.get_personnel", "line_number": 184, "usage_type": "call"}, {"api_name": "libs.personnel_utils", "line_number": 184, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 186, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 186, "usage_type": "name"}, {"api_name": "libs.personnel_utils.get_personnel", "line_number": 188, "usage_type": "call"}, {"api_name": "libs.personnel_utils", "line_number": 188, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 190, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 190, "usage_type": "name"}, {"api_name": "libs.personnel_utils.get_personnel", "line_number": 192, "usage_type": "call"}, {"api_name": "libs.personnel_utils", "line_number": 192, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 194, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 194, "usage_type": "name"}, {"api_name": "libs.nhi_utils.APPLY_TYPE", "line_number": 194, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 194, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 195, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 195, "usage_type": "name"}, {"api_name": "libs.nhi_utils.PHARMACY_APPLY_TYPE", "line_number": 195, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 195, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 196, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 196, "usage_type": "name"}, {"api_name": "libs.nhi_utils.SHARE_TYPE", "line_number": 196, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 196, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 197, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 197, "usage_type": "name"}, {"api_name": "libs.nhi_utils.TREAT_TYPE", "line_number": 197, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 197, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 198, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 198, "usage_type": "name"}, {"api_name": "libs.nhi_utils.INJURY_TYPE", "line_number": 198, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 198, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 199, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 199, "usage_type": "name"}, {"api_name": "libs.nhi_utils.ABNORMAL_CARD_WITH_HINT", "line_number": 199, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 199, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 200, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 200, "usage_type": "name"}, {"api_name": "libs.nhi_utils.CARD", "line_number": 200, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 200, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 201, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 201, "usage_type": "name"}, {"api_name": "libs.nhi_utils.COURSE", "line_number": 201, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 201, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 203, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 203, "usage_type": "name"}, {"api_name": "libs.nhi_utils.UPLOAD_TYPE", "line_number": 203, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 203, "usage_type": "name"}, {"api_name": "libs.ui_utils.set_combo_box", "line_number": 204, "usage_type": "call"}, {"api_name": "libs.ui_utils", "line_number": 204, "usage_type": "name"}, {"api_name": "libs.nhi_utils.TREAT_AFTER_CHECK", "line_number": 204, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 204, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 229, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 229, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 230, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 230, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 231, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 231, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 232, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 232, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 233, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 233, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 234, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 234, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 235, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 235, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 236, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 236, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 237, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 237, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 238, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 238, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 239, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 239, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 240, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 240, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 241, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 241, "usage_type": "name"}, {"api_name": "libs.system_utils.set_combo_box_item", "line_number": 244, "usage_type": "call"}, {"api_name": "libs.system_utils", "line_number": 244, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 244, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 244, "usage_type": "name"}, {"api_name": "libs.system_utils.set_combo_box_item", "line_number": 245, "usage_type": "call"}, {"api_name": "libs.system_utils", "line_number": 245, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 245, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 245, "usage_type": "name"}, {"api_name": "libs.system_utils.set_combo_box_item", "line_number": 246, "usage_type": "call"}, {"api_name": "libs.system_utils", "line_number": 246, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 246, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 246, "usage_type": "name"}, {"api_name": "libs.system_utils.set_combo_box_item", "line_number": 247, "usage_type": "call"}, {"api_name": "libs.system_utils", "line_number": 247, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 247, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 247, "usage_type": "name"}, {"api_name": "libs.system_utils.set_combo_box_item", "line_number": 248, "usage_type": "call"}, {"api_name": "libs.system_utils", "line_number": 248, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 248, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 248, "usage_type": "name"}, {"api_name": "libs.case_utils.extract_security_xml", "line_number": 256, "usage_type": "call"}, {"api_name": "libs.case_utils", "line_number": 256, "usage_type": "name"}, {"api_name": "libs.case_utils.extract_security_xml", "line_number": 257, "usage_type": "call"}, {"api_name": "libs.case_utils", "line_number": 257, "usage_type": "name"}, {"api_name": "libs.case_utils.extract_security_xml", "line_number": 258, "usage_type": "call"}, {"api_name": "libs.case_utils", "line_number": 258, "usage_type": "name"}, {"api_name": "libs.case_utils.extract_security_xml", "line_number": 259, "usage_type": "call"}, {"api_name": "libs.case_utils", "line_number": 259, "usage_type": "name"}, {"api_name": "libs.case_utils.extract_security_xml", "line_number": 260, "usage_type": "call"}, {"api_name": "libs.case_utils", "line_number": 260, "usage_type": "name"}, {"api_name": "libs.case_utils.extract_security_xml", "line_number": 261, "usage_type": "call"}, {"api_name": "libs.case_utils", "line_number": 261, "usage_type": "name"}, {"api_name": "libs.case_utils.extract_security_xml", "line_number": 262, "usage_type": "call"}, {"api_name": "libs.case_utils", "line_number": 262, "usage_type": "name"}, {"api_name": "libs.case_utils.extract_security_xml", "line_number": 263, "usage_type": "call"}, {"api_name": "libs.case_utils", "line_number": 263, "usage_type": "name"}, {"api_name": "libs.case_utils.extract_security_xml", "line_number": 264, "usage_type": "call"}, {"api_name": "libs.case_utils", "line_number": 264, "usage_type": "name"}, {"api_name": "libs.cshis_utils.UPLOAD_TYPE_DICT", "line_number": 272, "usage_type": "attribute"}, {"api_name": "libs.cshis_utils", "line_number": 272, "usage_type": "name"}, {"api_name": "libs.cshis_utils.TREAT_AFTER_CHECK_DICT", "line_number": 273, "usage_type": "attribute"}, {"api_name": "libs.cshis_utils", "line_number": 273, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 279, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 279, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 280, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 280, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 281, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 281, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 282, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 282, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 283, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 283, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 285, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 285, "usage_type": "name"}, {"api_name": "libs.nhi_utils.ABNORMAL_CARD", "line_number": 286, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 286, "usage_type": "name"}, {"api_name": "libs.nhi_utils.ABNORMAL_CARD_DICT", "line_number": 287, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 287, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 291, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 291, "usage_type": "name"}, {"api_name": "libs.nhi_utils.ABNORMAL_CARD", "line_number": 292, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 292, "usage_type": "name"}, {"api_name": "libs.nhi_utils.ABNORMAL_CARD_DICT", "line_number": 293, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 293, "usage_type": "name"}, {"api_name": "libs.nhi_utils.ABNORMAL_CARD_WITH_HINT", "line_number": 295, "usage_type": "attribute"}, {"api_name": "libs.nhi_utils", "line_number": 295, "usage_type": "name"}, {"api_name": "libs.nhi_utils.CARD", "line_number": 295, "usage_type": "attribute"}, {"api_name": "libs.string_utils.xstr", "line_number": 299, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 299, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 300, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 300, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 302, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 302, "usage_type": "name"}, {"api_name": "libs.number_utils.get_integer", "line_number": 302, "usage_type": "call"}, {"api_name": "libs.number_utils", "line_number": 302, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 324, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 324, "usage_type": "name"}, {"api_name": "libs.nhi_utils.get_treat_code", "line_number": 325, "usage_type": "call"}, {"api_name": "libs.nhi_utils", "line_number": 325, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 329, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 329, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QTableWidgetItem", "line_number": 334, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 334, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 340, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 340, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 359, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 359, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 360, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 360, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 361, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 361, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QTableWidgetItem", "line_number": 367, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 367, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 383, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 383, "usage_type": "name"}, {"api_name": "libs.string_utils.xstr", "line_number": 384, "usage_type": "call"}, {"api_name": "libs.string_utils", "line_number": 384, "usage_type": "name"}, {"api_name": "libs.case_utils.update_xml", "line_number": 436, "usage_type": "call"}, {"api_name": "libs.case_utils", "line_number": 436, "usage_type": "name"}, {"api_name": "libs.case_utils.update_xml", "line_number": 439, "usage_type": "call"}, {"api_name": "libs.case_utils", "line_number": 439, "usage_type": "name"}]}
+{"seq_id": "469096002", "text": "\nfrom django.views.generic import View\n\nfrom modules.response import Response\nfrom .task_model import TaskModel\n\nimport json\nimport sys\nsys.path.append(\"../../modules\")\n\nclass Task(View, Response):\n\n def __init__(self):\n\n super().__init__()\n self.model = TaskModel()\n \n def get(self, request):\n\n uuid = request.GET.get(\"uuid\")\n print(uuid)\n task_list = self.model.fetch_research_tasks(uuid)\n return self.response(task_list)\n\n\n def post(self):\n return self.response({})\n\n\nclass CompleteTask(View, Response):\n\n def __init__(self):\n super().__init__()\n self.model = TaskModel()\n\n\n def post(self, request):\n\n research_id = request.POST.get(\"researchId\")\n task_uid = request.POST.get(\"taskUid\")\n complete = True if request.POST.get(\"complete\") == 'true' else False\n\n result = self.model.complete_task(research_id, task_uid, complete)\n\n return self.response({\"result\": result})\n\nclass UpdateTask(View, Response):\n\n def __init__(self):\n super().__init__()\n self.model = TaskModel()\n\n \n def post(self, request):\n\n research_id = request.POST.get(\"researchId\")\n task_uid = request.POST.get(\"taskUid\")\n actual = json.loads(request.POST.get(\"actual\"))\n progress = int(request.POST.get(\"progress\"))\n\n result = self.model.update_task(research_id, task_uid, actual, progress)\n return self.response({\"result\": result})", "sub_path": "content/django/research_management_api/task/controllers/task.py", "file_name": "task.py", "file_ext": "py", "file_size_in_byte": 1483, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sys.path.append", "line_number": 9, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "django.views.generic.View", "line_number": 11, "usage_type": "name"}, {"api_name": "modules.response.Response", "line_number": 11, "usage_type": "name"}, {"api_name": "task_model.TaskModel", "line_number": 16, "usage_type": "call"}, {"api_name": "django.views.generic.View", "line_number": 30, "usage_type": "name"}, {"api_name": "modules.response.Response", "line_number": 30, "usage_type": "name"}, {"api_name": "task_model.TaskModel", "line_number": 34, "usage_type": "call"}, {"api_name": "django.views.generic.View", "line_number": 47, "usage_type": "name"}, {"api_name": "modules.response.Response", "line_number": 47, "usage_type": "name"}, {"api_name": "task_model.TaskModel", "line_number": 51, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 58, "usage_type": "call"}]}
+{"seq_id": "451827291", "text": "from argparse import ArgumentParser\nfrom pathlib import Path\nfrom statistics import mean, median\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport cv2\n\nfrom modules.scenedetect_gui.frame_timecode import FrameTimecode\nfrom videotools.Film import Film\n\n\ndef read_timecode(timecode):\n temp = timecode.split(':')\n seconds = float(temp[2])\n minutes = float(temp[1])\n hours = float(temp[0])\n if minutes > 0:\n if hours > 0:\n m = (hours * minutes) + minutes\n time = (m * 60) + seconds\n else:\n m = minutes\n time = (m * 60) + seconds\n else:\n time = seconds\n return time\n\n\n# Accepts list of strings containing either begin and end frames or begin and end timecodes in format \"[00:00:00, 00:00:01]\"\ndef calculate_durations(data, frame=True, fps=24):\n y = []\n if frame:\n [y.append((d[1].frame - d[0].frame)/fps) for d in data]\n else:\n for line in data:\n if isinstance(line, str):\n temp = line.split(',')\n begin = read_timecode(temp[0])\n end = read_timecode(temp[1])\n elif isinstance(line, list):\n begin = read_timecode(line[0])\n end = read_timecode(line[1])\n y.append(end - begin)\n return y\n\n\ndef create_shot_numbers(number):\n x = []\n [x.append(num + 1) for num in range(0, number)]\n return x\n\n\ndef load_data(data, frame=True):\n x, y = None, Film()\n if isinstance(data, str):\n with open(str(data), 'r') as d:\n loaded_data = d.readlines()\n y = calculate_durations(loaded_data, frame=frame)\n x = create_shot_numbers(len(loaded_data))\n elif isinstance(data, list):\n if isinstance(data[0], list) or isinstance(data[0], tuple):\n if len(data[0]) == 2:\n if isinstance(data[0][0], FrameTimecode):\n fps = data[0][0].get_framerate()\n new_data = []\n for line in data:\n temp = [str(line[0].get_timecode()), str(line[1].get_timecode())]\n new_data.append(temp)\n data = new_data\n y.shot_lengths = calculate_durations(data, frame=False, fps=fps)\n elif isinstance(data[0][0], str):\n y.shot_lengths = calculate_durations(data, frame=False)\n\n else:\n y.shot_lengths = data\n x = create_shot_numbers(len(data))\n if y:\n x = create_shot_numbers(len(y.shot_lengths))\n y.average_shot_length = mean(y.shot_lengths)\n y.median_shot_length = median(y.shot_lengths)\n return x, y\n\n\ndef visualize(data, frame=True, show=True, save=False):\n x, y = load_data(data, frame)\n\n # Produce shot length graph\n plt.plot(x, y.shot_lengths)\n plt.xlabel('Shot Numbers')\n plt.ylabel('Duration in seconds')\n plt.title('Shot lengths')\n plt.ylim(min(y.shot_lengths), max(y.shot_lengths))\n if show:\n plt.show()\n root_path = get_root_path(__file__)\n if save:\n save_path = Path(root_path).joinpath('{}/data/qt/graph.png'.format(str(root_path)))\n plt.savefig(str(save_path), bbox_inches='tight')\n plt.cla()\n\n # Produce histogram of shot lengths\n hist = np.histogram(y.shot_lengths)\n n, bins, patches = plt.hist(hist)\n plt.xlabel('Value')\n plt.ylabel('Frequency')\n plt.title('Shot Length Histogram')\n maxfreqs = [max(x) for x in n]\n maxfreq = max(maxfreqs)\n plt.ylim(ymax=np.ceil(maxfreq / 10) * 10 if maxfreq % 10 else maxfreq + 10)\n if show:\n plt.show()\n if save:\n save_path = Path(root_path).joinpath('{}/data/qt/hist.png'.format(str(root_path)))\n plt.savefig(str(save_path), bbox_inches='tight')\n\n return y\n\n\ndef debug():\n new_data = []\n root = Path(__file__).parent\n path = Path(root).joinpath('data/test_files/results.csv')\n if not path.parent.exists():\n Path.mkdir(path.parent, parents=True)\n visualize(str(path), frame=False)\n\n\nif __name__ == '__main__':\n ap = ArgumentParser()\n ap.add_argument('--video', '-v')\n ap.add_argument('--data', '-d')\n args = vars(ap.parse_args())\n if not args['video'] and not args['data']:\n print('Either field \"video\" or field \"data\" must contain a value.')\n elif args['video'] and not args['data']:\n print('Attempting to load shot information from video. This may take a while.')\n try:\n cap = cv2.VideoCapture(str(args['video']))\n except:\n print('Unable to open video.')\n exit()\n data = detect_shots(args['video'], verbose=True, write_to_csv=True,\n outpath='/home/amos/MyPrograms/DissertationWork/data/test_files/results.csv')\n visualize(data)\n elif args['data'] and not args['video']:\n visualize(args['data'])\n else:\n print('Please enter either \"video\" or \"data\", not both. Using \"data\" path.')\n try:\n visualize(args['data'])\n except:\n print('Unable to load data from file. Now loading video entered in field \"video.\" This may take a while.')\n try:\n data = detect_shots(args['video'])\n visualize(data)\n except:\n print('Also unable to analyze video. Please check both paths to ensure their accuracy.')\n exit()\n", "sub_path": "owl/visualize.py", "file_name": "visualize.py", "file_ext": "py", "file_size_in_byte": 5415, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "videotools.Film.Film", "line_number": 55, "usage_type": "call"}, {"api_name": "modules.scenedetect_gui.frame_timecode.FrameTimecode", "line_number": 64, "usage_type": "argument"}, {"api_name": "statistics.mean", "line_number": 80, "usage_type": "call"}, {"api_name": "statistics.median", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cla", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "numpy.histogram", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 106, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 107, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 107, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 110, "usage_type": "name"}, {"api_name": "numpy.ceil", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 122, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 123, "usage_type": "call"}, {"api_name": "pathlib.Path.mkdir", "line_number": 125, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 125, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 130, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 139, "usage_type": "call"}]}
+{"seq_id": "66588432", "text": "#!/usr/bin/env python\n# encoding: utf-8\n\nimport os\nimport cv2\nimport numpy as np\nimport pandas as pd\nimport matplotlib.image as mpimg\n\n\nCORRECTION = 0.2\nIMAGE_HEIGHT, IMAGE_WIDTH, IMAGE_CHANNELS = 66, 200, 3\nINPUT_SHAPE = (IMAGE_HEIGHT, IMAGE_WIDTH, IMAGE_CHANNELS)\n\n\ndef load_csv(data_dir):\n \"\"\"\n Load training data\n \"\"\"\n data_df = pd.read_csv(os.path.join(data_dir, 'driving_log.csv'))\n X = data_df[['center', 'left', 'right']].values\n y = data_df['steering'].values\n return X, y\n\n\ndef crop(image):\n \"\"\"\n Crop the image (removing the sky at the top and the car front at the bottom)\n \"\"\"\n return image[40:-20, :, :]\n\n\ndef resize(image):\n \"\"\"\n Resize the image to the input shape used by the network model\n \"\"\"\n return cv2.resize(image, (IMAGE_WIDTH, IMAGE_HEIGHT), cv2.INTER_AREA)\n\n\ndef rgb2yuv(image):\n \"\"\"\n Convert the image from RGB to YUV\n \"\"\"\n return cv2.cvtColor(image, cv2.COLOR_RGB2YUV)\n\n\ndef preprocess(image):\n \"\"\"\n Combine all preprocess functions into one\n \"\"\"\n image = crop(image)\n image = resize(image)\n image = rgb2yuv(image)\n return image\n\n\ndef pick_image(images, angles):\n \"\"\"\n Randomly choose an image from the center, left or right\n \"\"\"\n i = np.random.choice(3)\n return images[i], angles[i]\n\n\n\ndef random_flip(image, angle):\n \"\"\"\n Randomly flip the image to left or right, and adjust the steering angle.\n \"\"\"\n if np.random.rand() < 0.5:\n image = cv2.flip(image, 1)\n angle = -angle\n return image, angle\n\n\ndef random_brightness(image):\n \"\"\"\n Randomly adjust brightness of the image.\n \"\"\"\n # HSV (Hue, Saturation, Value) is also called HSB ('B' for Brightness).\n hsv = cv2.cvtColor(image, cv2.COLOR_RGB2HSV)\n ratio = 1.0 + 0.4 * (np.random.rand() - 0.5)\n hsv[:,:,2] = hsv[:,:,2] * ratio\n return cv2.cvtColor(hsv, cv2.COLOR_HSV2RGB)\n\n\ndef random_translation(image, angle, range_x=100, range_y=10):\n \"\"\"\n Randomly shift the image virtially and horizontally (translation).\n \"\"\"\n trans_x = range_x * (np.random.rand() - 0.5)\n trans_y = range_y * (np.random.rand() - 0.5)\n angle += trans_x * 0.002\n trans_m = np.float32([[1, 0, trans_x], [0, 1, trans_y]])\n height, width = image.shape[:2]\n image = cv2.warpAffine(image, trans_m, (width, height))\n\n return image, angle\n\n\ndef augment(images, angles):\n \"\"\"\n Augment images through flip, shift and brightness tuning\n \"\"\"\n # 1. randomly pick up a image\n image, angle = pick_image(images, angles)\n\n # 2. randomly flip the image\n image, angle = random_flip(image, angle)\n\n # 3. randomly adjust shift\n image, angle = random_translation(image, angle)\n\n # 4. randomly adjust brightness\n image = random_brightness(image)\n\n return image, angle\n\n\ndef load_images(img_dir, img_names, angle):\n images = [mpimg.imread(os.path.join(img_dir, img_names[i].strip().split('/')[-1])) for i in range(3)]\n angles = [angle, angle+CORRECTION, angle-CORRECTION]\n return images, angles\n\n\ndef flip_images(images, angles):\n return [cv2.flip(i, 1) for i in images], [a*(-1) for a in angles]\n\n\ndef batch_generator(img_dir, X_data, y_data, batch_size=40, is_training=True):\n batch_images = np.empty([batch_size, IMAGE_HEIGHT, IMAGE_WIDTH, IMAGE_CHANNELS])\n batch_angles = np.empty(batch_size)\n\n while True:\n i = 0\n for idx in np.random.permutation(X_data.shape[0]):\n image_names = X_data[idx]\n angle = y_data[idx]\n images, angles = load_images(img_dir, image_names, angle)\n if is_training and np.random.rand() < 0.6:\n image, angle = augment(images, angles)\n image = preprocess(image)\n else:\n image = preprocess(images[0])\n\n batch_images[i] = image\n batch_angles[i] = angle\n i += 1\n if i >= batch_size:\n break\n yield batch_images, batch_angles\n\n\ndef random_show_image(img_dir, X_data, y_data):\n import matplotlib.pyplot as plt\n\n idx = np.random.choice(X_data.shape[0])\n image_names = X_data[idx]\n angle = y_data[idx]\n\n images, angles = load_images(img_dir, image_names, angle)\n image, _ = augment(images, angles)\n image = preprocess(image)\n\n plt.imshow(image)\n plt.show()\n", "sub_path": "utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 4341, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pandas.read_csv", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 37, "usage_type": "call"}, {"api_name": "cv2.INTER_AREA", "line_number": 37, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 44, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2YUV", "line_number": 44, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 61, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 70, "usage_type": "attribute"}, {"api_name": "cv2.flip", "line_number": 71, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 81, "usage_type": "call"}, {"api_name": "cv2.COLOR_RGB2HSV", "line_number": 81, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 82, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 84, "usage_type": "call"}, {"api_name": "cv2.COLOR_HSV2RGB", "line_number": 84, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 91, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 92, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 94, "usage_type": "call"}, {"api_name": "cv2.warpAffine", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.image.imread", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.image", "line_number": 121, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path", "line_number": 121, "usage_type": "attribute"}, {"api_name": "cv2.flip", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.random.permutation", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 136, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 140, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 157, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 165, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 166, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 166, "usage_type": "name"}]}
+{"seq_id": "537114766", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Nov 15 18:06:09 2013\n\n@author: root\n\"\"\"\nimport numpy as np\n\n\n\ndef findSum(value):\n temp = np.copy(value - 1)\n sumval = 0\n for ii in range(value):\n sumval = sumval + temp\n temp = temp - 1\n return(sumval)\n \ndef compare(startpt, endpt, data):\n import cv2\n from matplotlib import pyplot as plt\n import skimage.morphology as skmorph\n def gauss_kern(Img):\n \"\"\" Returns a normalized 2D gauss kernel array for convolutions \"\"\"\n h2,h1 = Img.shape \n x, y = np.mgrid[0:h2, 0:h1]\n x = x-h2/2\n y = y-h1/2\n sigma = 3.5\n # sigma = 15\n g = np.exp( -( x**2 + y**2 ) / (2*sigma**2) );\n return g / g.sum()\n \n def myshow(img):\n \n def onClick(event):\n print(img[event.ydata,event.xdata]) \n # plt.close('all')\n plt.figure() \n plt.ion() \n plt.imshow(img,cmap='gray'),plt.show()\n fig = plt.gcf()\n # on mouse click the value at the image location is displayed in output screen\n _ = fig.canvas.mpl_connect('button_press_event', onClick)\n \n def myshow2(img):\n \n def onClick(event):\n print(img[event.ydata,event.xdata]) \n # plt.close('all')\n plt.figure() \n plt.ion() \n plt.imshow(img),plt.show()\n fig = plt.gcf()\n # on mouse click the value at the image location is displayed in output screen\n _ = fig.canvas.mpl_connect('button_press_event', onClick)\n \n \n img = imgout = np.zeros((200,200))\n rotimg = np.zeros((200,200), np.uint8)\n lengthofrect = np.int32(np.sqrt((startpt[0] - endpt[0])**2 + (startpt[1] - endpt[1])**2))\n angleofrect = np.math.atan2((startpt[0] - endpt[0]), (startpt[1] - endpt[1]))\n widthofrect = 25\n img[95:105, 100 - lengthofrect/2: 100 + lengthofrect/2] = 1\n \n \n gau = gauss_kern(img)\n #Img_smooth = signal.convolve(Img,g,mode='same')\n imgfft = np.fft.rfft2(img)\n gfft = np.fft.rfft2(gau)\n fftimage = np.multiply(imgfft, gfft)\n img_smooth =np.real(np.fft.ifftshift( np.fft.irfft2(fftimage)))\n imgout = img_smooth>0.1\n \n #imgborder = imgout - cv2.dilate(np.uint8(imgout), None, iterations=1)\n #bordercoord = np.argwhere(imgborder)\n \n imgcoord = np.float64(np.argwhere(imgout))\n imgcoordorigin = imgrotcoord = imgrotcoordorigin = np.zeros((imgcoord.shape))\n rowmean = imgcoord[:,0].mean()\n colmean = imgcoord[:,1].mean()\n imgcoordorigin[:, 0] = imgcoord[:, 0] - rowmean\n imgcoordorigin[:, 1] = imgcoord[:, 1] - colmean\n \n imgrotcoordorigin[:, 0] = imgcoordorigin[:, 1] * np.math.sin(angleofrect) + imgcoordorigin[:, 0] * np.math.cos(angleofrect)\n imgrotcoordorigin[:, 1] = imgcoordorigin[:, 1] * np.math.cos(angleofrect) - imgcoordorigin[:, 0] * np.math.sin(angleofrect)\n imgrotcoord[:, 0] = imgrotcoordorigin[:, 0] + rowmean\n imgrotcoord[:, 1] = imgrotcoordorigin[:, 1] + colmean\n imgrotcoord = np.int64(imgrotcoord)\n strelplus = np.ones((3,3),np.uint8)\n strelplus[0,0] = strelplus[2,0] = strelplus[0,2] = strelplus[2,2] = 0\n #rotimg = np.uint8(rotimg)\n rotimg[imgrotcoord[:,0], imgrotcoord[:,1]] = 1\n #rotimg = cv2.dilate(rotimg, strelplus, iterations = 1)\n #rotimg = cv2.erode(rotimg, strelplus, iterations = 1)\n rotimg = np.uint8(rotimg)\n rotimg = skmorph.closing(rotimg, np.ones((3, 3), np.uint8))\n \n #enlarge the image to gather the details of crop region\n enlargimg = skmorph.dilation(rotimg, np.ones((9, 9), np.uint8))\n enlargimgcoord = np.argwhere(enlargimg)\n encoord = np.copy(enlargimgcoord)\n #normalize the coordinate\n enlargimgcoord[:, 0] = enlargimgcoord[:, 0] - enlargimgcoord[:, 0].mean()\n enlargimgcoord[:, 1] = enlargimgcoord[:, 1] - enlargimgcoord[:, 1].mean()\n #midpoint of skeleton line\n midpt = np.array( [startpt[0] + endpt[0], startpt[1] + endpt[1]])/2\n #add midpoint of skeleton to get the image coordinates of data\n \n imgpartcoord = np.zeros(enlargimgcoord.shape)\n imgpartcoord[:, 0] = enlargimgcoord[:, 0] + midpt[0]\n imgpartcoord[:, 1] = enlargimgcoord[:, 1] + midpt[1]\n imgpartcoord = np.int32(imgpartcoord)\n # image part to compare\n outcoord = np.argwhere(rotimg)\n imgpart = np.zeros((200,200))\n imgpart[encoord[:, 0], encoord[:, 1]] = data[imgpartcoord[:, 0], imgpartcoord[:, 1]]\n numerator = np.sum(imgpart[outcoord[:,0], outcoord[:,1]]) * 2 \n denomenator = np.argwhere(imgpart).shape[0] + np.argwhere(rotimg).shape[0]\n dicescore = numerator/denomenator\n coverage = np.float64(np.argwhere(imgpart).shape[0])/np.float64(enlargimgcoord.shape[0]) \n return(dicescore, imgpart, coverage, imgpartcoord)\n\n#urotimg, srotimg, vrotimg = np.linalg.svd(rotimg)\n#uimgpart, simgpart, vimgpart = np.linalg.svd(imgpart)\n\n#evalrotimg, princomprotimg = principalComponents(rotimg)\n#evalimgpart, princompimgpart = principalComponents(imgpart)\n\n\n\n\n\n\n", "sub_path": "gaussianRectangle.py", "file_name": "gaussianRectangle.py", "file_ext": "py", "file_size_in_byte": 5002, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "numpy.copy", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.mgrid", "line_number": 26, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ion", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ion", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 60, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.math.atan2", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.math", "line_number": 62, "usage_type": "attribute"}, {"api_name": "numpy.fft.rfft2", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 69, "usage_type": "attribute"}, {"api_name": "numpy.fft.rfft2", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 70, "usage_type": "attribute"}, {"api_name": "numpy.multiply", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.real", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.fft.ifftshift", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 72, "usage_type": "attribute"}, {"api_name": "numpy.fft.irfft2", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.math.sin", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.math", "line_number": 85, "usage_type": "attribute"}, {"api_name": "numpy.math.cos", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.math.cos", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.math", "line_number": 86, "usage_type": "attribute"}, {"api_name": "numpy.math.sin", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 90, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 96, "usage_type": "call"}, {"api_name": "skimage.morphology.closing", "line_number": 97, "usage_type": "call"}, {"api_name": "skimage.morphology", "line_number": 97, "usage_type": "name"}, {"api_name": "numpy.ones", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 97, "usage_type": "attribute"}, {"api_name": "skimage.morphology.dilation", "line_number": 100, "usage_type": "call"}, {"api_name": "skimage.morphology", "line_number": 100, "usage_type": "name"}, {"api_name": "numpy.ones", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 100, "usage_type": "attribute"}, {"api_name": "numpy.argwhere", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 121, "usage_type": "call"}]}
+{"seq_id": "612323363", "text": "import json\nimport logging\nimport os\n\nfrom tabulate import tabulate\nfrom jsonschema.exceptions import ValidationError\nfrom buildtest.buildsystem.parser import BuildspecParser\nfrom buildtest.config import load_settings\nfrom buildtest.defaults import BUILDSPEC_CACHE_FILE, BUILDSPEC_DEFAULT_PATH\nfrom buildtest.exceptions import BuildTestError\nfrom buildtest.utils.file import is_file, walk_tree, resolve_path\n\nlogger = logging.getLogger(__name__)\n\n\nclass BuildspecCache:\n\n table = {}\n filter_fields = [\"type\", \"executor\", \"tags\"]\n default_format_fields = [\"name\", \"type\", \"executor\", \"tags\", \"description\"]\n format_fields = default_format_fields + [\"file\"]\n\n def __init__(self, rebuild, filterfields, formatfields, roots):\n self.filter = filterfields\n self.format = formatfields\n self.roots = roots\n self.paths = []\n self.rebuild = rebuild\n self.cache = {}\n\n self.load_paths()\n self.build()\n\n self.check_filter_fields()\n self.check_format_fields()\n self.find_buildspecs()\n\n def get_paths(self):\n \"\"\"Returns a list of root buildspec roots\"\"\"\n\n return self.paths\n\n def get_cache(self):\n \"\"\"Returns cache file as loaded dictionary\"\"\"\n\n return self.cache\n\n def load_paths(self):\n \"\"\"Add all paths to search for buildspecs. We must read configuration file\n and check property ``buildspec_roots`` for list of directories to search.\n We check all directories exist, if any fail we don't add them to path.\n In addition, we add the default buildspec path where we find tutorials\n and general tests.\n \"\"\"\n\n config_opts = load_settings()\n buildspec_paths = config_opts.get(\"buildspec_roots\") or []\n\n # self.file_roots will store files specified by --roots option\n self.file_roots = []\n\n if self.roots:\n buildspec_paths += self.roots\n\n # only load default buildspecs if 'load_default_buildspecs' set to True\n if config_opts.get(\"load_default_buildspecs\"):\n self.paths += BUILDSPEC_DEFAULT_PATH\n\n # if buildspec_roots defined in configuration, resolve path and if path exist add\n # to list of paths to search for buildspecs\n if buildspec_paths:\n\n for root in buildspec_paths:\n path = resolve_path(root, exist=False)\n if not os.path.exists(path):\n print(f\"Path: {path} does not exist!\")\n\n if is_file(path):\n self.file_roots.append(path)\n\n self.paths.append(path)\n\n def build(self):\n \"\"\"This method will build buildspec cache file. If user requests to\n rebuild cache we remove the file and recreate cache. If cache file\n exists, we simply load from cache\n \"\"\"\n\n # implements buildtest buildspec find --rebuild which removes cache file\n # before finding all buildspecs. We only remove file if file exists\n if self.rebuild and is_file(BUILDSPEC_CACHE_FILE):\n try:\n os.remove(BUILDSPEC_CACHE_FILE)\n print(f\"Clearing cache file: {BUILDSPEC_CACHE_FILE}\")\n except OSError as msg:\n raise BuildTestError(msg)\n\n # if cache file is not found, then we will build cache by searching\n # all buildspecs paths and traverse directory to find all .yml files\n\n if not is_file(BUILDSPEC_CACHE_FILE):\n self.build_cache()\n\n with open(BUILDSPEC_CACHE_FILE, \"r\") as fd:\n self.cache = json.loads(fd.read())\n\n def _discover_buildspecs(self):\n \"\"\"This method retrieves buildspecs based on ``self.paths`` which is a\n list of directory paths to search. If --root is specified for specifying\n buildspec roots we process each argument, if its a file we add file,\n if its a directory we recursively find all .yml files\n \"\"\"\n\n buildspecs = []\n # add all buildspecs from each repo. walk_tree will find all .yml files\n # recursively and add them to list\n for path in self.paths:\n buildspec = walk_tree(path, \".yml\")\n buildspecs += buildspec\n\n # if --root specifies a file we add buildspecs only if they end in .yml extension\n if self.file_roots:\n for filename in self.file_roots:\n if filename.endswith(\".yml\"):\n buildspecs += filename\n else:\n print(\n f\"File: {filename} does not end in .yml extension, skipping file\"\n )\n\n print(f\"\\nBuildspec Paths: {self.paths} \\n\")\n\n # remove any files in .buildtest directory of root of repo.\n buildspecs = [\n buildspec\n for buildspec in buildspecs\n if os.path.basename(os.path.dirname(buildspec)) != \".buildtest\"\n ]\n return buildspecs\n\n def _write_buildspec_cache(self):\n \"\"\"This method is responsible for writing buildspec cache to file\"\"\"\n\n with open(BUILDSPEC_CACHE_FILE, \"w\") as fd:\n json.dump(self.update_cache, fd, indent=2)\n\n print(f\"\\nDetected {len(self.invalid_buildspecs)} invalid buildspecs \\n\")\n\n # write invalid buildspecs to file if any found\n if self.invalid_buildspecs:\n buildspec_error_file = os.path.join(\n os.path.dirname(BUILDSPEC_CACHE_FILE), \"buildspec.error\"\n )\n\n with open(buildspec_error_file, \"w\") as fd:\n for file, msg in self.invalid_buildspecs.items():\n fd.write(f\"buildspec:{file} \\n\\n\")\n fd.write(f\"{msg} \\n\")\n\n print(f\"Writing invalid buildspecs to file: {buildspec_error_file} \")\n print(\"\\n\\n\")\n\n def _validate_buildspecs(self, buildspecs):\n \"\"\"Given a list of buildspec files, validate each buildspec using BuildspecParser\n and return a list of valid buildspecs. Any invalid buildspecs are added to\n separate list\n \"\"\"\n valid_buildspecs = []\n self.count = 0\n for buildspec in buildspecs:\n self.count += 1\n\n try:\n parse = BuildspecParser(buildspec)\n # any buildspec that raises SystemExit or ValidationError imply\n # buildspec is not valid, we add this to invalid list along with\n # error message and skip to next buildspec\n except (BuildTestError, ValidationError) as err:\n self.invalid_buildspecs[buildspec] = err\n continue\n\n valid_buildspecs.append(parse)\n\n if self.count % 5 == 0:\n print(f\"Validated {self.count}/{len(buildspecs)} buildspecs\")\n\n print(f\"Validated {self.count}/{len(buildspecs)} buildspecs\")\n return valid_buildspecs\n\n def build_cache(self):\n \"\"\"This method will rebuild the buildspec cache file by recursively searching\n all .yml files specified by input argument ``paths`` which is a list of directory\n roots. The buildspecs are validated and cache file is updated\"\n\n :param paths: A list of directory roots to process buildspecs files.\n :type paths: list\n :return: Rebuild cache file\n \"\"\"\n\n self.update_cache = {}\n self.update_cache[\"unique_tags\"] = []\n self.update_cache[\"unique_executors\"] = []\n self.update_cache[\"buildspecs\"] = {}\n self.update_cache[\"executor\"] = {}\n self.update_cache[\"tags\"] = {}\n self.invalid_buildspecs = {}\n\n for path in self.paths:\n self.update_cache[path] = {}\n\n buildspecs = self._discover_buildspecs()\n print(f\"Found {len(buildspecs)} buildspecs \")\n\n # validate each buildspec and return a list of valid buildspec parsers that\n # is an instance of BuildspecParser class\n parsers = self._validate_buildspecs(buildspecs)\n\n # for every parsers (valid buildspecs) we update cache to build an index\n for parser in parsers:\n\n recipe = parser.recipe[\"buildspecs\"]\n\n path_root = [\n path\n for path in self.paths\n if os.path.commonprefix([parser.buildspec, path]) == path\n ]\n path_root = path_root[0]\n\n if not self.update_cache[\"buildspecs\"].get(path_root):\n self.update_cache[\"buildspecs\"][path_root] = {}\n\n if not self.update_cache[\"buildspecs\"][path_root].get(parser.buildspec):\n self.update_cache[\"buildspecs\"][path_root][parser.buildspec] = {}\n\n for name in recipe.keys():\n\n self.update_cache[\"buildspecs\"][path_root][parser.buildspec][\n name\n ] = recipe[name]\n tags = recipe[name].get(\"tags\")\n executor = recipe[name].get(\"executor\")\n description = recipe[name].get(\"description\")\n\n if tags:\n\n if isinstance(tags, str):\n self.update_cache[\"unique_tags\"].append(tags)\n\n if not self.update_cache[\"tags\"].get(tags):\n self.update_cache[\"tags\"][tags] = {}\n\n self.update_cache[\"tags\"][tags][name] = description\n\n elif isinstance(tags, list):\n self.update_cache[\"unique_tags\"] += tags\n\n # for every tagname, build a tags to testname association\n for tag in tags:\n if not self.update_cache[\"tags\"].get(tag):\n self.update_cache[\"tags\"][tag] = {}\n\n self.update_cache[\"tags\"][tag][name] = description\n\n if executor:\n self.update_cache[\"unique_executors\"].append(executor)\n\n if not self.update_cache[\"executor\"].get(executor):\n self.update_cache[\"executor\"][executor] = {}\n\n self.update_cache[\"executor\"][executor][name] = description\n\n self.update_cache[\"unique_tags\"] = list(set(self.update_cache[\"unique_tags\"]))\n self.update_cache[\"unique_executors\"] = list(\n set(self.update_cache[\"unique_executors\"])\n )\n self._write_buildspec_cache()\n\n def check_filter_fields(self):\n \"\"\" This method checks filter fields are valid. The filter fields are specified\n as ``buildtest buildspec find --filter =,=,...\n \"\"\"\n\n self.executor_filter = None\n self.tags_filter = None\n self.type_filter = None\n\n if self.filter:\n\n filter_error = False\n # check if filter keys are accepted filter fields, if not we raise error\n for key in self.filter.keys():\n if key not in self.filter_fields:\n print(f\"Invalid filter key: {key}\")\n filter_error = True\n\n # raise error if any filter field is invalid\n if filter_error:\n raise BuildTestError(f\"Invalid filter fields format {self.filter}\")\n\n self.executor_filter = self.filter.get(\"executor\")\n self.tags_filter = self.filter.get(\"tags\")\n self.type_filter = self.filter.get(\"type\")\n\n def check_format_fields(self):\n \"\"\" This method will check if all format fields are valid. Format fields\n are passed comma separated as --format field1,field2,field3,...\n \"\"\"\n\n for field in self.default_format_fields:\n self.table[field] = []\n\n if self.format:\n\n format_error = False\n for key in self.format.split(\",\"):\n if key not in self.format_fields:\n print(f\"Invalid format field: {key}\")\n format_error = True\n\n if format_error:\n raise BuildTestError(f\"Invalid format fields format {self.format}\")\n\n # if --format option specified we setup cache dictionary based on format\n # fields that are added to list\n self.table = {}\n for field in self.format.split(\",\"):\n self.table[field] = []\n\n def _filter_buildspecs(self, executor, tags, schema_type):\n \"\"\" This method will return a boolean True/False that determines if\n buildspec test entry is skipped as part of filter process. The filter\n are done based on executor, tags, type field. ``True`` indicates test\n needs to be skipped.\n\n :param executor: 'executor; field from buildspec recipe\n :type executor: str, required\n :param tags: 'tags' field from buildspec recipe\n :type tags: str or list, required\n :param schema_type: 'type' field from buildspec recipe\n :type schema_type: str, required\n :return: boolean to determine if we need to skip buildspec\n :rtype: bool\n \"\"\"\n\n # skip all entries that dont match filtered executor\n if self.executor_filter and self.executor_filter != executor:\n return True\n\n # if skip all entries that dont match filtered tag. We only search if --filter tag=value is set\n if self.tags_filter:\n # if tags is not set in buildspec cache we default to empty list which and this condition should always be true\n if self.tags_filter not in tags:\n return True\n\n if self.type_filter and self.type_filter != schema_type:\n return True\n\n return False\n\n def find_buildspecs(self):\n \"\"\" This method will find buildspecs based on cache content. We skip any\n tests based on executor filter, tag filter or type filter and build\n a table of tests that will be printed using print_buildspecs method.\n \"\"\"\n\n for path in self.cache[\"buildspecs\"].keys():\n for buildspecfile in self.cache[\"buildspecs\"][path].keys():\n for test in self.cache[\"buildspecs\"][path][buildspecfile].keys():\n\n test_recipe = self.cache[\"buildspecs\"][path][buildspecfile][test]\n schema_type = test_recipe.get(\"type\")\n executor = test_recipe.get(\"executor\")\n # if tags not defined in cache we set to empty list for comparison with tag_filter\n tags = test_recipe.get(\"tags\") or []\n description = test_recipe.get(\"description\")\n\n # filters buildspecs by executor, tags, type field. The return\n # is a boolean, if its True we skip the test\n if self._filter_buildspecs(executor, tags, schema_type):\n continue\n\n if self.format:\n for field in self.table.keys():\n if field == \"type\":\n self.table[field].append(schema_type)\n\n elif field == \"file\":\n self.table[field].append(buildspecfile)\n elif field == \"name\":\n self.table[field].append(test)\n # description, tags, executor have matching format fields with buildspec properties\n else:\n self.table[field].append(test_recipe.get(field))\n\n else:\n self.table[\"name\"].append(test)\n self.table[\"type\"].append(schema_type)\n self.table[\"executor\"].append(executor)\n self.table[\"tags\"].append(tags)\n self.table[\"description\"].append(description)\n\n def get_buildspecfiles(self):\n \"\"\" This method implements ``buildtest buildspec find --buildspec-files``\n which reports all buildspec files in cache.\n\n :param cache: content of cache as dictionary\n :type cache: dict\n \"\"\"\n\n table = {\"buildspecs\": []}\n files = []\n\n for path in self.cache[\"buildspecs\"].keys():\n files += self.cache[\"buildspecs\"][path].keys()\n\n table[\"buildspecs\"] = files\n print(tabulate(table, headers=table.keys(), tablefmt=\"grid\"))\n\n def get_tags(self):\n \"\"\" This method implements ``buildtest buildspec find --tags`` which\n reports a list of unique tags from all buildspecs in cache file.\n\n :param cache: content of cache as dictionary\n :type cache: dict\n \"\"\"\n\n table = {\"Tags\": []}\n\n table[\"Tags\"] = self.cache[\"unique_tags\"]\n print(tabulate(table, headers=table.keys(), tablefmt=\"grid\"))\n\n def get_executors(self):\n \"\"\" This method implements ``buildtest buildspec find --list-executors``\n which reports all executors from cache.\n\n :param cache: content of cache as dictionary\n :type cache: dict\n \"\"\"\n\n table = {\"executors\": []}\n table[\"executors\"] = self.cache[\"unique_executors\"]\n print(tabulate(table, headers=table.keys(), tablefmt=\"grid\"))\n\n def print_by_executors(self):\n \"\"\" This method prints executors by tests and implements\n ``buildtest buildspec find --test-by-tags`` command\n \"\"\"\n\n table = {\"executor\": [], \"name\": [], \"description\": []}\n\n for executor_name in self.cache[\"executor\"].keys():\n for test_name, description in self.cache[\"executor\"][executor_name].items():\n table[\"executor\"].append(executor_name)\n table[\"name\"].append(test_name)\n table[\"description\"].append(description)\n\n print(tabulate(table, headers=table.keys(), tablefmt=\"grid\"))\n\n def print_by_tags(self):\n \"\"\" This method prints tags by tests and implements\n ``buildtest buildspec find --test-by-tags`` command\n \"\"\"\n\n table = {\"tags\": [], \"name\": [], \"description\": []}\n\n for tagname in self.cache[\"tags\"].keys():\n for test_name, description in self.cache[\"tags\"][tagname].items():\n table[\"tags\"].append(tagname)\n table[\"name\"].append(test_name)\n table[\"description\"].append(description)\n\n print(tabulate(table, headers=table.keys(), tablefmt=\"grid\"))\n\n def print_buildspecs(self):\n \"\"\"Print buildspec table\"\"\"\n\n print(tabulate(self.table, headers=self.table.keys(), tablefmt=\"grid\"))\n\n @staticmethod\n def print_filter_fields():\n \"\"\"This method prints filter fields available for buildspec cache. This\n method implements command ``buildtest buildspec find --helpfilter``\"\"\"\n\n filter_field_table = [\n [\"executor\", \"Filter by executor name\", \"STRING\"],\n [\"tags\", \"Filter by tag name \", \"STRING\"],\n [\"type\", \"Filter by schema type \", \"STRING\"],\n ]\n\n print(\n tabulate(\n filter_field_table,\n headers=[\"Field\", \"Description\", \"Type\"],\n tablefmt=\"simple\",\n )\n )\n\n @staticmethod\n def print_format_fields():\n \"\"\"This method prints format fields available for buildspec cache. This\n method implements command ``buildtest buildspec find --helpformat``\"\"\"\n\n format_fields = [\n [\"name\", \"Format by test name\"],\n [\"tags\", \"Format by tag name \"],\n [\"type\", \"Format by schema type\"],\n [\"executor\", \"Format by executor type\"],\n [\"description\", \"Format by description\"],\n [\"file\", \"Format by file\"],\n ]\n\n print(\n tabulate(\n format_fields, headers=[\"Field\", \"Description\"], tablefmt=\"simple\",\n )\n )\n\n def print_paths(self):\n \"\"\" This method print buildspec paths, this implements command\n ``buildtest buildspec find --paths``\n \"\"\"\n\n for path in self.get_paths():\n print(path)\n\n\ndef func_buildspec_find(args):\n \"\"\" Entry point for ``buildtest buildspec find``. This method\n will attempt to read for buildspec cache file (BUILDSPEC_CACHE_FILE)\n if found and print a list of all buildspecs. Otherwise, it will\n find and load all buildspecs and validate them using BuildspecParser class.\n BuildspecParser will raise SystemError or ValidationError if a buildspec\n is invalid which will be added to list of invalid buildspecs. Finally we\n print a list of all valid buildspecs and any invalid buildspecs are\n written to file along with error message.\n\n :param args: Input argument from command line passed from argparse\n :return: A list of valid buildspecs found in all repositories.\n \"\"\"\n\n bp_cache = BuildspecCache(\n rebuild=args.rebuild,\n filterfields=args.filter,\n formatfields=args.format,\n roots=args.root,\n )\n\n # implements buildtest buildspec find --tags\n if args.tags:\n bp_cache.get_tags()\n return\n\n # implements buildtest buildspec find --buildspec-files\n if args.buildspec_files:\n bp_cache.get_buildspecfiles()\n return\n\n if args.paths:\n bp_cache.print_paths()\n return\n\n # implements buildtest buildspec find --executors\n if args.executors:\n bp_cache.get_executors()\n return\n\n if args.group_by_executor:\n bp_cache.print_by_executors()\n return\n\n if args.group_by_tags:\n bp_cache.print_by_tags()\n return\n\n # implements buildtest buildspec find --helpfilter\n if args.helpfilter:\n bp_cache.print_filter_fields()\n return\n\n # implements buildtest buildspec find --helpformat\n if args.helpformat:\n bp_cache.print_format_fields()\n return\n\n bp_cache.print_buildspecs()\n", "sub_path": "buildtest/menu/buildspec.py", "file_name": "buildspec.py", "file_ext": "py", "file_size_in_byte": 22001, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.getLogger", "line_number": 13, "usage_type": "call"}, {"api_name": "buildtest.config.load_settings", "line_number": 56, "usage_type": "call"}, {"api_name": "buildtest.defaults.BUILDSPEC_DEFAULT_PATH", "line_number": 67, "usage_type": "name"}, {"api_name": "buildtest.utils.file.resolve_path", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path", "line_number": 75, "usage_type": "attribute"}, {"api_name": "buildtest.utils.file.is_file", "line_number": 78, "usage_type": "call"}, {"api_name": "buildtest.utils.file.is_file", "line_number": 91, "usage_type": "call"}, {"api_name": "buildtest.defaults.BUILDSPEC_CACHE_FILE", "line_number": 91, "usage_type": "argument"}, {"api_name": "os.remove", "line_number": 93, "usage_type": "call"}, {"api_name": "buildtest.defaults.BUILDSPEC_CACHE_FILE", "line_number": 93, "usage_type": "argument"}, {"api_name": "buildtest.defaults.BUILDSPEC_CACHE_FILE", "line_number": 94, "usage_type": "name"}, {"api_name": "buildtest.exceptions.BuildTestError", "line_number": 96, "usage_type": "call"}, {"api_name": "buildtest.utils.file.is_file", "line_number": 101, "usage_type": "call"}, {"api_name": "buildtest.defaults.BUILDSPEC_CACHE_FILE", "line_number": 101, "usage_type": "argument"}, {"api_name": "buildtest.defaults.BUILDSPEC_CACHE_FILE", "line_number": 104, "usage_type": "argument"}, {"api_name": "json.loads", "line_number": 105, "usage_type": "call"}, {"api_name": "buildtest.utils.file.walk_tree", "line_number": 118, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 137, "usage_type": "call"}, {"api_name": "os.path", "line_number": 137, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 137, "usage_type": "call"}, {"api_name": "buildtest.defaults.BUILDSPEC_CACHE_FILE", "line_number": 144, "usage_type": "argument"}, {"api_name": "json.dump", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 151, "usage_type": "call"}, {"api_name": "os.path", "line_number": 151, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 152, "usage_type": "call"}, {"api_name": "buildtest.defaults.BUILDSPEC_CACHE_FILE", "line_number": 152, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 152, "usage_type": "attribute"}, {"api_name": "buildtest.buildsystem.parser.BuildspecParser", "line_number": 174, "usage_type": "call"}, {"api_name": "buildtest.exceptions.BuildTestError", "line_number": 178, "usage_type": "name"}, {"api_name": "jsonschema.exceptions.ValidationError", "line_number": 178, "usage_type": "name"}, {"api_name": "os.path.commonprefix", "line_number": 226, "usage_type": "call"}, {"api_name": "os.path", "line_number": 226, "usage_type": "attribute"}, {"api_name": "buildtest.exceptions.BuildTestError", "line_number": 299, "usage_type": "call"}, {"api_name": "buildtest.exceptions.BuildTestError", "line_number": 322, "usage_type": "call"}, {"api_name": "tabulate.tabulate", "line_number": 418, "usage_type": "call"}, {"api_name": "tabulate.tabulate", "line_number": 431, "usage_type": "call"}, {"api_name": "tabulate.tabulate", "line_number": 443, "usage_type": "call"}, {"api_name": "tabulate.tabulate", "line_number": 458, "usage_type": "call"}, {"api_name": "tabulate.tabulate", "line_number": 473, "usage_type": "call"}, {"api_name": "tabulate.tabulate", "line_number": 478, "usage_type": "call"}, {"api_name": "tabulate.tabulate", "line_number": 492, "usage_type": "call"}, {"api_name": "tabulate.tabulate", "line_number": 514, "usage_type": "call"}]}
+{"seq_id": "573467990", "text": "from sklearn.feature_extraction.text import CountVectorizer\nfrom sklearn.feature_extraction.text import TfidfVectorizer\nfrom sklearn.model_selection import cross_val_score\nfrom sklearn.metrics import accuracy_score\nimport datetime\n\nconst_interpunction = [',', '.', ';', '-', '?', '!', ':', '|', '_', '@', '~', '#', '^', '(', ')', '{', '}', '[', ']', '\\\\', '/', '+']\n\nconst_stopwords = ['i', 'me', 'my', 'myself', 'we', 'our', 'ours', 'ourselves', 'you', \"you're\", \"you've\", \"you'll\",\n \"you'd\", 'your', 'yours', 'yourself', 'yourselves', 'he', 'him', 'his', 'himself', 'she', \"she's\",\n 'her', 'hers', 'herself', 'it', \"it's\", 'its', 'itself', 'they', 'them', 'their', 'theirs',\n 'themselves', 'what', 'which', 'who', 'whom', 'this', 'that', \"that'll\", 'these', 'those', 'am',\n 'is', 'are', 'was', 'were', 'be', 'been', 'being', 'have', 'has', 'had', 'having', 'do', 'does',\n 'did', 'doing', 'a', 'an', 'the', 'and', 'but', 'if', 'or', 'because', 'as', 'until', 'while', 'of',\n 'at', 'by', 'for', 'with', 'about', 'against', 'between', 'into', 'through', 'during', 'before',\n 'after', 'above', 'below', 'to', 'from', 'up', 'down', 'in', 'out', 'on', 'off', 'over', 'under',\n 'again', 'further', 'then', 'once', 'here', 'there', 'when', 'where', 'why', 'how', 'all', 'any',\n 'both', 'each', 'few', 'more', 'most', 'other', 'some', 'such', 'no', 'nor', 'not', 'only', 'own',\n 'same', 'so', 'than', 'too', 'very', 's', 't', 'can', 'will', 'just', 'don', \"don't\", 'should',\n \"should've\", 'now', 'd', 'll', 'm', 'o', 're', 've', 'y', 'ain', 'aren', \"aren't\", 'couldn',\n \"couldn't\", 'didn', \"didn't\", 'doesn', \"doesn't\", 'hadn', \"hadn't\", 'hasn', \"hasn't\", 'haven',\n \"haven't\", 'isn', \"isn't\", 'ma', 'mightn', \"mightn't\", 'mustn', \"mustn't\", 'needn', \"needn't\",\n 'shan', \"shan't\", 'shouldn', \"shouldn't\", 'wasn', \"wasn't\", 'weren', \"weren't\", 'won', \"won't\",\n 'wouldn', \"wouldn't\"]\n\n\nclass Title:\n def __init__(self, text, clickbait):\n self.text = text\n self.clickbait = clickbait\n\n'''\n Method finds occurrences of single character in string\n'''\ndef find_char_occurrences(s, ch):\n return [i for i, letter in enumerate(s) if letter == ch]\n\n'''\n Method finds occurrences of substring in string\n'''\ndef find_substring_occurrences(text, substring):\n return [i for i in range(len(text)) if text.startswith(substring, i)]\n\n\n'''\nMethod used to remove interpunction signs from const_interpunction,\nbut also to remove sign ' in places where it is used as quotation mark\n'''\n\n\ndef remove_interpunction(text):\n\n # removing signs from const_interpunction list\n for interpunction in const_interpunction:\n text = text.replace(interpunction, ' ')\n\n # removing ' character\n c = \"'\"\n occurences = find_char_occurrences(text, c)\n for position in occurences:\n # if ' is appearing as first or as last character in string\n if (position == 0) or (position + 1 == len(text)):\n text = text[:position:] + ' ' + text[position + 1::]\n else:\n # if somewhere around ' is an space, before or after\n if (text[position - 1] == ' ') or (text[position + 1] == ' '):\n text = text[:position:] + ' ' + text[position + 1::]\n\n return text\n\n'''\n Method used to remove stopwords from text\n'''\ndef remove_stopwords(text):\n for stopword in const_stopwords:\n if stopword in text:\n occurences = find_substring_occurrences(text, stopword)\n\n # going through every start index of every occurrence of stopword in text\n for position in occurences:\n remove = False\n if position == 0:\n # if it is the first word in string\n if text[position + len(stopword)] == ' ':\n remove = True\n elif position + len(stopword) == len(text):\n # if it is the last word in string\n if text[position - 1] == ' ':\n remove = True\n else:\n # if it is not part of another word\n if (text[position - 1] == ' ') and (text[position + len(stopword)] == ' '):\n remove = True\n\n if remove:\n text = text[:position:] + text[position + len(stopword)::]\n\n return text\n\ndef text_preprocessing(fileName, isTraining):\n naslovi = []\n clickbates = []\n with open(fileName) as train_json:\n titles = train_json.read().split(\"},{\")\n first_line = True\n for title in titles:\n # removing keyword\n if first_line:\n if isTraining:\n title = title[15:]\n else:\n title = title[14:]\n first_line = False\n else:\n if isTraining:\n title = title[13:]\n else:\n title = title[12:]\n clickbait = title[0]\n\n if isTraining:\n naslov = title[11:-1].lower()\n else:\n naslov = title[10:-1].lower()\n # text = remove_stopwords(naslov)\n # text = remove_interpunction(text)\n\n naslovi.append(naslov)\n clickbates.append(clickbait)\n\n naslovi[-1] = naslovi[-1][:-2]\n\n # for naslov in naslovi:\n # print(naslov)\n\n return naslovi, clickbates\n\n# def text_preprocessing(fileName):\n# with open(fileName) as train_json:\n# titles = train_json.read().split(\"},{\")\n# first_line = True\n# title_objects = []\n# for title in titles:\n# # removing keyword\n# if first_line:\n# title = title[14:]\n# first_line = False\n# else:\n# title = title[12:]\n#\n# clickbait = title[0]\n#\n# lower_title_text = title[10:-1].lower()\n# text = remove_stopwords(lower_title_text)\n# text = remove_interpunction(text)\n# title_objects.append(Title(text, clickbait))\n#\n# title_objects[-1].text = title_objects[-1].text[:-2]\n#\n# for t in title_objects:\n# print(t.clickbait)\n# print(t.text)\n#\n# X_train = []\n# Y_train = []\n# for title in title_objects:\n# X_train.append(title.text)\n# # Y_train.append(title.clickbait)\n#\n# return X_train, Y_train\n\ndef vectorisation(training, test):\n # vectorizer = CountVectorizer()\n # vectorizer = CountVectorizer(training, stop_words=const_stopwords)\n vectorizer = TfidfVectorizer()\n vector_training = vectorizer.fit_transform(training)\n vector_test = vectorizer.transform(test)\n return vector_training.toarray(), vector_test.toarray()\n\nif __name__ == '__main__':\n # if len(sys.argv) != 3:\n # print(\"Bad argument list, enter in following form:\")\n # print(\"python .py \")\n # exit()\n # X_train, Y_train = text_preprocessing(sys.argv[1], True)\n # X_test, Y_test = text_preprocessing(sys.argv[2], True)\n\n a = datetime.datetime.now()\n X_train, Y_train = text_preprocessing('resources/train.json', True)\n X_test, Y_test = text_preprocessing('resources/preview.json', False)\n\n X_train, X_test = vectorisation(X_train, X_test)\n\n from sklearn.svm import SVC\n\n classifier = SVC(kernel='linear', random_state=0)\n classifier.fit(X_train, Y_train)\n\n Y_Pred = classifier.predict(X_test)\n\n # scores = cross_val_score(classifier, X_test, Y_test, cv=5)\n # print(scores)\n\n\n print(accuracy_score(Y_test, Y_Pred))\n b = datetime.datetime.now()\n print(b - a)\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "domaci_03/svm.py", "file_name": "svm.py", "file_ext": "py", "file_size_in_byte": 7929, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sklearn.feature_extraction.text.TfidfVectorizer", "line_number": 175, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 188, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 188, "usage_type": "attribute"}, {"api_name": "sklearn.svm.SVC", "line_number": 196, "usage_type": "call"}, {"api_name": "sklearn.metrics.accuracy_score", "line_number": 205, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 206, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 206, "usage_type": "attribute"}]}
+{"seq_id": "52646267", "text": "import asyncio\nfrom http.server import BaseHTTPRequestHandler\nfrom io import BytesIO\n\n\nclass HTTPRequest(BaseHTTPRequestHandler):\n def __init__(self, request_bytes):\n self.rfile = BytesIO(request_bytes)\n self.raw_requestline = self.rfile.readline()\n self.error_code = self.error_message = None\n self.parse_request()\n\n def send_error(self, code, message):\n self.error_code = code\n self.error_message = message\n\n\n@asyncio.coroutine\ndef handle_request(reader, writer):\n data = b\"\"\n\n while True:\n try:\n data += yield from asyncio.wait_for(reader.read(100), timeout=10)\n except asyncio.TimeoutError:\n print(\"TimeoutError!\")\n writer.write(b\"Timeout!\")\n writer.close()\n return\n\n header_end_ix = data.find(b\"\\r\\n\\r\\n\")\n if header_end_ix != -1:\n print(\"Detected end of header\")\n break\n print(\"Not at end of header yet\")\n\n request = HTTPRequest(data[:header_end_ix])\n if request.error_code is not None:\n reply = \"Error {}\".format(request.error_code)\n else:\n host = request.headers.get(\"Host\")\n if host is None:\n print(\"It's a non-hosty one\")\n else:\n print(\"It has the host {!r}\".format(host))\n\n reply = \"You're trying to access {!r}\".format(host)\n\n print('Send: {!r}'.format(reply))\n writer.write(reply.encode(\"utf8\"))\n\n print('Close the client socket')\n writer.close()\n\nloop = asyncio.get_event_loop()\ncoro = asyncio.start_server(handle_request, '127.0.0.1', 8888, loop=loop)\nserver = loop.run_until_complete(coro)\n\n# Serve requests until CTRL+c is pressed\nprint('Serving on {}'.format(server.sockets[0].getsockname()))\ntry:\n loop.run_forever()\nexcept KeyboardInterrupt:\n pass\n\n# Close the server\nserver.close()\nloop.run_until_complete(server.wait_closed())\nloop.close()\n\n", "sub_path": "httpserver.py", "file_name": "httpserver.py", "file_ext": "py", "file_size_in_byte": 1913, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "http.server.BaseHTTPRequestHandler", "line_number": 6, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 8, "usage_type": "call"}, {"api_name": "asyncio.wait_for", "line_number": 24, "usage_type": "call"}, {"api_name": "asyncio.TimeoutError", "line_number": 25, "usage_type": "attribute"}, {"api_name": "asyncio.coroutine", "line_number": 18, "usage_type": "attribute"}, {"api_name": "asyncio.get_event_loop", "line_number": 55, "usage_type": "call"}, {"api_name": "asyncio.start_server", "line_number": 56, "usage_type": "call"}]}
+{"seq_id": "470602223", "text": "#!/usr/bin/env python3\nimport argparse, subprocess, sys, ue4cli\n\n# Parse our command-line arguments\nparser = argparse.ArgumentParser()\nparser.add_argument(\"--upload\", default=None, help=\"Upload built package to the specified remote\")\nargs = parser.parse_args()\n\n# Query ue4cli for the UE4 version string\nue4 = ue4cli.UnrealManagerFactory.create()\nversionFull = ue4.getEngineVersion()\nversionShort = ue4.getEngineVersion('short')\nversionMinor = int(ue4.getEngineVersion('minor'))\n\n# Verify that the detected version of UE4 is new enough\nif versionMinor < 19:\n print('Error: UE4Capture requires Unreal Engine 4.19 or newer, detected version {}.'.format(versionFull), file=sys.stderr)\n sys.exit(1)\n\n# Build the Conan package, using the short (major.minor) Engine version as the channel name\nchannel = versionShort\nif subprocess.call([\"conan\", \"create\", \".\", \"adamrehn/{}\".format(channel), \"--profile\", \"ue4\"]) != 0:\n sys.exit(1)\n\n# Upload the package to the specified remote if the user provided one\nif args.upload != None:\n if subprocess.call([\"conan\", \"upload\", \"MediaIPC-ue4/*@adamrehn/*\", \"--all\", \"--confirm\", \"-r\", args.upload]) != 0:\n sys.exit(1)\n", "sub_path": "recipes/MediaIPC/build.py", "file_name": "build.py", "file_ext": "py", "file_size_in_byte": 1170, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 5, "usage_type": "call"}, {"api_name": "ue4cli.UnrealManagerFactory.create", "line_number": 10, "usage_type": "call"}, {"api_name": "ue4cli.UnrealManagerFactory", "line_number": 10, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 17, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 18, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 22, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 23, "usage_type": "call"}, {"api_name": "subprocess.call", "line_number": 27, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 28, "usage_type": "call"}]}
+{"seq_id": "409497424", "text": "import json\nfrom requests.exceptions import ConnectionError\nfrom schema import Schema, And, Use, Optional\nimport requests\nfrom os.path import expanduser\nimport os\n\nvul_schema = Schema(\n {\n 'name': str,\n 'tool': str,\n 'description': str,\n 'project': str,\n 'target': str,\n 'scan': str,\n Optional('cwe'): And(Use(int)),\n Optional('observation'): str,\n Optional('severity'): And(Use(int), lambda n: 0 <= n <= 3),\n Optional('remediation'): str,\n Optional('evidences'): list\n },\n ignore_extra_keys=False\n)\n\nevidence_schema = Schema(\n {\n 'name': str,\n 'url': str,\n 'vulnId': str,\n Optional('param'): str,\n Optional('log'): str,\n Optional('attack'): str,\n Optional('otherInfo'): str,\n Optional('evidence'): str,\n Optional('data'): str\n },\n ignore_extra_keys=False\n)\n\n\ndef threatplaybook_con(threatplaybook):\n \"\"\"\n :param threatplaybook: URL of ThreatPlaybook API Server.\n :return: If Connection to http://threat-playbook/graph is successful, returns True.\n \"\"\"\n try:\n r = requests.get(url='{}/graph'.format(threatplaybook))\n if r.status_code == 200:\n return True\n else:\n return False\n except ConnectionError:\n return False\n\n\ndef config_file():\n \"\"\"\n Creates a ThreatPlaybook config file to store Authorization Token\n :return: Path of config file\n \"\"\"\n # directory = expanduser(path='~/.threatplaybook')\n config_file_path = 'config'\n open(config_file_path,'a').close()\n return config_file_path\n\n\ndef clean_string(string):\n \"\"\"\n Cleans the string before creating a GraphQL query. `\\n`, `\\\\`, `\\r`, etc.. cause issues.\n :param string: String that needs to be formatted\n :return: Formatted string compatible with GraphQL Query.\n \"\"\"\n cleaned_string = str(string).replace(\"\\n\", \" \").replace('\"', \" \").replace(\"\\\\\", '').replace(\"\\r\", \"\")\n return cleaned_string\n\n\ndef _post_req(url, email, password):\n \"\"\"\n Posts non-graphql requests to ThreatPlaybook API Server. Currently used for:\n * Create User\n * Login\n :param url: ThreatPlaybook API URL\n :param email: E-mail of the User\n :param password: Password set by the User\n :return: Returns with response\n \"\"\"\n headers = {'content-type': 'application/json'}\n auth = {\"email\": email, \"password\": password}\n try:\n r = requests.post(url=url, headers=headers, data=json.dumps(auth))\n if r.status_code == 500:\n return {'error': 'Server Error'}\n return r.json()\n except ConnectionError:\n return {'error': 'Unable to contact Threatplaybook API server'}\n\n\ndef _post_query(**kwargs):\n \"\"\"\n Posts GraphQL requests to ThreatPlaybook API Server.\n :param threatplaybook: URL of ThreatPlaybook API Server\n :param token: authorization token\n :param query: GraphQL Query\n :return: Returns with response\n \"\"\"\n threatplaybook = kwargs.get('threatplaybook')\n token = kwargs.get('token')\n query = kwargs.get('query')\n url = '{}/graph'.format(threatplaybook)\n if token:\n headers = {'content-type': 'application/json', 'authorization': token}\n try:\n r = requests.post(url=url, json={'query': query}, headers=headers)\n return r.json()\n except Exception as e:\n return {'error': e.message}\n else:\n return {'error': 'Token not found in config file'}\n\n\ndef create_scan(target):\n \"\"\"\n Creates CreateScan mutation query\n :param target: Target name\n :return: CreateScan mutation query\n \"\"\"\n create_scan_query = \"\"\"\n mutation {\n createScan(target: \"%s\") {\n scan {\n name\n createdOn\n }\n }\n }\n \"\"\" % target\n return create_scan_query\n\n\ndef create_evidence(evidence):\n \"\"\"\n Validates evidence dictionary and creates CreateVulnerabilityEvidence mutation query\n :param evidence: Evidence dictionary of a Vulnerability\n :return: CreateVulnerabilityEvidence mutation query\n \"\"\"\n valid_evidence = evidence_schema.validate(evidence)\n create_evidence_query = \"\"\"\n mutation {\n createVulnerabilityEvidence(\n evidence: {\n name: \"%s\"\n log: \"%s\"\n data: \"%s\"\n url: \"%s\"\n param: \"%s\"\n attack: \"%s\"\n evidence: \"%s\"\n otherInfo: \"%s\"\n vulnId: \"%s\"\n })\n {\n vulnEvidence {\n name\n }\n }\n } \n \"\"\" % (valid_evidence.get('name', 'Unknown'),\n valid_evidence.get('log'),\n valid_evidence.get('data'),\n valid_evidence.get('url'),\n valid_evidence.get('param'),\n valid_evidence.get('attack'),\n valid_evidence.get('evidence'),\n valid_evidence.get('otherInfo'),\n valid_evidence.get('vulnId'))\n return create_evidence_query\n\n\ndef create_vulnerability(vul_dict):\n \"\"\"\n Validates vulnerability dictionary and creates CreateVulnerability mutation query\n :param vul_dict: Vulnerability dictionary\n :return: CreateVulnerability mutation query\n \"\"\"\n valid_vul = vul_schema.validate(vul_dict)\n create_vulnerability_query = \"\"\"\n mutation {\n createVulnerability(\n vuln: {\n name: \"%s\"\n tool: \"%s\"\n description: \"%s\"\n project: \"%s\"\n target: \"%s\"\n scan: \"%s\"\n cwe: %d\n observation: \"%s\"\n severity: %d\n remediation: \"%s\"\n })\n {\n vulnerability {\n name\n id\n }\n }\n }\n \"\"\" % (valid_vul.get('name', 'Unknown'),\n valid_vul.get('tool'),\n valid_vul.get('description', ''),\n valid_vul.get('project'),\n valid_vul.get('target'),\n valid_vul.get('scan'),\n valid_vul.get('cwe', 0),\n valid_vul.get('observation', ''),\n valid_vul.get('severity', 0),\n valid_vul.get('remediation', ''),)\n return create_vulnerability_query\n", "sub_path": "robot/threat_playbook/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 6421, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "schema.Schema", "line_number": 8, "usage_type": "call"}, {"api_name": "schema.Optional", "line_number": 16, "usage_type": "call"}, {"api_name": "schema.Optional", "line_number": 17, "usage_type": "call"}, {"api_name": "schema.Optional", "line_number": 18, "usage_type": "call"}, {"api_name": "schema.Optional", "line_number": 19, "usage_type": "call"}, {"api_name": "schema.Optional", "line_number": 20, "usage_type": "call"}, {"api_name": "schema.And", "line_number": 16, "usage_type": "call"}, {"api_name": "schema.Use", "line_number": 16, "usage_type": "call"}, {"api_name": "schema.And", "line_number": 18, "usage_type": "call"}, {"api_name": "schema.Use", "line_number": 18, "usage_type": "call"}, {"api_name": "schema.Schema", "line_number": 25, "usage_type": "call"}, {"api_name": "schema.Optional", "line_number": 30, "usage_type": "call"}, {"api_name": "schema.Optional", "line_number": 31, "usage_type": "call"}, {"api_name": "schema.Optional", "line_number": 32, "usage_type": "call"}, {"api_name": "schema.Optional", "line_number": 33, "usage_type": "call"}, {"api_name": "schema.Optional", "line_number": 34, "usage_type": "call"}, {"api_name": "schema.Optional", "line_number": 35, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 47, "usage_type": "call"}, {"api_name": "requests.exceptions.ConnectionError", "line_number": 52, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 90, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 90, "usage_type": "call"}, {"api_name": "requests.exceptions.ConnectionError", "line_number": 94, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 113, "usage_type": "call"}]}
+{"seq_id": "29088729", "text": "#!/usr/bin/env python3\n#-*-coding:Utf-8 -*\n\n#!/usr/bin/env python3\n#-*-coding:Utf-8 -*\n\n\"\"\"ms_make_fa_adc_groups.\n\ncreate a 2 csv file with the adc and fa means of either the \nnon placebo gor the placebo group, for each time pf each patient.\n\nthose file will look like:\n\npatient;LESIONS;DAWM;NAWM\nXXXX;X;X;X\nXXXX;X;X;X\nXXXX;X;X;X...\n\nYou can specified a maximum number of processus to use\nwith the --nbproc option. The default value 0 is equivalent\nto use all the available CPU core.\n\nGiven paths have to fit have to contains directories that fit:\n\n 'root/Center-XX/Patient-XX/VX/'\n\nA groups.csv file must be in the root directories from where your launching the script. \nIt have to fit:\n\nXX-XX;1\nXX-XX;2\nXX-XX;1...\n\nUsage:\n ms_make_fa_adc_groups \n [ --nbproc= | -p ]\n ...\n\n ms_make_fa_adc_groups -h | --help\n ms_make_fa_adc_groups --version\n\nOptions:\n -h --help\n Show this screen.\n --version\n Show version.\n -p , --nbproc \n Number of processus to create [default: 0]\n\"\"\"\n\nimport os\nimport re\nimport csv\nimport shutil\n\ndef make_groups(patient_directory):\n center = re.sub(r\".*Center-(?P\\d\\d).*\", r\"\\g\", patient_directory)\n patient = re.sub(r\".*Patient-(?P\\d\\d)$\", r\"\\g\", patient_directory)\n\n group = \"0\"\n patient_group_id = center + \"-\" + patient\n\n with open(\"groups.csv\", newline='') as f:\n c = csv.reader(f, delimiter=\";\")\n for row in c:\n m = re.search(center + \"-\" + patient, row[0])\n if m:\n group = row[1]\n # dbg\n print(patient_group_id, group)\n\n for root, dirs, files, in os.walk(patient_directory):\n for d in sorted(dirs):\n m = re.search(r\"(V\\d).*\", d)\n if not m:\n continue\n\n study_time = m.group(1)\n pID = center + patient\n prefix_file_name = pID + study_time + \"_\"\n\n ADC_FA_MEANS = os.path.join(patient_directory, d, prefix_file_name + \"ADC_FA_MEANS.csv\")\n ADC_MEANS = os.path.join(\"groups\", \"ADC_MEANS_\" + study_time +\"_GROUP\" + group + \".csv\")\n FA_MEANS = os.path.join(\"groups\", \"FA_MEANS_\" + study_time +\"_GROUP\" + group + \".csv\")\n\n with open(ADC_FA_MEANS, newline='') as f:\n c = csv.reader(f, delimiter=';')\n rows = list(c)\n\n with open(ADC_MEANS, 'a', newline='') as f:\n c = csv.writer(f, delimiter=';') \n c.writerow([pID,\n rows[0][1],\n rows[1][1],\n rows[2][1],])\n \n with open(FA_MEANS, 'a', newline='') as f:\n c = csv.writer(f, delimiter=';') \n c.writerow([pID,\n rows[3][1],\n rows[4][1],\n rows[5][1],])\n\n\n#--Main\nfrom docopt import docopt\nfrom concurrent.futures import ProcessPoolExecutor\n\narguments = docopt(__doc__, version='ms_patient_process 1.0')\n\ntry:\n nbproc = arguments['--nbproc']\n nbproc = int(nbproc)\n assert nbproc >= 0\nexcept (ValueError, AssertionError):\n print(\"--nbproc must be an positive integer.\\nDefault value (number of CPU core) will be used.\\n\")\n nbproc = 0\n\nif nbproc == 0:\n nbproc = None\n\n# Get all the patient dirs and put them in a list.\npatient_dirs = list()\nfor path in arguments['']:\n for root, dirs, files in os.walk(path):\n if re.search(r\".*Center-\\d\\d.*Patient-\\d\\d$\", root):\n patient_dirs.append(root)\n\nif os.path.exists(\"groups\"):\n shutil.rmtree(\"groups\")\nos.makedirs(\"groups\")\n\n# Use a pool of nbproc processus to do the work on one each patient\nwith ProcessPoolExecutor(max_workers=nbproc) as executor:\n for patient_dir in patient_dirs:\n executor.submit(make_groups, patient_dir)\n#--End of main", "sub_path": "ms-scripts/ms_make_fa_adc_groups.py", "file_name": "ms_make_fa_adc_groups.py", "file_ext": "py", "file_size_in_byte": 3844, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "re.sub", "line_number": 57, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 58, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 64, "usage_type": "call"}, {"api_name": "re.search", "line_number": 66, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 72, "usage_type": "call"}, {"api_name": "re.search", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "csv.reader", "line_number": 87, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 91, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 98, "usage_type": "call"}, {"api_name": "docopt.docopt", "line_number": 109, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 125, "usage_type": "call"}, {"api_name": "re.search", "line_number": 126, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 129, "usage_type": "call"}, {"api_name": "os.path", "line_number": 129, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 130, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 131, "usage_type": "call"}, {"api_name": "concurrent.futures.ProcessPoolExecutor", "line_number": 134, "usage_type": "call"}]}
+{"seq_id": "484584135", "text": "import pandas as pd\nimport argparse\nimport string\n\nimport nltk\n\nfrom nltk import pos_tag, RegexpParser, ne_chunk\nfrom nltk.chunk import conlltags2tree, tree2conlltags\nfrom nltk.tree import Tree\nfrom nltk.tokenize import word_tokenize, PunktSentenceTokenizer\nfrom nltk.corpus import state_union\nfrom nltk.stem import WordNetLemmatizer, PorterStemmer\n\nclass ExtractTuple:\n def __init__(self):\n self.accepted_puntuations = ['-']\n self.rejected_words_ingredients = ['cup', 'tablespoon', 'teaspoon', 'tablespoons']\n self.lemmatizer = WordNetLemmatizer()\n self.stemmer = PorterStemmer()\n\n def pos_tag(self, text):\n return nltk.pos_tag(text)\n\n def extract_ingredient(self, text):\n # ner = self.ner_tagger.ner_tag(text)\n text = [char for char in text.split(' ') if char != '']\n tags = self.pos_tag(text)\n print()\n print(tags)\n\n i = 0\n ingredient = []\n tuples = []\n continoues = False\n while i < len(tags):\n if tags[i][1].startswith('NN'):\n continoues = True\n if tags[i][0] not in self.rejected_words_ingredients:\n ingredient.append(self.lemmatizer.lemmatize(tags[i][0]))\n # elif tags[i][1] == 'JJ' or tags[i][1] == 'VBP' or tags[i][1] == 'VBN':\n # continoues = True\n # if tags[i][0] not in self.rejected_words_ingredients:\n # ingredient.append(self.lemmatizer.lemmatize(tags[i][0]))\n elif tags[i][0] in self.accepted_puntuations and continoues:\n continoues = True\n if tags[i][0] not in self.rejected_words_ingredients:\n ingredient.append(self.lemmatizer.lemmatize(tags[i][0]))\n else:\n continoues = False\n if ingredient != []:\n ingredient = ' '.join(ingredient)\n if not (ingredient.endswith('ed') or ingredient.endswith('ing') or ingredient in string.punctuation):\n tuples.append(ingredient)\n ingredient = []\n\n i += 1\n\n if ingredient != []:\n ingredient = ' '.join(ingredient)\n if not (ingredient.endswith('ed') or ingredient.endswith('ing') or ingredient in string.punctuation):\n tuples.append(ingredient)\n ingredient = []\n\n return tuples\n\n def extract(self, text, ingredients):\n tuples = []\n text = [char for char in text.split(' ') if char != '']\n tags = self.pos_tag(text)\n\n print(tags)\n\n action = None\n i = 0\n ingredient = []\n continoues = False\n while i < len(tags):\n if tags[i][1] == 'NNP':\n action = tags[i][0]\n continoues = False\n if ingredient != []:\n ingredient = ' '.join(ingredient)\n if action != None:\n tuples.append([action, ingredient])\n ingredient = []\n elif tags[i][1] == 'JJ' or tags[i][1] == 'VBD':\n continoues = True\n ingredient.append(tags[i][0])\n elif tags[i][1].startswith('NN'):\n continoues = True\n ingredient.append(tags[i][0])\n elif tags[i][1].startswith('VB'):\n continoues = False\n if ingredient != []:\n ingredient = ' '.join(ingredient)\n if action != None:\n tuples.append([action, ingredient])\n ingredient = []\n elif tags[i][0] in self.accepted_puntuations and continoues:\n continoues = True\n ingredient.append(tags[i][0])\n else:\n continoues = False\n if ingredient != []:\n ingredient = ' '.join(ingredient)\n if action != None:\n tuples.append([action, ingredient])\n ingredient = []\n\n # print(tags[i], ingredient)\n i += 1\n return tuples\n\n\nparser = argparse.ArgumentParser()\nparser.add_argument('--file', '-f', help='Path to cleaned dataset', default='dataset/processed_recipes.tsv', type=str)\nargs = parser.parse_args()\n\ndf = pd.read_csv(args.file, sep='\\t', lineterminator='\\n')\n\nindex = 1\ntitle = df.title[index]\ningredients = df.ingredients[index].split('|')\ningredients = [ing for ing in ingredients if 'Add' not in ing]\ndirections = df.directions[index]\n\nextractor = ExtractTuple()\n\nfor i in range(len(ingredients)):\n tuples = '|'.join(extractor.extract_ingredient(ingredients[i]))\n print(tuples)\n ingredients[i] = tuples\n\nprint(directions)\n\nprint()\ntuples = extractor.extract(directions, ' '.join(ingredients))\n\nfor tup in tuples:\n print(tup)\n", "sub_path": "tuple_extractor.py", "file_name": "tuple_extractor.py", "file_ext": "py", "file_size_in_byte": 4862, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "nltk.stem.WordNetLemmatizer", "line_number": 18, "usage_type": "call"}, {"api_name": "nltk.stem.PorterStemmer", "line_number": 19, "usage_type": "call"}, {"api_name": "nltk.pos_tag", "line_number": 22, "usage_type": "call"}, {"api_name": "string.punctuation", "line_number": 52, "usage_type": "attribute"}, {"api_name": "string.punctuation", "line_number": 60, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 115, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 119, "usage_type": "call"}]}
+{"seq_id": "441418115", "text": "from django.shortcuts import render\nfrom django.http import JsonResponse\nfrom apps.index import models\nimport time\nimport datetime\n\n\n# Create your views here.\n\ndef show(request):\n if request.method == 'GET':\n return render(request,'vip/vip.html')\n elif request.method == 'POST':\n vipdays = request.POST.get('vipdays','')\n user_id = request.POST.get('user_id','')\n if not all([vipdays,user_id]):\n return JsonResponse({'msg':0,'errormsg':'信息不完整或有误!'})\n try:\n vipuser_obj = models.Users.objects.get(userid=user_id)\n # 当前用户的viptime时间戳 小于当前 说明vip已到期\n # 当前用户的viptime时间戳 为空 说明vip未开通过\n # vip 到期时间 = 当前时间+充值的天数\n if (not vipuser_obj.viptime) or float(vipuser_obj.viptime) < time.time() :\n viptime = delay_time(time_num=time.time(),days=vipdays)\n else:\n # vip 到期时间 = 当前vip 到期时间+ 充值的天数\n viptime = delay_time(time_num=vipuser_obj.viptime,days=vipdays)\n vipuser_obj.viptime = viptime\n vipuser_obj.save(force_update=True)\n return render(request,'vip/vip.html',{\n 'jscode': 'function showSuccessMessages(){swal(\"充值成功!\", \"VIP时间已充值成功\", \"success\");}showSuccessMessages()',\n })\n except Exception as e:\n print(e)\n return render(request,'vip/vip.html',{\n 'jscode': 'function showSuccessMessages(){swal(\"充值失败!\", \"请联系管理员处理\", \"error\");}showSuccessMessages()',\n })\n\n\n#时间戳处理函数\ndef delay_time(time_num,days):\n '''\n :param time_num: 时间戳\n :param days: 要推迟的天数\n :return: 推迟之后时间戳\n '''\n try:\n new_viptime = str(datetime.datetime.fromtimestamp(float(time_num)) + datetime.timedelta(days=int(days)))[:-7]\n time_tuple = time.strptime(new_viptime, '%Y-%m-%d %H:%M:%S')\n viptime = str(time.mktime(time_tuple))\n except Exception as e:\n print(e)\n new_viptime = str(datetime.datetime.fromtimestamp(float(time_num)) + datetime.timedelta(days=int(days)))\n time_tuple = time.strptime(new_viptime, '%Y-%m-%d %H:%M:%S')\n viptime = str(time.mktime(time_tuple))\n return viptime", "sub_path": "blog_admin/apps/vip/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2430, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.shortcuts.render", "line_number": 12, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 17, "usage_type": "call"}, {"api_name": "apps.index.models.Users.objects.get", "line_number": 19, "usage_type": "call"}, {"api_name": "apps.index.models.Users", "line_number": 19, "usage_type": "attribute"}, {"api_name": "apps.index.models", "line_number": 19, "usage_type": "name"}, {"api_name": "time.time", "line_number": 23, "usage_type": "call"}, {"api_name": "time.time", "line_number": 24, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 30, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 35, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 48, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 48, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 48, "usage_type": "call"}, {"api_name": "time.strptime", "line_number": 49, "usage_type": "call"}, {"api_name": "time.mktime", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 53, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 53, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 53, "usage_type": "call"}, {"api_name": "time.strptime", "line_number": 54, "usage_type": "call"}, {"api_name": "time.mktime", "line_number": 55, "usage_type": "call"}]}
+{"seq_id": "363102644", "text": "# create the mask for the pilatus2M\n# import stuff to make plotting easier (NOT advised to do for actual\n# deployment)\nimport os.path\nimport numpy as np\n\nfrom PIL import Image\nfrom matplotlib.pyplot import ion, imshow, figure, clf, clim\n\n# read some data via databroker\nfrom SciStreams.interfaces.databroker.databases import databases\n\n# this is a gui\nfrom SciStreams.tools.MaskCreator import MaskCreator\n\nfrom SciStreams.config import mask_config\n\nfrom SciStreams.data.Mask import MasterMask, MaskGenerator\nfrom SciStreams.detectors.mask_generators import generate_mask\n\n# interactive mode\nion()\n\ncmsdb = databases['cms:data']\n\n# choose the detector key\ndetector_key = \"pilatus2M_image\"\n\n# some optional filters to limit searches\ncmsdb.add_filter(start_time=\"2017-09-13\", stop_time=\"2017-09-14\")\nhdrs = (list(cmsdb(sample_name=\"AgBH_Sep13_2017_JH_test\")))\n\nimgs = cmsdb.get_images(hdrs, detector_key)\n\n# 0 : GISAXS , 1 : SAXS\nind = 0\nstartdoc = hdrs[ind]['start']\nimg = imgs[ind]\n\nmsk = MaskCreator(data=img)\nmsk.set_clim(0, 10)\n\nmsg = \"Use the Mask GUI. Press Enter Here when done\"\nmsg += \" and code will continue\"\nmsg += \"\\n Type resume() when done.\"\n#input(msg)\n\ndef resume():\n mask = msk.mask\n\n ''' There are a few components to this:\n 1. Get the detector positions during the creation of this mask\n 2. Save the origin of detector in mask image. Here it's 0,0 since\n mask is the same shape. However, it can be other if the mask arises\n from a lager stitched image.\n 3. Set the scl of the pixels (needed when transforming pixel to lab\n coordinates for detector)\n 4. Determine what parameters define this mask. For example:\n bs_phi : the phi orientation of the beam stop. If this changed, then\n the mask is no longer valid\n bsx, bsy : beamstop x, y positions\n maybe positions of other obstructions?\n sample-detector distance : the shadow will change with these\n 5. Finally, save all this into a npz file.\n\n The mask will then be loaded by a MasterMask object, which willl be loaded\n by a MaskGenerator object. The MaskGenerator object takes a y,x coordinate\n to generate a future mask. This y, x coordinate is the detector lab\n position.\n\n '''\n # now get detector position\n refmotorx = startdoc['motor_SAXSx']\n refmotory = startdoc['motor_SAXSy']\n\n refpoint_lab = refmotory, refmotorx\n # set this ref point to the 0, 0 coordinate\n refpoint = 0, 0\n # pilatus pixel conversion\n scl = .172, .172\n # prepare the filename\n #filename = os.path.expanduser(\"mask_pilatus2M_master_1.npz\")\n for i in range(1000):\n fname = \"~/research/projects/SciAnalysis-data\"\n fname = fname + \"/masks/pilatus2M_image/mask_pilatus2M_master_{}.npz\".format(i)\n filename = os.path.expanduser(fname)\n if not os.path.isfile(filename):\n break\n kwargs = dict()\n # kwargs.update(startdoc)\n kwargs['mask'] = mask\n kwargs['refpoint'] = refpoint\n kwargs['refpoint_lab'] = refpoint_lab\n kwargs['scl'] = scl\n # the motor positions used to define the mask\n kwargs['motor_bsphi'] = startdoc['motor_bsphi']\n kwargs['motor_bsx'] = startdoc['motor_bsx']\n kwargs['motor_bsy'] = startdoc['motor_bsy']\n kwargs['detector_SAXS_distance_m'] = startdoc['detector_SAXS_distance_m']\n # saving here (uncomment)\n # saving is done hwere\n np.savez(filename, **kwargs)\n\n ''' As a test, we can try reading the mask '''\n # now read\n detector_key = 'pilatus2M_image'\n fname = \"~/research/projects/SciAnalysis-data\"\n fname = fname + \"/masks/pilatus2M_image/mask_pilatus2M_master_1.npz\"\n filename = os.path.expanduser(fname)\n # test that it loads fine\n master_mask = MasterMask(filename)\n blem_fname = mask_config[detector_key]['blemish']['filename']\n blemish = np.array(Image.open(blem_fname))\n\n mmg = MaskGenerator(master_mask, blemish)\n mask = mmg.generate(-72.99992548, -65.00001532)\n\n ''' As a further test, we can test that the SciStreams library is also reading\n it properly'''\n # GISAXS\n md = dict()\n md.update(**startdoc)\n md.update(detector_key=detector_key)\n\n mask = generate_mask(**md)['mask']\n\n # plot them to see they make sense\n figure(2)\n clf()\n imshow(mask)\n clim(0, 1)\n\n figure(3)\n clf()\n imshow(img)\n clim(0, 100)\n", "sub_path": "SciStreams/examples/mask_creation_simple_gisaxs.py", "file_name": "mask_creation_simple_gisaxs.py", "file_ext": "py", "file_size_in_byte": 4431, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "matplotlib.pyplot.ion", "line_number": 22, "usage_type": "call"}, {"api_name": "SciStreams.interfaces.databroker.databases.databases", "line_number": 24, "usage_type": "name"}, {"api_name": "SciStreams.tools.MaskCreator.MaskCreator", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path.path.expanduser", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 86, "usage_type": "name"}, {"api_name": "os.path.path.isfile", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 87, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 87, "usage_type": "name"}, {"api_name": "numpy.savez", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path.path.expanduser", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 109, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 109, "usage_type": "name"}, {"api_name": "SciStreams.data.Mask.MasterMask", "line_number": 111, "usage_type": "call"}, {"api_name": "SciStreams.config.mask_config", "line_number": 112, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 113, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 113, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 113, "usage_type": "name"}, {"api_name": "SciStreams.data.Mask.MaskGenerator", "line_number": 115, "usage_type": "call"}, {"api_name": "SciStreams.detectors.mask_generators.generate_mask", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clim", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clim", "line_number": 136, "usage_type": "call"}]}
+{"seq_id": "97784368", "text": "import requests\nimport json\nfrom pprint import pprint\nfrom pymongo import MongoClient, GEO2D\nimport time\nimport dateutil.parser\nimport urllib2\n\n\n\natlas = MongoClient('mongodb+srv://dbLina:isen2020@cluster0.io2qf.mongodb.net/BicycleStation?retryWrites=true&w=majority')\n\ndb = atlas.bicycle\n\n#Indexes \ndb.datas.create_index([('station_id', 1),('date', -1)], unique=True)\ndb.stations1.create_index([('geometry','2dsphere')]) \n\n# def get_user_lat_lon():\n# try:\n# return json.load(urllib2.urlopen('http://ipinfo.io/json'))\n# except urllib2.HTTPError:\n# return False\n# coord = get_user_lat_lon()\n\ndef get_station_id(id_ext):\n tps = db.station1.find_one({ 'source.id_ext':id_ext }, { '_id': 1 })\n return tps['_id']\n\ndef get_vlille():\n url = \"https://opendata.lillemetropole.fr/api/records/1.0/search/?dataset=vlille-realtime&q=&rows=-1&facet=libelle&facet=nom&facet=commune&facet=etat&facet=type&facet=etatconnexion\"\n response = requests.request(\"GET\", url)\n response_json = json.loads(response.text.encode('utf8'))\n return response_json.get(\"records\", [])\n\n\n#inp = input(' Enter 0 for manual coordinates input, enter 1 for auto-geolocalisation : ')\n\n#if inp==0:\nprint('Entrer latitude: ')\nlat = input()\nprint('Entrer longitude: ')\nlon = input()\nprint('Your position is: ', (lat,lon))\n \n# elif inp==1:\n# coord = get_user_lat_lon()\n# print 'You are in '+ coord['city']+ ' ('+ coord['loc'] + ')'\n\n\ndef get_nearest_station(lat,lon):\n nearest= db.stations1.find({'geometry': { \n '$near': { '$geometry': {\n 'type': \"Point\" ,\n 'coordinates': [ lat, lon ]},\n '$maxDistance': 100,\n '$minDistance': 0 }\n }})\n return(list(nearest))\n \n\nprint(get_nearest_station(lat, lon))\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "user_program.py", "file_name": "user_program.py", "file_ext": "py", "file_size_in_byte": 1752, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pymongo.MongoClient", "line_number": 11, "usage_type": "call"}, {"api_name": "requests.request", "line_number": 32, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 33, "usage_type": "call"}]}
+{"seq_id": "568452269", "text": "import ast\nimport constant\nimport csv\nimport multiprocessing as mp\nimport logging\nimport outlierdet\nimport stats\nimport similarity\n\nlogging.basicConfig()\nlogger = logging.getLogger(__name__)\nlogger.setLevel(logging.INFO)\n\n\n##TODO: incorparate website addresses\n\n## final file----------------------------------\n\nclass GC_Record:\n def __init__(self, id, address, parish, keyword, type, geocode, gt_geocode,\n distance=0, category=0, multi_gc_type=0, outlier_list=()):\n \"\"\"A class to represent a geocoded address.\n\n Possible values for 'category' are:\n\n 0 addresses with no geocodes\n 1 addresses that match exactly and you get one exact location in OSM\n 2 addresses that match exactly and you get several locations in OSM\n 3 addresses where you get an approximate string match to addresses in 1\n 4 addresses where you get an approximate string match to addresses in 2\n\n Possible values for 'multi_gc_type' are:\n\n 0 not applicable\n 1 priority typed single address\n 2 invalid multigc address set\n 3 direct averaging multi gc address set\n 4 outliers address set\n\n \"\"\"\n self.id = id\n self.address = address\n self.parish = parish\n self.keyword = keyword\n self.type = type\n self.geocode = geocode\n self.gt_geocode = gt_geocode\n self.category = category\n self.distance = distance\n self.multi_gc_type = multi_gc_type\n self.outlier_list = outlier_list\n\n\n# =============================================================================\n\ndef retrieve_finnish_parishes(parish_file, village_file):\n parish_dict = {}\n with open(parish_file, 'r') as data:\n reader = csv.DictReader(data)\n for line in reader:\n parish_dict[line['id']] = line['name']\n\n village_parish_dict = {}\n with open(village_file, 'r') as data:\n reader = csv.DictReader(data)\n for line in reader:\n village_parish_dict[line['id']] = parish_dict[line['parish']]\n\n return village_parish_dict\n\n\n# =============================================================================\n\ndef retrieve_data(input_file, id, parish_file='', village_file=''):\n id = constant.ID\n # Parish retrieval for Finnish dataset\n parishes_dict = {}\n if constant.FINNISH:\n parishes_dict = retrieve_finnish_parishes(parish_file, village_file)\n\n record_dict = {}\n with open(input_file, 'r') as data:\n reader = csv.DictReader(data)\n\n for line in reader:\n\n if line[id] not in record_dict.keys():\n record_dict[line[id]] = list()\n\n # ground truth retrieval\n gt_gc = ''\n if 'gt_geocode' in line.keys():\n gt_gc = line['gt_geocode']\n\n # Parish retrieval\n parish = ''\n if constant.FINNISH and line['id'] in parishes_dict.keys():\n parish = parishes_dict[line['id']]\n # IOS dataset has parishes stated as source\n # not incorporating parishes to ios\n # elif not constant.FINNISH:\n\n r = GC_Record(line[id], line['address'].strip(), parish,\n line['keyword'], line['type'], line['geocode'], gt_gc)\n record_dict[line[id]].append(r)\n\n logging.info(' Record size of {} is {}'\n .format(input_file, str(len(record_dict))))\n return record_dict\n\n\n# =============================================================================\n\ndef retrieve_unique_records(record_dict):\n unique_record_list = list()\n if constant.PARALLEL:\n pool = mp.Pool(processes=constant.PARALLEL_PROCESSES)\n for id, record_list in record_dict.items():\n # unique_record_list.append(parallel_code(id, record_list, constant.T))\n pool.apply_async(parallel_code,\n args=(id, record_list, constant.T),\n callback=unique_record_list.append)\n pool.close()\n pool.join()\n else:\n for id, record_list in record_dict.items():\n unique_record_list.append(parallel_code(id, record_list, constant.T))\n logging.info('Unique record list size {}'.format(len(unique_record_list)))\n\n return unique_record_list\n\n\n# =============================================================================\n\ndef parallel_code(id, record_list, types_priority_list):\n # logging.info(\n # ' Processing birth record {} {}'.format(id, record_list[0].address))\n\n # If a single geocode is available or no geocodes available\n if len(record_list) == 1:\n if record_list[0].geocode != '':\n # logging.info(' Single geocode of type {}'.format(record_list[0].type))\n record_list[0].category = 1\n # else:\n # logging.info(' No geocodes found')\n return record_list[0]\n\n # If multiple geocodes are available\n else:\n # logging.info(' Multiple geocodes found')\n\n # Check if records exist in priority list\n priority_typed_record_list = list()\n for type in types_priority_list:\n for record in record_list:\n if type == record.type:\n priority_typed_record_list.append(record)\n if len(priority_typed_record_list) != 0:\n break\n\n # If no records are of priority types, get the average of records\n if len(priority_typed_record_list) == 0:\n # logging.info(' No geocodes from prioritized types')\n return process_multiple_geocodes(record_list)\n\n # If only one records is from priority types, get the particular record\n elif len(priority_typed_record_list) == 1:\n # logging.info(' Single geocode from prioritized types {}'\n # .format(priority_typed_record_list[0].type))\n priority_typed_record_list[0].category = 2\n priority_typed_record_list[0].multi_gc_type = 1\n return priority_typed_record_list[0]\n\n # If more than one valid types are available, get their average\n else:\n # logging.info(' Multiple geocodes from prioritized types')\n return process_multiple_geocodes(priority_typed_record_list)\n\n\n# =============================================================================\n\ndef process_multiple_geocodes(record_list):\n \"\"\"Return\n\n ARGUMENTS:\n record_list Record list to be averaged\n avg_stat_dict Dictionary holding stats of records\n avg_gc_dict Dictionary to keep track of unique addresses\n\n DESCRIPTION:\n Aqqqqqqqqqqqqqqqqqqqqqqq\n \"\"\"\n address = record_list[0].address\n parish = record_list[0].parish\n # If the address string is not yet averaged\n\n geocode_list = list()\n tmp_gc = set() # To make the gc list unique\n for record in record_list:\n if record.geocode in tmp_gc:\n continue\n tmp_gc.add(record.geocode)\n geocode_list.append(ast.literal_eval(record.geocode))\n\n geocode_list, multi_gc_type, outlier_list = outlierdet. \\\n retrieve_geocodes_outliers_removed(geocode_list, (address + '_p_ ' +\n parish),\n constant.OUTLIER_FUNCTION)\n if multi_gc_type == 0:\n logging.info('tadaaaaaaaaaaaaaaaaaa')\n if len(geocode_list) == 0:\n return GC_Record(record_list[0].id, address, parish, '', '', '', '', 0, 0,\n multi_gc_type, outlier_list)\n\n lat_list = list()\n lng_list = list()\n for geocode in geocode_list:\n lat_list.append(geocode[0])\n lng_list.append(geocode[1])\n lat_val = sum(lat_list) / len(lat_list)\n lng_val = sum(lng_list) / len(lng_list)\n\n # logging.info(' Resulting lat lng {}'.format(str([lat_val, lng_val])))\n gc = str([lat_val, lng_val])\n\n return GC_Record(record_list[0].id, address, parish, 'avg', 'avg', gc,\n record_list[0].gt_geocode, 0, 2, multi_gc_type, outlier_list)\n\n\n# =============================================================================\n\ndef validate_mispelled_values(unique_record_list, output_file):\n non_geocoded_record_dict = {}\n geocoded_record_dict = {}\n\n for record in unique_record_list:\n if record.geocode == '':\n if record.address != '':\n non_geocoded_record_dict[record.address] = None\n stats.initial_empty_address_set.add(record.address + record.parish)\n else:\n geocoded_record_dict[record.address] = record\n\n similarity.match_similar_strings_parallel(geocoded_record_dict,\n non_geocoded_record_dict)\n similarity.gazetteer_matches(non_geocoded_record_dict)\n\n w = csv.writer(open(output_file, \"w\"))\n w.writerow(['address_str', 'matching_string',\n 'similarity_score'])\n\n for address, geocoded_similar_record in non_geocoded_record_dict.items():\n if geocoded_similar_record == None:\n w.writerow([address, '', ''])\n else:\n w.writerow([address, geocoded_similar_record[0].address,\n geocoded_similar_record[1]])\n\n # Assigning approximated values\n for record in unique_record_list:\n\n # Processing geocode empty records\n if record.geocode == '' and record.address != '':\n\n # Check if the parishes are the same before assigning\n if non_geocoded_record_dict[record.address] != None and (\n record.parish == non_geocoded_record_dict[record.address][0].parish):\n\n record.keyword = non_geocoded_record_dict[record.address][0].address\n record.geocode = non_geocoded_record_dict[record.address][0].geocode\n record.type = non_geocoded_record_dict[record.address][0].type\n\n # Setting the geocode category\n if non_geocoded_record_dict[record.address][0].category == 1:\n record.category = 3\n elif non_geocoded_record_dict[record.address][0].category == 2:\n record.category = 4\n else:\n raise Exception('A geocoded address should be either 1 or 2')\n\n # Stats\n stats.prediction_address_set.add(record.address + record.parish)\n else:\n stats.no_prediction_address_set.add(record.address + record.parish)\n\n\n# =============================================================================\n\ndef analyze_results(unique_record_list, name=''):\n empty_address_count = 0\n\n empty_gc_count = 0\n empty_set = set()\n\n count_exactly_single = 0\n count_exactly_single_set = set()\n count_exactly_multiple = 0\n count_exactly_multiple_set = set()\n count_approx_from_exactly_single = 0\n count_approx_from_exactly_single_set = set()\n count_approx_from_exactly_multiple = 0\n count_approx_from_exactly_multiple_set = set()\n\n unique_address_set = set()\n\n # Assigning approximated values and analyzing the results\n for record in unique_record_list:\n if record.geocode != '' and record.gt_geocode != '':\n record.distance = outlierdet.dist_fun(ast.literal_eval(record.geocode),\n ast.literal_eval(record.gt_geocode))\n\n unique = (record.address + record.parish)\n if record.address == '':\n empty_address_count += 1\n if record.geocode == '':\n empty_gc_count += 1\n empty_set.add(unique)\n elif record.category == 1:\n count_exactly_single += 1\n count_exactly_single_set.add(unique)\n elif record.category == 2:\n count_exactly_multiple += 1\n count_exactly_multiple_set.add(unique)\n elif record.category == 3:\n count_approx_from_exactly_single += 1\n count_approx_from_exactly_single_set.add(unique)\n elif record.category == 4:\n count_approx_from_exactly_multiple += 1\n count_approx_from_exactly_multiple_set.add(unique)\n\n logging.info('-------- GENERIC STATS ----------')\n logging.info('Unique address count {}'.format(len(stats.unique_address_set)))\n logging.info('Initial No gc address count {}'.format(len(\n stats.no_gc_address_set)))\n logging.info('Initial Single gc address count {}'.format(len(\n stats.single_gc_address_set)))\n logging.info('Initial Multiple gc address count {}'.format(len(\n stats.multiple_gc_address_set)))\n logging.info('Single word address count {}'.format(len(\n stats.single_word_address)))\n logging.info('Multi word address count {}'.format(len(\n stats.multi_word_address)))\n\n logging.info('-------- OUTLIER STATS UNIQUE----------')\n logging.info('Priority typed single address count {}'\n .format(len(stats.priority_typed_single_address_set)))\n logging.info('Invalid multiple goecode address count {}'\n .format(len(stats.invalid_multigc_address_set)))\n logging.info('Direct averaging geocode count {}'\n .format(len(stats.direct_averaging_multigc_address_set)))\n logging.info('Address count with Outliers {}'\n .format(len(stats.outliers_address_set)))\n\n logging.info('--------STRING PREDICTION STATS----------')\n logging.info(\n 'After outliers unique empty locations {}'.format(\n len(stats.initial_empty_address_set)))\n logging.info('Unique no predictions {}'.format(len(\n stats.no_prediction_address_set)))\n logging.info('Unique predictions {}'.format(len(\n stats.prediction_address_set)))\n\n logging.info('-------- FINAL STATS----------')\n logging.info('Empty addresses in dataset {}'.format(empty_address_count))\n logging.info('Final addresses with empty geocodes {}'.format(empty_gc_count))\n logging.info(\n 'Locations with exactly single geocode {}'.format(count_exactly_single))\n logging.info('Locations with exactly multiple geocode {}'.format(\n count_exactly_multiple))\n logging.info('Locations approximated from exactly single geocode {}'.format(\n count_approx_from_exactly_single))\n logging.info('Locations approximated from exactly multiple geocode {}'.format(\n count_approx_from_exactly_multiple))\n\n logging.info('Unique Final empty locations {}'.format(len(empty_set)))\n logging.info(\n 'Unique Locations with exactly single geocode {}'.format(\n len(count_exactly_single_set)))\n logging.info('Unique Locations with exactly multiple geocode {}'.format(\n len(count_exactly_multiple_set)))\n logging.info(\n 'Unique Locations approximated from exactly single geocode {}'.format(\n len(count_approx_from_exactly_single_set)))\n logging.info(\n 'Unique Locations approximated from exactly multiple geocode {}'.format(\n len(count_approx_from_exactly_multiple_set)))\n\n if constant.FINNISH:\n stats.analyze_groundtruth(unique_record_list)\n\n\n# =============================================================================\n\ndef write_to_file(unique_record_list, output_file):\n w = csv.writer(open(output_file, \"w\"))\n w.writerow(['id', 'location', 'keyword', 'type', 'geocode', 'category'])\n for record in unique_record_list:\n w.writerow([record.id, record.address, record.keyword, record.type,\n record.geocode, record.category])\n\n\n# =============================================================================\n\ndef retreive_rule_based_gc(input_file, parish_file='', village_file=''):\n output_file1 = 'out/sim/village' + constant.get_name_ext() + '.csv'\n output_file2 = 'out/gc/village' + constant.get_name_ext() + '.csv'\n\n logging.info('----------------------------{}'.format(constant.get_name_ext()))\n\n stats.initialize()\n id = constant.ID\n record_dict = retrieve_data(input_file, id, parish_file, village_file)\n stats.retrieve_stats_before_outlier(record_dict)\n unique_record_list = retrieve_unique_records(record_dict)\n stats.retrieve_stats_after_outlier(unique_record_list)\n validate_mispelled_values(unique_record_list, output_file1)\n analyze_results(unique_record_list)\n write_to_file(unique_record_list, output_file2)\n\n\ndef set_finnish(is_finnish):\n if is_finnish:\n constant.LLCRNRLON = 19.0\n constant.LLCRNRLAT = 59.3\n constant.URCRNRLON = 31.60\n constant.URCRNRLAT = 63.2\n # constant.URCRNRLAT = 70.2\n constant.FINNISH = True\n constant.T = ['hamlet', 'village', 'suburb', 'yes', 'residential']\n constant.ID = 'id'\n\n else:\n constant.LLCRNRLON = -6.8\n constant.LLCRNRLAT = 57.0\n constant.URCRNRLON = -5.60\n constant.URCRNRLAT = 57.75\n constant.FINNISH = False\n constant.T = ['hamlet', 'village', 'yes', 'residential']\n constant.ID = 'iosbirth_identifier'\n\nif (__name__ == '__main__'):\n logging.info('Main')\n #### osm source - types_priority_list = ['hamlet','residential', 'yes', 'town']\n #### arcgis - types_priority_list = ['Locality','POI']\n\n set_finnish(False)\n input_file = '../_data_20190704/osm_address_0705.csv'\n retreive_rule_based_gc('../_data_20190704/osm_address_0705.csv')\n\n # set_finnish(True)\n # input_file = '../_data_20190704/osm_village_0709_wp.csv'\n # parish_file = '../_data_raw/finnish/ascii/parish-ascii.csv'\n # village_file = '../_data_raw/finnish/ascii/village-ascii.csv'\n # retreive_rule_based_gc(input_file, parish_file, village_file)\n", "sub_path": "Programs/Python/new_gc/rule_based/geocode_retrieval.py", "file_name": "geocode_retrieval.py", "file_ext": "py", "file_size_in_byte": 16468, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.basicConfig", "line_number": 10, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 12, "usage_type": "attribute"}, {"api_name": "csv.DictReader", "line_number": 59, "usage_type": "call"}, {"api_name": "csv.DictReader", "line_number": 65, "usage_type": "call"}, {"api_name": "constant.ID", "line_number": 75, "usage_type": "attribute"}, {"api_name": "constant.FINNISH", "line_number": 78, "usage_type": "attribute"}, {"api_name": "csv.DictReader", "line_number": 83, "usage_type": "call"}, {"api_name": "constant.FINNISH", "line_number": 97, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 107, "usage_type": "call"}, {"api_name": "constant.PARALLEL", "line_number": 116, "usage_type": "attribute"}, {"api_name": "multiprocessing.Pool", "line_number": 117, "usage_type": "call"}, {"api_name": "constant.PARALLEL_PROCESSES", "line_number": 117, "usage_type": "attribute"}, {"api_name": "constant.T", "line_number": 121, "usage_type": "attribute"}, {"api_name": "constant.T", "line_number": 127, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 128, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 203, "usage_type": "call"}, {"api_name": "outlierdet.retrieve_geocodes_outliers_removed", "line_number": 205, "usage_type": "call"}, {"api_name": "constant.OUTLIER_FUNCTION", "line_number": 208, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 210, "usage_type": "call"}, {"api_name": "stats.initial_empty_address_set.add", "line_number": 240, "usage_type": "call"}, {"api_name": "stats.initial_empty_address_set", "line_number": 240, "usage_type": "attribute"}, {"api_name": "similarity.match_similar_strings_parallel", "line_number": 244, "usage_type": "call"}, {"api_name": "similarity.gazetteer_matches", "line_number": 246, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 248, "usage_type": "call"}, {"api_name": "stats.prediction_address_set.add", "line_number": 282, "usage_type": "call"}, {"api_name": "stats.prediction_address_set", "line_number": 282, "usage_type": "attribute"}, {"api_name": "stats.no_prediction_address_set.add", "line_number": 284, "usage_type": "call"}, {"api_name": "stats.no_prediction_address_set", "line_number": 284, "usage_type": "attribute"}, {"api_name": "outlierdet.dist_fun", "line_number": 309, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 309, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 310, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 331, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 332, "usage_type": "call"}, {"api_name": "stats.unique_address_set", "line_number": 332, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 333, "usage_type": "call"}, {"api_name": "stats.no_gc_address_set", "line_number": 334, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 335, "usage_type": "call"}, {"api_name": "stats.single_gc_address_set", "line_number": 336, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 337, "usage_type": "call"}, {"api_name": "stats.multiple_gc_address_set", "line_number": 338, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 339, "usage_type": "call"}, {"api_name": "stats.single_word_address", "line_number": 340, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 341, "usage_type": "call"}, {"api_name": "stats.multi_word_address", "line_number": 342, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 344, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 345, "usage_type": "call"}, {"api_name": "stats.priority_typed_single_address_set", "line_number": 346, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 347, "usage_type": "call"}, {"api_name": "stats.invalid_multigc_address_set", "line_number": 348, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 349, "usage_type": "call"}, {"api_name": "stats.direct_averaging_multigc_address_set", "line_number": 350, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 351, "usage_type": "call"}, {"api_name": "stats.outliers_address_set", "line_number": 352, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 354, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 355, "usage_type": "call"}, {"api_name": "stats.initial_empty_address_set", "line_number": 357, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 358, "usage_type": "call"}, {"api_name": "stats.no_prediction_address_set", "line_number": 359, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 360, "usage_type": "call"}, {"api_name": "stats.prediction_address_set", "line_number": 361, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 363, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 364, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 365, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 366, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 368, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 370, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 372, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 375, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 376, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 379, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 381, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 384, "usage_type": "call"}, {"api_name": "constant.FINNISH", "line_number": 388, "usage_type": "attribute"}, {"api_name": "stats.analyze_groundtruth", "line_number": 389, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 395, "usage_type": "call"}, {"api_name": "constant.get_name_ext", "line_number": 405, "usage_type": "call"}, {"api_name": "constant.get_name_ext", "line_number": 406, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 408, "usage_type": "call"}, {"api_name": "constant.get_name_ext", "line_number": 408, "usage_type": "call"}, {"api_name": "stats.initialize", "line_number": 410, "usage_type": "call"}, {"api_name": "constant.ID", "line_number": 411, "usage_type": "attribute"}, {"api_name": "stats.retrieve_stats_before_outlier", "line_number": 413, "usage_type": "call"}, {"api_name": "stats.retrieve_stats_after_outlier", "line_number": 415, "usage_type": "call"}, {"api_name": "constant.LLCRNRLON", "line_number": 423, "usage_type": "attribute"}, {"api_name": "constant.LLCRNRLAT", "line_number": 424, "usage_type": "attribute"}, {"api_name": "constant.URCRNRLON", "line_number": 425, "usage_type": "attribute"}, {"api_name": "constant.URCRNRLAT", "line_number": 426, "usage_type": "attribute"}, {"api_name": "constant.FINNISH", "line_number": 428, "usage_type": "attribute"}, {"api_name": "constant.T", "line_number": 429, "usage_type": "attribute"}, {"api_name": "constant.ID", "line_number": 430, "usage_type": "attribute"}, {"api_name": "constant.LLCRNRLON", "line_number": 433, "usage_type": "attribute"}, {"api_name": "constant.LLCRNRLAT", "line_number": 434, "usage_type": "attribute"}, {"api_name": "constant.URCRNRLON", "line_number": 435, "usage_type": "attribute"}, {"api_name": "constant.URCRNRLAT", "line_number": 436, "usage_type": "attribute"}, {"api_name": "constant.FINNISH", "line_number": 437, "usage_type": "attribute"}, {"api_name": "constant.T", "line_number": 438, "usage_type": "attribute"}, {"api_name": "constant.ID", "line_number": 439, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 442, "usage_type": "call"}]}
+{"seq_id": "566083893", "text": "\"\"\"empty message\n\nRevision ID: 5d3ded72de8\nRevises: f59bfd5312e\nCreate Date: 2015-11-30 13:51:40.741708\n\n\"\"\"\n\n# revision identifiers, used by Alembic.\nrevision = '5d3ded72de8'\ndown_revision = 'f59bfd5312e'\n\nfrom alembic import op\nimport sqlalchemy as sa\n\n\ndef upgrade():\n ### commands auto generated by Alembic - please adjust! ###\n op.add_column('user', sa.Column('active', sa.Boolean(), nullable=True))\n op.add_column('user', sa.Column('staff', sa.Boolean(), nullable=True))\n ### end Alembic commands ###\n\n\ndef downgrade():\n ### commands auto generated by Alembic - please adjust! ###\n op.drop_column('user', 'staff')\n op.drop_column('user', 'active')\n ### end Alembic commands ###\n", "sub_path": "migrations/versions/5d3ded72de8_.py", "file_name": "5d3ded72de8_.py", "file_ext": "py", "file_size_in_byte": 708, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "alembic.op.add_column", "line_number": 19, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 19, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 19, "usage_type": "call"}, {"api_name": "sqlalchemy.Boolean", "line_number": 19, "usage_type": "call"}, {"api_name": "alembic.op.add_column", "line_number": 20, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 20, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 20, "usage_type": "call"}, {"api_name": "sqlalchemy.Boolean", "line_number": 20, "usage_type": "call"}, {"api_name": "alembic.op.drop_column", "line_number": 26, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 26, "usage_type": "name"}, {"api_name": "alembic.op.drop_column", "line_number": 27, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 27, "usage_type": "name"}]}
+{"seq_id": "336749775", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('fileapi', '0001_initial'),\n ]\n\n operations = [\n migrations.AlterField(\n model_name='fileinfo',\n name='file_id',\n field=models.AutoField(serialize=False, primary_key=True),\n ),\n ]\n", "sub_path": "fileapi/migrations/0002_auto_20150820_0703.py", "file_name": "0002_auto_20150820_0703.py", "file_ext": "py", "file_size_in_byte": 414, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}]}
+{"seq_id": "534278760", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n# CircuitPython\n\n# SPDX-FileCopyrightText: 2021 s-light\n# SPDX-License-Identifier: MIT\n# Author Stefan Krüger (s-light)\n\n\"\"\"TLC5971 / TLC59711 Test BCData.\"\"\"\n\n__doc__ = \"\"\"\ntlc59711_test_bcdata.py.\n\ntest brightness correction data (BC)\n\"\"\"\n\nimport time\n\nimport board\nimport busio\nimport supervisor\n\nimport adafruit_tlc59711\n\n##########################################\nPIXEL_COUNT = 16 * 8\n\nspi = busio.SPI(board.SCK, MOSI=board.MOSI)\npixels = adafruit_tlc59711.TLC59711(spi, pixel_count=PIXEL_COUNT)\n\n##########################################\n\n\ndef main_loop():\n \"\"\"Loop.\"\"\"\n new_value = input()\n if \"v\" in new_value:\n try:\n value = int(new_value[1:])\n except ValueError as e:\n print(\"Exception: \", e)\n pixels.set_pixel_all_16bit_value(value, value, value)\n else:\n Ioclmax, IoutR, IoutG, IoutB = (18, 18, 11, 13)\n try:\n Ioclmax, IoutR, IoutG, IoutB = new_value.split(\";\")\n Ioclmax = float(Ioclmax)\n IoutR = float(IoutR)\n IoutG = float(IoutG)\n IoutB = float(IoutB)\n except ValueError as e:\n print(\"Exception: \", e)\n BCValues = adafruit_tlc59711.TLC59711.calculate_BCData(\n Ioclmax=Ioclmax,\n IoutR=IoutR,\n IoutG=IoutG,\n IoutB=IoutB,\n )\n pixels.bcr = BCValues[0]\n pixels.bcg = BCValues[1]\n pixels.bcb = BCValues[2]\n print(\n \"bcr: {:>3}\\n\"\n \"bcg: {:>3}\\n\"\n \"bcb: {:>3}\\n\"\n \"\".format(\n pixels.bcr,\n pixels.bcg,\n pixels.bcb,\n )\n )\n pixels.update_BCData()\n pixels.show()\n # prepare new input\n print(\"\\nenter new values:\")\n\n\ndef test_main():\n \"\"\"Test Main.\"\"\"\n print(42 * \"*\", end=\"\")\n print(__doc__, end=\"\")\n print(42 * \"*\")\n # print()\n # time.sleep(0.5)\n # print(42 * '*')\n\n print(\"set pixel all to 100, 100, 100\")\n pixels.set_pixel_all((5000, 5000, 5000))\n # calculate bc values\n Ioclmax = adafruit_tlc59711.TLC59711.calculate_Ioclmax(Riref=2.7)\n print(\"Ioclmax = {}\".format(Ioclmax))\n Riref = adafruit_tlc59711.TLC59711.calculate_Riref(Ioclmax=Ioclmax)\n print(\"Riref = {}\".format(Riref))\n BCValues = adafruit_tlc59711.TLC59711.calculate_BCData(\n Ioclmax=Ioclmax,\n IoutR=18,\n IoutG=11,\n IoutB=13,\n )\n # (127, 77, 91)\n print(\"BCValues = {}\".format(BCValues))\n pixels.bcr = BCValues[0]\n pixels.bcg = BCValues[1]\n pixels.bcb = BCValues[2]\n pixels.update_BCData()\n pixels.show()\n time.sleep(0.1)\n\n if supervisor.runtime.serial_connected:\n print(\n \"\\n\"\n \"this script offers two things to be changed:\\n\"\n \"- value for all channels\\n\"\n \"example: 'v10'\\n\"\n \"example: 'v65535'\\n\"\n \"- (global) brightness control:\\n\"\n \"use format: 'Ioclmax; IoutR; IoutG; IoutB'\\n\"\n \"example: '18; 7; 15; 17'\"\n \"\\n\"\n )\n while True:\n if supervisor.runtime.serial_bytes_available:\n main_loop()\n\n\n##########################################\n# main loop\n\nif __name__ == \"__main__\":\n test_main()\n", "sub_path": "examples/tlc59711_test_bcdata.py", "file_name": "tlc59711_test_bcdata.py", "file_ext": "py", "file_size_in_byte": 3299, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "busio.SPI", "line_number": 28, "usage_type": "call"}, {"api_name": "board.SCK", "line_number": 28, "usage_type": "attribute"}, {"api_name": "board.MOSI", "line_number": 28, "usage_type": "attribute"}, {"api_name": "adafruit_tlc59711.TLC59711", "line_number": 29, "usage_type": "call"}, {"api_name": "adafruit_tlc59711.TLC59711.calculate_BCData", "line_number": 53, "usage_type": "call"}, {"api_name": "adafruit_tlc59711.TLC59711", "line_number": 53, "usage_type": "attribute"}, {"api_name": "adafruit_tlc59711.TLC59711.calculate_Ioclmax", "line_number": 90, "usage_type": "call"}, {"api_name": "adafruit_tlc59711.TLC59711", "line_number": 90, "usage_type": "attribute"}, {"api_name": "adafruit_tlc59711.TLC59711.calculate_Riref", "line_number": 92, "usage_type": "call"}, {"api_name": "adafruit_tlc59711.TLC59711", "line_number": 92, "usage_type": "attribute"}, {"api_name": "adafruit_tlc59711.TLC59711.calculate_BCData", "line_number": 94, "usage_type": "call"}, {"api_name": "adafruit_tlc59711.TLC59711", "line_number": 94, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 107, "usage_type": "call"}, {"api_name": "supervisor.runtime", "line_number": 109, "usage_type": "attribute"}, {"api_name": "supervisor.runtime", "line_number": 122, "usage_type": "attribute"}]}
+{"seq_id": "74912037", "text": "#!/usr/bin/env python\nimport argparse\nfrom glob import glob\nimport os\nfrom tqdm.notebook import tqdm\nimport shutil\n\n\ndef create_dir(_path):\n if not os.path.exists(_path):\n os.mkdir(_path)\n\n\ndef merge_dir(_dir):\n for class_name in tqdm(class_names):\n count = 0\n class_dir = os.path.join(dest_dir, class_name)\n create_dir(class_dir)\n source_dir = os.path.join(_dir, class_name)\n src_paths = glob(os.path.join(source_dir, '*.jpg'))\n for src_ipath in tqdm(src_paths):\n dest_ipath = os.path.join(class_dir, f\"{count}.jpg\")\n shutil.copy(src_ipath, dest_ipath)\n count += 1\n src_paths = glob(os.path.join(source_dir, '*.png'))\n for src_ipath in tqdm(src_paths):\n dest_ipath = os.path.join(class_dir, f\"{count}.png\")\n shutil.copy(src_ipath, dest_ipath)\n count += 1\n\n\ndef get_input_args():\n ''' \n 1. Read command line arguments and convert them into the apropriate data type. \n 2. Returns a data structure containing everything that have been read, or the default values \n for the paramater that haven't been explicitly specified.\n '''\n parser = argparse.ArgumentParser()\n \n parser.add_argument('--src_dir', type=str, default=os.path.join(os.getcwd(), 'src'),\n help='The folder of the collected images')\n\n parser.add_argument('--dataset_dir', type=str, default=os.path.join(os.getcwd(), 'dataset'),\n help='The folder of the base dataset')\n\n parser.add_argument('--dest_dir', type=str, default=os.path.join(os.getcwd(), 'data'),\n help='Target folder for the new combined dataset')\n\n in_args = parser.parse_args()\n\n return in_args\n\n\nif __name__ == '__main__':\n args = get_input_args()\n src_dir = args.src_dir if os.path.exists(args.src_dir) else os.path.join(os.getcwd(), 'src')\n dataset_dir = args.dataset_dir if os.path.exists(args.src_dir) else os.path.join(os.getcwd(), 'dataset')\n class_names = os.listdir(dataset_dir)\n print('Found Classes:', class_names)\n dest_dir = args.dest_dir if os.path.exists(args.dest_dir) else os.path.join(os.getcwd(), 'data')\n\n create_dir(dest_dir)\n merge_dir(src_dir)\n merge_dir(dataset_dir)\n\n", "sub_path": "screw_classification/data_join.py", "file_name": "data_join.py", "file_ext": "py", "file_size_in_byte": 2295, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.path.exists", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 11, "usage_type": "call"}, {"api_name": "tqdm.notebook.tqdm", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "tqdm.notebook.tqdm", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 23, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "tqdm.notebook.tqdm", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 28, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 56, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 57, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 57, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 60, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 60, "usage_type": "call"}]}
+{"seq_id": "302609873", "text": "\"\"\"\nThis is the core class for supporting a DTM based around dask-dataframes.\n\nEssentially the key here is to recognize that a sparse matrix can be\nrepresented well by a trio of data-frames with linked indices and the DTM\nclass contains sets of common operations for this class of data.\n\"\"\"\nfrom .DTM_part_methods import *\nfrom dask import delayed\nimport dask.dataframe as dd\nimport pandas as pd\nimport numpy as np\nimport difflib\nimport glob\nimport os\n\n\ndef _prep_fnames(doc_globstring, term_globstring):\n \"\"\"prepares lists of file names to help with writing updates\n\n Parameters\n ----------\n doc_globstring : str\n globstring corresponding to doc_df\n term_globstring : str\n globstring corresponding to term_df\n\n Returns\n -------\n None\n \"\"\"\n\n # TODO note that this currently doesn't \"truly\" find the unique\n # component of each file name. It compares with the globstring and\n # it may be the case that the resulting files still have some shared\n # component (e.g. if globstring is doc_*.csv and all doc files are\n # called doc_id_*.csv), this should be fixed\n\n # get file lists\n doc_flist = glob.glob(doc_globstring)\n term_flist = glob.glob(term_globstring)\n\n doc_flist.sort()\n term_flist.sort()\n\n # get base names\n doc_flist = [os.path.basename(f) for f in doc_flist]\n term_flist = [os.path.basename(f) for f in term_flist]\n\n # extract patterns to populate count_map\n doc_fpat = [\"\".join([r.replace(\"+ \", \"\") for r in\n difflib.ndiff(doc_globstring, f) if \"+\" in r])\n for f in doc_flist]\n term_fpat = [\"\".join([r.replace(\"+ \", \"\") for r in\n difflib.ndiff(term_globstring, f) if \"+\" in r])\n for f in term_flist]\n\n if len(doc_fpat) == 1:\n doc_fpat = None\n if len(term_fpat) == 1:\n term_fpat = None\n\n return doc_fpat, term_fpat\n\n\nclass DTM(object):\n \"\"\"Core class for handling DTM data\n\n Parameters\n ----------\n doc_df : dask-dataframe\n dataframe containing doc metadata and id/index\n term_df : dask-dataframe\n dataframe containing term metadata and id/index\n count_df : dask-dataframe\n dataframe containing counts and doc+term id/index\n doc_index : str\n label for doc id/index\n term_index : str\n label for term id/index\n doc_fpat : list or None\n file pattern stored to produce doc files, will be:\n /doc_[i]\n term_fpat : list or None\n file pattern stored to produce term files, will be:\n /term_[j]\n\n Attributes\n ----------\n doc_df : dask-dataframe\n dataframe containing doc metadata and id/index\n term_df : dask-dataframe\n dataframe containing term metadata and id/index\n count_df : dask-dataframe\n dataframe containing counts and doc+term id/index\n doc_index : str\n label for doc id/index\n term_index : str\n label for term id/index\n doc_fpat : list or None\n file pattern stored to produce doc files, will be:\n /doc_[i]\n term_fpat : list or None\n file pattern stored to produce term files, will be:\n /term_[j]\n\n Notes\n -----\n If file patterns are stored the count files will be\n /count_[i]_[j]\n \"\"\"\n\n def __init__(self, doc_df, term_df, count_df, doc_index, term_index,\n doc_fpat=None, term_fpat=None):\n\n self.doc_df = doc_df\n self.term_df = term_df\n self.count_df = count_df\n\n self.doc_index = doc_index\n self.term_index = term_index\n\n self.doc_fpat = doc_fpat\n self.term_fpat = term_fpat\n\n self.npartitions = (doc_df.npartitions,\n term_df.npartitions)\n\n\n def copy(self):\n \"\"\"generates a copy of the DTM and returns that\n\n Returns\n -------\n copy of current DTM\n \"\"\"\n\n dtm = DTM(doc_df=self.doc_df.copy(),\n term_df=self.term_df.copy(),\n count_df=self.count_df.copy(),\n doc_index=self.doc_index,\n term_index=self.term_index,\n doc_fpat=self.doc_fpat,\n term_fpat=self.term_fpat)\n\n return dtm\n\n\n def persist(self):\n \"\"\"persists the underlying data-frames\n\n Returns\n -------\n copy of current DTM\n \"\"\"\n\n dtm = self.copy()\n\n dtm.doc_df = dtm.doc_df.persist()\n dtm.term_df = dtm.term_df.persist()\n dtm.count_df = dtm.count_df.persist()\n\n return dtm\n\n\n\n def to_csv(self, out_dir=None, doc_urlpath=None,\n term_urlpath=None, count_urlpath=None, **kwargs):\n \"\"\"writes the current DTM to the specified files\n\n Parameters\n ----------\n out_dir : str or None\n location directory where results should be stored\n doc_urlpath : str or None\n urlpath for doc_df\n term_urlpath : str or None\n urlpath for term_df\n count_urlpath : str or None\n urlpath for count_df\n \"\"\"\n\n# self = self.persist()\n\n if not out_dir:\n out_dir = \".\"\n if self.doc_fpat:\n doc_fpat = self.doc_fpat\n else:\n doc_fpat = [(\"%d\" % d).zfill(2) for d in\n range(self.npartitions[0])]\n if self.term_fpat:\n term_fpat = self.term_fpat\n else:\n term_fpat = [(\"%d\" % d).zfill(2) for d in\n range(self.npartitions[1])]\n\n if not doc_urlpath:\n doc_urlpath = [os.path.join(out_dir, \"doc_%s.csv\" % f)\n for f in doc_fpat]\n if not term_urlpath:\n term_urlpath = [os.path.join(out_dir, \"term_%s.csv\" % f)\n for f in term_fpat]\n if not count_urlpath:\n count_urlpath = [os.path.join(out_dir, \"count_%s_%s.csv\" %\n (d_f, t_f))\n for d_f in doc_fpat\n for t_f in term_fpat]\n\n self.doc_df.to_csv(doc_urlpath, index=False, **kwargs)\n self.term_df.to_csv(term_urlpath, index=False, **kwargs)\n self.count_df.to_csv(count_urlpath, index=False, **kwargs)\n\n\n def repartition_doc(self, npartitions):\n \"\"\"repartition the DTM along the doc axis\n\n Parameters\n ----------\n npartitions : scalar\n number of partitions for new DTM\n\n Returns\n -------\n updated DTM\n\n Notes\n -----\n - Only supports shrinking the number of partitions\n - We attempt to optimize the partitions so that the actual docs\n are spread as evenly as possible. Therefore we base the new\n partitions on the doc counts within the old partitions.\n \"\"\"\n\n dtm = self.copy()\n\n if npartitions > dtm.doc_df.npartitions:\n raise ValueError(\"repartition currently only supports shrinking \\\n the existing number of partitions. Therefore \\\n new npartitions must be less than existing.\")\n\n doc_fn = lambda x: x[[dtm.doc_index]].count()\n doc_count = dtm.doc_df.map_partitions(doc_fn).compute()\n D = doc_count.sum()\n stp = D / npartitions\n doc_cum_count = doc_count.cumsum()\n\n # prep partition info\n partitions = []\n for n in range(npartitions):\n chk = len(doc_cum_count[doc_cum_count < n * stp])\n partitions.append(chk)\n partitions.append(len(doc_cum_count))\n\n term_part = dtm.term_df.npartitions\n\n # init fn\n del_concat = delayed(pd.concat)\n\n # prep delayed data\n doc_del = dtm.doc_df.to_delayed()\n count_del = dtm.count_df.to_delayed()\n\n doc_del_l = []\n count_del_l = []\n\n # collapse partitions\n for n in range(1, npartitions + 1):\n\n n_start = partitions[n-1]\n n_stop = partitions[n]\n doc_del_n = doc_del[n_start:n_stop]\n\n doc_del_l.append(del_concat(doc_del_n))\n\n for t in range(term_part):\n\n t_start = (n_start * term_part + t)\n t_stop = (n_stop * term_part + t)\n t_stp = term_part\n\n count_del_nt = count_del[t_start:t_stop:t_stp]\n count_del_l.append(del_concat(count_del_nt))\n\n dtm.doc_df = dd.from_delayed(doc_del_l)\n dtm.count_df = dd.from_delayed(count_del_l)\n\n # if we are storing fname patterns combine these as well\n if dtm.doc_fpat is not None:\n\n n_doc_fpat = []\n for n in range(1, npartitions + 1):\n n_start = partitions[n-1]\n n_stop = partitions[n]\n\n doc_fpat_part = dtm.doc_fpat[n_start:n_stop]\n if len(doc_fpat_part) > 1:\n n_doc_fpat.append(doc_fpat_part[0] + \"_T_\" +\n doc_fpat_part[-1])\n elif len(doc_fpat_part) == 1:\n n_doc_fpat.append(doc_fpat_part[0])\n else:\n raise ValueError(\"doc_fpat doesn't align with parititons\")\n\n dtm.doc_fpat = n_doc_fpat\n\n dtm.npartitions = (dtm.doc_df.npartitions,\n dtm.term_df.npartitions)\n\n return dtm\n\n\n def repartition_term(self):\n \"\"\"repartitions the DTM along the term axis\n\n Returns\n -------\n updated DTM\n\n Notes\n -----\n This only support collapsing the term partitions to 1\n \"\"\"\n\n dtm = self.copy()\n\n # collapsing terms is simple\n dtm.term_df = dtm.term_df.repartition(npartitions=1)\n\n # now we need to collapse each term partition\n del_concat = delayed(pd.concat)\n\n count_del = dtm.count_df.to_delayed()\n\n Dp, Vp = dtm.npartitions\n\n count_del_l = []\n\n for d in range(Dp):\n\n count_del_nt = count_del[(d * Vp):((d + 1) * Vp)]\n count_del_l.append(del_concat(count_del_nt))\n\n dtm.count_df = dd.from_delayed(count_del_l)\n\n # now reset fpatterns if provided\n# if dtm.term_fpat is not None:\n# dtm.term_fpat = [dtm.term_fpat[0] + \"_T_\" + dtm.term_fpat[1]]\n # since we've collapsed to term counts we shouldn't have a pattern\n # any longer\n dtm.term_fpat = None\n\n # finally reset npartitions\n dtm.npartitions = (dtm.doc_df.npartitions,\n dtm.term_df.npartitions)\n\n return dtm\n\n\n def repartition(self, npartitions=None, axis=\"doc\"):\n \"\"\"repartitions along the provided axis\n\n Parameters\n ----------\n npartitions : scalar or None\n number of new partitions, must not be None if axis == doc\n axis : str\n label for axis over which to repartition\n\n Returns\n -------\n updated DTM\n \"\"\"\n\n if axis == \"doc\":\n if npartitions is None:\n raise ValueError(\"If repartitioning over docs, npartitions \\\n must be provided\")\n else:\n return self.repartition_doc(npartitions)\n\n elif axis == \"term\":\n if npartitions is not None:\n raise ValueError(\"If repartitioning over terms, npartitions \\\n should not be provided\")\n else:\n return self.repartition_term()\n\n\n def map_doc(dtm, func, alt=None, term=False, count=False,\n kwds_l=None, **kwargs):\n \"\"\"maps the provided method over the doc partitions\n\n Parameters\n ----------\n func : python function\n method to apply to each doc partition\n alt : str or None\n if provided, this corresponds to a dask-df with comparable\n partitions to doc_df (currently contained in kwargs), the\n underlying df for each partition will be passed into func\n along with doc_df\n term : bool\n whether or not the term_df partitions will be passed\n into func along with doc_df (as a list)\n count : bool\n whether or not the comparable count_df partitions will be passed\n into func along with doc_df (as a list)\n kwds_l : list or None\n if provided, this corresponds to a list of key-words which\n are partition specific, therefore it should be the same length\n as the term partitions\n\n Returns\n -------\n updated DTM\n\n Notes\n -----\n We assume that the doc partitions are bigger than the term\n partitions, therefore the outer loop will be handled by dask and\n any additional partitions (over counts or terms) will be passed in\n as lists of values\n \"\"\"\n\n dtm = dtm.copy()\n\n Dp, Vp = dtm.npartitions\n\n if kwds_l is None:\n kwds_l = [{} for i in range(Dp)]\n if len(kwds_l) != Dp:\n raise ValueError(\"kwds_l needs to be same length as doc \\\n partitions\")\n\n if alt is not None:\n alt_df = kwargs.pop(alt)\n alt_del = alt_df.to_delayed()\n\n doc_del = dtm.doc_df.to_delayed()\n term_del = dtm.term_df.to_delayed()\n count_del = dtm.count_df.to_delayed()\n\n del_l = []\n del_fn = delayed(func)\n\n for i, doc_i in enumerate(doc_del):\n if alt is not None:\n kwargs[alt] = alt_del[i]\n if term:\n kwargs[\"term_df\"] = term_del\n if count:\n kwargs[\"count_df\"] = count_del[(i*Vp):((i+1)*Vp)]\n kwds = {**kwds_l[i], **kwargs}\n del_l.append(del_fn(doc_i, **kwds))\n\n dtm.doc_df = dd.from_delayed(del_l)\n\n return dtm\n\n\n def map_term(self, func, alt=None, doc=False, count=False,\n kwds_l=None, **kwargs):\n \"\"\"maps the provided method over the term partitions\n\n Parameters\n ----------\n func : python function\n method to apply to each term partition\n alt : str or None\n if provided, this corresponds to a dask-df with comparable\n partitions to term_df (currently contained in kwargs), the\n underlying df for each partition will be passed into func\n along with term_df\n doc : bool\n whether or not the doc_df partitions will be passed\n into func along with term_df\n count : bool\n whether or not the comparable count_df partitions will be passed\n into func along with term_df\n kwds_l : list or None\n if provided, this corresponds to a list of key-words which\n are partition specific, therefore it should be the same length\n as the term partitions\n\n Returns\n -------\n updated DTM\n\n Notes\n -----\n We assume that the term partitions are smaller than the doc\n partitions so if we need doc or count info for func, we will do\n the outer loop over term partitions without using delayed and then\n reserve delayed/parallel estimation for within the function\n \"\"\"\n\n dtm = self.copy()\n\n Dp, Vp = dtm.npartitions\n\n if kwds_l is None:\n kwds_l = [{} for i in range(Vp)]\n if len(kwds_l) != Vp:\n raise ValueError(\"kwds_l needs to be same length as term \\\n partitions\")\n\n if alt is not None:\n alt_df = kwargs.pop(alt)\n alt_del = alt_df.to_delayed()\n\n term_del = dtm.term_df.to_delayed()\n\n del_l = []\n\n # if we don't need the counts and docs we can do this completely\n # in parallel, otherwise we need to save the parallel component\n # for the inner loop\n if not (doc or count):\n for i, term_i in enumerate(term_del):\n if alt is not None:\n kwargs[alt] = alt_del[i]\n kwds = {**kwds_l[i], **kwargs}\n del_l.append(delayed(func)(term_i, **kwds))\n\n else:\n\n # TODO there likely is a better way to do this part...\n del_fn = delayed(lambda x: x)\n\n for i, term_i in enumerate(term_del):\n if alt is not None:\n kwargs[alt] = alt_del[i]\n if doc:\n kwargs[\"doc_df\"] = dtm.doc_df\n if count:\n tmp = dtm.count_df.partitions[i:(Dp*Vp):Vp]\n kwargs[\"count_df\"] = tmp\n kwds = {**kwds_l[i], **kwargs}\n term_i = func(term_i, **kwds)\n del_l.append(del_fn(term_i))\n\n dtm.term_df = dd.from_delayed(del_l)\n\n return dtm\n\n\n def map_count(self, func, alt=None, alt_doc=None, alt_term=None,\n doc=False, term=False, kwds_l=None, **kwargs):\n \"\"\"maps the provided method over the count partitions\n\n Parameters\n ----------\n func : python function\n method to apply to each count partition\n alt : str or None\n if provided, this corresponds to a dask-df with comparable\n partitions to count_df (currently contained in kwargs), the\n underlying df for each partition will be passed into func\n along with count_df\n alt_doc : str or None\n if provided, this corresponds to a dask-df with comparable\n partitions to doc_df (currently contained in kwargs), the\n underlying df for each partition will be passed into func\n along with count_df\n alt_term : str or None\n if provided, this corresponds to a dask-df with comparable\n partitions to term_df (currently contained in kwargs), the\n underlying df for each partition will be passed into func\n along with count_df\n doc : bool\n whether or not the comparable doc_df partition will be passed\n into func along with count_df\n term : bool\n whether or not the comparable term_df partition will be passed\n into func along with count_df\n kwds_l : list or None\n if provided, this corresponds to a list of key-words which\n are partition specific, therefore it should be the same length\n as the count partitions\n\n Returns\n -------\n updated DTM\n \"\"\"\n\n dtm = self.copy()\n\n Dp, Vp = dtm.npartitions\n\n if kwds_l is None:\n kwds_l = [{} for i in range(Dp * Vp)]\n if len(kwds_l) != (Dp * Vp):\n raise ValueError(\"kwds_l needs to be same length as count \\\n partitions\")\n\n # prep delayed values for additional metadata\n if alt is not None:\n alt_df = kwargs.pop(alt)\n alt_del = alt_df.to_delayed()\n if alt_doc is not None:\n alt_doc_df = kwargs.pop(alt_doc)\n alt_doc_del = alt_doc_df.to_delayed()\n if alt_term is not None:\n alt_term_df = kwargs.pop(alt_term)\n alt_term_del = alt_term_df.to_delayed()\n\n doc_del = dtm.doc_df.to_delayed()\n term_del = dtm.term_df.to_delayed()\n count_del = dtm.count_df.to_delayed()\n\n del_l = []\n del_fn = delayed(func)\n\n for i, doc_i in enumerate(doc_del):\n for j, term_j in enumerate(term_del):\n q = (i * Vp) + j\n if alt is not None:\n kwargs[alt] = alt_del[q]\n if alt_doc is not None:\n kwargs[alt_doc] = alt_doc_del[i]\n if alt_term is not None:\n kwargs[alt_term] = alt_term_del[j]\n if doc:\n kwargs[\"doc_df\"] = doc_i\n if term:\n kwargs[\"term_df\"] = term_j\n kwds = {**kwds_l[q], **kwargs}\n del_l.append(del_fn(count_del[q], **kwds))\n\n dtm.count_df = dd.from_delayed(del_l)\n\n return dtm\n\n\n def map_partitions(self, func, axis=\"doc\", **kwargs):\n \"\"\"wrapper function for generally mapping over partitions along\n a provided axis\n\n Parameters\n ----------\n func : python funct\n method to apply to corresponding axis over the partitions\n axis : str\n label for axis to map\n\n Returns\n -------\n updated DTM\n \"\"\"\n\n if axis == \"doc\":\n return self.map_doc(func, **kwargs)\n elif axis == \"term\":\n return self.map_term(func, **kwargs)\n elif axis == \"count\":\n return self.map_count(func, **kwargs)\n\n\n def intersect_index(self, doc=False, term=False, count=False):\n \"\"\"intersects the specified axes to populate any updates which might\n drop observations\n\n Parameters\n ----------\n doc : bool\n whether to constrain doc_df based on counts\n term : bool\n whether to constrain term_df based on counts\n count : bool\n whether to constrain count_df based on doc_df and term_df\n\n Returns\n -------\n updated DTM\n \"\"\"\n\n dtm = self.copy()\n\n # first subset dfs based on updates\n if count:\n dtm = dtm.map_partitions(intersect_count, axis=\"count\", doc=True,\n term=True, doc_index=dtm.doc_index,\n term_index=dtm.term_index)\n if doc:\n dtm = dtm.map_partitions(intersect_doc, axis=\"doc\", count=True,\n doc_index=dtm.doc_index)\n if term:\n dtm = dtm.map_partitions(intersect_term, axis=\"term\", count=True,\n term_index=self.term_index)\n\n return dtm\n\n\n def reset_index(self):\n \"\"\"resets each index, this should be run after any series of\n operations which may change the data in DTM where the index matters\n\n Returns\n -------\n updated DTM\n \"\"\"\n\n # TODO currently we have to touch the metadata twice, ideally we\n # should only have to touch it once here, but that is going to require\n # that the we can map functions which return multiple values...\n\n dtm = self.copy()\n\n # now generate cumulative counts to share new df size between\n # partitions\n doc_fn = lambda x: x[[dtm.doc_index]].count()\n doc_count = dtm.doc_df.map_partitions(doc_fn).cumsum()\n term_fn = lambda x: x[[dtm.term_index]].count()\n term_count = dtm.term_df.map_partitions(term_fn).cumsum()\n\n # now reset\n dtm = dtm.map_partitions(reset_ind_count, axis=\"count\", doc=True,\n term=True, alt_doc=\"doc_count\",\n alt_term=\"term_count\", doc_count=doc_count,\n term_count=term_count,\n doc_index=dtm.doc_index,\n term_index=dtm.term_index)\n dtm = dtm.map_partitions(reset_ind_mdata, axis=\"doc\",\n alt=\"mdata_count\", mdata_count=doc_count,\n mdata_index=dtm.doc_index)\n dtm = dtm.map_partitions(reset_ind_mdata, axis=\"term\",\n alt=\"mdata_count\", mdata_count=term_count,\n mdata_index=dtm.term_index)\n\n return dtm\n\n\n def add(self, new_dtm, **kwargs):\n \"\"\"sum two DTMs with same doc and terms\n\n Parameters\n ----------\n new_dtm : DTM instance\n another DTM which we assume has comparable dimensions/partitions\n to the current one\n\n Returns\n -------\n updated DTM\n \"\"\"\n\n # TODO this currently doesn't really compare dataframes\n\n # all metadata must be the same between DTMs to add\n if len(self.doc_df) != len(new_dtm.doc_df):\n raise ValueError(\"main and new DTM must share same doc_df\")\n elif len(self.term_df) != len(new_dtm.term_df):\n raise ValueError(\"main and new DTM must share same term_df\")\n\n kwargs[\"new_count\"] = new_dtm.count_df\n return self.map_partitions(add_part, alt=\"new_count\",\n axis=\"count\", doc_index=self.doc_index,\n term_index=self.term_index, **kwargs)\n\n\n def sort(self):\n \"\"\"sorts the DTM according to doc and term ids\n\n Returns\n -------\n updated sorted DTM\n \"\"\"\n\n dtm = self.copy()\n\n doc_fn = lambda x: x.sort_values(self.doc_index)\n dtm.doc_df = dtm.doc_df.map_partitions(doc_fn)\n term_fn = lambda x: x.sort_values(self.term_index)\n dtm.term_df = dtm.term_df.map_partitions(term_fn)\n count_fn = lambda x: x.sort_values([self.doc_index, self.term_index])\n dtm.count_df = dtm.count_df.map_partitions(count_fn)\n\n return dtm\n\n\n def merge(self, new_df, axis=\"doc\", **kwargs):\n \"\"\"merge another dask-dataframe along the specified axis\n\n Parameters\n ----------\n new_df : dask data-frame\n new dask-dataframe which we wish to merge to one axis\n axis : str\n label for axis which we wish to merge new_df\n\n Returns\n -------\n updated DTM\n \"\"\"\n\n if axis == \"doc\":\n comp_df = self.doc_df\n lab = \"doc_df\"\n elif axis == \"term\":\n comp_df = self.term_df\n lab = \"term_df\"\n elif axis == \"count\":\n comp_df = self.count_df\n lab = \"count_df\"\n if comp_df.npartitions != new_df.npartitions:\n raise ValueError(\"new_df and %s must have same npartitions\" % lab)\n\n kwargs[\"new_df\"] = new_df\n\n dtm = self.map_partitions(merge_part, alt=\"new_df\", axis=axis,\n **kwargs)\n\n return dtm\n\n\n def sample(self, n=None, frac=None, axis=\"doc\", full_sample=True,\n **kwargs):\n \"\"\"sample from provided axis\n\n Parameters\n ----------\n n : int, optional\n Number of items from axis to return. Cannot be used with frace.\n Default = 1 if frac = None\n frac : float, optional\n Fraction of axis items to return. Cannot be used with n\n axis : str\n label for axis which we wish to merge new_df\n full_index : bool\n whether to sample from each partition indvidually or as a whole\n\n Returns\n -------\n updated DTM\n \"\"\"\n\n if axis == \"doc\":\n comp_df = self.doc_df\n lab = \"doc_df\"\n elif axis == \"term\":\n comp_df = self.term_df\n lab = \"term_df\"\n elif axis == \"count\":\n comp_df = self.count_df\n lab = \"count_df\"\n\n # if we do full sample then we need to prep n for each partition\n if full_sample:\n if n and frac:\n raise ValueError(\"n and frac can't both be provided\")\n elif n:\n n = int(n / self.npartitions[0])\n elif frac:\n if frac < 0 or frac > 1:\n raise ValueError(\"frac must be between 0 and 1\")\n else:\n n = int(frac * len(comp_df))\n kwds_l = [{\"n\": n} for i in range(comp_df.npartitions)]\n else:\n kwargs[\"n\"] = n\n kwargs[\"frac\"] = frac\n\n dtm = self.map_partitions(sample_part, axis=axis, kwds_l=kwds_l,\n **kwargs)\n\n # NOTE we need to persist here because sample is stochastic, if we\n # don't persist then future operations may get different results\n dtm = dtm.persist()\n\n return dtm\n\n\n def ttpartition(self, CV_partitions=None, out_dir=None,\n out_dir_pattern=None):\n \"\"\"partitions the DTM into a list of tuples where each tuple is\n a training/testing partition pair\n\n Parameters\n ----------\n CV_partitions : None or scalar\n if provided, we first repartition the data and then the list\n will be of length CV_partitions, otherwise CV_partitions is\n just the current number of partitions\n out_dir : None or str\n we write each partition (instead of returning the list) if\n provided\n out_dir_pattern : None or str\n if provided, this is used as a pattern to fill in test and CV\n labels, otherwise we just use %s_CV%d % (type, part)\n\n Returns\n -------\n list of DTMs or None\n \"\"\"\n\n # if a CV_partition count isn't provided, just partition over the\n # docs\n if CV_partitions is None:\n CV_partitions = self.npartitions[0]\n dtm = self\n else:\n dtm = self.copy()\n dtm = dtm.repartition(CV_partitions)\n\n if out_dir_pattern is None:\n out_dir_pattern = \"%s_CV%d\"\n\n res = []\n\n Dp, Vp = dtm.npartitions\n\n for part in range(CV_partitions):\n\n # prep train_dtm\n cpart_ind = list(range(Dp * Vp))\n dpart_ind = list(range(Dp))\n\n ctrain_ind = cpart_ind[:(part * Vp)] + cpart_ind[((part+1) * Vp):]\n ctest_ind = cpart_ind[(part * Vp):((part+1) * Vp)]\n dtrain_ind = dpart_ind[:part] + dpart_ind[(part+1):]\n dtest_ind = dpart_ind[part:(part+1)]\n\n train_dtm = dtm.copy()\n train_dtm.count_df = train_dtm.count_df.partitions[ctrain_ind]\n train_dtm.doc_df = train_dtm.doc_df.partitions[dtrain_ind]\n if train_dtm.doc_fpat is not None:\n train_dtm.doc_fpat = (train_dtm.doc_fpat[:part] +\n train_dtm.doc_fpat[(part+1):])\n train_dtm.npartitions = (Dp - 1, Vp)\n train_dtm = train_dtm.reset_index()\n\n # prep test_dtm\n test_dtm = dtm.copy()\n test_dtm.count_df = test_dtm.count_df.partitions[ctest_ind]\n test_dtm.doc_df = test_dtm.doc_df.partitions[dtest_ind]\n if test_dtm.doc_fpat is not None:\n test_dtm.doc_fpat = test_dtm.doc_fpat[part:(part+1)]\n test_dtm.npartitions = (1, Vp)\n test_dtm = testm_dtm.reset_index()\n\n # store output\n if out_dir:\n\n train_dir = os.path.join(out_dir,\n out_dir_pattern % (\"train\", part))\n test_dir = os.path.join(out_dir,\n out_dir_pattern % (\"test\", part))\n\n os.makedirs(train_dir, exist_ok=True)\n os.makedirs(test_dir, exist_ok=True)\n\n train_dtm.to_csv(out_dir=train_dir)\n test_dtm.to_csv(out_dir=test_dir)\n\n else:\n res.append((train_dtm, test_dtm))\n\n if len(res) > 0:\n return res\n\n\ndef read_csv(in_dir=None, doc_globstring=None, term_globstring=None,\n count_globstring=None, doc_flist=None, term_flist=None,\n count_flist=None, doc_index=\"doc_id\", term_index=\"term_id\",\n keep_fname=True, blocksize=None, **kwargs):\n \"\"\"reads the csvs for each partition and populates DTM\n\n Parameters\n ----------\n in_dir : str or None\n if provided, we assume that all the files in this directory\n correspond to a DTM and populate globstrings accordingly\n doc_globstring : str or None\n globstring for doc_df files\n term_globstring : str or None\n globstring for term_df files\n count_globstring : str or None\n globstring for count files\n doc_flist : list or None\n list for doc_df files\n term_list : list or None\n list for term_df files\n count_list : list or None\n list for count files\n doc_index : str\n label for doc axis index\n term_index : str\n label for term axis index\n keep_fname : bool\n whether to keep a record of the filename patterns (for writing\n updated DTM)\n blocksize : scalar or None\n blocksize for dask dfs. Given that we want our partitions to align\n we default this to None (so each partition corresponds to a file)\n\n Returns\n -------\n populated DTM object\n \"\"\"\n\n if not in_dir:\n in_dir = \".\"\n if not doc_flist:\n if not doc_globstring:\n doc_globstring = os.path.join(in_dir, \"doc_*.csv\")\n doc_flist = glob.glob(doc_globstring)\n if not term_flist:\n if not term_globstring:\n term_globstring = os.path.join(in_dir, \"term_*.csv\")\n term_flist = glob.glob(term_globstring)\n if not count_flist:\n if not count_globstring:\n count_globstring = os.path.join(in_dir, \"count_*.csv\")\n count_flist = glob.glob(count_globstring)\n\n # load doc id info\n doc_flist.sort()\n if len(doc_flist) == 1:\n doc_df = dd.from_pandas(pd.read_csv(doc_flist[0], **kwargs),\n npartitions=1)\n else:\n doc_df = dd.read_csv(doc_flist, blocksize=blocksize, **kwargs)\n\n # load term id info\n term_flist.sort()\n if len(term_flist) == 1:\n term_df = dd.from_pandas(pd.read_csv(term_flist[0], **kwargs),\n npartitions=1)\n else:\n term_df = dd.read_csv(term_flist, blocksize=blocksize, **kwargs)\n\n # load counts\n count_flist.sort()\n if len(count_flist) == 1:\n count_df = dd.from_pandas(pd.read_csv(count_flist[0], **kwargs),\n npartitions=1)\n else:\n count_df = dd.read_csv(count_flist, blocksize=blocksize, **kwargs)\n\n # prep doc/term fpatterns\n if keep_fname:\n doc_fpat, term_fpat = _prep_fnames(doc_globstring, term_globstring)\n else:\n doc_fpat, term_fpat = None, None\n\n dtm = DTM(doc_df=doc_df, term_df=term_df, count_df=count_df,\n doc_index=doc_index, term_index=term_index,\n doc_fpat=doc_fpat, term_fpat=term_fpat)\n\n return dtm\n", "sub_path": "DiSTL/DTM__dep00/DTM.py", "file_name": "DTM.py", "file_ext": "py", "file_size_in_byte": 34402, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "glob.glob", "line_number": 40, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "difflib.ndiff", "line_number": 52, "usage_type": "call"}, {"api_name": "difflib.ndiff", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 199, "usage_type": "call"}, {"api_name": "os.path", "line_number": 199, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 202, "usage_type": "call"}, {"api_name": "os.path", "line_number": 202, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 205, "usage_type": "call"}, {"api_name": "os.path", "line_number": 205, "usage_type": "attribute"}, {"api_name": "dask.delayed", "line_number": 258, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 258, "usage_type": "attribute"}, {"api_name": "dask.dataframe.from_delayed", "line_number": 285, "usage_type": "call"}, {"api_name": "dask.dataframe", "line_number": 285, "usage_type": "name"}, {"api_name": "dask.dataframe.from_delayed", "line_number": 286, "usage_type": "call"}, {"api_name": "dask.dataframe", "line_number": 286, "usage_type": "name"}, {"api_name": "dask.delayed", "line_number": 331, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 331, "usage_type": "attribute"}, {"api_name": "dask.dataframe.from_delayed", "line_number": 344, "usage_type": "call"}, {"api_name": "dask.dataframe", "line_number": 344, "usage_type": "name"}, {"api_name": "dask.delayed", "line_number": 445, "usage_type": "call"}, {"api_name": "dask.dataframe.from_delayed", "line_number": 457, "usage_type": "call"}, {"api_name": "dask.dataframe", "line_number": 457, "usage_type": "name"}, {"api_name": "dask.delayed", "line_number": 524, "usage_type": "call"}, {"api_name": "dask.delayed", "line_number": 529, "usage_type": "call"}, {"api_name": "dask.dataframe.from_delayed", "line_number": 543, "usage_type": "call"}, {"api_name": "dask.dataframe", "line_number": 543, "usage_type": "name"}, {"api_name": "dask.delayed", "line_number": 613, "usage_type": "call"}, {"api_name": "dask.dataframe.from_delayed", "line_number": 631, "usage_type": "call"}, {"api_name": "dask.dataframe", "line_number": 631, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 945, "usage_type": "call"}, {"api_name": "os.path", "line_number": 945, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 947, "usage_type": "call"}, {"api_name": "os.path", "line_number": 947, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 950, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 951, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1006, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1006, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 1007, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1010, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1010, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 1011, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1014, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1014, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 1015, "usage_type": "call"}, {"api_name": "dask.dataframe.from_pandas", "line_number": 1020, "usage_type": "call"}, {"api_name": "dask.dataframe", "line_number": 1020, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 1020, "usage_type": "call"}, {"api_name": "dask.dataframe.read_csv", "line_number": 1023, "usage_type": "call"}, {"api_name": "dask.dataframe", "line_number": 1023, "usage_type": "name"}, {"api_name": "dask.dataframe.from_pandas", "line_number": 1028, "usage_type": "call"}, {"api_name": "dask.dataframe", "line_number": 1028, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 1028, "usage_type": "call"}, {"api_name": "dask.dataframe.read_csv", "line_number": 1031, "usage_type": "call"}, {"api_name": "dask.dataframe", "line_number": 1031, "usage_type": "name"}, {"api_name": "dask.dataframe.from_pandas", "line_number": 1036, "usage_type": "call"}, {"api_name": "dask.dataframe", "line_number": 1036, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 1036, "usage_type": "call"}, {"api_name": "dask.dataframe.read_csv", "line_number": 1039, "usage_type": "call"}, {"api_name": "dask.dataframe", "line_number": 1039, "usage_type": "name"}]}
+{"seq_id": "583602540", "text": "import json\nimport requests\nimport sys\nimport socket\nimport os\nimport logging as log\n\nACCESS_TOKEN = os.environ['ACCESS_TOKEN']\nACCESS_SECRET = os.environ['ACCESS_SECRET']\nCONSUMER_KEY = os.environ['CONSUMER_KEY']\nCONSUMER_SECRET = os.environ['CONSUMER_SECRET']\nHEADERS = {\n \"Content-type\": \"application/json\",\n \"Authorization\": f\"Bearer {os.environ['BEARER']}\"\n}\n\nTWITTER_IP = \"TWITTER_IP\"\nTWITTER_PORT = \"TWITTER_PORT\"\n\n\ndef create_headers(bearer_token):\n headers = {\"Authorization\": \"Bearer {}\".format(bearer_token)}\n return headers\n\n\ndef get_rules(headers):\n response = requests.get(\n \"https://api.twitter.com/2/tweets/search/stream/rules\", headers=headers\n )\n if response.status_code != 200:\n raise Exception(\n \"Cannot get rules (HTTP {}): {}\".format(response.status_code, response.text)\n )\n print(json.dumps(response.json()))\n return response.json()\n\n\ndef delete_all_rules(headers, rules):\n if rules is None or \"data\" not in rules:\n return None\n\n ids = list(map(lambda rule: rule[\"id\"], rules[\"data\"]))\n payload = {\"delete\": {\"ids\": ids}}\n response = requests.post(\n \"https://api.twitter.com/2/tweets/search/stream/rules\",\n headers=headers,\n json=payload\n )\n if response.status_code != 200:\n raise Exception(\n \"Cannot delete rules (HTTP {}): {}\".format(\n response.status_code, response.text\n )\n )\n print(json.dumps(response.json()))\n\n\ndef set_rules(headers):\n # You can adjust the rules if needed\n sample_rules = [\n {\"value\": \"(covid OR brasil OR vacina OR governo) lang:pt\", \"tag\": \"politica\"},\n {\"value\": \"(bbb OR meme OR futebol OR tv) lang:pt\", \"tag\": \"entretenimento\"},\n ]\n payload = {\"add\": sample_rules}\n response = requests.post(\n \"https://api.twitter.com/2/tweets/search/stream/rules\",\n headers=headers,\n json=payload,\n )\n if response.status_code != 201:\n raise Exception(\n \"Cannot add rules (HTTP {}): {}\".format(response.status_code, response.text)\n )\n print(json.dumps(response.json()))\n\n\ndef get_stream(ip, port, headers):\n\n TCP_IP = ip\n TCP_PORT = port\n conn = None\n s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n\n s.bind((TCP_IP, TCP_PORT))\n s.listen(1)\n\n log.info(\"Aguardando uma conexão TCP...\")\n conn, addr = s.accept()\n\n log.info(\"Conectado... Começando a coletar tweets...\")\n\n response = requests.get(\n \"https://api.twitter.com/2/tweets/search/stream\", headers=headers, stream=True,\n )\n print(response.status_code)\n if response.status_code != 200:\n raise Exception(\n \"Cannot get stream (HTTP {}): {}\".format(\n response.status_code, response.text\n )\n )\n for response_line in response.iter_lines():\n if response_line:\n try:\n full_tweet = json.loads(response_line)\n tweet_text = full_tweet['data']['text']\n log.debug(\"Tweet Text: \" + tweet_text)\n log.debug(\"------------------------------------------\")\n b = bytes(tweet_text + '\\n', 'utf-8')\n conn.send(b)\n except Exception as e:\n log.error(f\"Error: {e}\")\n\nif __name__ == \"__main__\":\n rules = get_rules(HEADERS)\n delete = delete_all_rules(HEADERS, rules)\n set_rules(HEADERS)\n get_stream(ip=os.environ[TWITTER_IP], port=int(os.environ[TWITTER_PORT]), headers=HEADERS)", "sub_path": "groups/group1/tp1 - SparkStreaming/services/twitter/client.py", "file_name": "client.py", "file_ext": "py", "file_size_in_byte": 3539, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.environ", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 9, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 10, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 14, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 27, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 34, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 44, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 55, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 65, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 74, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 82, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 82, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 82, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 87, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 90, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 92, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 105, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 107, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 108, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 112, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 118, "usage_type": "attribute"}]}
+{"seq_id": "226062615", "text": "from concurrent import futures\nimport time\nimport pandas as pd\nimport grpc\nimport csv\nimport communicate_pb2\nimport communicate_pb2_grpc\nimport string\nimport random\nfrom datetime import datetime\nimport os\nimport sys\nfrom hashlib import sha1\nimport json\nimport glob\nimport socket\n\ndef get_my_ip_address(remote_server=\"google.com\"):\n \"\"\"\n Return the/a network-facing IP number for this system.\n \"\"\"\n with socket.socket(socket.AF_INET, socket.SOCK_DGRAM) as s:\n s.connect((remote_server, 80))\n return s.getsockname()[0]\n\nsalas = {}\ndf = pd.DataFrame(columns=['LogID', 'NomeSala', 'Horario', 'NumJogada', 'Tabuleiro', 'Ganhador'])\ndf_snapshot = pd.DataFrame(columns=['Jogos'])\n# numPartida = 1\ncounter = 1\ncounterServerLog = 0\ncounterSnap = 0\nporta = None\nmyPort = None\nnext = None\nprev = None\nipNext = ''\nipPrev = ''\nflag = False\nnodes = {}\nmyIp = get_my_ip_address()\nmyHash = None\nnumNos = 2\nmyHashRange = []\nimFirst = False\n\n\ndef geraLog(stringLength=10):\n \"\"\" Generate a random string of fixed length \"\"\"\n letters = string.ascii_lowercase\n return ''.join(random.choice(letters) for i in range(stringLength))\n\n\ndef logAppend(nomeSala, numJogada, tabuleiro, ganhador=''):\n global counterServerLog\n global df\n global counter\n global df_snapshot\n global salas\n global counterSnap\n global flag\n\n if counter % 3 == 0:\n counterServerLog += 1\n if counterServerLog % 4 == 0 or flag:\n file = './ServerLog/' + myPort + '/serverLog' + str(counterServerLog - 4) + '.csv'\n os.remove(file)\n os.remove('./Snapshots/' + myPort + '/snapshot' + str(counterSnap - 3) + '.csv')\n # counterServerLog = 0\n # counterSnap = 0\n flag = True\n\n df = pd.DataFrame(columns=['LogID', 'NomeSala', 'Horario', 'NumJogada', 'Tabuleiro', 'Ganhador'])\n df.drop(columns=['LogID', 'NomeSala', 'Horario', 'NumJogada', 'Tabuleiro', 'Ganhador'])\n\n if not os.path.exists('./Snapshots/' + myPort):\n os.mkdir('./Snapshots/' + myPort)\n\n snapName = './Snapshots/' + myPort + '/snapshot' + str(counterSnap) + '.csv'\n df_snapshot = pd.DataFrame.from_dict(salas, orient=\"index\",\n columns=['Tabuleiro', 'NumJogada', 'NumJogadores', 'Jogador1', 'Jogador2','vencedor'])\n df_snapshot['Salas'] = salas.keys()\n df_snapshot.to_csv(snapName, index=None, header=True)\n counterSnap += 1\n\n if not os.path.exists('./ServerLog/' + myPort):\n os.mkdir('./ServerLog/' + myPort)\n name = './ServerLog/' + myPort + '/serverLog' + str(counterServerLog) + '.csv'\n\n horario = datetime.now()\n horario = horario.strftime(\"%d/%m/%Y - %H:%M:%S\")\n\n while True:\n logID = geraLog()\n if not (logID in str(df['LogID'])):\n break\n\n df = df.append(\n {'LogID': logID, 'NomeSala': nomeSala, 'Horario': horario, 'NumJogada': numJogada, 'Tabuleiro': tabuleiro,\n 'Ganhador': ganhador}, ignore_index=True)\n df.to_csv(name, index=None, header=True)\n counter += 1\n print(counter, counterServerLog, counterSnap)\n # return\n\n\ndef atribui_posicao(nomeSala, pos):\n global salas\n\n tabuleiro = salas[nomeSala][0]\n numJogada = salas[nomeSala][1]\n\n if (numJogada % 2) == 0:\n tabuleiro = tabuleiro.replace(pos, 'X')\n else:\n tabuleiro = tabuleiro.replace(pos, 'O')\n print(tabuleiro)\n\n logAppend(nomeSala, numJogada, tabuleiro)\n\n salas[nomeSala][1] += 1\n return tabuleiro\n\n\ndef winner(nomeSala):\n global salas\n user1 = salas[nomeSala][3]\n user2 = salas[nomeSala][4]\n tabuleiro = salas[nomeSala][0]\n numJogada = salas[nomeSala][1]\n sign = ''\n for i in ['X', 'O']:\n # horizontal\n if tabuleiro[0] == tabuleiro[1] and tabuleiro[1] == tabuleiro[2] and tabuleiro[2] == i:\n sign = i\n elif tabuleiro[3] == tabuleiro[4] and tabuleiro[4] == tabuleiro[5] and tabuleiro[5] == i:\n sign = i\n elif tabuleiro[6] == tabuleiro[7] and tabuleiro[7] == tabuleiro[8] and tabuleiro[8] == i:\n sign = i\n # vertical\n elif tabuleiro[0] == tabuleiro[3] and tabuleiro[3] == tabuleiro[6] and tabuleiro[6] == i:\n sign = i\n elif tabuleiro[1] == tabuleiro[4] and tabuleiro[4] == tabuleiro[7] and tabuleiro[7] == i:\n sign = i\n elif tabuleiro[2] == tabuleiro[5] and tabuleiro[5] == tabuleiro[8] and tabuleiro[8] == i:\n sign = i\n # diagonal\n elif tabuleiro[0] == tabuleiro[4] and tabuleiro[4] == tabuleiro[8] and tabuleiro[8] == i:\n sign = i\n elif tabuleiro[6] == tabuleiro[4] and tabuleiro[4] == tabuleiro[2] and tabuleiro[2] == i:\n sign = i\n\n if sign == 'X':\n logAppend(nomeSala, numJogada, tabuleiro, user1)\n return user1\n elif sign == \"0\":\n logAppend(nomeSala, numJogada, tabuleiro, user2)\n return user2\n elif numJogada == 9:\n logAppend(nomeSala, numJogada, tabuleiro, 'empate')\n return 'empate'\n return None\n\n\ndef verificaCadastro(username, password):\n sucesso = True\n df = pd.read_csv('cadastros.csv')\n\n for index, row in df.iterrows():\n if str(row['usuario']) == username and str(row['senha']) == password:\n return sucesso\n return not sucesso\n\n\ndef conectaUsuario(username):\n sucesso = True\n with open('usuarios_conectados.csv', 'r') as file:\n reader = csv.reader(file)\n\n # verifica se ja tem um conectado com o mesmo usuario\n for row in reader:\n try:\n if row[0] == username:\n return not sucesso\n except IndexError:\n pass\n\n with open(r'usuarios_conectados.csv', 'a', newline='') as f:\n writer = csv.writer(f)\n writer.writerow([username])\n return sucesso\n\n\ndef desconectaUsuarios(nomeSala):\n global salas\n usuariosConectados = [salas[nomeSala][3], salas[nomeSala][4]]\n\n users = pd.read_csv('usuarios_conectados.csv')\n for usuario in usuariosConectados:\n users = users[users.usuarios != usuario]\n\n users.to_csv('usuarios_conectados.csv', encoding='utf-8', index=False)\n\n\ndef validaUsername(username):\n sucesso = True\n df = pd.read_csv('cadastros.csv')\n\n if username in str(df['usuario']):\n return not sucesso\n else:\n return sucesso\n\n\ndef cadastraUsuario(username, password):\n df = pd.read_csv('cadastros.csv')\n df = df.append({'usuario': username, 'senha': password}, ignore_index=True)\n df.to_csv(r'cadastros.csv', index=None, header=True)\n\n\nclass Communicate(communicate_pb2_grpc.CommunicateServicer):\n\n def ReqNum(self, request, context):\n global salas\n global next\n global prev\n global imFirst\n\n nomeSala = request.nome\n hashSala = int(sha1(nomeSala.encode('utf-8')).hexdigest(), 16) % 360\n\n if imFirst:\n if (0 <= hashSala <= myHashRange[0]) or (myHashRange[1] <= hashSala <= 359):\n numJogadores = salas[nomeSala][2]\n return communicate_pb2.NumJogadores(numJogadores=numJogadores)\n else:\n response = next.ReqNum(communicate_pb2.NomeSala(nome=request.nome))\n return communicate_pb2.NumJogadores(numJogadores=response.numJogadores)\n else:\n if myHashRange[1] >= hashSala >= myHashRange[0]:\n numJogadores = salas[nomeSala][2]\n return communicate_pb2.NumJogadores(numJogadores=numJogadores)\n elif hashSala > myHashRange[1]:\n response = next.ReqNum(communicate_pb2.NomeSala(nome=request.nome))\n return communicate_pb2.NumJogadores(numJogadores=response.numJogadores)\n elif hashSala < myHashRange[0]:\n response = prev.ReqNum(communicate_pb2.NomeSala(nome=request.nome))\n return communicate_pb2.NumJogadores(numJogadores=response.numJogadores)\n\n def Login(self, request, context):\n conectado_com_sucesso = 1\n usuario_nao_existe = 2\n usuario_ja_conectado = 3\n client = communicate_pb2.LoginRequest()\n client.username, client.password = request.username, request.password\n\n usuarioExiste = verificaCadastro(client.username, client.password)\n if usuarioExiste:\n possivelConectar = conectaUsuario(client.username)\n else:\n return communicate_pb2.APIResponse(responseCode=usuario_nao_existe)\n\n if possivelConectar:\n return communicate_pb2.APIResponse(responseCode=conectado_com_sucesso)\n else:\n return communicate_pb2.APIResponse(responseCode=usuario_ja_conectado)\n\n def AttPos(self, request, context):\n global next\n global prev\n global imFirst\n nomeSala = request.nome\n hashSala = int(sha1(nomeSala.encode('utf-8')).hexdigest(), 16) % 360\n\n if imFirst:\n if (0 <= hashSala <= myHashRange[0]) or (myHashRange[1] <= hashSala <= 359):\n # joga normal\n pos = request.pos\n salas[nomeSala][0] = atribui_posicao(nomeSala, pos)\n return communicate_pb2.Tabuleiro(tabuleiro=salas[nomeSala][0])\n else:\n response = next.AttPos(communicate_pb2.Jogada(nome=nomeSala, pos=request.pos))\n return communicate_pb2.Tabuleiro(tabuleiro=response.tabuleiro)\n else:\n if myHashRange[1] >= hashSala >= myHashRange[0]:\n # joga normal\n pos = request.pos\n salas[nomeSala][0] = atribui_posicao(nomeSala, pos)\n return communicate_pb2.Tabuleiro(tabuleiro=salas[nomeSala][0])\n elif hashSala > myHashRange[1]:\n response = next.AttPos(communicate_pb2.Jogada(nome=nomeSala, pos=request.pos))\n return communicate_pb2.Tabuleiro(tabuleiro=response.tabuleiro)\n elif hashSala < myHashRange[0]:\n response = prev.AttPos(communicate_pb2.Jogada(nome=nomeSala, pos=request.pos))\n return communicate_pb2.Tabuleiro(tabuleiro=response.tabuleiro)\n\n\n\n def ChecaVez(self, request, context):\n global salas\n global next\n global prev\n global imFirst\n\n nomeSala = request.nome\n hashSala = int(sha1(nomeSala.encode('utf-8')).hexdigest(), 16) % 360\n\n if imFirst:\n if (0 <= hashSala <= myHashRange[0]) or (myHashRange[1] <= hashSala <= 359):\n tabuleiro = salas[nomeSala][0]\n numJogada = salas[nomeSala][1]\n\n if (numJogada % 2) == 0:\n return communicate_pb2.Vez(vez=salas[nomeSala][3], tabuleiro=tabuleiro)\n return communicate_pb2.Vez(vez=salas[nomeSala][4], tabuleiro=tabuleiro)\n else:\n response = next.ChecaVez(communicate_pb2.NomeSala(nome=request.nome))\n return communicate_pb2.Vez(vez=response.vez, tabuleiro=response.tabuleiro)\n else:\n if myHashRange[1] >= hashSala >= myHashRange[0]:\n tabuleiro = salas[nomeSala][0]\n numJogada = salas[nomeSala][1]\n\n if (numJogada % 2) == 0:\n return communicate_pb2.Vez(vez=salas[nomeSala][3], tabuleiro=tabuleiro)\n return communicate_pb2.Vez(vez=salas[nomeSala][4], tabuleiro=tabuleiro)\n elif hashSala > myHashRange[1]:\n response = next.ChecaVez(communicate_pb2.NomeSala(nome=request.nome))\n return communicate_pb2.Vez(vez=response.vez, tabuleiro=response.tabuleiro)\n elif hashSala < myHashRange[0]:\n response = prev.ChecaVez(communicate_pb2.NomeSala(nome=request.nome))\n return communicate_pb2.Vez(vez=response.vez, tabuleiro=response.tabuleiro)\n\n\n def ChecaVencedor(self, request, context):\n global imFirst\n global salas\n global next\n global prev\n\n nomeSala = request.nome\n hashSala = int(sha1(nomeSala.encode('utf-8')).hexdigest(), 16) % 360\n\n if imFirst:\n if (0 <= hashSala <= myHashRange[0]) or (myHashRange[1] <= hashSala <= 359):\n vencedor = winner(nomeSala)\n if vencedor is not None:\n desconectaUsuarios(nomeSala)\n salas[nomeSala][5] = vencedor\n return communicate_pb2.Vencedor(vencedor=vencedor)\n else:\n return communicate_pb2.Vencedor(vencedor='ninguem')\n else:\n response = next.ChecaVencedor(communicate_pb2.NomeSala(nome=request.nome))\n return communicate_pb2.Vencedor(vencedor=response.vencedor)\n else:\n if myHashRange[1] >= hashSala >= myHashRange[0]:\n vencedor = winner(nomeSala)\n if vencedor is not None:\n desconectaUsuarios(nomeSala)\n salas[nomeSala][5] = vencedor\n return communicate_pb2.Vencedor(vencedor=vencedor)\n else:\n return communicate_pb2.Vencedor(vencedor='ninguem')\n elif hashSala > myHashRange[1]:\n response = next.ChecaVencedor(communicate_pb2.NomeSala(nome=request.nome))\n return communicate_pb2.Vencedor(vencedor=response.vencedor)\n elif hashSala < myHashRange[0]:\n response = prev.ChecaVencedor(communicate_pb2.NomeSala(nome=request.nome))\n return communicate_pb2.Vencedor(vencedor=response.vencedor)\n\n\n def CriaNovoJogo(self, request, context):\n nomeSala = request.nome\n hashSala = int(sha1(nomeSala.encode('utf-8')).hexdigest(), 16) % 360\n\n global imFirst\n global salas\n global next\n global prev\n\n if imFirst:\n if (0 <= hashSala <= myHashRange[0]) or (myHashRange[1] <= hashSala <= 359):\n if nomeSala not in salas:\n salas[nomeSala] = ['123456789', 0, 1, request.nomeJogador, '', '']\n print(salas)\n return communicate_pb2.APIResponse(responseCode=1)\n else:\n return communicate_pb2.APIResponse(responseCode=0)\n else:\n response = next.CriaNovoJogo(communicate_pb2.SalaEJogador(nome=request.nome, nomeJogador=request.nomeJogador))\n return communicate_pb2.APIResponse(responseCode=response.responseCode)\n else:\n if myHashRange[1] >= hashSala >= myHashRange[0]:\n if nomeSala not in salas:\n salas[nomeSala] = ['123456789', 0, 1, request.nomeJogador, '', '']\n print(salas)\n return communicate_pb2.APIResponse(responseCode=1)\n else:\n return communicate_pb2.APIResponse(responseCode=0)\n elif hashSala > myHashRange[1]:\n response = next.CriaNovoJogo(communicate_pb2.SalaEJogador(nome=request.nome, nomeJogador=request.nomeJogador))\n return communicate_pb2.APIResponse(responseCode=response.responseCode)\n elif hashSala < myHashRange[0]:\n response = prev.CriaNovoJogo(communicate_pb2.SalaEJogador(nome=request.nome, nomeJogador=request.nomeJogador))\n return communicate_pb2.APIResponse(responseCode=response.responseCode)\n\n\n def ConectaSala(self, request, context):\n nomeSala = request.nome\n hashSala = int(sha1(nomeSala.encode('utf-8')).hexdigest(), 16) % 360\n\n global imFirst\n global salas\n global next\n global prev\n\n if imFirst:\n if (0 <= hashSala <= myHashRange[0]) or (myHashRange[1] <= hashSala <= 359):\n if nomeSala in salas:\n salas[nomeSala][2] += 1\n salas[nomeSala][4] = request.nomeJogador\n print(salas)\n return communicate_pb2.APIResponse(responseCode=1)\n else:\n return communicate_pb2.APIResponse(responseCode=0)\n else:\n response = next.ConectaSala(communicate_pb2.SalaEJogador(nome=request.nome, nomeJogador=request.nomeJogador))\n return communicate_pb2.APIResponse(responseCode=response.responseCode)\n else:\n if myHashRange[1] >= hashSala >= myHashRange[0]:\n if nomeSala in salas:\n salas[nomeSala][2] += 1\n salas[nomeSala][4] = request.nomeJogador\n print(salas)\n return communicate_pb2.APIResponse(responseCode=1)\n else:\n return communicate_pb2.APIResponse(responseCode=0)\n elif hashSala > myHashRange[1]:\n response = next.ConectaSala(\n communicate_pb2.SalaEJogador(nome=request.nome, nomeJogador=request.nomeJogador))\n return communicate_pb2.APIResponse(responseCode=response.responseCode)\n elif hashSala < myHashRange[0]:\n response = prev.ConectaSala(\n communicate_pb2.SalaEJogador(nome=request.nome, nomeJogador=request.nomeJogador))\n return communicate_pb2.APIResponse(responseCode=response.responseCode)\n\n\n def SignUp(self, request, context):\n sucesso = validaUsername(request.username)\n if sucesso:\n cadastraUsuario(request.username, request.password)\n return communicate_pb2.APIResponse(responseCode=1)\n else:\n return communicate_pb2.APIResponse(responseCode=0)\n\n def TestaConexao(self, request, context):\n global myHash\n return communicate_pb2.String(message=myHash)\n\n def ConectaDeVolta(self, request, context):\n global next\n global prev\n global myHash\n global myHashRange\n global imFirst\n if prev is None:\n channel = grpc.insecure_channel(request.ip + ':' + request.port)\n prev = communicate_pb2_grpc.CommunicateStub(channel)\n print('Comunicacao em duas vias com ' + request.ip + ':' + request.port + ' estabelecida com sucesso!')\n print('Prev: ' + request.ip + ':' + request.port)\n if int(request.hash) > int(myHash):\n imFirst = True\n print(imFirst)\n myHashRange.append(min(int(myHash), int(request.hash) + 1))\n myHashRange.append(max(int(myHash), int(request.hash) + 1))\n print(myHashRange)\n return communicate_pb2.Endereco(ip=myIp, port=myPort)\n\n def VerificaNextPrev(self, request, context):\n global ipNext\n global ipPrev\n return communicate_pb2.NPResponse(enderecoNext=ipNext, enderecoPrev=ipPrev)\n\n def ConexaoDisponivel(self, request, context):\n global nodes\n global myHash\n if nodes[myHash][2]:\n return communicate_pb2.DispResponse(disponivel=True, nextDisponivel='')\n else:\n endereco = None\n for key in nodes:\n if nodes[key][2]:\n endereco = nodes[key][0] + ':' + nodes[key][1]\n break\n return communicate_pb2.DispResponse(disponivel=False, nextDisponivel=endereco)\n\n def PegaDicionario(self, request, context):\n global nodes\n string_nodes = json.dumps(nodes) # transforma dict pra string -- pra destransformar json.loads(dict)\n return communicate_pb2.Dicionario(dicionario=string_nodes)\n\n def AtualizaDics(self, request, context):\n global nodes\n global next\n global myHash\n nodes = json.loads(request.dicionario)\n #print(nodes)\n if next is not None:\n next.AtualizaDics(communicate_pb2.Dicionario(dicionario=json.dumps(nodes)))\n return communicate_pb2.APIResponse(responseCode=1)\n else:\n return communicate_pb2.APIResponse(responseCode=1)\n\n\n# conecta em um servidor da lista, ou seja, esse server vira client do q ele digitou o ip\n# so pode mandar mensagens /// pra automatizar, tira o input e coloca pra ele ler um dontpad contendo a lista\n# VERIFICA SE NO JA TEM NEXT E PREV, SE TIVER, MANDAR CONECTAR COM OUTRO\ndef conectaNoAleatorio(myHash):\n global ipNext\n global ipPrev\n global nodes\n global myIp\n global myPort\n global next\n ipStub = input(\"Digite o IP de qualquer servidor na lista: \")\n if ipStub == '':\n nodes[myHash] = [myIp, myPort, True]\n #print(nodes)\n return None\n portStub = input(\"Digite a porta de qualquer servidor na lista: \")\n\n channel = grpc.insecure_channel(ipStub + ':' + portStub)\n next = communicate_pb2_grpc.CommunicateStub(channel)\n\n response = next.ConexaoDisponivel(communicate_pb2.Empty())\n if response.disponivel:\n response_dic = next.PegaDicionario(communicate_pb2.Empty())\n nodes = json.loads(response_dic.dicionario)\n for key in nodes:\n nodes[key][2] = False\n nodes[myHash] = [myIp, myPort, True]\n response_atualiza_dic = next.AtualizaDics(communicate_pb2.Dicionario(dicionario=json.dumps(nodes)))\n else:\n channel.close()\n endereco = response.nextDisponivel\n channel = grpc.insecure_channel(endereco)\n next = communicate_pb2_grpc.CommunicateStub(channel)\n response_dic = next.PegaDicionario(communicate_pb2.Empty())\n nodes = json.loads(response_dic.dicionario)\n for key in nodes:\n nodes[key][2] = False\n nodes[myHash] = [myIp, myPort, True]\n response_atualiza_dic = next.AtualizaDics(communicate_pb2.Dicionario(dicionario=json.dumps(nodes)))\n #print(nodes)\n\n\n# recebe o stub e envia pede pro outro conectar nele, criando a comunicacao em duas vias\n# recebe 1 caso tenha sido bem sucedido\ndef conectaDeVolta():\n global myPort\n global myIp\n global next\n global prev\n global myHash\n if next is not None:\n response = next.ConectaDeVolta(communicate_pb2.EnderecoAndHash(ip=myIp, port=myPort, hash=int(myHash)))\n #print('Comunicacao em duas vias com ' + response.ip + ':' + response.port + ' estabelecida com sucesso!')\n\n\ndef ordenaNos():\n global nodes\n import collections\n dicAux = {}\n\n for key in nodes:\n dicAux[int(key)] = nodes[key]\n\n od = collections.OrderedDict(sorted(dicAux.items()))\n return od\n\n\ndef conectaNext():\n global myHash\n global nodes\n global numNos\n global next\n hashList = list(nodes)\n\n nextindex = hashList.index(int(myHash)) + 1\n if nextindex == numNos:\n nextindex = 0\n nextkeyvalue = hashList[nextindex]\n nextkeyaddress = nodes[nextkeyvalue][0] + ':' + nodes[nextkeyvalue][1]\n\n channel = grpc.insecure_channel(nextkeyaddress)\n next = communicate_pb2_grpc.CommunicateStub(channel)\n\n\ndef xesque(x):\n tabuleiro = x['Tabuleiro']\n numJogada = x['NumJogada']\n numJogadores = x['NumJogadores']\n jogador1 = x['Jogador1']\n jogador2 = x['Jogador2']\n vencedor = x['vencedor']\n nomeSala = x['Salas']\n\n if pd.isna(vencedor):\n global salas\n salas[nomeSala] = [tabuleiro, numJogada, numJogadores, jogador1, jogador2, '']\n\n\ndef recuperaEstadoSnapshot():\n global myPort\n\n if os.path.exists('./Snapshots/' + myPort):\n listOfFiles = glob.glob('./Snapshots/' + str(myPort) + '/*.csv')\n lastFile = max(listOfFiles, key=os.path.getctime)\n\n df = pd.read_csv(lastFile)\n df.apply(xesque,axis = 1)\n\ndef get_my_ip_address(remote_server=\"google.com\"):\n \"\"\"\n Return the/a network-facing IP number for this system.\n \"\"\"\n with socket.socket(socket.AF_INET, socket.SOCK_DGRAM) as s:\n s.connect((remote_server, 80))\n return s.getsockname()[0]\n\ndef serve():\n global myPort\n global myHash\n global myIp\n global numNos\n global nodes\n global next\n\n print(\"My ip: \" + myIp)\n\n server = grpc.server(futures.ThreadPoolExecutor(max_workers=10))\n communicate_pb2_grpc.add_CommunicateServicer_to_server(Communicate(), server)\n myPort = sys.argv[1]\n myHash = str(int(sha1((myIp + myPort).encode('utf-8')).hexdigest(), 16) % 360)\n server.add_insecure_port(myIp + ':' + myPort)\n server.start()\n print(myHash)\n\n # acha no aleatorio na lista de possiveis nos e conecta, atribuindo ele como next\n conectaNoAleatorio(myHash)\n\n while True:\n time.sleep(3)\n if len(nodes) == numNos:\n nodes = ordenaNos()\n conectaNext()\n break\n\n conectaDeVolta()\n\n recuperaEstadoSnapshot()\n\n global salas\n print(salas)\n\n #response = next.TestaConexao(communicate_pb2.Empty())\n #print(myHash + '->' + response.message)\n\n try:\n while True:\n time.sleep(86400)\n except KeyboardInterrupt:\n server.stop(0)\n\n\nif __name__ == '__main__':\n serve()\n", "sub_path": "server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 24875, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "socket.socket", "line_number": 22, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 22, "usage_type": "attribute"}, {"api_name": "socket.SOCK_DGRAM", "line_number": 22, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 27, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 28, "usage_type": "call"}, {"api_name": "string.ascii_lowercase", "line_number": 50, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 51, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 67, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 68, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 77, "usage_type": "call"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 80, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 80, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 87, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 90, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 90, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 167, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 178, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 189, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 198, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 207, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 216, "usage_type": "call"}, {"api_name": "communicate_pb2_grpc.CommunicateServicer", "line_number": 221, "usage_type": "attribute"}, {"api_name": "hashlib.sha1", "line_number": 230, "usage_type": "call"}, {"api_name": "communicate_pb2.NumJogadores", "line_number": 235, "usage_type": "call"}, {"api_name": "communicate_pb2.NomeSala", "line_number": 237, "usage_type": "call"}, {"api_name": "communicate_pb2.NumJogadores", "line_number": 238, "usage_type": "call"}, {"api_name": "communicate_pb2.NumJogadores", "line_number": 242, "usage_type": "call"}, {"api_name": "communicate_pb2.NomeSala", "line_number": 244, "usage_type": "call"}, {"api_name": "communicate_pb2.NumJogadores", "line_number": 245, "usage_type": "call"}, {"api_name": "communicate_pb2.NomeSala", "line_number": 247, "usage_type": "call"}, {"api_name": "communicate_pb2.NumJogadores", "line_number": 248, "usage_type": "call"}, {"api_name": "communicate_pb2.LoginRequest", "line_number": 254, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 261, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 264, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 266, "usage_type": "call"}, {"api_name": "hashlib.sha1", "line_number": 273, "usage_type": "call"}, {"api_name": "communicate_pb2.Tabuleiro", "line_number": 280, "usage_type": "call"}, {"api_name": "communicate_pb2.Jogada", "line_number": 282, "usage_type": "call"}, {"api_name": "communicate_pb2.Tabuleiro", "line_number": 283, "usage_type": "call"}, {"api_name": "communicate_pb2.Tabuleiro", "line_number": 289, "usage_type": "call"}, {"api_name": "communicate_pb2.Jogada", "line_number": 291, "usage_type": "call"}, {"api_name": "communicate_pb2.Tabuleiro", "line_number": 292, "usage_type": "call"}, {"api_name": "communicate_pb2.Jogada", "line_number": 294, "usage_type": "call"}, {"api_name": "communicate_pb2.Tabuleiro", "line_number": 295, "usage_type": "call"}, {"api_name": "hashlib.sha1", "line_number": 306, "usage_type": "call"}, {"api_name": "communicate_pb2.Vez", "line_number": 314, "usage_type": "call"}, {"api_name": "communicate_pb2.Vez", "line_number": 315, "usage_type": "call"}, {"api_name": "communicate_pb2.NomeSala", "line_number": 317, "usage_type": "call"}, {"api_name": "communicate_pb2.Vez", "line_number": 318, "usage_type": "call"}, {"api_name": "communicate_pb2.Vez", "line_number": 325, "usage_type": "call"}, {"api_name": "communicate_pb2.Vez", "line_number": 326, "usage_type": "call"}, {"api_name": "communicate_pb2.NomeSala", "line_number": 328, "usage_type": "call"}, {"api_name": "communicate_pb2.Vez", "line_number": 329, "usage_type": "call"}, {"api_name": "communicate_pb2.NomeSala", "line_number": 331, "usage_type": "call"}, {"api_name": "communicate_pb2.Vez", "line_number": 332, "usage_type": "call"}, {"api_name": "hashlib.sha1", "line_number": 342, "usage_type": "call"}, {"api_name": "communicate_pb2.Vencedor", "line_number": 350, "usage_type": "call"}, {"api_name": "communicate_pb2.Vencedor", "line_number": 352, "usage_type": "call"}, {"api_name": "communicate_pb2.NomeSala", "line_number": 354, "usage_type": "call"}, {"api_name": "communicate_pb2.Vencedor", "line_number": 355, "usage_type": "call"}, {"api_name": "communicate_pb2.Vencedor", "line_number": 362, "usage_type": "call"}, {"api_name": "communicate_pb2.Vencedor", "line_number": 364, "usage_type": "call"}, {"api_name": "communicate_pb2.NomeSala", "line_number": 366, "usage_type": "call"}, {"api_name": "communicate_pb2.Vencedor", "line_number": 367, "usage_type": "call"}, {"api_name": "communicate_pb2.NomeSala", "line_number": 369, "usage_type": "call"}, {"api_name": "communicate_pb2.Vencedor", "line_number": 370, "usage_type": "call"}, {"api_name": "hashlib.sha1", "line_number": 375, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 387, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 389, "usage_type": "call"}, {"api_name": "communicate_pb2.SalaEJogador", "line_number": 391, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 392, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 398, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 400, "usage_type": "call"}, {"api_name": "communicate_pb2.SalaEJogador", "line_number": 402, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 403, "usage_type": "call"}, {"api_name": "communicate_pb2.SalaEJogador", "line_number": 405, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 406, "usage_type": "call"}, {"api_name": "hashlib.sha1", "line_number": 411, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 424, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 426, "usage_type": "call"}, {"api_name": "communicate_pb2.SalaEJogador", "line_number": 428, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 429, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 436, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 438, "usage_type": "call"}, {"api_name": "communicate_pb2.SalaEJogador", "line_number": 441, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 442, "usage_type": "call"}, {"api_name": "communicate_pb2.SalaEJogador", "line_number": 445, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 446, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 453, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 455, "usage_type": "call"}, {"api_name": "communicate_pb2.String", "line_number": 459, "usage_type": "call"}, {"api_name": "grpc.insecure_channel", "line_number": 468, "usage_type": "call"}, {"api_name": "communicate_pb2_grpc.CommunicateStub", "line_number": 469, "usage_type": "call"}, {"api_name": "communicate_pb2.Endereco", "line_number": 478, "usage_type": "call"}, {"api_name": "communicate_pb2.NPResponse", "line_number": 483, "usage_type": "call"}, {"api_name": "communicate_pb2.DispResponse", "line_number": 489, "usage_type": "call"}, {"api_name": "communicate_pb2.DispResponse", "line_number": 496, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 500, "usage_type": "call"}, {"api_name": "communicate_pb2.Dicionario", "line_number": 501, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 507, "usage_type": "call"}, {"api_name": "communicate_pb2.Dicionario", "line_number": 510, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 510, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 511, "usage_type": "call"}, {"api_name": "communicate_pb2.APIResponse", "line_number": 513, "usage_type": "call"}, {"api_name": "grpc.insecure_channel", "line_number": 533, "usage_type": "call"}, {"api_name": "communicate_pb2_grpc.CommunicateStub", "line_number": 534, "usage_type": "call"}, {"api_name": "communicate_pb2.Empty", "line_number": 536, "usage_type": "call"}, {"api_name": "communicate_pb2.Empty", "line_number": 538, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 539, "usage_type": "call"}, {"api_name": "communicate_pb2.Dicionario", "line_number": 543, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 543, "usage_type": "call"}, {"api_name": "grpc.insecure_channel", "line_number": 547, "usage_type": "call"}, {"api_name": "communicate_pb2_grpc.CommunicateStub", "line_number": 548, "usage_type": "call"}, {"api_name": "communicate_pb2.Empty", "line_number": 549, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 550, "usage_type": "call"}, {"api_name": "communicate_pb2.Dicionario", "line_number": 554, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 554, "usage_type": "call"}, {"api_name": "communicate_pb2.EnderecoAndHash", "line_number": 567, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 579, "usage_type": "call"}, {"api_name": "grpc.insecure_channel", "line_number": 596, "usage_type": "call"}, {"api_name": "communicate_pb2_grpc.CommunicateStub", "line_number": 597, "usage_type": "call"}, {"api_name": "pandas.isna", "line_number": 609, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 617, "usage_type": "call"}, {"api_name": "os.path", "line_number": 617, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 618, "usage_type": "call"}, {"api_name": "os.path", "line_number": 619, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 621, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 628, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 628, "usage_type": "attribute"}, {"api_name": "socket.SOCK_DGRAM", "line_number": 628, "usage_type": "attribute"}, {"api_name": "grpc.server", "line_number": 642, "usage_type": "call"}, {"api_name": "concurrent.futures.ThreadPoolExecutor", "line_number": 642, "usage_type": "call"}, {"api_name": "concurrent.futures", "line_number": 642, "usage_type": "name"}, {"api_name": "communicate_pb2_grpc.add_CommunicateServicer_to_server", "line_number": 643, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 644, "usage_type": "attribute"}, {"api_name": "hashlib.sha1", "line_number": 645, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 654, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 672, "usage_type": "call"}]}
+{"seq_id": "303701722", "text": "from __future__ import unicode_literals\n\nfrom django.db import models\nfrom mongoengine import *\n\n# Create your models here.\nimport json\n\n\nclass DailyUpdate(Document):\n opratorName = StringField(max_length=100)\n startKM = IntField()\n endKM = IntField()\n date = DateTimeField()\n operatorId = StringField(max_length=10)\n serviceHour = StringField(max_length=5)\n\n def convert(self,document):\n if(document is not None):\n return DailyUpdate(document['opratorName'],document['startKM'],document['endKM'],document['date'],document['operatorId'],document['serviceHour'])\n else:\n return None\n\nclass ServiceData(EmbeddedDocument):\n created = DateTimeField()\n KMstand = IntField()\n itemListReplaced = ListField()\n comment = StringField(max_length=500)\n cost = FloatField()\n\n def toDATA(self):\n data = {}\n data[\"created\"] = self.created.strftime('%Y-%m-%d %H:%M')\n data[\"KMstand\"] = self.KMstand\n data[\"itemListReplaced\"] = self.itemListReplaced\n data[\"comment\"] = self.comment\n data[\"cost\"] = self.cost\n return data\n\n def convert(self,document):\n if(document is not None):\n return ServiceData(document['created'],document['KMstand'],document['itemListReplaced'],document['comment'],document['cost'])\n else:\n return None\n\n def setData(self, _created, _KMstand, _itemListReplaced, _comment, _cost):\n self.created = _created\n self.KMstand = _KMstand\n self.itemListReplaced = _itemListReplaced\n self.comment = _comment\n self.cost = _cost\n\nclass TripDetail(EmbeddedDocument):\n tripFrom = StringField(max_lenght=1024)\n tripTo = StringField(max_lenght=1024)\n revenueGenerated = FloatField()\n runningCost = FloatField()\n journeyDuration = StringField(max_lenght=1024);\n journeyDate = DateTimeField();\n odoMeasure = IntField()\n\n def toDATA(self):\n data = {}\n data[\"tripFrom\"] = self.tripFrom\n data[\"tripTo\"] = self.tripTo\n data[\"revenueGenerated\"] = self.revenueGenerated\n data[\"runningCost\"] = self.runningCost\n data[\"journeyDuration\"] = self.journeyDuration\n data[\"journeyDate\"] = self.journeyDate.strftime('%Y-%m-%d %H:%M')\n data[\"odoMeasure\"] = self.odoMeasure\n\n return data\n\n def setData(self, _tripFrom, _tripTo, _revenueGenerated, _runningCost, _journeyDuration, _journeyDate, _odoMeasure):\n self.tripFrom = _tripFrom\n self.tripTo = _tripTo\n self.revenueGenerated = _revenueGenerated\n self.runningCost = _runningCost\n self.journeyDuration = _journeyDuration\n self.journeyDate = _journeyDate\n self.odoMeasure = _odoMeasure\n\nclass Vehical(Document):\n purchaseDate = DateTimeField()\n vn = StringField(max_length=20)\n chasisNumber = StringField(max_length=20)\n serviceHistory = EmbeddedDocumentListField(ServiceData)\n tripHistory = EmbeddedDocumentListField(TripDetail)\n\n def toJSON(self):\n data = {}\n data[\"purchaseDate\"] = self.purchaseDate.strftime('%Y-%m-%d %H:%M')\n data[\"vn\"] = self.vn\n data[\"chasisNumber\"] = self.chasisNumber\n if(len(self.serviceHistory)>0):\n sHistory = []\n for s in self.serviceHistory:\n sHistory.append(s.toDATA())\n data[\"serviceHistory\"] = sHistory\n else:\n data[\"serviceHistory\"] = \"[]\"\n\n if(len(self.tripHistory)>0):\n trips = []\n for trip in self.tripHistory:\n trips.append(trip.toDATA())\n data[\"tripHistory\"] = trips\n else:\n data[\"tripHistory\"] = \"[]\"\n\n return json.dumps(data)", "sub_path": "inventoryMangement/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 3728, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "json.dumps", "line_number": 111, "usage_type": "call"}]}
+{"seq_id": "45691896", "text": "from django.shortcuts import render\nfrom django.shortcuts import render_to_response\nfrom django.template.loader import render_to_string\nfrom django.http import HttpResponse\nfrom articulo.models import Accesorio, GafaSol, EspejueloGraduado, LenteContacto, CristalGraduado\nfrom rest_framework import viewsets\nfrom rest_framework import viewsets, status\nfrom rest_framework.response import Response\nfrom articulo.serializers import *\nfrom datetime import datetime\nfrom rest_framework.decorators import list_route, detail_route\n\nclass VentaViewSet(viewsets.ModelViewSet):\n queryset = Venta.objects.all()\n serializer_class = VentaSerializer\n\n def list(self, request, **kwargs):\n try:\n esp = query_ventas_by_args(**request.query_params)\n serializer = VentaSerializer(esp['items'], context={'request': request}, many=True)\n result = dict()\n result['data'] = serializer.data\n result['draw'] = esp['draw']\n result['recordsTotal'] = esp['total']\n result['recordsFiltered'] = esp['count']\n\n return Response(result, status=status.HTTP_200_OK, template_name=None, content_type=None)\n\n except Exception as e:\n return Response(e, status=status.HTTP_404_NOT_FOUND, template_name=None, content_type=None)\n\n # @detail_route(methods=['get'], url_name='pormes', url_path='pormes')\n # def pormes(self, request, **kwargs):\n # fecha = datetime.today()\n # Mes = kwargs['pk']\n # Anno = fecha.year\n # ventas = Venta.objects.filter(fecha__year=Anno)\n # ventas = ventas.filter(fecha__month=Mes)\n # result = VentaSerializer(ventas, context={'request': request}, many=True)\n # return Response(result.data, status=status.HTTP_200_OK, template_name=None, content_type=None)\n\n @detail_route(methods=['get'], url_name='ventasEnAnnoActual', url_path='ventasEnAnnoActual')\n def ventasEnAnnoActual(self, request, **kwargs):\n kwarg = kwargs['pk']\n fecha = datetime.today()\n AnnoActual = fecha.year\n MesActual = fecha.month\n\n ventasDeAnnoActual = Venta.objects.filter(fecha__lte=fecha,fecha__year=AnnoActual)\n ventas = VentaSerializer(ventasDeAnnoActual, context={'request': request}, many=True)\n\n # Rellenando con cero el total de ventas en cada mes para luego calcularlas\n totalesPorMes = {}\n for mes in range(1, MesActual + 1):\n totalesPorMes[mes] = 0\n\n sumaTotal = 0\n\n for venta in ventas.data:\n fechaVenta = venta['fecha']\n mesVenta = fechaVenta[5:7]\n mes = int(mesVenta)\n total = venta['total']\n totalesPorMes[mes] += total\n sumaTotal += total\n\n return Response({'sumaTotal':sumaTotal,'totalesPorMes':totalesPorMes,'mesActual':MesActual,'annoActual':AnnoActual}, status=status.HTTP_200_OK, template_name=None, content_type=None)\n\n @detail_route(methods=['get'], url_name='ventasPorAnno', url_path='ventasPorAnno')\n def ventasPorAnno(self, request, **kwargs):\n kwarg = kwargs['pk']\n fecha = datetime.today()\n annoActual = fecha.year\n\n ventasAll = Venta.objects.filter(fecha__lte=fecha)\n ventas = VentaSerializer(ventasAll, context={'request': request}, many=True)\n\n totalesPorAnno = {}\n sumaTotal = 0\n annoMenor = annoActual\n\n # Rellenando con ceros la suma por annos del arreglo\n for venta in ventas.data:\n fechaVenta = venta['fecha']\n annoVenta = fechaVenta[0:4]\n anno = int(annoVenta)\n totalesPorAnno[anno] = 0\n if(annoMenor > anno):\n annoMenor = anno\n\n # LLenando las posiciones de los annos con sus totales de venta\n for venta in ventas.data:\n fechaVenta = venta['fecha']\n annoVenta = fechaVenta[0:4]\n anno = int(annoVenta)\n total = venta['total']\n totalesPorAnno[anno] += total\n sumaTotal += total\n\n return Response(\n {'sumaTotal': sumaTotal, 'totalesPorAnno': totalesPorAnno, 'annoActual': annoActual, 'annoMenor':annoMenor},\n status=status.HTTP_200_OK, template_name=None, content_type=None)\n\n\ndef addVentas(request):\n return render_to_response('addVenta.html')\n\ndef ventas(request):\n return render_to_response('ventas.html')\n\n\n", "sub_path": "venta/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 4407, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "rest_framework.viewsets.ModelViewSet", "line_number": 13, "usage_type": "attribute"}, {"api_name": "rest_framework.viewsets", "line_number": 13, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 27, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 27, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 27, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 30, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_404_NOT_FOUND", "line_number": 30, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 30, "usage_type": "name"}, {"api_name": "datetime.datetime.today", "line_number": 45, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 45, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 67, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 67, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 67, "usage_type": "name"}, {"api_name": "rest_framework.decorators.detail_route", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 72, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 72, "usage_type": "name"}, {"api_name": "rest_framework.response.Response", "line_number": 100, "usage_type": "call"}, {"api_name": "rest_framework.status.HTTP_200_OK", "line_number": 102, "usage_type": "attribute"}, {"api_name": "rest_framework.status", "line_number": 102, "usage_type": "name"}, {"api_name": "rest_framework.decorators.detail_route", "line_number": 69, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 106, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 109, "usage_type": "call"}]}
+{"seq_id": "335363608", "text": "#!/usr/bin/env python\n\n# Copyright 2020 Google Inc. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Google Cloud Game Servers sample for listing game server configs.\n\nExample usage:\n python list_configs.py --project-id --deployment-id \n\"\"\"\n\nimport argparse\n\nfrom google.cloud import gaming\n\n\n# [START cloud_game_servers_config_list]\ndef list_configs(project_id, deployment_id):\n \"\"\"Lists the existing game server deployments.\"\"\"\n\n client = gaming.GameServerConfigsServiceClient()\n\n # Location is hard coded as global, as game server configs can\n # only be created in global. This is done for all operations on\n # game server configs.\n response = client.list_game_server_configs(\n parent=f\"projects/{project_id}/locations/global/gameServerDeployments/{deployment_id}\"\n )\n\n for config in response.game_server_configs:\n print(f\"Name: {config.name}\")\n\n return response.game_server_configs\n# [END cloud_game_servers_config_list]\n\n\nif __name__ == \"__main__\":\n parser = argparse.ArgumentParser()\n parser.add_argument('--project-id', help='Your cloud project ID.', required=True)\n parser.add_argument('--deployment-id', help='Your game server deployment ID.', required=True)\n\n args = parser.parse_args()\n\n list_configs(args.project_id, args.deployment_id)\n", "sub_path": "samples/snippets/list_configs.py", "file_name": "list_configs.py", "file_ext": "py", "file_size_in_byte": 1861, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "google.cloud.gaming.GameServerConfigsServiceClient", "line_number": 32, "usage_type": "call"}, {"api_name": "google.cloud.gaming", "line_number": 32, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 49, "usage_type": "call"}]}
+{"seq_id": "300236382", "text": "import tempfile\n\nimport pytest\n\nfrom ravenpackapi import RPApi\nfrom ravenpackapi.models.job import Job\n\n\nclass TestDatafile(object):\n api = RPApi()\n\n @pytest.mark.slow\n @pytest.mark.datafile\n def test_small_async_download(self):\n ds = self.api.get_dataset(dataset_id='swiss20')\n job = ds.request_datafile(\n start_date='2018-01-01 18:00:00',\n end_date='2018-01-02 18:00:00',\n )\n assert isinstance(job, Job)\n with tempfile.NamedTemporaryFile() as fp:\n job.save_to_file(filename=fp.name)\n", "sub_path": "ravenpackapi/tests/test_datafile.py", "file_name": "test_datafile.py", "file_ext": "py", "file_size_in_byte": 565, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "ravenpackapi.RPApi", "line_number": 10, "usage_type": "call"}, {"api_name": "ravenpackapi.models.job.Job", "line_number": 20, "usage_type": "argument"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 21, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pytest.mark", "line_number": 13, "usage_type": "attribute"}]}
+{"seq_id": "441576702", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Apr 21 13:29:46 2020\n高速公路客货运距分析研究-以宁夏回族自治区为例\n@author: 18120900\n\"\"\"\n\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport matplotlib.gridspec as gridspec\nimport matplotlib.ticker as ticker\nimport seaborn as sns\nimport numpy as np\nfrom scipy.interpolate import make_interp_spline\nfrom statsmodels.graphics.tsaplots import plot_acf, plot_pacf\nfrom sklearn.metrics.pairwise import cosine_similarity\nfrom minepy import MINE\nfrom scipy import stats\nfrom PIL import Image\nimport io\n\n\nclass SeabornFig2Grid:\n\n def __init__(self, seaborngrid, fig, subplot_spec):\n self.fig = fig\n self.sg = seaborngrid\n self.subplot = subplot_spec\n if isinstance(self.sg, sns.axisgrid.FacetGrid) or isinstance(self.sg, sns.axisgrid.PairGrid):\n self._movegrid()\n elif isinstance(self.sg, sns.axisgrid.JointGrid):\n self._movejointgrid()\n self._finalize()\n\n def _movegrid(self):\n \"\"\" Move PairGrid or Facetgrid \"\"\"\n self._resize()\n n = self.sg.axes.shape[0]\n m = self.sg.axes.shape[1]\n self.subgrid = gridspec.GridSpecFromSubplotSpec(n, m, subplot_spec=self.subplot)\n for i in range(n):\n for j in range(m):\n self._moveaxes(self.sg.axes[i, j], self.subgrid[i, j])\n\n def _movejointgrid(self):\n \"\"\" Move Jointgrid \"\"\"\n h = self.sg.ax_joint.get_position().height\n h2 = self.sg.ax_marg_x.get_position().height\n r = int(np.round(h / h2))\n self._resize()\n self.subgrid = gridspec.GridSpecFromSubplotSpec(r + 1, r + 1, subplot_spec=self.subplot)\n\n self._moveaxes(self.sg.ax_joint, self.subgrid[1:, :-1])\n self._moveaxes(self.sg.ax_marg_x, self.subgrid[0, :-1])\n self._moveaxes(self.sg.ax_marg_y, self.subgrid[1:, -1])\n\n def _moveaxes(self, ax, gs):\n # https://stackoverflow.com/a/46906599/4124317\n ax.remove()\n ax.figure = self.fig\n self.fig.axes.append(ax)\n self.fig.add_axes(ax)\n ax._subplotspec = gs\n ax.set_position(gs.get_position(self.fig))\n ax.set_subplotspec(gs)\n\n def _finalize(self):\n plt.close(self.sg.fig)\n self.fig.canvas.mpl_connect(\"resize_event\", self._resize)\n self.fig.canvas.draw()\n\n def _resize(self, evt=None):\n self.sg.fig.set_size_inches(self.fig.get_size_inches())\n\n\ndef dealdata(data, groupby, index):\n pivoted_data = (data\n .groupby(groupby).sum().reset_index()\n .assign(avgmiles=lambda x: round(x['totalmiles'] / x['count'], 2))\n .pivot_table(values='avgmiles', index=index, columns=['veh'], aggfunc=np.mean)\n .reset_index())\n return pivoted_data\n\n\ndef statistic(data):\n data.describe()\n data.skew()\n data.kurt()\n\n\ndef drawyeartrendgram(data, lblname):\n f, axes = plt.subplots(3, 2, sharex='all', sharey='all', figsize=(5.12, 4))\n plt.subplots_adjust(0.07, 0.07, 0.98, 0.98, 0, 0)\n for i in range(len(lblname)):\n y = data.iloc[i, 2:]\n y = (y - y.min()) / (y.max() - y.min())\n axes[i // 3, i % 3].plot(data.columns[2:], y, 'k-', label=lblname[i])\n axes[i // 3, i % 3].legend(prop=font_CN, loc='upper right', frameon=False)\n seticksandlegend(loc='upper right', frameon=False)\n savefig2tif(f, '各车型变化趋势图.tif')\n\n\ndef hourtrendgram(pivot_data, lblname):\n f = plt.figure(figsize=(2.6, 2.5))\n ax = plt.axes()\n plt.subplots_adjust(0.1, 0.17, 0.98, 0.98, 0.2, 0.2)\n x = pivot_data.iloc[:, 0]\n x_new = np.linspace(x.min(), x.max(), 300)\n for i in range(len(lblname)):\n y = pivot_data.iloc[:, i + 1]\n y_smooth = make_interp_spline(x, y)(x_new)\n plt.plot(x_new, y_smooth, color='k', linestyle=dict_linestyle[i], linewidth=1, label=lblname[i])\n ax.spines['top'].set_visible(False)\n ax.spines['right'].set_visible(False)\n seticksandlegend('Hour/h', 'Miles/km', loc='upper right', ncol=2, frameon=False)\n savefig2tif(f, '小时趋势变化图.tif')\n\n\ndef hourlineplot(data, lblname):\n f = plt.figure(figsize=(2.6, 2.5))\n ax = plt.axes()\n plt.subplots_adjust(0.18, 0.15, 0.98, 0.85, 0.2, 0.2)\n r1 = [1, 2, 3, 4] if len(lblname) == 4 else [2, 3, 4, 5, 6]\n data['veh'] = data['veh'].replace(r1, lblname)\n data['avgmiles'] = round(data['totalmiles'] / data['count'], 2)\n veh = '客车' if len(lblname) == 4 else '货车'\n\n sns.lineplot('time', 'avgmiles', style='veh', color='k', data=data)\n\n ax.spines['top'].set_visible(False)\n ax.spines['right'].set_visible(False)\n plt.xlabel('Hour/h', fontdict=font_EN)\n plt.ylabel('Miles/km', fontdict=font_EN)\n plt.legend(prop=font_CN,\n loc='lower left',\n ncol=3,\n frameon=False,\n columnspacing=0.5,\n bbox_to_anchor=(0, 1),\n borderaxespad=0)\n savefig2tif(f, veh + '平均运距小时变化趋势图.tif')\n\n\ndef drawfig(pivot_data, lblname):\n veh = '客车' if len(lblname) == 4 else '货车'\n\n f = plt.figure(figsize=(3, 2.2))\n ax = plt.axes()\n plt.subplots_adjust(0.15, 0.2, 0.98, 0.98, 0.2, 0.25)\n for i in range(len(lblname)):\n plt.plot(pivot_data.iloc[:, 0],\n pivot_data.iloc[:, i + 1].rolling(7).mean(),\n color='k',\n linestyle=dict_linestyle[i],\n linewidth=1,\n label=lblname[i])\n ax.spines['top'].set_visible(False)\n ax.spines['right'].set_visible(False)\n loc = 'best' if veh == '货车' else 'upper center'\n seticksandlegend(xlabel='日期', ylabel='平均行驶里程/km', ncol=-(-len(lblname) // 2), loc=loc, rotation=15)\n savefig2tif(f, veh + '平均运距年度变化趋势图.tif')\n\n f = plt.figure(figsize=(3, 2))\n ax = plt.axes()\n plt.subplots_adjust(0.17, 0.17, 0.96, 0.96, 0.2, 0.25)\n for i in range(len(lblname)):\n sns.distplot(pivot_data.iloc[:, i + 1],\n hist=False,\n kde_kws={\"shade\": False, \"linestyle\": dict_linestyle[i]},\n color='k',\n label=lblname[i])\n ax.spines['top'].set_visible(False)\n ax.spines['right'].set_visible(False)\n plt.gca().yaxis.set_major_formatter(ticker.FormatStrFormatter('%.2f'))\n seticksandlegend(xlabel='平均行驶里程/km', ylabel='频率', loc='upper right')\n savefig2tif(f, veh + '平均运距年度分布图.tif')\n\n\ndef drawKDEfig(data, lblname):\n sns.set_style(style='white')\n fig = plt.figure(figsize=(5.12, 4))\n gs = gridspec.GridSpec(2, 3, None, 0.1, 0.1, 0.98, 0.98, 0.2, 0.25)\n for i in range(len(lblname)):\n temp_data = data if i == 0 else data[data.veh == i + 1]\n axes = sns.jointplot(x='miles',\n y='weight',\n data=temp_data,\n kind='kde',\n color='k',\n xlim=(-50, 500),\n height=4,\n space=0)\n ylabel = 'weight/t' if i in [0, 3] else ''\n axes.set_axis_labels(lblname[i] + '\\'s haul distance/km', ylabel, fontdict=font_EN)\n SeabornFig2Grid(axes, fig, gs[i])\n seticksandlegend()\n savefig2tif(fig, '货重和运距联合密度分布图.tif')\n\n\ndef drawbarplot(data):\n plt.figure(figsize=(3, 2.5))\n ax = plt.axes()\n plt.subplots_adjust(0.18, 0.15, 0.98, 0.85, 0.2, 0.2)\n tick_label = data.iloc[:, 0]\n x = np.arange(len(tick_label))\n width = 0.3\n ax.bar(x - width/2, data.iloc[:, 1], width=width, color='white', edgecolor='k', hatch='///', label='宁夏')\n ax.bar(x + width/2, data.iloc[:, 2], width=width, color='white', edgecolor='k', hatch='+++', label='黑龙江')\n ax.set_ylabel('平均行驶里程/km', fontdict=font_CN)\n ax.set_xlabel('车型', fontdict=font_CN)\n ax.set_xticks(x)\n ax.set_xticklabels(tick_label)\n ax.legend(prop=font_CN)\n\n\ndef seticksandlegend(xlabel='', ylabel='', loc='upper center', ncol=1, frameon=False, rotation=0):\n plt.xlabel(xlabel, fontdict=font_CN)\n plt.ylabel(ylabel, fontdict=font_CN)\n plt.xticks(fontproperties='Times New Roman', size=9, rotation=rotation)\n plt.yticks(fontproperties='Times New Roman', size=9)\n plt.legend(prop=font_CN,\n loc=loc,\n ncol=ncol,\n frameon=frameon,\n columnspacing=0.5,\n labelspacing=0.2,\n borderaxespad=0)\n\n\ndef savefig2tif(fig, filename):\n png1 = io.BytesIO()\n fig.savefig(png1, format=\"png\")\n Image.open(png1).save(filename)\n png1.close()\n\n\ndef calMIC(data):\n for i in range(5):\n mine = MINE(alpha=0.6, c=15)\n miles = data[data.veh == (i + 2)].iloc[:, 1]\n weight = data[data.veh == (i + 2)].iloc[:, 2]\n mine.compute_score(miles, weight)\n print(\"Without noise:\", \"MIC\", mine.mic())\n\n\ndef correlationanalysis(data):\n _, axes = plt.subplots(2, len(data.columns) - 1, sharex='all', sharey='all', figsize=(5.12, 4))\n plt.subplots_adjust(0.12, 0.07, 0.97, 0.97, 0, 0)\n for i in range(len(data.columns) - 1):\n plot_acf(data.iloc[:, i + 1], ax=axes[0, i], title='')\n plot_pacf(data.iloc[:, i + 1], ax=axes[1, i], title='')\n axes[0, 0].set_ylabel('ACF', fontdict=font_EN)\n axes[1, 0].set_ylabel('PACF', fontdict=font_EN)\n\n\ndef divisionoftimeperiod(data):\n \"\"\"\n 时段划分模型\n @param data: 数据集\n @return: None\n \"\"\"\n title = data.columns.values.tolist()\n # 生成21*21的矩阵(7天,每天三个纬度),即一天与其他天的余弦值\n result = cosine_similarity(data.values, data.values)\n df = pd.DataFrame(columns=title, data=result, index=title)\n plt.figure(figsize=(5, 4))\n plt.subplots_adjust(0.09, 0.1, 0.98, 0.98, 0.2, 0.2)\n sns.heatmap(df, cmap='coolwarm')\n plt.show()\n\n\ndef calSimilarity(p, q):\n assert len(p) == len(q), \"两个向量的维数不同\"\n print('余弦相似性', '欧氏距离', '皮尔逊相关系数')\n for i in range(len(p)):\n cc = cosine_similarity(p[i].reshape(1, -1), q[i].reshape(1, -1))[0]\n oo = np.sqrt(np.sum(np.square(p[i] - q[i])))\n pp = stats.pearsonr(p[i], q[i])\n print(cc, oo, pp)\n\n\nif __name__ == \"__main__\":\n plt.rcParams['font.sans-serif'] = ['simsun'] # 中文字体设置-黑体\n plt.rcParams['xtick.direction'] = 'in' # 设置x轴刻度线方向,朝内还是朝外\n plt.rcParams['ytick.direction'] = 'in' # 设置y轴刻度线方向,朝内还是朝外\n plt.rcParams['lines.linewidth'] = 1 # 设置线宽\n plt.rcParams['axes.unicode_minus'] = False # 解决保存图像是负号'-'显示为方块的问题\n\n font_EN = {'family': 'Times New Roman', 'weight': 'normal', 'size': 9}\n font_CN = {'family': 'simsun', 'weight': 'normal', 'size': 9}\n dict_markers = {0: 'x', 1: 'D', 2: '^', 3: 'o', 4: 's'}\n dict_linestyle = {0: '-', 1: '--', 2: '-.', 3: ':', 4: (0, (5, 5))}\n label = ['1类', '2类', '3类', '4类', '二轴', '三轴', '四轴', '五轴', '六轴']\n # 读取数据\n datafilepath = \"Data.xlsx\"\n sheet = pd.read_excel(datafilepath, [0, 1, 2, 3, 4, 5, 6, 7])\n #\n # 计算统计量\n # statistic(dealdata(sheet[1], ['date', 'veh'], ['date'])) # 客车\n # statistic(dealdata(sheet[2], ['date', 'veh'], ['date'])) # 货车\n #\n # # 画各个车型从2014到2018的变化趋势图\n # drawyeartrendgram(sheet[0], label)\n #\n # 各个车型年度变化趋势图、平均运距直方图\n drawfig(dealdata(sheet[1], ['date', 'veh'], ['date']), label[:4]) # 客车\n drawfig(dealdata(sheet[2], ['date', 'veh'], ['date']), label[4:]) # 货车\n #\n # # 各个车型小时变化趋势图\n # hourtrendgram(dealdata(sheet[1], ['time', 'veh'], ['time']), label[:4]) # 客车-平滑曲线\n # hourtrendgram(dealdata(sheet[2], ['time', 'veh'], ['time']), label[4:]) # 货车-平滑曲线\n # hourlineplot(sheet[1], label[:4]) # 客车-含有置信区间\n # hourlineplot(sheet[2], label[4:]) # 货车*含有置信区间\n #\n # # 黑龙江和宁夏结果对比\n # drawbarplot(sheet[4])\n #\n # # 相关性分析图\n # correlationanalysis(dealdata(sheet[1], ['date', 'veh'], ['date'])) # 客车\n # correlationanalysis(dealdata(sheet[2], ['date', 'veh'], ['date'])) # 货车\n #\n # # 货车平均运距与载重的kde图\n # drawKDEfig(sheet[3], ['All Truck', 'Axis-II', 'Axis-III', 'Axis-IV', 'Axis-V', 'Axis-VI'])\n # # 计算车货总重和平均运距之间的互相关信息MIC\n # calMIC(sheet[3])\n #\n # # 计算余弦相似性\n # divisionoftimeperiod(sheet[5])\n #\n # 计算变量之间的相似性\n # calSimilarity(sheet[6].values[:, 1:], sheet[7].values[:, 1:])\n\n plt.show()\n", "sub_path": "高速公路客货运距分析研究/高速公路客货运距分析研究.py", "file_name": "高速公路客货运距分析研究.py", "file_ext": "py", "file_size_in_byte": 12965, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "seaborn.axisgrid", "line_number": 29, "usage_type": "attribute"}, {"api_name": "seaborn.axisgrid", "line_number": 31, "usage_type": "attribute"}, {"api_name": "matplotlib.gridspec.GridSpecFromSubplotSpec", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.gridspec", "line_number": 40, "usage_type": "name"}, {"api_name": "numpy.round", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.gridspec.GridSpecFromSubplotSpec", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.gridspec", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 80, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 93, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 93, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 106, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 108, "usage_type": "call"}, {"api_name": "scipy.interpolate.make_interp_spline", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 120, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 120, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 121, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 121, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 122, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 122, "usage_type": "name"}, {"api_name": "seaborn.lineplot", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 147, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 148, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 148, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 149, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 151, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 163, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 163, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 164, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 165, "usage_type": "name"}, {"api_name": "seaborn.distplot", "line_number": 167, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 174, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 174, "usage_type": "name"}, {"api_name": "matplotlib.ticker.FormatStrFormatter", "line_number": 174, "usage_type": "call"}, {"api_name": "matplotlib.ticker", "line_number": 174, "usage_type": "name"}, {"api_name": "seaborn.set_style", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 181, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 181, "usage_type": "name"}, {"api_name": "matplotlib.gridspec.GridSpec", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.gridspec", "line_number": 182, "usage_type": "name"}, {"api_name": "seaborn.jointplot", "line_number": 185, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 201, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 201, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 202, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 202, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 203, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 203, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 205, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 217, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 217, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 218, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 218, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 219, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 219, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 220, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 220, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 221, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 221, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 231, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 233, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 233, "usage_type": "name"}, {"api_name": "minepy.MINE", "line_number": 239, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 247, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 247, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 248, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 248, "usage_type": "name"}, {"api_name": "statsmodels.graphics.tsaplots.plot_acf", "line_number": 250, "usage_type": "call"}, {"api_name": "statsmodels.graphics.tsaplots.plot_pacf", "line_number": 251, "usage_type": "call"}, {"api_name": "sklearn.metrics.pairwise.cosine_similarity", "line_number": 264, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 265, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 266, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 266, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 267, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 267, "usage_type": "name"}, {"api_name": "seaborn.heatmap", "line_number": 268, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 269, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 269, "usage_type": "name"}, {"api_name": "sklearn.metrics.pairwise.cosine_similarity", "line_number": 276, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 277, "usage_type": "call"}, {"api_name": "scipy.stats.pearsonr", "line_number": 278, "usage_type": "call"}, {"api_name": "scipy.stats", "line_number": 278, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 283, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 283, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 284, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 284, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 285, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 285, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 286, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 286, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 287, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 287, "usage_type": "name"}, {"api_name": "pandas.read_excel", "line_number": 296, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 333, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 333, "usage_type": "name"}]}
+{"seq_id": "300310141", "text": "# -*- coding: utf-8 -*-\nimport scrapy\nfrom scrapy.linkextractors import LinkExtractor\nfrom scrapy.spiders import CrawlSpider, Rule\nfrom Sun.items import SunItem\n\n\nclass DongguanSpider(CrawlSpider):\n\tname = 'dongguan'\n\tallowed_domains = ['wz.sun0769.com']\n\tstart_urls = ['http://wz.sun0769.com/index.php/question/questionType?type=4']\n\n\trules = (\n\t\t# LinkExtractor():从响应中根据正则提取链接,follow指后面网页是否匹配此正则\n\t\t# 获取详情页url\n\t\tRule(LinkExtractor(allow=r'html/question/\\d+/\\d+.shtml'), callback='parse_item', follow=False),\n\t\tRule(LinkExtractor(allow=r'questionType'), follow=True),\n\t)\n\t\t# '''http://wz.sun0769.com/index.php/question/questionType?type=4&page=30'''\n \t# http://wz.sun0769.com/html/question/201708/343828.shtml\n \t# http://wz.sun0769.com/html/question/201708/343815.shtml\n\t\n\tdef parse_item(self, response):\n\t\t# 测试是否获取到了url\n\t\t# print(\"*\"*20)\n\t\t# print(response.url)\n\t\t# 实例化模型对象\n\t\titem = SunItem()\n\t\titem['id'] = response.xpath(\"/html/body/div[6]/div/div[1]/div[1]/strong/text()\").extract_first().split(':')[-1]\n\t\titem['detail_url'] = response.url\n\t\t# 获取标题\n\t\titem['title'] = response.xpath(\"/html/head/title/text()\").extract_first().split(\"_\")[0]\n\t\t# 获取内容,分为有图片和没有图片,内容为列表,用extract(),并用.join(),进行字符串拼接\n\t\tcontent = ''.join(response.xpath(\"/html/body/div[6]/div/div[2]/div[1]/text()\").extract())\n\t\t# 字符串的strip()方法,取出两边空格\n\t\tif content.strip() == '':\n\t\t\t# 如果为空,证明提交内容有图片,重新xpath\n\t\t\tcontent = ''.join(response.xpath(\"/html/body/div[6]/div/div[2]/div[1]/div[3]/text()\").extract())\n\t\t\n\t\titem['content'] = content\n\t\t# 将数据返回给引擎\n\t\tyield item\n", "sub_path": "Sun/Sun/spiders/dongguan.py", "file_name": "dongguan.py", "file_ext": "py", "file_size_in_byte": 1758, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "scrapy.spiders.CrawlSpider", "line_number": 8, "usage_type": "name"}, {"api_name": "scrapy.spiders.Rule", "line_number": 16, "usage_type": "call"}, {"api_name": "scrapy.linkextractors.LinkExtractor", "line_number": 16, "usage_type": "call"}, {"api_name": "scrapy.spiders.Rule", "line_number": 17, "usage_type": "call"}, {"api_name": "scrapy.linkextractors.LinkExtractor", "line_number": 17, "usage_type": "call"}, {"api_name": "Sun.items.SunItem", "line_number": 28, "usage_type": "call"}]}
+{"seq_id": "44927642", "text": "\"\"\"Initial expense accounting database population\"\"\"\nfrom datetime import date\n\nfrom sqlalchemy import create_engine\nfrom sqlalchemy.orm import sessionmaker\n\nfrom database_setup import Base, Account, Transaction\n\nengine = create_engine('sqlite:///expense_accounting.db')\nBase.metadata.bind = engine\nDBSession = sessionmaker(bind=engine)\n\nsession = DBSession()\n\naccount_payable = Account(name=\"payable\")\naccount_receivable = Account(name=\"receivable\")\naccount_cash_register = Account(name=\"cash_register\")\naccount_rafael = Account(name=\"rafael\")\naccount_mirjam = Account(name=\"mirjam\")\naccount_mama = Account(name=\"mama\")\naccount_manuela = Account(name=\"manuela\")\n\nsession.add(account_manuela)\nsession.add(account_mama)\nsession.add(account_receivable)\nsession.add(account_mirjam)\nsession.add(account_rafael)\nsession.add(account_cash_register)\nsession.add(account_payable)\nsession.flush()\ntransaction_1 = Transaction(account_credit=account_mama.id,\n account_debit=account_payable.id,\n amount=15.16,\n reason=\"Einkauf\",\n transaction_date=date.today())\nsession.add(transaction_1)\n\ntransaction_2 = Transaction(account_credit=account_cash_register.id,\n account_debit=account_mama.id,\n amount=15.16,\n reason=\"Überweisung\",\n transaction_date=date.today())\nsession.add(transaction_2)\n\ntransaction_3 = Transaction(account_credit=account_mama.id,\n account_debit=account_mama.id,\n amount=15.16,\n reason=\"Überweisung\",\n transaction_date=date.today())\nsession.add(transaction_3)\n\nsession.commit()\n", "sub_path": "populate_expense_accounting_db.py", "file_name": "populate_expense_accounting_db.py", "file_ext": "py", "file_size_in_byte": 1808, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 9, "usage_type": "call"}, {"api_name": "database_setup.Base.metadata", "line_number": 10, "usage_type": "attribute"}, {"api_name": "database_setup.Base", "line_number": 10, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.sessionmaker", "line_number": 11, "usage_type": "call"}, {"api_name": "database_setup.Account", "line_number": 15, "usage_type": "call"}, {"api_name": "database_setup.Account", "line_number": 16, "usage_type": "call"}, {"api_name": "database_setup.Account", "line_number": 17, "usage_type": "call"}, {"api_name": "database_setup.Account", "line_number": 18, "usage_type": "call"}, {"api_name": "database_setup.Account", "line_number": 19, "usage_type": "call"}, {"api_name": "database_setup.Account", "line_number": 20, "usage_type": "call"}, {"api_name": "database_setup.Account", "line_number": 21, "usage_type": "call"}, {"api_name": "database_setup.Transaction", "line_number": 31, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 35, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 35, "usage_type": "name"}, {"api_name": "database_setup.Transaction", "line_number": 38, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 42, "usage_type": "name"}, {"api_name": "database_setup.Transaction", "line_number": 45, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 49, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 49, "usage_type": "name"}]}
+{"seq_id": "342223122", "text": "# -*- coding:utf-8 -*-\n\nfrom tornado.web import Application,RequestHandler\nfrom tornado.gen import IOLoop,coroutine\nfrom tornado.httpclient import AsyncHTTPClient,HTTPRequest\nfrom tornado.options import parse_command_line,options,define\nfrom tornado.escape import native_str\nfrom tornado.log import gen_log\n\ndefine('port',default=8888,type=int,help='listen at the port,default is 8888')\n\n\nclass ProxyHandler(RequestHandler):\n\n def prepare(self):\n gen_log.info('headers:{}'.format(self.request.headers))\n\n @coroutine\n def get(self, *args, **kwargs):\n\n yield self.forward(*args,**kwargs)\n\n @coroutine\n def post(self, *args, **kwargs):\n\n gen_log.info('data:{}'.format(self.request.body))\n yield self.forward(*args, **kwargs)\n\n\n\n\n @coroutine\n def forward(self,*args,**kwargs):\n #headers=[\"%s: %s\" % (native_str(k), native_str(v))\n # for k, v in self.request.headers.get_all()]\n headers=self.request.headers\n #print(native_str(headers['User-Agent']))\n headers['User-Agent']='Mozilla/5.0 (Android; U; Android; en-US; rv:0.9.4)'\n path=self.request.uri.replace('{}://{}'.format(self.request.protocol,self.request.host),'')\n if self.request.host.endswith(':'):\n uri = '{}://{}{}{}'.format(self.request.protocol,self.request.host,80,path)\n else:\n uri='{}://{}{}'.format(self.request.protocol,self.request.host,path)\n # res=yield AsyncHTTPClient.fetch(uri,self.request.headers)\n #headers=self.request.headers\n #self.request.headers['Content-Type']='Custome Proxy'\n response = yield AsyncHTTPClient().fetch(\n HTTPRequest(url=uri,\n method=self.request.method,\n body=self.request.body if self.request.method.upper() != \"GET\" else None,\n headers=headers,\n follow_redirects=True)\n )\n gen_log.info(response.body)\n self.write(response.body)\n\ndef main():\n parse_command_line()\n AsyncHTTPClient.configure('tornado.curl_httpclient.CurlAsyncHTTPClient')\n settings={\n 'debug':True,\n 'default_handler_class':ProxyHandler\n }\n app=Application(handlers=[\n (r'/',ProxyHandler)\n ],**settings)\n app.listen(options.port)\n gen_log.info('listen at {}'.format(options.port))\n IOLoop.current().start()\n\n\nif __name__=='__main__':\n main()\n\n", "sub_path": "002_python_study/tornado_proxy_demo/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 2433, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "tornado.options.define", "line_number": 10, "usage_type": "call"}, {"api_name": "tornado.web.RequestHandler", "line_number": 13, "usage_type": "name"}, {"api_name": "tornado.log.gen_log.info", "line_number": 16, "usage_type": "call"}, {"api_name": "tornado.log.gen_log", "line_number": 16, "usage_type": "name"}, {"api_name": "tornado.gen.coroutine", "line_number": 18, "usage_type": "name"}, {"api_name": "tornado.log.gen_log.info", "line_number": 26, "usage_type": "call"}, {"api_name": "tornado.log.gen_log", "line_number": 26, "usage_type": "name"}, {"api_name": "tornado.gen.coroutine", "line_number": 23, "usage_type": "name"}, {"api_name": "tornado.httpclient.AsyncHTTPClient", "line_number": 47, "usage_type": "call"}, {"api_name": "tornado.httpclient.HTTPRequest", "line_number": 48, "usage_type": "call"}, {"api_name": "tornado.log.gen_log.info", "line_number": 54, "usage_type": "call"}, {"api_name": "tornado.log.gen_log", "line_number": 54, "usage_type": "name"}, {"api_name": "tornado.gen.coroutine", "line_number": 32, "usage_type": "name"}, {"api_name": "tornado.options.parse_command_line", "line_number": 58, "usage_type": "call"}, {"api_name": "tornado.httpclient.AsyncHTTPClient.configure", "line_number": 59, "usage_type": "call"}, {"api_name": "tornado.httpclient.AsyncHTTPClient", "line_number": 59, "usage_type": "name"}, {"api_name": "tornado.web.Application", "line_number": 64, "usage_type": "call"}, {"api_name": "tornado.options.options.port", "line_number": 67, "usage_type": "attribute"}, {"api_name": "tornado.options.options", "line_number": 67, "usage_type": "name"}, {"api_name": "tornado.log.gen_log.info", "line_number": 68, "usage_type": "call"}, {"api_name": "tornado.log.gen_log", "line_number": 68, "usage_type": "name"}, {"api_name": "tornado.options.options.port", "line_number": 68, "usage_type": "attribute"}, {"api_name": "tornado.options.options", "line_number": 68, "usage_type": "name"}, {"api_name": "tornado.gen.IOLoop.current", "line_number": 69, "usage_type": "call"}, {"api_name": "tornado.gen.IOLoop", "line_number": 69, "usage_type": "name"}]}
+{"seq_id": "369286301", "text": "from django.conf.urls import url\nfrom django.views.decorators.cache import never_cache\n\nfrom .views import (AccountCreateView, AccountDeleteView,\n AccountEditView, AccountListView, ContactView,\n InstitutionPaymentsView, MigrateDataView,\n MigrateOptionsView, MigrateVersionView,\n PendingAccountDeleteView, ResponsiblePartyCreateView,\n ResponsiblePartyDeleteView, ResponsiblePartyEditView,\n ResponsiblePartyListView, ShareDataView,\n ShareThirdPartiesView, SnapshotCSVExportView,\n SnapshotCSVDownloadView, SnapshotPDFExportView,\n SnapshotPDFDownloadView)\n\napp_name = 'manage'\n\nurlpatterns = [\n url(r'^contact/$', ContactView.as_view(), name='institution-contact'),\n\n url(r'^payments/$', InstitutionPaymentsView.as_view(),\n name='institution-payments'),\n\n # Responsible Party views:\n url(r'^responsible-party/$',\n ResponsiblePartyListView.as_view(),\n name='responsible-party-list'),\n\n url(r'^responsible-party/create/$',\n ResponsiblePartyCreateView.as_view(),\n name='responsible-party-create'),\n\n url(r'^responsible-party/(?P\\d+)/edit/$',\n ResponsiblePartyEditView.as_view(),\n name='responsible-party-edit'),\n\n url(r'^responsible-party/(?P\\d+)/delete/$',\n ResponsiblePartyDeleteView.as_view(),\n name='responsible-party-delete'),\n\n # User/Account views:\n url(r'^user/$', never_cache(AccountListView.as_view()),\n name='account-list'),\n\n url(r'^user/create/$', AccountCreateView.as_view(),\n name='account-create'),\n\n url(r'^user/(?P\\d+)/edit/$', AccountEditView.as_view(),\n name='account-edit'),\n\n url(r'^user/(?P\\d+)/delete/$', AccountDeleteView.as_view(),\n name='account-delete'),\n\n url(r'^pending-user/(?P\\d+)/delete/$',\n PendingAccountDeleteView.as_view(),\n name='pending-account-delete'),\n\n # Share Data views:\n url(r'^share-data/snapshot-archive/$', ShareDataView.as_view(),\n name='share-data'),\n\n url(r'^share-data/snapshot-archive/(?P[^/]+)/csv/$',\n never_cache(SnapshotCSVExportView.as_view()),\n name='snapshot-export-csv'),\n url(r'^share-data/snapshot-archive/(?P[^/]+)/csv/download/(?P[^/]+)/$',\n never_cache(SnapshotCSVDownloadView.as_view()),\n name='snapshot-download-csv'),\n\n url(r'^share-data/snapshot-archive/(?P[^/]+)/pdf/$',\n never_cache(SnapshotPDFExportView.as_view()),\n name='snapshot-export-pdf'),\n url(r'^share-data/snapshot-archive/(?P[^/]+)/pdf/download/(?P[^/]+)/$',\n never_cache(SnapshotPDFDownloadView.as_view()),\n name='snapshot-download-pdf'),\n\n url(r'^share-data/$', ShareThirdPartiesView.as_view(),\n name='share-third-parties'),\n\n # Migration views:\n url(r'^migrate/$', MigrateOptionsView.as_view(),\n name='migrate-options'),\n\n url(r'^migrate/data/(?P\\d+)/$', MigrateDataView.as_view(),\n name='migrate-data'),\n\n url(r'^migrate/version/(?P\\d+)/$', MigrateVersionView.as_view(),\n name='migrate-version')\n]\n", "sub_path": "stars/apps/tool/manage/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 3268, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.conf.urls.url", "line_number": 18, "usage_type": "call"}, {"api_name": "views.ContactView.as_view", "line_number": 18, "usage_type": "call"}, {"api_name": "views.ContactView", "line_number": 18, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 20, "usage_type": "call"}, {"api_name": "views.InstitutionPaymentsView.as_view", "line_number": 20, "usage_type": "call"}, {"api_name": "views.InstitutionPaymentsView", "line_number": 20, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 24, "usage_type": "call"}, {"api_name": "views.ResponsiblePartyListView.as_view", "line_number": 25, "usage_type": "call"}, {"api_name": "views.ResponsiblePartyListView", "line_number": 25, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 28, "usage_type": "call"}, {"api_name": "views.ResponsiblePartyCreateView.as_view", "line_number": 29, "usage_type": "call"}, {"api_name": "views.ResponsiblePartyCreateView", "line_number": 29, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 32, "usage_type": "call"}, {"api_name": "views.ResponsiblePartyEditView.as_view", "line_number": 33, "usage_type": "call"}, {"api_name": "views.ResponsiblePartyEditView", "line_number": 33, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 36, "usage_type": "call"}, {"api_name": "views.ResponsiblePartyDeleteView.as_view", "line_number": 37, "usage_type": "call"}, {"api_name": "views.ResponsiblePartyDeleteView", "line_number": 37, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 41, "usage_type": "call"}, {"api_name": "django.views.decorators.cache.never_cache", "line_number": 41, "usage_type": "call"}, {"api_name": "views.AccountListView.as_view", "line_number": 41, "usage_type": "call"}, {"api_name": "views.AccountListView", "line_number": 41, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 44, "usage_type": "call"}, {"api_name": "views.AccountCreateView.as_view", "line_number": 44, "usage_type": "call"}, {"api_name": "views.AccountCreateView", "line_number": 44, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 47, "usage_type": "call"}, {"api_name": "views.AccountEditView.as_view", "line_number": 47, "usage_type": "call"}, {"api_name": "views.AccountEditView", "line_number": 47, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 50, "usage_type": "call"}, {"api_name": "views.AccountDeleteView.as_view", "line_number": 50, "usage_type": "call"}, {"api_name": "views.AccountDeleteView", "line_number": 50, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 53, "usage_type": "call"}, {"api_name": "views.PendingAccountDeleteView.as_view", "line_number": 54, "usage_type": "call"}, {"api_name": "views.PendingAccountDeleteView", "line_number": 54, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 58, "usage_type": "call"}, {"api_name": "views.ShareDataView.as_view", "line_number": 58, "usage_type": "call"}, {"api_name": "views.ShareDataView", "line_number": 58, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 61, "usage_type": "call"}, {"api_name": "django.views.decorators.cache.never_cache", "line_number": 62, "usage_type": "call"}, {"api_name": "views.SnapshotCSVExportView.as_view", "line_number": 62, "usage_type": "call"}, {"api_name": "views.SnapshotCSVExportView", "line_number": 62, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 64, "usage_type": "call"}, {"api_name": "django.views.decorators.cache.never_cache", "line_number": 65, "usage_type": "call"}, {"api_name": "views.SnapshotCSVDownloadView.as_view", "line_number": 65, "usage_type": "call"}, {"api_name": "views.SnapshotCSVDownloadView", "line_number": 65, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 68, "usage_type": "call"}, {"api_name": "django.views.decorators.cache.never_cache", "line_number": 69, "usage_type": "call"}, {"api_name": "views.SnapshotPDFExportView.as_view", "line_number": 69, "usage_type": "call"}, {"api_name": "views.SnapshotPDFExportView", "line_number": 69, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 71, "usage_type": "call"}, {"api_name": "django.views.decorators.cache.never_cache", "line_number": 72, "usage_type": "call"}, {"api_name": "views.SnapshotPDFDownloadView.as_view", "line_number": 72, "usage_type": "call"}, {"api_name": "views.SnapshotPDFDownloadView", "line_number": 72, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 75, "usage_type": "call"}, {"api_name": "views.ShareThirdPartiesView.as_view", "line_number": 75, "usage_type": "call"}, {"api_name": "views.ShareThirdPartiesView", "line_number": 75, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 79, "usage_type": "call"}, {"api_name": "views.MigrateOptionsView.as_view", "line_number": 79, "usage_type": "call"}, {"api_name": "views.MigrateOptionsView", "line_number": 79, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 82, "usage_type": "call"}, {"api_name": "views.MigrateDataView.as_view", "line_number": 82, "usage_type": "call"}, {"api_name": "views.MigrateDataView", "line_number": 82, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 85, "usage_type": "call"}, {"api_name": "views.MigrateVersionView.as_view", "line_number": 85, "usage_type": "call"}, {"api_name": "views.MigrateVersionView", "line_number": 85, "usage_type": "name"}]}
+{"seq_id": "175606274", "text": "import configparser\n# import sys\n# import os\nconfig = configparser.ConfigParser()\nconfig.read(\"config.cfg\")\nconfig.sections()\nconfig\n\n\n# Ez az alap config lekérdező, kifejtve\ndef configSectionMap(section):\n global config\n dict1 = {}\n options = config.options(section)\n for option in options:\n try:\n dict1[option] = config.get(section, option)\n if dict1[option] == -1:\n print(\"skip: %s\" % option)\n except:\n print(\"exception on %s!\" % option)\n dict1[option] = None\n return dict1\n\n\n# Ez meg a leegyszerűsített változata\ndef configLekeres(szekcio2,mezo):\n Szekcio = configSectionMap(szekcio2)\n # print(Cim) #Ez kilistázza a szekció mezőit és értékeit\n # print(str(mezo[szekcio]))\n if Szekcio[mezo] == 'True':\n return True\n elif Szekcio[mezo] == 'False':\n return False\n elif Szekcio[mezo] == 'None':\n return None\n else:\n return Szekcio[mezo]\n\n\ndef configModosito(szekcio3, mezo1, ujErtek):\n # lets create that config file for next time...\n global config\n cfgfile = open(\"config.cfg\", 'w')\n if ujErtek is not None:\n config.set(szekcio3, mezo1, ujErtek)\n config.write(cfgfile)\n cfgfile.close()\n\n\n# ==================== Na innentől csak teszt funkciók vannak ====================\ndef boolTesztelo():\n oprendszer = configLekeres('alap', 'oprendszer')\n print(\"Oprendszer tipusa:\", type(oprendszer))\n if not oprendszer:\n print(\"Natív False\")\n print(type(oprendszer))\n elif oprendszer:\n oprendszer = bool(oprendszer)\n print(\"Natív True\")\n print(type(oprendszer))\n else:\n print(\"Oprendszer:\"+str(oprendszer)+\"<\")\n\n\ndef tesztIrasOlvasas():\n boolTesztelo()\n configModosito(\"alap\", \"oprendszer\", 'True')\n boolTesztelo()\n configModosito(\"alap\", \"oprendszer\", 'False')\n boolTesztelo()\n", "sub_path": "lib/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1924, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "configparser.ConfigParser", "line_number": 4, "usage_type": "call"}]}
+{"seq_id": "563823918", "text": "from django.contrib import admin\nfrom enquete.models import Pergunta, Escolha\n\nclass EscolhaTabular(admin.TabularInline):\n model = Escolha\n extra = 3 \n\nclass EnqueteAdmin(admin.ModelAdmin):\n inlines = [EscolhaTabular]\n search_fields = ['pergunta']\n date_hierarchy = 'data_hora'\n list_filter = ['data_hora']\n\nadmin.site.register(Pergunta, EnqueteAdmin)\n", "sub_path": "site1/.enquete/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 370, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.contrib.admin.TabularInline", "line_number": 4, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 4, "usage_type": "name"}, {"api_name": "enquete.models.Escolha", "line_number": 5, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 8, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 14, "usage_type": "call"}, {"api_name": "enquete.models.Pergunta", "line_number": 14, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 14, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 14, "usage_type": "name"}]}
+{"seq_id": "637136536", "text": "import json\nfrom AboutMe.AboutMe import AboutMe\n\n\nclass Edit(AboutMe):\n def edit():\n # This variable is used to keep track whether data is available in database\n found = False\n uid = input('Please enter your unique identification number:')\n # This list is used to hold the data temporarily for editing\n datalist = []\n with open('../sample.json') as file:\n for obj in file:\n data = json.loads(obj)\n if uid in data.values():\n print(data)\n found = True\n objt = AboutMe(data['uid'], data['name'], data['profession'],\n data['contact'], data['skills'], data['DOB'],\n data['description'])\n else:\n datalist.append(data)\n\n if found:\n objt.edit_data()\n data = objt.__dict__\n datalist.append(data)\n with open('../sample.json', 'w') as file:\n for data in datalist:\n json.dump(data, file)\n file.write('\\n')\n\n else:\n print('Unique identification number is not found')\n", "sub_path": "Assign_1/Edit/Edit.py", "file_name": "Edit.py", "file_ext": "py", "file_size_in_byte": 1221, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "AboutMe.AboutMe.AboutMe", "line_number": 5, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 14, "usage_type": "call"}, {"api_name": "AboutMe.AboutMe.AboutMe", "line_number": 18, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 30, "usage_type": "call"}]}
+{"seq_id": "596878300", "text": "#Author : Dhaval Harish Sharma\n#Red ID : 824654344\n#Assignment 4, Question 2(A)\n\"\"\"Apply Butterworth low pass filter with an appropriate order and cutoff frequency \n(you choose) to the above three images. Present the Fourier spectrum (log magnitude) and \nfiltered images in your report and discuss the effect of this filter on the above three \nimages including the effect of the cutoff frequency.\"\"\"\n\n#Importing the required libraries\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport math\nfrom skimage.color import rgb2gray\n\n#Taking the image from user\nprint(\"The available images: 1) Blonde1 2) Blonde2 3) Zebra\")\ntemp = input(\"Please enter the image to filter:\")\nif temp == \"1\":\n in_img = plt.imread(\"Blonde1.jpg\")\nelif temp == \"2\":\n in_img = plt.imread(\"Blonde2.jpg\")\nelif temp == \"3\":\n in_img = plt.imread(\"Zebra.jpg\")\n in_img = (rgb2gray(in_img) * 255).astype(np.uint8)\nelse:\n print(\"Please enter correct name!\")\n\n#Padding zeros to make the dimensions of image a power of 2\ndef next_pow(num):\n return math.ceil(math.log(num,2))\n\nnextpow = next_pow(max(in_img.shape[0], in_img.shape[1]))\npadded_img = np.zeros(shape = (2**nextpow, 2**nextpow), dtype = np.uint8)\nfor i in range(in_img.shape[0]):\n for j in range(in_img.shape[1]):\n padded_img[i][j] = in_img[i][j]\n\n#Making the butterworth low pass filter\nx = np.linspace(-padded_img.shape[0] // 2, padded_img.shape[0] // 2 - 1, padded_img.shape[0])\ny = np.linspace(-padded_img.shape[1] // 2, padded_img.shape[1] //2 - 1, padded_img.shape[1])\n[u, v] = np.meshgrid(x, y)\nr_sq = u**2 + v**2\n#bwlpf = 1/(1 + (r_sq/(r)**2)**p)\nbwlpf = 1/(1 + (r_sq/(50)**2)**3)\n\n#Applying FFT on the images\npadded_img_fft = np.fft.fftshift(np.fft.fft2(padded_img))\npadded_img_filtered = padded_img_fft * bwlpf\npadded_img_inverse = np.fft.ifft2(padded_img_filtered)\n\n#Getting the output image from padded image\nout_img = np.zeros(shape = (in_img.shape[0], in_img.shape[1]), dtype = padded_img_inverse.dtype)\nfor i in range(in_img.shape[0]):\n for j in range(in_img.shape[1]):\n out_img[i][j] = padded_img_inverse[i][j]\n\n#Printing the output image\nfig, ax = plt.subplots(nrows = 2, ncols = 3)\nax[0][0].imshow(in_img, cmap = 'gray')\nax[0][1].imshow(bwlpf, cmap = plt.cm.gray)\nax[0][2].imshow((np.log(1 + np.abs(padded_img_fft))).astype(int), cmap = plt.cm.gray)\nax[1][0].imshow((np.log(1 + np.abs(padded_img_filtered))).astype(int), cmap = plt.cm.gray)\nax[1][1].imshow(np.abs(padded_img_inverse).astype(int), cmap = plt.cm.gray)\nax[1][2].imshow(np.abs(out_img).astype(int), cmap = plt.cm.gray)", "sub_path": "Assignment 4/Question2A.py", "file_name": "Question2A.py", "file_ext": "py", "file_size_in_byte": 2565, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "matplotlib.pyplot.imread", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imread", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imread", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "skimage.color.rgb2gray", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 24, "usage_type": "attribute"}, {"api_name": "math.ceil", "line_number": 30, "usage_type": "call"}, {"api_name": "math.log", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 33, "usage_type": "attribute"}, {"api_name": "numpy.linspace", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.fft.fftshift", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 47, "usage_type": "attribute"}, {"api_name": "numpy.fft.fft2", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.fft.ifft2", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 49, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 60, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "numpy.log", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 61, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "numpy.log", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 62, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "numpy.abs", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 63, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "numpy.abs", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 64, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}]}
+{"seq_id": "488677475", "text": "from config import Config\nfrom app.routingmodel import RoutingModel\nfrom app.command_executor import OsCommandExecutor\nfrom app.command_executor import PrintCommandExecutor\nfrom app import routablemodel\nimport os\n\nclass ViewModel:\n\n def __init__(self):\n if Config.TEST_ENVIRONMENT: \n command_executor = PrintCommandExecutor()\n else:\n command_executor = OsCommandExecutor()\n self.routable_model = routablemodel.RoutableModel(command_executor)\n\n def get_devices(self):\n devices = []\n for k,v in self.routable_model.leases.items():\n print (v)\n devices.append(\n {\n 'mac' : v.ethernet,\n 'ipv4' : v.ip, \n 'description':v.hostname,\n 'route_to': self.get_route(v.ip)\n })\n return devices\n\n def set_route(self, ip, route:int):\n self.routable_model.set_route(ip, route)\n\n def get_route(self, ip):\n return self.routable_model.get_route(ip)\n", "sub_path": "router/routeconfigurator/app/viewmodel.py", "file_name": "viewmodel.py", "file_ext": "py", "file_size_in_byte": 1028, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "config.Config.TEST_ENVIRONMENT", "line_number": 11, "usage_type": "attribute"}, {"api_name": "config.Config", "line_number": 11, "usage_type": "name"}, {"api_name": "app.command_executor.PrintCommandExecutor", "line_number": 12, "usage_type": "call"}, {"api_name": "app.command_executor.OsCommandExecutor", "line_number": 14, "usage_type": "call"}, {"api_name": "app.routablemodel.RoutableModel", "line_number": 15, "usage_type": "call"}, {"api_name": "app.routablemodel", "line_number": 15, "usage_type": "name"}]}
+{"seq_id": "292627411", "text": "import datetime\nimport os\nimport pprint\nimport time\nimport threading\nimport torch as th\nimport dill\nimport numpy as np\nfrom types import SimpleNamespace as SN\nfrom utils.logging import Logger\nfrom utils.timehelper import time_left, time_str\nfrom os.path import dirname, abspath\n\nfrom learners import REGISTRY as le_REGISTRY\nfrom runners import REGISTRY as r_REGISTRY\nfrom controllers import REGISTRY as mac_REGISTRY\nfrom components.episode_buffer import ReplayBuffer\nfrom components.transforms import OneHot\n\nimport collections\ndef recursive_dict_update(d, u):\n for k, v in u.items():\n if isinstance(v, collections.Mapping):\n d[k] = recursive_dict_update(d.get(k, {}), v)\n else:\n d[k] = v\n return d\n\nimport yaml\ndef _get_config(params, arg_name, subfolder):\n config_name = None\n for _i, _v in enumerate(params):\n if _v.split(\"=\")[0] == arg_name:\n config_name = _v.split(\"=\")[1]\n del params[_i]\n break\n\n if config_name is not None:\n with open(os.path.join(os.path.dirname(__file__), \"config\", subfolder, \"{}.yaml\".format(config_name)), \"r\") as f:\n try:\n config_dict = yaml.load(f)\n except yaml.YAMLError as exc:\n assert False, \"{}.yaml error: {}\".format(config_name, exc)\n return config_dict\n\ndef run(_run, _config, _log):\n\n # check args sanity\n _config = args_sanity_check(_config, _log)\n\n args = SN(**_config)\n args.device = \"cuda\" if args.use_cuda else \"cpu\"\n\n # setup loggers\n logger = Logger(_log)\n\n _log.info(\"Experiment Parameters:\")\n experiment_params = pprint.pformat(_config,\n indent=4,\n width=1)\n _log.info(\"\\n\\n\" + experiment_params + \"\\n\")\n\n # configure tensorboard logger\n unique_token = \"{}__{}\".format(args.name, datetime.datetime.now().strftime(\"%Y-%m-%d_%H-%M-%S\"))\n args.unique_token = unique_token\n if args.use_tensorboard:\n tb_logs_direc = os.path.join(dirname(dirname(abspath(__file__))), \"results\", \"tb_logs\")\n tb_exp_direc = os.path.join(tb_logs_direc, \"{}\").format(unique_token)\n logger.setup_tb(tb_exp_direc)\n\n # sacred is on by default\n logger.setup_sacred(_run)\n\n # Run and train\n if args.meta == \"reptile\":\n run_reptile(args=args, logger=logger, _log=_log, _run=_run)\n\n else:\n run_sequential(args=args, logger=logger)\n\n # Clean up after finishing\n print(\"Exiting Main\")\n\n print(\"Stopping all threads\")\n for t in threading.enumerate():\n if t.name != \"MainThread\":\n print(\"Thread {} is alive! Is daemon: {}\".format(t.name, t.daemon))\n t.join(timeout=1)\n print(\"Thread joined\")\n\n print(\"Exiting script\")\n\n # Making sure framework really exits\n os._exit(os.EX_OK)\n\n\ndef evaluate_sequential(args, runner):\n\n for _ in range(args.test_nepisode):\n runner.run(test_mode=True)\n\n if args.save_replay:\n runner.save_replay()\n\n runner.close_env()\n\ndef run_sequential(args, logger):\n\n # Init runner so we can get env info\n runner = r_REGISTRY[args.runner](args=args, logger=logger)\n\n # Set up schemes and groups here\n env_info = runner.get_env_info()\n args.n_agents = env_info[\"n_agents\"]\n args.n_actions = env_info[\"n_actions\"]\n args.obs_decoder = dill.loads(env_info[\"obs_decoder\"]) if env_info[\"obs_decoder\"] is not None else None\n args.avail_actions_encoder = dill.loads(env_info[\"avail_actions_encoder_grid\"]) if env_info[\"avail_actions_encoder_grid\"] is not None else None\n\n args.state_shape = env_info[\"state_shape\"]\n\n # Default/Base scheme\n scheme = {\n \"state\": {\"vshape\": env_info[\"state_shape\"]},\n \"obs\": {\"vshape\": env_info[\"obs_shape\"], \"group\": \"agents\", \"vshape_decoded\": env_info.get(\"obs_shape_decoded\", env_info[\"obs_shape\"])},\n \"actions\": {\"vshape\": (1,), \"group\": \"agents\", \"dtype\": th.long},\n \"avail_actions\": {\"vshape\": (env_info[\"n_actions\"],), \"group\": \"agents\", \"dtype\": th.int},\n \"reward\": {\"vshape\": (1,)},\n \"terminated\": {\"vshape\": (1,), \"dtype\": th.uint8},\n }\n groups = {\n \"agents\": args.n_agents\n }\n preprocess = {\n \"actions\": (\"actions_onehot\", [OneHot(out_dim=args.n_actions)])\n }\n\n buffer = ReplayBuffer(scheme, groups, args.buffer_size, env_info[\"episode_limit\"] + 1,\n preprocess=preprocess,\n device=\"cpu\" if args.buffer_cpu_only else args.device)\n\n # Setup multiagent controller here\n mac = mac_REGISTRY[args.mac](buffer.scheme, groups, args)\n\n # Give runner the scheme\n runner.setup(scheme=scheme, groups=groups, preprocess=preprocess, mac=mac)\n\n # Learner\n learner = le_REGISTRY[args.learner](mac, buffer.scheme, logger, args)\n\n if args.use_cuda:\n learner.cuda()\n\n if args.checkpoint_path != \"\":\n\n timesteps = []\n timestep_to_load = 0\n\n if not os.path.isdir(args.checkpoint_path):\n logger.console_logger.info(\"Checkpoint directiory {} doesn't exist\".format(args.checkpoint_path))\n return\n\n # Go through all files in args.checkpoint_path\n for name in os.listdir(args.checkpoint_path):\n full_name = os.path.join(args.checkpoint_path, name)\n # Check if they are dirs the names of which are numbers\n if os.path.isdir(full_name) and name.isdigit():\n timesteps.append(int(name))\n\n if args.load_step == 0:\n # choose the max timestep\n timestep_to_load = max(timesteps)\n else:\n # choose the timestep closest to load_step\n timestep_to_load = min(timesteps, key=lambda x: abs(x - args.load_step))\n\n model_path = os.path.join(args.checkpoint_path, str(timestep_to_load))\n\n logger.console_logger.info(\"Loading model from {}\".format(model_path))\n learner.load_models(model_path)\n runner.t_env = timestep_to_load\n\n if args.evaluate or args.save_replay:\n evaluate_sequential(args, runner)\n return\n\n # start training\n episode = 0\n last_test_T = -args.test_interval - 1\n last_log_T = 0\n model_save_time = 0\n\n start_time = time.time()\n last_time = start_time\n\n logger.console_logger.info(\"Beginning training for {} timesteps\".format(args.t_max))\n\n\n\n\n while runner.t_env <= args.t_max:\n th.cuda.empty_cache()\n # Run for a whole episode at a time\n episode_batch = runner.run(test_mode=False)\n buffer.insert_episode_batch(episode_batch)\n del episode_batch\n\n if buffer.can_sample(args.batch_size):\n episode_sample = buffer.sample(args.batch_size)\n\n # Truncate batch to only filled timesteps\n max_ep_t = episode_sample.max_t_filled()\n episode_sample = episode_sample[:, :max_ep_t]\n\n if episode_sample.device != args.device:\n episode_sample.to(args.device)\n\n\n\n learner.train(episode_sample, runner.t_env, episode)\n th.cuda.empty_cache()\n\n # Execute test runs once in a while\n n_test_runs = max(1, args.test_nepisode // runner.batch_size)\n if (runner.t_env - last_test_T) / args.test_interval >= 1.0:\n\n logger.console_logger.info(\"t_env: {} / {}\".format(runner.t_env, args.t_max))\n logger.console_logger.info(\"Estimated time left: {}. Time passed: {}\".format(\n time_left(last_time, last_test_T, runner.t_env, args.t_max), time_str(time.time() - start_time)))\n last_time = time.time()\n\n last_test_T = runner.t_env\n for _ in range(n_test_runs):\n runner.run(test_mode=True)\n th.cuda.empty_cache()\n\n if args.save_model and (runner.t_env - model_save_time >= args.save_model_interval or model_save_time == 0):\n model_save_time = runner.t_env\n save_path = os.path.join(args.local_results_path, \"models\", args.unique_token, str(runner.t_env))\n #\"results/models/{}\".format(unique_token)\n os.makedirs(save_path, exist_ok=True)\n logger.console_logger.info(\"Saving models to {}\".format(save_path))\n\n # learner should handle saving/loading -- delegate actor save/load to mac,\n # use appropriate filenames to do critics, optimizer states\n learner.save_models(save_path)\n\n episode += args.batch_size_run\n th.cuda.empty_cache()\n\n if (runner.t_env - last_log_T) >= args.log_interval:\n logger.log_stat(\"episode\", episode, runner.t_env)\n logger.print_recent_stats()\n last_log_T = runner.t_env\n\n runner.close_env()\n logger.console_logger.info(\"Finished Training\")\n\n\ndef update_env(args, logger):\n temp = args.env_args[\"map_name\"]\n args.env_args[\"map_name\"] = args.env_args[\"map_name2\"]\n args.env_args[\"map_name2\"] = temp\n runner = r_REGISTRY[args.runner](args=args, logger=logger)\n env_info = runner.get_env_info()\n args.n_agents = env_info[\"n_agents\"]\n args.n_actions = env_info[\"n_actions\"]\n args.obs_decoder = dill.loads(env_info[\"obs_decoder\"]) if env_info[\"obs_decoder\"] is not None else None\n args.avail_actions_encoder = dill.loads(env_info[\"avail_actions_encoder_grid\"]) if env_info[\n \"avail_actions_encoder_grid\"] is not None else None\n\n args.state_shape = env_info[\"state_shape\"]\n\n # Default/Base scheme\n scheme = {\n \"state\": {\"vshape\": env_info[\"state_shape\"]},\n \"obs\": {\"vshape\": env_info[\"obs_shape\"], \"group\": \"agents\",\n \"vshape_decoded\": env_info.get(\"obs_shape_decoded\", env_info[\"obs_shape\"])},\n \"actions\": {\"vshape\": (1,), \"group\": \"agents\", \"dtype\": th.long},\n \"avail_actions\": {\"vshape\": (env_info[\"n_actions\"],), \"group\": \"agents\", \"dtype\": th.int},\n \"reward\": {\"vshape\": (1,)},\n \"terminated\": {\"vshape\": (1,), \"dtype\": th.uint8},\n }\n groups = {\n \"agents\": args.n_agents\n }\n preprocess = {\n \"actions\": (\"actions_onehot\", [OneHot(out_dim=args.n_actions)])\n }\n\n buffer = ReplayBuffer(scheme, groups, args.buffer_size, env_info[\"episode_limit\"] + 1,\n preprocess=preprocess,\n device=\"cpu\" if args.buffer_cpu_only else args.device)\n\n # Setup multiagent controller here\n mac = mac_REGISTRY[args.mac](buffer.scheme, groups, args)\n\n # Give runner the scheme\n runner.setup(scheme=scheme, groups=groups, preprocess=preprocess, mac=mac)\n\n # Learner\n learner = le_REGISTRY[args.learner](mac, buffer.scheme, logger, args)\n\n if args.use_cuda:\n learner.cuda()\n return runner, buffer, learner\n\ndef args_sanity_check(config, _log):\n\n # set CUDA flags\n # config[\"use_cuda\"] = True # Use cuda whenever possible!\n if config[\"use_cuda\"] and not th.cuda.is_available():\n config[\"use_cuda\"] = False\n _log.warning(\"CUDA flag use_cuda was switched OFF automatically because no CUDA devices are available!\")\n\n if config[\"test_nepisode\"] < config[\"batch_size_run\"]:\n config[\"test_nepisode\"] = config[\"batch_size_run\"]\n else:\n config[\"test_nepisode\"] = (config[\"test_nepisode\"]//config[\"batch_size_run\"]) * config[\"batch_size_run\"]\n\n return config\n\n\ndef run_reptile(args, logger, _log, _run):\n\n loggers = {}\n runners = {}\n macs = {}\n learners = {}\n buffers = {}\n\n agent_state_dict = None\n\n import yaml\n #from .main import _get_config\n # compile all the relevant task configs\n task_configs = {}\n\n class Bunch(object):\n def __init__(self, adict):\n self.__dict__.update(adict)\n\n r = np.random.RandomState(args.seed)\n for k, v in sorted(args.tasks.items()): # important for reproducibility of seeds!\n\n # Get the defaults from default.yaml\n with open(os.path.join(os.path.dirname(__file__), \"config\", \"default.yaml\"), \"r\") as f:\n try:\n config_dict = yaml.load(f)\n except yaml.YAMLError as exc:\n assert False, \"default.yaml error: {}\".format(exc)\n\n # Load algorithm and env base configs\n params = [\"\", \"--config={}\".format(v.pop(\"config\")), \"--env-config={}\".format(v.pop(\"env-config\"))]\n alg_config = _get_config(params, \"--config\", \"algs\")\n env_config = _get_config(params, \"--env-config\", \"envs\")\n\n # config_dict = {**config_dict, **env_config, **alg_config}\n config_dict = recursive_dict_update(config_dict, env_config)\n config_dict = recursive_dict_update(config_dict, alg_config)\n config_dict = recursive_dict_update(config_dict, v)\n\n # from src.utils.dict2namedtuple import convert\n config_dict.pop(\"no-mongo\")\n config_dict[\"seed\"] = r.randint(0, 2**31-1) # have to set manually\n config_dict[\"env_args\"][\"seed\"] = r.randint(0, 2**31-1)\n config_dict[\"device\"] = args.device\n config_dict[\"unique_token\"] = \"{}__{}\".format(args.unique_token,\n k)\n task_configs[k] = Bunch(config_dict)\n\n def setup_components(logger,\n agent_state_dict):\n task_names = []\n for task_name, _ in task_configs.items():\n task_names.append(task_name)\n\n # set up tasks based on the configs\n for task_name, task_config in task_configs.items():\n\n task_args = task_config\n\n from copy import deepcopy\n logger = Logger(_log)\n # sacred is on by default\n logger.setup_sacred(_run)\n # logger = deepcopy(meta_logger)\n logger.prefix = task_name\n loggers[task_name] = logger\n\n # Init runner so we can get env info\n runner = r_REGISTRY[task_args.runner](args=task_args,\n logger=logger)\n runners[task_name] = runner\n\n # Set up schemes and groups here\n env_info = runner.get_env_info()\n task_args.n_agents = env_info[\"n_agents\"]\n task_args.n_actions = env_info[\"n_actions\"]\n task_args.obs_decoder = dill.loads(env_info[\"obs_decoder\"]) if env_info[\"obs_decoder\"] is not None else None\n task_args.avail_actions_encoder = dill.loads(env_info[\"avail_actions_encoder_grid\"]) if env_info[\n \"avail_actions_encoder_grid\"] is not None else None\n task_args.db_url = args.db_url\n task_args.db_name = args.db_name\n task_args.state_shape = env_info[\"state_shape\"]\n task_args.state_decoder = dill.loads(env_info[\"state_decoder\"]) if env_info[\"state_decoder\"] is not None else None\n task_args.obs_decoder = dill.loads(env_info[\"obs_decoder\"]) if env_info[\"obs_decoder\"] is not None else None\n\n # Default/Base scheme\n scheme = {\n \"state\": {\"vshape\": env_info[\"state_shape\"]},\n \"obs\": {\"vshape\": env_info[\"obs_shape\"], \"group\": \"agents\",\n \"vshape_decoded\": env_info.get(\"obs_shape_decoded\", env_info[\"obs_shape\"])},\n \"actions\": {\"vshape\": (1,), \"group\": \"agents\", \"dtype\": th.long},\n \"avail_actions\": {\"vshape\": (env_info[\"n_actions\"],), \"group\": \"agents\", \"dtype\": th.int},\n \"reward\": {\"vshape\": (1,)},\n \"terminated\": {\"vshape\": (1,), \"dtype\": th.uint8},\n }\n groups = {\n \"agents\": task_args.n_agents\n }\n preprocess = {\n \"actions\": (\"actions_onehot\", [OneHot(out_dim=task_args.n_actions)])\n }\n\n buffer = ReplayBuffer(scheme, groups, task_args.buffer_size, env_info[\"episode_limit\"] + 1,\n preprocess=preprocess,\n device=\"cpu\" if task_args.buffer_cpu_only else args.device)\n buffers[task_name] = buffer\n\n # Setup multiagent controller here\n mac = mac_REGISTRY[task_args.mac](buffer.scheme, groups, task_args)\n\n #point model to same object\n macs[task_name] = mac\n mac.agent = macs[task_names[0]].agent\n\n # Give runner the scheme\n runner.setup(scheme=scheme, groups=groups, preprocess=preprocess, mac=mac)\n\n # Learner\n learner = le_REGISTRY[task_args.learner](mac, buffer.scheme, logger, task_args)\n learners[task_name] = learner\n\n if task_args.use_cuda:\n learner.cuda()\n\n #if agent_state_dict is None:\n # agent_state_dict = mac.agent.state_dict()\n # else:\n # # copy all weights that have same dimensions\n # sd = mac.agent.state_dict()\n # for k, v in agent_state_dict.items():\n # if (k in sd) and (v.shape == sd[k].shape):\n # setattr(mac.agent, k, v)\n\n\n if task_args.checkpoint_path != \"\":\n\n timesteps = []\n timestep_to_load = 0\n\n if not os.path.isdir(task_args.checkpoint_path):\n logger.console_logger.info(\"Checkpoint directory {} doesn't exist\".format(task_args.checkpoint_path))\n return\n\n # Go through all files in args.checkpoint_path\n for name in os.listdir(task_args.checkpoint_path):\n full_name = os.path.join(task_args.checkpoint_path, name)\n # Check if they are dirs the names of which are numbers\n if os.path.isdir(full_name) and name.isdigit():\n timesteps.append(int(name))\n\n if task_args.load_step == 0:\n # choose the max timestep\n timestep_to_load = max(timesteps)\n else:\n # choose the timestep closest to load_step\n timestep_to_load = min(timesteps, key=lambda x: abs(x - task_args.load_step))\n\n model_path = os.path.join(task_args.checkpoint_path, str(timestep_to_load))\n\n logger.console_logger.info(\"Loading model from {}\".format(model_path))\n learner.load_models(model_path)\n runner.t_env = timestep_to_load\n\n if task_args.evaluate or task_args.save_replay:\n evaluate_sequential(task_args, runner)\n return\n return\n\n\n from copy import deepcopy\n # agent_state_dict = setup_components(logger, agent_state_dict)\n setup_components(logger, agent_state_dict)\n\n # start reptile training\n episode_ctrs = {k:0 for k, _ in sorted(task_configs.items())}\n last_test_Ts = {k:-v.test_interval - 1 for k, v in sorted(task_configs.items())}\n last_times = {k:time.time() for k, v in sorted(task_configs.items())}\n model_save_times = {k:0 for k, _ in sorted(task_configs.items())}\n start_time = time.time()\n\n logger.console_logger.info(\"Beginning REPTILE training ...\")\n\n previous_task_id = None\n unfinished_tasks = {k for k, v in task_configs.items() if episode_ctrs[k] <= v.t_max}\n while len(unfinished_tasks):\n # INNER LOOP\n unfinished_tasks = {k for k, v in task_configs.items() if episode_ctrs[k] <=v.t_max}\n\n # pick task\n from random import randint\n task_id = sorted(list(unfinished_tasks))[randint(0, len(unfinished_tasks)-1)]\n\n logger.console_logger.info(\"Chose task {} at global counter {}\".format(task_id, sum(episode_ctrs.values())))\n\n # roll out task a couple of times\n for t in range(args.n_task_rollouts[task_id]):\n episode_batch = runners[task_id].run(test_mode=False)\n buffers[task_id].insert_episode_batch(episode_batch)\n # train on task\n episode_ctrs[task_id] += 1\n if episode_ctrs[task_id] >= task_configs[task_id].t_max:\n break\n\n # reset mac weights\n # copy all weights that have same dimensions from last chosen task (not sure whether this is not redundant)\n if previous_task_id is not None:\n sd = macs[task_id].agent.state_dict()\n for k, v in macs[previous_task_id].agent.state_dict().items():\n if (k in sd) and (v.shape == sd[k].shape):\n setattr(macs[task_id].agent, k, v)\n\n # train\n for t in range(args.n_task_trains[task_id]):\n\n if buffers[task_id].can_sample(task_configs[task_id].batch_size):\n episode_sample = buffers[task_id].sample(task_configs[task_id].batch_size)\n max_ep_t = episode_sample.max_t_filled()\n episode_sample = episode_sample[:, :max_ep_t]\n if episode_sample.device != task_configs[task_id].device:\n episode_sample.to(task_configs[task_id].device)\n\n learners[task_id].train(episode_sample,\n runners[task_id].t_env,\n episode_ctrs[task_id])\n\n # update weights of same dimensions using simple rule (otherwise: formulate as a gradient procedure)\n import operator\n for _task_id, _ in sorted(task_configs.items()):\n mac_state_dict = macs[task_id].agent.state_dict()\n if _task_id != task_id:\n _mac_state_dict = macs[_task_id].agent.state_dict()\n for k, v in _mac_state_dict.items():\n if (k in mac_state_dict) and (v.shape == mac_state_dict[k].shape):\n new_weights = operator.attrgetter(k)(macs[_task_id].agent) + args.reptile_epsilon * (mac_state_dict[k] - v)\n setattr(macs[_task_id].agent, k, new_weights)\n # agent_state_dict[k] += args.reptile_epsilon * (mac_state_dict[k] - macs[_task_id].agent.state_dict()[k])\n\n\n for task_id, task_config in task_configs.items():\n # Execute test runs once in a while\n n_test_runs = max(1, task_configs[task_id].test_nepisode // runners[task_id].batch_size)\n if (runners[task_id].t_env - last_test_Ts[task_id]) / task_configs[task_id].test_interval >= 1.0:\n loggers[task_id].console_logger.info(\"Now testing: {}\".format(task_id))\n loggers[task_id].console_logger.info(\"t_env: {} / {}\".format(runners[task_id].t_env,\n task_configs[task_id].t_max))\n loggers[task_id].console_logger.info(\"Estimated time left: {}. Time passed: {}\".format(\n time_left(last_times[task_id],\n last_test_Ts[task_id],\n runners[task_id].t_env,\n task_configs[task_id].t_max),\n time_str(time.time() - start_time)))\n last_times[task_id] = time.time()\n\n last_test_Ts[task_id] = runners[task_id].t_env\n for _ in range(n_test_runs):\n runners[task_id].run(test_mode=True)\n\n previous_task_id = task_id\n\n for task_id, task_config in task_configs.items():\n if task_config.save_model and \\\n (runners[task_id].t_env - model_save_times[task_id] >= task_config.save_model_interval or\n model_save_times[task_id] == 0):\n model_save_times[task_id] = runners[task_id].t_env\n save_path = os.path.join(task_config.local_results_path,\n \"models\",\n task_config.unique_token,\n str(runners[task_id].t_env))\n #\"results/models/{}\".format(unique_token)\n os.makedirs(save_path, exist_ok=True)\n logger.console_logger.info(\"Saving models to {}\".format(save_path))\n\n # learner should handle saving/loading -- delegate actor save/load to mac,\n # use appropriate filenames to do critics, optimizer states\n learners[task_id].save_models(save_path)", "sub_path": "src/run.py", "file_name": "run.py", "file_ext": "py", "file_size_in_byte": 24362, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "collections.Mapping", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 39, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 41, "usage_type": "call"}, {"api_name": "yaml.YAMLError", "line_number": 42, "usage_type": "attribute"}, {"api_name": "types.SimpleNamespace", "line_number": 51, "usage_type": "call"}, {"api_name": "utils.logging.Logger", "line_number": 55, "usage_type": "call"}, {"api_name": "pprint.pformat", "line_number": 58, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 64, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path", "line_number": 68, "usage_type": "attribute"}, {"api_name": "threading.enumerate", "line_number": 85, "usage_type": "call"}, {"api_name": "os._exit", "line_number": 94, "usage_type": "call"}, {"api_name": "os.EX_OK", "line_number": 94, "usage_type": "attribute"}, {"api_name": "runners.REGISTRY", "line_number": 110, "usage_type": "name"}, {"api_name": "dill.loads", "line_number": 116, "usage_type": "call"}, {"api_name": "dill.loads", "line_number": 117, "usage_type": "call"}, {"api_name": "torch.long", "line_number": 125, "usage_type": "attribute"}, {"api_name": "torch.int", "line_number": 126, "usage_type": "attribute"}, {"api_name": "torch.uint8", "line_number": 128, "usage_type": "attribute"}, {"api_name": "components.transforms.OneHot", "line_number": 134, "usage_type": "call"}, {"api_name": "components.episode_buffer.ReplayBuffer", "line_number": 137, "usage_type": "call"}, {"api_name": "controllers.REGISTRY", "line_number": 142, "usage_type": "name"}, {"api_name": "learners.REGISTRY", "line_number": 148, "usage_type": "name"}, {"api_name": "os.path.isdir", "line_number": 158, "usage_type": "call"}, {"api_name": "os.path", "line_number": 158, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 164, "usage_type": "call"}, {"api_name": "os.path", "line_number": 164, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 166, "usage_type": "call"}, {"api_name": "os.path", "line_number": 166, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 176, "usage_type": "call"}, {"api_name": "os.path", "line_number": 176, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 192, "usage_type": "call"}, {"api_name": "torch.cuda.empty_cache", "line_number": 201, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 201, "usage_type": "attribute"}, {"api_name": "torch.cuda.empty_cache", "line_number": 220, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 220, "usage_type": "attribute"}, {"api_name": "utils.timehelper.time_left", "line_number": 228, "usage_type": "call"}, {"api_name": "utils.timehelper.time_str", "line_number": 228, "usage_type": "call"}, {"api_name": "time.time", "line_number": 228, "usage_type": "call"}, {"api_name": "time.time", "line_number": 229, "usage_type": "call"}, {"api_name": "torch.cuda.empty_cache", "line_number": 234, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 234, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 238, "usage_type": "call"}, {"api_name": "os.path", "line_number": 238, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 240, "usage_type": "call"}, {"api_name": "torch.cuda.empty_cache", "line_number": 248, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 248, "usage_type": "attribute"}, {"api_name": "runners.REGISTRY", "line_number": 263, "usage_type": "name"}, {"api_name": "dill.loads", "line_number": 267, "usage_type": "call"}, {"api_name": "dill.loads", "line_number": 268, "usage_type": "call"}, {"api_name": "torch.long", "line_number": 278, "usage_type": "attribute"}, {"api_name": "torch.int", "line_number": 279, "usage_type": "attribute"}, {"api_name": "torch.uint8", "line_number": 281, "usage_type": "attribute"}, {"api_name": "components.transforms.OneHot", "line_number": 287, "usage_type": "call"}, {"api_name": "components.episode_buffer.ReplayBuffer", "line_number": 290, "usage_type": "call"}, {"api_name": "controllers.REGISTRY", "line_number": 295, "usage_type": "name"}, {"api_name": "learners.REGISTRY", "line_number": 301, "usage_type": "name"}, {"api_name": "torch.cuda.is_available", "line_number": 311, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 311, "usage_type": "attribute"}, {"api_name": "numpy.random.RandomState", "line_number": 342, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 342, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 346, "usage_type": "call"}, {"api_name": "os.path", "line_number": 346, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 346, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 348, "usage_type": "call"}, {"api_name": "yaml.YAMLError", "line_number": 349, "usage_type": "attribute"}, {"api_name": "utils.logging.Logger", "line_number": 383, "usage_type": "call"}, {"api_name": "runners.REGISTRY", "line_number": 391, "usage_type": "name"}, {"api_name": "dill.loads", "line_number": 399, "usage_type": "call"}, {"api_name": "dill.loads", "line_number": 400, "usage_type": "call"}, {"api_name": "dill.loads", "line_number": 405, "usage_type": "call"}, {"api_name": "dill.loads", "line_number": 406, "usage_type": "call"}, {"api_name": "torch.long", "line_number": 413, "usage_type": "attribute"}, {"api_name": "torch.int", "line_number": 414, "usage_type": "attribute"}, {"api_name": "torch.uint8", "line_number": 416, "usage_type": "attribute"}, {"api_name": "components.transforms.OneHot", "line_number": 422, "usage_type": "call"}, {"api_name": "components.episode_buffer.ReplayBuffer", "line_number": 425, "usage_type": "call"}, {"api_name": "controllers.REGISTRY", "line_number": 431, "usage_type": "name"}, {"api_name": "learners.REGISTRY", "line_number": 441, "usage_type": "name"}, {"api_name": "os.path.isdir", "line_number": 462, "usage_type": "call"}, {"api_name": "os.path", "line_number": 462, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 467, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 468, "usage_type": "call"}, {"api_name": "os.path", "line_number": 468, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 470, "usage_type": "call"}, {"api_name": "os.path", "line_number": 470, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 480, "usage_type": "call"}, {"api_name": "os.path", "line_number": 480, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 499, "usage_type": "call"}, {"api_name": "time.time", "line_number": 501, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 513, "usage_type": "call"}, {"api_name": "operator.attrgetter", "line_number": 556, "usage_type": "call"}, {"api_name": "utils.timehelper.time_left", "line_number": 569, "usage_type": "call"}, {"api_name": "utils.timehelper.time_str", "line_number": 573, "usage_type": "call"}, {"api_name": "time.time", "line_number": 573, "usage_type": "call"}, {"api_name": "time.time", "line_number": 574, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 587, "usage_type": "call"}, {"api_name": "os.path", "line_number": 587, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 592, "usage_type": "call"}]}
+{"seq_id": "83395507", "text": "import cv2\n\nfrom PyQt5.QtCore import *\nfrom PyQt5.QtGui import *\nfrom PyQt5.QtWidgets import *\nimport numpy as np\nimport utils\nimport math\nimport skimage.segmentation as seg\nfrom enum import Enum\n\n\nclass BaseViewer(QLabel):\n # Signals\n zoom_changed: pyqtSignal = pyqtSignal([tuple, list])\n #move_drag: pyqtSignal = pyqtSignal([tuple, list])\n\n def __init__(self, parent, window):\n super().__init__()\n self.window = window\n self.zoom = 1.0\n self.pos_00 = [0, 0]\n self.image = None\n self.image_orig = None\n self.res_image = None\n self.image_name = None\n self.start_photo = None\n self.image_slic = None\n\n\n\n def paintEvent(self, event):\n if self.image is not None:\n painter = QPainter(self)\n img_to_draw = self.get_visual_image3()\n if img_to_draw is None:\n self.count_hw_pos(-100, 0, 0)\n img_to_draw = self.get_visual_image3()\n pixmap = QPixmap(utils.cvImage2QImage(self.image))\n painter.drawPixmap(QRect(0, 0, img_to_draw.shape[1], img_to_draw.shape[0]), pixmap)\n\n '''\n def wheelEvent(self, event):\n if self.image_orig is not None:\n if -self.pos_00[0] + event.pos().y() > self.zoom_hw[0] or -self.pos_00[1] + event.pos().x() > self.zoom_hw[\n 1]:\n return\n if event.angleDelta().y() > 0:\n self.count_hw_pos(0.5, event.pos().y(), event.pos().x())\n else:\n self.count_hw_pos(-0.5, event.pos().y(), event.pos().x())\n\n self.update()\n\n\n def normalize_pos00(self):\n if self.pos_00[0] > 0 or self.pos_00[1] > 0:\n self.pos_00 = [0, 0]\n '''\n\n\n def get_visual_image3(self):\n r0 = -self.pos_00[0]\n c0 = -self.pos_00[1]\n\n r1 = min(r0 + self.height(), self.zoom_hw[0])\n c1 = min(c0 + self.width(), self.zoom_hw[1])\n\n if r0 >= r1 - 100 or c0 >= c1 - 100:\n return None\n\n # final_size = (self.zoom_hw[1], self.zoom_hw[0])\n if self.res_image is None or self.res_image.shape[0] != self.zoom_hw[0] or self.res_image.shape[1] != self.zoom_hw[1]:\n part_size = (c1 - c0, r1 - r0)\n img_h, img_w = self.image.shape[:2]\n r0, r1 = round(r0 / self.zoom_hw[0] * img_h), round(r1 / self.zoom_hw[0] * img_h)\n c0, c1 = round(c0 / self.zoom_hw[1] * img_w), round(c1 / self.zoom_hw[1] * img_w)\n self.res_image = cv2.resize(self.image[r0:r1, c0:c1], part_size, interpolation=cv2.INTER_CUBIC)\n res_image = self.res_image\n else:\n res_image = self.res_image[r0:r1, c0:c1]\n #print('draw shape: ', self.image.shape)\n return res_image\n\n\n def count_hw_pos(self, delta_zoom, r, c):\n self.zoom += delta_zoom\n self.zoom += delta_zoom\n\n if self.zoom > 6:\n self.zoom = 6\n\n if self.zoom < 1:\n self.zoom = 1\n\n if self.zoom == 1:\n h_scale = float(self.image.shape[0]) / self.height()\n w_scale = float(self.image.shape[1]) / self.width()\n final_scale = max(h_scale, w_scale)\n h = math.floor(self.image.shape[0] / final_scale)\n w = math.floor(self.image.shape[1] / final_scale)\n self.pos_00 = [0, 0]\n else:\n h = math.floor(self.image.shape[0] * self.zoom)\n w = math.floor(self.image.shape[1] * self.zoom)\n\n img_abs_r = -self.pos_00[0] + r\n img_abs_c = -self.pos_00[1] + c\n\n img_rel_r = img_abs_r / self.zoom_hw[0]\n img_rel_c = img_abs_c / self.zoom_hw[1]\n\n new_img_abs_r = img_rel_r * h\n new_img_abs_c = img_rel_c * w\n\n widget_r0 = math.floor(new_img_abs_r - r)\n widget_c0 = math.floor(new_img_abs_c - c)\n\n self.pos_00[0] = -widget_r0\n self.pos_00[1] = -widget_c0\n self.normalize_pos00()\n\n self.zoom_hw = (h, w)\n self.zoom_changed.emit(self.zoom_hw, self.pos_00)\n\n\n def setPhoto(self, image):\n self.image = image\n self.image_orig = image\n self.background = np.zeros(self.image.shape, dtype=np.uint8)\n self.image_trimap = np.ones(self.image.shape, dtype=np.uint8) * 255\n self.contruct_visualization_image()\n self.count_hw_pos(-5, 0, 0) # init fields\n self.stack = [self.image_trimap.copy()]\n self.undo_key = 0\n self.during_stack_index = 0\n self.colors = Enum(value='', names=[('background', (255, 255, 255)), ('green', (0, 255, 0)), ('blue', (255, 0, 0)),\n ('yellow', (0, 255, 255)), ('black', (1, 1, 1)), ('red', (0, 0, 255))])\n\n def updatePhoto(self, image):\n self.image = image\n self.contruct_visualization_image()\n self.count_hw_pos(-5, 0, 0) # init fields\n\n\n def widget_to_img_pos(self, r, c):\n if self.zoom == 1:\n h_scale = float(self.image.shape[0]) / self.height()\n w_scale = float(self.image.shape[1]) / self.width()\n point_zoom = max(h_scale, w_scale)\n img_abs_r = -self.pos_00[0] + math.floor(r * point_zoom)\n img_abs_c = -self.pos_00[1] + math.floor(c * point_zoom)\n else:\n img_abs_r = -self.pos_00[0] + r\n img_abs_c = -self.pos_00[1] + c\n\n img_rel_r = img_abs_r / self.zoom_hw[0]\n img_rel_c = img_abs_c / self.zoom_hw[1]\n\n img_abs_r = math.floor(img_rel_r * self.image_orig.shape[0])\n img_abs_c = math.floor(img_rel_c * self.image_orig.shape[1])\n\n return img_abs_r, img_abs_c\n\n\n def contruct_visualization_image(self):\n if self.window.image.isChecked():\n trimap_overlay = ((self.image_trimap[:, :, 0] != 0) | (self.image_trimap[:, :, 1] != 0) | (\n self.image_trimap[:, :, 2] != 0)).astype(np.float32) \\\n * self.window.transp / 100.0\n trimap_overlay = np.repeat(np.expand_dims(trimap_overlay, 2), 3, axis=2)\n self.image = self.image * (1.0 - trimap_overlay) + trimap_overlay * self.image_trimap\n self.image = self.image.astype(np.uint8)\n if self.window.boundary.isChecked():\n self.image = cv2.bitwise_or(self.background, self.image)\n else:\n trimap_overlay = ((self.image_trimap[:, :, 0] != 0) | (self.image_trimap[:, :, 1] != 0) | (\n self.image_trimap[:, :, 2] != 0)).astype(np.float32) \\\n * self.window.transp / 100.0\n trimap_overlay = np.repeat(np.expand_dims(trimap_overlay, 2), 3, axis=2)\n self.image = self.background * (1.0 - trimap_overlay) + trimap_overlay * self.image_trimap\n self.image = self.image.astype(np.uint8)\n\n\n def boundary(self):\n self.background = np.zeros(self.image.shape, dtype=np.uint8)\n if self.window.denoise.isChecked():\n image = cv2.medianBlur(self.image_orig, 25)\n else:\n image = self.image_orig\n self.image_slic = seg.slic(image, n_segments=self.window.number_of_parts) * 255 // self.window.number_of_parts\n boundaries = seg.find_boundaries(self.image_slic, mode='outer').astype(np.uint8) * 255\n for i in range(3):\n self.background[:, :, i] = boundaries\n\n\n def save_mask(self, save_path):\n try:\n if self.image_trimap:\n cv2.imwrite(save_path, self.image_trimap)\n except AttributeError:\n pass\n\n\n def undo(self):\n if self.undo_key + 1 < len(self.stack):\n self.during_stack_index = len(self.stack) - 2 - self.undo_key\n self.image_trimap = self.stack[self.during_stack_index]\n self.undo_key += 1\n self.updatePhoto(self.image_orig)\n self.update()\n\n\n def redo(self):\n if self.undo_key >= -1 and self.during_stack_index + 1 < len(self.stack):\n self.image_trimap = self.stack[self.during_stack_index + 1]\n self.during_stack_index += 1\n self.undo_key -= 1\n self.updatePhoto(self.image_orig)\n self.update()\n\n\n def change_color(self):\n current_color = self.colors[self.window.selection_criterion.currentText()].value\n new_color = self.colors[self.window.change_criterion.currentText()].value\n self.image_trimap[np.where((self.image_trimap == current_color).all(axis=2))] = new_color\n trimap_for_stack = self.image_trimap.copy()\n self.stack.append(trimap_for_stack)\n if len(self.stack) > 10:\n self.stack = self.stack[1:]\n self.updatePhoto(self.image_orig)\n self.update()\n\n\n\n", "sub_path": "src_progect/base_viewer.py", "file_name": "base_viewer.py", "file_ext": "py", "file_size_in_byte": 8766, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "utils.cvImage2QImage", "line_number": 39, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 78, "usage_type": "call"}, {"api_name": "cv2.INTER_CUBIC", "line_number": 78, "usage_type": "attribute"}, {"api_name": "math.floor", "line_number": 100, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 101, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 104, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 105, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 116, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 130, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 131, "usage_type": "attribute"}, {"api_name": "enum.Enum", "line_number": 137, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 151, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 152, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 160, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 169, "usage_type": "attribute"}, {"api_name": "numpy.repeat", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 173, "usage_type": "attribute"}, {"api_name": "cv2.bitwise_or", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 178, "usage_type": "attribute"}, {"api_name": "numpy.repeat", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 182, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 186, "usage_type": "attribute"}, {"api_name": "cv2.medianBlur", "line_number": 188, "usage_type": "call"}, {"api_name": "skimage.segmentation.slic", "line_number": 191, "usage_type": "call"}, {"api_name": "skimage.segmentation", "line_number": 191, "usage_type": "name"}, {"api_name": "skimage.segmentation.find_boundaries", "line_number": 192, "usage_type": "call"}, {"api_name": "skimage.segmentation", "line_number": 192, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 192, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 200, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 226, "usage_type": "call"}]}
+{"seq_id": "624529591", "text": "import pytest\nimport json\nfrom pathlib import Path\nfrom libraries.helper.json_helper import get_json_file, get_json_value, write_json_file, update_json_file\n\npath_to_folder = Path(__file__).parent\ndefault_test_json = {\"test\": \"Hello from json\"}\n\n\ndef reset_json_file(json_file_path):\n with open(str(json_file_path), 'w') as outfile:\n json.dump(default_test_json, outfile, indent=2)\n\n\ndef test_gid_204905_get_json_file():\n \"\"\"\n Description:\n Verify this unit test\n\n Prerequisites: NA\n\n Test Data: NA\n\n Steps:\n 1) Run this unit test\n ER: This unit test passes\n Notes: NA\n\n Projects: BI Internal SW Tools\n \"\"\"\n test_json_path = path_to_folder / \"unit_test_data/test_json.json\"\n reset_json_file(test_json_path)\n test_json = get_json_file(test_json_path)\n assert test_json == default_test_json\n\n\ndef test_gid_204906_get_json_file_negative():\n \"\"\"\n Description:\n Verify this unit test\n\n Prerequisites: NA\n\n Test Data: NA\n\n Steps:\n 1) Run this unit test\n ER: This unit test passes\n Notes: NA\n\n Projects: BI Internal SW Tools\n \"\"\"\n with pytest.raises(Exception):\n get_json_file(\"/nonexistent_path\")\n\n\ndef test_gid_204907_get_json_value():\n \"\"\"\n Description:\n Verify this unit test\n\n Prerequisites: NA\n\n Test Data: NA\n\n Steps:\n 1) Run this unit test\n ER: This unit test passes\n Notes: NA\n\n Projects: BI Internal SW Tools\n \"\"\"\n test_json_path = path_to_folder / \"unit_test_data/test_json.json\"\n reset_json_file(test_json_path)\n json_object = get_json_file(test_json_path)\n get_json_value('test', json_object)\n\n\n# We don't error handle if the key isn't found, just pass back None from jmespath\ndef test_gid_204908_get_json_value_negative():\n \"\"\"\n Description:\n Verify this unit test\n\n Prerequisites: NA\n\n Test Data: NA\n\n Steps:\n 1) Run this unit test\n ER: This unit test passes\n Notes: NA\n\n Projects: BI Internal SW Tools\n \"\"\"\n test_json_path = path_to_folder / \"unit_test_data/test_json.json\"\n json_object = get_json_file(test_json_path)\n assert get_json_value('nonexistent_key', json_object) is None\n\n\n# Reset the test json file and check the contents, write to it\n# then check the contents, then reset the json file again\ndef test_gid_204909_write_json_file():\n \"\"\"\n Description:\n Verify this unit test\n\n Prerequisites: NA\n\n Test Data: NA\n\n Steps:\n 1) Run this unit test\n ER: This unit test passes\n Notes: NA\n\n Projects: BI Internal SW Tools\n \"\"\"\n test_json_path = path_to_folder / \"unit_test_data/test_json.json\"\n reset_json_file(test_json_path)\n json_object = get_json_file(test_json_path)\n assert json_object == default_test_json\n\n json_object_write = {\"test\": \"testing write_json_file()\"}\n write_json_file(test_json_path, json_object_write)\n json_object = get_json_file(test_json_path)\n assert json_object == json_object_write\n\n reset_json_file(test_json_path)\n\n\ndef test_gid_204910_write_json_file_negative():\n \"\"\"\n Description:\n Verify this unit test\n\n Prerequisites: NA\n\n Test Data: NA\n\n Steps:\n 1) Run this unit test\n ER: This unit test passes\n Notes: NA\n\n Projects: BI Internal SW Tools\n \"\"\"\n test_json_path = path_to_folder / \"nonexistent_folder/nonexistent_json.json\"\n\n json_object_write = {\"test\": \"testing write_json_file()\"}\n with pytest.raises(Exception):\n write_json_file(test_json_path, json_object_write)\n\n\ndef test_gid_204911_update_json_file():\n \"\"\"\n Description:\n Verify this unit test\n\n Prerequisites: NA\n\n Test Data: NA\n\n Steps:\n 1) Run this unit test\n ER: This unit test passes\n Notes: NA\n\n Projects: BI Internal SW Tools\n \"\"\"\n test_json_path = path_to_folder / \"unit_test_data/test_json.json\"\n reset_json_file(test_json_path)\n json_object = get_json_file(test_json_path)\n assert json_object == default_test_json\n\n update_json_file(test_json_path, \"test\", \"testing update_json_file()\")\n json_object = get_json_file(test_json_path)\n expected_json = default_test_json\n expected_json[\"test\"] = \"testing update_json_file()\"\n assert json_object == expected_json\n\n reset_json_file(test_json_path)\n\n\n# Error log for this one could be better, currently erroring out at the\n# logger level instead of the at the Exception raise after the log\ndef test_gid_204912_update_json_file_negatives():\n \"\"\"\n Description:\n Verify this unit test\n\n Prerequisites: NA\n\n Test Data: NA\n\n Steps:\n 1) Run this unit test\n ER: This unit test passes\n Notes: NA\n\n Projects: BI Internal SW Tools\n \"\"\"\n test_json_path = path_to_folder / \"unit_test_data/test_json.json\"\n reset_json_file(test_json_path)\n json_object = get_json_file(test_json_path)\n assert json_object == default_test_json\n\n with pytest.raises(Exception):\n update_json_file(test_json_path, \"nonexistent_key\", \"testing update_json_file()\")\n", "sub_path": "libraries/xFramework_unit_tests/test_json_helper.py", "file_name": "test_json_helper.py", "file_ext": "py", "file_size_in_byte": 5190, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pathlib.Path", "line_number": 6, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 12, "usage_type": "call"}, {"api_name": "libraries.helper.json_helper.get_json_file", "line_number": 33, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 53, "usage_type": "call"}, {"api_name": "libraries.helper.json_helper.get_json_file", "line_number": 54, "usage_type": "call"}, {"api_name": "libraries.helper.json_helper.get_json_file", "line_number": 75, "usage_type": "call"}, {"api_name": "libraries.helper.json_helper.get_json_value", "line_number": 76, "usage_type": "call"}, {"api_name": "libraries.helper.json_helper.get_json_file", "line_number": 97, "usage_type": "call"}, {"api_name": "libraries.helper.json_helper.get_json_value", "line_number": 98, "usage_type": "call"}, {"api_name": "libraries.helper.json_helper.get_json_file", "line_number": 121, "usage_type": "call"}, {"api_name": "libraries.helper.json_helper.write_json_file", "line_number": 125, "usage_type": "call"}, {"api_name": "libraries.helper.json_helper.get_json_file", "line_number": 126, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 151, "usage_type": "call"}, {"api_name": "libraries.helper.json_helper.write_json_file", "line_number": 152, "usage_type": "call"}, {"api_name": "libraries.helper.json_helper.get_json_file", "line_number": 173, "usage_type": "call"}, {"api_name": "libraries.helper.json_helper.update_json_file", "line_number": 176, "usage_type": "call"}, {"api_name": "libraries.helper.json_helper.get_json_file", "line_number": 177, "usage_type": "call"}, {"api_name": "libraries.helper.json_helper.get_json_file", "line_number": 205, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 208, "usage_type": "call"}, {"api_name": "libraries.helper.json_helper.update_json_file", "line_number": 209, "usage_type": "call"}]}
+{"seq_id": "605837828", "text": "from dataclasses import dataclass, field\nfrom typing import Dict, Optional, Tuple\n\nfrom electionguard.ballot import SubmittedBallot\n\nfrom .auxiliary import AuxiliaryDecrypt\nfrom .decryption import (\n compute_compensated_decryption_share_for_ballot,\n compute_decryption_share,\n compute_compensated_decryption_share,\n compute_decryption_share_for_ballots,\n reconstruct_decryption_share,\n reconstruct_decryption_share_for_ballot,\n)\nfrom .decryption_share import DecryptionShare, CompensatedDecryptionShare\nfrom .decrypt_with_shares import decrypt_ballots, decrypt_tally\nfrom .election import CiphertextElectionContext\nfrom .election_polynomial import compute_lagrange_coefficient\nfrom .group import ElementModP, ElementModQ\nfrom .guardian import Guardian\nfrom .key_ceremony import ElectionPublicKey\nfrom .rsa import rsa_decrypt\nfrom .tally import (\n CiphertextTally,\n PlaintextTally,\n)\nfrom .logs import log_info, log_warning\nfrom .types import BALLOT_ID, GUARDIAN_ID\n\nAVAILABLE_GUARDIAN_ID = GUARDIAN_ID\nMISSING_GUARDIAN_ID = GUARDIAN_ID\n\nGUARDIAN_PUBLIC_KEY = ElementModP\nSHARE_LOOKUP = Dict[AVAILABLE_GUARDIAN_ID, DecryptionShare]\nCOMPENSATED_SHARE_LOOKUP = Dict[AVAILABLE_GUARDIAN_ID, CompensatedDecryptionShare]\n\n# pylint: disable=too-many-instance-attributes\n@dataclass\nclass DecryptionMediator:\n \"\"\"\n The Decryption Mediator composes partial decryptions from each Guardian\n to form a decrypted representation of an election tally\n \"\"\"\n\n _encryption: CiphertextElectionContext\n\n # Tally to Decrypt\n _ciphertext_tally: CiphertextTally\n _ciphertext_ballots: Dict[BALLOT_ID, SubmittedBallot]\n\n # Tally\n _tally_shares: Dict[AVAILABLE_GUARDIAN_ID, DecryptionShare] = field(\n default_factory=lambda: {}\n )\n\n # Ballot\n _ballot_shares: Dict[\n AVAILABLE_GUARDIAN_ID, Dict[BALLOT_ID, DecryptionShare]\n ] = field(default_factory=lambda: {})\n\n # Guardians\n _available_guardians: Dict[AVAILABLE_GUARDIAN_ID, Guardian] = field(\n default_factory=lambda: {}\n )\n _missing_guardians: Dict[MISSING_GUARDIAN_ID, ElectionPublicKey] = field(\n default_factory=lambda: {}\n )\n _lagrange_coefficients: Dict[\n MISSING_GUARDIAN_ID, Dict[AVAILABLE_GUARDIAN_ID, ElementModQ]\n ] = field(default_factory=lambda: {})\n \"\"\"\n A collection of lagrange coefficients `w_{i,j}` computed by available guardians for each missing guardian\n \"\"\"\n\n def announce(\n self, guardian: Guardian\n ) -> Optional[Tuple[DecryptionShare, Dict[BALLOT_ID, DecryptionShare]]]:\n \"\"\"\n Announce that a Guardian is present and participating in the decryption.\n A Decryption Share will be generated for the Guardian\n\n :param guardian: The guardian who will participate in the decryption.\n :return: decryption shares for tally and ballot for this `Guardian` or `None` if there is an error.\n \"\"\"\n\n # Only allow a guardian to announce once\n if guardian.object_id in self._available_guardians:\n log_info(f\"guardian {guardian.object_id} already announced\")\n return (\n self._tally_shares[guardian.object_id],\n self._ballot_shares[guardian.object_id],\n )\n\n # Compute the tally and ballot decryption shares\n tally_share = compute_decryption_share(\n guardian, self._ciphertext_tally, self._encryption\n )\n if tally_share is None:\n log_warning(\n f\"announce could not generate tally decryption share for {guardian.object_id}\"\n )\n return None\n self._tally_shares[guardian.object_id] = tally_share\n\n # Compute the ballot decryption shares\n ballot_shares = compute_decryption_share_for_ballots(\n guardian, list(self._ciphertext_ballots.values()), self._encryption\n )\n if ballot_shares is None:\n log_warning(\n f\"announce could not generate ballot decryption share for {guardian.object_id}\"\n )\n return None\n self._ballot_shares[guardian.object_id] = ballot_shares\n\n # Mark guardian in attendance and check their keys\n self._mark_available(guardian)\n if not self._validate_missing_guardian_keys(guardian):\n return None\n\n return (tally_share, ballot_shares)\n\n # pylint: disable=too-many-return-statements\n def get_plaintext_tally(\n self, decrypt: AuxiliaryDecrypt = rsa_decrypt\n ) -> Optional[PlaintextTally]:\n \"\"\"\n Get the plaintext tally for the election by composing each Guardian's\n decrypted representation of each selection into a decrypted representation\n\n :return: a `PlaintextTally` or `None`\n \"\"\"\n\n # Make sure a Quorum of Guardians have announced\n if len(self._available_guardians) < self._encryption.quorum:\n log_warning(\n \"cannot get plaintext tally with less than quorum available guardians\"\n )\n return None\n\n # If all Guardians are present decrypt the tally\n if len(self._available_guardians) == self._encryption.number_of_guardians:\n return decrypt_tally(\n self._ciphertext_tally,\n self._tally_shares,\n self._encryption.crypto_extended_base_hash,\n )\n\n # If guardians are missing, compensate then decrypt\n self._compute_missing_shares_for_tally(decrypt)\n\n if len(self._tally_shares) != self._encryption.number_of_guardians:\n log_warning(\"get plaintext tally failed with share length mismatch\")\n return None\n\n return decrypt_tally(\n self._ciphertext_tally,\n self._tally_shares,\n self._encryption.crypto_extended_base_hash,\n )\n\n def _mark_available(self, guardian: Guardian) -> None:\n \"\"\"\n This guardian removes itself from the\n missing list since it generated a valid share\n \"\"\"\n self._available_guardians[guardian.object_id] = guardian\n if guardian.object_id in self._missing_guardians.keys():\n self._missing_guardians.pop(guardian.object_id)\n\n def _validate_missing_guardian_keys(self, guardian: Guardian) -> bool:\n \"\"\"\n Check the guardian's collections of keys and ensure the public keys\n match for the missing guardians\n \"\"\"\n\n # Check this guardian's collection of public keys\n # for other guardians that have not announced\n missing_guardians: Dict[MISSING_GUARDIAN_ID, ElectionPublicKey] = {\n guardian_id: public_key\n for guardian_id, public_key in guardian.guardian_election_public_keys()\n if guardian_id not in self._available_guardians\n }\n\n # Check that the public keys match for any missing guardians already reported\n # note this check naively assumes that the first guardian to annouce is telling the truth\n # but for this implementation it is simply a sanity check on the input data.\n # a consuming application should implement better validation of the guardian state\n # before announcing a guardian is available for decryption.\n for guardian_id, public_key in missing_guardians.items():\n if guardian_id in self._missing_guardians.keys():\n if self._missing_guardians[guardian_id] != public_key:\n log_warning(\n (\n f\"announce guardian: {guardian.object_id} \"\n f\"expected public key mismatch for missing {guardian_id}\"\n )\n )\n return False\n else:\n self._missing_guardians[guardian_id] = missing_guardians[guardian_id]\n return True\n\n def _compute_missing_shares_for_tally(\n self, decrypt: AuxiliaryDecrypt = rsa_decrypt\n ) -> None:\n # If missing guardians compensate for the missing guardians\n missing_tally_shares: Dict[MISSING_GUARDIAN_ID, DecryptionShare] = {}\n for missing_guardian_id, public_key in self._missing_guardians.items():\n if missing_guardian_id in self._tally_shares:\n continue\n self._compute_lagrange_coefficients(missing_guardian_id)\n compensated_shares = self._get_compensated_shares_for_tally(\n missing_guardian_id, decrypt\n )\n if compensated_shares is None:\n log_warning(\n f\"get plaintext tally failed compensating for {missing_guardian_id}\"\n )\n return\n\n missing_decryption_share = reconstruct_decryption_share(\n missing_guardian_id,\n public_key,\n self._ciphertext_tally,\n compensated_shares,\n self._lagrange_coefficients[missing_guardian_id],\n )\n missing_tally_shares[missing_guardian_id] = missing_decryption_share\n\n if missing_tally_shares is None:\n log_warning(\n \"get plaintext tally failed with computing missing decryption shares\"\n )\n return\n\n # Combine all tally shares\n self._tally_shares.update(missing_tally_shares)\n\n def _get_compensated_shares_for_tally(\n self, missing_guardian_id: str, decrypt: AuxiliaryDecrypt = rsa_decrypt\n ) -> Optional[Dict[AVAILABLE_GUARDIAN_ID, CompensatedDecryptionShare]]:\n \"\"\"\n Compensate for a missing guardian by reconstructing the share using the available guardians.\n\n :param missing_guardian_id: the guardian that failed to `announce`.\n :return: a collection of `CompensatedDecryptionShare` generated from all available guardians\n or `None if there is an error\n \"\"\"\n\n compensated_decryptions: Dict[\n AVAILABLE_GUARDIAN_ID, CompensatedDecryptionShare\n ] = {}\n # Loop through each of the available guardians\n # and calculate a partial for the missing one\n for (\n available_gaurdian_id,\n available_guardian,\n ) in self._available_guardians.items():\n # Compute the tally decryption shares\n tally_share = compute_compensated_decryption_share(\n available_guardian,\n missing_guardian_id,\n self._ciphertext_tally,\n self._encryption,\n decrypt,\n )\n if tally_share is None:\n log_warning(f\"compensation failed for missing: {missing_guardian_id}\")\n break\n compensated_decryptions[available_gaurdian_id] = tally_share\n\n # Verify generated the correct number of partials\n if len(compensated_decryptions) != len(self._available_guardians):\n log_warning(\n f\"compensate mismatch partial decryptions for missing guardian {missing_guardian_id}\"\n )\n return None\n\n return compensated_decryptions\n\n def get_plaintext_ballots(\n self, decrypt: AuxiliaryDecrypt = rsa_decrypt\n ) -> Optional[Dict[BALLOT_ID, PlaintextTally]]:\n \"\"\"\n Get the plaintext spoiled ballots for the election by composing each Guardian's\n decrypted representation of each selection into a decrypted representation\n\n :return: a Plaintext Spoiled Ballots or `None`\n \"\"\"\n\n # Make sure a Quorum of Guardians have announced\n if len(self._available_guardians) < self._encryption.quorum:\n log_warning(\"cannot decrypt with less than quorum available guardians\")\n return None\n\n # If all Guardians are present decrypt the ballots\n if len(self._available_guardians) == self._encryption.number_of_guardians:\n return decrypt_ballots(\n self._ciphertext_ballots,\n self._ballot_shares,\n self._encryption.crypto_extended_base_hash,\n )\n\n # If guardians are missing, compensate then decrypt\n for ballot_id in self._ciphertext_ballots.keys():\n self._compute_missing_shares_for_ballot(ballot_id, decrypt)\n\n if (\n self._count_ballot_shares(ballot_id)\n != self._encryption.number_of_guardians\n ):\n log_warning(\"get plaintext ballot failed with share length mismatch\")\n return None\n\n return decrypt_ballots(\n self._ciphertext_ballots,\n self._ballot_shares,\n self._encryption.crypto_extended_base_hash,\n )\n\n def _count_ballot_shares(self, ballot_id: str) -> int:\n count = 0\n for ballot_shares in self._ballot_shares.values():\n if ballot_shares.get(ballot_id):\n count += 1\n return count\n\n def _compute_missing_shares_for_ballot(\n self, ballot_id: str, decrypt: AuxiliaryDecrypt = rsa_decrypt\n ) -> None:\n \"\"\"\n Compute the missing decryption shares for all the guardians who are missing\n and add to the shares of the available guardians\n \"\"\"\n # If missing guardians compensate for the missing guardians\n for missing_guardian_id, public_key in self._missing_guardians.items():\n self._compute_lagrange_coefficients(missing_guardian_id)\n compensated_shares = self._get_compensated_shares_for_ballot(\n ballot_id, missing_guardian_id, decrypt\n )\n if compensated_shares is None:\n log_warning(\n f\"get plaintext ballot failed compensating for {missing_guardian_id}\"\n )\n return\n\n missing_decryption_share = reconstruct_decryption_share_for_ballot(\n missing_guardian_id,\n public_key,\n self._ciphertext_ballots[ballot_id],\n compensated_shares,\n self._lagrange_coefficients[missing_guardian_id],\n )\n\n self._ballot_shares[missing_guardian_id][\n ballot_id\n ] = missing_decryption_share\n\n def _get_compensated_shares_for_ballot(\n self,\n ballot_id: str,\n missing_guardian_id: str,\n decrypt: AuxiliaryDecrypt = rsa_decrypt,\n ) -> Optional[Dict[AVAILABLE_GUARDIAN_ID, CompensatedDecryptionShare]]:\n \"\"\"\n Compensate for a missing guardian by reconstructing the share using the available guardians.\n\n :param ballot_id: The id of the ballot to get the share of\n :param missing_guardian_id: the guardian that failed to `announce`.\n :return: a collection of `CompensatedDecryptionShare` generated from all available guardians\n or `None if there is an error\n \"\"\"\n\n compensated_decryptions: Dict[\n AVAILABLE_GUARDIAN_ID, CompensatedDecryptionShare\n ] = {}\n # Loop through each of the available guardians\n # and calculate a partial for the missing one\n for (\n available_gaurdian_id,\n available_guardian,\n ) in self._available_guardians.items():\n # Compute the tally decryption shares\n ballot_share = compute_compensated_decryption_share_for_ballot(\n available_guardian,\n missing_guardian_id,\n self._ciphertext_ballots[ballot_id],\n self._encryption,\n decrypt,\n )\n if ballot_share is None:\n log_warning(f\"compensation failed for missing: {missing_guardian_id}\")\n break\n compensated_decryptions[available_gaurdian_id] = ballot_share\n\n # Verify generated the correct number of partials\n if len(compensated_decryptions) != len(self._available_guardians):\n log_warning(\n f\"compensate mismatch partial decryptions for missing guardian {missing_guardian_id}\"\n )\n return None\n\n return compensated_decryptions\n\n def _compute_lagrange_coefficients(self, missing_guardian_id: str) -> None:\n \"\"\"Compute lagrange coefficients for each of the available guardians\"\"\"\n if self._lagrange_coefficients.get(missing_guardian_id):\n return\n\n lagrange_coefficients: Dict[AVAILABLE_GUARDIAN_ID, ElementModQ] = {}\n for available_guardian in self._available_guardians.values():\n lagrange_coefficients[\n available_guardian.object_id\n ] = compute_lagrange_coefficient(\n available_guardian.sequence_order,\n *[\n guardian.sequence_order\n for guardian in self._available_guardians.values()\n if guardian.object_id != available_guardian.object_id\n ],\n )\n self._lagrange_coefficients[missing_guardian_id] = lagrange_coefficients\n", "sub_path": "src/electionguard/decryption_mediator.py", "file_name": "decryption_mediator.py", "file_ext": "py", "file_size_in_byte": 16978, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "types.GUARDIAN_ID", "line_number": 30, "usage_type": "name"}, {"api_name": "types.GUARDIAN_ID", "line_number": 31, "usage_type": "name"}, {"api_name": "group.ElementModP", "line_number": 33, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 34, "usage_type": "name"}, {"api_name": "decryption_share.DecryptionShare", "line_number": 34, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 35, "usage_type": "name"}, {"api_name": "decryption_share.CompensatedDecryptionShare", "line_number": 35, "usage_type": "name"}, {"api_name": "election.CiphertextElectionContext", "line_number": 45, "usage_type": "name"}, {"api_name": "tally.CiphertextTally", "line_number": 48, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 49, "usage_type": "name"}, {"api_name": "types.BALLOT_ID", "line_number": 49, "usage_type": "name"}, {"api_name": "electionguard.ballot.SubmittedBallot", "line_number": 49, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 52, "usage_type": "name"}, {"api_name": "decryption_share.DecryptionShare", "line_number": 52, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 52, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 57, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 58, "usage_type": "name"}, {"api_name": "types.BALLOT_ID", "line_number": 58, "usage_type": "name"}, {"api_name": "decryption_share.DecryptionShare", "line_number": 58, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 59, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 62, "usage_type": "name"}, {"api_name": "guardian.Guardian", "line_number": 62, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 62, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 65, "usage_type": "name"}, {"api_name": "key_ceremony.ElectionPublicKey", "line_number": 65, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 65, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 68, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 69, "usage_type": "name"}, {"api_name": "group.ElementModQ", "line_number": 69, "usage_type": "name"}, {"api_name": "dataclasses.field", "line_number": 70, "usage_type": "call"}, {"api_name": "guardian.Guardian", "line_number": 76, "usage_type": "name"}, {"api_name": "guardian.object_id", "line_number": 87, "usage_type": "attribute"}, {"api_name": "logs.log_info", "line_number": 88, "usage_type": "call"}, {"api_name": "guardian.object_id", "line_number": 88, "usage_type": "attribute"}, {"api_name": "guardian.object_id", "line_number": 90, "usage_type": "attribute"}, {"api_name": "guardian.object_id", "line_number": 91, "usage_type": "attribute"}, {"api_name": "decryption.compute_decryption_share", "line_number": 95, "usage_type": "call"}, {"api_name": "logs.log_warning", "line_number": 99, "usage_type": "call"}, {"api_name": "guardian.object_id", "line_number": 100, "usage_type": "attribute"}, {"api_name": "guardian.object_id", "line_number": 103, "usage_type": "attribute"}, {"api_name": "decryption.compute_decryption_share_for_ballots", "line_number": 106, "usage_type": "call"}, {"api_name": "logs.log_warning", "line_number": 110, "usage_type": "call"}, {"api_name": "guardian.object_id", "line_number": 111, "usage_type": "attribute"}, {"api_name": "guardian.object_id", "line_number": 114, "usage_type": "attribute"}, {"api_name": "typing.Optional", "line_number": 77, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 77, "usage_type": "name"}, {"api_name": "decryption_share.DecryptionShare", "line_number": 77, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 77, "usage_type": "name"}, {"api_name": "types.BALLOT_ID", "line_number": 77, "usage_type": "name"}, {"api_name": "auxiliary.AuxiliaryDecrypt", "line_number": 125, "usage_type": "name"}, {"api_name": "rsa.rsa_decrypt", "line_number": 125, "usage_type": "name"}, {"api_name": "logs.log_warning", "line_number": 136, "usage_type": "call"}, {"api_name": "decrypt_with_shares.decrypt_tally", "line_number": 143, "usage_type": "call"}, {"api_name": "logs.log_warning", "line_number": 153, "usage_type": "call"}, {"api_name": "decrypt_with_shares.decrypt_tally", "line_number": 156, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 126, "usage_type": "name"}, {"api_name": "tally.PlaintextTally", "line_number": 126, "usage_type": "name"}, {"api_name": "guardian.Guardian", "line_number": 162, "usage_type": "name"}, {"api_name": "guardian.object_id", "line_number": 167, "usage_type": "attribute"}, {"api_name": "guardian.object_id", "line_number": 168, "usage_type": "attribute"}, {"api_name": "guardian.object_id", "line_number": 169, "usage_type": "attribute"}, {"api_name": "guardian.Guardian", "line_number": 171, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 179, "usage_type": "name"}, {"api_name": "key_ceremony.ElectionPublicKey", "line_number": 179, "usage_type": "name"}, {"api_name": "guardian.guardian_election_public_keys", "line_number": 181, "usage_type": "call"}, {"api_name": "logs.log_warning", "line_number": 193, "usage_type": "call"}, {"api_name": "guardian.object_id", "line_number": 195, "usage_type": "attribute"}, {"api_name": "auxiliary.AuxiliaryDecrypt", "line_number": 205, "usage_type": "name"}, {"api_name": "rsa.rsa_decrypt", "line_number": 205, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 208, "usage_type": "name"}, {"api_name": "decryption_share.DecryptionShare", "line_number": 208, "usage_type": "name"}, {"api_name": "logs.log_warning", "line_number": 217, "usage_type": "call"}, {"api_name": "decryption.reconstruct_decryption_share", "line_number": 222, "usage_type": "call"}, {"api_name": "logs.log_warning", "line_number": 232, "usage_type": "call"}, {"api_name": "auxiliary.AuxiliaryDecrypt", "line_number": 241, "usage_type": "name"}, {"api_name": "rsa.rsa_decrypt", "line_number": 241, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 251, "usage_type": "name"}, {"api_name": "decryption_share.CompensatedDecryptionShare", "line_number": 252, "usage_type": "name"}, {"api_name": "decryption.compute_compensated_decryption_share", "line_number": 261, "usage_type": "call"}, {"api_name": "logs.log_warning", "line_number": 269, "usage_type": "call"}, {"api_name": "logs.log_warning", "line_number": 275, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 242, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 242, "usage_type": "name"}, {"api_name": "decryption_share.CompensatedDecryptionShare", "line_number": 242, "usage_type": "name"}, {"api_name": "auxiliary.AuxiliaryDecrypt", "line_number": 283, "usage_type": "name"}, {"api_name": "rsa.rsa_decrypt", "line_number": 283, "usage_type": "name"}, {"api_name": "logs.log_warning", "line_number": 294, "usage_type": "call"}, {"api_name": "decrypt_with_shares.decrypt_ballots", "line_number": 299, "usage_type": "call"}, {"api_name": "logs.log_warning", "line_number": 313, "usage_type": "call"}, {"api_name": "decrypt_with_shares.decrypt_ballots", "line_number": 316, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 284, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 284, "usage_type": "name"}, {"api_name": "types.BALLOT_ID", "line_number": 284, "usage_type": "name"}, {"api_name": "tally.PlaintextTally", "line_number": 284, "usage_type": "name"}, {"api_name": "auxiliary.AuxiliaryDecrypt", "line_number": 330, "usage_type": "name"}, {"api_name": "rsa.rsa_decrypt", "line_number": 330, "usage_type": "name"}, {"api_name": "logs.log_warning", "line_number": 343, "usage_type": "call"}, {"api_name": "decryption.reconstruct_decryption_share_for_ballot", "line_number": 348, "usage_type": "call"}, {"api_name": "auxiliary.AuxiliaryDecrypt", "line_number": 364, "usage_type": "name"}, {"api_name": "rsa.rsa_decrypt", "line_number": 364, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 375, "usage_type": "name"}, {"api_name": "decryption_share.CompensatedDecryptionShare", "line_number": 376, "usage_type": "name"}, {"api_name": "decryption.compute_compensated_decryption_share_for_ballot", "line_number": 385, "usage_type": "call"}, {"api_name": "logs.log_warning", "line_number": 393, "usage_type": "call"}, {"api_name": "logs.log_warning", "line_number": 399, "usage_type": "call"}, {"api_name": "typing.Optional", "line_number": 365, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 365, "usage_type": "name"}, {"api_name": "decryption_share.CompensatedDecryptionShare", "line_number": 365, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 411, "usage_type": "name"}, {"api_name": "group.ElementModQ", "line_number": 411, "usage_type": "name"}, {"api_name": "election_polynomial.compute_lagrange_coefficient", "line_number": 415, "usage_type": "call"}, {"api_name": "guardian.sequence_order", "line_number": 418, "usage_type": "attribute"}, {"api_name": "guardian.object_id", "line_number": 420, "usage_type": "attribute"}, {"api_name": "dataclasses.dataclass", "line_number": 38, "usage_type": "name"}]}
+{"seq_id": "502586360", "text": "#----------------------------------------------------------------------------\n# IMPORTING MODULES\n#----------------------------------------------------------------------------\n\nfrom __future__ import print_function, division\n\nimport torch\nimport torch.nn as nn\nimport torch.optim as optim\nfrom torch.optim import lr_scheduler\nimport torch.nn.functional as F\nfrom torch.utils.data import Dataset, DataLoader\nfrom torchvision import transforms, utils\nfrom torch.autograd import Variable\n\nfrom PIL import Image\n\nimport numpy as np\nimport torchvision\nfrom torchvision import datasets, models, transforms\nimport time\nimport os\nimport copy\nimport pickle\n\n#---------------------------------------------------------------------------\n# IMPORTANT PARAMETERS\n#---------------------------------------------------------------------------\n\nroot_dir = \"../Dataset/\"\ndevice = torch.device(\"cuda:0\" if torch.cuda.is_available() else \"cpu\")\nepochs = 16\nbatch_size = 32\nmaxFaces = 15\nnumClasses = 3\n\n#---------------------------------------------------------------------------\n# DATASET AND LOADERS\n#---------------------------------------------------------------------------\n\nneg_train = sorted(os.listdir('../Dataset/emotiw/train/'+'Negative/'))\nneu_train = sorted(os.listdir('../Dataset/emotiw/train/'+'Neutral/'))\npos_train = sorted(os.listdir('../Dataset/emotiw/train/'+'Positive/'))\n\ntrain_filelist = neg_train + neu_train + pos_train\n\nval_filelist = []\ntest_filelist = []\n\nwith open('../Dataset/val_list', 'rb') as fp:\n val_filelist = pickle.load(fp)\n\nwith open('../Dataset/test_list', 'rb') as fp:\n test_filelist = pickle.load(fp)\n\nfor i in train_filelist:\n if i[0] != 'p' and i[0] != 'n':\n train_filelist.remove(i)\n \nfor i in val_filelist:\n if i[0] != 'p' and i[0] != 'n':\n val_filelist.remove(i)\n\nfor i in range(len(train_filelist)):\n train_filelist[i] = 'train/' + train_filelist[i]\n\nfor i in range(len(val_filelist)):\n val_filelist[i] = 'val/' + val_filelist[i]\n\nfor i in range(len(test_filelist)):\n test_filelist[i] = 'val/' + test_filelist[i]\n\nfull_train_filelist = train_filelist + val_filelist\nfull_val_filelist = test_filelist\n\ndataset_sizes = [len(full_train_filelist), len(full_val_filelist)]\nprint(dataset_sizes)\n\ntrain_global_data_transform = transforms.Compose([\n transforms.RandomResizedCrop(224),\n transforms.RandomHorizontalFlip(),\n transforms.ToTensor(),\n transforms.Normalize([0.485, 0.456, 0.406], [0.229, 0.224, 0.225])\n])\n\nval_global_data_transform = transforms.Compose([\n transforms.Resize(256),\n transforms.CenterCrop(224),\n transforms.ToTensor(),\n transforms.Normalize([0.485, 0.456, 0.406], [0.229, 0.224, 0.225])\n ])\n\nclass EmotiWDataset(Dataset):\n \n def __init__(self, filelist, root_dir, transformGlobal=transforms.ToTensor(), transformFaces = transforms.ToTensor()):\n \"\"\"\n Args:\n filelist: List of names of image/feature files.\n root_dir: Dataset directory\n transform (callable, optional): Optional transformer to be applied\n on an image sample.\n \"\"\"\n \n self.filelist = filelist\n self.root_dir = root_dir\n self.transformGlobal = transformGlobal\n self.transformFaces = transformFaces\n \n def __len__(self):\n return (len(self.filelist)) \n \n def __getitem__(self, idx):\n \n folder_name, filename = self.filelist[idx].split('/')[0], self.filelist[idx].split('/')[1].split('.')[0]\n # filename = self.filelist[idx].split('.')[0]\n labeldict = {'neg':'Negative',\n 'neu':'Neutral',\n 'pos':'Positive',\n 'Negative': 0,\n 'Neutral': 1,\n 'Positive':2}\n\n labelname = labeldict[filename.split('_')[0]]\n\n #IMAGE \n\n image = Image.open(self.root_dir+'emotiw/'+folder_name+'/'+labelname+'/'+filename+'.jpg')\n if self.transformGlobal:\n image = self.transformGlobal(image)\n if image.shape[0] == 1:\n image_1 = np.zeros((3, 224, 224), dtype = float)\n image_1[0] = image\n image_1[1] = image\n image_1[2] = image\n image = image_1\n image = torch.FloatTensor(image.tolist()) \n \n #SAMPLE\n sample = {'image': image, 'label':labeldict[labelname]}\n return sample\n\n\ntrain_dataset = EmotiWDataset(full_train_filelist, root_dir, transformGlobal=train_global_data_transform)\n\ntrain_dataloader = DataLoader(train_dataset, shuffle=True, batch_size=batch_size, num_workers=0)\n\nval_dataset = EmotiWDataset(full_val_filelist, root_dir, transformGlobal = val_global_data_transform)\n\nval_dataloader = DataLoader(val_dataset, shuffle =True, batch_size = batch_size, num_workers = 0)\n\nprint('Dataset Loaded')\n\n#---------------------------------------------------------------------------\n# MODEL DEFINITION\n#---------------------------------------------------------------------------\n\n\nmodel_ft = models.densenet161(pretrained=True)\nnum_ftrs = model_ft.classifier.in_features\nmodel_ft.classifier = nn.Linear(num_ftrs, 3)\n\nmodel_ft = model_ft.to(device)\n\n#---------------------------------------------------------------------------\n# TRAINING\n#---------------------------------------------------------------------------\n\ndef train_model(model, criterion, optimizer, scheduler, num_epochs = 25):\n \n since = time.time()\n \n best_model_wts = copy.deepcopy(model.state_dict())\n best_acc = 0.0\n \n for epoch in range(num_epochs):\n print(\"Epoch {}/{}\".format(epoch, num_epochs - 1))\n print('-' * 10)\n \n for phase in range(2):\n if phase == 0:\n dataloaders = train_dataloader\n scheduler.step()\n model.train()\n else:\n dataloaders = val_dataloader\n model.eval()\n \n running_loss = 0.0\n running_corrects = 0\n \n for i_batch, sample_batched in enumerate(dataloaders):\n inputs = sample_batched['image']\n labels = sample_batched['label']\n\n inputs = inputs.to(device)\n labels = labels.to(device)\n \n optimizer.zero_grad()\n \n with torch.set_grad_enabled(phase == 0):\n outputs = model(inputs)\n _, preds = torch.max(outputs, 1)\n loss = criterion(outputs, labels)\n \n if phase == 0:\n loss.backward()\n optimizer.step()\n \n running_loss += loss.item() * inputs.size(0)\n running_corrects += torch.sum(preds == labels.data)\n \n epoch_loss = running_loss / dataset_sizes[phase]\n epoch_acc = running_corrects.double() / dataset_sizes[phase]\n \n print('{} Loss: {:.4f} Acc: {:.4f}'.format(\n phase, epoch_loss, epoch_acc))\n \n if phase == 1 and epoch_acc > best_acc:\n best_acc = epoch_acc\n best_model_wts = copy.deepcopy(model.state_dict())\n torch.save(model_ft, '../TrainedModels/FullDataset/DenseNet161_EmotiW')\n\n \n print()\n time_elapsed = time.time() - since\n print('Training complete in {: .0f}m {:0f}s'.format(\n time_elapsed // 60, time_elapsed % 60))\n print('Best val Acc: {:.4f}'.format(best_acc))\n \n model.load_state_dict(best_model_wts)\n return model\n\ncriterion = nn.CrossEntropyLoss()\n\noptimizer_ft = optim.SGD(model_ft.parameters(), lr = 0.001, momentum=0.9)\n\nexp_lr_scheduler = lr_scheduler.StepLR(optimizer_ft, step_size=7, gamma=0.1)\n\nmodel_ft = train_model(model_ft, criterion, optimizer_ft, \n exp_lr_scheduler, num_epochs=epochs)\n\ntorch.save(model_ft, '../TrainedModels/FullDataset/DenseNet161_EmotiW')\n", "sub_path": "Models_FullTrained/FullDataset_DenseNet161_EmotiW.py", "file_name": "FullDataset_DenseNet161_EmotiW.py", "file_ext": "py", "file_size_in_byte": 8091, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "torch.device", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 41, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 42, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 43, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 51, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 54, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 79, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 79, "usage_type": "name"}, {"api_name": "torchvision.transforms.RandomResizedCrop", "line_number": 80, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 80, "usage_type": "name"}, {"api_name": "torchvision.transforms.RandomHorizontalFlip", "line_number": 81, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 81, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 82, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 82, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 83, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 83, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 86, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 86, "usage_type": "name"}, {"api_name": "torchvision.transforms.Resize", "line_number": 87, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 87, "usage_type": "name"}, {"api_name": "torchvision.transforms.CenterCrop", "line_number": 88, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 88, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 89, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 89, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 90, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 90, "usage_type": "name"}, {"api_name": "torch.utils.data.Dataset", "line_number": 93, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 95, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 95, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 127, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 127, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 131, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 136, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 145, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 149, "usage_type": "call"}, {"api_name": "torchvision.models.densenet161", "line_number": 158, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 158, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 160, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 160, "usage_type": "name"}, {"api_name": "time.time", "line_number": 170, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 172, "usage_type": "call"}, {"api_name": "torch.set_grad_enabled", "line_number": 200, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 202, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 210, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 220, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 221, "usage_type": "call"}, {"api_name": "time.time", "line_number": 225, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 233, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 233, "usage_type": "name"}, {"api_name": "torch.optim.SGD", "line_number": 235, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 235, "usage_type": "name"}, {"api_name": "torch.optim.lr_scheduler.StepLR", "line_number": 237, "usage_type": "call"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 237, "usage_type": "name"}, {"api_name": "torch.save", "line_number": 242, "usage_type": "call"}]}
+{"seq_id": "267735381", "text": "# Copyright © 2019 Province of British Columbia\n#\n# Licensed under the Apache License, Version 2.0 (the 'License');\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an 'AS IS' BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"Service to manage PayBC communication.\"\"\"\nimport base64\nimport datetime\nfrom typing import Any, Dict\n\nfrom flask import current_app\n\nfrom pay_api.utils.enums import AuthHeaderType, ContentType\n\nfrom .oauth_service import OAuthService\n\n\nclass PayBcService(OAuthService):\n \"\"\"Service to manage PayBC communication.\"\"\"\n\n def __init__(self):\n \"\"\"Init.\"\"\"\n super()\n\n def create_payment_records(self, invoice_request):\n \"\"\"Create payment related records in PayBC.\"\"\"\n current_app.logger.debug('Inside create invoice')\n return invoice\n\n def get_token(self):\n \"\"\"Generate oauth token from payBC which will be used for all communication.\"\"\"\n current_app.logger.debug('Getting token')\n return token_response\n\n def create_party(self, access_token, invoice_request):\n \"\"\"Create a party record in PayBC.\"\"\"\n current_app.logger.debug('Creating party Record')\n return party_response.json()\n\n def create_account(self, access_token, party, invoice_request):\n \"\"\"Create account record in PayBC.\"\"\"\n current_app.logger.debug('Creating account')\n return account_response.json()\n\n def create_site(self, access_token, account, invoice_request):\n \"\"\"Create site in PayBC.\"\"\"\n current_app.logger.debug('Creating site ')\n return site_response.json()\n\n def create_contact(self, access_token, site, invoice_request):\n \"\"\"Create contact in PayBC.\"\"\"\n current_app.logger.debug('Creating site contact')\n return contact_response.json()\n\n def create_invoice(self, access_token, site, invoice_request):\n \"\"\"Create invoice in PayBC.\"\"\"\n current_app.logger.debug('Creating PayBC Invoice Record')\n return invoice_response.json()\n", "sub_path": "pay-api/src/pay_api/services/paybc.py", "file_name": "paybc.py", "file_ext": "py", "file_size_in_byte": 7777, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "oauth_service.OAuthService", "line_number": 26, "usage_type": "name"}, {"api_name": "flask.current_app.logger.debug", "line_number": 35, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 35, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 35, "usage_type": "name"}, {"api_name": "flask.current_app.logger.debug", "line_number": 44, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 44, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 44, "usage_type": "name"}, {"api_name": "flask.current_app.logger.debug", "line_number": 49, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 49, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 49, "usage_type": "name"}, {"api_name": "flask.current_app.config.get", "line_number": 50, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 50, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 50, "usage_type": "name"}, {"api_name": "base64.b64encode", "line_number": 51, "usage_type": "call"}, {"api_name": "flask.current_app.config.get", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 52, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 52, "usage_type": "name"}, {"api_name": "pay_api.utils.enums.AuthHeaderType.BASIC", "line_number": 55, "usage_type": "attribute"}, {"api_name": "pay_api.utils.enums.AuthHeaderType", "line_number": 55, "usage_type": "name"}, {"api_name": "pay_api.utils.enums.ContentType.FORM_URL_ENCODED", "line_number": 55, "usage_type": "attribute"}, {"api_name": "pay_api.utils.enums.ContentType", "line_number": 55, "usage_type": "name"}, {"api_name": "flask.current_app.logger.debug", "line_number": 57, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 57, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 57, "usage_type": "name"}, {"api_name": "flask.current_app.logger.debug", "line_number": 62, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 62, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 62, "usage_type": "name"}, {"api_name": "flask.current_app.config.get", "line_number": 63, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 63, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 63, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 64, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 64, "usage_type": "name"}, {"api_name": "pay_api.utils.enums.AuthHeaderType.BEARER", "line_number": 68, "usage_type": "attribute"}, {"api_name": "pay_api.utils.enums.AuthHeaderType", "line_number": 68, "usage_type": "name"}, {"api_name": "pay_api.utils.enums.ContentType.JSON", "line_number": 68, "usage_type": "attribute"}, {"api_name": "pay_api.utils.enums.ContentType", "line_number": 68, "usage_type": "name"}, {"api_name": "flask.current_app.logger.debug", "line_number": 69, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 69, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 69, "usage_type": "name"}, {"api_name": "flask.current_app.logger.debug", "line_number": 74, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 74, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 74, "usage_type": "name"}, {"api_name": "flask.current_app.config.get", "line_number": 75, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 75, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 75, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 77, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 77, "usage_type": "name"}, {"api_name": "pay_api.utils.enums.AuthHeaderType.BEARER", "line_number": 82, "usage_type": "attribute"}, {"api_name": "pay_api.utils.enums.AuthHeaderType", "line_number": 82, "usage_type": "name"}, {"api_name": "pay_api.utils.enums.ContentType.JSON", "line_number": 82, "usage_type": "attribute"}, {"api_name": "pay_api.utils.enums.ContentType", "line_number": 82, "usage_type": "name"}, {"api_name": "flask.current_app.logger.debug", "line_number": 83, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 83, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 83, "usage_type": "name"}, {"api_name": "flask.current_app.logger.debug", "line_number": 88, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 88, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 88, "usage_type": "name"}, {"api_name": "flask.current_app.config.get", "line_number": 89, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 89, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 89, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 91, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 91, "usage_type": "name"}, {"api_name": "pay_api.utils.enums.AuthHeaderType.BEARER", "line_number": 103, "usage_type": "attribute"}, {"api_name": "pay_api.utils.enums.AuthHeaderType", "line_number": 103, "usage_type": "name"}, {"api_name": "pay_api.utils.enums.ContentType.JSON", "line_number": 103, "usage_type": "attribute"}, {"api_name": "pay_api.utils.enums.ContentType", "line_number": 103, "usage_type": "name"}, {"api_name": "flask.current_app.logger.debug", "line_number": 105, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 105, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 105, "usage_type": "name"}, {"api_name": "flask.current_app.logger.debug", "line_number": 110, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 110, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 110, "usage_type": "name"}, {"api_name": "flask.current_app.config.get", "line_number": 111, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 111, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 111, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 113, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 113, "usage_type": "name"}, {"api_name": "pay_api.utils.enums.AuthHeaderType.BEARER", "line_number": 122, "usage_type": "attribute"}, {"api_name": "pay_api.utils.enums.AuthHeaderType", "line_number": 122, "usage_type": "name"}, {"api_name": "pay_api.utils.enums.ContentType.JSON", "line_number": 122, "usage_type": "attribute"}, {"api_name": "pay_api.utils.enums.ContentType", "line_number": 122, "usage_type": "name"}, {"api_name": "flask.current_app.logger.debug", "line_number": 124, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 124, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 124, "usage_type": "name"}, {"api_name": "flask.current_app.logger.debug", "line_number": 129, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 129, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 129, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 130, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 130, "usage_type": "attribute"}, {"api_name": "flask.current_app.config.get", "line_number": 132, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 132, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 132, "usage_type": "name"}, {"api_name": "pay_api.utils.enums.AuthHeaderType.BEARER", "line_number": 159, "usage_type": "attribute"}, {"api_name": "pay_api.utils.enums.AuthHeaderType", "line_number": 159, "usage_type": "name"}, {"api_name": "pay_api.utils.enums.ContentType.JSON", "line_number": 159, "usage_type": "attribute"}, {"api_name": "pay_api.utils.enums.ContentType", "line_number": 159, "usage_type": "name"}, {"api_name": "flask.current_app.logger.debug", "line_number": 161, "usage_type": "call"}, {"api_name": "flask.current_app.logger", "line_number": 161, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 161, "usage_type": "name"}]}
+{"seq_id": "2313348", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Mar 1 16:12:05 2021\n\n@author: student\n\"\"\"\nimport pygame\nimport random\nfrom pygame.locals import *\n\n\npygame.init()\n\n(width, height) = (600, 400)\nscreen = pygame.display.set_mode((width, height))\npygame.display.set_caption('Clash_of_Brothers')\n\nfont_main = pygame.font.SysFont('Constantia', 30)\n\nclass Screen():\n \n def __init__(self, image_link):\n self.img = pygame.image.load(image_link).convert()\n self.img = pygame.transform.scale(self.img, (width, height))\n \n self.sprite_img = pygame.sprite.Sprite()\n self.sprite_img.image = self.img\n self.sprite_img.rect = self.img.get_rect\n \n def add_scenario(self, rect, punto, text, loc, font_color = (139, 0, 0)):\n font = pygame.font.SysFont('Constantia', punto)\n pygame.draw.rect(self.img, (211, 211, 211), rect)\n blit_text(self.sprite_img.image, text, loc, font, font_color)\n\nclicked = False \n\nclass Button():\n \n button_col = (0, 0, 0)\n hover_col = (70, 70, 70)\n click_col = (35, 35, 35)\n text_col = (34, 139, 234)\n \n def __init__(self, posx, posy, w, h, img, punto, text):\n self.x = posx\n self.y = posy\n self.width = w\n self.height = h\n self.screen = img\n self.font = pygame.font.SysFont('Constantia', punto)\n self.text = text\n \n def draw_button(self):\n \n global clicked\n action = False\n \n pos = pygame.mouse.get_pos()\n button_rect = Rect(self.x, self.y, self.width, self.height)\n \n if button_rect.collidepoint(pos):\n if pygame.mouse.get_pressed()[0] == 1:\n clicked = True\n pygame.draw.rect(self.screen, self.click_col, button_rect)\n elif pygame.mouse.get_pressed()[0] == 0 and clicked == True:\n clicked = False\n action = True\n else:\n pygame.draw.rect(self.screen, self.hover_col, button_rect)\n else:\n pygame.draw.rect(self.screen, self.button_col, button_rect)\n \n pygame.draw.line(self.screen, (0, 0, 139), (self.x, self.y), (self.x + self.width, self.y), 2)\n pygame.draw.line(self.screen, (0, 0, 139), (self.x, self.y), (self.x, self.y + self.height), 2)\n pygame.draw.line(self.screen, (0, 0, 139), (self.x, self.y + self.height), (self.x + self.width, self.y + self.height), 2)\n pygame.draw.line(self.screen, (0, 0, 139), (self.x + self.width, self.y), (self.x + self.width, self.y + self.height), 2)\n \n text_img = self.font.render(self.text, True, self.text_col)\n text_len = text_img.get_width()\n self.screen.blit(text_img, (self.x + int(self.width / 2) - int(text_len / 2), self.y + 12))\n return action\n\nclass Text_Box():\n \n text_color = (34, 139, 234)\n \n def __init__(self, posx, posy, w, h, img, font, text = ''):\n self.x = posx\n self.y = posy\n self.width = w\n self.height = h\n self.rect = Rect(posx, posy, w, h)\n self.screen = img\n self.font = font\n self.text = text\n self.text_surface = self.font.render(text, True, self.text_color)\n self.active = False\n \n def handle_event(self, event):\n text_rectangle = Rect(self.x, self.y, self.width, self.height)\n if event.type == pygame.MOUSEBUTTONDOWN:\n if text_rectangle.collidepoint(event.pos):\n self.active = not self.active\n else:\n self.active = False\n if event.type == pygame.KEYDOWN:\n if self.active:\n if event.key == pygame.K_RETURN:\n final_text = self.text\n self.text = ''\n return final_text\n elif event.key == pygame.K_BACKSPACE:\n self.text = self.text[:-1]\n else:\n self.text += event.unicode\n self.text_surface = self.font.render(self.text, True, self.text_color)\n \n def draw_text_box(self):\n text_box_rect = Rect(self.x, self.y, self.width, self.height)\n pygame.draw.rect(self.screen, (0, 0, 0), text_box_rect)\n \n pygame.draw.line(self.screen, (0, 0, 139), (self.x, self.y), (self.x + self.width, self.y), 2)\n pygame.draw.line(self.screen, (0, 0, 139), (self.x, self.y), (self.x, self.y + self.height), 2)\n pygame.draw.line(self.screen, (0, 0, 139), (self.x, self.y + self.height), (self.x + self.width, self.y + self.height), 2)\n pygame.draw.line(self.screen, (0, 0, 139), (self.x + self.width, self.y), (self.x + self.width, self.y + self.height), 2)\n \n text_len = self.text_surface.get_width()\n self.screen.blit(self.text_surface, (self.x + int(self.width / 2) - int(text_len / 2), self.y + 10))\n \ndef blit_text(surface, text, pos, font, color=pygame.Color('black')):\n words = [word.split(' ') for word in text.splitlines()] # 2D array where each row is a list of words.\n space = font.size(' ')[0] # The width of a space.\n max_width, max_height = width,height\n x, y = pos\n for line in words:\n for word in line:\n word_surface = font.render(word, 0, color)\n word_width, word_height = word_surface.get_size()\n if x + word_width >= max_width:\n x = pos[0] # Reset the x.\n y += word_height # Start on new row.\n surface.blit(word_surface, (x, y))\n x += word_width + space\n x = pos[0] # Reset the x.\n y += word_height # Start on new row.\n \nupper_pos = (0, 300, 600, 50)\nlower_pos = (0, 350, 600, 50)\n\n#topkapi 1 (check)\ntopkapi1 = Screen(r'topkapı_palace.jpg')\ntopkapi1_text = \"Welcome to Clash of Brothers!\"\nstart = Button(50, 280, 150, 44, topkapi1.sprite_img.image, 30, \"Start\")\nleave = Button(400, 280, 150, 44, topkapi1.sprite_img.image, 30, \"Leave\") \n#----------------------------------------------------------------------------------------------------------------------------------- \n\n#topkapi 2 (check())\ntopkapi2 = Screen(r'topkapı_palace_2.jpg') \ntopkapi2_text = \"You are an heir of the Ottoman Empire! Please pick a name for your character.\"\ntopkapi2_box = Text_Box(50, 300, 500, 30, topkapi2.sprite_img.image, pygame.font.SysFont('Constantia', 25))\n#----------------------------------------------------------------------------------------------------------------------------------- \n\n#karaman (check)\nkaraman = Screen(r'karaman.jpg')\ncontinue_button = Button(225, 280, 120, 40, karaman.sprite_img.image, 30, \"Continue\") \ndef gen_karaman_text(name):\n return \"Hello, Shahzade \" + name + \". You are ruling the province of Karaman which is close to the capital, Constantinople. \" \\\n \"Your father has always wanted you to rule the empire after his death. The governor of Konya, your older brother, is \" \\\n \"your political rival. He is jealous of you because of your close relationship with the emperor.\" \n#----------------------------------------------------------------------------------------------------------------------------------- \n \n#letter (check)\nletter = Screen(r'letter.jpg')\nletter_text = \"Here is a letter from Constantinople my Shahzade! Our great emperor has passed away. We should take off \" \\\n \"as soon as possible to arrive at the court before your brother. If your brother reaches to the capital before us, \" \\\n \"he might take over the throne.\" \nletter_humane = Button(0, 300, 600, 50, letter.sprite_img.image, 22, \"What a lamentable day! We can take off later. I am in mourning.\")\nletter_pragmatic = Button(0, 350, 600, 50, letter.sprite_img.image, 22, \"You are right, we should leave soon. I will be the emperor!\")\n#----------------------------------------------------------------------------------------------------------------------------------- \n\n#bursa and room (check)\nbursa = Screen(r'bursa.jpg')\nbursa_hum_text = \"You are near the Bursa province which is very close the capital Constantinople. You learned that your brother's men \" \\\n \"had managed to delay the arrival of the letter that informed you of the emperor's death. Hence, your brother is the emperor now. \" \\\n \"He wants your head! What will you do now?\"\nbursa_humane = Button(0, 300, 600, 50, bursa.sprite_img.image, 22, \"Let's retreat to my province, Karaman. Maybe, we can negotiate with him.\")\nbursa_pragmatic = Button(0, 350, 600, 50, bursa.sprite_img.image, 22, \"Attack Bursa and start a revolt!\")\n\n\nkaraman_room = Screen(r'karaman_room.jpg')\nkr_text = \"Your brother has taken over the throne my Shahzade! He has ordered your execution! What should we do now?\"\nkr_humane = Button(0, 300, 300, 50, karaman_room.sprite_img.image, 19, \"We should consider seeking help Egypt.\")\nkr_pragmatic = Button(300, 300, 300, 50, karaman_room.sprite_img.image, 22, \"Lets form alliances in Italy.\")\nkr_braveheart = Button(0, 350, 600, 50, karaman_room.sprite_img.image, 22, \"Prepare the army! We will fight\")\n#----------------------------------------------------------------------------------------------------------------------------------- \n\n#karaman_2 (check)\n\nkaraman_room2 = Screen(r'karaman_room.jpg')\nkr2_text = \"The new emperor, your brother, liked your retreat move. He called you to Constantinople and promised that he will \" \\\n \"not order your execution.\"\n \nkr2_humane = Button(0, 300, 600, 50, karaman_room2.sprite_img.image, 22, \"He is my brother. I believe him. Pack up my stuff. We will go to Constantinople.\")\nkr2_pragmatic = Button(0, 350, 600, 50, karaman_room2.sprite_img.image, 22, \"I don't believe him. He is a filthy liar. I will raise a strong army against him!\")\n\nkaraman_room2_2 = Screen(r'karaman_room.jpg')\nkr2_2_text = \"Whom should we seek help from?\"\nkr2_2_humane = Button(0, 300, 600, 50, karaman_room2_2.sprite_img.image, 22, \"Our Muslim brother in Egypt can help us.\")\nkr2_2_pragmatic = Button(0, 350, 600, 50, karaman_room2_2.sprite_img.image, 22, \"Lets form alliances in Italy.\")\n#----------------------------------------------------------------------------------------------------------------------------------- \n\n#Constantinople_hum (check)\n\nconstantinople_hum = Screen(r'constantinople_hum.jpg')\nconstantinople_hum2 = Screen(r'constantinople_bad.jpg')\nconst_text_good = \"Your brother has kept his promise and declared you as his second man. You lived a happy life.\" \\\n \"Your brothership and loyalty towards each other will always be remembered in the history of the Ottoman Empire!\"\nconst_text_bad = \"You noticed that something is going wrong. Suddenly, 8 to 10 executioners entered your room. They will strangle you!\"\nconst_hum_good = Button(0, 350, 600, 50, constantinople_hum.sprite_img.image, 22, \"What a beautiful life I have! My brother and I will never betray each other.\")\nconst_hum_bad = Button(0, 350, 600, 50, constantinople_hum2.sprite_img.image, 22, \"I will beat them with my bare hands!\")\n#-----------------------------------------------------------------------------------------------------------------------------------\n\n#Strangle (check)\ntopkapi_escape = Screen(r'topkapi_escape.jpg')\ntopkapi_esc_text = \"You are a strong warrior. You managed to beat all of the executioners and ran away from the palace. What will you do\" \\\n \"now my Shahzade?\"\nescape_hum = Button(0, 300, 600, 50, topkapi_escape.sprite_img.image, 22, \"I will seek help from our Muslims brothers in Egypt\")\nescape_prag = Button(0, 350, 600, 50, topkapi_escape.sprite_img.image, 22, \"I can secure support from European states. Let's go to Italy\")\n\ntopkapi_early_dead = Screen(r'executioners.jpg')\ntopkapi_early_dead_text = \"You have been brutally killed by your brother's executioners. You will always be remembered as a loyal and honorable \" \\\n \"shahzade. You have counted on your brother but he betrayed you. He will always be remembered as a \" \\\n \"cruel and wicked leader.\"\n\nmain_men = Button(0, 350, 600, 50, topkapi_early_dead.sprite_img.image, 22, \"Return to main menu\")\n#-----------------------------------------------------------------------------------------------------------------------------------\n\n#Attack Bursa (check)\nbursa_fail = Screen(r'bursa_lost.jpg')\nbursa_fail_text= \"Our army is losing my Shahzade! The capital army is very strong. We should retreat, but where? What should we do?\"\n\nbursa_egypt = Button(0, 300, 600, 50, bursa_fail.sprite_img.image, 22, \"Let's go to Egypt. One day, we shall return to home!\")\nbursa_italy = Button(0, 350, 600, 50, bursa_fail.sprite_img.image, 22, \"Let's seek help from the Pope. We can beat my brother with the help of the Pope.\")\n\nbursa_after_istanbul_text = \"Which country should we seek help from my Shahzade?\"\nbursa_success = Screen(r'bursa_capture.jpg')\ndef bursa_success_text(name):\n return \"You have defeated the emperor's army! People of Bursa are cheering and celebrating the victory of TRUE emperor! Hail emperor \" + name + \"!\"\nbursa_direct= Button(0, 300, 600, 50, bursa_success.sprite_img.image, 22, \"Either Constantinople will conquer me or I will conquer Constantinople!\")\nbursa_direct2 = Button(0, 350, 600, 50, bursa_success.sprite_img.image, 22, \"Time to attack Constantinople!\")\n#-----------------------------------------------------------------------------------------------------------------------------------\n\n#Istanbul Siege One Army (check)\nistanbul_siege_1 = Screen(r'istanbul_walls.jpg')\nistanbul_siege_1_text = \"It is very difficult to conquer Constantinople my Shahzade! There are huge walls, professional guards, and the capital, \" \\\n \"army. Are you sure you want to attack the city without any support from other nations?\"\nistanbul_siege1_support_egypt = Button(300, 300, 300,50, istanbul_siege_1.sprite_img.image, 22, \"Let's go to Egypt to form an alliance.\")\nistanbul_siege1_support_italy = Button(0, 300, 300, 50, istanbul_siege_1.sprite_img.image, 22, \"Let's seek support from Europe\")\nistanbul_siege1_direct = Button(0, 350,600, 50, istanbul_siege_1.sprite_img.image, 22, \"I have defeated my brother once. I can defeat him again!\")\n#-----------------------------------------------------------------------------------------------------------------------------------\n\n#Istanbul Siege Result (check)\nistanbul_siege_loss_1 = Screen(r'istanbul_siege.jpg')\nistanbul_win = Screen(r'istanbul_win.jpg')\n\nistanbul_siege_loss_text1 = \"Our army has been defeated my Shahzade! We should consider running away.\"\nsiege_loss_1_egypt = Button(0, 300, 600, 50, istanbul_siege_loss_1.sprite_img.image, 22, \"Let's head to Egypt.\")\nsiege_loss_1_italy = Button(0, 350, 600, 50, istanbul_siege_loss_1.sprite_img.image, 22, \"Let's head to Italy.\")\n\n\nistanbul_siege_win = \"With your mighty army you have managed to take over the throne my emperor! Will you forgive your brother?\"\nsiege_win_brutal = Button(0, 300, 600, 50, istanbul_win.sprite_img.image, 22, \"Send the executioners! He deserved it.\")\nsiege_win_humane = Button(0, 350, 600, 50, istanbul_win.sprite_img.image, 22, \"He is my brother. I won't be like him.\")\n#-----------------------------------------------------------------------------------------------------------------------------------\n\n#Letter_runaway\nletter_run_away = Screen(r'large_army.jpg')\nletter_run_away_text= \"Your brother has raised a very large army. He is coming at us! We can't win this battle. We should take off soon. \" \\\n \"Where should we go sir?\"\n#escape_hum\n#escape_prag\n#-----------------------------------------------------------------------------------------------------------------------------------\n \n#Karaman Room Braveheart (check)\nbraveheart_dead = Screen(r'braveheart_dead.jpg')\nbraveheart_dead_text = \"We lost my shahzade. You are killed by an archer during the combat. You will always be remembered as a brave \" \\\n \"and benevolent historical figure. Your loyalty, charisma, and courage will never be forgotten.\" \n#main_men button\n \nbraveheart_alive = Screen(r'sivas_province.jpg')\nbraveheart_alive_text = \"You managed to escape from the siege my shahzade! We should go to Egypt as the emperor's army is very powerful in \" \\\n \"the West.\"\nalive_hum = Button(0, 300, 600, 50, braveheart_alive.sprite_img.image, 22, \"Will I be able to return to my home again?\")\nalive_prag = Button(0, 350, 600, 50, braveheart_alive.sprite_img.image, 22, \"Let's head to Cairo and return with a more powerful army!\")\n#-----------------------------------------------------------------------------------------------------------------------------------\n\n#Egypt road (check)\nbrigands = Screen(r'medieval_brigand.jpg')\nbrigand_text = \"Oh no! Brigands cut off your escape. They want you to drop your weapon and give them money. Be careful my Shahzade. \" \\\n \"They might be your brother's men. \"\n\nkill_brigands = Button(0, 300, 600, 50, brigands.sprite_img.image, 22, \"We must fight them!\")\npay_brigands = Button(0, 350, 600, 50, brigands.sprite_img.image, 22, \"Pay them\")\n#-------------------------------------------------------------------------------------------------------------------------------\n\n#Brigand Fight (check)\nbrigand_win = Screen(r'brigand_win.jpg')\nbrigand_win_text = \"You beated them my Shahzade. You are a skilled warrior! Let's continue our journey to Egypt'\"\n#continue_button\n\nbrigand_lose = Screen(r'brigand_lose.jpg')\nbrigand_lose_text = \"Your are killed by brigands. No news will be heard from you. No one knows where your corpse is. You will always known \" \\\n \"as a traitor who sought help from Mamluks (Egypt)\" \n \n#return to main menu\n#-------------------------------------------------------------------------------------------------------------------------------\n\n#Pay Brigand (check)\npay_brigand_win = Screen(r'brigand_win.jpg')\npay_brigand_win_text = \"After taking all of your money, brigands allowed you to travel as you wish.\"\n#continue_button\n\npay_brigand_lose = Screen(r'brigand_captive.jpg')\npay_brigand_lose_text = \"Turns out that these men are your brother's spies. They handcuffed you and took you as a captive. They will deliver you over \" \\\n \"to your brother in Constantinople!\"\nescape = Button(0, 350, 600, 50, pay_brigand_lose.sprite_img.image, 22, \"I will escape!\")\n#-------------------------------------------------------------------------------------------------------------------------------\n\n#Escape Result (check)\nexecutioner = Screen(r'executioners.jpg')\nbrigand_ending = \"You could not escape. Executioners have been sent to your prison. You are dead and will always be remember as a traitor \" \\\n \"who sought help from Mamluks. (Egypt)\"\nescaped_text = \"You have managed to escape and met Mamluk Sultan. He welcomes you with great hospitality. How will you convince him for a military alliance against \" \\\n \"your brother.\"\n#-------------------------------------------------------------------------------------------------------------------------------\n \n#Mamluks (check)\nmamluks = Screen(r'mamluks.jpg')\nmamluks_text = \"You have met Mamluk Sultan! He welcomes you with great hospitality. How will you convince him for a military alliance against \" \\\n \"your brother.\"\nmamluk_traitor = Button(0, 300, 600, 50, mamluks.sprite_img.image, 22, \"I might give you some province if you support me.\")\nmamluk_hum = Button(0, 350, 600, 50, mamluks.sprite_img.image, 22, \"I will guarantee the friendship between the Mamluks and Ottomans if you help me.\")\n#-------------------------------------------------------------------------------------------------------------------------------\n\n# (check)\nmamluks_result = Screen(r'egypt.jpg')\n#Mamluk Good\nmamluk_gtraitor_text = \"Your offer sounded very interesting to Mamluk Sultan. He will provide military support. Time to conquer Constantinople!\"\nmamluk_ghum_text = \"Mamluk Sultan values the friendship between the Ottomans and Mamluks. He will provide military support. Prepare for the siege!\"\n#continue button\n\n#Mamluk Bad\nmamluk_prisoner = Screen(r'prisoner.jpg')\nmamluk_bad_hum_text = \"Delivering you over to your brother sounded like a more profitable business for the Mamluk Sultan. Executioners have been sent \" \\\n \"to you and you are dead.\"\n \n#main menu\n#-------------------------------------------------------------------------------------------------------------------------------\n\n#Way to Italy Direct and Pirates (check)\nway_to_italy = Screen(r'way_to_italy.jpg')\nway_to_italy_text = \"The trade ship is ready to take off to Italy. You are hiding your identity from other people.\"\n#continue_button\n\npirates = Screen(r'pirate.jpg')\npirate_text = \"Uh oh! Pirates have taken all the ship crew as captives. We should escape! What will you do now my Shahzade?\"\npay_pirates = Button(0, 300, 600, 50, pirates.sprite_img.image, 22, \"I will reveal my identity and offer them treasure in return for leaving me.\")\nrevolt = Button(0, 350, 600, 50, pirates.sprite_img.image, 22, \"Start a revolt against pirates!\")\n#-------------------------------------------------------------------------------------------------------------------------------\n\n#Pirate Success (check)\n#if pay \nitaly_coast = Screen(r'italy_coast.jpg')\nitaly_coast_text = \"Pirates have decided to release you. The pope paid the pirates for your release. You arrived at Italy!\"\npay_good_news = Button(0, 350, 600, 50, italy_coast.sprite_img.image, 22, \"The Pope is the best host I can ever ask for.\")\n#if revolt\npirate_revolt_win = Screen(r'pirate_revolt_win.jpg')\npirate_revolt_win_text = \"You are a true leader and warrior! You and the ship crew managed to beat pirates.\"\nrevolt_good_news = Button(0, 350,600, 50, pirate_revolt_win.sprite_img.image, 22, \"Time to meet the Pope!\")\n#-------------------------------------------------------------------------------------------------------------------------------\n\n#Pirate Failure (check)\n#if pay\nsold_to_brother = Screen(r'pirate_captive.jpg')\nsold_to_brother_text = \"Bad news! Pirates have decided to deliver you over your brother.\"\nexecutioner2 = Screen(r'executioners.jpg')\n\n#Executioner screen\nsold_to_brother_dead = \"You have been executed by your brother's men.\"\n#Return to main menu\n\n#if revolt\npirate_dead = Screen(r'pirate_dead.jpg')\npirate_dead_text = \"You have been killed by pirates. No news will be heard of you ever again. Your corpse will be in the bottom of the sea.\"\n#Return to main menu \n#-------------------------------------------------------------------------------------------------------------------------------\n\n#Pope (check)\npope = Screen(r'Pope.jpg')\npope_text = \"Pope: Welcome sir! We know that the Ottoman throne should be yours. We believe that such a mighty empire as the Ottomans should \" \\\n \"serve Jesus and Lord. If you follow the path of Jesus, we can raise an army for you.\"\nchristian_button = Button(0, 350, 600, 50, pope.sprite_img.image, 22, \"I will fight for Jesus and take over the Ottoman throne!\")\nmuslim = Button(0, 300, 600, 50, pope.sprite_img.image, 22, \"This is treason against my home country. I shall not accept this offer!\")\nold_man = Screen(r'old_man.jpg')\npope_text2 = \"Pope refused to offer you an army. Nor did he allow you to travel as you wished. You died of old age in Italy\"\n#Return to main menu\n#-------------------------------------------------------------------------------------------------------------------------------\n\n#Pope Army (check)\nchristian_army = Screen(r'christian_army.jpg')\n#if christian\nchristian_army_prag = \"You are ruling the papal army as a Christian general. You are known as a traitor within the borders of the Ottoman empire. \" \\\n \"Time to take over the throne my lord!\"\n#if muslim\nchristian_army_hum = \"The Pope admired your boldness and principled attitude. He decided to give you an army in return for a truce agreement\" \\\n \"between the Christian states and the Ottoman empire if you take the lead of the country.\"\n#-------------------------------------------------------------------------------------------------------------------------------\n\n#istanbul Siege Only One Army\nsiege_only_egypt = Screen(r'istanbul_siege_2.jpg')\nsiege_only_egypt_text = \"We have support from the Egyptian army! However, Constantinople's defense is very powerful. What should we do \" \\\n \"my Shahzade?\"\nto_italy = Button(0, 350, 600, 50, siege_only_egypt.sprite_img.image, 22, \"We should ask the Pope for help. Let's go to Rome.\")\nattack_one_army_egypt = Button(0, 300, 600, 50, siege_only_egypt.sprite_img.image, 22, \"Our army is strong enough. Let's attack the city!\")\n\nsiege_only_italy = Screen(r'istanbul_siege_2.jpg')\nsiege_only_italy_text = \"The Pope is supporting us my Shahzade! However, Constantinople's defense is very powerful. What should we do?\"\nto_egypt = Button(0, 350, 600, 50, siege_only_italy.sprite_img.image, 22, \"Let's seek help from Mamluks.\")\nattack_one_army_italy = Button(0, 300, 600, 50, siege_only_italy.sprite_img.image, 22, \"Our army is strong enough. Let's attack the city!\")\n#-------------------------------------------------------------------------------------------------------------------------------\n\n#Istanbul Siege Two Armies\ntwo_army_siege = Screen(r'large_army.jpg')\ntwo_army_siege_text = \"We have a powerful army now! Let's attack my Shahzade.\"\ntwo_army_siege_button = Button(0, 350, 600, 50, two_army_siege.sprite_img.image, 22, \"Prepare for the siege!\")\n#-------------------------------------------------------------------------------------------------------------------------------\n\n#if istanbul win brutal \ntyrant_king = Screen(r'tyrant_king.jpg')\ntyrant_king_text_christ = \"You are not only a traitor who have been converted to Christianity but also a ruthless leader who executed \" \\\n \"his brother. You will not be known as a virtuous leader.\"\ntyrant_king_text = \"You become a tyrant after the execution of your brother. You will always be known as a sibling slayer.\"\ntyrant_king_button = Button(0 , 350, 600, 50, tyrant_king.sprite_img.image, 22, \"I have no toleration for disobedience!\")\n\n#if Istanbul win \ngood_king_christ = Screen(r'good_king.jpg')\ngood_king_christ_text = \"You will always known as a sinful emperor who have received help from the Pope to take over the Ottoman throne. \" \\\n \"However, your affection towards your brother will always be remembered.\"\n \ngood_king_screen = Screen(r'benevolent_king.jpg')\ndef good_king_gen(name):\n return \"You become the emperor! Shall I call you emperor \" + name + \"? You affection towards your family will always be remembered. \" \\\n \"You are not only a benevolent leader but also a true military commander who succesfully managed to take over the throne.\" \ngood_king = Button(0, 350, 600, 50, good_king_screen.sprite_img.image, 22, \"Wise leaders should rule their countries with benevolence!\")\ngood_king_christian = Button(0, 350, 600, 50, good_king_christ.sprite_img.image, 22, \"I am a benevolent leader at least.\")\n#-------------------------------------------------------------------------------------------------------------------------------\n\n#if istanbul loss brutal \nfinal_loss_only_egypt = Screen(r'final_loss.jpg')\nfinal_loss_egypt_text = \"Our army has been destroyed. Mamluk's support was not enough. You have been captured during the siege and executed.\"\n\nfinal_loss_only_italy = Screen(r'traitor.jpg')\nfinal_loss_italy_text = \"People of the Ottoman empire are happy about your execution. You are a traitor who has converted to Christianity \" \\\n \"to receive help from the Pope.\"\n\nfinal_loss_only_italy_2 = Screen(r'medieval_execution.jpg')\nfinal_loss_italy_text2 = \"Our army has been destroyed. The Pope's support did not help us conquer Constantinople. You have been captured during \" \\\n \"the siege and exectued.\"\n\nboth_loss = Screen(r'heroic_execution.jpg')\nfinal_loss_both_army = \"Your brother has won the historical battle. Even though you had support from two nations, the capital army fought really well. \" \\\n \"Your brother has ordered your execution.\"\n#return to main menu\n#-------------------------------------------------------------------------------------------------------------------------------\n\n\ncurrent_screen = \"topkapi1\"\n\nwhile True:\n \n if current_screen == \"topkapi1\":\n army = 1\n dice = None\n egypt_army = False\n pope_army = False\n brother_kill = False\n christian = False\n bursa_direct_clicked = False\n escaped = False\n ambition_point = 0 \n \n screen.blit(topkapi1.img, (0,0))\n topkapi1.add_scenario(Rect(105, 40, 400, 30), 30, topkapi1_text, (160, 48))\n if start.draw_button():\n current_screen = \"topkapi2\"\n \n if current_screen == \"topkapi2\":\n screen.blit(topkapi2.img, (0,0))\n topkapi2.add_scenario(Rect(0, 40, 600, 50), 25, topkapi2_text, (60,48)) \n \n if current_screen == \"karaman\":\n screen.blit(karaman.img, (0,0))\n karaman_text = gen_karaman_text(char_name)\n karaman.add_scenario(Rect(0, 0, 600, 100), 25, karaman_text, (15, 10))\n if continue_button.draw_button():\n current_screen = \"letter\"\n \n if current_screen == \"letter\":\n screen.blit(letter.img, (0,0))\n letter.add_scenario(Rect(0, 0, 600, 90), 25, letter_text, (15, 10))\n if letter_humane.draw_button():\n current_screen = \"karaman_room\"\n if letter_pragmatic.draw_button():\n current_screen = \"bursa\"\n \n if current_screen == \"karaman_room\":\n screen.blit(karaman_room.img, (0,0))\n karaman_room.add_scenario(Rect(0, 0, 600, 70), 25, kr_text, (15, 10))\n if kr_humane.draw_button():\n current_screen = \"brigand\"\n if kr_pragmatic.draw_button():\n current_screen = \"pre-pirate\"\n if kr_braveheart.draw_button():\n \n if dice == None:\n dice = random.random()\n if dice <= 0.35:\n current_screen = \"braveheart_dead\"\n else:\n current_screen = \"braveheart_alive\"\n dice = None\n \n if current_screen == \"braveheart_dead\": \n screen.blit(braveheart_dead.img, (0,0))\n braveheart_dead.add_scenario(Rect(0, 0, 600, 70), 25, braveheart_dead_text, (15, 10))\n mainmen2 = Button(0, 350, 600, 50, braveheart_dead.sprite_img.image, 22, \"Return to main menu\")\n if mainmen2.draw_button():\n current_screen = \"topkapi1\"\n \n if current_screen == \"braveheart_alive\":\n screen.blit(braveheart_alive.img, (0,0))\n braveheart_alive.add_scenario(Rect(0, 0, 600, 60), 25, braveheart_alive_text, (15, 10))\n if alive_hum.draw_button():\n current_screen = \"brigand\"\n if alive_prag.draw_button():\n ambition_point = 0.1\n current_screen = \"brigand\"\n \n if current_screen == \"bursa\":\n screen.blit(bursa.img, (0,0))\n bursa.add_scenario(Rect(0, 0, 600, 100), 25, bursa_hum_text, (15, 10))\n if bursa_humane.draw_button():\n current_screen = \"karaman_room2\"\n if bursa_pragmatic.draw_button():\n if dice == None:\n dice = random.random()\n if dice <= 0.5:\n current_screen = \"bursa_success\"\n else:\n army -= 1\n current_screen = \"bursa_fail\"\n dice = None\n \n if current_screen == \"bursa_success\":\n screen.blit(bursa_success.img, (0,0))\n bursa_text_success = bursa_success_text(char_name)\n bursa_success.add_scenario(Rect(0, 0, 600, 60), 25, bursa_text_success, (15, 10))\n if bursa_direct.draw_button():\n ambition_pont = 0.1\n current_screen = \"istanbul_walls\"\n bursa_direct_clicked = True\n if bursa_direct2.draw_button():\n current_screen = \"istanbul_walls\"\n \n if current_screen == \"istanbul_walls\":\n screen.blit(istanbul_siege_1.img, (0,0))\n istanbul_siege_1.add_scenario(Rect(0, 0, 600, 70), 25, istanbul_siege_1_text, (15,10))\n if istanbul_siege1_support_egypt.draw_button():\n current_screen = \"brigand\"\n if istanbul_siege1_support_italy.draw_button():\n current_screen = \"pre-pirate\"\n if istanbul_siege1_direct.draw_button():\n if dice == None:\n dice = random.random()\n if dice <= 0.05 + ambition_point:\n current_screen = \"istanbul_win\"\n else:\n army -= 1\n current_screen = \"istanbul_loss_early\"\n dice = None\n \n if current_screen == \"istanbul_win\":\n screen.blit(istanbul_win.img, (0,0))\n istanbul_win.add_scenario(Rect(0, 0, 600, 60), 25, istanbul_siege_win, (15, 10))\n if siege_win_brutal.draw_button():\n current_screen = \"tyrant_king\"\n if siege_win_humane.draw_button():\n current_screen = \"good_king\"\n \n if current_screen == \"istanbul_loss_early\":\n screen.blit(istanbul_siege_loss_1.img, (0,0))\n istanbul_siege_loss_1.add_scenario(Rect(0, 0, 600, 60), 25, istanbul_siege_loss_text1, (15, 10))\n if siege_loss_1_egypt.draw_button():\n current_screen = \"brigand\"\n if siege_loss_1_italy.draw_button():\n current_screen = \"pre-pirate\"\n \n if current_screen == \"bursa_fail\":\n screen.blit(bursa_fail.img, (0,0))\n bursa_fail.add_scenario(Rect(0, 0, 600, 60), 25, bursa_fail_text, (15, 10))\n if bursa_egypt.draw_button():\n current_screen = \"brigand\"\n if bursa_italy.draw_button():\n current_screen = \"pre-pirate\"\n \n if current_screen == \"karaman_room2\":\n screen.blit(karaman_room2.img, (0,0))\n karaman_room2.add_scenario(Rect(0, 0, 600, 60), 25, kr2_text, (15, 10))\n if kr2_humane.draw_button():\n army -= 1\n current_screen = \"constantinople_early\"\n if kr2_pragmatic.draw_button():\n current_screen = \"karaman_room2_2\"\n \n if current_screen == \"karaman_room2_2\":\n screen.blit(karaman_room2_2.img, (0,0))\n karaman_room2_2.add_scenario(Rect(0, 0, 600, 40), 25, kr2_2_text, (15, 10))\n if kr2_2_humane.draw_button():\n current_screen = \"brigand\"\n if kr2_2_pragmatic.draw_button():\n current_screen = \"pre-pirate\"\n \n if current_screen == \"constantinople_early\":\n if dice == None:\n dice = random.random()\n if dice <= 0.1: \n screen.blit(constantinople_hum.img, (0,0))\n constantinople_hum.add_scenario(Rect(0, 0, 600, 80), 25, const_text_good, (15, 10))\n if const_hum_good.draw_button():\n current_screen = \"topkapi1\"\n \n else:\n screen.blit(constantinople_hum2.img, (0,0))\n constantinople_hum2.add_scenario(Rect(0, 0, 600, 60), 25, const_text_bad, (15, 10))\n if const_hum_bad.draw_button():\n dice = random.random()\n if dice <= 0.05:\n current_screen = \"topkapi_escape\"\n else:\n current_screen = \"topkapi_early_death\"\n \n if current_screen == \"topkapi_escape\":\n screen.blit(topkapi_escape.img, (0,0))\n topkapi_escape.add_scenario(Rect(0, 0, 600, 60), 25, topkapi_esc_text, (15, 10))\n if escape_hum.draw_button():\n current_screen = \"brigand\"\n if escape_prag.draw_button():\n current_screen = \"pre-pirate\"\n \n if current_screen == \"topkapi_early_death\":\n screen.blit(topkapi_early_dead.img, (0,0))\n topkapi_early_dead.add_scenario(Rect(0, 0, 600, 90), 25, topkapi_early_dead_text, (15, 10))\n if main_men.draw_button():\n current_screen = \"topkapi1\"\n \n if current_screen == \"brigand\":\n screen.blit(brigands.img, (0,0))\n brigands.add_scenario(Rect(0, 0, 600, 60), 25, brigand_text, (15, 10))\n if kill_brigands.draw_button():\n dice = random.random()\n if dice <= 0.4:\n current_screen = \"brigand_win\"\n else:\n current_screen = \"brigand_lose\"\n if pay_brigands.draw_button():\n dice = random.random()\n if dice <= 0.6:\n current_screen = \"pay_brigand_win\"\n else:\n current_screen = \"pay_brigand_lose\"\n \n if current_screen == \"brigand_win\":\n screen.blit(brigand_win.img, (0,0))\n brigand_win.add_scenario(Rect(0, 0, 600, 60), 25, brigand_win_text, (15, 10))\n continue2 = Button(0, 350, 600, 50, brigand_win.sprite_img.image, 22, \"Continue\") \n if continue2.draw_button():\n current_screen = \"egypt\"\n \n if current_screen == \"brigand_lose\":\n screen.blit(brigand_lose.img, (0,0))\n brigand_lose.add_scenario(Rect(0, 0, 600, 80), 25, brigand_lose_text, (15, 10))\n mainmen3 = Button(0, 350, 600, 50, brigand_lose.sprite_img.image, 22, \"Return to main menu\")\n if mainmen3.draw_button():\n current_screen = \"topkapi1\"\n \n \n if current_screen == \"pay_brigand_win\":\n screen.blit(pay_brigand_win.img, (0,0))\n pay_brigand_win.add_scenario(Rect(0, 0, 600, 60), 25, pay_brigand_win_text, (15, 10))\n continue3 = Button(0, 350, 600, 50, pay_brigand_win.sprite_img.image, 30, \"Continue\") \n if continue3.draw_button():\n current_screen = \"egypt\"\n \n if current_screen == \"pay_brigand_lose\":\n screen.blit(pay_brigand_lose.img, (0,0))\n pay_brigand_lose.add_scenario(Rect(0, 0, 600, 60), 25, pay_brigand_lose_text, (15, 10))\n if escape.draw_button():\n if dice == None:\n dice = random.random()\n if dice <= 0.3:\n escaped = True\n current_screen = \"egypt\"\n else:\n current_screen = \"executed\"\n dice = None\n \n if current_screen == \"executed\":\n screen.blit(executioner.img, (0,0))\n executioner.add_scenario(Rect(0, 0, 600, 80), 25, brigand_ending, (15, 10))\n mainmen4 = Button(0, 350, 600, 50, executioner.sprite_img.image, 22, \"Return to main menu\")\n if mainmen4.draw_button():\n current_screen = \"topkapi1\"\n \n \n if current_screen == \"egypt\":\n screen.blit(mamluks.img, (0,0))\n \n if not escaped:\n mamluks.add_scenario(Rect(0, 0, 600, 60), 25, mamluks_text, (15, 10))\n else:\n mamluks.add_scenario(Rect(0, 0, 600, 60), 25, escaped_text, (15, 10))\n \n if mamluk_traitor.draw_button():\n traitor = True\n if dice == None:\n dice = random.random()\n if dice <= 0.9:\n current_screen = \"egypt_positive\"\n army += 1\n egypt_army = True\n else:\n current_screen = \"egypt_negative\"\n dice = None\n if mamluk_hum.draw_button():\n traitor = False\n if dice == None:\n dice = random.random()\n if dice <= 0.2:\n current_screen = \"egypt_positive\"\n army += 1\n egypt_army = True\n else:\n current_screen = \"egypt_negative\"\n dice = None\n \n if current_screen == \"egypt_positive\":\n screen.blit(mamluks_result.img, (0,0))\n if traitor:\n mamluks_result.add_scenario(Rect(0, 0, 600, 60), 25, mamluk_gtraitor_text, (15, 10))\n else:\n mamluks_result.add_scenario((Rect(0, 0, 600, 60)), 25, mamluk_ghum_text, (15, 10)) \n continue4 = Button(0, 350, 600, 50, mamluks_result.sprite_img.image, 22, \"Great news from Mamluks!\")\n if continue4.draw_button():\n if not pope_army:\n current_screen = \"istanbul_siege_one_army\"\n else:\n current_screen = \"istanbul_siege_two_army\"\n\n if current_screen == \"egypt_negative\":\n screen.blit(mamluk_prisoner.img, (0,0))\n mamluk_prisoner.add_scenario(Rect(0, 0, 600, 70), 25, mamluk_bad_hum_text, (15, 10))\n mainmen5 = Button(0, 350, 600, 50, mamluk_prisoner.sprite_img.image, 22, \"Return to main menu\")\n if mainmen5.draw_button():\n current_screen = \"topkapi1\"\n \n if current_screen == \"pre-pirate\":\n screen.blit(way_to_italy.img, (0,0))\n way_to_italy.add_scenario(Rect(0, 0, 600, 60), 25, way_to_italy_text, (15, 10))\n continue5 = Button(0, 350, 600, 50, way_to_italy.sprite_img.image, 22, \"I am so excited to meet the Pope!\")\n if continue5.draw_button():\n current_screen = \"pirates\"\n \n if current_screen == \"pirates\":\n screen.blit(pirates.img, (0,0))\n pirates.add_scenario(Rect(0, 0, 600, 60), 25, pirate_text, (15, 10))\n if pay_pirates.draw_button():\n dice = random.random()\n if dice <= 0.5:\n current_screen = \"pay_pirate_success\"\n else:\n current_screen = \"pay_pirate_fail\"\n if revolt.draw_button():\n dice = random.random()\n if dice <= 0.5:\n current_screen = \"pirate_revolt_success\"\n else:\n current_screen = \"pirate_revolt_fail\"\n \n if current_screen == \"pay_pirate_success\":\n screen.blit(italy_coast.img, (0,0))\n italy_coast.add_scenario(Rect(0, 0, 600, 60), 25, italy_coast_text, (15, 10))\n if pay_good_news.draw_button():\n current_screen = \"italy\"\n \n if current_screen == \"pay_pirate_fail\":\n screen.blit(sold_to_brother.img, (0,0))\n sold_to_brother.add_scenario(Rect(0, 0, 600, 60), 25, sold_to_brother_text, (15, 10))\n continue6 = Button(0, 350, 600, 50, sold_to_brother.sprite_img.image, 22, \"I don't want to die in a ship!\")\n if continue6.draw_button():\n current_screen = \"executioner2\"\n \n if current_screen == \"executioner2\":\n screen.blit(executioner2.img, (0,0))\n executioner2.add_scenario(Rect(0, 0, 600, 60), 25, sold_to_brother_dead, (15, 10))\n mainmen6 = Button(0, 350, 600, 50, executioner2.sprite_img.image, 22, \"What a tragic ending!\")\n if mainmen6.draw_button():\n current_screen = \"topkapi1\"\n \n if current_screen == \"pirate_revolt_success\":\n screen.blit(pirate_revolt_win.img, (0,0))\n pirate_revolt_win.add_scenario(Rect(0, 0, 600, 60), 25, pirate_revolt_win_text, (15, 10))\n if revolt_good_news.draw_button():\n current_screen = \"italy\"\n \n if current_screen == \"pirate_revolt_fail\":\n screen.blit(pirate_dead.img, (0,0))\n pirate_dead.add_scenario(Rect(0, 0, 600, 60), 25, pirate_dead_text, (15, 10))\n mainmen7 = Button(0, 350, 600, 50, pirate_dead.sprite_img.image, 22, \"Hope I will go to heaven at least!\")\n if mainmen7.draw_button():\n current_screen = \"topkapi1\"\n \n if current_screen == \"italy\":\n screen.blit(pope.img, (0,0))\n pope.add_scenario(Rect(0, 0, 600, 80), 25, pope_text, (15, 10))\n if christian_button.draw_button():\n current_screen = \"christian_army\"\n christian = True\n pope_army = True\n army += 1\n if muslim.draw_button():\n dice = random.random()\n if dice <= 0.5:\n current_screen = \"muslim_good\"\n army += 1\n pope_army = True\n else:\n current_screen = \"muslim_bad\"\n \n if current_screen == \"christian_army\":\n screen.blit(christian_army.img, (0, 0))\n christian_army.add_scenario(Rect(0, 0, 600, 70), 25, christian_army_prag, (15, 10))\n continue7 = Button(0, 350, 600, 50, christian_army.sprite_img.image, 22, \"I will be the next Ottoman emperor!\")\n if continue7.draw_button():\n if egypt_army:\n current_screen = \"istanbul_siege_two_army\"\n else:\n current_screen = \"istanbul_siege_one_army\"\n \n if current_screen == \"muslim_good\":\n screen.blit(christian_army_hum.img, (0,0))\n christian_army_hum.add_scenario(Rect(0, 0, 600, 60), 25, christian_army_hum, (15, 10))\n continue8 = Button(0, 350, 600, 50, christian_army.sprite_img.image, 22, \"I will conquer Constantinople!\")\n if continue8.draw_button():\n if egypt_army:\n current_screen = \"istanbul_siege_two_army\"\n else:\n current_screen = \"istanbul_siege_one_army\"\n \n if current_screen == \"muslim_bad\":\n screen.blit(old_man.img, (0,0))\n old_man.add_scenario(Rect(0, 0, 600, 60), 25, pope_text2, (15, 10))\n mainmen8 = Button(0, 350, 600, 50, old_man.sprite_img.image, 22, \"I lived a decent and honorable life at least.\")\n if mainmen8.draw_button():\n current_screen = \"topkapi1\"\n \n \n if current_screen == \"istanbul_siege_one_army\":\n if egypt_army:\n screen.blit(siege_only_egypt.img, (0,0))\n siege_only_egypt.add_scenario(Rect(0, 0, 600, 60), 25, siege_only_egypt_text, (15, 10))\n if to_italy.draw_button():\n current_screen = \"pre-pirate\"\n if attack_one_army_egypt.draw_button():\n dice = random.random()\n if army == 2:\n ambition_point = 0.1\n else:\n ambition_point = 0\n if dice <= 0.35 + ambition_point:\n current_screen = \"istanbul_win\"\n else:\n current_screen = \"istanbul_loss\"\n if pope_army:\n screen.blit(siege_only_italy.img, (0,0))\n siege_only_italy.add_scenario(Rect(0, 0, 600, 60), 25, siege_only_italy_text, (15, 10))\n if to_egypt.draw_button():\n current_screen = \"brigand\"\n if attack_one_army_italy.draw_button():\n dice = random.random()\n if army == 2:\n ambition_point = 0.1\n else:\n ambition_point = 0\n if dice <= 0.35 + ambition_point:\n current_screen = \"istanbul_win\"\n else:\n current_screen = \"istanbul_loss\"\n \n if current_screen == \"istanbul_siege_two_army\":\n screen.blit(two_army_siege.img, (0,0))\n two_army_siege.add_scenario(Rect(0, 0, 600, 60), 25, two_army_siege_text, (15, 10))\n if two_army_siege_button.draw_button():\n print(army)\n if army == 3:\n dice = random.random()\n if dice <= 0.9:\n current_screen = \"istanbul_win\"\n else:\n current_screen = \"istanbul_loss\"\n if army == 2:\n dice = random.random()\n if dice <= 0.8:\n current_screen = \"istanbul_win\"\n else:\n current_screen = \"istanbul_loss\"\n \n if current_screen == \"istanbul_loss\":\n if egypt_army and pope_army:\n screen.blit(both_loss.img, (0,0))\n both_loss.add_scenario(Rect(0, 0, 600, 60), 25, final_loss_both_army, (15, 10))\n mainmenfinal = Button(0, 350, 600, 50, both_loss.sprite_img.image, 22, \"I did my best to take over the thone.\")\n elif egypt_army:\n screen.blit(final_loss_only_egypt.img, (0,0))\n final_loss_only_egypt.add_scenario(Rect(0, 0, 600, 60), 25, final_loss_egypt_text, (15, 10))\n mainmenfinal = Button(0, 350, 600, 50, final_loss_only_egypt.sprite_img.image, 22, \"I didn't bend the knee at least.\")\n else:\n if christian:\n screen.blit(final_loss_only_italy.img, (0,0))\n final_loss_only_italy.add_scenario(Rect(0, 0, 600, 60), 25, final_loss_italy_text, (15, 10))\n mainmenfinal = Button(0, 350, 600, 50, final_loss_only_italy.sprite_img.image, 22, \"I will pray to Jesus and the Lord.\")\n else:\n screen.blit(final_loss_only_italy_2.img, (0,0))\n final_loss_only_italy_2.add_scenario(Rect(0, 0, 600, 60), 25, final_loss_italy_text2, (15, 10))\n mainmenfinal = Button(0, 350, 600, 50, final_loss_only_italy_2.sprite_img.image, 22, \"I will pray to Allah for the last time.\") \n if mainmenfinal.draw_button():\n current_screen = \"topkapi1\"\n \n if current_screen == \"tyrant_king\":\n screen.blit(tyrant_king.img, (0,0))\n if christian:\n tyrant_king.add_scenario(Rect(0, 0, 600, 70), 25, tyrant_king_text_christ, (15, 10))\n else:\n tyrant_king.add_scenario(Rect(0, 0, 600, 75), 25, tyrant_king_text, (15, 10))\n if tyrant_king_button.draw_button():\n current_screen = \"topkapi1\"\n \n if current_screen == \"good_king\":\n if christian:\n screen.blit(good_king_christ.img, (0, 0))\n good_king_christ.add_scenario(Rect(0, 0, 600, 70), 25, good_king_christ_text, (15, 10))\n if good_king_christian.draw_button():\n current_screen = \"topkapi1\"\n else:\n screen.blit(good_king_screen.img, (0, 0))\n good_king_text = good_king_gen(char_name)\n good_king_screen.add_scenario(Rect(0, 0, 600, 70), 25, good_king_text, (15, 10))\n if good_king.draw_button():\n current_screen = \"topkapi1\"\n \n if leave.draw_button():\n pygame.quit()\n quit()\n \n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n pygame.quit()\n quit() \n if current_screen == \"topkapi2\":\n final_text = topkapi2_box.handle_event(event)\n topkapi2_box.draw_text_box()\n if final_text != None:\n current_screen = \"karaman\"\n char_name = final_text\n print(char_name)\n \n pygame.display.flip()\n pygame.display.update()", "sub_path": "Clash_of_Brothers.py", "file_name": "Clash_of_Brothers.py", "file_ext": "py", "file_size_in_byte": 50612, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pygame.init", "line_number": 13, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 16, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 17, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 17, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 19, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 24, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pygame.transform.scale", "line_number": 25, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 25, "usage_type": "attribute"}, {"api_name": "pygame.sprite.Sprite", "line_number": 27, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 32, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 32, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 33, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 33, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 51, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 51, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pos", "line_number": 59, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 59, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pressed", "line_number": 63, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 63, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 65, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 65, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pressed", "line_number": 66, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 66, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 70, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 70, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 72, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 72, "usage_type": "attribute"}, {"api_name": "pygame.draw.line", "line_number": 74, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 74, "usage_type": "attribute"}, {"api_name": "pygame.draw.line", "line_number": 75, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 75, "usage_type": "attribute"}, {"api_name": "pygame.draw.line", "line_number": 76, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 76, "usage_type": "attribute"}, {"api_name": "pygame.draw.line", "line_number": 77, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 77, "usage_type": "attribute"}, {"api_name": "pygame.MOUSEBUTTONDOWN", "line_number": 102, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 107, "usage_type": "attribute"}, {"api_name": "pygame.K_RETURN", "line_number": 109, "usage_type": "attribute"}, {"api_name": "pygame.K_BACKSPACE", "line_number": 113, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 121, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 121, "usage_type": "attribute"}, {"api_name": "pygame.draw.line", "line_number": 123, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 123, "usage_type": "attribute"}, {"api_name": "pygame.draw.line", "line_number": 124, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 124, "usage_type": "attribute"}, {"api_name": "pygame.draw.line", "line_number": 125, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 125, "usage_type": "attribute"}, {"api_name": "pygame.draw.line", "line_number": 126, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 126, "usage_type": "attribute"}, {"api_name": "pygame.Color", "line_number": 131, "usage_type": "call"}, {"api_name": "pygame.font.SysFont", "line_number": 161, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 161, "usage_type": "attribute"}, {"api_name": "random.random", "line_number": 527, "usage_type": "call"}, {"api_name": "random.random", "line_number": 557, "usage_type": "call"}, {"api_name": "random.random", "line_number": 585, "usage_type": "call"}, {"api_name": "random.random", "line_number": 636, "usage_type": "call"}, {"api_name": "random.random", "line_number": 647, "usage_type": "call"}, {"api_name": "random.random", "line_number": 671, "usage_type": "call"}, {"api_name": "random.random", "line_number": 677, "usage_type": "call"}, {"api_name": "random.random", "line_number": 710, "usage_type": "call"}, {"api_name": "random.random", "line_number": 737, "usage_type": "call"}, {"api_name": "random.random", "line_number": 748, "usage_type": "call"}, {"api_name": "random.random", "line_number": 788, "usage_type": "call"}, {"api_name": "random.random", "line_number": 794, "usage_type": "call"}, {"api_name": "random.random", "line_number": 842, "usage_type": "call"}, {"api_name": "random.random", "line_number": 885, "usage_type": "call"}, {"api_name": "random.random", "line_number": 900, "usage_type": "call"}, {"api_name": "random.random", "line_number": 916, "usage_type": "call"}, {"api_name": "random.random", "line_number": 922, "usage_type": "call"}, {"api_name": "pygame.quit", "line_number": 972, "usage_type": "call"}, {"api_name": "pygame.event.get", "line_number": 975, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 975, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 976, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 977, "usage_type": "call"}, {"api_name": "pygame.display.flip", "line_number": 987, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 987, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 988, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 988, "usage_type": "attribute"}]}
+{"seq_id": "595853893", "text": "from google.appengine.ext import webapp\r\nfrom google.appengine.ext import db\r\nfrom google.appengine.ext.webapp import util\r\nfrom google.appengine.ext.webapp import template\r\nfrom google.appengine.ext.db import djangoforms\r\nfrom google.appengine.ext.mapreduce.control import start_map\r\nfrom django import forms\r\n\r\nimport os\r\n\r\nfrom base import BaseHandler, QuizHandler\r\nimport links\r\nfrom models import Quiz, Question, Score\r\n\r\nclass QuizForm(djangoforms.ModelForm):\r\n class Meta:\r\n model = Quiz\r\n exclude = ['updated']\r\n \r\n attempts = forms.IntegerField(\r\n min_value = 1,\r\n widget = forms.TextInput(attrs={'class': 'mini'})\r\n )\r\n \r\n title = forms.CharField(\r\n widget = forms.TextInput()\r\n )\r\n \r\nclass AddHandler(QuizHandler): \r\n def get(self):\r\n self.values['form'] = QuizForm()\r\n self.values['action'] = links.Quiz.add()\r\n self.output('quiz_edit.html')\r\n \r\n def post(self):\r\n form = QuizForm(data=self.request.POST)\r\n self.values['form'] = form\r\n self.values['action'] = links.Quiz.add()\r\n \r\n if (form.is_valid()):\r\n quiz_entity = form.save()\r\n self.redirect(links.Quiz.edit(quiz_entity.key().id()))\r\n else:\r\n self.output('quiz_edit.html')\r\n\r\nclass EditHandler(QuizHandler):\r\n values = {\r\n 'saved': True\r\n }\r\n \r\n def get(self, id):\r\n self.fetch(id)\r\n self.values['form'] = QuizForm(instance=self.quiz_entity)\r\n self.values['action'] = links.Quiz.edit(int(id))\r\n self.values['questions'] = Question.get_for_quiz(self.quiz_entity)\r\n self.output('quiz_edit.html')\r\n \r\n def post(self, id):\r\n self.fetch(id)\r\n form = QuizForm(data=self.request.POST, instance=self.quiz_entity)\r\n self.values['action'] = links.Quiz.edit(int(id))\r\n self.values['form'] = form\r\n \r\n if form.is_valid():\r\n quiz_entity = form.save(commit=False)\r\n quiz_entity.version += 1\r\n quiz_entity.put()\r\n \r\n self.output('quiz_edit.html')\r\n\r\nclass DeleteHandler(QuizHandler):\r\n def get(self, id):\r\n self.fetch(id)\r\n self.values['action'] = links.Quiz.delete(int(id))\r\n self.values['prompt'] = \\\r\n 'Are you absolutely sure that you would like to delete the ' \\\r\n 'quiz, \"%s,\" forever?' % self.quiz_entity.title \r\n self.values['back'] = os.environ.get('HTTP_REFERER',\r\n links.Quiz.edit(int(id)))\r\n self.output('confirm.html')\r\n \r\n def post(self, id): \r\n self.fetch(id)\r\n self.quiz_entity.is_deleting = True\r\n self.quiz_entity.put()\r\n\r\n for model in ['Question', 'Attempt', 'Score', 'Snapshot']:\r\n start_map('Delete %s(s)' % model,\r\n 'jobs.cleanup.delete',\r\n 'google.appengine.ext.mapreduce.input_readers.DatastoreInputReader',\r\n {\r\n 'entity_kind': 'models.%s' % model,\r\n 'quiz_id': int(id)\r\n });\r\n\r\n self.quiz_entity.delete()\r\n self.redirect('/')\r\n\r\nclass ArchiveHandler(QuizHandler):\r\n def get(self, id):\r\n self.fetch(id)\r\n self.values['action'] = links.Quiz.archive(int(id))\r\n self.values['prompt'] = \\\r\n 'Are you sure that you would like archive all scores for the ' \\\r\n 'quiz, \"%s,\" forever?' % self.quiz_entity.title \r\n self.values['back'] = os.environ.get('HTTP_REFERER',\r\n links.Quiz.roster(int(id)))\r\n self.output('confirm.html')\r\n\r\n def post(self, id):\r\n for model in ['Score', 'Attempt']:\r\n start_map('Archive %s(s)' % model,\r\n 'jobs.cleanup.archive',\r\n 'google.appengine.ext.mapreduce.input_readers.DatastoreInputReader',\r\n {\r\n 'entity_kind': 'models.%s' % model,\r\n 'quiz_id': int(id)\r\n });\r\n\r\n self.redirect(links.Quiz.roster(int(id)))\r\n\r\nclass RosterHandler(QuizHandler):\r\n def get(self, id):\r\n self.fetch(id)\r\n\r\n scores_query = Score.all()\r\n scores_query.filter('is_archived =', False)\r\n scores_query.filter('quiz =', self.quiz_entity)\r\n scores_query.order('last_name')\r\n\r\n self.values['link_archived'] = links.Quiz.archived(int(id))\r\n self.values['link_archive'] = links.Quiz.archive(int(id))\r\n self.values['scores'] = scores_query.fetch(100)\r\n\r\n self.output('quiz_roster.html')\r\n\r\nclass ArchivedHandler(QuizHandler):\r\n def get(self, id):\r\n self.fetch(id)\r\n\r\n scores_query = Score.all()\r\n scores_query.filter('is_archived =', True)\r\n scores_query.filter('quiz =', self.quiz_entity)\r\n scores_query.order('updated')\r\n scores_query.order('last_name')\r\n\r\n self.values['link_back'] = links.Quiz.roster(int(id))\r\n self.values['scores'] = scores_query.fetch(100)\r\n\r\n self.output('quiz_archived.html')\r\n\r\ndef main():\r\n application = webapp.WSGIApplication([\r\n ('/quiz/add', AddHandler),\r\n (r'/quiz/edit/(.*)', EditHandler),\r\n #(r'/quiz/delete/(.*)', DeleteHandler),\r\n (r'/quiz/archive/(.*)', ArchiveHandler),\r\n (r'/quiz/roster/(.*)', RosterHandler),\r\n (r'/quiz/archived/(.*)', ArchivedHandler)\r\n ], debug=True)\r\n util.run_wsgi_app(application)\r\n\r\nif __name__ == '__main__':\r\n main()", "sub_path": "handlers/quiz/edit.py", "file_name": "edit.py", "file_ext": "py", "file_size_in_byte": 5613, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "google.appengine.ext.db.djangoforms.ModelForm", "line_number": 15, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.db.djangoforms", "line_number": 15, "usage_type": "name"}, {"api_name": "models.Quiz", "line_number": 17, "usage_type": "name"}, {"api_name": "django.forms.IntegerField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 20, "usage_type": "name"}, {"api_name": "django.forms.TextInput", "line_number": 22, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 22, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 25, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 25, "usage_type": "name"}, {"api_name": "django.forms.TextInput", "line_number": 26, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 26, "usage_type": "name"}, {"api_name": "base.QuizHandler", "line_number": 29, "usage_type": "name"}, {"api_name": "links.Quiz.add", "line_number": 32, "usage_type": "call"}, {"api_name": "links.Quiz", "line_number": 32, "usage_type": "attribute"}, {"api_name": "links.Quiz.add", "line_number": 38, "usage_type": "call"}, {"api_name": "links.Quiz", "line_number": 38, "usage_type": "attribute"}, {"api_name": "links.Quiz.edit", "line_number": 42, "usage_type": "call"}, {"api_name": "links.Quiz", "line_number": 42, "usage_type": "attribute"}, {"api_name": "base.QuizHandler", "line_number": 46, "usage_type": "name"}, {"api_name": "links.Quiz.edit", "line_number": 54, "usage_type": "call"}, {"api_name": "links.Quiz", "line_number": 54, "usage_type": "attribute"}, {"api_name": "models.Question.get_for_quiz", "line_number": 55, "usage_type": "call"}, {"api_name": "models.Question", "line_number": 55, "usage_type": "name"}, {"api_name": "links.Quiz.edit", "line_number": 61, "usage_type": "call"}, {"api_name": "links.Quiz", "line_number": 61, "usage_type": "attribute"}, {"api_name": "base.QuizHandler", "line_number": 71, "usage_type": "name"}, {"api_name": "links.Quiz.delete", "line_number": 74, "usage_type": "call"}, {"api_name": "links.Quiz", "line_number": 74, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 78, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 78, "usage_type": "attribute"}, {"api_name": "links.Quiz.edit", "line_number": 79, "usage_type": "call"}, {"api_name": "links.Quiz", "line_number": 79, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.mapreduce.control.start_map", "line_number": 88, "usage_type": "call"}, {"api_name": "base.QuizHandler", "line_number": 99, "usage_type": "name"}, {"api_name": "links.Quiz.archive", "line_number": 102, "usage_type": "call"}, {"api_name": "links.Quiz", "line_number": 102, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 106, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 106, "usage_type": "attribute"}, {"api_name": "links.Quiz.roster", "line_number": 107, "usage_type": "call"}, {"api_name": "links.Quiz", "line_number": 107, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.mapreduce.control.start_map", "line_number": 112, "usage_type": "call"}, {"api_name": "links.Quiz.roster", "line_number": 120, "usage_type": "call"}, {"api_name": "links.Quiz", "line_number": 120, "usage_type": "attribute"}, {"api_name": "base.QuizHandler", "line_number": 122, "usage_type": "name"}, {"api_name": "models.Score.all", "line_number": 126, "usage_type": "call"}, {"api_name": "models.Score", "line_number": 126, "usage_type": "name"}, {"api_name": "links.Quiz.archived", "line_number": 131, "usage_type": "call"}, {"api_name": "links.Quiz", "line_number": 131, "usage_type": "attribute"}, {"api_name": "links.Quiz.archive", "line_number": 132, "usage_type": "call"}, {"api_name": "links.Quiz", "line_number": 132, "usage_type": "attribute"}, {"api_name": "base.QuizHandler", "line_number": 137, "usage_type": "name"}, {"api_name": "models.Score.all", "line_number": 141, "usage_type": "call"}, {"api_name": "models.Score", "line_number": 141, "usage_type": "name"}, {"api_name": "links.Quiz.roster", "line_number": 147, "usage_type": "call"}, {"api_name": "links.Quiz", "line_number": 147, "usage_type": "attribute"}, {"api_name": "google.appengine.ext.webapp.WSGIApplication", "line_number": 153, "usage_type": "call"}, {"api_name": "google.appengine.ext.webapp", "line_number": 153, "usage_type": "name"}, {"api_name": "google.appengine.ext.webapp.util.run_wsgi_app", "line_number": 161, "usage_type": "call"}, {"api_name": "google.appengine.ext.webapp.util", "line_number": 161, "usage_type": "name"}]}
+{"seq_id": "460591662", "text": "from tri import *\nimport time as t\nimport matplotlib.pyplot as plt\n\n\ndef chargeClassement(fichier):\n f=open(fichier,'r',encoding='utf8')\n fichier=f.readlines()\n f.close() #ne pas oublier de fermer le fichier...\n L=[]\n for ligne in fichier:\n ligne=ligne.split('\\t') # coupe aux tabulations\n L1=[]\n L1.append(ligne[1])\n L1.append(ligne[2])\n L1.append(ligne[4])\n L.append(L1)\n return L\n\n\ndef convertirTemps(temps:str):\n '''temps est un str de la forme \"06h 09' 39''\" '''\n t_course=temps.split('h') #on peut aussi couper à h\n #print (t_course)\n heure=int(t_course[0])\n t_course2=t_course[1].split(\"'\")\n #print (t_course2)\n duree=int(t_course2[1])+60*int(t_course2[0])+3600*heure\n return duree\n\n\n\ndef classement(fichier):\n L=chargeClassement(fichier)\n for element in L:\n element[2]=convertirTemps(element[2])\n return L\n\nLG=classement('classement_general.txt')\nL18=classement('etape_18.txt')\n\n\n#Q4\ndef ajoutTemps(liste1:list,liste2:list)->list:\n LGN=[]\n for x in liste1:\n for y in liste2:\n if x[1]==y[1]:\n LGN+=[[x[0],x[1],y[2]+x[2]]]\n return LGN\n\n#Q5\n####Tri par insertion modifié\ndef tri_insertion_modifie(T):\n n=len(T)\n for i in range(1,n):\n j=i\n v=T[i][-1]\n v2=T[i]\n while j>0 and v=d-1:\n return\n else:\n m,a=partition_modifie(a,g,d)\n tri_rapide_modifie(a,g,m)\n tri_rapide_modifie(a,m+1,d)\n\n####Tri par fusion modifié\ndef fusion_modifie(a0,a,g,m,d):\n i,j=g,m\n for k in range(g,d):\n if i=d-1:\n return\n else:\n m=(g+d)//2\n tri_fusion_modifie(a,g,m)\n tri_fusion_modifie(a,m,d)\n a0[g:d]=a[g:d]\n fusion_modifie(a0,a,g,m,d)\n\n\n\n\n\ndef test_tri_modifie1():\n LGN=ajoutTemps(LG,L18)\n LGN0=LGN[:]\n LGN=sorted(LGN,key=lambda colonnes:colonnes[2])\n LGN1=tri_insertion_modifie(LGN0)\n print(LGN[0:6],'\\n\\n',LGN1[0:6])\n assert LGN==LGN1\n\n\ndef test_tri_modifie2():\n LGN=ajoutTemps(LG,L18)\n LGN1=LGN[:]\n LGN=sorted(LGN,key=lambda colonnes:colonnes[2])\n tri_fusion_modifie(LGN1,0,len(LGN1))\n print(LGN[0:6],'\\n\\n',LGN1[0:6])\n assert LGN==LGN1\n\ndef test_tri_modifie3():\n LGN=ajoutTemps(LG,L18)\n LGN1=LGN[:]\n LGN=sorted(LGN,key=lambda colonnes:colonnes[2])\n tri_rapide_modifie(LGN1,0,len(LGN1))\n print(LGN[0:6],'\\n\\n',LGN1[0:6],'\\n\\n',ajoutTemps(LG,L18)[0:6])\n assert LGN==LGN1\n\n\n\n\ndef update_classement_general(liste1:list,liste2:list)->list:\n LGN=ajoutTemps(LG,L18)\n sorted(LGN,key=lambda colonnes:colonnes[2])\n return LGN\n\n\n\n\n\n\n#LGN=update_classement_general(LG,L18)\n\n\n####Comparaison des temps de calcul\ndef comparer_tri(LGN,nom_de_fichier):\n temps_insertion=[]\n temps_rapide=[]\n temps_fusion=[]\n temps_sorted=[]\n liste_k=[]\n for k in range(2,len(LGN),100):\n liste_k.append(k)\n LGT1=LGN[:k+1]\n LGT2=LGN[:k+1]\n LGT3=LGN[:k+1]\n LGT4=LGN[:k+1]\n tic=t.time()\n tri_fusion(LGT1,0,len(LGT1))\n toc=t.time()\n temps_fusion.append(toc-tic)\n tic=t.time()\n tri_insertion(LGT2)\n toc=t.time()\n temps_insertion.append(toc-tic)\n tic=t.time()\n tri_rapide(LGT3,0,len(LGT3))\n toc=t.time()\n temps_rapide.append(toc-tic)\n tic=t.time()\n sorted(LGT4)\n toc=t.time()\n temps_sorted.append(toc-tic)\n\n plt.clf()\n #pdb.set_trace()\n plt.plot(liste_k,temps_fusion,'g-',label='fusion')\n plt.plot(liste_k,temps_rapide,'b*-',label='rapide')\n plt.plot(liste_k,temps_insertion,'r--',label='insertion')\n plt.plot(liste_k,temps_sorted,'k.-',label='sorted')\n plt.legend()\n plt.savefig(nom_de_fichier)\n\n\n\ndef comparer_tri_liste(LGN,nom_de_fichier):\n temps_insertion=[]\n temps_rapide=[]\n temps_fusion=[]\n temps_sorted=[]\n for k in range(2,len(LGN)):\n LGT1=LGN[:k+1]\n LGT2=LGN[:k+1]\n LGT3=LGN[:k+1]\n LGT4=LGN[:k+1]\n tic=t.time()\n tri_fusion_modifie(LGT1,0,len(LGT1))\n toc=t.time()\n temps_fusion.append(toc-tic)\n tic=t.time()\n tri_insertion_modifie(LGT2)\n toc=t.time()\n temps_insertion.append(toc-tic)\n tic=t.time()\n tri_rapide_modifie(LGT3,0,len(LGT3))\n toc=t.time()\n temps_rapide.append(toc-tic)\n tic=t.time()\n sorted(LGT4,key=lambda colonnes:colonnes[2])\n toc=t.time()\n temps_sorted.append(toc-tic)\n\n plt.clf()\n plt.plot(list(range(2,len(LGN))),temps_fusion,'g-',label='fusion')\n plt.plot(list(range(2,len(LGN))),temps_rapide,'b*-',label='rapide')\n plt.plot(list(range(2,len(LGN))),temps_insertion,'r--',label='insertion')\n plt.plot(list(range(2,len(LGN))),temps_sorted,'k.-',label='sorted')\n plt.legend()\n plt.savefig(nom_de_fichier)\n\nLGN=update_classement_general(LG,L18)\ncomparer_tri_liste(LGN,'tp09_durif_compare_tri1.png')\n\n\nimport random\ndef generer_liste(n,N):\n x = [random.randint(0, n) for p in range(0, N)]\n return(x)\n\nN=int(1e3)\n# for k in range(1,N,100):\n# L=generer_liste(N,N)\n\nL=generer_liste(N,N)\ncomparer_tri(L,'tp09_durif_compare_tri2.png')\n\n\n\n\n\n\n", "sub_path": "Exercices/S1_08_Tris/08_TourDeFrance/programmes/tri_emilien.py", "file_name": "tri_emilien.py", "file_ext": "py", "file_size_in_byte": 5829, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "time.time", "line_number": 173, "usage_type": "call"}, {"api_name": "time.time", "line_number": 175, "usage_type": "call"}, {"api_name": "time.time", "line_number": 177, "usage_type": "call"}, {"api_name": "time.time", "line_number": 179, "usage_type": "call"}, {"api_name": "time.time", "line_number": 181, "usage_type": "call"}, {"api_name": "time.time", "line_number": 183, "usage_type": "call"}, {"api_name": "time.time", "line_number": 185, "usage_type": "call"}, {"api_name": "time.time", "line_number": 187, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 190, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 190, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 192, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 192, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 193, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 193, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 194, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 194, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 195, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 195, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 196, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 196, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 197, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 197, "usage_type": "name"}, {"api_name": "time.time", "line_number": 211, "usage_type": "call"}, {"api_name": "time.time", "line_number": 213, "usage_type": "call"}, {"api_name": "time.time", "line_number": 215, "usage_type": "call"}, {"api_name": "time.time", "line_number": 217, "usage_type": "call"}, {"api_name": "time.time", "line_number": 219, "usage_type": "call"}, {"api_name": "time.time", "line_number": 221, "usage_type": "call"}, {"api_name": "time.time", "line_number": 223, "usage_type": "call"}, {"api_name": "time.time", "line_number": 225, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.clf", "line_number": 228, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 228, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 229, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 229, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 230, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 230, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 231, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 231, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 232, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 232, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 233, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 233, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 234, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 234, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 242, "usage_type": "call"}]}
+{"seq_id": "535532391", "text": "# -*- coding: utf-8 -*-\n# Copyright (c) 2010-2012 OpenStack, LLC.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or\n# implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\nimport os.path\nfrom eventlet import Timeout\nfrom webob import Request,Response\nfrom webob.exc import HTTPServerError,HTTPNoContent\nimport uuid\nimport time\nfrom swift.common.bufferedhttp import jresponse\nfrom swift.common.utils import get_logger,split_path,json\n\nfrom swift.common.middleware.userdb import db_insert,db_update,db_delete,db_values,task_db_delete,task_db_values\n\nclass UserOpMiddleware(object):\n\n def __init__(self, app, conf):\n self.app = app\n self.dbdir = conf.get('devices', '/mnt/cloudfs-object').strip()\n self.logger = get_logger(conf, log_route='catch-errors')\n\n def __call__(self, env, start_response):\n \n if 'swift.trans_id' not in env:\n tx_id = 'tx'+uuid.uuid4().hex\n env['swift.trans_id'] = tx_id\n\n new_env = env.copy() \n req = Request(new_env)\n vers,account,container,obj = split_path(req.path,1, 4,True)\n \n dbpath = '%s/%s.db' % (self.dbdir,account)\n if 'register' != container and not os.path.exists(dbpath):\n return jresponse('-1','user db file not found',req,404)(env,start_response)\n if 'GET_OP_HISTORY' == req.GET.get('op'):\n if req.GET.get('recent'):\n desc_flag = True\n limit = int(req.GET.get('recent')) \n data = db_values(dbpath,desc_flag,limit)\n else:\n data = db_values(dbpath)\n \n op_list = json.dumps(data)\n return Response(body=op_list, request=req)(env,start_response)\n \n elif 'DELETE_HISTORY' == req.GET.get('op'):\n \n if req.GET.get('recent'):\n desc_flag = True\n limit = int(req.GET.get('recent'))\n db_delete(dbpath,desc_flag,limit) \n else:\n db_delete(dbpath)\n return jresponse('0','',req,204)(env,start_response)\n \n elif 'GET_OP_TASK' == req.GET.get('op'):\n if req.GET.get('tx_id'):\n \n tx_id = req.GET.get('tx_id') \n data = task_db_values(dbpath,tx_id)\n if len(data) >0:\n op_list = json.dumps(data)\n if data[0][1] == 'success':\n return jresponse('0','success',req,201)(env,start_response)\n else:\n return jresponse('-1',data[0][2],req,400)(env,start_response)\n else: \n op_list = ''\n return jresponse('-1','task not found',req,400)(env,start_response)\n else:\n return jresponse('-1','error params',req,400)(env,start_response)\n \n if 'register' != container:\n path = ''\n type = ''\n if account and container and obj:\n path = obj\n type = 'object'\n elif account and container:\n path = container\n type = 'container'\n elif account:\n path = account\n type = 'account'\n \n method = req.method\n tenant = account\n swifttime = str(time.time())\n tx_id = req.environ.get('swift.trans_id')\n url = req.url\n qs = req.environ.get('QUERY_STRING','')\n env['fwuser_info'] = {} \n env['fwuser_info']['status'] = ''\n env['fwuser_info']['comment'] = ''\n db_insert(dbpath, tx_id, path, type,method, tenant, qs, swifttime, status='', comment='')\n\n resp = self.app(env, start_response)\n\n if 'register' != container:\n if env.get('fwuser_info'):\n estatus = env.get('fwuser_info').get('status','')\n comment = env['fwuser_info'].get('comment','')\n db_update(dbpath, estatus, comment, tx_id)\n \n return resp\n\ndef filter_factory(global_conf, **local_conf):\n conf = global_conf.copy()\n conf.update(local_conf)\n\n def userop_filter(app):\n return UserOpMiddleware(app, conf)\n return userop_filter\n", "sub_path": "swift/common/middleware/userop.py", "file_name": "userop.py", "file_ext": "py", "file_size_in_byte": 4740, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "swift.common.utils.get_logger", "line_number": 32, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 37, "usage_type": "call"}, {"api_name": "webob.Request", "line_number": 41, "usage_type": "call"}, {"api_name": "swift.common.utils.split_path", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path.path.exists", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 45, "usage_type": "name"}, {"api_name": "swift.common.bufferedhttp.jresponse", "line_number": 46, "usage_type": "call"}, {"api_name": "swift.common.middleware.userdb.db_values", "line_number": 51, "usage_type": "call"}, {"api_name": "swift.common.middleware.userdb.db_values", "line_number": 53, "usage_type": "call"}, {"api_name": "swift.common.utils.json.dumps", "line_number": 55, "usage_type": "call"}, {"api_name": "swift.common.utils.json", "line_number": 55, "usage_type": "name"}, {"api_name": "webob.Response", "line_number": 56, "usage_type": "call"}, {"api_name": "swift.common.middleware.userdb.db_delete", "line_number": 63, "usage_type": "call"}, {"api_name": "swift.common.middleware.userdb.db_delete", "line_number": 65, "usage_type": "call"}, {"api_name": "swift.common.bufferedhttp.jresponse", "line_number": 66, "usage_type": "call"}, {"api_name": "swift.common.middleware.userdb.task_db_values", "line_number": 72, "usage_type": "call"}, {"api_name": "swift.common.utils.json.dumps", "line_number": 74, "usage_type": "call"}, {"api_name": "swift.common.utils.json", "line_number": 74, "usage_type": "name"}, {"api_name": "swift.common.bufferedhttp.jresponse", "line_number": 76, "usage_type": "call"}, {"api_name": "swift.common.bufferedhttp.jresponse", "line_number": 78, "usage_type": "call"}, {"api_name": "swift.common.bufferedhttp.jresponse", "line_number": 81, "usage_type": "call"}, {"api_name": "swift.common.bufferedhttp.jresponse", "line_number": 83, "usage_type": "call"}, {"api_name": "time.time", "line_number": 100, "usage_type": "call"}, {"api_name": "swift.common.middleware.userdb.db_insert", "line_number": 107, "usage_type": "call"}, {"api_name": "swift.common.middleware.userdb.db_update", "line_number": 115, "usage_type": "call"}]}
+{"seq_id": "495604869", "text": "import logging\n\nfrom graph.proto import *\nfrom tile.tui import *\nfrom tile.grid import Grid\nimport argparse\n\nlogger = logging.getLogger()\nlogname = 'gameplay.log'\nlogging.basicConfig(filename=logname,\n filemode='a',\n format='%(asctime)s,%(msecs)d %(module)s %(levelno)s %(message)s',\n datefmt='%H:%M:%S',\n level=logging.DEBUG)\n\ndef main():\n # Run the prototype code for graph representation\n # proto()\n\n parser = argparse.ArgumentParser(description='A Turing complete... Thing.')\n parser.add_argument('-x', '--width', dest='width', metavar='N', type=int, default=10,\n help='Width of the grid')\n parser.add_argument('-y', '--height', dest='height', metavar='N', type=int, default=10,\n help='Height of the grid')\n box_parser = parser.add_mutually_exclusive_group(required=False)\n box_parser.add_argument('--box-draw', dest='box_draw', action='store_true',\n help='Enable UTF-8 box drawing characters')\n box_parser.add_argument('--no-box-draw', dest='box_draw', action='store_false',\n help='Disable UTF-8 box drawing characters')\n box_parser.set_defaults(box_draw=True)\n args = parser.parse_args()\n\n logger.debug(args)\n\n # Create a grid\n g = Grid(args.width, args.height)\n\n # Start up the UI\n t = TermUI(args, g)\n # Block until we quit the UI\n t.start()\n\nif __name__ == '__main__':\n main()\n", "sub_path": "src/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1516, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.getLogger", "line_number": 8, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 10, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 14, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 20, "usage_type": "call"}, {"api_name": "tile.grid.Grid", "line_number": 36, "usage_type": "call"}]}
+{"seq_id": "114852001", "text": "from __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nimport os\nimport logging\nfrom math import ceil\nimport sys\n\nimport numpy as np\nimport tensorflow as tf\n\nVGG_MEAN = [103.939, 116.779, 123.68]\n\n\nclass FCN8VGG:\n\n def __init__(self, vgg16_npy_path=None):\n if vgg16_npy_path is None:\n path = sys.modules[self.__class__.__module__].__file__\n # print path\n path = os.path.abspath(os.path.join(path, os.pardir))\n # print path\n path = os.path.join(path, \"vgg16.npy\")\n vgg16_npy_path = path\n logging.info(\"Load npy file from '%s'.\", vgg16_npy_path)\n # load the vgg16 parameters\n if not os.path.isfile(vgg16_npy_path):\n logging.error((\"File '%s' not found. Download it from \"\n \"ftp://mi.eng.cam.ac.uk/pub/mttt2/\"\n \"models/vgg16.npy\"), vgg16_npy_path)\n sys.exit(1)\n\n self.data_dict = np.load(vgg16_npy_path, encoding='latin1').item()\n # data_dict structure:\n # keys: 'conv5_1', 'fc6', 'conv5_3', 'conv5_2', 'fc8', 'fc7', 'conv4_1',\n # 'conv4_2', 'conv4_3', 'conv3_3', 'conv3_2', 'conv3_1', 'conv1_1', 'conv1_2', \n # 'conv2_2', 'conv2_1'] \n # each value is a list, [0] is the W, [1] is the bias\n\n self.wd = 5e-4\n print(\"npy file loaded\")\n\n def build(self, rgbm, train=False, num_classes=2, random_init_fc8=False,\n debug=False):\n \"\"\"\n Build the VGG model using loaded weights\n Parameters\n ----------\n rgbm: image batch tensor\n Image in rgb shap. Scaled to Intervall [0, 255]\n the last channel is the mask channel, we limit them to have only 0 or 255 value, 255 \n means the foreground and the 0 means the background\n train: bool\n Whether to build train or inference graph\n num_classes: int\n How many classes should be predicted (by fc8)\n random_init_fc8 : bool\n Whether to initialize fc8 layer randomly.\n Finetuning is required in this case.\n debug: bool\n Whether to print additional Debug Information.\n \"\"\"\n # Convert RGB to BGR\n\n with tf.name_scope('Processing'):\n\n red, green, blue, mask = tf.split(rgbm, 4, 3)\n #red, green, blue = tf.split(rgb, 3, 3)\n # assert red.get_shape().as_list()[1:] == [224, 224, 1]\n # assert green.get_shape().as_list()[1:] == [224, 224, 1]\n # assert blue.get_shape().as_list()[1:] == [224, 224, 1]\n # concatenate the fourth dimension, which is the color layer\n # the first dimension is the image dimension (different images)\n bgrm = tf.concat([\n blue - VGG_MEAN[0],\n green - VGG_MEAN[1],\n red - VGG_MEAN[2],\n mask\n ], 3)\n # now bgr is: [batch, in_height, in_width, in_channels]\n\n if debug:\n # print all input dimension size, summarize = 4\n bgrm = tf.Print(bgrm, [tf.shape(bgrm)],\n message='Shape of input image: ',\n summarize=4, first_n=1)\n\n self.conv1_1 = self._conv_layer(bgrm, \"conv1_1\")\n self.conv1_2 = self._conv_layer(self.conv1_1, \"conv1_2\")\n self.pool1 = self._max_pool(self.conv1_2, 'pool1', debug)\n\n self.conv2_1 = self._conv_layer(self.pool1, \"conv2_1\")\n self.conv2_2 = self._conv_layer(self.conv2_1, \"conv2_2\")\n self.pool2 = self._max_pool(self.conv2_2, 'pool2', debug)\n\n self.conv3_1 = self._conv_layer(self.pool2, \"conv3_1\")\n self.conv3_2 = self._conv_layer(self.conv3_1, \"conv3_2\")\n self.conv3_3 = self._conv_layer(self.conv3_2, \"conv3_3\")\n self.pool3 = self._max_pool(self.conv3_3, 'pool3', debug)\n\n self.conv4_1 = self._conv_layer(self.pool3, \"conv4_1\")\n self.conv4_2 = self._conv_layer(self.conv4_1, \"conv4_2\")\n self.conv4_3 = self._conv_layer(self.conv4_2, \"conv4_3\")\n self.pool4 = self._max_pool(self.conv4_3, 'pool4', debug)\n\n self.conv5_1 = self._conv_layer(self.pool4, \"conv5_1\")\n self.conv5_2 = self._conv_layer(self.conv5_1, \"conv5_2\")\n self.conv5_3 = self._conv_layer(self.conv5_2, \"conv5_3\")\n self.pool5 = self._max_pool(self.conv5_3, 'pool5', debug)\n\n self.fc6 = self._fc_layer(self.pool5, \"fc6\")\n\n if train:\n self.fc6 = tf.nn.dropout(self.fc6, 0.5)\n\n self.fc7 = self._fc_layer(self.fc6, \"fc7\")\n if train:\n self.fc7 = tf.nn.dropout(self.fc7, 0.5)\n\n if random_init_fc8:\n self.score_fr = self._score_layer(self.fc7, \"score_fr\",\n num_classes)\n else:\n self.score_fr = self._fc_layer(self.fc7, \"score_fr\",\n num_classes=num_classes,\n relu=False)\n\n self.pred = tf.argmax(self.score_fr, dimension=3)\n\n self.upscore2 = self._upscore_layer(self.score_fr,\n shape=tf.shape(self.pool4),\n num_classes=num_classes,\n debug=debug, name='upscore2',\n ksize=4, stride=2)\n self.score_pool4 = self._score_layer(self.pool4, \"score_pool4\",\n num_classes=num_classes)\n self.fuse_pool4 = tf.add(self.upscore2, self.score_pool4)\n\n self.upscore4 = self._upscore_layer(self.fuse_pool4,\n shape=tf.shape(self.pool3),\n num_classes=num_classes,\n debug=debug, name='upscore4',\n ksize=4, stride=2)\n self.score_pool3 = self._score_layer(self.pool3, \"score_pool3\",\n num_classes=num_classes)\n self.fuse_pool3 = tf.add(self.upscore4, self.score_pool3)\n\n self.upscore32 = self._upscore_layer(self.fuse_pool3,\n shape=tf.shape(bgrm),\n num_classes=num_classes,\n debug=debug, name='upscore32',\n ksize=16, stride=8)\n\n self.pred_up = tf.argmax(self.upscore32, dimension=3)\n if debug:\n tf.Print(self.pred_up, [tf.shape(self.pred_up)],\n message='Shape of output image: ',\n summarize=4, first_n=1)\n\n\n\n\n def _max_pool(self, bottom, name, debug):\n\n pool = tf.nn.max_pool(bottom, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1],\n padding='SAME', name=name)\n\n if debug:\n # tf.Print returns same tensor as input pool\n pool = tf.Print(pool, [tf.shape(pool)],\n message='Shape of %s' % name,\n summarize=4, first_n=1)\n return pool\n\n def _conv_layer(self, bottom, name):\n with tf.variable_scope(name) as scope:\n if (name == \"conv1_1\"):\n filt_rgb = self.get_conv_filter(name)\n filt_mask = tf.Variable(tf.truncated_normal([3, 3, 1, 64], dtype=tf.float32,\n stddev=1e-1))\n filt = tf.concat([filt_rgb,filt_mask],2)\n else:\n filt = self.get_conv_filter(name)\n conv = tf.nn.conv2d(bottom, filt, [1, 1, 1, 1], padding='SAME')\n\n conv_biases = self.get_bias(name)\n bias = tf.nn.bias_add(conv, conv_biases)\n\n relu = tf.nn.relu(bias)\n # Add summary to Tensorboard\n _activation_summary(relu)\n return relu\n\n def _fc_layer(self, bottom, name, num_classes=None,\n relu=True, debug=False):\n with tf.variable_scope(name) as scope:\n shape = bottom.get_shape().as_list()\n\n if name == 'fc6':\n filt = self.get_fc_weight_reshape(name, [7, 7, 512, 4096])\n elif name == 'score_fr':\n name = 'fc8' # Name of score_fr layer in VGG Model\n filt = self.get_fc_weight_reshape(name, [1, 1, 4096, 1000],\n num_classes=num_classes)\n else:\n filt = self.get_fc_weight_reshape(name, [1, 1, 4096, 4096])\n\n self._add_wd_and_summary(filt, self.wd, \"fc_wlosses\")\n\n conv = tf.nn.conv2d(bottom, filt, [1, 1, 1, 1], padding='SAME')\n conv_biases = self.get_bias(name, num_classes=num_classes)\n bias = tf.nn.bias_add(conv, conv_biases)\n\n if relu:\n bias = tf.nn.relu(bias)\n _activation_summary(bias)\n\n if debug:\n bias = tf.Print(bias, [tf.shape(bias)],\n message='Shape of %s' % name,\n summarize=4, first_n=1)\n return bias\n\n def _score_layer(self, bottom, name, num_classes):\n with tf.variable_scope(name) as scope:\n # get number of input channels\n in_features = bottom.get_shape()[3].value\n shape = [1, 1, in_features, num_classes]\n # He initialization Sheme\n if name == \"score_fr\":\n num_input = in_features\n stddev = (2 / num_input)**0.5\n elif name == \"score_pool4\":\n stddev = 0.001\n elif name == \"score_pool3\":\n stddev = 0.0001\n # Apply convolution\n w_decay = self.wd\n\n weights = self._variable_with_weight_decay(shape, stddev, w_decay,\n decoder=True)\n conv = tf.nn.conv2d(bottom, weights, [1, 1, 1, 1], padding='SAME')\n # Apply bias\n conv_biases = self._bias_variable([num_classes], constant=0.0)\n bias = tf.nn.bias_add(conv, conv_biases)\n\n _activation_summary(bias)\n\n return bias\n\n def _upscore_layer(self, bottom, shape,\n num_classes, name, debug,\n ksize=4, stride=2):\n strides = [1, stride, stride, 1]\n with tf.variable_scope(name):\n in_features = bottom.get_shape()[3].value\n\n if shape is None:\n # Compute shape out of Bottom\n in_shape = tf.shape(bottom)\n\n h = ((in_shape[1] - 1) * stride) + 1\n w = ((in_shape[2] - 1) * stride) + 1\n new_shape = [in_shape[0], h, w, num_classes]\n else:\n new_shape = [shape[0], shape[1], shape[2], num_classes]\n output_shape = tf.stack(new_shape)\n\n logging.debug(\"Layer: %s, Fan-in: %d\" % (name, in_features))\n f_shape = [ksize, ksize, num_classes, in_features]\n\n # create\n num_input = ksize * ksize * in_features / stride\n stddev = (2 / num_input)**0.5\n\n weights = self.get_deconv_filter(f_shape)\n self._add_wd_and_summary(weights, self.wd, \"fc_wlosses\")\n deconv = tf.nn.conv2d_transpose(bottom, weights, output_shape,\n strides=strides, padding='SAME')\n\n if debug:\n deconv = tf.Print(deconv, [tf.shape(deconv)],\n message='Shape of %s' % name,\n summarize=4, first_n=1)\n\n _activation_summary(deconv)\n return deconv\n\n def get_deconv_filter(self, f_shape):\n width = f_shape[0]\n heigh = f_shape[0]\n f = ceil(width/2.0)\n c = (2 * f - 1 - f % 2) / (2.0 * f)\n bilinear = np.zeros([f_shape[0], f_shape[1]])\n for x in range(width):\n for y in range(heigh):\n value = (1 - abs(x / f - c)) * (1 - abs(y / f - c))\n bilinear[x, y] = value\n weights = np.zeros(f_shape)\n for i in range(f_shape[2]):\n weights[:, :, i, i] = bilinear\n\n init = tf.constant_initializer(value=weights,\n dtype=tf.float32)\n var = tf.get_variable(name=\"up_filter\", initializer=init,\n shape=weights.shape)\n return var\n\n def get_conv_filter(self, name):\n init = tf.constant_initializer(value=self.data_dict[name][0],\n dtype=tf.float32)\n shape = self.data_dict[name][0].shape\n print('Layer name: %s' % name)\n print('Layer shape: %s' % str(shape))\n var = tf.get_variable(name=\"filter\", initializer=init, shape=shape)\n if not tf.get_variable_scope().reuse:\n weight_decay = tf.multiply(tf.nn.l2_loss(var), self.wd,\n name='weight_loss')\n tf.add_to_collection(tf.GraphKeys.REGULARIZATION_LOSSES,\n weight_decay)\n _variable_summaries(var)\n return var\n\n def get_bias(self, name, num_classes=None):\n bias_wights = self.data_dict[name][1]\n shape = self.data_dict[name][1].shape\n if name == 'fc8':\n bias_wights = self._bias_reshape(bias_wights, shape[0],\n num_classes)\n shape = [num_classes]\n init = tf.constant_initializer(value=bias_wights,\n dtype=tf.float32)\n var = tf.get_variable(name=\"biases\", initializer=init, shape=shape)\n _variable_summaries(var)\n return var\n\n def get_fc_weight(self, name):\n init = tf.constant_initializer(value=self.data_dict[name][0],\n dtype=tf.float32)\n shape = self.data_dict[name][0].shape\n var = tf.get_variable(name=\"weights\", initializer=init, shape=shape)\n if not tf.get_variable_scope().reuse:\n weight_decay = tf.multiply(tf.nn.l2_loss(var), self.wd,\n name='weight_loss')\n tf.add_to_collection(tf.GraphKeys.REGULARIZATION_LOSSES,\n weight_decay)\n _variable_summaries(var)\n return var\n\n def _bias_reshape(self, bweight, num_orig, num_new):\n \"\"\" Build bias weights for filter produces with `_summary_reshape`\n\n \"\"\"\n n_averaged_elements = num_orig//num_new\n avg_bweight = np.zeros(num_new)\n for i in range(0, num_orig, n_averaged_elements):\n start_idx = i\n end_idx = start_idx + n_averaged_elements\n avg_idx = start_idx//n_averaged_elements\n if avg_idx == num_new:\n break\n avg_bweight[avg_idx] = np.mean(bweight[start_idx:end_idx])\n return avg_bweight\n\n def _summary_reshape(self, fweight, shape, num_new):\n \"\"\" Produce weights for a reduced fully-connected layer.\n\n FC8 of VGG produces 1000 classes. Most semantic segmentation\n task require much less classes. This reshapes the original weights\n to be used in a fully-convolutional layer which produces num_new\n classes. To archive this the average (mean) of n adjanced classes is\n taken.\n\n Consider reordering fweight, to perserve semantic meaning of the\n weights.\n\n Args:\n fweight: original weights\n shape: shape of the desired fully-convolutional layer\n num_new: number of new classes\n\n\n Returns:\n Filter weights for `num_new` classes.\n \"\"\"\n num_orig = shape[3]\n shape[3] = num_new\n assert(num_new < num_orig)\n # // is “floor” division (rounds down to nearest whole number)\n n_averaged_elements = num_orig//num_new \n # avg_fweight is a new tensor with num_new output channels\n avg_fweight = np.zeros(shape)\n for i in range(0, num_orig, n_averaged_elements):\n start_idx = i\n end_idx = start_idx + n_averaged_elements\n avg_idx = start_idx//n_averaged_elements\n if avg_idx == num_new:\n break\n avg_fweight[:, :, :, avg_idx] = np.mean(\n fweight[:, :, :, start_idx:end_idx], axis=3)\n return avg_fweight\n\n def _variable_with_weight_decay(self, shape, stddev, wd, decoder=False):\n \"\"\"Helper to create an initialized Variable with weight decay.\n\n Note that the Variable is initialized with a truncated normal\n distribution.\n A weight decay is added only if one is specified.\n\n Args:\n name: name of the variable\n shape: list of ints\n stddev: standard deviation of a truncated Gaussian\n wd: add L2Loss weight decay multiplied by this float. If None, weight\n decay is not added for this Variable.\n\n Returns:\n Variable Tensor\n \"\"\"\n\n initializer = tf.truncated_normal_initializer(stddev=stddev)\n var = tf.get_variable('weights', shape=shape,\n initializer=initializer)\n\n collection_name = tf.GraphKeys.REGULARIZATION_LOSSES\n if wd and (not tf.get_variable_scope().reuse):\n weight_decay = tf.multiply(\n tf.nn.l2_loss(var), wd, name='weight_loss')\n tf.add_to_collection(collection_name, weight_decay)\n _variable_summaries(var)\n return var\n\n def _add_wd_and_summary(self, var, wd, collection_name=None):\n if collection_name is None:\n collection_name = tf.GraphKeys.REGULARIZATION_LOSSES\n if wd and (not tf.get_variable_scope().reuse):\n weight_decay = tf.multiply(\n tf.nn.l2_loss(var), wd, name='weight_loss')\n tf.add_to_collection(collection_name, weight_decay)\n _variable_summaries(var)\n return var\n\n def _bias_variable(self, shape, constant=0.0):\n initializer = tf.constant_initializer(constant)\n var = tf.get_variable(name='biases', shape=shape,\n initializer=initializer)\n _variable_summaries(var)\n return var\n\n def get_fc_weight_reshape(self, name, shape, num_classes=None):\n print('Layer name: %s' % name)\n print('Layer shape: %s' % shape)\n weights = self.data_dict[name][0]\n weights = weights.reshape(shape)\n if num_classes is not None:\n weights = self._summary_reshape(weights, shape,\n num_new=num_classes)\n init = tf.constant_initializer(value=weights,\n dtype=tf.float32)\n var = tf.get_variable(name=\"weights\", initializer=init, shape=shape)\n return var\n\n\ndef _activation_summary(x):\n \"\"\"Helper to create summaries for activations.\n\n Creates a summary that provides a histogram of activations.\n Creates a summary that measure the sparsity of activations.\n\n Args:\n x: Tensor\n Returns:\n nothing\n \"\"\"\n # Remove 'tower_[0-9]/' from the name in case this is a multi-GPU training\n # session. This helps the clarity of presentation on tensorboard.\n tensor_name = x.op.name\n # tensor_name = re.sub('%s_[0-9]*/' % TOWER_NAME, '', x.op.name)\n tf.summary.histogram(tensor_name + '/activations', x)\n tf.summary.scalar(tensor_name + '/sparsity', tf.nn.zero_fraction(x))\n\n\ndef _variable_summaries(var):\n \"\"\"Attach a lot of summaries to a Tensor.\"\"\"\n if not tf.get_variable_scope().reuse:\n name = var.op.name\n logging.info(\"Creating Summary for: %s\" % name)\n with tf.name_scope('summaries'):\n mean = tf.reduce_mean(var)\n tf.summary.scalar(name + '/mean', mean)\n with tf.name_scope('stddev'):\n stddev = tf.sqrt(tf.reduce_sum(tf.square(var - mean)))\n tf.summary.scalar(name + '/sttdev', stddev)\n tf.summary.scalar(name + '/max', tf.reduce_max(var))\n tf.summary.scalar(name + '/min', tf.reduce_min(var))\n tf.summary.histogram(name, var)\n\ndef pixel_wise_cross_entropy(logits, labels, num_classes):\n # self.cross_entropy = tf.reduce_mean( \n # tf.nn.softmax_cross_entropy_with_logits(\n # labels=tf.reshape(self.label, [-1, 21]), logits=tf.reshape(self.deconv1, [-1,21])))\n # print(logits.get_shape().as_list())\n\n # logits = tf.reshape(logits, (-1, num_classes))\n # print(logits.get_shape().as_list())\n print(labels.get_shape().as_list())\n\n cross_entropy = tf.reduce_mean(tf.nn.sparse_softmax_cross_entropy_with_logits(\n labels=tf.to_int32(labels/255), logits=logits)) \n # weight decay\n lambda_ = 5**(-4)\n #l2_loss = tf.add_n([tf.nn.l2_loss(v) for v in tf.trainable_variables()])\n l2_loss = tf.add_n(tf.get_collection(tf.GraphKeys.REGULARIZATION_LOSSES), name='l2_weight_loss')\n cross_entropy += lambda_*l2_loss\n return cross_entropy\n\n", "sub_path": "Step2-Segmentation/fcn8_vgg_ours.py", "file_name": "fcn8_vgg_ours.py", "file_ext": "py", "file_size_in_byte": 21237, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sys.modules", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.pardir", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "logging.info", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 29, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 34, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.split", "line_number": 68, "usage_type": "call"}, {"api_name": "tensorflow.concat", "line_number": 75, "usage_type": "call"}, {"api_name": "tensorflow.Print", "line_number": 85, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 85, "usage_type": "call"}, {"api_name": "tensorflow.nn.dropout", "line_number": 115, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 115, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.dropout", "line_number": 119, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 119, "usage_type": "attribute"}, {"api_name": "tensorflow.argmax", "line_number": 129, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 132, "usage_type": "call"}, {"api_name": "tensorflow.add", "line_number": 138, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 141, "usage_type": "call"}, {"api_name": "tensorflow.add", "line_number": 147, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 150, "usage_type": "call"}, {"api_name": "tensorflow.argmax", "line_number": 155, "usage_type": "call"}, {"api_name": "tensorflow.Print", "line_number": 157, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 157, "usage_type": "call"}, {"api_name": "tensorflow.nn.max_pool", "line_number": 166, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 166, "usage_type": "attribute"}, {"api_name": "tensorflow.Print", "line_number": 171, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 171, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 177, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 180, "usage_type": "call"}, {"api_name": "tensorflow.truncated_normal", "line_number": 180, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 180, "usage_type": "attribute"}, {"api_name": "tensorflow.concat", "line_number": 182, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d", "line_number": 185, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 185, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.bias_add", "line_number": 188, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 188, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.relu", "line_number": 190, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 190, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 197, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d", "line_number": 211, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 211, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.bias_add", "line_number": 213, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 213, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.relu", "line_number": 216, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 216, "usage_type": "attribute"}, {"api_name": "tensorflow.Print", "line_number": 220, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 220, "usage_type": "call"}, {"api_name": "tensorflow.variable_scope", "line_number": 226, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d", "line_number": 243, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 243, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.bias_add", "line_number": 246, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 246, "usage_type": "attribute"}, {"api_name": "tensorflow.variable_scope", "line_number": 256, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 261, "usage_type": "call"}, {"api_name": "tensorflow.stack", "line_number": 268, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 270, "usage_type": "call"}, {"api_name": "tensorflow.nn.conv2d_transpose", "line_number": 279, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 279, "usage_type": "attribute"}, {"api_name": "tensorflow.Print", "line_number": 283, "usage_type": "call"}, {"api_name": "tensorflow.shape", "line_number": 283, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 293, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 295, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 300, "usage_type": "call"}, {"api_name": "tensorflow.constant_initializer", "line_number": 304, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 305, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 306, "usage_type": "call"}, {"api_name": "tensorflow.constant_initializer", "line_number": 311, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 312, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 316, "usage_type": "call"}, {"api_name": "tensorflow.get_variable_scope", "line_number": 317, "usage_type": "call"}, {"api_name": "tensorflow.multiply", "line_number": 318, "usage_type": "call"}, {"api_name": "tensorflow.nn.l2_loss", "line_number": 318, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 318, "usage_type": "attribute"}, {"api_name": "tensorflow.add_to_collection", "line_number": 320, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 320, "usage_type": "attribute"}, {"api_name": "tensorflow.constant_initializer", "line_number": 332, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 333, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 334, "usage_type": "call"}, {"api_name": "tensorflow.constant_initializer", "line_number": 339, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 340, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 342, "usage_type": "call"}, {"api_name": "tensorflow.get_variable_scope", "line_number": 343, "usage_type": "call"}, {"api_name": "tensorflow.multiply", "line_number": 344, "usage_type": "call"}, {"api_name": "tensorflow.nn.l2_loss", "line_number": 344, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 344, "usage_type": "attribute"}, {"api_name": "tensorflow.add_to_collection", "line_number": 346, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 346, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 356, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 363, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 393, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 400, "usage_type": "call"}, {"api_name": "tensorflow.truncated_normal_initializer", "line_number": 422, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 423, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 426, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable_scope", "line_number": 427, "usage_type": "call"}, {"api_name": "tensorflow.multiply", "line_number": 428, "usage_type": "call"}, {"api_name": "tensorflow.nn.l2_loss", "line_number": 429, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 429, "usage_type": "attribute"}, {"api_name": "tensorflow.add_to_collection", "line_number": 430, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 436, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable_scope", "line_number": 437, "usage_type": "call"}, {"api_name": "tensorflow.multiply", "line_number": 438, "usage_type": "call"}, {"api_name": "tensorflow.nn.l2_loss", "line_number": 439, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 439, "usage_type": "attribute"}, {"api_name": "tensorflow.add_to_collection", "line_number": 440, "usage_type": "call"}, {"api_name": "tensorflow.constant_initializer", "line_number": 445, "usage_type": "call"}, {"api_name": "tensorflow.get_variable", "line_number": 446, "usage_type": "call"}, {"api_name": "tensorflow.constant_initializer", "line_number": 459, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 460, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable", "line_number": 461, "usage_type": "call"}, {"api_name": "tensorflow.summary.histogram", "line_number": 480, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 480, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.scalar", "line_number": 481, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 481, "usage_type": "attribute"}, {"api_name": "tensorflow.nn.zero_fraction", "line_number": 481, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 481, "usage_type": "attribute"}, {"api_name": "tensorflow.get_variable_scope", "line_number": 486, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 488, "usage_type": "call"}, {"api_name": "tensorflow.name_scope", "line_number": 489, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 490, "usage_type": "call"}, {"api_name": "tensorflow.summary.scalar", "line_number": 491, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 491, "usage_type": "attribute"}, {"api_name": "tensorflow.name_scope", "line_number": 492, "usage_type": "call"}, {"api_name": "tensorflow.sqrt", "line_number": 493, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 493, "usage_type": "call"}, {"api_name": "tensorflow.square", "line_number": 493, "usage_type": "call"}, {"api_name": "tensorflow.summary.scalar", "line_number": 494, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 494, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.scalar", "line_number": 495, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 495, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_max", "line_number": 495, "usage_type": "call"}, {"api_name": "tensorflow.summary.scalar", "line_number": 496, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 496, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_min", "line_number": 496, "usage_type": "call"}, {"api_name": "tensorflow.summary.histogram", "line_number": 497, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 497, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_mean", "line_number": 509, "usage_type": "call"}, {"api_name": "tensorflow.nn.sparse_softmax_cross_entropy_with_logits", "line_number": 509, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 509, "usage_type": "attribute"}, {"api_name": "tensorflow.to_int32", "line_number": 510, "usage_type": "call"}, {"api_name": "tensorflow.add_n", "line_number": 514, "usage_type": "call"}, {"api_name": "tensorflow.get_collection", "line_number": 514, "usage_type": "call"}, {"api_name": "tensorflow.GraphKeys", "line_number": 514, "usage_type": "attribute"}]}
+{"seq_id": "10528827", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Oct 19 18:26:08 2015\n\n@author: yuugangyang\n\"\"\"\n\nfrom time import time\nimport pylab\nimport matplotlib\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits.mplot3d import Axes3D\nfrom matplotlib.ticker import NullFormatter\nimport numpy as np\nimport scipy\nimport math\nfrom sklearn import manifold, datasets\nfrom sklearn.metrics.pairwise import euclidean_distances\nfrom sklearn.decomposition import PCA\nmatplotlib.interactive(True)\nfrom sklearn.manifold.spectral_embedding_ import _graph_is_connected\nfrom compute_diffusion_map import *\nfrom thresholdDistanceMatrix import *\nfrom sampleDataExample import *\nplt.rc('text',usetex=True)\n# Next line to silence pyflakes. This import is needed.\nAxes3D\n\nn_points = 800\n#X, color = datasets.samples_generator.make_swiss_roll(n_points)\n#\nplt.close('all')\nX, color = getCornerPlane(n_points)\nX, color = getTwinPeaks(n_points)\nfig = plt.figure(1)\nax = fig.add_subplot(111,projection='3d')\nax.scatter(X[:, 0], X[:, 1], X[:, 2], c=color, cmap=plt.cm.Spectral)\nax.set_xlabel('x1')\nax.set_ylabel('x2')\nax.set_zlabel('x3')\n\n# now we calculate the pair-wise distance\ndist = euclidean_distances(X,X)\nsort_dist = np.sort(dist)\ndist_thresh,epi = thresholdDistanceMatrix(dist,epi=0.5,mode='manual')\n#dist_thresh,epi = thresholdDistanceMatrix(dist)\nplt.figure(10)\nplt.imshow(dist_thresh)\nplt.colorbar()\nif not _graph_is_connected(dist_thresh):\n raise ValueError('Graph is disconnected')\n\n\nembedding, result = compute_diffusion_map(dist_thresh,alpha=0.,n_components=10)\n\nsum_vec = np.sum(dist_thresh,axis=1)\n\nsum_vec = np.power(sum_vec,-0.5)\nnorm_mat = np.diag(sum_vec)\n\n\nM = np.dot(norm_mat,np.dot(dist_thresh,norm_mat))\n\n#M = dist_thresh\n\nfor i in range(M.shape[0]):\n M[i] = M[i]/np.sum(M[i])\n\n# after normalization, M is not sysmetric \n#M = np.transpose(M)\n \nw,v = scipy.sparse.linalg.eigs(M,k=10)\nv = np.dot(v,np.diag(w))\n\nfig = plt.figure(6)\nax = fig.add_subplot(111,projection='3d')\nax.scatter(v[:,1],v[:,2],v[:,3],c=color, cmap=plt.cm.Spectral)\n\nfig = plt.figure(8)\nax = fig.add_subplot(111,projection='3d')\nax.scatter(v[:,1],v[:,2],c=color, cmap=plt.cm.Spectral)\n\nfig = plt.figure(7)\nax = fig.add_subplot(111)\nax.scatter(embedding[:,0],embedding[:,1],c=color, cmap=plt.cm.Spectral)\n\nfig = plt.figure(9)\nax = fig.add_subplot(111,projection='3d')\nax.scatter(embedding[:,0],embedding[:,1],embedding[:,2],c=color, cmap=plt.cm.Spectral)\n#\n", "sub_path": "diffusionmapping.py", "file_name": "diffusionmapping.py", "file_ext": "py", "file_size_in_byte": 2417, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "matplotlib.interactive", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rc", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "mpl_toolkits.mplot3d.Axes3D", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 37, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "sklearn.metrics.pairwise.euclidean_distances", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "sklearn.manifold.spectral_embedding_._graph_is_connected", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 67, "usage_type": "call"}, {"api_name": "scipy.sparse.linalg.eigs", "line_number": 72, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 72, "usage_type": "attribute"}, {"api_name": "numpy.dot", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 77, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 81, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 83, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 83, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 85, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 87, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 87, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 89, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}]}
+{"seq_id": "296225278", "text": "\nfrom random import choice\nfrom requests import exceptions as req_exc\nfrom bs4 import BeautifulSoup\nfrom lxml.html.clean import clean_html, Cleaner\n\nuser_agent_list = ['Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/53.0.2785.143 Safari/537.36',\n 'Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/53.0.2785.143 Safari/537.36',\n 'Mozilla/5.0 (Windows NT 10.0; WOW64; rv:49.0) Gecko/20100101 Firefox/49.0',\n 'Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/54.0.2840.71 Safari/537.36']\n\ndomain = 'http://www.example.com'\n\ndef random_user_agent():\n agent = choice(user_agent_list)\n return {'User-Agent':agent}\n\ndef make_request(url):\n try:\n r = requests.get(url,headers=random_user_agent())\n return r\n except req_exc.InvalidURL as invalid_url:\n with open('Classification.csv','a',encoding='utf-8') as class_file:\n class_file.write('\"{}\",\"{}\"\\n'.format(url,'Invalid URL'))\n except req_exc.ConnectionError as conn_error:\n with open('Classification.csv','a',encoding='utf-8') as class_file:\n class_file.write('\"{}\",\"{}\"\\n'.format(url, 'Connection Error'))\n except req_exc.TooManyRedirects as too_many_redirects:\n with open('Classification.csv','a',encoding='utf-8') as class_file:\n class_file.write('\"{}\",\"{}\"\\n'.format(url, 'Too Many Redirects'))\n except req_exc.MissingSchema as missing_schema:\n with open('Classification.csv','a',encoding='utf-8') as class_file:\n class_file.write('\"{}\",\"{}\"'.format(url,'Missing Schema'))\n except:\n pass\n\ndef generate_link_list(response,clean_html):\n url = response.url\n empty_link_list = []\n paragraphs_with_links = []\n soup = BeautifulSoup(clean_html,'html5lib')\n paragraphs = soup.find_all('p')\n for paragraph in paragraphs:\n does_paragraph_have_link = paragraph.find('a')\n if does_paragraph_have_link != None:\n if 'http://www.example.com' in str(does_paragraph_have_link):\n paragraphs_with_links.append(paragraph)\n for paragraphs_with_link in paragraphs_with_links:\n paragraph_link = paragraphs_with_link.find('a')\n paragraph_text = paragraphs_with_link.get_text()\n link_text = paragraph_link.get_text()\n if len(str(paragraph_text)) > len(str(link_text)):\n with open('Classification.csv','a',encoding='utf-8') as class_file:\n class_file.write('\"{}\",\"{}\"\\n'.format(url,'Content'))\n else:\n with open('Classification.csv','a',encoding='utf-8') as class_file:\n class_file.write('\"{}\",\"{}\"\\n'.format(url,'Navigational'))\n\n\n\ndef clean_html(html):\n try:\n cleaner = Cleaner(page_structure=False, scripts=True, style=True,\n remove_tags=['span', 'div', 'li', 'ul'],kill_tags=['img'])\n new_html = cleaner.clean_html(html)\n return new_html\n except:\n pass\n\ndef main():\n url_list = [line.rstrip() for line in open(r'urls.txt')]\n for url in url_list:\n try:\n response = requests.get(url)\n cleaned_html = clean_html(response.text)\n generate_link_list(response,cleaned_html)\n except:\n pass\n\nmain()\n\n", "sub_path": "classify_link.py", "file_name": "classify_link.py", "file_ext": "py", "file_size_in_byte": 3356, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "random.choice", "line_number": 15, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 20, "usage_type": "call"}, {"api_name": "requests.exceptions.InvalidURL", "line_number": 22, "usage_type": "attribute"}, {"api_name": "requests.exceptions", "line_number": 22, "usage_type": "name"}, {"api_name": "requests.exceptions.ConnectionError", "line_number": 25, "usage_type": "attribute"}, {"api_name": "requests.exceptions", "line_number": 25, "usage_type": "name"}, {"api_name": "requests.exceptions.TooManyRedirects", "line_number": 28, "usage_type": "attribute"}, {"api_name": "requests.exceptions", "line_number": 28, "usage_type": "name"}, {"api_name": "requests.exceptions.MissingSchema", "line_number": 31, "usage_type": "attribute"}, {"api_name": "requests.exceptions", "line_number": 31, "usage_type": "name"}, {"api_name": "bs4.BeautifulSoup", "line_number": 41, "usage_type": "call"}, {"api_name": "lxml.html.clean.clean_html", "line_number": 41, "usage_type": "argument"}, {"api_name": "lxml.html.clean.Cleaner", "line_number": 63, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 74, "usage_type": "call"}, {"api_name": "lxml.html.clean.clean_html", "line_number": 75, "usage_type": "call"}]}
+{"seq_id": "348063623", "text": "from pymongo import MongoClient\nimport time\nimport datetime\n\ncollection = MongoClient().robot.weibo\n\ndef insert_fake_browser_based_robots():\n robots = []\n for x in range(100):\n robots.append({\n 'id': x,\n 'type': 1,\n 'server_id':300000,\n 'profile_path' : 'whatever',\n 'timestamp': int(time.time()),\n 'timestamp_str': str(datetime.datetime.now()).split('.')[0],\n 'manage_status': 0,\n 'server_status': 0,\n 'weibo_status': 0,\n 'weibo_uid' : 1000000000+ x,\n 'nick' : 'whatever',\n 'avatar' : 'whaterver'\n })\n\n collection.insert_many(robots)\n\nif __name__ == '__main__':\n insert_fake_browser_based_robots()\n", "sub_path": "src/apps/unittest/robot.py", "file_name": "robot.py", "file_ext": "py", "file_size_in_byte": 757, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pymongo.MongoClient", "line_number": 5, "usage_type": "call"}, {"api_name": "time.time", "line_number": 15, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 16, "usage_type": "attribute"}]}
+{"seq_id": "423246454", "text": "#!/usr/bin/env python\n\nimport argparse\nimport sys\nimport boto\nfrom dateutil import parser\nfrom datetime import datetime, timedelta\nimport pytz\nimport time\nimport re\nfrom Queue import Queue\nfrom threading import Thread,Lock\nimport requests\nfrom requests.auth import HTTPBasicAuth\nfrom xml.etree import ElementTree\nimport ConfigParser\nimport os\nimport re\nimport logging\n\nclass S3Backup:\n\tdef __init__(self, config_file=None, num_threads=20):\n\t\t# disable the boto logs\n\t\tlogging.getLogger('boto').setLevel(logging.ERROR)\n\t\tlogging.getLogger('requests').setLevel(logging.ERROR)\n\t\tself.maximum_retention_days = 31\n\n\t\tself.queue_done = False\n\t\tself.num_threads = num_threads\n\t\t\n\t\tself.config = self.loadConfig(config_file)\n\n\t\tself.portal_host = self.config['portal_host']\n\t\tself.portal_user = self.config['portal_user']\n\t\tself.portal_pass = self.config['portal_pass']\n\t\tself.s3_src_bucket = self.config['s3_src_bucket']\n\t\tself.s3_src_prefix= self.config['s3_src_prefix']\n\t\tself.s3_dst_bucket = self.config['s3_dst_bucket']\n\t\tself.s3_dst_prefix = self.config['s3_dst_prefix']\n\t\tself.aws_access_key_id = self.config['aws_access_key_id']\n\t\tself.aws_secret_access_key = self.config['aws_secret_access_key']\n\t\tself.queue = Queue()\n\n\t\tself.backup_time = datetime.utcnow().replace(tzinfo = pytz.timezone('UTC'))\n\t\tself.valid_backup_set_regex = re.compile('.*/[0-9]{14,14}/.*')\n\t\tself.total_bytes = 0\n\t\tself.total_keys = 0\n\t\tself.total_keys_deleted = 0\n\n\t\tself.sites = self.getSites()\n\n\tdef loadConfig(self, config_file):\n\t\tconfig = ConfigParser.ConfigParser()\n\t\tif not os.path.isfile(config_file):\n\t\t\tprint('Please create mindtouch.s3.files.backup.ini file with the following contents:')\n\t\t\tprint('')\n\t\t\tprint('[Credentials]')\n\t\t\tprint('portal_host = portal.int.mindtouch.us')\n\t\t\tprint('portal_user = portal')\n\t\t\tprint('portal_pass = some_password')\n\t\t\tprint('src_s3_bucket = some_s3_bucket')\n\t\t\tprint('src_s3_prefix = some_s3_prefix')\n\t\t\tprint('dst_s3_bucket = some_s3_bucket')\n\t\t\tprint('dst_s3_prefix = some_s3_prefix')\n\t\t\tprint('aws_access_key_id = 12345')\n\t\t\tprint('aws_secret_access_key = abcdef')\n\t\t\tsys.exit(1)\n\n\t\tconfig.read(config_file)\n\n\t\toptions = {}\n\t\tsection = 'Credentials'\n\t\tfor option in config.options(section):\n\t\t\toptions[option] = config.get(section, option)\n\t\t\t\n\t\t# verify options\n\t\tif not 'portal_host' in options or not 'portal_user' in options or not 'portal_pass' in options or not 's3_src_bucket' in options or not 's3_src_prefix' in options or not 's3_dst_bucket' in options or not 's3_dst_prefix' in options or not 'aws_access_key_id' in options or not 'aws_secret_access_key' in options:\n\t\t\traise ValueError(\"%s contains invalid or missing data\" % config_file)\n\n\t\treturn options\n\t\n\tdef backup(self):\n\t\tself.startThreads(worker=self.copyKeysWorker)\n\t\tfor site in self.sites:\n\t\t\tself.backupSite(site)\n\t\t\twhile self.getQueueSize() > 0:\n\t\t\t\tself.printQueueSize()\n\t\t\t\ttime.sleep(1)\n\t\tself.finishThreads()\n\n\n\tdef backupSite(self, site, mode='DAILY'):\n\t\tlogging.info(\"[%s] Starting backup\" % site)\n\n\t\tlast_full_date = self.findLastBackup(site=site, type='FULL')\n\t\tif last_full_date == None:\n\t\t\t# never done a full backup\n\t\t\tlogging.debug(\"[%s] Unable to find the last full backup, proceeding with full backup\" % site)\n\t\t\tself.queueKeysForCopy(site, backup_mode='FULL', modified_since=None)\n\t\telif last_full_date <= self.backup_time - timedelta(days=self.maximum_retention_days):\n\t\t\t# oldest backup is older than 31 days, do a full backup\n\t\t\tlogging.debug(\"[%s] Oldest full backup is too old, proceeding with full backup\" % site)\n\t\t\tself.queueKeysForCopy(site, backup_mode='FULL', modified_since=None)\n\t\telse:\n\t\t\tlogging.debug(\"[%s] Proceeding with daily backup, last full backup is: %s\" % (site, last_full_date))\n\t\t\tstart_date = self.findLastBackup(site=site, type='DAILY')\n\t\t\tif start_date == None:\n\t\t\t\tstart_date = last_full_date\n\n\t\t\tlogging.debug(\"[%s] Backing up files newer than: %s\" % (site, start_date))\n\t\t\tself.queueKeysForCopy(site, backup_mode='DAILY', modified_since=start_date)\n\n\t\tlogging.info(\"[%s] Completed backup\" % site)\n\n\tdef findLastBackup(self, site, type):\n\t\ttype = type.upper()\n\t\tif type != \"FULL\" and type != \"DAILY\":\n\t\t\traise ValueError(\"type must be 'full' or 'daily'\")\n\n\t\tconn = boto.connect_s3(aws_access_key_id=self.aws_access_key_id, aws_secret_access_key=self.aws_secret_access_key)\n\t\tbucket = conn.get_bucket(bucket_name=self.s3_dst_bucket) \t\n\t\tprefix = self.getDstPrefixForSite(site) + type + '/'\n\n\t\tlatest_date = None\n\t\tfor key in bucket.list(delimiter='/', prefix=prefix):\n\t\t\t# keys should look like: \n\t\t\t# parse the right-most component \n\t\t\tdate = self.parse_time(key.name.split('/')[-2])\n\t\t\tdate = date.replace(tzinfo = pytz.timezone('UTC'))\n\t\t\tif latest_date == None:\n\t\t\t\tlatest_date = date\n\t\t\telse:\n\t\t\t\t# compare the date to see if it's newer\n\t\t\t\tif date > latest_date:\n\t\t\t\t\tlatest_date = date\n\n\t\treturn latest_date\n\n\t\n\tdef getSites(self):\n\t\tportal_uri = \"http://%s/%s\" % (self.portal_host, 'api/sites')\n\t\tsites = []\n\t\ttry:\n\t\t\tlogging.debug(\"connecting to Portal: %s, user: %s\" % (portal_uri, self.portal_user))\n\t\t\tr = requests.get(portal_uri, auth=HTTPBasicAuth(self.portal_user, self.portal_pass))\n\t\t\tr.raise_for_status()\n\t\t\tfor site in ElementTree.fromstring(r.content):\n\t\t\t\tsites.append(site.attrib['id'])\n\t\texcept:\n\t\t\traise \n\n\t\treturn sorted(sites)\n\n\tdef getSrcPrefix(self):\n\t\tif len(self.s3_src_prefix) > 0:\n\t\t\treturn self.s3_src_prefix + '/'\n\t\telse:\n\t\t\treturn ''\n\n\tdef getDstPrefix(self):\n\t\tif len(self.s3_dst_prefix) > 0:\n\t\t\treturn self.s3_dst_prefix + '/'\n\t\telse:\n\t\t\treturn ''\n\n\t\t\n\tdef getSrcPrefixForSite(self, site_id):\n\t\treturn self.getSrcPrefix() + site_id + '/'\n\n\n\tdef getDstPrefixForSite(self, site_id):\n\t\treturn self.getDstPrefix() + site_id + '/'\n\n\tdef getDstKeyName(self, key_name, site_id, backup_mode):\n\t\tif len(self.getSrcPrefix()) > 0:\n\t\t\tkey_name = replace(self.getSrcPrefix(), self.getDstPrefix())\n\t\telse:\n\t\t\tkey_name = self.getDstPrefix() + key_name\n\n\t\tkey_name = key_name.replace(site_id + '/', site_id + '/' + backup_mode + '/' + self.backup_time.strftime(\"%Y%m%d%H%M%S\") + '/')\n\t\treturn key_name\n\n\tdef queueKeysForCopy(self, site_id, backup_mode, modified_since=None):\n\t\tconn = boto.connect_s3(aws_access_key_id=self.aws_access_key_id, aws_secret_access_key=self.aws_secret_access_key)\n\t\tsrc_bucket = conn.get_bucket(bucket_name=self.s3_src_bucket) \t\n\t\tcount = 0\n\t\tfor key in src_bucket.list(prefix=self.getSrcPrefixForSite(site_id)):\n\t\t\tif modified_since == None or self.parse_time(key.last_modified) >= modified_since:\n\t\t\t\tcount = count + 1\n\t\t\t\tdst_key_name = self.getDstKeyName(key_name=key.name, site_id=site_id, backup_mode=backup_mode)\n\t\t\t\tif count % 1000 == 0:\n\t\t\t\t\tlogging.debug(\"[%s]: Queued: %d\" % (site_id, count))\n\n\t\t\t\tself.total_bytes += key.size\n\t\t\t\tself.total_keys += 1\n\t\t\t\tself.queue.put([key.name, dst_key_name])\n\n\tdef copyKeysWorker(self):\n\t\ttry:\n\t\t\tconn = boto.connect_s3(aws_access_key_id=self.aws_access_key_id, aws_secret_access_key=self.aws_secret_access_key)\n\t\t\tdst_bucket = conn.get_bucket(bucket_name=self.s3_dst_bucket)\n\t\t\twhile not self.queue_done:\n\t\t\t\tsrc_key, dst_key = self.queue.get()\n\t\t\t\tdst_bucket.copy_key(dst_key, self.s3_src_bucket, src_key)\n\t\t\t\tself.queue.task_done()\n\t\texcept:\n\t\t\traise\n\t\t\t\n\tdef getQueueSize(self):\n\t\treturn self.queue.qsize()\n\n\tdef printQueueSize(self):\n\t\tlogging.debug(\"Items remaining: %d\" % self.getQueueSize())\n\n\tdef startThreads(self, worker):\n\t\tfor i in range(self.num_threads):\n\t\t\tt = Thread(target=worker)\n\t\t\tt.daemon = True\n\t\t\tt.start()\n\t\t\n\tdef finishThreads(self):\n\t\tself.queue.join()\n\t\tself.queue_done = True\n\n\n\tdef delete_old_backups(self):\n\t\tself.queue_done = False\n\t\tself.startThreads(worker=self.deleteOldKeysWorker)\n\t\tfor site in self.sites:\n\t\t\tself.queueKeysForDelete(site)\n\t\t\twhile self.getQueueSize() > 0:\n\t\t\t\tself.printQueueSize()\n\t\t\t\ttime.sleep(1)\n\t\tself.finishThreads()\n\n\tdef queueKeysForDelete(self, site):\n\t\tconn = boto.connect_s3(aws_access_key_id=self.aws_access_key_id, aws_secret_access_key=self.aws_secret_access_key)\n\t\tdst_bucket = conn.get_bucket(bucket_name=self.s3_dst_bucket)\n\n\t\tcount = 0\n\t\tbackups_to_delete = []\n\n\t\t# find all full backups older than 31 days (self.maximum_retention_days) and queue for delete\n\t\ttype = 'FULL'\n\t\tprefix = self.getDstPrefixForSite(site) + type + '/'\n\t\tfor key in dst_bucket.list(prefix=prefix, delimiter='/'):\n\t\t\t# make sure the key looks like /DAILY/YYYYMMDDHHMMSS/ just to be careful\n\t\t\tif self.valid_backup_set_regex.match(key.name) != None:\t\n\t\t\t\tdate = self.parse_time(key.name[key.name.index(type + '/')+ len(type) + 1:].split('/')[0])\n\t\t\t\tdate = date.replace(tzinfo = pytz.timezone('UTC'))\n\t\t\t\tdelta = self.backup_time - date\n\n\t\t\t\tif (self.backup_time - date) >= timedelta(days=self.maximum_retention_days): \n\t\t\t\t\tlogging.debug(\"[%s] Deleting backup: %s\" % (site, key.name))\n\t\t\t\t\tbackups_to_delete.append(key.name)\n\t\t\t\t\t\n\n\t\t# find all daily backups older than the latest full backup and queue for delete\n\t\tlast_full_date = self.findLastBackup(site=site, type='FULL')\n\t\tif last_full_date != None: \n\t\t\ttype = 'DAILY'\n\t\t\tprefix = self.getDstPrefixForSite(site) + type + '/'\n\t\t\tfor key in dst_bucket.list(prefix=prefix, delimiter='/'):\n\t\t\t\t# make sure the key looks like /DAILY/YYYYMMDDHHMMSS/ just to be careful\n\t\t\t\tif self.valid_backup_set_regex.match(key.name) != None:\t\n\t\t\t\t\tdate = self.parse_time(key.name[key.name.index(type + '/')+ len(type) + 1:].split('/')[0])\n\t\t\t\t\tdate = date.replace(tzinfo = pytz.timezone('UTC'))\n\n\t\t\t\t\tif date < last_full_date:\n\t\t\t\t\t\tlogging.debug(\"[%s] Deleting: %s\" % (site, key.name))\n\t\t\t\t\t\tbackups_to_delete.append(key.name)\n\t\t\t\t\t\t\n\t\tfor backup in backups_to_delete:\n\t\t\tfor key in dst_bucket.list(prefix=backup):\n\t\t\t\tcount += 1\n\t\t\t\tif count % 1000 == 0:\n\t\t\t\t\tlogging.debug(\"[%s] Deleting: %d\" % (site, count))\n\t\t\t\tself.queue.put(key.name)\n\t\t\t\tself.total_keys_deleted += 1\n\n\t\t\t# also delete the parent folder\n\t\t\tself.queue.put(backup)\n\t\t\tself.total_keys_deleted += 1\n\n\n\tdef deleteOldKeysWorker(self):\n\t\ttry:\n\t\t\tconn = boto.connect_s3(aws_access_key_id=self.aws_access_key_id, aws_secret_access_key=self.aws_secret_access_key)\n\t\t\tdst_bucket = conn.get_bucket(bucket_name=self.s3_dst_bucket)\n\t\t\twhile not self.queue_done:\n\t\t\t\tdst_key = self.queue.get()\n\t\t\t\tdst_bucket.delete_key(dst_key)\n\t\t\t\tself.queue.task_done()\n\t\texcept:\n\t\t\traise\n\n\tdef parse_time(self, time):\n\t\ttry:\n\t\t\tret = parser.parse(time)\n\t\texcept ValueError:\n\t\t\tret = datetime.utcfromtimestamp(time)\n\t\treturn ret\n\n\ndef sizeof_fmt(num):\n\tfor x in ['bytes','KB','MB','GB']:\n\t\tif num < 1024.0 and num > -1024.0:\n\t\t\treturn \"%3.1f%s\" % (num, x)\n\t\tnum /= 1024.0\n\treturn \"%3.1f%s\" % (num, 'TB')\n\ndef define_arguments():\n\targumentParser = argparse.ArgumentParser(description='Copies file attachments between sites')\n\targumentParser.add_argument('-t', '--threads', help='Number of threads to spawn when copy files (Default: 20)', default=20)\n\targumentParser.add_argument('-c', '--config', help='Alternate config file (Default: mindtouch.s3.backup.files.ini', default=None)\n\targumentParser.add_argument('-v', '--verbose', help='Sets level of verbosity (repeat for higher verbosity)', action='count', default=None)\n\targumentParser.add_argument('-q', '--quiet', help='Will not send output to stdout', action='store_true', default=False)\n\targumentParser.add_argument('-l', '--log', help='Log to a file instead of the console', default=None)\n\treturn argumentParser\n\ndef parse_arguments(argumentParser):\n\tsettings = {}\n\targuments = argumentParser.parse_args()\n\tsettings['threads'] = arguments.threads\n\tsettings['verbosity'] = set_verbosity(arguments.verbose, arguments.quiet)\n\tsettings['config_file' ] = set_config_file(arguments.config)\n\tset_log(arguments.log, settings['verbosity'])\n\t\n\treturn settings\n\ndef set_config_file(config):\n\tif config == None:\n\t\t# assume current directory\n\t\treturn os.path.join(os.path.dirname(__file__), 'mindtouch.s3.files.backup.ini')\n\telse:\n\t\treturn os.path.abspath(config)\n\ndef set_verbosity(verbose, quiet):\n\tverbosity = 'CRITICAL'\n\tif quiet:\n\t\tverbosity = 'SILENT'\n\telse:\n\t\tif verbose == 1:\n\t\t\tverbosity = 'ERROR'\n\t\telif verbose == 2:\n\t\t\tverbosity = 'WARNING'\n\t\telif verbose == 3:\n\t\t\tverbosity = 'INFO'\n\t\telif verbose >= 4:\n\t\t\tverbosity = 'DEBUG'\n\treturn verbosity\n\ndef set_log(log, verbosity):\n\tif not verbosity == 'SILENT':\n\t\tif log:\n\t\t\tlogging.basicConfig(filename=log,\n\t\t\t\tlevel=verbosity,\n\t\t\t\tformat='%(asctime)s [%(levelname)s] - %(message)s')\n\t\telse:\n\t\t\tlogging.basicConfig(level=verbosity,\n\t\t\t\tformat='%(asctime)s [%(levelname)s] - %(message)s')\n\n\nif __name__ == '__main__':\n\targumentParser = define_arguments()\n\tsettings = parse_arguments(argumentParser)\n\n\tbackup = S3Backup(num_threads=settings['threads'], config_file=settings['config_file'])\n\n\tstart_time = datetime.utcnow()\n\tlogging.info('START BACKUP ALL')\n\tbackup.backup()\n\tend_time = datetime.utcnow()\n\tlogging.info('END BACKUP ALL')\n\tlogging.info(\"Copied: %d keys, %s in %d seconds\" % (backup.total_keys, sizeof_fmt(backup.total_bytes), (end_time - start_time).total_seconds()))\n\n\tstart_time = datetime.utcnow()\n\tlogging.info('START PRUNE OLD ALL')\n\tbackup.delete_old_backups()\n\tend_time = datetime.utcnow()\n\tlogging.info('END PRUNE OLD ALL') \n\tlogging.info(\"Deleted: %d keys in %d seconds\" % (backup.total_keys_deleted, (end_time - start_time).total_seconds()))\n", "sub_path": "modules/mindtouch_s3_files_backup/files/mindtouch.s3.files.backup/src/backup.py", "file_name": "backup.py", "file_ext": "py", "file_size_in_byte": 13274, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.getLogger", "line_number": 24, "usage_type": "call"}, {"api_name": "logging.ERROR", "line_number": 24, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 25, "usage_type": "call"}, {"api_name": "logging.ERROR", "line_number": 25, "usage_type": "attribute"}, {"api_name": "Queue.Queue", "line_number": 42, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 44, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 44, "usage_type": "name"}, {"api_name": "pytz.timezone", "line_number": 44, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 45, "usage_type": "call"}, {"api_name": "ConfigParser.ConfigParser", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 67, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 88, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 93, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 98, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 100, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 102, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 105, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 110, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 113, "usage_type": "call"}, {"api_name": "boto.connect_s3", "line_number": 120, "usage_type": "call"}, {"api_name": "pytz.timezone", "line_number": 129, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 144, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 145, "usage_type": "call"}, {"api_name": "requests.auth.HTTPBasicAuth", "line_number": 145, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.fromstring", "line_number": 147, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 147, "usage_type": "name"}, {"api_name": "boto.connect_s3", "line_number": 184, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 192, "usage_type": "call"}, {"api_name": "boto.connect_s3", "line_number": 200, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 213, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 217, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 233, "usage_type": "call"}, {"api_name": "boto.connect_s3", "line_number": 237, "usage_type": "call"}, {"api_name": "pytz.timezone", "line_number": 250, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 253, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 254, "usage_type": "call"}, {"api_name": "pytz.timezone", "line_number": 267, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 270, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 277, "usage_type": "call"}, {"api_name": "boto.connect_s3", "line_number": 288, "usage_type": "call"}, {"api_name": "dateutil.parser.parse", "line_number": 299, "usage_type": "call"}, {"api_name": "dateutil.parser", "line_number": 299, "usage_type": "name"}, {"api_name": "datetime.datetime.utcfromtimestamp", "line_number": 301, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 301, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 313, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 334, "usage_type": "call"}, {"api_name": "os.path", "line_number": 334, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 334, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 336, "usage_type": "call"}, {"api_name": "os.path", "line_number": 336, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 356, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 360, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 370, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 370, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 371, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 373, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 373, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 374, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 375, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 377, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 377, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 378, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 380, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 380, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 381, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 382, "usage_type": "call"}]}
+{"seq_id": "408929384", "text": "import dash_core_components as dcc\nimport dash_bootstrap_components as dbc\nimport dash_html_components as html\nfrom dash.dependencies import Input, Output, State\nimport plotly.graph_objs as go\nimport plotly.express as px\nimport pandas as pd\nfrom app import app\n\nimport dateutil.relativedelta\nfrom datetime import date\n\n# Imports for Machine Learning\nimport numpy as np\nfrom pmdarima.arima import AutoARIMA\nfrom pmdarima.arima.utils import ndiffs\nfrom tqdm.notebook import tqdm\nfrom sklearn.metrics import mean_squared_error\nimport yfinance as yf\nfrom statsmodels.tools.eval_measures import rmse\nimport statsmodels.api as sm\nimport itertools\nfrom statsmodels.tsa.arima_model import ARIMA, ARMA\nimport warnings\nwarnings.filterwarnings(\"ignore\")\nimport time\n\n\n# This will create the card for ticker\ndef make_card(alert_message, color, style_dict={\"textAlign\": \"center\"}):\n return dbc.Card([ dbc.Alert(alert_message, color = color)], style = style_dict)\n\n# children=html.Div(id='loading-output'\n\nlayout = html.Div([\n html.Br(),\n html.H1('Train your Model!', style={\"textAlign\": \"center\"}),\n html.Br(),\n dbc.Row([dbc.Col(make_card(\"Search for a stock to train\", \"primary\"))]),\n html.Br(),\n html.Div([dcc.Input(id='input-box', placeholder='MSFT', type='text', style={\"textAlign\": \"center\"}),\n html.Br(),\n html.Br(),\n html.Div(html.Button('Train Model', id='button'), style={\"textAlign\": \"center\"}),\n html.Br(),\n dcc.Loading(id='loading-output', type='graph', fullscreen=True ,style={\"textAlign\": \"center\"})], style={\"textAlign\": \"center\"}),\n html.Br(),\n html.Div(id='projected-results', style={\"textAlign\": \"center\", 'marginLeft':'40%', 'marginRight':'40%','font-weight':'bold'}),\n html.Br(),\n html.Div(id='model-graph', style={\"textAlign\": \"center\"}),\n html.Br(),\n html.Div(id='model-summary-title', style={\"textAlign\": \"center\"}),\n html.Br(),\n html.Div(id='model-summary-0', style={\"textAlign\": \"center\",'marginLeft':'15%', 'marginRight':'15%'}),\n html.Br(),\n html.Div(id='model-summary-1', style={\"textAlign\": \"center\", 'whiteSpace': 'pre-wrap','marginLeft':'15%', 'marginRight':'15%'}),\n])\n\n@app.callback(\n [Output('model-graph', 'children'), \n Output('model-summary-0', 'children'), \n Output('model-summary-1', 'children'), \n Output('loading-output', 'children'), \n Output('model-summary-title', 'children'),\n Output('projected-results', 'children')],\n [Input('button', 'n_clicks')],\n [State('input-box', 'value')])\ndef update_output(n_clicks, value):\n ticker = value.upper()\n df = yf.Ticker(ticker).history(period='Max')\n # df.reset_index(inplace=True)\n df = df.filter(['Close'])\n\n # Define the p, d and q parameters to take any value between 0 and 3\n d_value = ndiffs(df['Close'], test='adf')\n d = range(0,d_value+1)\n p = range(0,4)\n q = range(0,3)\n # p = d = q = range(0,3)\n # Generate all different combinations of p, q and q\n pdq = list(itertools.product(p, d, q))\n\n warnings.filterwarnings(\"ignore\")\n aic= []\n parameters = []\n for param in pdq:\n #for param in pdq:\n try:\n mod = sm.tsa.statespace.SARIMAX(df, order=param, enforce_stationarity=True, enforce_invertibility=True)\n results = mod.fit()\n # save results in lists\n aic.append(results.aic)\n parameters.append(param)\n #seasonal_param.append(param_seasonal)\n # print('ARIMA{} - AIC:{}'.format(param, results.aic))\n except:\n continue\n # find lowest aic \n index_min = min(range(len(aic)), key=aic.__getitem__) \n\n print('The optimal model is: ARIMA{} -AIC{}'.format(parameters[index_min], aic[index_min]))\n\n model = ARIMA(df, order=parameters[index_min])\n model_fit = model.fit(disp=0)\n\n # Updating Indexed\n last_date = df.index[-1]\n days = 5\n for day in range(1, days+1):\n newEntry = last_date + pd.Timedelta(day, unit='D')\n forecastValue = model_fit.forecast(days)[0][day-1]\n dfSeries = df.append(pd.DataFrame({'Close': forecastValue}, index=[newEntry]))\n\n fig = go.Figure(data=[go.Scatter(x=dfSeries.index, y=dfSeries['Close'])])\n graph = dcc.Graph(figure=fig)\n\n # Print Summary for Website\n for i in range(2):\n if i == 0:\n html = model_fit.summary().tables[i].as_html()\n df_table = pd.read_html(html, header=0, index_col=0)[0]\n df_table = df_table.reset_index()\n # df_table = df_table.dropna\n table1 = dbc.Table.from_dataframe(df_table, striped=True, bordered=True, hover=True)\n else:\n html = model_fit.summary().tables[i].as_html()\n df_table = pd.read_html(html, header=0, index_col=0)[0]\n df_table = df_table.reset_index()\n table2 = dbc.Table.from_dataframe(df_table, striped=True, bordered=True, hover=True)\n \n # Create the forecasted values results to print\n days=[]\n predList = []\n for i in range(5):\n days.append(i+1)\n predValue = '$ ' + str(round(model_fit.forecast(5)[0][i],2))\n predList.append(predValue)\n\n df_pred_table = pd.DataFrame({'Number of Days in Future':days,\n 'Close Price': predList})\n\n table3 = dbc.Table.from_dataframe(df_pred_table, striped=True, bordered=True, hover=True, size='sm')\n\n empty_string = ''\n table_title = 'Arima Summary Results'\n return graph, table1, table2, empty_string, table_title, table3", "sub_path": "00. Archive/Sam/Test/apps/machineLearning.py", "file_name": "machineLearning.py", "file_ext": "py", "file_size_in_byte": 5518, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "warnings.filterwarnings", "line_number": 25, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.Card", "line_number": 31, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.Alert", "line_number": 31, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 35, "usage_type": "call"}, {"api_name": "dash_html_components.Br", "line_number": 36, "usage_type": "call"}, {"api_name": "dash_html_components.H1", "line_number": 37, "usage_type": "call"}, {"api_name": "dash_html_components.Br", "line_number": 38, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.Row", "line_number": 39, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.Col", "line_number": 39, "usage_type": "call"}, {"api_name": "dash_html_components.Br", "line_number": 40, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 41, "usage_type": "call"}, {"api_name": "dash_core_components.Input", "line_number": 41, "usage_type": "call"}, {"api_name": "dash_html_components.Br", "line_number": 42, "usage_type": "call"}, {"api_name": "dash_html_components.Br", "line_number": 43, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 44, "usage_type": "call"}, {"api_name": "dash_html_components.Button", "line_number": 44, "usage_type": "call"}, {"api_name": "dash_html_components.Br", "line_number": 45, "usage_type": "call"}, {"api_name": "dash_core_components.Loading", "line_number": 46, "usage_type": "call"}, {"api_name": "dash_html_components.Br", "line_number": 47, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 48, "usage_type": "call"}, {"api_name": "dash_html_components.Br", "line_number": 49, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 50, "usage_type": "call"}, {"api_name": "dash_html_components.Br", "line_number": 51, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 52, "usage_type": "call"}, {"api_name": "dash_html_components.Br", "line_number": 53, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 54, "usage_type": "call"}, {"api_name": "dash_html_components.Br", "line_number": 55, "usage_type": "call"}, {"api_name": "dash_html_components.Div", "line_number": 56, "usage_type": "call"}, {"api_name": "yfinance.Ticker", "line_number": 70, "usage_type": "call"}, {"api_name": "pmdarima.arima.utils.ndiffs", "line_number": 75, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 81, "usage_type": "call"}, {"api_name": "warnings.filterwarnings", "line_number": 83, "usage_type": "call"}, {"api_name": "statsmodels.api.tsa.statespace.SARIMAX", "line_number": 89, "usage_type": "call"}, {"api_name": "statsmodels.api.tsa", "line_number": 89, "usage_type": "attribute"}, {"api_name": "statsmodels.api", "line_number": 89, "usage_type": "name"}, {"api_name": "statsmodels.tsa.arima_model.ARIMA", "line_number": 103, "usage_type": "call"}, {"api_name": "pandas.Timedelta", "line_number": 110, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 112, "usage_type": "call"}, {"api_name": "plotly.graph_objs.Figure", "line_number": 114, "usage_type": "call"}, {"api_name": "plotly.graph_objs", "line_number": 114, "usage_type": "name"}, {"api_name": "plotly.graph_objs.Scatter", "line_number": 114, "usage_type": "call"}, {"api_name": "dash_core_components.Graph", "line_number": 115, "usage_type": "call"}, {"api_name": "pandas.read_html", "line_number": 121, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.Table.from_dataframe", "line_number": 124, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.Table", "line_number": 124, "usage_type": "attribute"}, {"api_name": "pandas.read_html", "line_number": 127, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.Table.from_dataframe", "line_number": 129, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.Table", "line_number": 129, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 139, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.Table.from_dataframe", "line_number": 142, "usage_type": "call"}, {"api_name": "dash_bootstrap_components.Table", "line_number": 142, "usage_type": "attribute"}, {"api_name": "app.app.callback", "line_number": 59, "usage_type": "call"}, {"api_name": "app.app", "line_number": 59, "usage_type": "name"}, {"api_name": "dash.dependencies.Output", "line_number": 60, "usage_type": "call"}, {"api_name": "dash.dependencies.Output", "line_number": 61, "usage_type": "call"}, {"api_name": "dash.dependencies.Output", "line_number": 62, "usage_type": "call"}, {"api_name": "dash.dependencies.Output", "line_number": 63, "usage_type": "call"}, {"api_name": "dash.dependencies.Output", "line_number": 64, "usage_type": "call"}, {"api_name": "dash.dependencies.Output", "line_number": 65, "usage_type": "call"}, {"api_name": "dash.dependencies.Input", "line_number": 66, "usage_type": "call"}, {"api_name": "dash.dependencies.State", "line_number": 67, "usage_type": "call"}]}
+{"seq_id": "286348548", "text": "import scrapy\nfrom caijiSpider.items import MovieItem\n\nimport json, re\n\nclass DoubanMovie(scrapy.Spider):\n\t\n\tname = 'douban_movie'\n\n\theaders = {\n\t\t'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/53.0.2785.143 Safari/537.36'\n\t}\n\tstart_url = 'https://movie.douban.com/j/new_search_subjects?sort=R&range=0,10&tags=&start=0'\n\n\tdef start_requests(self):\n\t\tyield scrapy.Request(self.start_url, headers=self.headers)\n\n\tdef parse(self, response):\n\t\titem = MovieItem()\n\n\t\tdata = response.body.decode('utf8')\n\n\t\tif data:\n\t\t\tmovies = json.loads(data)['data']\n\n\t\t\tfor movie in movies:\n\t\t\t\titem['movie_name'] = movie['title']\n\t\t\t\titem['score'] = movie['rate']\n\t\t\t\titem['director'] = movie['directors']\n\n\t\t\t\tyield item\n\n\t\t\tpage_num = re.search(r'start=(\\d+)', response.url).group(1)\n\t\t\tpage_num = 'start=' + str(int(page_num) + 20)\n\t\t\tnext_url = re.sub(r'start=\\d+', page_num, response.url)\n\t\t\tyield scrapy.Request(next_url, headers=self.headers)", "sub_path": "ScrapyLearn/caijiSpider/caijiSpider/spiders/douban_movie.py", "file_name": "douban_movie.py", "file_ext": "py", "file_size_in_byte": 986, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "scrapy.Spider", "line_number": 6, "usage_type": "attribute"}, {"api_name": "scrapy.Request", "line_number": 16, "usage_type": "call"}, {"api_name": "caijiSpider.items.MovieItem", "line_number": 19, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 24, "usage_type": "call"}, {"api_name": "re.search", "line_number": 33, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 35, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 36, "usage_type": "call"}]}
+{"seq_id": "411362833", "text": "from pyspark import SparkConf\nfrom pyspark.sql import SparkSession\nimport traceback\n\nimport os\nos.environ['PYSPARK_PYTHON']='/Library/Frameworks/Python.framework/Versions/3.7/bin/python3'\n\nappname = \"test\" # 任务名称\nmaster = \"local\" # 单机模式设置\n'''\nlocal: 所有计算都运行在一个线程当中,没有任何并行计算,通常我们在本机执行一些测试代码,或者练手,就用这种模式。\nlocal[K]: 指定使用几个线程来运行计算,比如local[4]就是运行4个worker线程。通常我们的cpu有几个core,就指定几个线程,最大化利用cpu的计算能力\nlocal[*]: 这种模式直接帮你按照cpu最多cores来设置线程数了。\n'''\nspark_driver_host = \"10.0.0.248\"\n\ntry:\n # conf = SparkConf().setAppName(appname).setMaster(master).set(\"spark.driver.host\", spark_driver_host) # 集群\n conf = SparkConf().setAppName(appname).setMaster(master) # 本地\n spark = SparkSession.builder.config(conf=conf).getOrCreate()\n sc = spark.sparkContext\n words = sc.parallelize(\n [\"scala\",\n \"java\",\n \"hadoop\",\n \"spark\",\n \"akka\",\n \"spark vs hadoop\",\n \"pyspark\",\n \"pyspark and spark\"\n ])\n counts = words.count()\n print(\"Number of elements in RDD is %i\" % counts)\n sc.stop()\n print('计算成功!')\nexcept:\n sc.stop()\n traceback.print_exc() # 返回出错信息\n print('连接出错!')", "sub_path": "python/spark/spark_run.py", "file_name": "spark_run.py", "file_ext": "py", "file_size_in_byte": 1448, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.environ", "line_number": 6, "usage_type": "attribute"}, {"api_name": "pyspark.SparkConf", "line_number": 19, "usage_type": "call"}, {"api_name": "pyspark.sql.SparkSession.builder.config", "line_number": 20, "usage_type": "call"}, {"api_name": "pyspark.sql.SparkSession.builder", "line_number": 20, "usage_type": "attribute"}, {"api_name": "pyspark.sql.SparkSession", "line_number": 20, "usage_type": "name"}, {"api_name": "traceback.print_exc", "line_number": 38, "usage_type": "call"}]}
+{"seq_id": "144970215", "text": "#encoding:utf-8\r\nimport sys\r\nimport datetime\r\nimport os\r\n\r\n# @version 1.0\r\n# @author libin\r\n# @function 该脚本的主要作用是按天从impala数据库当中拉取数据\r\ndef main():\r\n start_date = '2017-01-01'\r\n start_date = datetime.datetime.strptime(start_date,'%Y-%m-%d')\r\n end_date = '2017-07-03'\r\n end_date = datetime.datetime.strptime(end_date,'%Y-%m-%d')\r\n next_date = start_date\r\n while (end_date - next_date).days > 0:\r\n string_date = next_date.strftime('%Y-%m-%d')\r\n load_data(string_date)\r\n\r\n next_date = next_date + datetime.timedelta(days=1)\r\n\r\ndef load_data(data):\r\n string = \"\"\"\r\n impala-shell -i 10.30.1.88 -k -B -q \"\r\n set request_pool=root.bi.da;\r\n select * from tlbb3d_audit.audit_chat_info where dt='{date}'; \" -o ./audit_data/audit_{date}.txt --print_header --output_delimiter=',' \"\"\"\r\n\r\n os.system(string)", "sub_path": "src/audit/get_data_from_impala.py", "file_name": "get_data_from_impala.py", "file_ext": "py", "file_size_in_byte": 886, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 11, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 11, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 13, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 13, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 19, "usage_type": "call"}, {"api_name": "os.system", "line_number": 27, "usage_type": "call"}]}
+{"seq_id": "463746587", "text": "import pickle\n\nfrom flask import flash, render_template, request\nfrom flask_login.utils import login_required\nfrom sklearn.ensemble import RandomForestClassifier, VotingClassifier\nfrom sklearn.linear_model import LogisticRegression\n \n\nfrom app import app\nfrom app.forms import PatientInfo\n\nensemble = pickle.load(open(\"ensemble.pickle\", \"rb\"))\n\n\n@app.route(\"/patient_info\", methods=[\"GET\", \"POST\"])\n@login_required\ndef patient_info():\n form = PatientInfo()\n if request.method == \"POST\":\n data = [\n [\n form.age.data,\n form.gender.data,\n form.chest_pain.data,\n form.resting_bp.data,\n form.sereum_cholestoral.data,\n form.fasting_blood_sugar.data,\n form.resting_ecg.data,\n form.max_heart_rate.data,\n form.exercise_induced_angina.data,\n form.oldpeak.data,\n form.slope.data,\n form.ca.data,\n form.thal.data,\n ]\n ]\n try:\n result = ensemble.predict(data)\n if result[0] == 0:\n flash(\"The patient doesn't have Heart Disease\", \"info\")\n else:\n flash(\"The patient has Heart Disease\", \"info\")\n except Exception as e:\n print(e)\n flash(\"Something went wrong\", \"danger\")\n return render_template(\"patient_info.html\", title=\"Patient Info - HDP\", form=form)\n", "sub_path": "app/routes/patient_info.py", "file_name": "patient_info.py", "file_ext": "py", "file_size_in_byte": 1476, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pickle.load", "line_number": 12, "usage_type": "call"}, {"api_name": "app.forms.PatientInfo", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 19, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 19, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 40, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 42, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 45, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 46, "usage_type": "call"}, {"api_name": "app.app.route", "line_number": 15, "usage_type": "call"}, {"api_name": "app.app", "line_number": 15, "usage_type": "name"}, {"api_name": "flask_login.utils.login_required", "line_number": 16, "usage_type": "name"}]}
+{"seq_id": "287104893", "text": "from netCDF4 import Dataset\nimport numpy as np\nimport os\nimport sys\n\ndef readMat(directory, filename):\n\n ncfile = os.path.join(directory, filename)\n if not os.path.isfile(ncfile):\n sys.exit('File not found ' +ncfile + \". Exiting.\")\n print('Reading ' + ncfile)\n\n # Load dataset\n dset = Dataset(ncfile)\n\n # Extract data from NetCDF file\n mat = np.array(dset.variables['mat'][:])\n dset.close()\n print('Size of matrix ' + str(mat.shape))\n \n return mat\n", "sub_path": "common/io/readMat.py", "file_name": "readMat.py", "file_ext": "py", "file_size_in_byte": 488, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.path.join", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 10, "usage_type": "call"}, {"api_name": "netCDF4.Dataset", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 17, "usage_type": "call"}]}
+{"seq_id": "322024743", "text": "import os\nimport numpy as np\nfrom dotmap import DotMap\nfrom functools import partial\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nfrom types import SimpleNamespace\nfrom lucent.optvis import transform, param, render\n\nfrom celebalucid.models.inceptionv1 import InceptionV1\nfrom celebalucid.utils import load_layer_info\nfrom celebalucid import base_url\n\n\nclass ModelManipulator(InceptionV1):\n def __init__(self, pt_url):\n self.bn = '-bn' in pt_url\n super(ModelManipulator, self).__init__(n_features=40,\n pretrained=False,\n redirected_ReLU=True,\n bn=self.bn)\n self._load_weights_from_url(pt_url)\n self.layer_info = load_layer_info()\n self.weights = self._load_weights()\n self.neurons = {}\n self._activations = {}\n self._register_activation_fw_hooks()\n\n # Assign device\n self.device = torch.device(\"cuda:0\" if torch.cuda.is_available() else \"cpu\")\n self.to(self.device).eval()\n\n def switch_to(self, str_model):\n url = os.path.join(base_url, str_model+'.pt')\n self._load_weights_from_url(url)\n self.weights = self._load_weights()\n\n def lucid(self, layer_n_channel, size=224, thresholds=[512], progress=False):\n layer_n_channel = self._correct_layer_n_channel(layer_n_channel)\n transforms = transform.standard_transforms.copy()\n transforms.append(lambda x: x * 255 - 117)\n def param_f(): return param.image(224, fft=True, decorrelate=True)\n try:\n img = render.render_vis(self,\n layer_n_channel,\n show_image=False,\n preprocess=False,\n progress=progress,\n transforms=transforms,\n thresholds=thresholds,\n param_f=param_f)[0][0]\n except AssertionError:\n raise AssertionError(\n 'Invalid layer {}. Retrieve the list of layers with `model.layer_info`.'.format(\n layer_n_channel)\n )\n\n img = (img*255).astype(np.uint8)\n return img\n\n def stream(self, x):\n x = x.to(self.device)\n self.forward(x)\n self.neurons = DotMap(self._activations)\n\n def set_weights(self, layer, reference, mode='both'):\n valid_modes = ['both', 'weight', 'bias']\n if mode not in valid_modes:\n raise ValueError('Invalid mode: {}. Please provide a mode from {}'\\\n .format(mode, ', '.join(valid_modes)))\n if mode == 'both':\n self.set_weights(layer, reference, 'weight')\n self.set_weights(layer, reference, 'bias')\n return\n layer = self._correct_layer_n_channel(layer)\n layer, neuron_i = self._split_layer_name(layer)\n targets = self._extract_target_weights(reference, layer, neuron_i, mode)\n state_dict = self.state_dict()\n new_state_dict = self._change_state_dict(state_dict, layer,\n neuron_i, targets, mode)\n self.load_state_dict(new_state_dict)\n\n\n # ========================================================================\n # Private functions\n # ========================================================================\n\n def _extract_target_weights(self, reference, layer, i, mode='weight'):\n if type(reference) == type(self):\n target = self._get_layer(reference, layer, i, mode)\n else:\n baseline = self._get_layer(self, layer, i, mode)\n target = torch.ones_like(baseline, dtype=baseline.dtype) * reference\n return target\n \n\n def _change_state_dict(self, state_dict, layer, i, value, mode='weight'):\n if i is None:\n state_dict[layer+'.'+mode] = value\n else:\n state_dict[layer+'.'+mode][i] = value\n return state_dict\n\n def _get_layer(self, model, layer, i, mode='weight'):\n states = model.state_dict()\n layer = states[layer+'.'+mode]\n if i is not None:\n layer = layer[i]\n return layer\n\n def _split_layer_name(self, layer):\n if ':' not in layer:\n return layer, None\n else:\n layer, neuron_i = layer.split(':')\n neuron_i = int(neuron_i)\n return layer, neuron_i\n\n def _correct_layer_n_channel(self, layer_n_channel):\n if ':' in layer_n_channel:\n layer, channel = layer_n_channel.split(':')\n layer += '_pre_relu_conv' if layer != 'logits' else ''\n return ':'.join([layer, channel])\n else:\n layer = layer_n_channel\n layer += '_pre_relu_conv' if layer != 'logits' else ''\n return layer\n\n def _register_activation_fw_hooks(self):\n for name, module in self.named_modules():\n if name.endswith('_pre_relu_conv') or name == 'logits':\n name = name.replace('_pre_relu_conv', '')\n args = partial(self._save_activation, name)\n module.register_forward_hook(args)\n\n def _save_activation(self, name, module, m_in, m_out):\n self._activations[name] = m_out\n\n def _load_weights(self):\n extracted_weights = DotMap()\n for name, weights in self.named_parameters():\n if \"_bn\" in name:\n continue\n layer_name, weight_format = name.split('.')\n weight_format = weight_format[0]\n layer_name = layer_name.replace('_pre_relu_conv', '')\n extracted_weights[layer_name][weight_format] = weights.detach(\n ).cpu().numpy()\n\n return extracted_weights\n\n def _load_weights_from_url(self, url):\n self.load_state_dict(torch.hub.load_state_dict_from_url(\n url, progress=True), strict=False)\n", "sub_path": "celebalucid/models/manipulator.py", "file_name": "manipulator.py", "file_ext": "py", "file_size_in_byte": 6057, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "celebalucid.models.inceptionv1.InceptionV1", "line_number": 16, "usage_type": "name"}, {"api_name": "celebalucid.utils.load_layer_info", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 31, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "celebalucid.base_url", "line_number": 35, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "lucent.optvis.transform.standard_transforms.copy", "line_number": 41, "usage_type": "call"}, {"api_name": "lucent.optvis.transform.standard_transforms", "line_number": 41, "usage_type": "attribute"}, {"api_name": "lucent.optvis.transform", "line_number": 41, "usage_type": "name"}, {"api_name": "lucent.optvis.param.image", "line_number": 43, "usage_type": "call"}, {"api_name": "lucent.optvis.param", "line_number": 43, "usage_type": "name"}, {"api_name": "lucent.optvis.render.render_vis", "line_number": 45, "usage_type": "call"}, {"api_name": "lucent.optvis.render", "line_number": 45, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 59, "usage_type": "attribute"}, {"api_name": "dotmap.DotMap", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.ones_like", "line_number": 94, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 134, "usage_type": "call"}, {"api_name": "dotmap.DotMap", "line_number": 141, "usage_type": "call"}, {"api_name": "torch.hub.load_state_dict_from_url", "line_number": 154, "usage_type": "call"}, {"api_name": "torch.hub", "line_number": 154, "usage_type": "attribute"}]}
+{"seq_id": "322074815", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nimport scipy.optimize\nfrom linregress import linregress\nfrom uncertainties import ufloat\n\n#Emissionvermögen als Steigung von Udata über Tdiff\nT,weiß,kupfer,schwarz,spiegel=np.genfromtxt('data.txt',unpack=True)\nxdata,Udata=np.genfromtxt('dataStrahlungsintensität.txt',unpack=True)\n\n#OFFSET U\nUa=0.0146*10**-3\nUend=0.0095*10**-3\nUoffset=(np.mean([Ua,Uend]))\nprint(Uoffset,'Uoffset')\n\nUdata*=10**-3\nweiß=weiß*10**-3-Uoffset\nkupfer=kupfer*10**-3-Uoffset\nschwarz=schwarz*10**-3-Uoffset\nprint(schwarz,'schwarz')\nspiegel=spiegel*10**-3-Uoffset\n\nT_0=273.15+21.2\nT=273.15+T\nprint(T**4,'\\n',T_0**4)\nTdiff=T**4-T_0**4\nprint(Tdiff,'Tdiff')\n\n\n# print(Tdiff)\n\n#WEIẞ\nwepsi,wepsierr,wb,wberr=linregress(Tdiff,weiß)\n\nq=np.linspace(0,1.2e10,100)\n\nplt.plot(Tdiff,weiß,'kx',label='Messwerte')\nplt.plot(q,wepsi*q+wb,'k-',label='Ausgleichsgerade')\nplt.xlabel(r'$T^4-\\,T_0^4 \\ / \\ \\mathrm{K^4}$')\nplt.ylabel(r'$U \\ / \\ \\mathrm{V}$')\nplt.legend(loc='best')\nplt.tight_layout()\nplt.savefig('plotWEISS.pdf')\nplt.close()\n\n#KUPFER\nkepsi,kepsierr,kb,kberr=linregress(Tdiff,kupfer)\n\nplt.plot(Tdiff,kupfer,'kx',label='Messwerte')\nplt.plot(q,kepsi*q+kb,'k-',label='Ausgleichsgerade')\nplt.xlabel(r'$T^4-\\,T_0^4 \\ / \\ \\mathrm{K^4}$')\nplt.ylabel(r'$U \\ / \\ \\mathrm{V}$')\nplt.legend(loc='best')\nplt.tight_layout()\nplt.savefig('plotKUPFER.pdf')\nplt.close()\n\n#SCHWARZ\nsepsi,sepsierr,sb,sberr=linregress(Tdiff,schwarz)\n\nplt.plot(Tdiff,schwarz,'kx',label='Messwerte')\nplt.plot(q,sepsi*q+sb,'k-',label='Ausgleichsgerade')\nplt.xlabel(r'$T^4-\\,T_0^4 \\ / \\ \\mathrm{K^4}$')\nplt.ylabel(r'$U \\ / \\ \\mathrm{V}$')\nplt.legend(loc='best')\nplt.tight_layout()\nplt.savefig('plotSCHWARZ.pdf')\nplt.close()\n\n#SPIEGEL\nspepsi,spepsierr,spb,spberr=linregress(Tdiff,spiegel)\n\nplt.plot(Tdiff,spiegel,'kx',label='Messwerte')\nplt.plot(q,spepsi*q+spb,'k-',label='Ausgleichsgerade')\nplt.xlabel(r'$T^4-\\,T_0^4 \\ / \\ \\mathrm{K^4}$')\nplt.ylabel(r'$U \\ / \\ \\mathrm{V}$')\nplt.legend(loc='best')\nplt.tight_layout()\nplt.savefig('plotSPIEGEL.pdf')\nplt.close()\n\n#EMISSIONVERMÖGEN\n\nprint('sepsi,sepsierr',sepsi,sepsierr)\nprint('wepsi,wepsierr',wepsi,wepsierr)\nprint('spepsi,spepsierr',spepsi,spepsierr)\nprint('kepsi,kepsierr',kepsi,kepsierr)\n\nSepsi=ufloat(sepsi,sepsierr)\nWepsi=ufloat(wepsi,wepsierr)\nSPepsi=ufloat(spepsi,spepsierr)\nKepsi=ufloat(kepsi,kepsierr)\n\ndef dreisatz(x):\n return x/Sepsi\n\nprint( 'schwarz',1,'\\n',\n dreisatz(Wepsi),'\\n',\n dreisatz(SPepsi),'\\n',\n dreisatz(Kepsi),'\\n',\n )\n\n\n# #ABSTAND STRAHLUNGSINTENSITÄT\n# x,xerr,U,Uerr=linregress(xdata,Udata)\n#\n# q=np.linspace(0,12,100)\n# plt.plot(xdata,Udata, 'kx', label='Messwerte')\n# plt.plot(q,x*q+U,'k-',label='Ausgleichsgerade')\n# plt.xlabel(r'$r \\ / \\ \\mathrm{cm}$')\n# plt.ylabel(r'$U \\ / \\ \\mathrm{V}$')\n# plt.legend(loc='best')\n# plt.tight_layout()\n# plt.savefig('plot1.pdf')\n# plt.close()\n\n#ABSTAND STRAHLUNGSINTENSIT NEU\nfrom scipy.optimize import curve_fit\n\ndef theory(x,a,b):\n return a/x**2 + b\n\na,b= curve_fit(theory,xdata,Udata)\nprint(a,'\\n',b,'A,B')\n\nq=np.linspace(0,12,100)\nplt.plot(q,theory(q,*a), 'k-', label='Ausgleichsgerade')\nplt.plot(xdata,Udata,'kx',label='Messwerte')\nplt.xlabel(r'$\\frac{1}{r^2} \\ / \\ \\mathrm{cm^{-2}}$')\nplt.ylabel(r'$U \\ / \\ \\mathrm{V}$')\nplt.legend(loc='best')\nplt.tight_layout()\nplt.savefig('plot1neu.pdf')\nplt.close()\n\n\n\n\n\n#plots zu u über t\nq=np.linspace(300,380,100)\n\nwepsi,wepsierr,wb,wberr=linregress(T,weiß)\nprint('wepsi,wepsierr',wepsi,wepsierr)\nplt.plot(T,weiß,'kx',label='Messwerte')\nplt.plot(q,wepsi*q+wb,'k-',label='Ausgleichsgerade')\nplt.xlabel(r'$T \\ / \\ \\mathrm{K}$')\nplt.ylabel(r'$U \\ / \\ \\mathrm{V}$')\nplt.legend(loc='best')\nplt.tight_layout()\nplt.savefig('plotWEISS123.pdf')\nplt.close()\n\nsepsi,sepsierr,sb,sberr=linregress(T,schwarz)\nprint('sepsi,sepsierr',sepsi,sepsierr)\nplt.plot(T,schwarz,'kx',label='Messwerte')\nplt.plot(q,sepsi*q+sb,'k-',label='Ausgleichsgerade')\nplt.xlabel(r'$T \\ / \\ \\mathrm{K}$')\nplt.ylabel(r'$U \\ / \\ \\mathrm{V}$')\nplt.legend(loc='best')\nplt.tight_layout()\nplt.savefig('plotSCHWARZ123.pdf')\nplt.close()\n\nkepsi,kepsierr,kb,kberr=linregress(T,kupfer)\nprint('kepsi,kepsierr',kepsi,kepsierr)\nplt.plot(T,kupfer,'kx',label='Messwerte')\nplt.plot(q,kepsi*q+kb,'k-',label='Ausgleichsgerade')\nplt.xlabel(r'$T \\ / \\ \\mathrm{K}$')\nplt.ylabel(r'$U \\ / \\ \\mathrm{V}$')\nplt.legend(loc='best')\nplt.tight_layout()\nplt.savefig('plotKUPFER123.pdf')\nplt.close()\n\nspepsi,spepsierr,spb,spberr=linregress(T,spiegel)\nprint('spepsi,spepsierr',spepsi,spepsierr)\nplt.plot(T,spiegel,'kx',label='Messwerte')\nplt.plot(q,spepsi*q+spb,'k-',label='Ausgleichsgerade')\nplt.xlabel(r'$T \\ / \\ \\mathrm{K}$')\nplt.ylabel(r'$U \\ / \\ \\mathrm{V}$')\nplt.legend(loc='best')\nplt.tight_layout()\nplt.savefig('plotSPIEGEL123.pdf')\nplt.close()\n", "sub_path": "Protokolle/Das_Stefan-Boltzmann_Gesetz/latex-template/Auswertung/auswertung.py", "file_name": "auswertung.py", "file_ext": "py", "file_size_in_byte": 4769, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "numpy.genfromtxt", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.genfromtxt", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 14, "usage_type": "call"}, {"api_name": "linregress.linregress", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "linregress.linregress", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "linregress.linregress", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "linregress.linregress", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "uncertainties.ufloat", "line_number": 90, "usage_type": "call"}, {"api_name": "uncertainties.ufloat", "line_number": 91, "usage_type": "call"}, {"api_name": "uncertainties.ufloat", "line_number": 92, "usage_type": "call"}, {"api_name": "uncertainties.ufloat", "line_number": 93, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 142, "usage_type": "call"}, {"api_name": "linregress.linregress", "line_number": 144, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 146, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 147, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 148, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 148, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 149, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 149, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 151, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 152, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 152, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 153, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 153, "usage_type": "name"}, {"api_name": "linregress.linregress", "line_number": 155, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 157, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 158, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 158, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 159, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 159, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 160, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 161, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 161, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 162, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 162, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 163, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 163, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 164, "usage_type": "name"}, {"api_name": "linregress.linregress", "line_number": 166, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 168, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 168, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 169, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 169, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 170, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 170, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 171, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 171, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 172, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 172, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 173, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 173, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 174, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 174, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 175, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 175, "usage_type": "name"}, {"api_name": "linregress.linregress", "line_number": 177, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 179, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 179, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 180, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 181, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 181, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 182, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 184, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 185, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 185, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 186, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 186, "usage_type": "name"}]}
+{"seq_id": "404554170", "text": "from typing import Iterable\n\nfrom catalyst.metrics._topk_metric import TopKMetric\nfrom catalyst.metrics.functional._hitrate import hitrate\n\n\nclass HitrateMetric(TopKMetric):\n \"\"\"Calculates the hitrate.\n\n Args:\n topk: list of `topk` for hitrate@topk computing\n compute_on_call: if True, computes and returns metric value during metric call\n prefix: metric prefix\n suffix: metric suffix\n\n Compute mean value of hitrate and it's approximate std value.\n\n Examples:\n\n .. code-block:: python\n\n import torch\n from catalyst import metrics\n\n outputs = torch.Tensor([[4.0, 2.0, 3.0, 1.0], [1.0, 2.0, 3.0, 4.0]])\n targets = torch.Tensor([[0, 0, 1.0, 1.0], [0, 0, 0.0, 0.0]])\n metric = metrics.HitrateMetric(topk=[1, 2, 3, 4])\n metric.reset()\n\n metric.update(outputs, targets)\n metric.compute()\n # (\n # (0.0, 0.25, 0.25, 0.5), # mean for @01, @02, @03, @04\n # (0.0, 0.0, 0.0, 0.0) # std for @01, @02, @03, @04\n # )\n\n metric.compute_key_value()\n # {\n # 'hitrate01': 0.0,\n # 'hitrate01/std': 0.0,\n # 'hitrate02': 0.25,\n # 'hitrate02/std': 0.0,\n # 'hitrate03': 0.25,\n # 'hitrate03/std': 0.0,\n # 'hitrate04': 0.5,\n # 'hitrate04/std': 0.0\n # }\n\n metric.reset()\n metric(outputs, targets)\n # (\n # (0.0, 0.25, 0.25, 0.5), # mean for @01, @02, @03, @04\n # (0.0, 0.0, 0.0, 0.0) # std for @01, @02, @03, @04\n # )\n\n .. code-block:: python\n\n import torch\n from torch.utils.data import DataLoader, TensorDataset\n from catalyst import dl\n\n # sample data\n num_users, num_features, num_items = int(1e4), int(1e1), 10\n X = torch.rand(num_users, num_features)\n y = (torch.rand(num_users, num_items) > 0.5).to(torch.float32)\n\n # pytorch loaders\n dataset = TensorDataset(X, y)\n loader = DataLoader(dataset, batch_size=32, num_workers=1)\n loaders = {\"train\": loader, \"valid\": loader}\n\n # model, criterion, optimizer, scheduler\n model = torch.nn.Linear(num_features, num_items)\n criterion = torch.nn.BCEWithLogitsLoss()\n optimizer = torch.optim.Adam(model.parameters())\n scheduler = torch.optim.lr_scheduler.MultiStepLR(optimizer, [2])\n\n # model training\n runner = dl.SupervisedRunner(\n input_key=\"features\",\n output_key=\"logits\",\n target_key=\"targets\",\n loss_key=\"loss\"\n )\n runner.train(\n model=model,\n criterion=criterion,\n optimizer=optimizer,\n scheduler=scheduler,\n loaders=loaders,\n num_epochs=3,\n verbose=True,\n callbacks=[\n dl.BatchTransformCallback(\n transform=torch.sigmoid,\n scope=\"on_batch_end\",\n input_key=\"logits\",\n output_key=\"scores\"\n ),\n dl.CriterionCallback(\n input_key=\"logits\", target_key=\"targets\", metric_key=\"loss\"\n ),\n dl.AUCCallback(input_key=\"scores\", target_key=\"targets\"),\n dl.HitrateCallback(\n input_key=\"scores\", target_key=\"targets\", topk=(1, 3, 5)\n ),\n dl.MRRCallback(input_key=\"scores\", target_key=\"targets\", topk=(1, 3, 5)),\n dl.MAPCallback(input_key=\"scores\", target_key=\"targets\", topk=(1, 3, 5)),\n dl.NDCGCallback(input_key=\"scores\", target_key=\"targets\", topk=(1, 3)),\n dl.OptimizerCallback(metric_key=\"loss\"),\n dl.SchedulerCallback(),\n dl.CheckpointCallback(\n logdir=\"./logs\", loader_key=\"valid\", metric_key=\"loss\", minimize=True\n ),\n ]\n )\n\n .. note::\n Metric names depending on input parameters:\n\n - ``topk = (1,) or None`` ---> ``\"hitrate01\"``\n - ``topk = (1, 3)`` ---> ``\"hitrate01\"``, ``\"hitrate03\"``\n - ``topk = (1, 3, 5)`` ---> ``\"hitrate01\"``, ``\"hitrate03\"``, ``\"hitrate05\"``\n\n You can find them in ``runner.batch_metrics``, ``runner.loader_metrics`` or\n ``runner.epoch_metrics``.\n\n .. note::\n Please follow the `minimal examples`_ sections for more use cases.\n\n .. _`minimal examples`: https://github.com/catalyst-team/catalyst#minimal-examples # noqa: E501, W505\n \"\"\"\n\n def __init__(\n self,\n topk: Iterable[int] = None,\n compute_on_call: bool = True,\n prefix: str = None,\n suffix: str = None,\n ):\n \"\"\"Init HitrateMetric\"\"\"\n super().__init__(\n metric_name=\"hitrate\",\n metric_function=hitrate,\n topk=topk,\n compute_on_call=compute_on_call,\n prefix=prefix,\n suffix=suffix,\n )\n\n\n__all__ = [\"HitrateMetric\"]\n", "sub_path": "catalyst/metrics/_hitrate.py", "file_name": "_hitrate.py", "file_ext": "py", "file_size_in_byte": 5050, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "catalyst.metrics._topk_metric.TopKMetric", "line_number": 7, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 136, "usage_type": "name"}, {"api_name": "catalyst.metrics.functional._hitrate.hitrate", "line_number": 144, "usage_type": "name"}]}
+{"seq_id": "18887018", "text": "import xlrd\n\n\ndef getWorkbookSheet(source, index=0):\n wb = xlrd.open_workbook(source)\n return wb.sheet_by_index(index)\n\n\ndef getCategoryIndex(allCategories, category):\n return allCategories.index(category)\n\n\ndef getCellValue(sheet, row, category):\n if isinstance(category, str):\n index = getCategoryIndex(sheet.row_values(0), category)\n else:\n index = category\n return sheet.cell_value(row, index)\n\n\ndef createDictionary(sheet, keyColumns, valueColumns, keyFilter=None, keyCombiner=None, valueCombiner=None, valueUpdater=None, valueFinalizer=None):\n ret = {}\n for i in range(1, sheet.nrows):\n keyList = []\n for k in keyColumns:\n keyList.append(getCellValue(sheet, i, k))\n if keyFilter is not None:\n if not keyFilter(keyList):\n continue\n if keyCombiner is not None:\n key = keyCombiner(keyList)\n else:\n key = tuple(keyList)\n\n valueList = []\n for v in valueColumns:\n valueList.append(getCellValue(sheet, i, v))\n if valueCombiner is not None:\n value = valueCombiner(valueList)\n else:\n value = valueList\n\n if key in ret.keys():\n if valueUpdater is None:\n newVal = ret.get(key)\n for j in range(len(newVal)):\n newVal[j] = newVal[j] + value[j]\n else:\n newVal = valueUpdater(value, ret[key])\n ret[key] = newVal\n else:\n if valueUpdater is None:\n ret[key] = value\n else:\n ret[key] = valueUpdater(value, None)\n if valueFinalizer is not None:\n for k in ret.keys():\n ret[k] = valueFinalizer(ret[k])\n return ret\n", "sub_path": "XLSXReader.py", "file_name": "XLSXReader.py", "file_ext": "py", "file_size_in_byte": 1777, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "xlrd.open_workbook", "line_number": 5, "usage_type": "call"}]}
+{"seq_id": "401571042", "text": "#!/usr/bin/env python3\n\"\"\"\nAuthor : edwin\nDate : 2020-02-05\nPurpose: Homework1\n\"\"\"\n\nimport argparse\nimport os\nimport sys\n\n\n# --------------------------------------------------\ndef get_args():\n \"\"\"Get command-line arguments\"\"\"\n\n parser = argparse.ArgumentParser(\n description='Find position of vowel in string',\n formatter_class=argparse.ArgumentDefaultsHelpFormatter)\n#order in which they are declared matters\n parser.add_argument('vowel',\n metavar='vowel',\n help='A vowel to look for',\n choices = ['a', 'e', 'i', 'o', 'u', 'A', 'E', 'I', 'O', 'U'])\n#couldve written this as choices = ['aeiouAEIOU']\n parser.add_argument('text',\n metavar='text',\n help='The text to search')\n\n\n return parser.parse_args()\n\n\n# --------------------------------------------------\ndef main():\n \"\"\"Make a jazz noise here\"\"\"\n\n args = get_args()\n vowel = args.vowel\n text = args.text\n index = text.find(vowel)\n\n if vowel in text:\n print(f'Found \"{vowel}\" in \"{text}\" at index {index}.')\n else:\n print(f'\"{vowel}\" is not found in \"{text}\".')\n\n\n# --------------------------------------------------\nif __name__ == '__main__':\n main()\n", "sub_path": "assignments/01_strings/vpos.py", "file_name": "vpos.py", "file_ext": "py", "file_size_in_byte": 1290, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 17, "usage_type": "call"}, {"api_name": "argparse.ArgumentDefaultsHelpFormatter", "line_number": 19, "usage_type": "attribute"}]}
+{"seq_id": "159154949", "text": "import tkinter as tk\nimport MusEEG\nfrom MusEEG import Processor, eegData, parentDir\nfrom tkinter import filedialog\nfrom tkinter.scrolledtext import ScrolledText\nimport sys\nimport os\n\nimport queue\n\nimport matplotlib\nmatplotlib.use(\"TkAgg\")\nfrom matplotlib.backends.backend_tkagg import FigureCanvasTkAgg\nimport matplotlib.pyplot as plt\nimport matplotlib.animation as animation\n\nprocessor = Processor()\nprocessor.OSCstart()\nprocessor.defineOSCMessages()\nprocessor.sendMIDI = False\n\nlastbandpwr = [[0, 1, 2, 3], processor.baselinedB]\n\nclass MIDIOSCControl():\n\n def __init__(self, master=None, startRow=0, startColumn=0):\n self.master = master\n self.startRow = startRow\n self.startColumn = startColumn\n self.controlRow = startRow ##this gets overwritten if defineChordEntry() is on\n self.controlColumn = startColumn\n\n def defineChordEntry(self):\n def listToString(s):\n # initialize an empty string\n str1 = \" \"\n # return string\n return (str1.join(s))\n\n def stringToList(s):\n return s.split()\n\n self.chordEntryLbl = list()\n self.chordEntrybx = list()\n\n self.chordlist = [chord for chord in processor.mididict.values()]\n\n self.controlRow = self.startRow + len(self.chordlist) + 3\n\n for gesture in processor.cerebro.gestures:\n index = processor.cerebro.gestures.index(gesture)\n\n self.chordEntryLbl.append(tk.Label(self.master))\n self.chordEntryLbl[index][\"text\"] = gesture\n self.chordEntryLbl[index].grid(row=self.startRow+index, column=self.startColumn)\n\n # create entry box and set defaultchordlist as default\n self.chordEntrybx.append(tk.Entry(self.master))\n self.chordEntrybx[index].insert(0, listToString(self.chordlist[index]))\n self.chordEntrybx[index].grid(row=self.startRow + index, column=self.startColumn+1)\n\n # retrieve chords from list\n def defineChordList():\n for items in range(len(processor.cerebro.gestures)):\n self.chordlist[items] = stringToList(self.chordEntrybx[items].get())\n\n processor.updateMIDIdict(self.chordlist)\n\n # button to update chords\n self.updateChords = tk.Button(self.master, command=defineChordList)\n self.updateChords[\"text\"] = \"update MIDIdict\"\n # place the button under all the entry boxes\n self.updateChords.grid(row=self.startRow + len(self.chordlist) + 2, column=self.startColumn, columnspan=2)\n\n def saveChordDict():\n processor.saveMIDIdict(\n addressPath=filedialog.asksaveasfilename(initialdir=MusEEG.parentDir + '/data/MIDIdicts',\n title='save MIDI dictionary') + '.pickle')\n\n # button to save chords\n self.saveChords = tk.Button(self.master, command=saveChordDict)\n self.saveChords[\"text\"] = \"save MIDIdict\"\n # place the button under all the entry boxes\n self.saveChords.grid(row=self.startRow + len(self.chordlist) + 1, column=self.startColumn)\n\n def loadChordDict():\n processor.loadMIDIdict(\n addressPath=filedialog.askopenfilename(initialdir=MusEEG.parentDir + '/data/MIDIdicts',\n title='load MIDI dictionary'))\n\n for gesture in processor.cerebro.gestures:\n index = processor.cerebro.gestures.index(gesture)\n self.chordEntrybx[index].delete(0, 'end')\n self.chordEntrybx[index].insert(0, listToString(processor.mididict[gesture]))\n\n # button to load chords\n self.loadChords = tk.Button(self.master, command=loadChordDict)\n self.loadChords[\"text\"] = \"load MIDIdict\"\n # place the button under all the entry boxes\n self.loadChords.grid(row=self.startRow + len(self.chordlist) + 1, column=self.startColumn+1)\n\n def checkboxArpeggiate(self):\n self.arpeggiateVar = tk.BooleanVar()\n self.checkboxArp = tk.Checkbutton(self.master, text=\"arpeggiate?\", variable=self.arpeggiateVar)\n self.checkboxArp.grid(row=self.controlRow, column=self.startColumn, padx=5, pady=5)\n\n def checkboxScramble(self):\n self.scrambleVar = tk.BooleanVar()\n checkboxScramble = tk.Checkbutton(self.master, text=\"scramble?\", variable=self.scrambleVar)\n checkboxScramble.grid(row=self.controlRow+1, column=self.startColumn, padx=5, pady=5)\n\n def chordDuration(self):\n self.sustainlbl = tk.Label(self.master, text='sustain duration (in qtr notes)').grid(row=self.controlRow+1, column=self.startColumn+1)\n self.sustainbx = tk.Entry(self.master)\n self.sustainbx.insert(10, '8')\n self.sustainbx.grid(row=self.controlRow, column=self.startColumn+1)\n\n def numRepeats(self):\n self.numRepeatsLabel = tk.Label(self.master, text='number of repeats (for arpeggios)').grid(row=self.controlRow+3, column=self.startColumn+1)\n self.numRepBx = tk.Entry(self.master)\n self.numRepBx.insert(0, '8')\n self.numRepBx.grid(row=self.controlRow+2, column=self.startColumn+1)\n\n def updateAllButton(self):\n def updateAll():\n processor.scrambleBool = self.scrambleVar.get()\n processor.arpBool = self.arpeggiateVar.get()\n processor.durVal = self.sustainbx.get()\n processor.numRepeats = self.numRepBx.get()\n\n self.updateAllBttn = tk.Button(self.master, command=updateAll)\n self.updateAllBttn[\"text\"] = \"update these ^^^^^\"\n self.updateAllBttn.grid(row=self.controlRow+4, column=self.controlColumn, columnspan=2, padx=5, pady=5)\n\n def createWidgets(self):\n self.defineChordEntry()\n self.checkboxArpeggiate()\n self.checkboxScramble()\n self.chordDuration()\n self.numRepeats()\n self.updateAllButton()\n\nclass App(tk.Frame):\n\n def __init__(self, master=None):\n super().__init__(master)\n self.master = master\n self.pack()\n self.create_widgets()\n\n def buttonLoadModel(self):\n def loadModel():\n path = filedialog.askdirectory(initialdir=MusEEG.parentDir+'/data/savedModels', title='load classifier')\n processor.bigBrain.loadmodel(filename=path, loadScaler=True)\n\n loadModelBttn = tk.Button(self, command=loadModel)\n loadModelBttn[\"text\"] = \"Load a bigBrain\"\n loadModelBttn.grid(row=8, column=3, columnspan=1, padx=5, pady=5)\n\n def buttonLoadSmallModel(self):\n def loadModel():\n path = filedialog.askdirectory(initialdir=MusEEG.parentDir+'/data/savedModels', title='load classifier')\n processor.smallBrain.loadmodel(filename=path, loadScaler=True)\n\n loadModelBttn = tk.Button(self, command=loadModel)\n loadModelBttn[\"text\"] = \"Load a smallBrain\"\n loadModelBttn.grid(row=8, column=4, columnspan=1, padx=5, pady=5)\n\n def buttonStartProcessor(self):\n self.startProcessorBttn = tk.Button(self, command=self.on_click)\n self.startProcessorBttn[\"text\"] = \"Start Processor\"\n self.startProcessorBttn.grid(row=9, column=1, columnspan=1, padx=0, pady=0)\n\n def plotWindow(self):\n self.running = False\n self.ani = None\n\n self.fig = plt.Figure()\n self.ax1 = self.fig.add_subplot(111)\n\n self.lines = []\n for _ in range(0, eegData.nchannels):\n templine, = self.ax1.plot([], [], lw=2)\n self.lines.append(templine)\n\n # self.line, = self.ax1.plot([], [], lw=2)\n self.canvas = FigureCanvasTkAgg(self.fig, master=self)\n self.canvas.draw()\n self.canvas.get_tk_widget().grid(row=1, column=0, rowspan=4, columnspan=5)\n\n self.ax1.set_title('Raw EEG')\n self.ax1.set_ylim(-500, 6000)\n self.ax1.set_xlim(0, processor.client.windowSize/eegData.sampleRate)\n\n def bandPowerWindow(self):\n self.BPani = None\n\n self.bpfig = plt.Figure((3, 2))\n self.bpax1 = self.bpfig.add_subplot(111)\n\n\n self.bpline, = self.bpax1.plot([0, 1, 2, 3], processor.baselinedB, lw=2)\n\n # self.line, = self.ax1.plot([], [], lw=2)\n self.bpcanvas = FigureCanvasTkAgg(self.bpfig, master=self)\n self.bpcanvas.draw()\n self.bpcanvas.get_tk_widget().grid(row=1, column=6, rowspan=2, columnspan=2)\n\n self.bpax1.set_ylim(0, 40)\n self.bpax1.set_title('Band Power (dB)')\n self.bpax1.set_xlim(-1, 4)\n self.bpax1.set_xticklabels(['', 'theta', 'alpha', 'beta', 'gamma', ''])\n\n def smallBrainMonitor(self):\n self.sbani = None\n self.sbfig = plt.Figure((3, 2))\n self.sbax = self.sbfig.add_subplot(111)\n\n self.sblines = []\n for _ in range(0, eegData.nchannels):\n templine, = self.sbax.plot([], [], lw=2)\n self.sblines.append(templine)\n\n self.sbcanvas = FigureCanvasTkAgg(self.sbfig, master=self)\n self.sbcanvas.draw()\n self.sbcanvas.get_tk_widget().grid(row=5, column=6, rowspan=2, columnspan=2)\n\n self.sbax.set_ylim(-800, 800)\n # self.sbax.set_title('smallBrain monitor')\n self.sbax.set_xlim(-5, eegData.smallchunkSize+5)\n\n def bigBrainMonitor(self):\n self.bbani = None\n self.bbfig = plt.Figure((3, 2))\n self.bbax = self.bbfig.add_subplot(111)\n\n self.bblines = []\n for _ in range(0, eegData.nchannels):\n templine, = self.bbax.plot([], [], lw=2)\n self.bblines.append(templine)\n\n self.bbcanvas = FigureCanvasTkAgg(self.bbfig, master=self)\n self.bbcanvas.draw()\n self.bbcanvas.get_tk_widget().grid(row=3, column=6, rowspan=2, columnspan=2)\n\n self.bbax.set_ylim(-800, 800)\n self.bbax.set_title('bigBrain monitor')\n self.bbax.set_xlim(-5, eegData.chunkSize+5)\n\n def on_click(self):\n '''the button is a start, pause and unpause button all in one\n this method sorts out which of those actions to take'''\n if self.ani is None:\n processor.startStream()\n processor.runProcessorThread(target=processor.mainProcessorWithBackTrack)\n processor.bandPowerThread(asThread=True)\n # print('here on_click')\n # animation is not running; start it\n return self.start()\n\n def start(self):\n self.ani = animation.FuncAnimation(\n self.fig,\n self.update_graph,\n interval=processor.client.windowSize/processor.client.refreshScale/eegData.sampleRate*1000,\n repeat=False)\n self.bpani = animation.FuncAnimation(\n self.bpfig,\n self.update_graph_bp,\n interval=500,\n repeat=False)\n self.sbani = animation.FuncAnimation(\n self.sbfig,\n self.update_graph_sb,\n interval=250,\n repeat=False)\n self.bbani = animation.FuncAnimation(\n self.bbfig,\n self.update_graph_bb,\n interval=250,\n repeat=False)\n self.running = True\n # self.startProcessorBttn.config(text='Pause')\n self.ani._start()\n self.bpani._start()\n self.sbani._start()\n self.bbani._start()\n print('started animation')\n\n def get_data_raw(self):\n x, y = processor.client.getPlotData()\n return x, y\n\n def get_data_bp(self):\n global lastbandpwr\n try:\n bandpwr = processor.bandPowerQueue.get(block=False)\n lastbandpwr = bandpwr\n\n except queue.Empty:\n bandpwr = lastbandpwr\n # x = bandpwr[0]\n x = [0, 1, 2, 3]\n y = bandpwr[1]\n return x, y\n\n def update_graph(self, i):\n x, y = self.get_data_raw()\n\n for idx, line in enumerate(self.lines):\n line.set_data(x, y[:, idx])\n\n return self.lines\n\n def update_graph_bp(self, i):\n x, y = self.get_data_bp()\n # print(y)\n if y is not None:\n self.bpline.set_data(x, y)\n\n return self.bpline\n\n def update_graph_sb(self, i):\n try:\n y = processor.smallBrainMonitorQueue.get(block=False)\n for idx, line in enumerate(self.sblines):\n line.set_data(range(0, eegData.smallchunkSize), y[:, idx])\n\n except queue.Empty:\n pass\n\n return self.sblines\n\n def update_graph_bb(self, i):\n try:\n y = processor.bigBrainMonitorQueue.get(block=False)\n for idx, line in enumerate(self.bblines):\n line.set_data(range(0, eegData.chunkSize), y[:, idx])\n except queue.Empty:\n pass\n\n return self.bblines\n\n def commandWindow(self):\n self.cmd = ScrolledText(master=self, height=15, width=50, relief=\"solid\", bd =2)\n self.cmd.grid(row=5, column=0, rowspan=2, columnspan=6, padx=5, pady=5)\n\n self.add_timestamp()\n\n def add_timestamp(self):\n self.cmd.see(\"end\")\n self.after(1000, self.add_timestamp)\n\n def buttonConnect(self):\n def setup():\n device = self.deviceVar.get()\n if device == 'sim':\n processor.setDevice('sim')\n simPath = filedialog.askopenfilename(initialdir=MusEEG.parentDir+'/data/longRawTrainingSamples', title='choose a .csv file!')\n processor.simPath = simPath\n print('loaded ' + simPath + '!')\n\n else:\n processor.setDevice(device)\n processor.simulation = False\n\n self.connectBttn = tk.Button(self, command=setup)\n self.connectBttn[\"text\"] = \"setup\"\n self.connectBttn.grid(row=8, column=1, columnspan=1, padx=5, pady=5)\n\n def deviceDropDown(self):\n self.deviceLabel = tk.Label(self, text='pick a device above ^^^')\n self.deviceLabel.grid(row=9, column=0, columnspan=1, padx=0, pady=0)\n self.deviceVar = tk.StringVar(self)\n self.deviceVar.set(\"sim\")\n\n self.deviceMenu = tk.OptionMenu(self, self.deviceVar, *processor.deviceList).grid(row=8, column=0, columnspan=1, padx=5, pady=5)\n\n def quitButton(self):\n self.quitbttn = tk.Button(self, text=\"Shutdown\", command=self.quitProcessor).grid(row=9, column=4, padx=5, pady=5)\n\n def quitProcessor(self):\n # global processor\n # processor.client.done = True\n # processor.processorShutDown()\n print('starting animation shutdown')\n if self.running:\n self.bbani._stop()\n self.bpani._stop()\n self.ani._stop()\n self.sbani._stop()\n print('starting processor shutdown')\n processor.processorShutDown()\n # del processor\n print('processor object deleted')\n print('shutdown! feel free to quit')\n self.master.destroy()\n\n def midiOSCCOntrolButton(self):\n def createMIDIWindow():\n if not self.MIDIWindowOpen:\n submaster = tk.Toplevel(self)\n submaster.wm_title('MusEEG MIDI')\n midiOSC = MIDIOSCControl(master=submaster, startRow=0, startColumn=0)\n midiOSC.createWidgets()\n processor.sendMIDI = True\n self.MIDIWindowOpen = True\n\n self.MIDIWindowOpen = False\n button = tk.Button(self, text=\"Show MIDI Menu\", command=createMIDIWindow)\n button.grid(row=9, column=3)\n\n def create_widgets(self):\n self.winfo_toplevel().title(\"MusEEG (OSC)\")\n self.buttonStartProcessor()\n self.buttonLoadModel()\n self.plotWindow()\n self.commandWindow()\n self.bandPowerWindow()\n self.deviceDropDown()\n self.buttonConnect()\n self.smallBrainMonitor()\n self.bigBrainMonitor()\n self.midiOSCCOntrolButton()\n # self.buttonLoadSmallModel()\n # self.quitButton()\n\n self.master.protocol(\"WM_DELETE_WINDOW\", self.quitProcessor)\n # replace sys.stdout with our object\n sys.stdout = PrintLogger(self.cmd)\n #\n\n\nclass PrintLogger(): # createa file like object\n def __init__(self, textbox): # pass reference to text widget\n self.textbox = textbox # keep ref\n\n def write(self, text):\n self.textbox.insert(tk.END, text) # write text to textbox\n # could also scroll to end of textbox here to make sure always visible\n\n def flush(self): # needed for file like object\n pass\n\n\nif __name__ == \"__main__\":\n root = tk.Tk()\n root.lift()\n root.iconbitmap(os.path.join(parentDir, 'museeg-logo.ico'))\n app = App(master=root)\n\n\n\n # todo: the app isnt quitting properly\n while True:\n try:\n print('hello! this is the MusEEG log')\n print('classification results are printed here\\n\\n')\n flower = [\n \"/ __ __ /\",\n \"/ .' `...' `. /\",\n \"/ __| | |__ /\",\n \"/ .' \\ . / `. /\",\n \"/ | ./###\\. | /\",\n \"/ >---- |#####| ----< /\",\n \"/ | `\\###/' | /\",\n \"/ `.__ / . \\ __.' /\",\n \"/ | | | /\",\n \"/ `.___.^.___.' /\"]\n\n for line in flower:\n print(line)\n\n app.mainloop()\n break\n except UnicodeDecodeError:\n pass\n", "sub_path": "GUIappSC.py", "file_name": "GUIappSC.py", "file_ext": "py", "file_size_in_byte": 17646, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "matplotlib.use", "line_number": 12, "usage_type": "call"}, {"api_name": "MusEEG.Processor", "line_number": 17, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 53, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 58, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 70, "usage_type": "call"}, {"api_name": "tkinter.filedialog.asksaveasfilename", "line_number": 77, "usage_type": "call"}, {"api_name": "tkinter.filedialog", "line_number": 77, "usage_type": "name"}, {"api_name": "MusEEG.parentDir", "line_number": 77, "usage_type": "attribute"}, {"api_name": "tkinter.Button", "line_number": 81, "usage_type": "call"}, {"api_name": "tkinter.filedialog.askopenfilename", "line_number": 88, "usage_type": "call"}, {"api_name": "tkinter.filedialog", "line_number": 88, "usage_type": "name"}, {"api_name": "MusEEG.parentDir", "line_number": 88, "usage_type": "attribute"}, {"api_name": "tkinter.Button", "line_number": 97, "usage_type": "call"}, {"api_name": "tkinter.BooleanVar", "line_number": 103, "usage_type": "call"}, {"api_name": "tkinter.Checkbutton", "line_number": 104, "usage_type": "call"}, {"api_name": "tkinter.BooleanVar", "line_number": 108, "usage_type": "call"}, {"api_name": "tkinter.Checkbutton", "line_number": 109, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 113, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 114, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 119, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 120, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 131, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 143, "usage_type": "attribute"}, {"api_name": "tkinter.filedialog.askdirectory", "line_number": 153, "usage_type": "call"}, {"api_name": "tkinter.filedialog", "line_number": 153, "usage_type": "name"}, {"api_name": "MusEEG.parentDir", "line_number": 153, "usage_type": "attribute"}, {"api_name": "tkinter.Button", "line_number": 156, "usage_type": "call"}, {"api_name": "tkinter.filedialog.askdirectory", "line_number": 162, "usage_type": "call"}, {"api_name": "tkinter.filedialog", "line_number": 162, "usage_type": "name"}, {"api_name": "MusEEG.parentDir", "line_number": 162, "usage_type": "attribute"}, {"api_name": "tkinter.Button", "line_number": 165, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 170, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.Figure", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 178, "usage_type": "name"}, {"api_name": "MusEEG.eegData.nchannels", "line_number": 182, "usage_type": "attribute"}, {"api_name": "MusEEG.eegData", "line_number": 182, "usage_type": "name"}, {"api_name": "matplotlib.backends.backend_tkagg.FigureCanvasTkAgg", "line_number": 187, "usage_type": "call"}, {"api_name": "MusEEG.eegData.sampleRate", "line_number": 193, "usage_type": "attribute"}, {"api_name": "MusEEG.eegData", "line_number": 193, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.Figure", "line_number": 198, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 198, "usage_type": "name"}, {"api_name": "matplotlib.backends.backend_tkagg.FigureCanvasTkAgg", "line_number": 205, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.Figure", "line_number": 216, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 216, "usage_type": "name"}, {"api_name": "MusEEG.eegData.nchannels", "line_number": 220, "usage_type": "attribute"}, {"api_name": "MusEEG.eegData", "line_number": 220, "usage_type": "name"}, {"api_name": "matplotlib.backends.backend_tkagg.FigureCanvasTkAgg", "line_number": 224, "usage_type": "call"}, {"api_name": "MusEEG.eegData.smallchunkSize", "line_number": 230, "usage_type": "attribute"}, {"api_name": "MusEEG.eegData", "line_number": 230, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.Figure", "line_number": 234, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 234, "usage_type": "name"}, {"api_name": "MusEEG.eegData.nchannels", "line_number": 238, "usage_type": "attribute"}, {"api_name": "MusEEG.eegData", "line_number": 238, "usage_type": "name"}, {"api_name": "matplotlib.backends.backend_tkagg.FigureCanvasTkAgg", "line_number": 242, "usage_type": "call"}, {"api_name": "MusEEG.eegData.chunkSize", "line_number": 248, "usage_type": "attribute"}, {"api_name": "MusEEG.eegData", "line_number": 248, "usage_type": "name"}, {"api_name": "matplotlib.animation.FuncAnimation", "line_number": 262, "usage_type": "call"}, {"api_name": "matplotlib.animation", "line_number": 262, "usage_type": "name"}, {"api_name": "MusEEG.eegData.sampleRate", "line_number": 265, "usage_type": "attribute"}, {"api_name": "MusEEG.eegData", "line_number": 265, "usage_type": "name"}, {"api_name": "matplotlib.animation.FuncAnimation", "line_number": 267, "usage_type": "call"}, {"api_name": "matplotlib.animation", "line_number": 267, "usage_type": "name"}, {"api_name": "matplotlib.animation.FuncAnimation", "line_number": 272, "usage_type": "call"}, {"api_name": "matplotlib.animation", "line_number": 272, "usage_type": "name"}, {"api_name": "matplotlib.animation.FuncAnimation", "line_number": 277, "usage_type": "call"}, {"api_name": "matplotlib.animation", "line_number": 277, "usage_type": "name"}, {"api_name": "queue.Empty", "line_number": 300, "usage_type": "attribute"}, {"api_name": "MusEEG.eegData.smallchunkSize", "line_number": 327, "usage_type": "attribute"}, {"api_name": "MusEEG.eegData", "line_number": 327, "usage_type": "name"}, {"api_name": "queue.Empty", "line_number": 329, "usage_type": "attribute"}, {"api_name": "MusEEG.eegData.chunkSize", "line_number": 338, "usage_type": "attribute"}, {"api_name": "MusEEG.eegData", "line_number": 338, "usage_type": "name"}, {"api_name": "queue.Empty", "line_number": 339, "usage_type": "attribute"}, {"api_name": "tkinter.scrolledtext.ScrolledText", "line_number": 345, "usage_type": "call"}, {"api_name": "tkinter.filedialog.askopenfilename", "line_number": 359, "usage_type": "call"}, {"api_name": "tkinter.filedialog", "line_number": 359, "usage_type": "name"}, {"api_name": "MusEEG.parentDir", "line_number": 359, "usage_type": "attribute"}, {"api_name": "tkinter.Button", "line_number": 367, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 372, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 374, "usage_type": "call"}, {"api_name": "tkinter.OptionMenu", "line_number": 377, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 380, "usage_type": "call"}, {"api_name": "tkinter.Toplevel", "line_number": 402, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 410, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 430, "usage_type": "attribute"}, {"api_name": "tkinter.END", "line_number": 439, "usage_type": "attribute"}, {"api_name": "tkinter.Tk", "line_number": 447, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 449, "usage_type": "call"}, {"api_name": "MusEEG.parentDir", "line_number": 449, "usage_type": "argument"}, {"api_name": "os.path", "line_number": 449, "usage_type": "attribute"}]}
+{"seq_id": "100824204", "text": "from mutation import *\n\nnp.random.seed(1)\nrandom.seed(1)\n\ndef parse_args():\n import argparse\n parser = argparse.ArgumentParser(description='HIV sequence analysis')\n parser.add_argument('model_name', type=str,\n help='Type of language model (e.g., hmm, lstm)')\n parser.add_argument('--namespace', type=str, default='hiv',\n help='Model namespace')\n parser.add_argument('--dim', type=int, default=512,\n help='Embedding dimension')\n parser.add_argument('--batch-size', type=int, default=1000,\n help='Training minibatch size')\n parser.add_argument('--n-epochs', type=int, default=4,\n help='Number of training epochs')\n parser.add_argument('--seed', type=int, default=1,\n help='Random seed')\n parser.add_argument('--checkpoint', type=str, default=None,\n help='Model checkpoint')\n parser.add_argument('--train', action='store_true',\n help='Train model')\n parser.add_argument('--train-split', action='store_true',\n help='Train model on portion of data')\n parser.add_argument('--test', action='store_true',\n help='Test model')\n parser.add_argument('--embed', action='store_true',\n help='Analyze embeddings')\n parser.add_argument('--semantics', action='store_true',\n help='Analyze mutational semantic change')\n parser.add_argument('--combfit', action='store_true',\n help='Analyze combinatorial fitness')\n args = parser.parse_args()\n return args\n\ndef load_meta(meta_fnames):\n metas = {}\n for fname in meta_fnames:\n with open(fname) as f:\n for line in f:\n if not line.startswith('>'):\n continue\n accession = line[1:].rstrip()\n fields = line.rstrip().split('.')\n subtype, country, year, strain = (\n fields[0], fields[1], fields[2], fields[3]\n )\n\n if year == '-':\n year = None\n else:\n year = int(year)\n\n subtype = subtype.split('_')[-1]\n subtype = subtype.lstrip('>0123')\n\n keep_subtypes = {\n 'A', 'A1', 'A1A2', 'A1C', 'A1D', 'A2', 'A3', 'A6',\n 'AE', 'AG', 'B', 'C', 'BC', 'D',\n 'F', 'F1', 'F2', 'G', 'H', 'J',\n 'K', 'L', 'N', 'O', 'P', 'U',\n }\n if subtype not in keep_subtypes:\n subtype = 'Other'\n\n metas[accession] = {\n 'subtype': subtype,\n 'country': country,\n 'year': year,\n 'strain': strain,\n }\n return metas\n\ndef process(args, fnames, meta_fnames):\n metas = load_meta(meta_fnames)\n\n seqs = {}\n for fname in fnames:\n for record in SeqIO.parse(fname, 'fasta'):\n accession = record.description\n meta = metas[accession]\n meta['seqlen'] = len(str(record.seq))\n if args.namespace == 'hiva' and \\\n (not meta['subtype'].startswith('A')):\n continue\n if 'X' in record.seq:\n continue\n if record.seq not in seqs:\n seqs[record.seq] = []\n seqs[record.seq].append(meta)\n return seqs\n\ndef split_seqs(seqs, split_method='random'):\n train_seqs, test_seqs = {}, {}\n\n old_cutoff = 1900\n new_cutoff = 2008\n\n tprint('Splitting seqs...')\n for seq in seqs:\n # Pick validation set based on date.\n seq_dates = [\n meta['year'] for meta in seqs[seq]\n if meta['year'] is not None\n ]\n if len(seq_dates) == 0:\n test_seqs[seq] = seqs[seq]\n continue\n if len(seq_dates) > 0:\n oldest_date = sorted(seq_dates)[0]\n if oldest_date < old_cutoff or oldest_date >= new_cutoff:\n test_seqs[seq] = seqs[seq]\n continue\n train_seqs[seq] = seqs[seq]\n tprint('{} train seqs, {} test seqs.'\n .format(len(train_seqs), len(test_seqs)))\n\n return train_seqs, test_seqs\n\ndef setup(args):\n fnames = [ 'data/hiv/HIV-1_env_samelen.fa' ]\n meta_fnames = [ 'data/hiv/HIV-1_env_samelen.fa' ]\n\n seqs = process(args, fnames, meta_fnames)\n\n seq_len = max([ len(seq) for seq in seqs ]) + 2\n vocab_size = len(AAs) + 2\n\n model = get_model(args, seq_len, vocab_size,\n inference_batch_size=1000)\n\n return model, seqs\n\ndef interpret_clusters(adata):\n clusters = sorted(set(adata.obs['louvain']))\n for cluster in clusters:\n tprint('Cluster {}'.format(cluster))\n adata_cluster = adata[adata.obs['louvain'] == cluster]\n for var in [ 'year', 'country', 'subtype' ]:\n tprint('\\t{}:'.format(var))\n counts = Counter(adata_cluster.obs[var])\n for val, count in counts.most_common():\n tprint('\\t\\t{}: {}'.format(val, count))\n tprint('')\n\n cluster2subtype = {}\n for i in range(len(adata)):\n cluster = adata.obs['louvain'][i]\n if cluster not in cluster2subtype:\n cluster2subtype[cluster] = []\n cluster2subtype[cluster].append(adata.obs['subtype'][i])\n largest_pct_subtype = []\n for cluster in cluster2subtype:\n count = Counter(cluster2subtype[cluster]).most_common(1)[0][1]\n pct_subtype = float(count) / len(cluster2subtype[cluster])\n largest_pct_subtype.append(pct_subtype)\n tprint('\\tCluster {}, largest subtype % = {}'\n .format(cluster, pct_subtype))\n tprint('Purity, Louvain and subtype: {}'\n .format(np.mean(largest_pct_subtype)))\n\ndef plot_umap(adata):\n sc.tl.umap(adata, min_dist=1.)\n sc.pl.umap(adata, color='louvain', save='_hiv_louvain.png')\n sc.pl.umap(adata, color='subtype', save='_hiv_subtype.png')\n\ndef analyze_embedding(args, model, seqs, vocabulary):\n sorted_seqs = np.array([ str(s) for s in sorted(seqs.keys()) ])\n batch_size = 3000\n n_batches = math.ceil(len(sorted_seqs) / float(batch_size))\n for batchi in range(n_batches):\n start = batchi * batch_size\n end = (batchi + 1) * batch_size\n seqs_batch = { seq: seqs[seq] for seq in sorted_seqs[start:end] }\n seqs_batch = embed_seqs(args, model, seqs_batch, vocabulary,\n use_cache=False)\n for seq in seqs_batch:\n for meta in seqs[seq]:\n meta['embedding'] = seqs_batch[seq][0]['embedding'].mean(0)\n del seqs_batch\n\n X, obs = [], {}\n obs['n_seq'] = []\n obs['seq'] = []\n for seq in seqs:\n meta = seqs[seq][0]\n X.append(meta['embedding'])\n for key in meta:\n if key == 'embedding':\n continue\n if key not in obs:\n obs[key] = []\n obs[key].append(Counter([\n meta[key] for meta in seqs[seq]\n ]).most_common(1)[0][0])\n obs['n_seq'].append(len(seqs[seq]))\n obs['seq'].append(str(seq))\n X = np.array(X)\n\n adata = AnnData(X)\n for key in obs:\n adata.obs[key] = obs[key]\n\n sc.pp.neighbors(adata, n_neighbors=200, use_rep='X')\n sc.tl.louvain(adata, resolution=1.)\n\n sc.set_figure_params(dpi_save=500)\n plot_umap(adata)\n\n interpret_clusters(adata)\n\nif __name__ == '__main__':\n args = parse_args()\n\n AAs = [\n 'A', 'R', 'N', 'D', 'C', 'Q', 'E', 'G', 'H',\n 'I', 'L', 'K', 'M', 'F', 'P', 'S', 'T', 'W',\n 'Y', 'V', 'X', 'Z', 'J', 'U', 'B',\n ]\n vocabulary = { aa: idx + 1 for idx, aa in enumerate(sorted(AAs)) }\n\n model, seqs = setup(args)\n\n if 'esm' in args.model_name:\n args.checkpoint = args.model_name\n elif args.checkpoint is not None:\n model.model_.load_weights(args.checkpoint)\n tprint('Model summary:')\n tprint(model.model_.summary())\n\n if args.train:\n batch_train(args, model, seqs, vocabulary, batch_size=5000)\n\n if args.train_split or args.test:\n train_test(args, model, seqs, vocabulary, split_seqs)\n\n if args.embed:\n if args.checkpoint is None and not args.train:\n raise ValueError('Model must be trained or loaded '\n 'from checkpoint.')\n no_embed = { 'hmm' }\n if args.model_name in no_embed:\n raise ValueError('Embeddings not available for models: {}'\n .format(', '.join(no_embed)))\n analyze_embedding(args, model, seqs, vocabulary)\n\n if args.semantics:\n if args.checkpoint is None and not args.train:\n raise ValueError('Model must be trained or loaded '\n 'from checkpoint.')\n\n from escape import load_dingens2019\n tprint('Dingens et al. 2019...')\n seq_to_mutate, escape_seqs = load_dingens2019()\n positions = [ escape_seqs[seq][0]['pos'] for seq in escape_seqs ]\n min_pos, max_pos = min(positions), max(positions)\n analyze_semantics(\n args, model, vocabulary, seq_to_mutate, escape_seqs,\n min_pos=min_pos, max_pos=max_pos,\n prob_cutoff=0., beta=1., plot_acquisition=True,\n )\n\n if args.combfit:\n from combinatorial_fitness import load_haddox2018\n tprint('Haddox et al. 2018...')\n wt_seqs, seqs_fitness = load_haddox2018()\n strains = sorted(wt_seqs.keys())\n for strain in strains:\n analyze_comb_fitness(args, model, vocabulary,\n strain, wt_seqs[strain], seqs_fitness,\n prob_cutoff=0., beta=1.)\n", "sub_path": "bin/hiv.py", "file_name": "hiv.py", "file_ext": "py", "file_size_in_byte": 9871, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 8, "usage_type": "call"}, {"api_name": "escape.load_dingens2019", "line_number": 256, "usage_type": "call"}, {"api_name": "combinatorial_fitness.load_haddox2018", "line_number": 268, "usage_type": "call"}]}
+{"seq_id": "350779142", "text": "# -------------------------------------------------------------------------------\r\n# Name: Reading polar volume data\r\n# Purpose:\r\n#\r\n# Author: heistermann\r\n#\r\n# Created: 14.01.2013\r\n# Copyright: (c) heistermann 2013\r\n# Licence: MIT\r\n# -------------------------------------------------------------------------------\r\n#!/usr/bin/env python\r\n\r\nimport wradlib\r\nimport numpy as np\r\nimport matplotlib.pyplot as pl\r\n# just making sure that the plots immediately pop up\r\n#pl.interactive(True)\r\nimport datetime as dt\r\nimport os\r\nfrom osgeo import osr\r\n\r\n\r\ndef recipe_polar_volume_example():\r\n # read the data (sample file in wradlib/examples/data)\r\n raw = wradlib.io.read_OPERA_hdf5(os.path.dirname(__file__) + '/' + \"data/knmi_polar_volume.h5\")\r\n # this is the radar position tuple (longitude, latitude, altitude)\r\n sitecoords = (raw[\"where\"][\"lon\"][0], raw[\"where\"][\"lat\"][0], raw[\"where\"][\"height\"][0])\r\n # define your cartesian reference system\r\n # proj = wradlib.georef.create_osr(32632)\r\n proj = osr.SpatialReference()\r\n proj.ImportFromEPSG(32632)\r\n # containers to hold Cartesian bin coordinates and data\r\n xyz, data = np.array([]).reshape((-1, 3)), np.array([])\r\n # iterate over 14 elevation angles\r\n for i in range(14):\r\n # get the scan metadata for each elevation\r\n where = raw[\"dataset%d/where\" % (i + 1)]\r\n what = raw[\"dataset%d/data1/what\" % (i + 1)]\r\n # define arrays of polar coordinate arrays (azimuth and range)\r\n az = np.arange(0., 360., 360. / where[\"nrays\"])\r\n r = np.arange(where[\"rstart\"], where[\"rstart\"] + where[\"nbins\"] * where[\"rscale\"], where[\"rscale\"])\r\n # derive 3-D Cartesian coordinate tuples\r\n xyz_ = wradlib.vpr.volcoords_from_polar(sitecoords, where[\"elangle\"], az, r, proj)\r\n # get the scan data for this elevation\r\n # here, you can do all the processing on the 2-D polar level\r\n # e.g. clutter elimination, attenuation correction, ...\r\n data_ = what[\"offset\"] + what[\"gain\"] * raw[\"dataset%d/data1/data\" % (i + 1)]\r\n # transfer to containers\r\n xyz, data = np.vstack((xyz, xyz_)), np.append(data, data_.ravel())\r\n\r\n # generate 3-D Cartesian target grid coordinates\r\n maxrange = 200000.\r\n minelev = 0.1\r\n maxelev = 25.\r\n maxalt = 5000.\r\n horiz_res = 2000.\r\n vert_res = 250.\r\n trgxyz, trgshape = wradlib.vpr.make_3D_grid(sitecoords, proj, maxrange, maxalt, horiz_res, vert_res)\r\n\r\n # interpolate to Cartesian 3-D volume grid\r\n tstart = dt.datetime.now()\r\n gridder = wradlib.vpr.CAPPI(xyz, trgxyz, trgshape, maxrange, minelev, maxelev)\r\n vol = np.ma.masked_invalid(gridder(data).reshape(trgshape))\r\n print(\"3-D interpolation took:\", dt.datetime.now() - tstart)\r\n\r\n # diagnostic plot\r\n trgx = trgxyz[:, 0].reshape(trgshape)[0, 0, :]\r\n trgy = trgxyz[:, 1].reshape(trgshape)[0, :, 0]\r\n trgz = trgxyz[:, 2].reshape(trgshape)[:, 0, 0]\r\n wradlib.vis.plot_max_plan_and_vert(trgx, trgy, trgz, vol, unit=\"dBZH\", levels=range(-32, 60))\r\n\r\n\r\nif __name__ == '__main__':\r\n recipe_polar_volume_example()\r\n", "sub_path": "examples/recipe2_polar_volume_example.py", "file_name": "recipe2_polar_volume_example.py", "file_ext": "py", "file_size_in_byte": 3131, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "wradlib.io.read_OPERA_hdf5", "line_number": 25, "usage_type": "call"}, {"api_name": "wradlib.io", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "osgeo.osr.SpatialReference", "line_number": 30, "usage_type": "call"}, {"api_name": "osgeo.osr", "line_number": 30, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 41, "usage_type": "call"}, {"api_name": "wradlib.vpr.volcoords_from_polar", "line_number": 43, "usage_type": "call"}, {"api_name": "wradlib.vpr", "line_number": 43, "usage_type": "attribute"}, {"api_name": "numpy.vstack", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 49, "usage_type": "call"}, {"api_name": "wradlib.vpr.make_3D_grid", "line_number": 58, "usage_type": "call"}, {"api_name": "wradlib.vpr", "line_number": 58, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 61, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 61, "usage_type": "attribute"}, {"api_name": "wradlib.vpr.CAPPI", "line_number": 62, "usage_type": "call"}, {"api_name": "wradlib.vpr", "line_number": 62, "usage_type": "attribute"}, {"api_name": "numpy.ma.masked_invalid", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.ma", "line_number": 63, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 64, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 64, "usage_type": "attribute"}, {"api_name": "wradlib.vis.plot_max_plan_and_vert", "line_number": 70, "usage_type": "call"}, {"api_name": "wradlib.vis", "line_number": 70, "usage_type": "attribute"}]}
+{"seq_id": "218549704", "text": "import cv2\nimport numpy as np\nimport math\n\nimg = cv2.imread('img.jpg', 1)\nimgInfo = img.shape\nheight = imgInfo[0]\nwidth = imgInfo[1]\ncv2.imshow('src', img)\n#canny 1 gray 2 高斯滤波 3 canny\n# gray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)\n# imgG = cv2.GaussianBlur(gray, (3, 3), 0)\n# dst = cv2.Canny(img, 50, 50)\n\n#sobel 1算子模板 2 图片卷积 3阈值判决\n#[ 1 2 1 [ 1, 0. -1\n# 0 0 0 2, 0, -2\n#-1 -2 -1] 1, 0, -1]\n\ngray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)\ndst = np.zeros((height ,width, 1), np.uint8)\nfor i in range(0, height-2):\n for j in range(0, width-2):\n gy = gray[i,j]*1+gray[i,j+1]*2+gray[i,j+2]*1-gray[i+2,j]*1-gray[i+2,j+1]*2-gray[i+2,j+2]*1\n gx = gray[i,j]*1+gray[i+1,j]*2+gray[i+2,j]*1-gray[i,j+2]*1-gray[i+1,j+2]*2-gray[i+2,j+2]*1\n grad = math.sqrt(gx*gx+gy*gy)\n if grad > 50:\n dst[i, j] = 255\n else:\n dst[i, j] = 0\ncv2.imshow('dst', dst)\ncv2.waitKey()\n", "sub_path": "OpenCV-Notes/图像的特效变换/Edge.py", "file_name": "Edge.py", "file_ext": "py", "file_size_in_byte": 953, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "cv2.imread", "line_number": 5, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 20, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 21, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 31, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 32, "usage_type": "call"}]}
+{"seq_id": "483297869", "text": "# -*- coding: UTF-8 -*-\n\nfrom scrapy.selector import Selector\nfrom scrapy.contrib.linkextractors.sgml import SgmlLinkExtractor\nfrom scrapy.contrib.spiders import CrawlSpider, Rule\nfrom scrapy.http import Request\nfrom items import JDProduct\nfrom scrapy import log\nfrom BaseSpider import BaseSpider\nimport re\nimport json\n\n\nclass JdItemsSpider(BaseSpider):\n name = 'JdItems'\n allowed_domains = []\n # ['http://www.jd.com/', 'http://item.jd.com/']\n start_urls = ['http://www.jd.com/']\n\n def parse(self, response):\n urls = BaseSpider.get_urls(self)\n if urls:\n for url in urls:\n yield BaseSpider.make_request(self, url, self.parse_url)\n\n def parse_url(self, response):\n sel = Selector(response)\n title = sel.xpath(\"//title/text()\").extract()[0]\n item = JDProduct()\n extracted_info = self.get_skuid_and_name(sel)\n item[\"id\"] = int(extracted_info[0])\n item[\"name\"] = extracted_info[1]\n if item.get(\"id\"):\n request = Request(url=\"http://p.3.cn/prices/mgets?skuIds=J_\" + str(item[\"id\"]),\n callback=self.parse_price)\n request.meta[\"item\"] = item\n return request\n return None\n\n def parse_price(self, response):\n item = response.meta[\"item\"]\n json_object = json.loads(response.body)\n item[\"price\"] = float(json_object[0].get(\"p\"))\n return item\n\n def get_skuid_and_name(self, sel):\n skuid = 0\n name = \"\"\n scripts = BaseSpider.get_scripts(self, sel)\n for script in scripts:\n script_text = script.extract()\n m = re.search(r\"skuid:\\s(\\d+)\", script_text)\n if m:\n skuid = m.group(1)\n m2 = re.search(r\"name:\\s'(.*?)'\", script_text)\n if m2:\n name = BaseSpider.unicode_escape(self, m2.group(1))\n break\n return (skuid, name)\n\t\n", "sub_path": "tutorial/tutorial/spiders/JdItems.py", "file_name": "JdItems.py", "file_ext": "py", "file_size_in_byte": 1960, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "BaseSpider.BaseSpider", "line_number": 14, "usage_type": "name"}, {"api_name": "BaseSpider.BaseSpider.get_urls", "line_number": 21, "usage_type": "call"}, {"api_name": "BaseSpider.BaseSpider", "line_number": 21, "usage_type": "name"}, {"api_name": "BaseSpider.BaseSpider.make_request", "line_number": 24, "usage_type": "call"}, {"api_name": "BaseSpider.BaseSpider", "line_number": 24, "usage_type": "name"}, {"api_name": "scrapy.selector.Selector", "line_number": 27, "usage_type": "call"}, {"api_name": "items.JDProduct", "line_number": 29, "usage_type": "call"}, {"api_name": "scrapy.http.Request", "line_number": 34, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 42, "usage_type": "call"}, {"api_name": "BaseSpider.BaseSpider.get_scripts", "line_number": 49, "usage_type": "call"}, {"api_name": "BaseSpider.BaseSpider", "line_number": 49, "usage_type": "name"}, {"api_name": "re.search", "line_number": 52, "usage_type": "call"}, {"api_name": "re.search", "line_number": 55, "usage_type": "call"}, {"api_name": "BaseSpider.BaseSpider.unicode_escape", "line_number": 57, "usage_type": "call"}, {"api_name": "BaseSpider.BaseSpider", "line_number": 57, "usage_type": "name"}]}
+{"seq_id": "575174526", "text": "from PIL import Image\nimport pytesseract\nimport cv2\n\ndef hashCheck(frame):\n #frame = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB)\n frame1 = frame[400:430, 0:10]\n\n #frame1 = cv2.bitwise_not(frame1)\n frame1Vec = frame1.flatten()\n frame1VecSum = sum(frame1Vec)\n #print(frame1VecSum)\n cv2.imshow('frame', frame1)\n cv2.waitKey(0)\n cv2.destroyAllWindows()\n print(frame1VecSum)\n if frame1VecSum > 100000:\n ret = \"left\"\n # if ret == \"left\":\n # frame2 = frame[400:432, 1191:1200]\n # frame2 = cv2.bitwise_not(frame)\n # frame2Vec = frame2.flatten()\n # frame2VecSum = sum(frame2Vec)\n # cv2.imshow('frame', frame2)\n # cv2.waitKey(0)\n # ret = \"center\"\n else:\n frame2 = frame[330:355, 0:10]\n frame2Vec = frame2.flatten()\n frame2VecSum = sum(frame2Vec)\n cv2.imshow('frame', frame2)\n cv2.waitKey(0)\n cv2.destroyAllWindows()\n print(frame2VecSum)\n if frame2VecSum > 100000:\n ret = \"center\"\n else:\n ret = \"right\"\n #cv2.imshow('frame', frame2)\n #cv2.waitKey(0)\n #ret = \"right\"\n #frame = cv2.GaussianBlur(frame, (5, 5), 0)\n # thresh, frame = cv2.threshold(frame, 100, cv2.ADAPTIVE_THRESH_GAUSSIAN_C, 11, 2)\n # frame1 = cv2.resize(frame, (100,102))\n # frame2 = cv2.resize(frame, (200, 200)) #changed values\n # tessFrame = Image.fromarray(frame2)\n # ret = pytesseract.image_to_string(tessFrame, config=\"--psm 10000 -c tessedit_char_whitelist=0123456789\")\n # cv2.imshow('frame', frame2)\n # cv2.waitKey(0)\n #cv2.destroyAllWindows()\n # if ret == \"\":\n # #doesn't recognize 0 alone, so this is a measure to prevent that\n # tessFrame = Image.fromarray(frame1)\n # ret = pytesseract.image_to_string(tessFrame, config=\"--psm 10000 -c tessedit_char_whitelist=0123456789\")\n # cv2.imshow('frame1', frame1)\n # cv2.waitKey(0)\n return(ret)\n\nframe = cv2.imread(\"ScreenStuff/images/hashImages/30.png\")\n#frame = cv2.resize(frame, (1200, 800))\nprint(hashCheck(frame))\n", "sub_path": "ScreenStuff/ScreenStuff/textRecognition/hash.py", "file_name": "hash.py", "file_ext": "py", "file_size_in_byte": 2093, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "cv2.imshow", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 31, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 32, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 33, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 59, "usage_type": "call"}]}
+{"seq_id": "492699134", "text": "#coding:utf8\r\n'''\r\nCreated on 2014-1-17\r\n\r\n@author: CC\r\n'''\r\n\r\nfrom app.game.component.baseInfo.ItemBaseInfoComponent import ItemBaseInfoComponent\r\nfrom app.game.component.attribute.ItemAttributeComponent import ItemAttributeComponent\r\nfrom app.game.component.pack.ItemPackComponent import ItemPackComponent\r\nfrom app.share.dbopear import dbItems\r\nfrom app.dbfront.memmode import tb_item_admin\r\nimport datetime\r\n\r\nclass Item(object):\r\n\t'''物品类'''\r\n\r\n\tdef __init__(self,itemTemplateId=0,id=0,name=''):\r\n\t\t'''初始化物品类\r\n\t\t@param id: int 物品在数据库中的id\r\n\t\t@param itemTemplateId: int 物品的模板id\r\n\t\t@param selfExtraAttributeId: []int list 物品自身的附加属性\r\n\t\t'''\r\n\t\tself.baseInfo=ItemBaseInfoComponent(self,id,name,itemTemplateId)\r\n\t\tself.attribute=ItemAttributeComponent(self)\r\n\t\tself.pack=ItemPackComponent(self)\r\n\r\n\tdef initItemInstance(self,itemInstance):\r\n\t\t'''初始化实际物品信息\r\n\t\t'''\r\n\t\tself.baseInfo.setItemTemplateId(itemInstance['itemTemplateId'])\r\n\t\tself.attribute.setDurability(itemInstance['durability'])\r\n\t\tself.pack.setStack(itemInstance['stack'])\r\n\r\n\tdef getLJtype(self):\r\n\t\t'''获取零件类型'''\r\n\t\titeminfo=self.baseInfo.getItemTemplateInfo()#物品模板id信息\r\n\t\ttypeid=iteminfo.get('bodyType',0)\r\n\t\treturn typeid\r\n\r\n\tdef formatItemInfo(self):\r\n\t\t'''格式化物品信息'''\r\n\t\tdata=self.baseInfo.getItemTemplateInfo()\r\n\t\tdata['id']=self.baseInfo.getId()\r\n\t\tdata['templateId']=self.baseInfo.getItemTemplateId()\r\n\t\tdata['stack']=self.pack.getStack()\r\n\t\treturn data\r\n\r\n\tdef InsertItemIntoDB(self,characterId=0):\r\n\t\t'''将物品信息写入数据库'''\r\n\t\tif self.baseInfo.id:\r\n\t\t\treturn\r\n\t\titemTemplateId=self.baseInfo.itemTemplateId\r\n\t\tisBound=0\r\n\t\tdurability=-1\r\n\t\tstack=self.pack.getStack()\r\n\t\tdata={'characterId':characterId,'itemTemplateId':itemTemplateId,'isBound':isBound,'accesstime':datetime.datetime.now(),'durability':durability,'stack':stack}\r\n\t\tnewitemmode=tb_item_admin.new(data)\r\n\t\titemId=int(newitemmode._name.split(':')[1])\r\n\t\tself.baseInfo.setId(itemId)\r\n\t\treturn itemId\r\n\r\n\tdef destoryItemInDB(self):\r\n\t\t'''删除数据库中的自身的信息'''\r\n\t\tif self.baseInfo.id!=0:\r\n\t\t\treturn tb_item_admin.deleteMode(self.baseInfo.id)\r\n\t\treturn False", "sub_path": "app/game/core/Item.py", "file_name": "Item.py", "file_ext": "py", "file_size_in_byte": 2234, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "app.game.component.baseInfo.ItemBaseInfoComponent.ItemBaseInfoComponent", "line_number": 24, "usage_type": "call"}, {"api_name": "app.game.component.attribute.ItemAttributeComponent.ItemAttributeComponent", "line_number": 25, "usage_type": "call"}, {"api_name": "app.game.component.pack.ItemPackComponent.ItemPackComponent", "line_number": 26, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 57, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 57, "usage_type": "attribute"}, {"api_name": "app.dbfront.memmode.tb_item_admin.new", "line_number": 58, "usage_type": "call"}, {"api_name": "app.dbfront.memmode.tb_item_admin", "line_number": 58, "usage_type": "name"}, {"api_name": "app.dbfront.memmode.tb_item_admin.deleteMode", "line_number": 66, "usage_type": "call"}, {"api_name": "app.dbfront.memmode.tb_item_admin", "line_number": 66, "usage_type": "name"}]}
+{"seq_id": "648918596", "text": "from selenium import webdriver\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.common.keys import Keys\nfrom selenium.webdriver.support.ui import Select\nfrom selenium.common.exceptions import NoSuchElementException\nfrom selenium.common.exceptions import NoAlertPresentException\nimport unittest, time,os\nclass Baidu1(unittest.TestCase):\n\n\n def setUp(self):\n\n\n self.driver = webdriver.Chrome()\n self.driver.implicitly_wait(30)\n self.base_url = \"http://www.baidu.com/\"\n self.verificationErrors = []\n self.accept_next_alert = True\n\n\n\n # def test_baidusearch(self):\n #\n #\n # driver = self.driver\n # driver.get(self.base_url + \"/\")\n # driver.find_element_by_id(\"kw\").click()\n # driver.find_element_by_id(\"kw\").clear()\n#\n# 批量执行脚本\n# 构建测试套件\n# 完整的单元测试很少只执行一个测试用例,开发人员通常都需要编写多个测试用例才能对某一软件功能进行比较完\n# 整的测试,这些相关的测试用例称为一个测试用例集,在unittest中是用TestSuite\n# 类来表示的。\n# driver.find_element_by_id(\"kw\").send_keys(u\"测试\")\n# driver.find_element_by_id(\"su\").click()\n\n\n\n def test_hao(self):\n\n\n driver = self.driver\n driver.get(self.base_url + \"/\")\n driver.find_element_by_link_text(\"hao123\").click()\n time.sleep(2)\n try:\n self.assertEqual(driver.title,\"123\", driver.title)\n except:\n self.saveScreenShot(driver,\"hao.png\")\n\n def saveScreenShot(self,driver,filename):\n if not os.path.exists('./image'):\n os.makedirs('./image')\n now = time.strftime(\"%Y%m%d-%H%M%S\",time.localtime(time.time()))\n driver.get_screenshot_as_file('./image/'+now+filename)\n\n def is_element_present(self, how, what):\n\n\n try:\n self.driver.find_element( by =how, value=what)\n except NoSuchElementException as e:\n return False\n return True\n\n\n# 判断alert是否存在,可删除\n def is_alert_present(self):\n\n\n try:\n self.driver.switch_to_alert()\n except NoAlertPresentException as e:\n return False\n return True\n\n\n def close_alert_and_get_its_text(self):\n\n\n try:\n alert = self.driver.switch_to_alert()\n alert_text = alert.text\n if self.accept_next_alert:\n alert.accept()\n else:\n alert.dismiss()\n return alert_text\n finally: self.accept_next_alert = True\n\n\n# test fixture,清除环境\n def tearDown(self):\n\n\n self.driver.quit()\n self.assertEqual([], self.verificationErrors)\n\nif __name__ == \"__main__\":\n unittest.main()\n", "sub_path": "unittestwork/Tcase.py", "file_name": "Tcase.py", "file_ext": "py", "file_size_in_byte": 2781, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "unittest.TestCase", "line_number": 8, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 14, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 14, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 54, "usage_type": "call"}, {"api_name": "time.strftime", "line_number": 55, "usage_type": "call"}, {"api_name": "time.localtime", "line_number": 55, "usage_type": "call"}, {"api_name": "time.time", "line_number": 55, "usage_type": "call"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 63, "usage_type": "name"}, {"api_name": "selenium.common.exceptions.NoAlertPresentException", "line_number": 74, "usage_type": "name"}, {"api_name": "unittest.main", "line_number": 101, "usage_type": "call"}]}
+{"seq_id": "153864555", "text": "\"\"\" \ncode courtesy of \nhttps://github.com/erikwijmans/Pointnet2_PyTorch\n\"\"\"\n\nimport torch\nimport faiss\nimport numpy as np\nfrom threading import Thread\n\nfrom .._ext import sampling\nfrom .._ext import linalg\n\nif torch.cuda.is_available():\n from .faiss_setup import GPU_RES\n\n\ndef normalize_point_batch(pc, NCHW=True):\n \"\"\"\n normalize a batch of point clouds\n :param\n pc [B, N, 3] or [B, 3, N]\n NCHW if True, treat the second dimension as channel dimension\n :return\n pc normalized point clouds, same shape as input\n centroid [B, 1, 3] or [B, 3, 1] center of point clouds\n furthest_distance [B, 1, 1] scale of point clouds\n \"\"\"\n point_axis = 2 if NCHW else 1\n dim_axis = 1 if NCHW else 2\n centroid = torch.mean(pc, dim=point_axis, keepdim=True)\n pc = pc - centroid\n furthest_distance, _ = torch.max(\n torch.sqrt(torch.sum(pc ** 2, dim=dim_axis, keepdim=True)), dim=point_axis, keepdim=True)\n pc = pc / furthest_distance\n return pc, centroid, furthest_distance\n\n\ndef channel_shuffle(x, groups=2):\n '''Channel shuffle: [N,C,H,W] -> [N,g,C/g,H,W] -> [N,C/g,g,H,w] -> [N,C,H,W]'''\n N, C, H, W = x.size()\n g = groups\n return x.view(N, g, C/g, H, W).permute(0, 2, 1, 3, 4).reshape(N, C, H, W)\n\n\ndef jitter_perturbation_point_cloud(batch_data, sigma=0.005, clip=0.02, is_2D=False, NCHW=True):\n if NCHW:\n batch_data = batch_data.transpose(1, 2)\n\n batch_size = batch_data.shape[0]\n chn = 2 if is_2D else 3\n jittered_data = sigma * torch.randn_like(batch_data)\n for b in range(batch_size):\n jittered_data[b].clamp_(-clip[b].item(), clip[b].item())\n jittered_data[:, :, chn:] = 0\n jittered_data += batch_data\n if NCHW:\n jittered_data = jittered_data.transpose(1, 2)\n return jittered_data\n\n\ndef __swig_ptr_from_FloatTensor(x):\n assert x.is_contiguous()\n assert x.dtype == torch.float32\n return faiss.cast_integer_to_float_ptr(x.storage().data_ptr())\n\n\ndef __swig_ptr_from_LongTensor(x):\n assert x.is_contiguous()\n assert x.dtype == torch.int64, 'dtype=%s' % x.dtype\n return faiss.cast_integer_to_long_ptr(x.storage().data_ptr())\n\n\ndef search_index_pytorch(database, x, k):\n \"\"\"\n KNN search via Faiss\n :param\n database BxNxC\n x BxMxC\n :return\n D BxMxK\n I BxMxK\n \"\"\"\n Dptr = database.storage().data_ptr()\n if not (x.is_cuda or database.is_cuda):\n index = faiss.IndexFlatL2(database.size(-1))\n else:\n index = faiss.GpuIndexFlatL2(\n GPU_RES, database.size(-1)) # dimension is 3\n index.add_c(database.size(0), faiss.cast_integer_to_float_ptr(Dptr))\n\n assert x.is_contiguous()\n n, d = x.size()\n assert d == index.d\n\n D = torch.empty((n, k), dtype=torch.float32, device=x.device)\n I = torch.empty((n, k), dtype=torch.int64, device=x.device)\n\n torch.cuda.synchronize()\n xptr = __swig_ptr_from_FloatTensor(x)\n Iptr = __swig_ptr_from_LongTensor(I)\n Dptr = __swig_ptr_from_FloatTensor(D)\n index.search_c(n, xptr,\n k, Dptr, Iptr)\n torch.cuda.synchronize()\n index.reset()\n return D, I\n\n\nclass KNN(torch.autograd.Function):\n @staticmethod\n def forward(ctx, k, query, points):\n \"\"\"\n :param k: k in KNN\n query: BxMxC\n points: BxNxC\n :return:\n neighbors_points: BxMxK\n index_batch: BxMxK\n \"\"\"\n # selected_gt: BxkxCxM\n # process each batch independently.\n index_batch = []\n distance_batch = []\n for i in range(points.shape[0]):\n D_var, I_var = search_index_pytorch(points[i], query[i], k)\n GPU_RES.syncDefaultStreamCurrentDevice()\n index_batch.append(I_var) # M, k\n distance_batch.append(D_var) # M, k\n\n # B, M, K\n index_batch = torch.stack(index_batch, dim=0)\n distance_batch = torch.stack(distance_batch, dim=0)\n ctx.mark_non_differentiable(index_batch, distance_batch)\n return index_batch, distance_batch\n\n\ndef faiss_knn(k, query, points, NCHW=True):\n \"\"\"\n group batch of points to neighborhoods\n :param\n k: neighborhood size\n query: BxCxM or BxMxC\n points: BxCxN or BxNxC\n NCHW: if true, the second dimension is the channel dimension\n :return\n neighbor_points BxCxMxk (if NCHW) or BxMxkxC (otherwise)\n index_batch BxMxk\n distance_batch BxMxk\n \"\"\"\n if NCHW:\n batch_size, channels, num_points = points.size()\n points_trans = points.transpose(2, 1).contiguous()\n query_trans = query.transpose(2, 1).contiguous()\n else:\n points_trans = points.contiguous()\n query_trans = query.contiguous()\n\n batch_size, num_points, _ = points_trans.size()\n # BxMxk\n index_batch, distance_batch = KNN.apply(k, query_trans, points_trans)\n # BxNxC -> BxMxNxC\n points_expanded = points_trans.unsqueeze(dim=1).expand(\n (-1, query.size(1), -1, -1))\n # BxMxk -> BxMxkxC\n index_batch_expanded = index_batch.unsqueeze(dim=-1).expand(\n (-1, -1, -1, points_trans.size(-1)))\n # BxMxkxC\n neighbor_points = torch.gather(points_expanded, 2, index_batch_expanded)\n index_batch = index_batch\n if NCHW:\n # BxCxMxk\n neighbor_points = neighbor_points.permute(0, 3, 1, 2).contiguous()\n return neighbor_points, index_batch, distance_batch\n\n\ndef __batch_distance_matrix_general(A, B):\n \"\"\"\n :param\n A, B [B,N,C], [B,M,C]\n :return\n D [B,N,M]\n \"\"\"\n r_A = torch.sum(A * A, dim=2, keepdim=True)\n r_B = torch.sum(B * B, dim=2, keepdim=True)\n m = torch.matmul(A, B.permute(0, 2, 1))\n D = r_A - 2 * m + r_B.permute(0, 2, 1)\n return D\n\n\ndef group_knn(k, query, points, unique=True, NCHW=True):\n \"\"\"\n group batch of points to neighborhoods\n :param\n k: neighborhood size\n query: BxCxM or BxMxC\n points: BxCxN or BxNxC\n unique: neighborhood contains *unique* points\n NCHW: if true, the second dimension is the channel dimension\n :return\n neighbor_points BxCxMxk (if NCHW) or BxMxkxC (otherwise)\n index_batch BxMxk\n distance_batch BxMxk\n \"\"\"\n if NCHW:\n batch_size, channels, num_points = points.size()\n points_trans = points.transpose(2, 1).contiguous()\n query_trans = query.transpose(2, 1).contiguous()\n else:\n points_trans = points.contiguous()\n query_trans = query.contiguous()\n\n batch_size, num_points, _ = points_trans.size()\n assert(num_points >= k\n ), \"points size must be greater or equal to k\"\n\n D = __batch_distance_matrix_general(query_trans, points_trans)\n if unique:\n # prepare duplicate entries\n points_np = points_trans.detach().cpu().numpy()\n indices_duplicated = np.ones(\n (batch_size, 1, num_points), dtype=np.int32)\n\n for idx in range(batch_size):\n _, indices = np.unique(points_np[idx], return_index=True, axis=0)\n indices_duplicated[idx, :, indices] = 0\n\n indices_duplicated = torch.from_numpy(\n indices_duplicated).to(device=D.device, dtype=torch.float32)\n D += torch.max(D) * indices_duplicated\n\n # (B,M,k)\n distances, point_indices = torch.topk(-D, k, dim=-1, sorted=True)\n # (B,N,C)->(B,M,N,C), (B,M,k)->(B,M,k,C)\n knn_trans = torch.gather(points_trans.unsqueeze(1).expand(-1, query_trans.size(1), -1, -1),\n 2,\n point_indices.unsqueeze(-1).expand(-1, -1, -1, points_trans.size(-1)))\n\n if NCHW:\n knn_trans = knn_trans.permute(0, 3, 1, 2)\n\n return knn_trans, point_indices, -distances\n\n\nclass GatherFunction(torch.autograd.Function):\n @staticmethod\n def forward(ctx, features, idx):\n r\"\"\"\n Parameters\n ----------\n features : torch.Tensor\n (B, C, N) tensor\n idx : torch.Tensor\n (B, npoint) tensor of the features to gather\n Returns\n -------\n torch.Tensor\n (B, C, npoint) tensor\n \"\"\"\n features = features.contiguous()\n idx = idx.contiguous()\n idx = idx.to(dtype=torch.int32)\n\n B, npoint = idx.size()\n _, C, N = features.size()\n\n output = torch.empty(\n B, C, npoint, dtype=features.dtype, device=features.device)\n output = sampling.gather_forward(\n B, C, N, npoint, features, idx, output\n )\n\n ctx.save_for_backward(idx)\n ctx.C = C\n ctx.N = N\n return output\n\n @staticmethod\n def backward(ctx, grad_out):\n idx, = ctx.saved_tensors\n B, npoint = idx.size()\n\n grad_features = torch.zeros(\n B, ctx.C, ctx.N, dtype=grad_out.dtype, device=grad_out.device)\n grad_features = sampling.gather_backward(\n B, ctx.C, ctx.N, npoint, grad_out.contiguous(), idx, grad_features\n )\n\n return grad_features, None\n\n\ngather_points = GatherFunction.apply\n\n\nclass BallQuery(torch.autograd.Function):\n @staticmethod\n def forward(ctx, radius, nsample, xyz, new_xyz):\n r\"\"\"\n Parameters\n ----------\n radius : float\n radius of the balls\n nsample : int\n maximum number of features in the balls\n xyz : torch.Tensor\n (B, N, 3) xyz coordinates of the features\n new_xyz : torch.Tensor\n (B, npoint, 3) centers of the ball query\n Returns\n -------\n torch.Tensor\n (B, npoint, nsample) tensor with the indicies of the features that form the query balls\n \"\"\"\n return sampling.ball_query(new_xyz, xyz, radius, nsample)\n\n @staticmethod\n def backward(ctx, a=None):\n return None, None, None, None\n\n\nball_query = BallQuery.apply\n\n\nclass GroupingOperation(torch.autograd.Function):\n @staticmethod\n def forward(ctx, features, idx):\n r\"\"\"\n Parameters\n ----------\n features : torch.Tensor\n (B, C, N) tensor of features to group\n idx : torch.Tensor\n (B, npoint, nsample) tensor containing the indicies of features to group with\n Returns\n -------\n torch.Tensor\n (B, C, npoint, nsample) tensor\n \"\"\"\n B, nfeatures, nsample = idx.size()\n _, C, N = features.size()\n\n ctx.for_backwards = (idx, N)\n\n return sampling.group_points(features, idx)\n\n @staticmethod\n def backward(ctx, grad_out):\n r\"\"\"\n Parameters\n ----------\n grad_out : torch.Tensor\n (B, C, npoint, nsample) tensor of the gradients of the output from forward\n Returns\n -------\n torch.Tensor\n (B, C, N) gradient of the features\n None\n \"\"\"\n idx, N = ctx.for_backwards\n\n grad_features = sampling.group_points_grad(grad_out.contiguous(), idx, N)\n\n return grad_features, None\n\n\ngrouping_operation = GroupingOperation.apply\n\n\nclass QueryAndGroup(torch.nn.Module):\n r\"\"\"\n Groups with a ball query of radius\n Parameters\n ---------\n radius : float32\n Radius of ball\n nsample : int32\n Maximum number of features to gather in the ball\n \"\"\"\n\n def __init__(self, radius, nsample, use_xyz=True):\n super(QueryAndGroup, self).__init__()\n self.radius, self.nsample, self.use_xyz = radius, nsample, use_xyz\n\n def forward(self, xyz, new_xyz, features=None):\n r\"\"\"\n Parameters\n ----------\n xyz : torch.Tensor\n xyz coordinates of the features (B, N, 3)\n new_xyz : torch.Tensor\n centriods (B, npoint, 3)\n features : torch.Tensor\n Descriptors of the features (B, C, N)\n Returns\n -------\n new_features : torch.Tensor\n (B, 3 + C, npoint, nsample) tensor\n \"\"\"\n # (B, npoint, k)\n idx = ball_query(self.radius, self.nsample, xyz, new_xyz)\n # (B, 3, N)\n xyz_trans = xyz.transpose(1, 2).contiguous()\n grouped_xyz = grouping_operation(xyz_trans, idx) # (B, 3, npoint, nsample)\n grouped_xyz -= new_xyz.transpose(1, 2).unsqueeze(-1)\n\n if features is not None:\n grouped_features = grouping_operation(features, idx)\n if self.use_xyz:\n new_features = torch.cat(\n [grouped_xyz, grouped_features], dim=1\n ) # (B, C + 3, npoint, nsample)\n else:\n new_features = grouped_features\n else:\n assert (\n self.use_xyz\n ), \"Cannot have not features and not use xyz as a feature!\"\n new_features = grouped_xyz\n\n return new_features\n\n\nclass FurthestPointSampling(torch.autograd.Function):\n\n @staticmethod\n def forward(ctx, xyz, npoint, seedIdx):\n r\"\"\"\n Uses iterative furthest point sampling to select a set of npoint features that have the largest\n minimum distance\n Parameters\n ----------\n xyz : torch.Tensor\n (B, N, 3) tensor where N > npoint\n npoint : int32\n number of features in the sampled set\n Returns\n -------\n torch.LongTensor\n (B, npoint) tensor containing the indices\n\n \"\"\"\n B, N, _ = xyz.size()\n\n idx = torch.empty([B, npoint], dtype=torch.int32, device=xyz.device)\n temp = torch.full([B, N], 1e10, dtype=torch.float32, device=xyz.device)\n\n sampling.furthest_sampling(\n npoint, seedIdx, xyz, temp, idx\n )\n ctx.mark_non_differentiable(idx)\n return idx\n\n\n__furthest_point_sample = FurthestPointSampling.apply\n\n\ndef furthest_point_sample(xyz, npoint, NCHW=True, seedIdx=0):\n \"\"\"\n :param\n xyz (B, 3, N) or (B, N, 3)\n npoint a constant\n :return\n torch.LongTensor\n (B, npoint) tensor containing the indices\n torch.FloatTensor\n (B, npoint, 3) or (B, 3, npoint) point sets\"\"\"\n assert(xyz.dim() == 3), \"input for furthest sampling must be a 3D-tensor, but xyz.size() is {}\".format(xyz.size())\n # need transpose\n if NCHW:\n xyz = xyz.transpose(2, 1).contiguous()\n\n assert(xyz.size(2) == 3), \"furthest sampling is implemented for 3D points\"\n idx = __furthest_point_sample(xyz, npoint, seedIdx)\n sampled_pc = gather_points(xyz.transpose(2, 1).contiguous(), idx)\n if not NCHW:\n sampled_pc = sampled_pc.transpose(2, 1).contiguous()\n return idx, sampled_pc\n\n\nclass BatchSVDFunction(torch.autograd.Function):\n \"\"\"\n batched svd implemented by https://github.com/KinglittleQ/torch-batch-svd\n \"\"\"\n @staticmethod\n def forward(ctx, x):\n ctx.device = x.device\n if not torch.cuda.is_available():\n assert(RuntimeError), \"BatchSVDFunction only runs on gpu\"\n x = x.cuda()\n U, S, V = linalg.batch_svd_forward(x, True, 1e-7, 100)\n k = S.size(1)\n U = U[:, :, :k]\n V = V[:, :, :k]\n ctx.save_for_backward(x, U, S, V)\n U = U.to(ctx.device)\n S = S.to(ctx.device)\n V = V.to(ctx.device)\n return U, S, V\n\n @staticmethod\n def backward(ctx, grad_u, grad_s, grad_v):\n x, U, S, V = ctx.saved_variables\n\n grad_out = linalg.batch_svd_backward(\n [grad_u, grad_s, grad_v],\n x, True, True, U, S, V\n )\n\n return grad_out.to(device=ctx.device)\n\n\ndef batch_svd(x):\n \"\"\"\n input:\n x --- shape of [B, M, N], k = min(M,N)\n return:\n U, S, V = batch_svd(x) where x = USV^T\n U [M, k]\n V [N, k]\n S [B, k] in decending order\n \"\"\"\n assert(x.dim() == 3)\n return BatchSVDFunction.apply(x)\n\n\ndef batch_normals(points, base=None, nn_size=20, NCHW=True):\n \"\"\"\n compute normals vectors for batched points [B, C, M]\n If base is given, compute the normals of points using the neighborhood in base\n The direction of normal could flip.\n inputs:\n points: (B,C,M)\n base: (B,C,N)\n return:\n normals: (B,C,M)\n \"\"\"\n if base is None:\n base = points\n\n if NCHW:\n points = points.transpose(2, 1).contiguous()\n base = base.transpose(2, 1).contiguous()\n\n assert(nn_size < base.shape[1])\n batch_size, M, C = points.shape\n # B,M,k,C\n grouped_points, group_idx, _ = group_knn(nn_size, points, base, unique=True, NCHW=False)\n group_center = torch.mean(grouped_points, dim=2, keepdim=True)\n points = grouped_points - group_center\n allpoints = points.view(-1, nn_size, C).contiguous()\n # MB,C,k\n U, S, V = batch_svd(allpoints)\n # V is MBxCxC, last_u MBxC\n normals = V[:, :, -1]\n normals = normals.view(batch_size, M, C)\n if NCHW:\n normals = normals.transpose(1, 2)\n return normals\n\n\nif __name__ == '__main__':\n from ..utils import pc_utils\n cuda0 = torch.device('cuda:0')\n pc = pc_utils.read_ply(\"/home/ywang/Documents/points/point-upsampling/3PU/prepare_data/polygonmesh_base/build/data_PPU_output/training/112/angel4_aligned_2.ply\")\n pc = pc[:, :3]\n print(\"{} input points\".format(pc.shape[0]))\n pc_utils.save_ply(pc, \"./input.ply\", colors=None, normals=None)\n pc = torch.from_numpy(pc).requires_grad_().to(cuda0).unsqueeze(0)\n pc = pc.transpose(2, 1)\n\n # test furthest point\n idx, sampled_pc = furthest_point_sample(pc, 1250)\n output = sampled_pc.transpose(2, 1).cpu().squeeze()\n pc_utils.save_ply(output.detach(), \"./output.ply\", colors=None, normals=None)\n\n # test KNN\n knn_points, _, _ = group_knn(10, sampled_pc, pc, NCHW=True) # B, C, M, K\n labels = torch.arange(0, knn_points.size(2)).unsqueeze_(\n 0).unsqueeze_(0).unsqueeze_(-1) # 1, 1, M, 1\n labels = labels.expand(knn_points.size(0), -1, -1,\n knn_points.size(3)) # B, 1, M, K\n # B, C, P\n labels = torch.cat(torch.unbind(labels, dim=-1), dim=-1).squeeze().detach().cpu().numpy()\n knn_points = torch.cat(torch.unbind(knn_points, dim=-1),\n dim=-1).transpose(2, 1).squeeze(0).detach().cpu().numpy()\n pc_utils.save_ply_property(knn_points, labels, \"./knn_output.ply\", cmap_name='jet')\n\n from torch.autograd import gradcheck\n # test = gradcheck(furthest_point_sample, [pc, 1250], eps=1e-6, atol=1e-4)\n # print(test)\n test = gradcheck(gather_points, [pc.to(\n dtype=torch.float64), idx], eps=1e-6, atol=1e-4)\n\n print(test)\n", "sub_path": "pytorch_points/pytorch_points/network/operations.py", "file_name": "operations.py", "file_ext": "py", "file_size_in_byte": 18531, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "torch.cuda.is_available", "line_number": 14, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 14, "usage_type": "attribute"}, {"api_name": "torch.mean", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.sqrt", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.randn_like", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 64, "usage_type": "attribute"}, {"api_name": "faiss.cast_integer_to_float_ptr", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.int64", "line_number": 70, "usage_type": "attribute"}, {"api_name": "faiss.cast_integer_to_long_ptr", "line_number": 71, "usage_type": "call"}, {"api_name": "faiss.IndexFlatL2", "line_number": 86, "usage_type": "call"}, {"api_name": "faiss.GpuIndexFlatL2", "line_number": 88, "usage_type": "call"}, {"api_name": "faiss_setup.GPU_RES", "line_number": 89, "usage_type": "argument"}, {"api_name": "faiss.cast_integer_to_float_ptr", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.empty", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 96, "usage_type": "attribute"}, {"api_name": "torch.empty", "line_number": 97, "usage_type": "call"}, {"api_name": "torch.int64", "line_number": 97, "usage_type": "attribute"}, {"api_name": "torch.cuda.synchronize", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 99, "usage_type": "attribute"}, {"api_name": "torch.cuda.synchronize", "line_number": 105, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 105, "usage_type": "attribute"}, {"api_name": "torch.autograd", "line_number": 110, "usage_type": "attribute"}, {"api_name": "faiss_setup.GPU_RES.syncDefaultStreamCurrentDevice", "line_number": 127, "usage_type": "call"}, {"api_name": "faiss_setup.GPU_RES", "line_number": 127, "usage_type": "name"}, {"api_name": "torch.stack", "line_number": 132, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 133, "usage_type": "call"}, {"api_name": "torch.gather", "line_number": 169, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 184, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 185, "usage_type": "call"}, {"api_name": "torch.matmul", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 222, "usage_type": "attribute"}, {"api_name": "numpy.unique", "line_number": 225, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 228, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 229, "usage_type": "attribute"}, {"api_name": "torch.max", "line_number": 230, "usage_type": "call"}, {"api_name": "torch.topk", "line_number": 233, "usage_type": "call"}, {"api_name": "torch.gather", "line_number": 235, "usage_type": "call"}, {"api_name": "torch.autograd", "line_number": 245, "usage_type": "attribute"}, {"api_name": "torch.int32", "line_number": 262, "usage_type": "attribute"}, {"api_name": "torch.empty", "line_number": 267, "usage_type": "call"}, {"api_name": "_ext.sampling.gather_forward", "line_number": 269, "usage_type": "call"}, {"api_name": "_ext.sampling", "line_number": 269, "usage_type": "name"}, {"api_name": "torch.zeros", "line_number": 283, "usage_type": "call"}, {"api_name": "_ext.sampling.gather_backward", "line_number": 285, "usage_type": "call"}, {"api_name": "_ext.sampling", "line_number": 285, "usage_type": "name"}, {"api_name": "torch.autograd", "line_number": 295, "usage_type": "attribute"}, {"api_name": "_ext.sampling.ball_query", "line_number": 314, "usage_type": "call"}, {"api_name": "_ext.sampling", "line_number": 314, "usage_type": "name"}, {"api_name": "torch.autograd", "line_number": 324, "usage_type": "attribute"}, {"api_name": "_ext.sampling.group_points", "line_number": 344, "usage_type": "call"}, {"api_name": "_ext.sampling", "line_number": 344, "usage_type": "name"}, {"api_name": "_ext.sampling.group_points_grad", "line_number": 361, "usage_type": "call"}, {"api_name": "_ext.sampling", "line_number": 361, "usage_type": "name"}, {"api_name": "torch.nn", "line_number": 369, "usage_type": "attribute"}, {"api_name": "torch.cat", "line_number": 409, "usage_type": "call"}, {"api_name": "torch.autograd", "line_number": 423, "usage_type": "attribute"}, {"api_name": "torch.empty", "line_number": 444, "usage_type": "call"}, {"api_name": "torch.int32", "line_number": 444, "usage_type": "attribute"}, {"api_name": "torch.full", "line_number": 445, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 445, "usage_type": "attribute"}, {"api_name": "_ext.sampling.furthest_sampling", "line_number": 447, "usage_type": "call"}, {"api_name": "_ext.sampling", "line_number": 447, "usage_type": "name"}, {"api_name": "torch.autograd", "line_number": 480, "usage_type": "attribute"}, {"api_name": "torch.cuda.is_available", "line_number": 487, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 487, "usage_type": "attribute"}, {"api_name": "_ext.linalg.batch_svd_forward", "line_number": 490, "usage_type": "call"}, {"api_name": "_ext.linalg", "line_number": 490, "usage_type": "name"}, {"api_name": "_ext.linalg.batch_svd_backward", "line_number": 504, "usage_type": "call"}, {"api_name": "_ext.linalg", "line_number": 504, "usage_type": "name"}, {"api_name": "torch.mean", "line_number": 548, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 563, "usage_type": "call"}, {"api_name": "utils.pc_utils.read_ply", "line_number": 564, "usage_type": "call"}, {"api_name": "utils.pc_utils", "line_number": 564, "usage_type": "name"}, {"api_name": "utils.pc_utils.save_ply", "line_number": 567, "usage_type": "call"}, {"api_name": "utils.pc_utils", "line_number": 567, "usage_type": "name"}, {"api_name": "torch.from_numpy", "line_number": 568, "usage_type": "call"}, {"api_name": "utils.pc_utils.save_ply", "line_number": 574, "usage_type": "call"}, {"api_name": "utils.pc_utils", "line_number": 574, "usage_type": "name"}, {"api_name": "torch.arange", "line_number": 578, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 583, "usage_type": "call"}, {"api_name": "torch.unbind", "line_number": 583, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 584, "usage_type": "call"}, {"api_name": "torch.unbind", "line_number": 584, "usage_type": "call"}, {"api_name": "utils.pc_utils.save_ply_property", "line_number": 586, "usage_type": "call"}, {"api_name": "utils.pc_utils", "line_number": 586, "usage_type": "name"}, {"api_name": "torch.autograd.gradcheck", "line_number": 591, "usage_type": "call"}, {"api_name": "torch.float64", "line_number": 592, "usage_type": "attribute"}]}
+{"seq_id": "416500936", "text": "from setuptools import find_packages\nfrom setuptools import setup\n\nREQUIRED_PACKAGES = ['h5py>=2.7.0', 'Keras>=2.0.6', 'Theano>=0.9.0', 'matplotlib>=2.0.2', 'scikit-learn>=0.18', 'hyperopt>=0.1', 'google-cloud-storage']\n\nsetup(\n name='trainer',\n version='3.0',\n install_requires = REQUIRED_PACKAGES,\n packages = find_packages(),\n include_package_data = True,\n description='MRRDT CNN Cascade Trainer Package'\n)\n", "sub_path": "Vision/Face Detection/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 428, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "setuptools.setup", "line_number": 6, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 10, "usage_type": "call"}]}
+{"seq_id": "463648338", "text": "\"\"\"loads acoustic brainz features for mbids in msd_bb_mbid dataset.\"\"\"\nimport os.path\n\nfrom logzero import logger\nimport pandas as pd\n\nfrom hit_prediction_code.dataloaders import acousticbrainz\n\npath_prefix = 'data/hit_song_prediction_ismir2020/interim'\n\n\ndef _load_features(name, feature):\n songs = pd.read_csv(os.path.join(path_prefix, name + '.csv'))['mbid']\n mbids = set(songs)\n\n logger.info('Combine %s features for %s' % (feature, name))\n data = acousticbrainz.load_ab_features_as_df(mbids, feature)\n\n if feature == 'll':\n version_col = 'metadata.version.essentia_git_sha'\n else:\n version_col = 'metadata.version.lowlevel.essentia_git_sha'\n\n def is_clean_version(version):\n try:\n return not version.startswith('v2.1_beta1')\n except Exception as ex:\n logger.exception(ex)\n logger.error('version is: %s' % str(version))\n\n # Ignore sample if the value is no string\n return False\n\n data = data[data[version_col].apply(is_clean_version)]\n\n filename = name + '_ab_' + feature + '_features.parquet'\n\n logger.info('Store %s' % filename)\n data.to_parquet(os.path.join(path_prefix, filename))\n\n\n# combine all features\nlogger.info('Combine acousticbrainz features for msd_bb_mbid')\ndatasets = [\n 'msd_bb_mbid_cleaned_matches',\n 'msd_bb_mbid_exact_matches',\n 'msd_bb_mbid_non_matches',\n]\nfor filename in datasets:\n for feature_type in ['ll', 'hl']:\n _load_features(filename, feature_type)\n", "sub_path": "tools/hsp-s/merge_msd_bb_mbid_and_ab.py", "file_name": "merge_msd_bb_mbid_and_ab.py", "file_ext": "py", "file_size_in_byte": 1518, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pandas.read_csv", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 13, "usage_type": "name"}, {"api_name": "logzero.logger.info", "line_number": 16, "usage_type": "call"}, {"api_name": "logzero.logger", "line_number": 16, "usage_type": "name"}, {"api_name": "hit_prediction_code.dataloaders.acousticbrainz.load_ab_features_as_df", "line_number": 17, "usage_type": "call"}, {"api_name": "hit_prediction_code.dataloaders.acousticbrainz", "line_number": 17, "usage_type": "name"}, {"api_name": "logzero.logger.exception", "line_number": 28, "usage_type": "call"}, {"api_name": "logzero.logger", "line_number": 28, "usage_type": "name"}, {"api_name": "logzero.logger.error", "line_number": 29, "usage_type": "call"}, {"api_name": "logzero.logger", "line_number": 29, "usage_type": "name"}, {"api_name": "logzero.logger.info", "line_number": 38, "usage_type": "call"}, {"api_name": "logzero.logger", "line_number": 38, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 39, "usage_type": "name"}, {"api_name": "logzero.logger.info", "line_number": 43, "usage_type": "call"}, {"api_name": "logzero.logger", "line_number": 43, "usage_type": "name"}]}
+{"seq_id": "359452747", "text": "\"\"\"Mini-project 2\n@author: Richard Ballaux\"\"\"\n\nimport random\nimport math\nfrom PIL import Image\n\n\ndef build_random_function(min_depth, max_depth):\n \"\"\"Build a random function.\n\n Builds a random function of depth at least min_depth and depth at most\n max_depth. (See the assignment write-up for the definition of depth\n in this context)\n\n Args:\n min_depth: the minimum depth of the random function\n max_depth: the maximum depth of the random function\n\n Returns:\n The randomly generated function represented as a nested list.\n (See the assignment writ-eup for details on the representation of\n these functions)\n \"\"\"\n depth = random.randint(min_depth,max_depth)\n #print('depth',depth)\n functionNumber = random.randint(0,5)\n #print('functionNumber',functionNumber)\n if depth == 1:\n return random.choice([[\"x\"],[\"y\"]])\n #choose randomly between x or y\n else:\n #if depth is different from 1 go deeper\n if functionNumber == 0:\n #product\n return [\"prod\",build_random_function(depth-1,depth-1),build_random_function(depth-1,depth-1)]\n if functionNumber == 1:\n return [\"avg\",build_random_function(depth-1,depth-1),build_random_function(depth-1,depth-1)]\n if functionNumber == 2:\n return [\"cos_pi\",build_random_function(depth-1,depth-1)]\n if functionNumber == 3:\n return [\"sin_pi\",build_random_function(depth-1,depth-1)]\n\n if functionNumber == 4:\n #power 2\n return [\"power2\",build_random_function(depth-1,depth-1)]\n if functionNumber == 5:\n #sqrt\n return [\"sqrt\",build_random_function(depth-1,depth-1)]\n if functionNumber == 6:\n # just x\n return [\"x\"]\n if functionNumber == 7:\n # just y\n return [\"y\"]\n\n#print(build_random_function(3,5))\n\n\ndef evaluate_random_function(f, x, y):\n \"\"\"Evaluate the random function f with inputs x,y.\n\n The representation of the function f is defined in the assignment write-up.\n\n Args:\n f: the function to evaluate\n x: the value of x to be used to evaluate the function\n y: the value of y to be used to evaluate the function\n\n Returns:\n The function value\n\n Examples:\n >>> evaluate_random_function([\"x\"],-0.5, 0.75)\n -0.5\n >>> evaluate_random_function([\"y\"],0.1,0.02)\n 0.02\n\n \"\"\"\n if f[0] == \"prod\":\n return evaluate_random_function(f[1],x,y)*evaluate_random_function(f[2],x,y)\n if f[0] == \"avg\":\n return 0.5*(evaluate_random_function(f[1],x,y)+evaluate_random_function(f[2],x,y))\n if f[0] == \"cos_pi\":\n return math.cos(math.pi*evaluate_random_function(f[1],x,y))\n if f[0] == \"sin_pi\":\n return math.sin(math.pi*evaluate_random_function(f[1],x,y))\n if f[0] == \"x\":\n return x\n if f[0] == \"y\":\n return y\n if f[0] == \"power2\":\n return math.pow(evaluate_random_function(f[1],x,y),2)\n if f[0] == \"sqrt\":\n return math.sqrt(math.fabs(evaluate_random_function(f[1],x,y)))\n #filled_in_string = f[0].replace(\"x\",str(x))\n #filled_in_string = filled_in_string.replace(\"y\",str(y))\n #return eval(filled_in_string)\n\n\ndef remap_interval(val,\n input_interval_start,\n input_interval_end,\n output_interval_start,\n output_interval_end):\n \"\"\"Remap a value from one interval to another.\n\n Given an input value in the interval [input_interval_start,\n input_interval_end], return an output value scaled to fall within\n the output interval [output_interval_start, output_interval_end].\n\n Args:\n val: the value to remap\n input_interval_start: the start of the interval that contains all\n possible values for val\n input_interval_end: the end of the interval that contains all possible\n values for val\n output_interval_start: the start of the interval that contains all\n possible output values\n output_inteval_end: the end of the interval that contains all possible\n output values\n\n Returns:\n The value remapped from the input to the output interval\n\n Examples:\n >>> remap_interval(0.5, 0, 1, 0, 10)\n 5.0\n >>> remap_interval(5, 4, 6, 0, 2)\n 1.0\n >>> remap_interval(5, 4, 6, 1, 2)\n 1.5\n \"\"\"\n return (val-input_interval_start)*(output_interval_end-output_interval_start)/(input_interval_end-input_interval_start) + output_interval_start;\n\n\n\ndef color_map(val):\n \"\"\"Maps input value between -1 and 1 to an integer 0-255, suitable for use as an RGB color code.\n\n Args:\n val: value to remap, must be a float in the interval [-1, 1]\n\n Returns:\n An integer in the interval [0,255]\n\n Examples:\n >>> color_map(-1.0)\n 0\n >>> color_map(1.0)\n 255\n >>> color_map(0.0)\n 127\n >>> color_map(0.5)\n 191\n \"\"\"\n # NOTE: This relies on remap_interval, which you must provide\n color_code = remap_interval(val, -1, 1, 0, 255)\n return int(color_code)\n\n\ndef test_image(filename, x_size=350, y_size=350):\n \"\"\"Generate a test image with random pixels and save as an image file.\n\n Args:\n filename: string filename for image (should be .png)\n x_size, y_size: optional args to set image dimensions (default: 350)\n \"\"\"\n # Create image and loop over all pixels\n im = Image.new(\"RGB\", (x_size, y_size))\n pixels = im.load()\n for i in range(x_size):\n for j in range(y_size):\n x = remap_interval(i, 0, x_size, -1, 1)\n y = remap_interval(j, 0, y_size, -1, 1)\n pixels[i, j] = (random.randint(0, 255), # Red channel\n random.randint(0, 255), # Green channel\n random.randint(0, 255)) # Blue channel\n\n im.save(filename)\n\n\ndef generate_art(filename, x_size=350, y_size=350):\n \"\"\"Generate computational art and save as an image file.\n\n Args:\n filename: string filename for image (should be .png)\n x_size, y_size: optional args to set image dimensions (default: 350)\n \"\"\"\n # Functions for red, green, and blue channels - where the magic happens!\n red_function = build_random_function(7,9)\n print(red_function)\n green_function = build_random_function(7,9)\n print(green_function)\n blue_function = build_random_function(7,9)\n print(blue_function)\n\n # Create image and loop over all pixels\n im = Image.new(\"RGB\", (x_size, y_size))\n pixels = im.load()\n for i in range(x_size):\n for j in range(y_size):\n x = remap_interval(i, 0, x_size, -1, 1)\n y = remap_interval(j, 0, y_size, -1, 1)\n pixels[i, j] = (\n color_map(evaluate_random_function(red_function, x, y)),\n color_map(evaluate_random_function(green_function, x, y)),\n color_map(evaluate_random_function(blue_function, x, y))\n )\n\n im.save(filename)\n\n\nif __name__ == '__main__':\n import doctest\n #doctest.testmod()\n #doctest.run_docstring_examples(evaluate_random_function,globals(),verbose=True)\n for i in range(35,40):\n name = \"myart\"+str(i)+\".png\"\n generate_art(name,1920,1080)\n #print(evaluate_random_function([\"x**2+5*y\"],2.2,5.3))\n", "sub_path": "recursive_art.py", "file_name": "recursive_art.py", "file_ext": "py", "file_size_in_byte": 7493, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "random.randint", "line_number": 25, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 27, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 30, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 85, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 85, "usage_type": "attribute"}, {"api_name": "math.sin", "line_number": 87, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 87, "usage_type": "attribute"}, {"api_name": "math.pow", "line_number": 93, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 95, "usage_type": "call"}, {"api_name": "math.fabs", "line_number": 95, "usage_type": "call"}, {"api_name": "PIL.Image.new", "line_number": 170, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 170, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 176, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 177, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 178, "usage_type": "call"}, {"api_name": "PIL.Image.new", "line_number": 199, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 199, "usage_type": "name"}]}
+{"seq_id": "155255740", "text": "import os\nimport boto3\nimport tempfile\nimport logging\nfrom sqlalchemy import text\nfrom multiprocessing.pool import ThreadPool\n\nlogger = logging.getLogger()\n\nfrom indra.util import batch_iter\nfrom indra_db.util import get_primary_db\n\nBUCKET = 'bigmech'\n\n\ndef stream_text_content(batch_size=1000000):\n db = get_primary_db()\n query_for_ids = 'SELECT id from text_content WHERE id >= 6000000'\n id_stream = db.session.execute(text(query_for_ids))\n query_for_content = \"\"\"SELECT\n text_ref_id, source, format, text_type, content\n FROM\n text_content\n WHERE\n id in :tcids\n \"\"\"\n for tcid_batch in batch_iter(id_stream, batch_size):\n tcids = tuple(tcid[0] for tcid in tcid_batch)\n logger.info('gathering content')\n res = db.session.execute(text(query_for_content), {'tcids': tcids})\n yield res\n\n\nclient = boto3.client('s3')\n\n\ndef upload_file(x):\n key, content = x\n with tempfile.NamedTemporaryFile() as temp:\n with open(temp.name, 'wb') as f:\n f.write(content)\n client.upload_file(temp.name, BUCKET, key)\n\n\ndef content_key_generator(batch):\n for text_ref_id, source, format_, text_type, content in batch:\n key = os.path.join('content', str(text_ref_id), source, format_,\n text_type, f'content.zlib')\n content = content.tobytes()\n yield (key, content)\n\n\ndef do_it():\n content_stream = stream_text_content()\n for i, batch in enumerate(content_stream):\n logger.info(f'uploading content for batch: {i}')\n with ThreadPool(512) as pool:\n pool.map(upload_file, content_key_generator(batch))\n\n\nif __name__ == '__main__':\n do_it()\n", "sub_path": "adeft_indra/put_everything_on_s3.py", "file_name": "put_everything_on_s3.py", "file_ext": "py", "file_size_in_byte": 1822, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.getLogger", "line_number": 8, "usage_type": "call"}, {"api_name": "indra_db.util.get_primary_db", "line_number": 17, "usage_type": "call"}, {"api_name": "sqlalchemy.text", "line_number": 19, "usage_type": "call"}, {"api_name": "indra.util.batch_iter", "line_number": 27, "usage_type": "call"}, {"api_name": "sqlalchemy.text", "line_number": 30, "usage_type": "call"}, {"api_name": "boto3.client", "line_number": 34, "usage_type": "call"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "multiprocessing.pool.ThreadPool", "line_number": 57, "usage_type": "call"}]}
+{"seq_id": "117190972", "text": "# -*- coding: utf-8 -*-\r\nfrom django.http import HttpResponse\r\nfrom django.http import HttpResponseRedirect\r\nfrom django.shortcuts import render_to_response\r\n\r\n\r\n\r\nfrom django.db.models import Sum, Q, F, Count, TextField\r\nfrom django.db.models.functions import Cast\r\n\r\nfrom django.utils.safestring import mark_safe\r\nfrom django.utils.numberformat import format\r\n\r\nfrom django.contrib.auth.models import User, Group, UserManager\r\n\r\nimport re, datetime\r\n\r\nfrom django import template\r\n\r\nregister = template.Library()\r\n\r\n\r\nfrom devapp.models import *\r\n\r\nfrom panel.models import profileuser\r\n\r\nfrom commentaryapp.models import *\r\n\r\n#\r\n@register.inclusion_tag('techtask_get_status_inc.html', takes_context=True)\r\ndef techtask_get_status_inc(context, **kwargs):\r\n\trequest = context['request']\r\n\t#\r\n\trun=0\r\n\ttest=0\r\n\tinit=0\r\n\tstop=0\r\n\tsuccess=0\r\n\t#\r\n\ttry:\r\n\t\ttechtask_instance=kwargs['object'] #techtask_instance\r\n\texcept:\r\n\t\tpass\r\n\telse:\r\n\t\tobject=techtask_instance\r\n\t\t#\r\n\t\ttechstep_list=techstep.objects.filter(techtask=techtask_instance) \r\n\t\trun=techstep_list.filter(status='run').count()\r\n\t\ttest=techstep_list.filter(status='test').count()\r\n\t\tstop=techstep_list.filter(status='stop').count()\r\n\t\tinit=techstep_list.filter(status='init').count()\r\n\t\tsuccess=techstep_list.filter(status='success').count()\r\n\t\r\n\treturn {\r\n\t\t'request': request, \r\n\t\t'object': object, \r\n\t\t'run': run, \r\n\t\t'test': test, \r\n\t\t'stop': stop,\r\n\t\t'init': init,\r\n\t\t'success': success,\r\n\t\t}\r\n\r\n@register.inclusion_tag('techstep_get_status_inc.html', takes_context=True)\r\ndef techstep_get_status_inc(context, **kwargs):\r\n\trequest = context['request']\r\n\t#\r\n\ttry:\r\n\t\tobject=kwargs['object'] #techtask_instance\r\n\texcept:\r\n\t\tpass\r\n\treturn {'request': request, 'object': object,}\r\n\t\r\n@register.inclusion_tag('techtask_get_executor_inc.html', takes_context=True)\r\ndef techtask_get_executor_inc(context, **kwargs):\r\n\trequest = context['request']\r\n\t#\r\n\ttry:\r\n\t\tobject=kwargs['object'] #techtask_instance\r\n\texcept:\r\n\t\tpass\r\n\t\t\r\n\texecutor_list=profileuser.objects.filter(techstep__techtask=object)\r\n\t#executor_list=[]\r\n\t\r\n\treturn {'request': request, 'object': object, 'executor_list' : executor_list,}\r\n\r\n\t\r\n\r\n@register.simple_tag\r\ndef techtask_and_techstep_commentary_count(obj):\r\n\ta=commentary.objects.filter(app='devapp', mid='techtask', oid=obj.id)\r\n\tb=commentary.objects.filter(app='devapp', mid='techstep', oid__in=[x['id'] for x in obj.techstep_set.all().values('id')])\r\n\treturn a.count()+b.count()\r\n\r\n", "sub_path": "crm/devapp/templatetags/devapptag.py", "file_name": "devapptag.py", "file_ext": "py", "file_size_in_byte": 2464, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.template.Library", "line_number": 20, "usage_type": "call"}, {"api_name": "django.template", "line_number": 20, "usage_type": "name"}, {"api_name": "panel.models.profileuser.objects.filter", "line_number": 83, "usage_type": "call"}, {"api_name": "panel.models.profileuser.objects", "line_number": 83, "usage_type": "attribute"}, {"api_name": "panel.models.profileuser", "line_number": 83, "usage_type": "name"}]}
+{"seq_id": "409068124", "text": "import numpy as np\r\nimport argparse\r\nimport cv2\r\n\r\nap = argparse.ArgumentParser()\r\nap.add_argument(\"-i\",\"--image\",required=True,help=\"resmin pc'deki yolu\")\r\nap.add_argument(\"-p\",\"--prototxt\",required=True,help=\"Caffe modulu yolu, prototxt dosyası için\")\r\nap.add_argument(\"-m\",\"--model\",required=True,help=\"Caffe pre-trained model yolu\")\r\nap.add_argument(\"-c\",\"--confidence\",type=float, default=0.5,help=\"zayıf filtreleme için min olasılık\")\r\nargs = vars(ap.parse_args())\r\n\r\n\r\nprint(\"[INFO] model yuklenior....\")\r\nnet = cv2.dnn.readNetFromCaffe(args[\"prototxt\"],args[\"model\"])\r\n\r\nimage = cv2.imread(args[\"image\"])\r\n(h,w) = image.shape[:2]\r\nblob = cv2.dnn.blobFromImage(cv2.resize(image,(300,300),),1.0,(300,300),(104.0,177.0,123.0))\r\n\r\nprint(\"[INFO]nesne belirleme hesaplanıyor.....\")\r\nnet.setInput(blob)\r\ndetections = net.forward()\r\n\r\nfor i in range(0,detections.shape[2]):\r\n confidence = detections[0,0,i,2]\r\n\r\n if confidence > args[\"confidence\"]:\r\n box = detections[0,0,i,3:7] * np.array([w,h,w,h])\r\n (startX, startY,endX,endY) = box.astype(\"int\")\r\n\r\n text = \"{:.2f}%\".format(confidence*100)\r\n y = startY-10 if startY-10 > 10 else startY + 10\r\n cv2.rectangle(image,(startX,startY),(endX,endY),(0,0,255),2)\r\n cv2.putText(image,text,(startX,y),cv2.FONT_HERSHEY_SIMPLEX,0.45,(0,0,255),2)\r\n\r\n cv2.imshow(\"output\",image)\r\n cv2.waitKey(0)\r\n\r\n\r\n", "sub_path": "face-detect.py", "file_name": "face-detect.py", "file_ext": "py", "file_size_in_byte": 1401, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 5, "usage_type": "call"}, {"api_name": "cv2.dnn.readNetFromCaffe", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 14, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.dnn.blobFromImage", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 18, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 33, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 34, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 36, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 37, "usage_type": "call"}]}
+{"seq_id": "64183195", "text": "import math\nimport sys\nimport Atom\nimport Utility\nimport random\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits.mplot3d import Axes3D\nfrom mpl_toolkits.mplot3d.art3d import Poly3DCollection, Line3DCollection\n\n\ndef main():\n Atoms, Ligands = init()\n rCutOff = float(sys.argv[2])\n num = int(sys.argv[3])\n #plotLigands(Ligands)\n Vertices = buildLigandStructure(Ligands, rCutOff)\n vol = MonteCarlo(Ligands, Vertices, rCutOff, num)\n #plotBoundingBox(Ligands, Vertices)\n testNumRandomPoints(Ligands, Vertices)\n #testrCutOffGrowth(Ligands)\n #print(\"The volume is approximately: \" + str(vol) + \" Angstroms cubed\")\n return 1\n\ndef init():\n try:\n if len(sys.argv) != 4: #Checks to see if the correct amount of arguments have been passed in\n print(\"[Error] Not enough Arguments\")\n exit(1)\n else:\n filePath = sys.argv[1]\n with open(filePath, 'r') as file:\n atoms, ligands = Utility.Constructor(file)\n file.close()\n return atoms, ligands\n\n except EnvironmentError:\n print(\"[Error] No PQR file given\")\n exit(1)\n\ndef testNumRandomPoints(Ligands, Vertices):\n rCutOff = 5\n #nums = [1000, 10000, 100000, 1000000, 10000000]\n nums = [10000, 50000, 100000, 500000, 1000000, 1500000, 2000000]\n results = {}\n for num in nums:\n countin = 0\n sum = 0\n for _ in range(10):\n countin += 1\n vol = MonteCarlo(Ligands, Vertices, rCutOff, num)\n sum += vol\n if num not in results:\n results[num] = [vol]\n else:\n results[num].append(vol)\n print(\"Count: \" + str(countin))\n avg = sum / countin\n results[num].append(avg)\n print(\"For \" + str(num) + \" points randomly generated\")\n print(\"Results: \" + str(results[num]))\n print(\"The Avg. is \" + str(avg))\n analyzeResults(results)\n return 1\n\ndef analyzeResults(results):\n # Calculate the standard deviation for each key\n fig = plt.figure()\n for key in results:\n sum = 0\n mean = results[key].pop(len(results[key])-1)\n for res in results[key]:\n sum += (math.pow(res - mean, 2))\n stdDev = math.sqrt(sum / len(results[key]))\n #Min = min(results[key]) #finds the minimum value\n #Max = max(results[key]) #finds the maximum value\n x = [key, key, key]\n y = [(mean - stdDev), (mean + stdDev), mean]\n #x = [key]\n #y = [stdDev]\n plt.scatter(x, y)\n plt.show()\n\n\ndef testrCutOffGrowth(Ligands):\n num = 1000000\n rCutOff = 5\n results = {}\n for _ in range(10):\n sum = 0\n Vertices = buildLigandStructure(Ligands, rCutOff)\n for _ in range(5):\n vol = MonteCarlo(Ligands, Vertices, rCutOff, num)\n sum += vol\n avg = sum / 5\n results[rCutOff] = avg\n print(\"rCutOff: \" + str(rCutOff) + \" , Volume: \" + str(results[rCutOff]))\n rCutOff += 1\n analyzerCutOff(results)\n\ndef analyzerCutOff(results):\n fig = plt.figure()\n x = []\n y = []\n z = []\n for key in results:\n cubedRoot = results[key]**(1/3)\n x.append(key)\n y.append(results[key])\n z.append(cubedRoot)\n print(\"The cubed root is: \" + str(cubedRoot))\n plt.scatter(x, z)\n #plt.scatter(x, y)\n plt.xlabel(\"rCutOff\")\n plt.ylabel(\"Cubed Root of Volume\")\n plt.show()\n\ndef MonteCarlo(Ligands, Vertices, rCutOff, num):\n randPnts = generateRandomPoints(Vertices, num)\n hits = checkHits(Ligands, randPnts, rCutOff)\n volBox = calcVolBox(Vertices)\n volLigs = volBox * (float(hits) / len(randPnts))\n #print(\"The ratio of hits is: \" + str(float(hits) / len(randPnts)))\n #print(\"The volume of the box is: \" + str(volBox))\n #print(\"The volume of the Ligands is: \" + str(volLigs))\n return volLigs\n\ndef generateRandomPoints(Vertices, num):\n randPnts = []\n for _ in range(num):\n xval = random.uniform(Vertices[0][0], Vertices[7][0])\n yval = random.uniform(Vertices[0][1], Vertices[7][1])\n zval = random.uniform(Vertices[0][2], Vertices[7][2])\n randPnts.append((xval, yval, zval))\n return randPnts\n\ndef checkHits(Ligands, randPnts, rCutOff):\n hits = 0\n for point in randPnts:\n hits += checkPoint(Ligands, point, rCutOff)\n return hits\n\ndef checkPoint(Ligands, point, rCutOff):\n for lig in Ligands:\n dist = getDistanceFromLigand(lig, point)\n if dist <= rCutOff:\n return 1\n return 0\n\ndef getDistanceFromLigand(ligand, p):\n return math.sqrt(math.pow(ligand.X - p[0], 2) + math.pow(ligand.Y - p[1], 2) + math.pow(ligand.Z - p[2], 2))\n\ndef calcVolBox(Vertices):\n width = getDistBetweenPnts(Vertices[0], Vertices[1])\n length = getDistBetweenPnts(Vertices[0], Vertices[4])\n height = getDistBetweenPnts(Vertices[0], Vertices[2])\n vol = width * length * height\n return vol\n\ndef getDistBetweenPnts(p1, p2):\n return math.sqrt(math.pow(p2[0] - p1[0], 2) + math.pow(p2[1] - p1[1], 2) + math.pow(p2[2] - p1[2], 2))\n\ndef buildLigandStructure(Ligands, rCutOff):\n LigX, LigY, LigZ = getXYZLists(Ligands)\n # Get initial values\n Xmin = LigX[0] - rCutOff\n Xmax = LigX[0] + rCutOff\n Ymin = LigY[0] - rCutOff\n Ymax = LigY[0] + rCutOff\n Zmin = LigZ[0] - rCutOff\n Zmax = LigZ[0] + rCutOff\n\n for x in range(len(Ligands)):\n if LigX[x] - rCutOff < Xmin:\n Xmin = LigX[x] - rCutOff\n\n elif LigX[x] + rCutOff > Xmax:\n Xmax = LigX[x] + rCutOff\n for x in range(len(Ligands)):\n if LigY[x] - rCutOff < Ymin:\n Ymin = LigY[x] - rCutOff\n\n elif LigY[x] + rCutOff > Ymax:\n Ymax = LigY[x] + rCutOff\n for x in range(len(Ligands)):\n if LigZ[x] - rCutOff < Zmin:\n Zmin = LigZ[x] - rCutOff\n\n elif LigZ[x] + rCutOff > Zmax:\n Zmax = LigZ[x] + rCutOff\n\n Vertices = []\n Xs = [Xmin, Xmax]\n Ys = [Ymin, Ymax]\n Zs = [Zmin, Zmax]\n\n for x in Xs:\n for y in Ys:\n for z in Zs:\n Vertices.append((x, y, z))\n return Vertices\n\ndef plotBoundingBox(Ligands, v):\n fig = plt.figure()\n ax = fig.add_subplot(111, projection = '3d')\n LigX, LigY, LigZ = getXYZLists(Ligands)\n\n verts = [[v[0], v[2], v[3], v[1]],\n [v[4], v[5], v[7], v[6]],\n [v[0], v[4], v[6], v[2]],\n [v[0], v[1], v[5], v[4]],\n [v[2], v[3], v[7], v[6]],\n [v[1], v[3], v[7], v[5]]]\n\n collection = Poly3DCollection(verts, linewidths=1, edgecolors='r', alpha= 0.4)\n face_color = [0.75, 0.75, 0.75]\n collection.set_facecolor(face_color)\n ax.add_collection3d(collection)\n\n ax.scatter(LigX, LigY, LigZ, c='b', marker='^')\n\n ax.set_xlabel(\"X Label\")\n ax.set_ylabel(\"Y Label\")\n ax.set_zlabel(\"Z Label\")\n plt.show()\n\ndef plotLigands(Ligands):\n fig = plt.figure()\n ax = fig.add_subplot(111, projection = '3d')\n LigX, LigY, LigZ = getXYZLists(Ligands)\n ax.scatter(LigX, LigY, LigZ, c='b', marker='^')\n ax.set_xlabel(\"X Label\")\n ax.set_ylabel(\"Y Label\")\n ax.set_zlabel(\"Z Label\")\n plt.show()\n\ndef getXYZLists(list):\n x, y, z = [], [], []\n for element in list:\n x.append(element.X)\n y.append(element.Y)\n z.append(element.Z)\n return x, y, z\n\nmain()\n", "sub_path": "VolumeApprox/bin/Estimate.py", "file_name": "Estimate.py", "file_ext": "py", "file_size_in_byte": 7409, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sys.argv", "line_number": 13, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 14, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 26, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 30, "usage_type": "attribute"}, {"api_name": "Utility.Constructor", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "math.pow", "line_number": 72, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}, {"api_name": "random.uniform", "line_number": 130, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 131, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 132, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 150, "usage_type": "call"}, {"api_name": "math.pow", "line_number": 150, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 160, "usage_type": "call"}, {"api_name": "math.pow", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 203, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 203, "usage_type": "name"}, {"api_name": "mpl_toolkits.mplot3d.art3d.Poly3DCollection", "line_number": 214, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 224, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 224, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 227, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 227, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 234, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 234, "usage_type": "name"}]}
+{"seq_id": "359678964", "text": "import numpy as np\r\nimport matplotlib as mpl\r\nimport matplotlib.pyplot as plt\r\nimport os.path as osp\r\nfrom astropy.io import fits\r\nfrom scipy.special import legendre\r\n\r\n\r\ndef rebin(a, *args):\r\n shape = a.shape\r\n lenShape = len(shape)\r\n factor = np.asarray(shape) / np.asarray(args)\r\n evList = ['a.reshape('] + \\\r\n ['args[%d],factor[%d],' % (i, i) for i in range(lenShape)] + \\\r\n [')'] + ['.mean(%d)' % (i + 1) for i in range(lenShape)]\r\n print(''.join(evList))\r\n eval(''.join(evList))\r\n\r\n\r\ndef legendre_coef(Xi, l):\r\n if l % 2 == 1:\r\n raise ValueError(\"\\'l\\' has to be even\")\r\n delta_mu = 2 / 32\r\n temp = np.copy(Xi)\r\n nonzeros = [np.count_nonzero(e) for e in temp]\r\n nonzeros[nonzeros == 0] = 1\r\n mu_vec = np.linspace(-1, 1, 32)[:, None] # the vector of mu values\r\n p = legendre(l) # Legendre polynomial at order l\r\n mu_vec = np.array([p(mu) for mu in mu_vec]) # P_l(mu)\r\n temp = temp * mu_vec.transpose() * delta_mu # P_l(mu) * Xi(s, mu) * delta_mu\r\n temp = temp * (2 * l + 1) / 2\r\n return temp.sum(axis=1)\r\n\r\n\r\nmpl.rc(\"font\", family=\"serif\", size=14)\r\n\r\nfile = osp.join('/home', 'yujie', 'Documents', 'data', 'corr_North_dr12_3D.dat')\r\ndata = np.genfromtxt(file)\r\npi, sigma, xi, dxi = [data[:, i] for i in range(0, 4)]\r\nDD, DR, RR = [data[:, i] for i in range(4, 7)]\r\n\r\nPi, Sigma = np.meshgrid(pi[:32], pi[:32])\r\nS = np.sqrt(Pi ** 2 + Sigma ** 2)\r\ns_vec = np.sqrt(pi[:32]**2+pi[:32]**2)\r\nmu_vec = pi[:32]/s_vec\r\n\r\nXi = xi.reshape(32, 32)\r\n\r\n# leng = Xi.shape\r\n# Xi = np.lib.pad(Xi, ((leng[0], 0), (leng[1], 0)), 'reflect')\r\n\r\nprint(Xi.min(), Xi.mean(), Xi.max())\r\n\r\nlevels = np.arange(-0.01, 0.03, 0.002)\r\nplt.contourf(Xi.T, 20, origin='lower', interpolation='nearest', extent=[-200, 200, -200, 200], levels=levels)\r\nplt.xlabel(r'$\\sigma$')\r\nplt.ylabel(r'$\\pi$')\r\nplt.colorbar()\r\nplt.tight_layout()\r\nplt.show()\r\nfig, [ax0, ax1] = plt.subplots(1, 2, figsize=(10, 4))\r\nim = ax0.imshow((S ** 2 * Xi).T, origin='lower', interpolation='nearest',\r\n extent=[0, 200, 0, 200],\r\n vmin=0, vmax=200,\r\n cmap='magma')\r\nim = ax0.imshow((1 * Xi).T, origin='lower', interpolation='nearest',\r\n extent=[0, 200, 0, 200], vmax=0.03,\r\n cmap='magma')\r\nax0.set(aspect='equal',\r\n xlabel='$\\sigma$',\r\n ylabel='$\\pi$')\r\ncb = fig.colorbar(im, ax=ax0, label=r'$s^2 \\xi(s)$', fraction=0.05)\r\n\r\ncf = ax1.contourf(Sigma, Pi, S ** 2 * Xi, 20,\r\n vmin=0, vmax=200,\r\n cmap='magma')\r\nax1.set(aspect='equal',\r\n xlabel='$\\sigma$',\r\n ylabel='$\\pi$',\r\n xlim=(0, 200),\r\n ylim=(0, 200))\r\ncb = fig.colorbar(im, ax=ax1, label=r'$s^2 \\xi(s)$', fraction=0.05)\r\n\r\nfig.tight_layout()\r\n\r\nplt.show()\r\n\r\n\r\nXi = np.zeros((32,32))\r\nfor tuple in data:\r\n pi, sigma, xi, dxi =tuple[0:4]\r\n s = np.sqrt(pi ** 2 + sigma ** 2)\r\n mu=pi/s\r\n s_index=int(s*32/280)\r\n mu_index=int((mu+1)*32/2)\r\n Xi[s_index,mu_index] = xi\r\n\r\n# Xi1 = np.copy(Xi)\r\n# for i in range(1, 32):\r\n# temp = Xi[i, :]\r\n# zeros = np.nonzero(temp == 0)[0]\r\n# nonzeros = np.nonzero(temp != 0)[0]\r\n# nonzero_vals = temp[nonzeros]\r\n# if len(nonzeros) >= 10:\r\n# temp[temp == 0] = np.interp(zeros, nonzeros, nonzero_vals)\r\n# Xi1[i, :] = temp\r\n# Xi = Xi1\r\n\r\na = np.arange(0, 2.5, 0.5)\r\nb = np.arange(10).reshape(-1, 5)\r\nplt.plot(a, b[0])\r\nplt.show()\r\n\r\nhdu = fits.BinTableHDU.from_columns([\r\n fits.Column(name='s', array=s_vec, format='I'),\r\n fits.Column(name='tpcf0', array=legendre_coef(Xi, 0)*s_vec**2, format='D'),\r\n fits.Column(name='tpcf2', array=legendre_coef(Xi, 2)*s_vec**2, format='D'),\r\n fits.Column(name='tpcf4', array=legendre_coef(Xi, 4)*s_vec**2, format='D'),\r\n fits.Column(name='tpcf6', array=legendre_coef(Xi, 6)*s_vec**2, format='D')\r\n],\r\n name='legendre')\r\nhdu.writeto('cute_legendre.fits')\r\n\r\n", "sub_path": "plot3D.py", "file_name": "plot3D.py", "file_ext": "py", "file_size_in_byte": 3923, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "numpy.asarray", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.count_nonzero", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 27, "usage_type": "call"}, {"api_name": "scipy.special.legendre", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "name"}, {"api_name": "numpy.genfromtxt", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.contourf", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "astropy.io.fits.BinTableHDU.from_columns", "line_number": 114, "usage_type": "call"}, {"api_name": "astropy.io.fits.BinTableHDU", "line_number": 114, "usage_type": "attribute"}, {"api_name": "astropy.io.fits", "line_number": 114, "usage_type": "name"}, {"api_name": "astropy.io.fits.Column", "line_number": 115, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 115, "usage_type": "name"}, {"api_name": "astropy.io.fits.Column", "line_number": 116, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 116, "usage_type": "name"}, {"api_name": "astropy.io.fits.Column", "line_number": 117, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 117, "usage_type": "name"}, {"api_name": "astropy.io.fits.Column", "line_number": 118, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 118, "usage_type": "name"}, {"api_name": "astropy.io.fits.Column", "line_number": 119, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 119, "usage_type": "name"}]}
+{"seq_id": "537717994", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Feb 23 23:43:31 2018\n\nStudy different optimization tasks following\n https://docs.scipy.org/doc/scipy/reference/tutorial/optimize.html\n\n@author: mhaa\n\"\"\"\n\nimport numpy as np\nfrom scipy.optimize import minimize\nfrom scipy.optimize import minimize_scalar\nfrom scipy.optimize import basinhopping\n\nimport matplotlib.pyplot as plt\n\n\n\"\"\"\nA. Unconstrained optimization of multivariate scalar \nfunctions (minimization)\n\"\"\"\n\ndef rosen(x):\n \"\"\"The Rosenbrock function\"\"\"\n return sum(100.0*(x[1:]-x[:-1]**2.0)**2.0 + (1-x[:-1])**2.0)\n\n\ndef rosen_der(x):\n # Gradient of the Rosen function \n xm = x[1:-1]\n xm_m1 = x[:-2]\n xm_p1 = x[2:]\n der = np.zeros_like(x)\n der[1:-1] = 200*(xm-xm_m1**2) - 400*(xm_p1 - xm**2)*xm - 2*(1-xm)\n der[0] = -400*x[0]*(x[1]-x[0]**2) - 2*(1-x[0])\n der[-1] = 200*(x[-1]-x[-2]**2)\n return der\n\n\ndef ownfunc2d(x):\n # Own function with various local minima\n y = x[0]**2 + -10*x[0]*np.cos(x[0]) - x[1]**3 + 9*x[1]*np.sin(x[1]) + 50 \n return(y)\n\n\ndef nelder_mead():\n # Without gradient\n #\n # Nelder-Mead Simplex algorithm. \n # Probably the simplest way to minimize a \n # fairly well-behaved multivariate function. Requires only \n # function evaluations, good choice for simple \n # minimization problems. No gradients, may take \n # longer to find the minimum.\n #\n # Another option: method='powell'\n \n x0 = np.array([1.3, 0.7, 0.8, 1.9, 1.2])\n print('x0:', x0)\n res = minimize(rosen, x0, method='nelder-mead',\n options={'xtol': 1e-8, 'disp': True})\n print(res.x)\n\n x0 = np.array([0.0, 0.0])\n res = minimize(ownfunc2d, x0, method='nelder-mead',\n options={'xtol': 1e-8, 'disp': True})\n print(res.x)\n\n\ndef BFGS():\n # Takes gradient information\n # either directly or estimated numerically from the function \n print('BFGS')\n \n x0 = np.array([1.3, 0.7, 0.8, 1.9, 1.2])\n res = minimize(rosen, x0, method='BFGS', jac=rosen_der,\n options={'disp': True})\n print('With gradient information:', res.x)\n\n res = minimize(rosen, x0, method='BFGS', \n options={'disp': True})\n print('Gradient estimated numerically:', res.x)\n\n\ndef rosen_hess(x):\n x = np.asarray(x)\n H = np.diag(-400*x[:-1],1) - np.diag(400*x[:-1],-1)\n diagonal = np.zeros_like(x)\n diagonal[0] = 1200*x[0]**2-400*x[1]+2\n diagonal[-1] = 200\n diagonal[1:-1] = 202 + 1200*x[1:-1]**2 - 400*x[2:]\n H = H + np.diag(diagonal)\n return H\n\ndef rosen_hess_p(x, p):\n x = np.asarray(x)\n Hp = np.zeros_like(x)\n Hp[0] = (1200*x[0]**2 - 400*x[1] + 2)*p[0] - 400*x[0]*p[1]\n Hp[1:-1] = -400*x[:-2]*p[:-2]+(202+1200*x[1:-1]**2-400*x[2:])*p[1:-1] \\\n -400*x[1:-1]*p[2:]\n Hp[-1] = -400*x[-2]*p[-2] + 200*p[-1]\n return Hp\n\n\ndef Newton_CG():\n # Suitable for large-scale problems\n # \n print('Newton-CG')\n x0 = np.array([1.3, 0.7, 0.8, 1.9, 1.2])\n res = minimize(rosen, x0, method='Newton-CG',\n jac=rosen_der, hess=rosen_hess,\n options={'xtol': 1e-8, 'disp': True})\n print(res.x)\n \n # Also version with Hessian product\n # need Hp\n # For larger minimization problems, storing the entire \n # Hessian matrix can consume considerable time and memory. \n # The Newton-CG algorithm only needs the product of the\n # Hessian times an arbitrary vector.\n res = minimize(rosen, x0, method='Newton-CG',\n jac=rosen_der, hessp=rosen_hess_p,\n options={'xtol': 1e-8, 'disp': True})\n print(res.x)\n\n\n\ndef trust_region_version():\n # Methods suitable for large-scale problems\n # (problems with thousands of variables)\n #\n # Similar to the trust-ncg method, the trust-krylov method is \n # a method suitable for large-scale problems as it uses \n # the hessian only as linear operator by means of \n # matrix-vector products. It solves the quadratic subproblem \n # more accurately than the trust-ncg method.\n #\n print('Trust-region methods')\n x0 = np.array([1.3, 0.7, 0.8, 1.9, 1.2])\n \n print('trust-ncg')\n res = minimize(rosen, x0, method='trust-ncg',\n jac=rosen_der, hess=rosen_hess,\n options={'gtol': 1e-8, 'disp': True})\n print(res.x)\n\n print('trust-krylov, trust-exact don''t work, maybe need scipy update')\n #res = minimize(rosen, x0, method='trust-krylov',\n # jac=rosen_der, hess=rosen_hess,\n # options={'gtol': 1e-8, 'disp': True})\n #print(res.x)\n\n\n # Trust-exact\n # For medium-size problems, where storage and factorization \n # of Hessian are not critical. Also exact solution of trust-region subproblems\n #\n #res = minimize(rosen, x0, method='trust-exact',\n # jac=rosen_der, hess=rosen_hess,\n # options={'gtol': 1e-8, 'disp': True})\n #print(res.x)\n\n\n\"\"\"\nB. Constrained optimization of multivariate scalar \nfunctions (minimization)\n\"\"\"\n\ndef func(x, sign=1.0):\n \"\"\" \n Objective function \n Use sign = -1 to make the task as maximization\n \"\"\"\n return sign*(2*x[0]*x[1] + 2*x[0] - x[0]**2 - 2*x[1]**2)\n\n\ndef func_deriv(x, sign=1.0):\n \"\"\" Derivative of objective function \"\"\"\n dfdx0 = sign*(-2*x[0] + 2*x[1] + 2)\n dfdx1 = sign*(2*x[0] - 4*x[1])\n return np.array([ dfdx0, dfdx1 ])\n\n\ndef constr_opt():\n print('Constrained optimization')\n # Constraints\n # x^3 = y\n # y-1 >= 0\n # [x0, y0] = [-1.0, 1.0] initial guess\n # args=(-1.0) feed in the sign to get maximization problem\n # \n # SLSQP - Sequential Least SQuares Programming optimization\n #\n cons = ({'type': 'eq',\n 'fun' : lambda x: np.array([x[0]**3 - x[1]]),\n 'jac' : lambda x: np.array([3.0*(x[0]**2.0), -1.0])},\n {'type': 'ineq',\n 'fun' : lambda x: np.array([x[1] - 1]),\n 'jac' : lambda x: np.array([0.0, 1.0])})\n\n # First unconstrained optimization\n res = minimize(func, [-1.0,1.0], args=(-1.0,), jac=func_deriv,\n method='SLSQP', options={'disp': True})\n print(res.x)\n \n # With constrains\n res = minimize(func, [-1.0,1.0], args=(-1.0,), jac=func_deriv,\n constraints=cons, method='SLSQP', options={'disp': True})\n print(res.x)\n\n # With constrains, without gradient\n res = minimize(func, [-1.0,1.0], args=(-1.0,), \n constraints=cons, method='SLSQP', options={'disp': True})\n print(res.x)\n\n \n\n\n\n\"\"\"\nC. Least-squares minimization (least_squares)\n\"\"\"\n\n\n\"\"\"\nD. Univariate function minimizers (minimize_scalar)\n Unconstrained minimization (method='brent')\n Bounded minimization (method='bounded')\n\"\"\"\n\ndef ownfunc(x):\n y = x**2 + 10*x*np.cos(x) + 2.5\n return(y)\n\n\ndef minim_scalar():\n # Optimization methods for univariate functions\n # Bounded and unbounded minimization\n\n # First Bessel function\n #from scipy.special import j1\n #func = j1\n\n # Own 1D function defined above to minimize\n func = ownfunc\n \n x = np.linspace(-10, 15, 1000)\n plt.plot(x, func(x))\n plt.show()\n\n # Bounded minimization\n bounds = (-10, 0)\n res = minimize_scalar(func, bounds=bounds, method='bounded')\n print('Minimum at:', res.x, 'for bounds:', bounds, ', y =', res.fun)\n\n # Unbounded minimization\n bracket = (-10, 0) # Note bracket=(-10, 0) --> wrong minimimun\n #bracket = None\n res = minimize_scalar(func, bracket=bracket, method='brent')\n print('Minimum at:', res.x, 'unbounded, y =', res.fun)\n \n # Note: Own func w/ unbounded optimization leads to wrong minimum\n # --> Use bounds if known\n \n \n\n\"\"\"\nE. Custom minimizers\n\"\"\"\n\n\n\"\"\"\nF. Root finding\n\"\"\"\n\n\n\"\"\"\nG. Basin hopping: multivariate scalar functions\n\nhttps://docs.scipy.org/doc/scipy-0.18.1/reference/generated/scipy.optimize.basinhopping.html\n\"\"\"\n\ndef print_fun(x, f, accepted):\n # Custom callback function, prints the value of every \n # minimum found\n print(\"at x = %.3f minimum %.4f accepted %d\" % (x, f, int(accepted)))\n\n\n\ndef basin_hopping_1d():\n # Stochastic global optimization. Particularly useful \n # when the function has many minima separated by large \n # barriers.\n #\n # No way to determine if \n # the true global minimum has actually been found. Instead, \n # as a consistency check, the algorithm can be run from a \n # number of different random starting points to ensure the \n # lowest minimum found in each example has converged to the \n # global minimum. For this reason basinhopping will by \n # default simply run for the number of iterations niter \n # and return the lowest minimum found. \n\n # 1D function\n print('1D function')\n func = lambda x: np.cos(14.5 * x - 0.3) + (x + 0.2) * x\n x0=[1.]\n print('Starting x0:', x0)\n \n x = np.linspace(-2, 2, 100)\n plt.plot(x, func(x))\n plt.show()\n \n \n minimizer_kwargs = {\"method\": \"BFGS\"}\n ret = basinhopping(func, x0, minimizer_kwargs=minimizer_kwargs,\n niter=20, callback=print_fun)\n # niter: was 200 originally\n print(\"global minimum: x = %.4f, f(x0) = %.4f\" % (ret.x, ret.fun))\n\n\n\ndef basin_hopping_1d_own():\n # 1D function, own function (see above 'ownfunc')\n print('1D function, own function')\n func = ownfunc\n \n x = np.linspace(-10, 15, 1000)\n plt.plot(x, func(x))\n plt.ylim([-22, 5])\n plt.show()\n\n print(np.random.rand(1))\n\n x0=[1.0]\n # Random starting point in the range -10 ... 15\n # x0 = np.random.rand(1)*25.0 - 10.0\n print('Starting x0 =', x0)\n\n minimizer_kwargs = {\"method\": \"BFGS\"}\n ret = basinhopping(func, x0, stepsize=15.0, minimizer_kwargs=minimizer_kwargs,\n niter=5, callback=print_fun)\n print(\"global minimum: x = %.4f, f(x0) = %.4f\" % (ret.x, ret.fun))\n print(\"\\nNote: if x0=1.0 and stepsize=default=0.5, does not\\n \\\n find the global minimum at about x= -6.0\")\n print(\" --> Set stepsize = 15.0, then most of the time OK\")\n\n # input('press enter')\n\n\n\ndef basin_hopping_2d():\n # 2D function\n print('2D function')\n \n def func2d(x):\n f = np.cos(14.5 * x[0] - 0.3) + (x[1] + 0.2) * x[1] + (x[0] +\n 0.2) * x[0]\n df = np.zeros(2)\n df[0] = -14.5 * np.sin(14.5 * x[0] - 0.3) + 2. * x[0] + 0.2\n df[1] = 2. * x[1] + 0.2\n return f, df\n\n minimizer_kwargs = {\"method\":\"L-BFGS-B\", \"jac\":True}\n x0 = [1.0, 1.0]\n ret = basinhopping(func2d, x0, minimizer_kwargs=minimizer_kwargs,\n niter=200) # , callback=print_fun\n print(\"global minimum: x = [%.4f, %.4f], f(x0) = %.4f\" % (ret.x[0],\n ret.x[1],\n ret.fun))\n\n\n\n\ndef main():\n \n nelder_mead()\n print('---------') \n \n BFGS() \n print('---------') \n \n trust_region_version()\n print('---------') \n \n constr_opt()\n print('---------') \n\n minim_scalar() \n print('---------') \n\n basin_hopping_1d() \n print('---------') \n\n basin_hopping_1d_own() \n print('---------') \n\n basin_hopping_2d() \n print('---------') \n\n\n print('\\nDone')\n\nif __name__ == '__main__':\n main()\n\n\n\n", "sub_path": "optimization/optimization_examples.py", "file_name": "optimization_examples.py", "file_ext": "py", "file_size_in_byte": 11229, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "numpy.zeros_like", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 59, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 65, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 76, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 77, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 110, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 111, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 140, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 182, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 196, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 199, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 200, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 203, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 208, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 233, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 248, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 249, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 249, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 250, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 250, "usage_type": "name"}, {"api_name": "scipy.optimize.minimize_scalar", "line_number": 254, "usage_type": "call"}, {"api_name": "scipy.optimize.minimize_scalar", "line_number": 260, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 307, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 311, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 312, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 312, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 313, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 313, "usage_type": "name"}, {"api_name": "scipy.optimize.basinhopping", "line_number": 317, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 329, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 330, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 330, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 331, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 331, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 332, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 332, "usage_type": "name"}, {"api_name": "numpy.random.rand", "line_number": 334, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 334, "usage_type": "attribute"}, {"api_name": "scipy.optimize.basinhopping", "line_number": 342, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 358, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 360, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 361, "usage_type": "call"}, {"api_name": "scipy.optimize.basinhopping", "line_number": 367, "usage_type": "call"}]}
+{"seq_id": "25714422", "text": "\"\"\"\nCreated on Tue Nov 9 10:37:29 2021\n\n@author: Teresa\n\"\"\"\n\nimport sys\nsys.path.append('../marcos_client')\nimport experiment as ex\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.cm as cm\nimport scipy.signal as sig\nimport time \n\ndef rabiflops_standalone(\n init_gpa= False, \n larmorFreq=3.076, \n rfExAmp=0.6, \n rfReAmp=None, \n rfExPhase = 0,\n rfExTimeIni=10, \n rfExTimeEnd = 150, \n nExTime = 60, \n nReadout = 160,\n tAdq = 4*1e3,\n tEcho = 20*1e3,\n tRepetition = 500*1e3, \n plotSeq = 0, \n shimming=[-80, -100, 10]):\n\n# INITALISATION OF VARIABLES ################################################################################\n #CONTANTS\n tStart = 20\n txGatePre = 15\n txGatePost = 1\n oversamplingFactor=6\n shimming=np.array(shimming)*1e-4\n \n #ARRAY INITIALIZATIONS \n txTime=[]\n txAmp=[]\n txGateTime=[]\n txGateAmp=[]\n rxTime = []\n rxAmp = []\n dataAll =[]\n \n #RF PULSES\n if rfReAmp is None:\n rfReAmp = rfExAmp\n rfExPhase = rfExPhase*np.pi/180\n rfExAmp = rfExAmp*np.exp(1j*rfExPhase)\n rfRePhase = 0\n rfReAmp = rfReAmp *np.exp(1j*rfRePhase)\n #Excitation times\n rfExTime= np.linspace(rfExTimeIni, rfExTimeEnd, nExTime, endpoint=True)\n \n# DEFINITION OF PULSES ####################################################################################\n def rfPulse(tRef, rfAmp, rfDuration, txTimePrevious,txAmpPrevious, txGateTimePrevious, txGateAmpPrevious):\n txTime = np.array([tRef-rfDuration/2,tRef+rfDuration/2])\n txAmp = np.array([rfAmp,0.])\n txGateTime = np.array([txTime[0]-txGatePre,txTime[1]+txGatePost])\n txGateAmp = np.array([1,0])\n txTime = np.concatenate((txTimePrevious,txTime), axis=0)\n txAmp = np.concatenate((txAmpPrevious,txAmp ), axis=0)\n txGateTime = np.concatenate((txGateTimePrevious,txGateTime), axis=0)\n txGateAmp = np.concatenate((txGateAmpPrevious,txGateAmp), axis=0)\n return txTime, txAmp, txGateTime, txGateAmp\n \n def readoutGate(tRef,tRd,rxTimePrevious, rxAmpPrevious):\n rxTime = np.array([tRef-tRd/2, tRef+tRd/2])\n rxAmp = np.array([1,0])\n rxTime=np.concatenate((rxTimePrevious, rxTime), axis=0)\n rxAmp=np.concatenate((rxAmpPrevious, rxAmp), axis=0)\n return rxTime, rxAmp\n\n\n# SPECIFIC FUNCTIONS ####################################################################################\n def plotData(data, rfExTime, tAdqReal):\n plt.figure(1)\n colors = cm.rainbow(np.linspace(0, 0.8, len(rfExTime)))\n for indexExTime in range(nExTime):\n tPlot = np.linspace(-tAdqReal/2, tAdqReal/2, nReadout, endpoint ='True')*1e-3\n leg = 'Time = '+ str(np.round(rfExTime[indexExTime]))+ 'us'\n plt.plot(tPlot[5:], np.abs(data[indexExTime, 5:]), label = leg, color=colors[indexExTime])\n# plt.plot(tPlot[5:], np.real(data[indexExTime, 5:]))\n# plt.plot(tPlot[5:], np.imag(data[indexExTime, 5:]))\n plt.xlabel('t(ms)')\n plt.ylabel('A(mV)')\n plt.legend()\n \n def plotRabiFlop(data, rfExTime, tAdqReal):\n for indexExTime in range(nExTime):\n# np.max(np.abs(data[indexExTime, 5:]))\n if indexExTime == 0:\n maxEchoes = np.max(np.abs(data[indexExTime,5:]))\n else:\n maxEchoes=np.append(maxEchoes,np.max(np.abs(data[indexExTime, 5:])))\n plt.figure(2)\n plt.plot(rfExTime, maxEchoes)\n plt.xlabel('t(us)')\n plt.ylabel('A(mV)')\n titleRF= 'RF Amp = '+ str(np.real(rfExAmp))\n plt.title(titleRF)\n \n\n\n\n# SEQUENCE ############################################################################################\n\n for indexExTime in range(nExTime):\n \n rfReTime = 60\n \n txTime=[]\n txAmp=[]\n txGateTime=[]\n txGateAmp=[]\n rxTime = []\n rxAmp = []\n \n # INIT EXPERIMENT\n BW = nReadout/tAdq\n BWov = BW*oversamplingFactor\n samplingPeriod = 1/BWov\n expt = ex.Experiment(lo_freq=larmorFreq, rx_t=samplingPeriod, init_gpa=init_gpa, gpa_fhdo_offset_time=(1 / 0.2 / 3.1))\n samplingPeriod = expt.get_rx_ts()[0]\n BWReal = 1/samplingPeriod/oversamplingFactor\n tAdqReal = nReadout/BWReal \n tIni=20 #us initial time\n # Shimming\n expt.add_flodict({\n 'grad_vx': (np.array([tIni]),np.array([shimming[0]])), \n 'grad_vy': (np.array([tIni]),np.array([shimming[1]])), \n 'grad_vz': (np.array([tIni]),np.array([shimming[2]])),\n })\n # TR \n tRef = tStart+rfExTime[indexExTime]/2+tIni+100\n txTime, txAmp,txGateTime,txGateAmp = rfPulse(tRef,rfExAmp, rfExTime[indexExTime], txTime, txAmp, txGateTime, txGateAmp)\n tRef = tRef+tEcho/2\n txTime, txAmp, txGateTime, txGateAmp = rfPulse(tRef,rfReAmp, rfReTime, txTime, txAmp, txGateTime, txGateAmp)\n tRef = tRef+tEcho/2\n rxTime, rxAmp = readoutGate(tRef, tAdqReal, rxTime, rxAmp)\n \n expt.add_flodict({\n 'tx0': (txTime, txAmp),\n 'tx_gate': (txGateTime, txGateAmp), \n 'rx0_en': (rxTime, rxAmp),\n 'rx_gate': (rxTime, rxAmp),\n })\n tEnd = tRepetition\n expt.add_flodict({\n 'grad_vx': (np.array([tEnd]),np.array([0])), \n 'grad_vy': (np.array([tEnd]),np.array([0])), \n 'grad_vz': (np.array([tEnd]),np.array([0])),\n })\n\n if plotSeq == 0:\n print(indexExTime, '.- Running...')\n rxd, msgs = expt.run()\n expt.__del__()\n print(' End')\n data = sig.decimate(rxd['rx0']*13.788, oversamplingFactor, ftype='fir', zero_phase=True)\n dataAll = np.concatenate((dataAll, data), axis=0)\n elif plotSeq == 1:\n expt.plot_sequence()\n plt.show()\n expt.__del__()\n\n \n if plotSeq == 1:\n expt.plot_sequence()\n plt.show()\n expt.__del__()\n elif plotSeq == 0:\n data = np.reshape(dataAll, (nExTime, nReadout))\n plotData(data, rfExTime, tAdqReal)\n plotRabiFlop(data, rfExTime, tAdqReal)\n plt.show()\n\n# MAIN ######################################################################################################\nif __name__ == \"__main__\":\n rabiflops_standalone()\n", "sub_path": "seq_standalone_OLD/rabiflops3.py", "file_name": "rabiflops3.py", "file_ext": "py", "file_size_in_byte": 6523, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sys.path.append", "line_number": 8, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 52, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.cm.rainbow", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.cm", "line_number": 82, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 86, "usage_type": "name"}, {"api_name": "numpy.abs", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}, {"api_name": "numpy.real", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "experiment.Experiment", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 156, "usage_type": "call"}, {"api_name": "scipy.signal.decimate", "line_number": 164, "usage_type": "call"}, {"api_name": "scipy.signal", "line_number": 164, "usage_type": "name"}, {"api_name": "numpy.concatenate", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 168, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 168, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 174, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 174, "usage_type": "name"}, {"api_name": "numpy.reshape", "line_number": 177, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 180, "usage_type": "name"}]}
+{"seq_id": "652941569", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ]\n\n operations = [\n migrations.CreateModel(\n name='Exercise',\n fields=[\n ('id', models.AutoField(auto_created=True, serialize=False, primary_key=True, verbose_name='ID')),\n ('number', models.IntegerField(unique=True)),\n ('text', models.TextField()),\n ('status', models.BooleanField()),\n ],\n options={\n },\n bases=(models.Model,),\n ),\n migrations.CreateModel(\n name='Lector',\n fields=[\n ('id', models.AutoField(auto_created=True, serialize=False, primary_key=True, verbose_name='ID')),\n ('name', models.TextField()),\n ('info', models.TextField()),\n ],\n options={\n },\n bases=(models.Model,),\n ),\n migrations.CreateModel(\n name='Lesson',\n fields=[\n ('id', models.AutoField(auto_created=True, serialize=False, primary_key=True, verbose_name='ID')),\n ('number', models.IntegerField(unique=True)),\n ('text', models.TextField()),\n ],\n options={\n },\n bases=(models.Model,),\n ),\n migrations.CreateModel(\n name='Mooc',\n fields=[\n ('id', models.AutoField(auto_created=True, serialize=False, primary_key=True, verbose_name='ID')),\n ('title', models.TextField()),\n ('description', models.TextField()),\n ('pub_date', models.DateTimeField(verbose_name='date published')),\n ],\n options={\n },\n bases=(models.Model,),\n ),\n migrations.CreateModel(\n name='User',\n fields=[\n ('id', models.AutoField(auto_created=True, serialize=False, primary_key=True, verbose_name='ID')),\n ('name', models.TextField()),\n ('login', models.TextField()),\n ('password', models.TextField()),\n ('exercise', models.ForeignKey(to='moocs.Exercise')),\n ],\n options={\n },\n bases=(models.Model,),\n ),\n migrations.AddField(\n model_name='lesson',\n name='mooc',\n field=models.ForeignKey(to='moocs.Mooc'),\n preserve_default=True,\n ),\n migrations.AddField(\n model_name='lector',\n name='mooc',\n field=models.ForeignKey(to='moocs.Mooc'),\n preserve_default=True,\n ),\n migrations.AddField(\n model_name='exercise',\n name='lesson',\n field=models.ForeignKey(to='moocs.Lesson'),\n preserve_default=True,\n ),\n ]\n", "sub_path": "moocs/migrations/0001_initial.py", "file_name": "0001_initial.py", "file_ext": "py", "file_size_in_byte": 2986, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.BooleanField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 23, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 28, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 28, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 29, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 29, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 34, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 34, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 36, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 36, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 39, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 39, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 40, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 40, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 41, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 41, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 45, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 45, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 47, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 47, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 50, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 50, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 51, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 51, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 52, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 52, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 53, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 53, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 57, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 57, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 59, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 59, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 62, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 62, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 63, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 63, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 64, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 64, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 65, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 65, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 66, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 66, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 70, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 70, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 72, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 72, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 75, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 75, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 78, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 78, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 81, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 81, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 84, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 84, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 87, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 87, "usage_type": "name"}]}
+{"seq_id": "612361281", "text": "import numpy as np\n\nfrom .crf_dual_layer import DualCRFLayer, get_crf_training_loss\nimport lasagne\nimport time\nimport theano\nimport logging\nfrom .base_network import get_base_network\nimport sys\nimport theano.tensor as T\nfrom bionlp.utils.utils import theano_logsumexp\nlogging.basicConfig()\nlogger = logging.getLogger(__name__)\n\n\ndef setup_NN(worker, x_in, u_in, mask_in, y_in, params, numTags, emb_w):\n premodel = time.time()\n logger.info('Loading CRF-RNN network format with pairwise modeling. '\n 'Forward Backward will be used for calculating the marginals while training. '\n 'The output sequence will be calculated using posterior decoding.')\n\n crf_layer, pairwise, l_in, l_mask, l_u_in = get_base_network(x_in, u_in, mask_in, y_in, params, numTags, emb_w)\n t_out = T.tensor3()\n\n mid_out = lasagne.layers.get_output(crf_layer, deterministic=True)\n mid_output = theano.function(\n [l_in.input_var, l_u_in.input_var, l_mask.input_var], mid_out)\n logger.info(\"output shape for for unary mid layer {0}\".format(mid_output(\n x_in.astype('int32'), u_in.astype('float32'), mask_in.astype('float32')).shape))\n logger.info(\"output sum for for unary mid layer {0}\".format(np.sum(mid_output(\n x_in.astype('int32'), u_in.astype('float32'), mask_in.astype('float32'))[0, :, :], axis=1)))\n\n sum_layer = DualCRFLayer([crf_layer, pairwise, l_mask], mask_input=True)\n\n outp = lasagne.layers.get_output(sum_layer, deterministic=False)\n eval_out = lasagne.layers.get_output(sum_layer, deterministic=True)\n\n crf_output = theano.function(\n [l_in.input_var, l_u_in.input_var, l_mask.input_var], eval_out)\n lstm_output = crf_output # Included for future functionality\n print((\"output shape for theano net\", lstm_output(x_in.astype(\n 'int32'), u_in.astype('float32'), mask_in.astype('float32')).shape))\n eval_cost = T.mean((eval_out - t_out)**2)\n\n all_params = lasagne.layers.get_all_params(sum_layer, trainable=True)\n logger.info('Params :{0}'.format(all_params))\n logger.info(\n '\\'l1\\' and \\'l2\\' Regularization is applied to all trainable parameters in the network')\n l2_cost = params['l2'] * lasagne.regularization.apply_penalty(\n all_params, lasagne.regularization.l2)\n l1_cost = params['l1'] * lasagne.regularization.apply_penalty(\n all_params, lasagne.regularization.l1)\n regularization_losses = l2_cost + l1_cost\n\n num_params = lasagne.layers.count_params(sum_layer, trainable=True)\n print(('Number of parameters: {0}'.format(num_params)))\n\n crf_train_loss = get_crf_training_loss(\n sum_layer, pairwise, t_out, numTags, params, x_in, u_in, y_in, mask_in, l_in, l_u_in, l_mask)\n cost = T.mean(crf_train_loss) + regularization_losses\n cost_loss = T.mean(crf_train_loss)\n cost_regularization = regularization_losses\n\n updates = lasagne.updates.adagrad(\n cost, all_params, learning_rate=params['learning-rate'])\n if params['momentum'] == 2:\n updates = lasagne.updates.apply_nesterov_momentum(\n updates, all_params, momentum=0.9)\n logger.info('Using Nesterov\\'s momentum')\n if params['momentum'] == 1:\n updates = lasagne.updates.apply_momentum(\n updates, all_params, momentum=0.9)\n logger.info('Using Momentum')\n if params['momentum'] == 0:\n logger.warning('Not using any momentum')\n\n train = theano.function(\n [l_in.input_var, l_u_in.input_var, t_out, l_mask.input_var], cost, updates=updates)\n\n compute_cost = theano.function(\n [l_in.input_var, l_u_in.input_var, t_out, l_mask.input_var], cost)\n compute_cost_loss = theano.function(\n [l_in.input_var, l_u_in.input_var, t_out, l_mask.input_var], cost_loss)\n\n compute_cost_regularization = theano.function([], cost_regularization)\n acc_ = T.sum(T.eq(T.argmax(eval_out, axis=2), T.argmax(\n t_out, axis=2)) * l_mask.input_var) / T.sum(l_mask.input_var)\n compute_acc = theano.function(\n [l_in.input_var, l_u_in.input_var, t_out, l_mask.input_var], acc_)\n print(('Time to build and compile model {0}'.format(\n time.time() - premodel)))\n\n return {'crf_output': crf_output, 'lstm_output': lstm_output, 'train': train, 'compute_cost': compute_cost,\n 'compute_acc': compute_acc, 'compute_cost_loss': compute_cost_loss,\n 'compute_cost_regularization': compute_cost_regularization, 'final_layers': sum_layer}\n", "sub_path": "bionlp/taggers/rnn_feature/networks/dual_network.py", "file_name": "dual_network.py", "file_ext": "py", "file_size_in_byte": 4468, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.basicConfig", "line_number": 12, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 13, "usage_type": "call"}, {"api_name": "time.time", "line_number": 17, "usage_type": "call"}, {"api_name": "base_network.get_base_network", "line_number": 22, "usage_type": "call"}, {"api_name": "theano.tensor.tensor3", "line_number": 23, "usage_type": "call"}, {"api_name": "theano.tensor", "line_number": 23, "usage_type": "name"}, {"api_name": "lasagne.layers.get_output", "line_number": 25, "usage_type": "call"}, {"api_name": "lasagne.layers", "line_number": 25, "usage_type": "attribute"}, {"api_name": "theano.function", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 30, "usage_type": "call"}, {"api_name": "crf_dual_layer.DualCRFLayer", "line_number": 33, "usage_type": "call"}, {"api_name": "lasagne.layers.get_output", "line_number": 35, "usage_type": "call"}, {"api_name": "lasagne.layers", "line_number": 35, "usage_type": "attribute"}, {"api_name": "lasagne.layers.get_output", "line_number": 36, "usage_type": "call"}, {"api_name": "lasagne.layers", "line_number": 36, "usage_type": "attribute"}, {"api_name": "theano.function", "line_number": 38, "usage_type": "call"}, {"api_name": "theano.tensor.mean", "line_number": 43, "usage_type": "call"}, {"api_name": "theano.tensor", "line_number": 43, "usage_type": "name"}, {"api_name": "lasagne.layers.get_all_params", "line_number": 45, "usage_type": "call"}, {"api_name": "lasagne.layers", "line_number": 45, "usage_type": "attribute"}, {"api_name": "lasagne.regularization.apply_penalty", "line_number": 49, "usage_type": "call"}, {"api_name": "lasagne.regularization", "line_number": 49, "usage_type": "attribute"}, {"api_name": "lasagne.regularization", "line_number": 50, "usage_type": "attribute"}, {"api_name": "lasagne.regularization.apply_penalty", "line_number": 51, "usage_type": "call"}, {"api_name": "lasagne.regularization", "line_number": 51, "usage_type": "attribute"}, {"api_name": "lasagne.regularization", "line_number": 52, "usage_type": "attribute"}, {"api_name": "lasagne.layers.count_params", "line_number": 55, "usage_type": "call"}, {"api_name": "lasagne.layers", "line_number": 55, "usage_type": "attribute"}, {"api_name": "crf_dual_layer.get_crf_training_loss", "line_number": 58, "usage_type": "call"}, {"api_name": "theano.tensor.mean", "line_number": 60, "usage_type": "call"}, {"api_name": "theano.tensor", "line_number": 60, "usage_type": "name"}, {"api_name": "theano.tensor.mean", "line_number": 61, "usage_type": "call"}, {"api_name": "theano.tensor", "line_number": 61, "usage_type": "name"}, {"api_name": "lasagne.updates.adagrad", "line_number": 64, "usage_type": "call"}, {"api_name": "lasagne.updates", "line_number": 64, "usage_type": "attribute"}, {"api_name": "lasagne.updates.apply_nesterov_momentum", "line_number": 67, "usage_type": "call"}, {"api_name": "lasagne.updates", "line_number": 67, "usage_type": "attribute"}, {"api_name": "lasagne.updates.apply_momentum", "line_number": 71, "usage_type": "call"}, {"api_name": "lasagne.updates", "line_number": 71, "usage_type": "attribute"}, {"api_name": "theano.function", "line_number": 77, "usage_type": "call"}, {"api_name": "theano.function", "line_number": 80, "usage_type": "call"}, {"api_name": "theano.function", "line_number": 82, "usage_type": "call"}, {"api_name": "theano.function", "line_number": 85, "usage_type": "call"}, {"api_name": "theano.tensor.sum", "line_number": 86, "usage_type": "call"}, {"api_name": "theano.tensor", "line_number": 86, "usage_type": "name"}, {"api_name": "theano.tensor.eq", "line_number": 86, "usage_type": "call"}, {"api_name": "theano.tensor.argmax", "line_number": 86, "usage_type": "call"}, {"api_name": "theano.tensor.sum", "line_number": 87, "usage_type": "call"}, {"api_name": "theano.tensor", "line_number": 87, "usage_type": "name"}, {"api_name": "theano.function", "line_number": 88, "usage_type": "call"}, {"api_name": "time.time", "line_number": 91, "usage_type": "call"}]}
+{"seq_id": "96228634", "text": "#! /usr/bin/env python3\n\nimport csv\nimport json\nimport xml.etree.ElementTree as ET\n\n\n##############################################################################\n# Класс заполнения зарплаты\n##############################################################################\nclass Zarplata:\n def __init__(self):\n self.month = ['Січень',\n 'Лютий',\n 'Березень',\n 'Квітень',\n 'Травень',\n 'Червень',\n 'Липень',\n 'Серпень',\n 'Вересень',\n 'Жовтень',\n 'Листопад',\n 'Грудень']\n self.salary = [] # Зарплата {}\n with open('./data/Coefficient.csv', 'r') as theFile:\n reader = csv.DictReader(theFile, delimiter=';')\n i = 0\n for line in reader:\n self.salary.append({'nn': i,\n 'year': int(line['year']),\n 'month': int(line['month']),\n 'Coefficient': float(line['Coefficient'].replace(',', '.')),\n 'zp': 0,\n 'rc': 0,\n 'pension': None,\n 'experience': False})\n i += 1\n theFile.close()\n\n # Метод добавления зарплаты и расчета коэффициента\n def set_zp(self, year, month, zp, experience=True):\n for ll in range(0, len(self.salary)):\n if self.salary[ll].get('year') == year \\\n and self.salary[ll].get('month') == month \\\n and self.salary[ll].get('Coefficient') > 0:\n if year < 1995:\n self.salary[ll]['zp'] = round(zp, 5)\n else:\n self.salary[ll]['zp'] = round(zp, 2)\n\n self.salary[ll]['rc'] = round(zp / self.salary[ll].get('Coefficient'), 5)\n self.salary[ll]['pension'] = False\n self.salary[ll]['experience'] = experience\n\n def get_salary(self):\n return self.salary\n\n def set_pension(self, pens):\n for line in pens:\n self.salary[line.get('nn')]['pension'] = True\n\n # Очищаю для следующего ввода\n def erase_salary(self):\n for ll in range(len(self.salary)):\n self.salary[ll]['zp'] = 0\n self.salary[ll]['rc'] = 0\n self.salary[ll]['pension'] = None\n self.salary[ll]['experience'] = False\n\n # Метод добавления зарплаты из джейсона (наверное будет нужен)\n def set_zp_json(self, zp_json):\n for zz in json.loads(zp_json.decode()):\n self.set_zp(zz.get('year'), zz.get('month'), zz.get('zp'))\n\n # Метод загрузки XML из ПФУ справки ОК5 (можно взять на кабинете ПФУ)\n # Или csv без шапки (\"год\", \"месц(число)\", \"ЗП\")\n # - нужно для отладки, чтоб не набивать данные в ручную\n def import_ok5(self, file_name):\n self.erase_salary()\n ret = []\n if file_name[-3:] == 'xml':\n tree = ET.ElementTree(file=file_name)\n rtree = tree.getroot()\n revenues = rtree.findall('revenues')\n for i in revenues[0]:\n for revenue in i:\n yyyy = ''\n if revenue.tag == 'year':\n yyyy = revenue.text\n if revenue.tag == 'pensgrn':\n mm = 1\n for month in revenue:\n self.set_zp(int(yyyy), int(mm), float(month.text))\n ret.append({'year': int(yyyy), 'month': int(mm), 'zp': float(month.text)})\n mm += 1\n if file_name[-3:] == 'csv':\n with open(file_name) as cf:\n delim = cf.read(5)[-1]\n cf.seek(0)\n in_csv = csv.reader(cf, delimiter=delim)\n for line in in_csv:\n if line[2]:\n zp = float(line[2].replace(',', '.'))\n self.set_zp(int(line[0]), int(line[1]), zp)\n ret.append({'year': int(line[0]), 'month': int(line[1]), 'zp': zp})\n return ret\n\n\n##############################################################################\n# Класс определения лучшего периода\n##############################################################################\nclass GoodSalary:\n def __init__(self, salary, kz, sp1):\n self.kz = kz # Коефициент стажу\n self.rsm = int((kz * 1200) - (sp1 * 2) + sp1) # Расчет месяцев стажа с учетом спецстажа СПИСОК1\n self.salary_old = [] # зарплата {}\n self.salary_new = [] # зарплата {}\n self.start_p = 0 # Определяю начало персонификации\n\n # Разделяем на до 01.07.2000 и после\n for ii in salary:\n if ii.get('year') == 2000 and ii.get('month') == 7:\n self.start_p = ii.get('nn')\n if self.start_p == 0:\n if ii.get('experience'):\n self.salary_old.append(ii)\n else:\n if ii.get('experience'):\n self.salary_new.append(ii)\n self.period_a = [] # Возможные периоды для добавления заполняется методом add_period() из self.salary_old\n self.period_d = [] # Возможные периоды для удаления заполняется методом del_period() из self.salary_good\n self.salary_good = [] # Лучшая зарплата до 01.07.200\n self.salary_clear = [] # масив с чистой зарплаты по которой начислено пенсию\n self.add_p = False # Стоит ли добавлять период\n\n # Метод выбора лучшего период до 01.07.2000\n def add_period(self, add=True):\n for i_start in range(len(self.salary_old)):\n nn = None\n rc = 0\n for index, line in enumerate(self.salary_old[i_start:]):\n if not nn:\n rc = line.get('rc')\n nn = line.get('nn')\n else:\n if nn+1 == line.get('nn'):\n rc += line.get('rc')\n nn = line.get('nn')\n else:\n self.period_a.append([i_start, i_start + index+1, rc, self.salary_old[i_start], nn])\n break\n if index == 59:\n self.period_a.append([i_start, i_start + 59, rc, self.salary_old[i_start], nn])\n break\n\n self.period_a.sort(key=lambda x: x[2], reverse=True)\n if add:\n self.salary_good.extend(self.salary_old[self.period_a[0][0]: self.period_a[0][1]])\n self.salary_good.extend(self.salary_new)\n\n def d_period(self, start, end):\n sg = self.salary_good[:]\n sd = self.salary_good[start:start+end]\n\n for ll in sd:\n sg.remove(ll)\n\n rc = 0\n km = len(sg)\n if sd:\n for rs in sg:\n # km += 1\n rc += rs.get('rc')\n self.period_d.append([start, start+end, km, rc, rc/km,\n self.salary_good[start],\n sd[0].get('nn'), sd[-1].get('nn')])\n else:\n self.period_d.append([start, start+end, 0, 0, 0])\n\n def del_period(self):\n # отнимать можно 10% но не более 60 месяцев\n dm = 60 if int(self.rsm * 0.1) >= 60 else int(self.rsm * 0.1)\n month = len(self.salary_good)\n if month > 60:\n if int(month - dm) <= 60: # остатся может не меньше 60\n mm = int(month - 60)\n else:\n mm = dm\n\n for ii in range(1, mm+1):\n for cc in range(month - ii+1):\n self.d_period(cc, ii)\n\n self.period_d.sort(key=lambda x: x[4], reverse=True)\n self.salary_clear.extend(self.salary_good[:self.period_d[0][0]])\n self.salary_clear.extend(self.salary_good[self.period_d[0][1]:])\n else:\n self.salary_clear.extend(self.salary_good)\n self.period_d.append([0, len(self.salary_good)-1, 0, 0, 0])\n\n def clear(self):\n self.period_a = []\n self.period_d = []\n self.salary_good = []\n self.salary_clear = []\n\n def calc_kz(self, dd=True):\n self.clear()\n self.add_p = False\n self.add_period(add=self.add_p)\n self.del_period()\n try:\n if self.period_d[0][4] < self.period_a[0][2] / 60 or not dd:\n self.clear()\n self.add_p = True\n self.add_period(add=self.add_p)\n self.del_period()\n except IndexError:\n pass\n\n return self.add_p\n", "sub_path": "pension.py", "file_name": "pension.py", "file_ext": "py", "file_size_in_byte": 9522, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "csv.DictReader", "line_number": 27, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 73, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.ElementTree", "line_number": 83, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 83, "usage_type": "name"}, {"api_name": "csv.reader", "line_number": 101, "usage_type": "call"}]}
+{"seq_id": "508922149", "text": "import requests\nimport urllib.parse\n\nlink = 'https://randomuser.me/api/?'\n\nparametersDict = {\n 'gender': ('male', 'female', None),\n 'results': (1, 5000),\n}\n\n\n# https://randomuser.me/api/?password=special,upper,lower,number\n# https://randomuser.me/api/?seed=foobar\n\ndef getPersonList(listSize=10, gender=None):\n minSize, maxSize = parametersDict['results']\n # return {'min': minSize, 'max': maxSize, 'size': listSize}\n if listSize > maxSize or listSize < minSize:\n raise Exception('The list size cannot be greater than 5000 or less than 1')\n\n if gender not in parametersDict['gender']:\n raise Exception('This is not valid gender')\n\n parameters = {'results': listSize}\n if gender is not None:\n parameters['gender'] = gender\n\n response = requests.get(link + urllib.parse.urlencode(parameters))\n\n if response.status_code != 200:\n raise Exception('The serve raised error %s' % response.status_code)\n\n response = response.json()\n response = response['results'], response['info']['seed']\n\n return response\n\n\ndef getPersonBySeed(seed='random'):\n response = requests.get(link + urllib.parse.urlencode({'seed': seed}))\n return response.json()\n\n\n# 'password': instance.password\ndef filterPersons(personList):\n try:\n newPersonList = []\n for person in personList:\n location = person['location']\n newPersonList.append({\n 'gender': person['gender'],\n 'first_name': person['name']['first'],\n 'last_name': person['name']['last'],\n 'title': person['name']['title'],\n 'street': location['street']['name'] + str(location['street']['number']),\n 'city': location['city'],\n 'state': location['state'],\n 'country': location['country'],\n 'email': person['email'],\n 'username': person['login']['username'],\n 'password': person['login']['password'],\n 'phone': person['phone'],\n 'picture small': person['picture']['thumbnail'],\n 'picture medium': person['picture']['medium'],\n 'picture large': person['picture']['large'],\n 'nationality': person['nat']\n })\n\n return newPersonList\n except Exception as e:\n raise Exception(e.args)\n", "sub_path": "Backend/source/authentication/userManager.py", "file_name": "userManager.py", "file_ext": "py", "file_size_in_byte": 2391, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "requests.get", "line_number": 28, "usage_type": "call"}, {"api_name": "urllib.parse.parse.urlencode", "line_number": 28, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 28, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 28, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 40, "usage_type": "call"}, {"api_name": "urllib.parse.parse.urlencode", "line_number": 40, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 40, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 40, "usage_type": "name"}]}
+{"seq_id": "626621368", "text": "import sys\nimport re\nimport xml.etree.ElementTree as etree\nimport xlwt\nimport collections\n\nrule_dict = collections.defaultdict(list)\n\n\nwb = xlwt.Workbook()\nstyle = xlwt.easyxf('pattern: pattern solid, fore_colour gray25;'\n'font: colour black, bold True;'\n'borders: left thin, right thin, top thin, bottom thin;')\nstyle.alignment.wrap = 1\nws = wb.add_sheet('Objects', cell_overwrite_ok=True)\ni=1\n\nDG_list = []\nrule_list = []\nrule_count = 0\n\n\npaloalto = open(sys.argv[1], 'r').read()\n\nroot = etree.fromstring(paloalto)\n\nfor child in root.findall(\".//devices/entry/device-group/\"):\n DG = child.get('name')\n DG_list.append(DG)\nDG_Unique = set(DG_list)\nprint(len(DG_Unique))\n\nfor DG in DG_Unique:\n for child in root.findall(\".//entry[@name='\" + DG + \"']/pre-rulebase/security/rules/\"):\n rule = child.get('name')\n rule_list.append(rule)\n rule_count = len(rule_list)\n #print(rule_count)\n rule_dict[DG].append(rule_count)\n rule_list = []\n rule_count = 0\n#print(rule_dict)\n\nfor child in root.findall(\".//readonly/devices/entry/device-group/\"):\n DG = child.get('name')\n if child.findall('parent-dg'):\n for parent_dg in child.findall('parent-dg'):\n parent_dg_string = parent_dg.text\n DG_count = str(rule_dict.get(DG, \"none\"))\n PG_count = str(rule_dict.get(parent_dg_string, \"none\"))\n else:\n parent_dg_string = 'NONE'\n DG_count = str(rule_dict.get(DG, \"none\"))\n PG_count = 'N/A'\n ws.write(0, 0, 'Device-Group', style)\n ws.write(0, 1, 'Device-Group Count', style)\n ws.write(0, 2, 'Parent Device-Group', style)\n ws.write(0, 3, 'Parent DG Count', style)\n ws.write(i, 0, DG)\n ws.write(i, 1, DG_count)\n ws.write(i, 2, parent_dg_string)\n ws.write(i, 3, PG_count)\n i = i + 1\n parent_dg_string = ''\n DG_count = ''\n PG_count = ''\n\nwb.save('rule_count.xls')\n", "sub_path": "rule_count.py", "file_name": "rule_count.py", "file_ext": "py", "file_size_in_byte": 1884, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "collections.defaultdict", "line_number": 7, "usage_type": "call"}, {"api_name": "xlwt.Workbook", "line_number": 10, "usage_type": "call"}, {"api_name": "xlwt.easyxf", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 23, "usage_type": "attribute"}, {"api_name": "xml.etree.ElementTree.fromstring", "line_number": 25, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 25, "usage_type": "name"}]}
+{"seq_id": "166348972", "text": "\"\"\"Tests for checkin_timeline module\n\nFeel free to add more tests as you see fit.\n\n\"\"\"\nimport random\n\nfrom datetime import datetime, timedelta\n\nfrom checkin import CheckIn\nfrom checkin_timeline import CheckInTimeline\nfrom pokeball import Pokeball\n\n\ndef random_timed_checkins(count=100):\n \"\"\"A helper function that returns CheckIns with random times\n\n Note that the names and locations will all be the same.\n\n :param int count: The number of CheckIns to return\n\n :return: a list of CheckIn instances in a random order.\n \"\"\"\n # Generate a list of random, unique CheckIns\n checkins = []\n time = datetime.fromtimestamp(0)\n for _ in range(count):\n time += timedelta(minutes=random.randint(1, 59))\n c = CheckIn(\"Bob\", Pokeball.poke_ball, \"Pokemart\", str(time))\n checkins.append(c)\n\n # Shuffle up the CheckIns\n random.shuffle(checkins)\n\n return checkins\n\n\ndef random_timed_checkin_timeline(count=100):\n \"\"\"A helper function that returns an CheckInTimeline that contains\n CheckIns with random times.\n\n Note that the names and locations will all be the same.\n\n :param int count: The number of CheckIns in the CheckInTimeline\n\n :return: An CheckInTimeline with ``count`` CheckIns with\n random times.\n\n \"\"\"\n # Generate some checkins\n checkins = random_timed_checkins(count)\n\n # Add the checkins to our timeline\n timeline = CheckInTimeline()\n for t in checkins:\n timeline.add(t)\n\n return timeline\n\n\ndef test_add():\n \"\"\"Test that adding more CheckIns keeps them in order.\"\"\"\n # Generate checkins with random times\n timeline = random_timed_checkin_timeline()\n\n # Check that our checkins are in order\n # (Go look up zip())\n for prev, current in zip(timeline.checkins, timeline.checkins[1:]):\n assert prev.time < current.time\n\n\ndef test_window():\n \"\"\"Test that our window generator respects window_size.\"\"\"\n # Generate checkins with random times\n timeline = random_timed_checkin_timeline()\n\n # Defaults to one hour\n for window in timeline.windows():\n # Gotta be a tuple, though we don't know the length\n assert isinstance(window, tuple)\n assert len(window) > 0\n\n # Check the types\n for o in window:\n assert isinstance(o, CheckIn)\n\n # Double check that CheckIns in the window are sorted (for fun)\n for o1, o2 in zip(window, window[1:]):\n assert o1 < o2\n\n # Make sure each member is within an hour of the first.\n # We know they're sorted, so just check first and last.\n assert (window[0].time + timedelta(hours=1)) > window[-1].time\n\n\ndef test_rendezvous():\n \"\"\"Test our rendezvous generator\"\"\"\n timeline = CheckInTimeline()\n timeline.add(CheckIn(\"Alice\",\n Pokeball.poke_ball,\n \"Pokemart\",\n \"1970-01-02 02:53:00\"))\n\n timeline.add(CheckIn(\"Bob\",\n Pokeball.poke_ball,\n \"Pokemart\",\n \"1970-01-02 03:52:00\"))\n\n for i, agent_pair in enumerate(timeline.rendezvous()):\n # Gotta be a tuple of length 2\n assert isinstance(agent_pair, tuple)\n assert len(agent_pair) == 2\n\n # Unpack 'em\n a1, a2 = agent_pair\n\n # Check the types\n assert isinstance(a1, CheckIn)\n assert isinstance(a2, CheckIn)\n\n # Check that we've got our agents\n assert \"Alice\" in (a1.name, a2.name)\n assert \"Bob\" in (a1.name, a2.name)\n\n # We only looped one time, so 'i' was set to zero, and that's it.\n assert i == 0\n", "sub_path": "test_checkin_timeline.py", "file_name": "test_checkin_timeline.py", "file_ext": "py", "file_size_in_byte": 3636, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "datetime.datetime.fromtimestamp", "line_number": 26, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 26, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 28, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 28, "usage_type": "call"}, {"api_name": "checkin.CheckIn", "line_number": 29, "usage_type": "call"}, {"api_name": "pokeball.Pokeball.poke_ball", "line_number": 29, "usage_type": "attribute"}, {"api_name": "pokeball.Pokeball", "line_number": 29, "usage_type": "name"}, {"api_name": "random.shuffle", "line_number": 33, "usage_type": "call"}, {"api_name": "checkin_timeline.CheckInTimeline", "line_number": 54, "usage_type": "call"}, {"api_name": "checkin.CheckIn", "line_number": 85, "usage_type": "argument"}, {"api_name": "datetime.timedelta", "line_number": 93, "usage_type": "call"}, {"api_name": "checkin_timeline.CheckInTimeline", "line_number": 98, "usage_type": "call"}, {"api_name": "checkin.CheckIn", "line_number": 99, "usage_type": "call"}, {"api_name": "pokeball.Pokeball.poke_ball", "line_number": 100, "usage_type": "attribute"}, {"api_name": "pokeball.Pokeball", "line_number": 100, "usage_type": "name"}, {"api_name": "checkin.CheckIn", "line_number": 104, "usage_type": "call"}, {"api_name": "pokeball.Pokeball.poke_ball", "line_number": 105, "usage_type": "attribute"}, {"api_name": "pokeball.Pokeball", "line_number": 105, "usage_type": "name"}, {"api_name": "checkin.CheckIn", "line_number": 118, "usage_type": "argument"}, {"api_name": "checkin.CheckIn", "line_number": 119, "usage_type": "argument"}]}
+{"seq_id": "205003488", "text": "import argparse\nimport pyBigWig\nimport os\nimport sys\nimport numpy as np\nimport scipy.stats\nimport re\nimport lightgbm as lgb\nimport pickle\n\ndef calculate_mmm(input3d, axis=2): # input 3d output 3d; calculate max, min, mean along axis=2\n output3d=np.zeros((input3d.shape[0], input3d.shape[1], 3))\n output3d[:,:,0]= np.max(input3d, axis=axis)\n output3d[:,:,1]= np.min(input3d, axis=axis)\n output3d[:,:,2]= np.mean(input3d, axis=axis)\n return output3d\n\ndef convert_feature_tensor(input2d, reso=25): # e.g. input2d d0 * (25*n) -> output2d (d0*25) * n \n d0 = input2d.shape[0]\n output2d=input2d.reshape((d0, -1, reso))\n output2d=np.swapaxes(output2d,1,2)\n output2d=output2d.reshape((d0 * reso,-1))\n return output2d\n\nnum_epoch=1\nsize_batch=100000 # 0.1m; must be 25*n\nflank=5\nneighbor=2*flank+1\nflank_dna=1\nneighbor_dna=0 #2*flank_dna+1\n\npath1='../../data_challenge/signal_anchored_final/'\n\nchr_all=['chr1','chr2','chr3','chr4','chr5','chr6','chr7','chr8','chr9','chr10','chr11','chr12','chr13','chr14','chr15','chr16','chr17','chr18','chr19','chr20','chr21','chr22','chrX']\nnum_bp=[248956422,242193529,198295559,190214555,181538259,170805979,159345973,145138636,138394717,133797422,135086622,133275309,114364328,107043718,101991189,90338345,83257441,80373285,58617616,64444167,46709983,50818468,156040895]\n\nchr_len_cut={}\nfor i in np.arange(len(chr_all)):\n chr_len_cut[chr_all[i]]=int(np.floor(num_bp[i]/25.0)*25) # HERE I cut tails\n\nchr_len={}\nfor i in np.arange(len(chr_all)):\n chr_len[chr_all[i]]=num_bp[i]\n\nchr_len25={}\nfor i in np.arange(len(chr_all)):\n chr_len25[chr_all[i]]=int(np.ceil(num_bp[i]/25.0))\n\n## test cell\n#exclude_all=['C04','C17','C20','C24','C32','C34','C46','C48','C50']\nexclude_all=[]\nlist_dna=['A','C','G','T']\n\n# argv\ndef get_args():\n parser = argparse.ArgumentParser(description=\"lightgbm train\")\n parser.add_argument('-f', '--feature', default='M02', nargs='+', type=str,\n help='feature assay')\n parser.add_argument('-t', '--target', default='M18', nargs='+',type=str,\n help='target assay')\n parser.add_argument('-cv', '--crossvalidation', nargs='+', type=str,\n help='crossvalidation cell lines')\n args = parser.parse_args()\n return args\n\nargs=get_args()\n\nexclude_all.extend(args.crossvalidation)\ncell_train=args.crossvalidation[0]\ncell_vali=args.crossvalidation[1]\ncell_test=args.crossvalidation[2]\n\n## target\nlist_label_train=[]\nlist_label_vali=[]\nlist_label_test=[]\nfor the_assay in args.target:\n list_label_train.append(cell_train + the_assay)\n list_label_vali.append(cell_vali + the_assay)\n list_label_test.append(cell_test + the_assay)\n\n## row feature\nlist_feature_train=[]\nlist_feature_vali=[]\nlist_feature_test=[]\nfor the_assay in args.feature:\n list_feature_train.append(cell_train + the_assay)\n list_feature_vali.append(cell_vali + the_assay)\n list_feature_test.append(cell_test + the_assay)\n\n## col feature\nlist_feature_common=[]\nfor the_assay in args.target:\n for i in np.arange(1,52):\n the_cell = 'C' + '%02d' % i\n the_id = the_cell + the_assay\n if (the_cell not in exclude_all) and os.path.isfile(path1 + the_id + '.bigwig'):\n list_feature_common.append(the_id)\n\n# load pyBigWig\n#dict_label_test={}\n#for the_id in list_label_test:\n# dict_label_test[the_id]=pyBigWig.open(path1 + 'gold_' + the_id + '.bigwig')\n#dict_dna={}\n#for the_id in list_dna:\n# dict_dna[the_id]=pyBigWig.open(path1 + the_id + '.bigwig')\ndict_feature_common={}\ndict_orange_common={}\nfor the_id in list_feature_common:\n dict_feature_common[the_id]=pyBigWig.open(path1 + the_id + '.bigwig')\n dict_orange_common[the_id]=pyBigWig.open(path1 + 'orange_' + the_id + '.bigwig')\ndict_feature_test={}\ndict_orange_test={}\nfor the_id in list_feature_test:\n dict_feature_test[the_id]=pyBigWig.open(path1 + the_id + '.bigwig')\n dict_orange_test[the_id]=pyBigWig.open(path1 + 'orange_' + the_id + '.bigwig')\ndict_avg={}\ndict_avg_orange={}\nfor the_assay in args.feature:\n dict_avg[the_assay]=pyBigWig.open(path1 + 'avg_' + the_assay + '.bigwig')\n dict_avg_orange[the_assay]=pyBigWig.open(path1 + 'avg_orange_' + the_assay + '.bigwig')\nfor the_assay in args.target:\n dict_avg[the_assay]=pyBigWig.open(path1 + 'avg_' + the_assay + '.bigwig')\n dict_avg_orange[the_assay]=pyBigWig.open(path1 + 'avg_orange_' + the_assay + '.bigwig')\n\nprint('label_test',list_label_test)\nprint('feature_common',list_feature_common)\nprint('feature_test',list_feature_test)\n\nid_train_vali=list_label_train[0] + '_' + list_label_vali[0] \nfilename='./model/' + id_train_vali + '.model'\ngbm=pickle.load(open(filename, 'rb'))\n\n## prediction\nlist_chr=chr_all\n\nnum_dna=0\nnum_feature=len(list_feature_common)*2 + len(list_feature_test)*2\nnum_orange=len(list_feature_common)*2 + len(list_feature_test)*2\n\nfor the_chr in list_chr:\n print(the_chr)\n pred=np.zeros(chr_len25[the_chr])\n start=0\n while start < chr_len_cut[the_chr] - 25*flank:\n end = min(start + size_batch, chr_len_cut[the_chr])\n image=np.zeros((num_dna + num_feature, end-start), dtype='float32')\n # 2.1 dna\n num=0\n# for j in np.arange(len(list_dna)):\n# the_id=list_dna[j]\n# image[num,:] = dict_dna[the_id].values(the_chr,start,end)\n# num+=1\n # 2.2 feature & diff\n for j in np.arange(len(list_feature_common)):\n the_id=list_feature_common[j]\n # feature\n image[num,:] = dict_feature_common[the_id].values(the_chr,start,end)\n # diff\n the_assay=re.sub('C[0-9][0-9]','',the_id)\n image[num+1,:]=image[num,:] - dict_avg[the_assay].values(the_chr,start,end)\n num+=2\n for j in np.arange(len(list_feature_test)):\n the_id=list_feature_test[j]\n # feature\n image[num,:] = dict_feature_test[the_id].values(the_chr,start,end)\n # diff\n the_assay=re.sub('C[0-9][0-9]','',the_id)\n image[num+1,:]=image[num,:] - dict_avg[the_assay].values(the_chr,start,end)\n num+=2\n\n start_orange=int(start/25)\n end_orange=int(end/25)\n orange=np.zeros((num_orange, end_orange-start_orange), dtype='float32')\n # 2.3 orange & diff\n num=0\n for j in np.arange(len(list_feature_common)):\n the_id=list_feature_common[j]\n # feature\n orange[num,:] = dict_orange_common[the_id].values(the_chr,start_orange,end_orange)\n # diff\n the_assay=re.sub('C[0-9][0-9]','',the_id)\n orange[num+1,:]=orange[num,:] - dict_avg_orange[the_assay].values(the_chr,start_orange,end_orange)\n num+=2\n for j in np.arange(len(list_feature_test)):\n the_id=list_feature_test[j]\n # feature\n orange[num,:] = dict_orange_test[the_id].values(the_chr,start_orange,end_orange)\n # diff\n the_assay=re.sub('C[0-9][0-9]','',the_id)\n orange[num+1,:]=orange[num,:] - dict_avg_orange[the_assay].values(the_chr,start_orange,end_orange)\n num+=2\n\n # convert to 25bp features\n # dna - 25bp one hot\n# image1=convert_feature_tensor(image[:num_dna,:], reso=25) # this reso is different from train.py\n # assay - mmm features\n image2=image[num_dna:,:].reshape((num_feature, -1, 25)) # 3d - d0 * n * 25\n image2=calculate_mmm(image2) # 3d - d0 * n * 3\n image2=image2.reshape((num_feature, -1)) # 2d - d0 * (n*3)\n image2=convert_feature_tensor(image2, reso=3) # 2d - (d0*3) * n\n\n # convert to largespace; f1 neighbors -> f2 neighbors -> ...\n largespace=np.zeros((num_dna*25*neighbor_dna+num_feature*3*neighbor+num_orange*neighbor, \\\n int((end-start)/25)-2*flank))\n # dna - 25bp one hot\n# for n1 in np.arange(num_dna):\n# for n2 in np.arange(neighbor_dna):\n# tmp1 = n1*25*neighbor_dna + n2*25\n# tmp2 = tmp1 + 25\n# tmp = flank-flank_dna # neighbor_dna short so we need to shift it\n# largespace[tmp1:tmp2,:]=image1[n1*25:(n1+1)*25, \\\n# (n2+tmp):(int((end-start)/25)-2*flank+n2+tmp)]\n # assay - mmm features\n for n1 in np.arange(num_feature):\n for n2 in np.arange(neighbor):\n tmp1 = n1*3*neighbor + n2*3 + num_dna*25*neighbor_dna\n tmp2 = tmp1 + 3\n largespace[tmp1:tmp2,:]=image2[n1*3:(n1+1)*3, n2:(int((end-start)/25)-2*flank+n2)]\n # orange\n for n1 in np.arange(num_orange):\n for n2 in np.arange(neighbor):\n tmp1 = n1*neighbor + n2*1 + num_feature*3*neighbor + num_dna*25*neighbor_dna\n tmp2 = tmp1 + 1\n largespace[tmp1:tmp2,:]=orange[n1:(n1+1), n2:(int((end-start)/25)-2*flank+n2)]\n\n start_pred=int(start/25) + flank # skip both ends of a chromosome for now \n end_pred=int(end/25) - flank\n pred[start_pred:end_pred]=gbm.predict(largespace.T)\n del image\n del orange\n# del image1\n del image2\n del largespace\n\n start = end - 25*flank\n\n np.save('pred25bp_' + list_label_test[0] + '_' + the_chr, pred)\n\n## close bw\n#for the_id in dict_label_test.keys():\n# dict_label_test[the_id].close()\n#for the_id in dict_dna.keys():\n# dict_dna[the_id].close()\nfor the_id in dict_feature_common.keys():\n dict_feature_common[the_id].close()\nfor the_id in dict_feature_test.keys():\n dict_feature_test[the_id].close()\nfor the_id in dict_avg.keys():\n dict_avg[the_id].close()\nfor the_id in dict_orange_common.keys():\n dict_orange_common[the_id].close()\nfor the_id in dict_orange_test.keys():\n dict_orange_test[the_id].close()\nfor the_id in dict_avg_orange.keys():\n dict_avg_orange[the_id].close()\n\n\n\n\n", "sub_path": "code_challenge/template_lgbm_v1/pred25bp.py", "file_name": "pred25bp.py", "file_ext": "py", "file_size_in_byte": 9843, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "numpy.zeros", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.swapaxes", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 47, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 97, "usage_type": "call"}, {"api_name": "os.path", "line_number": 97, "usage_type": "attribute"}, {"api_name": "pyBigWig.open", "line_number": 110, "usage_type": "call"}, {"api_name": "pyBigWig.open", "line_number": 111, "usage_type": "call"}, {"api_name": "pyBigWig.open", "line_number": 115, "usage_type": "call"}, {"api_name": "pyBigWig.open", "line_number": 116, "usage_type": "call"}, {"api_name": "pyBigWig.open", "line_number": 120, "usage_type": "call"}, {"api_name": "pyBigWig.open", "line_number": 121, "usage_type": "call"}, {"api_name": "pyBigWig.open", "line_number": 123, "usage_type": "call"}, {"api_name": "pyBigWig.open", "line_number": 124, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 155, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 163, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 177, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 182, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 185, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 215, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 216, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 238, "usage_type": "call"}]}
+{"seq_id": "641435955", "text": "import keras\nimport preprocess as pre\nimport imdb\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport gc\nimport InitializeModel as im\nimport tf_idf as tfidf\nimport scipy.sparse as sp\nimport pickle\nfrom tf_idf import stemmed_words\n\n'''\nprint('Loading model')\nmodel = keras.models.load_model('models/simple_model.h5')\n'''\n\nprint('reading IMDB_data')\n(train_data, train_labels) = pre.loadData('dataset/aclImdb/train/pos', 'dataset/aclImdb/train/neg')\n(test_data, test_labels) = pre.loadData('dataset/aclImdb/test/pos', 'dataset/aclImdb/test/neg')\n'''\noptional:Analyzing Dataset\n'''\nprint(\"Categories:\", np.unique(train_labels))\nprint(\"Number of unique words:\", len(np.unique(np.hstack(train_data))))\n\n# 将word_index反转,实现将整数索引到单词的映射\n'''\n# Simple Vectoring data\nprint('Vectoring data')\nX_train = pre.vectorize_sequences(train_data)\nX_test = pre.vectorize_sequences(test_data)\n'''\n# TF-IDF Vectoring data\nprint('\\nVectoring train data')\nX_train, train_labels = tfidf.tf_idf_2doc(train_data, train_labels, feat=10000)\nprint('\\nVectoring test data')\nX_test, test_labels = tfidf.tf_idf_2doc(test_data, test_labels, feat=10000)\ndata = sp.vstack((X_train, X_test))\n\n# Vectoring label\nprint('\\nVectoring labels')\ny_train = np.asarray(train_labels).astype('float32')\ny_test = np.asarray(test_labels).astype('float32')\ntarget = np.append(y_train, y_test)\n'''\nX_val = X_test[: 10000]\npartial_x_train = X_test[10000:]\n\ny_val = y_test[: 10000]\npartial_y_train = y_test[10000:]\n'''\ntrain_x = data[10000:]\ntrain_y = target[10000:]\ntest_x = data[:10000]\ntest_y = target[:10000]\n\nprint('\\nInitializing model')\nmodel = im.default_modelConf(size=10000)\nprint('\\nTraining model')\nhistory = model.fit(train_x,\n train_y,\n epochs=11,\n batch_size=500,\n validation_data=(test_x, test_y))\nhistory_dict = history.history\nprint(history_dict.keys())\n\n# figure show\n# \"bo\" is for \"blue dot\"\nepochs = range(1, len(history.history['accuracy']) + 1)\nplt.plot(epochs, history.history['loss'], 'bo', label='Training loss')\nplt.plot(epochs, history.history['accuracy'], 'ro', label='Training acc')\n# b is for \"solid blue line\"\nplt.plot(epochs, history.history['val_loss'], 'b', label='Validation loss')\nplt.plot(epochs, history.history['val_accuracy'], 'r', label='Validation acc')\nplt.title('Loss and Acc')\nplt.xlabel('Epochs')\nplt.ylabel('Loss and Acc')\nplt.legend()\nplt.grid()\nplt.show()\nmodel.save('models/tfidf_default_40000.h5')\nplt.savefig('report/tfidf_default_40000.png')\nwith open('report/tfidf_default_40000.txt', 'wb') as fhis:\n pickle.dump(history.history, fhis)\nfhis.close()\ngc.collect()\n", "sub_path": "tfidf_train.py", "file_name": "tfidf_train.py", "file_ext": "py", "file_size_in_byte": 2680, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "preprocess.loadData", "line_number": 19, "usage_type": "call"}, {"api_name": "preprocess.loadData", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 25, "usage_type": "call"}, {"api_name": "tf_idf.tf_idf_2doc", "line_number": 36, "usage_type": "call"}, {"api_name": "tf_idf.tf_idf_2doc", "line_number": 38, "usage_type": "call"}, {"api_name": "scipy.sparse.vstack", "line_number": 39, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 39, "usage_type": "name"}, {"api_name": "numpy.asarray", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 45, "usage_type": "call"}, {"api_name": "InitializeModel.default_modelConf", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 73, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 80, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 80, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 81, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 81, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "pickle.dump", "line_number": 86, "usage_type": "call"}, {"api_name": "gc.collect", "line_number": 88, "usage_type": "call"}]}
+{"seq_id": "41241032", "text": "from graphics import *\nimport random\nfrom copy import copy\nfrom os import _exit\nimport threading\nfrom asyncio import Queue\n\n\ndef synchronized(method):\n def new_method(self, *arg, **kws):\n with self.lock:\n return method(self, *arg, **kws)\n return new_method\n\n\n############################################################\n# BLOCK CLASS\n############################################################\n\nclass Block(Rectangle):\n ''' Block class:\n Implement a block for a tetris piece\n Attributes: x - type: int\n y - type: int\n specify the position on the tetris board\n in terms of the square grid\n '''\n\n BLOCK_SIZE = 25 \n SIDE_LENGTH = 2\n\n lock = threading.Lock()\n\n def __init__(self, pos, color):\n \n self.x = int(pos.x)\n self.y = int(pos.y) # We take the top to be y=0 for calculation purposes\n self.color = color\n \n pos.y = Tetris.BOARD_HEIGHT- 2*pos.y\n pos.x *= 2 \n p2 = Point(pos.x+ self.SIDE_LENGTH, pos.y - self.SIDE_LENGTH ) \n Rectangle.__init__(self,pos,p2) # However in the Rectangle object, the lower left is (0,0)\n self.setFill(color)\n \n @synchronized \n def move(self, dx=0, dy=1):\n ''' Parameters: dx - type: int\n dy - type: int\n \n moves the block dx blocks in the x direction\n and dy blocks in the y direction\n '''\n \n self.x += int(dx)\n self.y += int(dy)\n #print(\"Moving block to x: {}, y: {}\".format(self.x, self.y))\n Rectangle.move(self, dx*self.SIDE_LENGTH ,-dy*self.SIDE_LENGTH)\n\n############################################################\n# SHAPE CLASS\n############################################################\n\nclass Shape():\n ''' Shape class:\n Base class for all the tetris shapes\n Attributes: blocks - type: list - the list of blocks making up the shape\n rotation_dir - type: int - the current rotation direction of the shape\n shift_rotation_dir - type: Boolean - whether or not the shape rotates\n '''\n\n def __init__(self, coords=None, color='blue'):\n self.blocks = []\n ### A boolean to indicate if a shape shifts rotation direction or not.\n ### Defaults to false since only 3 shapes shift rotation directions (I, S and Z)\n \n if(coords!=None): \n for pos in coords:\n self.blocks.append(Block(pos, color))\n\n\n def get_blocks(self):\n '''returns the list of blocks\n '''\n return self.blocks\n\n def draw(self, win):\n ''' Parameter: win - type: CanvasFrame\n\n Draws the shape \n ''' \n for block in self.blocks:\n # print(\"Drawing block at x: {}, y: {}\".format(block.x, block.y))\n block.draw(win)\n def deepcopy(self):\n ''' Returns a deep copy of self with new blocks and such '''\n new = Shape()\n for block in self.get_blocks():\n \n newBlock = Block(Point(block.x,block.y),block.color)\n new.blocks.append(newBlock)\n if(block==self.center_block):\n new.center_block=newBlock\n return new\n \n \n def test_move(self, point):\n '''Returns a shallow copy of self, moved\n '''\n tmp = self.deepcopy()\n tmp.move(point) \n return tmp\n\n def move(self, point):\n\n ''' Parameters: point \n point.x - dx - type: int\n point.y - dy - type: int\n\n moves the shape dx squares in the x direction\n and dy squares in the y direction\n '''\n #print(\"move x: {}, y: {}\".format(point.x, point.y))\n for block in self.blocks:\n block.move(point.x, point.y)\n\n\n def test_rotate(self, direction):\n ''' Parameters: x, y coordinates of rotation block\n Return value: Shape object, a deep copy of self\n \n This function returns a \"test rotation\" of the shape, a\n deep copy that was rotated and did not affect the original,\n but that can be tested to see if it will fit on the board.\n '''\n temp_shape = self.deepcopy()\n temp_shape.rotate(direction)\n return temp_shape \n\n def rotate(self, direction='Right'):\n ''' Parameters: board - type: Board object\n\n rotates the shape:\n 1. Get the rotation direction using the get_rotation_dir method\n 2. Compute the position of each block after rotation\n 3. Move the block to the new position\n \n ''' \n sign = 1 if direction=='Right' else -1 # If not Right, then Left\n \n c_x = self.center_block.x\n c_y = self.center_block.y\n\n for block in self.blocks:\n x_diff, y_diff = c_x-block.x, c_y-block.y # The block's x,y differences from center_block \n\n # To perform the rotation, we (abstractly) transform each block to a grid where\n # the center block is the origin. We then swap x, y differences from origin.\n # The resultant x diff (the original y diff) is negated for right rotation.\n # The results are then transformed back to the original grid.\n\n block.move( (sign*(y_diff) + c_x - block.x) , (-sign*(x_diff) + c_y - block.y))\n \n \n\n############################################################\n# ALL SHAPE CLASSES\n############################################################\n\n \nclass I_shape(Shape):\n def __init__(self, center):\n coords = [Point(center.x - 2, center.y),\n Point(center.x - 1, center.y),\n Point(center.x , center.y),\n Point(center.x + 1, center.y)]\n Shape.__init__(self, coords, 'blue')\n self.center_block = self.blocks[2]\n\nclass J_shape(Shape):\n def __init__(self, center):\n coords = [Point(center.x - 1, center.y),\n Point(center.x , center.y),\n Point(center.x + 1, center.y),\n Point(center.x + 1, center.y + 1)]\n Shape.__init__(self, coords, 'orange') \n self.center_block = self.blocks[1]\n\nclass L_shape(Shape):\n '''\n A four-piece L \n '''\n def __init__(self, center):\n coords = [Point(center.x - 1, center.y),\n Point(center.x , center.y),\n Point(center.x + 1, center.y),\n Point(center.x - 1, center.y + 1)]\n Shape.__init__(self, coords, 'cyan') \n self.center_block = self.blocks[0]\n\n\nclass O_shape(Shape):\n '''\n A four-piece square\n '''\n def __init__(self, center):\n coords = [Point(center.x , center.y),\n Point(center.x - 1, center.y),\n Point(center.x , center.y + 1),\n Point(center.x - 1, center.y + 1)]\n Shape.__init__(self, coords, 'red')\n self.center_block = self.blocks[0]\n\n def rotate(self, direction):\n # Override Shape's rotate method since O_Shape does not rotate\n return \n\nclass S_shape(Shape):\n \n def __init__(self, center):\n coords = [Point(center.x , center.y),\n Point(center.x , center.y + 1),\n Point(center.x + 1, center.y),\n Point(center.x - 1, center.y + 1)]\n Shape.__init__(self, coords, 'green')\n self.center_block = self.blocks[0]\n\n\nclass T_shape(Shape):\n def __init__(self, center):\n coords = [Point(center.x - 1, center.y),\n Point(center.x , center.y),\n Point(center.x + 1, center.y),\n Point(center.x , center.y + 1)]\n Shape.__init__(self, coords, 'yellow')\n self.center_block = self.blocks[1]\n\n\nclass Z_shape(Shape):\n def __init__(self, center):\n coords = [Point(center.x - 1, center.y),\n Point(center.x , center.y), \n Point(center.x , center.y + 1),\n Point(center.x + 1, center.y + 1)]\n Shape.__init__(self, coords, 'magenta')\n self.center_block = self.blocks[1]\n\n\n\n############################################################\n# BOARD CLASS\n############################################################\n\nclass Board(GraphWin):\n ''' Board class: it represents the Tetris board\n GraphWin is a sub-class of tk.Canvas, so Board is also a Canvas.\n\n Attributes: width - type:int - width of the board in squares\n height - type:int - height of the board in squares\n canvas - type:CanvasFrame - where the pieces will be drawn\n grid - type:Dictionary - keeps track of the current state of\n the board; stores the blocks for a given position\n '''\n\n lock = threading.Lock()\n \n def __init__(self, title, width=200, height=200):\n super().__init__(title, width, height)\n # create a canvas to draw the tetris shapes on\n super().setBackground('light gray')\n #super().setCoords(0,0,width/Block.BLOCK_SIZE,height/Block.BLOCK_SIZE)\n super().setCoords(0,0,Tetris.BOARD_WIDTH,Tetris.BOARD_HEIGHT)\n # The grid is a two dimensional list which holds a Block at every location a Block can be.\n self.blank_block = Block(Point(0,0), 'blue')\n self.grid = [[self.blank_block for y in range(Tetris.BB_HEIGHT)] for x in range(Tetris.BB_WIDTH)]\n self.game_over = False\n self.cant_move = False\n\n self.text_score = Text(Point(2,Tetris.BOARD_HEIGHT-1), \"Score: 0\")\n self.text_score.draw(self)\n\n def add_score(self, val):\n new_score = val + int(self.text_score.getText()[7:])\n self.text_score.setText(\"Score: \" + str(new_score))\n\n\n @synchronized \n def move_on_board(self,point=Point(0,1)):\n ''' Parameters: x - type:int\n y - type:int\n Return value: type: bool\n\n if there is already a block at that postion, can't move there\n return False\n\n otherwise return True\n '''\n if(point==None): return\n \n moved_shape = self.active_shape.test_move(point)\n \n for block in moved_shape.get_blocks():\n # Check if the translated block is valid, and either open or a part of the original shape\n if not(self.valid_block(block)):\n if(block.y>=Tetris.BB_HEIGHT): self.cant_move = True\n return False \n if not(self.open_block(block)) and not(self.grid[block.x][block.y] in self.active_shape.get_blocks()):\n self.cant_move = True\n return False\n \n self.remove_shape(self.active_shape)\n #self.active_shape.move(point)\n self._add_shape(moved_shape)\n \n return True\n @synchronized\n def rotate(self, direction):\n '''Parameters: direction string - 'Left' for CCW or 'Right' for CW\n Determines if the current shape can move in the given direction\n '''\n if self.cant_move == True: return False\n rotated_shape = self.active_shape.test_rotate(direction)\n\n for block in rotated_shape.get_blocks():\n if(\n not self.valid_block(block) or\n not(self.open_block(block) or \n self.grid[block.x][block.y] in self.active_shape.get_blocks())\n ):\n return False\n self.remove_shape(self.active_shape) # Removes from grid\n #self.active_shape.rotate(direction) #Transforms shape's internals\n self._add_shape(rotated_shape) # Adds back transformed shape to the grid\n return True \n\n def valid_block(self, block):\n #print(\"Testing valid: x: {}, y: {} \".format(block.x, block.y)) \n if(block.x<0 or block.y<0 or block.x>=Tetris.BB_WIDTH or block.y>=Tetris.BB_HEIGHT):\n return False\n return True\n\n def open_block(self, block):\n \n #print(\"Testing open: x: {}, y: {} \".format(block.x, block.y)) \n if self.grid[block.x][block.y] != self.blank_block:\n return False\n return True\n\n def add_shape(self, shape):\n ''' Parameter: shape - type:Shape\n \n add a shape to the board grid\n '''\n \n self.clean_rows()\n\n for block in shape.get_blocks():\n if not(self.valid_block(block)): \n raise RuntimeError(\"The block with coordinates x: {}, y: {} was out of bounds.\".format(block.x, block.y)) \n if not self.open_block(block):\n self.game_over = True \n return False # A shape could not be added so the game is over\n \n self.grid[block.x][block.y] = block\n \n self.active_shape = shape\n shape.draw(self)\n return True \n\n def _add_shape(self, shape):\n '''Does not check for validity'''\n for block in shape.get_blocks():\n self.grid[block.x][block.y] = block\n block.draw(self)\n self.active_shape = shape\n \n\n def remove_shape(self, shape):\n ''' Removes the shape from the board's grid\n '''\n for block in self.active_shape.get_blocks():\n self.grid[block.x][block.y] = self.blank_block\n block.undraw()\n\n def clean_rows(self):\n ''' removes all the complete rows and shifts down after deletions\n '''\n delete_ct = 0\n row = Tetris.BB_HEIGHT-1 # Start cleaning from the bottom\n while row>self.find_empty_row():\n if self.row_is_complete(row): \n self.delete_row(row)\n delete_ct += 1\n else: row -= 1\n \n if delete_ct>0: self.add_score(10*(delete_ct**3))\n\n def find_empty_row(self, num=1):\n \n if(num<=0): raise RuntimeError(\"Cannot find '{}'th empty row!\".format(num))\n for row in range(Tetris.BB_HEIGHT-1, -1, -1):\n if self.row_is_empty(row):\n if(num==1): return row\n else: num-=1\n return -1 # No empty rows!\n\n def delete_row(self, y):\n ''' Parameters: y - type:int\n\n remove all the blocks in row y\n '''\n \n for i in range(Tetris.BB_WIDTH):\n self.grid[i][y].undraw()\n self.grid[i][y] = self.blank_block\n \n self.move_down_rows(y-1)\n \n def row_is_empty(self, y):\n '''Returns True if the row is empty'''\n for col in range(Tetris.BB_WIDTH):\n block = self.grid[col][y] \n if block != self.blank_block: \n return False\n return True \n\n def row_is_complete(self, y): \n ''' Parameter: y - the row index - type: int\n Return value: type: bool\n '''\n \n for i in range(Tetris.BB_WIDTH):\n # print(\"x:\",i,\"y:\",y)\n if(self.grid[i][y]==self.blank_block): \n return False\n\n return True\n\n def move_down_rows(self, y_start):\n ''' Parameters: y_start - type:int \n Move down rows from y_start to the last non-empty row\n '''\n top = self.find_empty_row(2)\n if y_start<=top: return # No rows to move\n\n for row in range(y_start, top, -1):\n self.move_down_row(row)\n \n def move_down_row(self, y):\n '''Just moves down a given row''' \n if y>=Tetris.BB_HEIGHT-1: raise RuntimeError(\"Cannot move down the bottom row!\")\n for col in range(Tetris.BB_WIDTH):\n block = self.grid[col][y]\n self.grid[col][y] = self.blank_block\n block.move(0,1) #down\n self.grid[col][y+1] = block\n\n def show_game_over(self):\n ''' Call when the game has ended\n '''\n self.board.game_over = True\n \n #Display GAME OVER\n text = Text(Point(Tetris.BOARD_WIDTH/2,Tetris.BOARD_HEIGHT/2), \"Game over\")\n text.setFill('black')\n text.setSize(36)\n text.draw(self.board)\n\n\n \n \n\n############################################################\n# TETRIS CLASS\n############################################################\n\nclass Tetris():\n ''' Tetris class: Controls the game play\n Attributes:\n SHAPES - type: list (list of Shape classes)\n DIRECTION - type: dictionary - converts string direction to (dx, dy)\n BOARD_WIDTH - type:int - the width of the board\n BOARD_HEIGHT - type:int - the height of the board\n board - type:Board - the tetris board\n delay - type:int - the speed in milliseconds for moving the shapes\n current_shape - type: Shape - the current moving shape on the board\n '''\n \n \n SHAPES = [I_shape, J_shape, L_shape, O_shape, S_shape, T_shape, Z_shape]\n DIRECTION = {'Left': Point(-1, 0), 'Right': Point(1, 0), 'Down': Point(0,1)}\n # The true coordinates\n BOARD_WIDTH = 20 \n BOARD_HEIGHT = 40\n # Used for Block calculations\n BB_WIDTH = BOARD_WIDTH//Block.SIDE_LENGTH \n BB_HEIGHT = BOARD_HEIGHT//Block.SIDE_LENGTH \n \n def __init__(self, title, delay=800):\n self.queue = Queue(3000)\n self.board = Board(title, Block.BLOCK_SIZE*self.BOARD_WIDTH, Block.BLOCK_SIZE*self.BOARD_HEIGHT)\n self.delay = delay #ms\n # set the current shape to a random new shape\n if not self.create_new_shape(): raise RuntimeError(\"The initial shape could not be created.\")\n \n # Bind key-presses\n self.board.bind_all('', self.key_eval)\n \n\n def animate(self):\n \n #print(\"auto_move\")\n if self.board.cant_move:\n self.board.cant_move = False\n self.create_new_shape()\n \n if(self.board.game_over):\n self.board.show_game_over()\n else:\n #pass\n self.queue.put_nowait('Down')\n\n \n def create_new_shape(self):\n ''' Return value: type: Shape\n Create a random new shape that is centered\n at y = 0 and x = int(self.BOARD_WIDTH/2)\n set the current_shape with this shape\n '''\n shape = Tetris.SHAPES[random.randrange(0,len(Tetris.SHAPES))]\n center = Point(self.BB_WIDTH//2,0)\n self.current_shape = shape(center) # Creates new instance of whichever shape, passing in the center\n\n if not self.board.add_shape(self.current_shape): return False\n return True\n\n def key_eval(self, evnt):\n '''\n if the user presses the space bar 'space', the shape will move\n down until it can no longer move and is added to the board\n\n if the user presses the ['Ctrl','0'] key ,\n the shape should rotate ['Left','Right'].\n\n '''\n key = evnt.keysym # self.board.checkKey() \n \n if(self.board.game_over or key=='e' or key=='c'): os._exit(0)\n\n if(key==\"\"): return # If there was no key pressed, do nothing\n elif(key=='Control_R'):\n self.queue.put_nowait('Rotate Left')\n elif key=='KP_0':\n self.queue.put_nowait('Rotate Right')\n elif key=='space':\n self.queue.put_nowait('All Down')\n elif len(key)>1:\n self.queue.put_nowait(key)\n \n # print(key)\n\n def update(self):\n ''' Processes updates from self.queue '''\n while not self.queue.empty():\n if self.board.isClosed(): os._exit(0) \n item = self.queue.get_nowait()\n if item == None: continue\n elif item=='Rotate Right':\n self.board.rotate('Right')\n elif item=='Rotate Left':\n self.board.rotate('Left')\n elif item=='All Down':\n while not self.board.cant_move:\n self.board.move_on_board()\n else:\n direction = self.DIRECTION.get(item, None)\n if(direction != None): self.board.move_on_board(direction)\n\n \n################################################################\n# Start the game\n################################################################\n\ngame = Tetris(\"Tetris\")\n\ndef auto_move():\n ''' Moves the current_shape every self.delay ms\n '''\n if not(game.board.game_over):\n game.animate() \n game.board.getRoot().after(game.delay, auto_move)\n\ndef auto_update():\n if not(game.board.game_over):\n game.update()\n game.board.getRoot().after(100, auto_update)\n\ngame.board.getRoot().after(0, auto_move)\ngame.board.getRoot().after(0, auto_update)\ngame.board.getRoot().mainloop()\n", "sub_path": "tetris.py", "file_name": "tetris.py", "file_ext": "py", "file_size_in_byte": 20706, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "threading.Lock", "line_number": 32, "usage_type": "call"}, {"api_name": "threading.Lock", "line_number": 265, "usage_type": "call"}, {"api_name": "asyncio.Queue", "line_number": 500, "usage_type": "call"}, {"api_name": "random.randrange", "line_number": 530, "usage_type": "call"}, {"api_name": "os._exit", "line_number": 548, "usage_type": "call"}, {"api_name": "os._exit", "line_number": 565, "usage_type": "call"}]}
+{"seq_id": "479101087", "text": "# Copyright 2015 Juliano Martinez, Taylan Develioglu\n# All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport datetime\nimport shutil\nimport time\nimport tempfile\n\nfrom contextlib import contextmanager\nfrom functools import partial\n\nfrom haproxyadmin import haproxy\n\nfrom balancer.agent import Worker\nfrom balancer.agent.exceptions import InitializationError, ConfigurationDiffError\nfrom .haproxy_parser import (\n parse, compare, get_frontends, get_socket_dir,\n NO_CHANGES_FOUND, SHOULD_PATCH, SHOULD_RELOAD\n)\n\n@contextmanager\ndef temp_file(lst):\n temp = tempfile.NamedTemporaryFile(mode='w')\n temp.writelines(lst)\n temp.flush()\n yield(temp)\n temp.close()\n\n\ndef calculate_differences(from_file_path, to_file_path, logger):\n \"\"\"Given two haproxy configuration files, find the differences between them.\n \"\"\"\n from_ = parse(open(from_file_path, 'r'))\n to_ = parse(open(to_file_path, 'r'))\n\n result, servers_to_patch = compare(from_, to_, logger)\n if result == SHOULD_PATCH:\n # Find the diferences between the servers in the configuration file.\n return False, servers_to_patch\n elif result == SHOULD_RELOAD:\n return True, servers_to_patch\n elif result == NO_CHANGES_FOUND:\n return False, None\n\n\ndef timestamp_in_microsecs():\n \"\"\"Returns the timestamp in microseconds\n \"\"\"\n now = datetime.datetime.now()\n ts = time.mktime(now.timetuple()) * 1e3 + now.microsecond / 1e3\n return str(ts).replace('.', '')\n\n\nclass HAProxyWorker(Worker):\n def __init__(self, config, *args, **kwargs):\n super().__init__(config, *args, **kwargs)\n self.configp = partial(self.config.get, self.shortname())\n plugins = self.configp('plugins')\n self.__load_plugins__(plugins.split())\n\n def __plugins__(self, parsed_config):\n if self._plugins:\n frontends = get_frontends(parsed_config)\n for plugin in self._plugins:\n with plugin.Plugin(self.logger, self.config) as p:\n p.apply(frontends)\n\n @property\n def config_file_path(self):\n return self.configp('config_file_path')\n\n def will_start(self):\n active_config_file_path = self.download_active_config_file()\n\n if self.check_configuration(active_config_file_path):\n self.logger(\"New configuration file is valid!\")\n self.copy_file(active_config_file_path,\n self.config_file_path)\n parsed_config = parse(open(self.config_file_path, 'r'))\n self.__plugins__(parsed_config)\n\n else:\n self.logger(\"New configuration file is not valid :(\")\n self.logger(\"Keeping the existing configuration file.\")\n if self.check_configuration(self.config_file_path):\n self.logger(\"Existing configuration is valid!\")\n else:\n raise InitializationError(\n \"Existing configuration file is not valid!\")\n\n self.reload_service()\n self.logger(\"Initialization finished.\")\n\n def download_active_config_file(self):\n \"\"\"Downloads the active configuration for this load balancer.\n \"\"\"\n config_file_path = self.config_file_path \\\n + '.' + timestamp_in_microsecs()\n\n self.logger(\n \"Storing configuration in {}\".format(config_file_path))\n\n # Write the configuration file and immediately flushes the buffer. If we\n # don't flush here, we don't have anything copied to the destination\n # later on.\n with temp_file(self.lbconfig.body) as temp:\n self.copy_file(temp.name, config_file_path)\n\n return config_file_path\n\n def disable_server(self, message):\n \"\"\"Disable receiving traffic in the server. Under the hood, it should\n disable ECMP.\n \"\"\"\n self.logger(\"disable server\")\n return True\n\n def enable_server(self, message):\n \"\"\"Enable this server to receive traffic. Under the hood, it should\n enable ECMP.\n \"\"\"\n self.logger(\"enable server\")\n return True\n\n def reload_service(self, message=None):\n reload_command = self.configp('reload_command')\n return self._run_command_(reload_command)\n\n def disable_service(self, message=None):\n \"\"\"Stop HAProxy service in this box.\n \"\"\"\n stop_command = self.configp('stop_command')\n return self._run_command_(stop_command)\n\n def enable_service(self, message=None):\n \"\"\"Start HAProxy service in this box.\n \"\"\"\n start_command = self.configp('start_command')\n return self._run_command_(start_command)\n\n def check_configuration(self, file_path):\n check_command = self.configp('check_command').format(file_path=file_path)\n return self._run_command_(check_command)\n\n def copy_file(self, from_path, to_path):\n try:\n self.logger(\"Copying from {} to {}\".format(from_path, to_path))\n shutil.copyfile(from_path, to_path)\n self.logger(\"File copied.\")\n except IOError as e:\n self.logger(\n \"Could not copy {} to {}. Aborting\".format(from_path, to_path))\n\n def workflow(self, message):\n \"\"\"Replace configuration file and reload HAProxy.\n \"\"\"\n active_config_file_path = self.download_active_config_file()\n\n if not self.check_configuration(active_config_file_path):\n self.logger(\"New configuration file is not valid :( Aborting.\")\n return False\n\n try:\n should_reload, diff = calculate_differences(\n self.config_file_path, active_config_file_path,\n logger=self.logger)\n except ConfigurationDiffError as e:\n self.logger(e)\n self.logger(\"Couldn't compute configuration differences. Aborting.\")\n return False\n\n self.copy_file(active_config_file_path, self.config_file_path)\n parsed_config = parse(open(self.config_file_path, 'r'))\n\n # Plugins\n self.__plugins__(parsed_config)\n\n if should_reload:\n self.logger(\"Will reload service\")\n self.reload_service()\n elif diff:\n hap = haproxy.HAProxy(socket_dir=get_socket_dir(parsed_config))\n set_state = lambda d, s: \\\n hap.server(d['name'], d['pool'])[0].setstate(s)\n\n self.logger(\"Will not reload service\")\n for server in diff.get('enabled'):\n\n set_state(server, haproxy.STATE_ENABLE)\n\n # Here is our hook to Pavlos' lib to enable a server.\n self.logger(\"Enabling server {} in pool {}\".format(\n server['name'], server['pool']))\n for server in diff.get('disabled'):\n\n set_state(server, haproxy.STATE_DISABLE)\n\n # Here is our hook to Pavlos' lib to disable a server.\n self.logger(\"Disabling server {} in pool {}\".format(\n server['name'], server['pool']))\n else:\n self.logger(\"No changes were found.\")\n\n return True\n", "sub_path": "balancer/agent/lbtype/haproxy/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 7647, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "tempfile.NamedTemporaryFile", "line_number": 35, "usage_type": "call"}, {"api_name": "contextlib.contextmanager", "line_number": 33, "usage_type": "name"}, {"api_name": "haproxy_parser.parse", "line_number": 45, "usage_type": "call"}, {"api_name": "haproxy_parser.parse", "line_number": 46, "usage_type": "call"}, {"api_name": "haproxy_parser.compare", "line_number": 48, "usage_type": "call"}, {"api_name": "haproxy_parser.SHOULD_PATCH", "line_number": 49, "usage_type": "name"}, {"api_name": "haproxy_parser.SHOULD_RELOAD", "line_number": 52, "usage_type": "name"}, {"api_name": "haproxy_parser.NO_CHANGES_FOUND", "line_number": 54, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 61, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 61, "usage_type": "attribute"}, {"api_name": "time.mktime", "line_number": 62, "usage_type": "call"}, {"api_name": "balancer.agent.Worker", "line_number": 66, "usage_type": "name"}, {"api_name": "functools.partial", "line_number": 69, "usage_type": "call"}, {"api_name": "haproxy_parser.get_frontends", "line_number": 75, "usage_type": "call"}, {"api_name": "haproxy_parser.parse", "line_number": 91, "usage_type": "call"}, {"api_name": "balancer.agent.exceptions.InitializationError", "line_number": 100, "usage_type": "call"}, {"api_name": "shutil.copyfile", "line_number": 160, "usage_type": "call"}, {"api_name": "balancer.agent.exceptions.ConfigurationDiffError", "line_number": 179, "usage_type": "name"}, {"api_name": "haproxy_parser.parse", "line_number": 185, "usage_type": "call"}, {"api_name": "haproxyadmin.haproxy.HAProxy", "line_number": 194, "usage_type": "call"}, {"api_name": "haproxyadmin.haproxy", "line_number": 194, "usage_type": "name"}, {"api_name": "haproxy_parser.get_socket_dir", "line_number": 194, "usage_type": "call"}, {"api_name": "haproxyadmin.haproxy.STATE_ENABLE", "line_number": 201, "usage_type": "attribute"}, {"api_name": "haproxyadmin.haproxy", "line_number": 201, "usage_type": "name"}, {"api_name": "haproxyadmin.haproxy.STATE_DISABLE", "line_number": 208, "usage_type": "attribute"}, {"api_name": "haproxyadmin.haproxy", "line_number": 208, "usage_type": "name"}]}
+{"seq_id": "53815061", "text": "from django.db import models\nfrom django.contrib import admin\nfrom .models import *\nfrom martor.widgets import AdminMartorWidget\n\nclass PostAdmin (admin.ModelAdmin):\n list_display = (\"title\", \"date_created\")\n prepopulated_fields = {\"slug\":(\"title\",)}\n formfield_overrides = {\n models.TextField: {'widget': AdminMartorWidget},\n }\n\nclass BlogPostAdmin (admin.ModelAdmin):\n list_display = (\"title\", \"date_created\")\n prepopulated_fields = {\"slug\":(\"title\",)}\n formfield_overrides = {\n models.TextField: {'widget': AdminMartorWidget},\n }\n\nclass TutorialPostAdmin (admin.ModelAdmin):\n list_display = (\"title\", \"date_created\")\n prepopulated_fields = {\"slug\":(\"title\",)}\n formfield_overrides = {\n models.TextField: {'widget': AdminMartorWidget},\n }\n\n\nclass ClassAdmin (admin.ModelAdmin):\n list_display = (\"title\", \"date_created\")\n prepopulated_fields = {\"slug\":(\"title\",)}\n\nadmin.site.register(PublishedPost, PostAdmin)\nadmin.site.register(BlogPost, BlogPostAdmin)\nadmin.site.register(TutorialPost, TutorialPostAdmin)\nadmin.site.register(TutorialClass, ClassAdmin)\n", "sub_path": "site/blog/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 1120, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.contrib.admin.ModelAdmin", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 6, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 10, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 10, "usage_type": "name"}, {"api_name": "martor.widgets.AdminMartorWidget", "line_number": 10, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 13, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 17, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "martor.widgets.AdminMartorWidget", "line_number": 17, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 20, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 24, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 24, "usage_type": "name"}, {"api_name": "martor.widgets.AdminMartorWidget", "line_number": 24, "usage_type": "name"}, {"api_name": "django.contrib.admin.ModelAdmin", "line_number": 28, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 28, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 32, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 32, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 32, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 33, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 33, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 33, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 34, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 34, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 34, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 35, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 35, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 35, "usage_type": "name"}]}
+{"seq_id": "337362036", "text": "#!/usr/bin/env python\n\n############################################\n# Scripts for plotting functions\n# Contains functions:\n# get_RMS - used to calculate RMS for plotting statistics\n# get_FWHM - FWHM calculation method, not currently plugged in\n# plot_history - scatter line plot of loss vs epochs\n# plot_distributions_CCNC - plot energy distribution for truth, and for NN reco\n# plot_resolutions_CCNC - plot energy resoltuion for (NN reco - truth)\n# plot_2D_prediction - 2D plot of True vs Reco\n# plot_single_resolution - Resolution histogram, (NN reco - true) and can compare (old reco - true)\n# plot_compare_resolution - Histograms of resolutions for systematic sets, overlaid\n# plot_systematic_slices - \"Scatter plot\" with systematic sets on x axis and 68% resolution on y axis\n# plot_energy_slices - Scatter plot energy cut vs resolution\n##############################################\n\nimport numpy\nimport h5py\nimport os, sys\nimport matplotlib.pyplot as plt\nimport matplotlib.colors as colors\nfrom scipy.interpolate import UnivariateSpline\nfrom scipy import interpolate\nimport scipy.stats\nimport itertools\n\nimport matplotlib \nmatplotlib.rc('xtick', labelsize=20) \nmatplotlib.rc('ytick', labelsize=20) \n\nfont = {'family' : 'normal',\n 'weight' : 'normal',\n 'size' : 22}\n\nmatplotlib.rc('font', **font)\n\ndef get_RMS(resolution,weights=None):\n if weights is not None:\n import wquantiles as wq\n\n mean_array = numpy.ones_like(resolution)*numpy.mean(resolution)\n if weights is None:\n rms = numpy.sqrt( sum((mean_array - resolution)**2)/len(resolution) )\n else:\n rms = numpy.zeros_like(resolution)\n rms = numpy.sqrt( sum(weights*(mean_array - resolution)**2)/sum(weights) )\n return rms\n\ndef get_FWHM(resolution,bins):\n x_range = numpy.linspace(min(resolution),max(resolution),bins)\n y_values,bin_edges = numpy.histogram(resolution,bins=bins)\n spline = UnivariateSpline(x_range,y_values - max(y_values)/2.)\n r = spline.roots()\n if len(r) != 2:\n print(\"Root are weird\")\n print(r)\n r1 = 0\n r2 = 0\n else:\n r1, r2 = spline.roots()\n return r1, r2\n\ndef find_contours_2D(x_values,y_values,xbins,weights=None,c1=16,c2=84): \n \"\"\"\n Find upper and lower contours and median\n x_values = array, input for hist2d for x axis (typically truth)\n y_values = array, input for hist2d for y axis (typically reconstruction)\n xbins = values for the starting edge of the x bins (output from hist2d)\n c1 = percentage for lower contour bound (16% - 84% means a 68% band, so c1 = 16)\n c2 = percentage for upper contour bound (16% - 84% means a 68% band, so c2=84)\n Returns:\n x = values for xbins, repeated for plotting (i.e. [0,0,1,1,2,2,...]\n y_median = values for y value medians per bin, repeated for plotting (i.e. [40,40,20,20,50,50,...]\n y_lower = values for y value lower limits per bin, repeated for plotting (i.e. [30,30,10,10,20,20,...]\n y_upper = values for y value upper limits per bin, repeated for plotting (i.e. [50,50,40,40,60,60,...]\n \"\"\"\n if weights is not None:\n import wquantiles as wq\n y_values = numpy.array(y_values)\n indices = numpy.digitize(x_values,xbins)\n r1_save = []\n r2_save = []\n median_save = []\n for i in range(1,len(xbins)):\n mask = indices==i\n if len(y_values[mask])>0:\n if weights is None:\n r1, m, r2 = numpy.percentile(y_values[mask],[c1,50,c2])\n else:\n r1 = wq.quantile(y_values[mask],weights[mask],c1/100.)\n r2 = wq.quantile(y_values[mask],weights[mask],c2/100.)\n m = wq.median(y_values[mask],weights[mask])\n else:\n #print(i,'empty bin')\n r1 = numpy.nan\n m = numpy.nan\n r2 = numpy.nan\n median_save.append(m)\n r1_save.append(r1)\n r2_save.append(r2)\n median = numpy.array(median_save)\n lower = numpy.array(r1_save)\n upper = numpy.array(r2_save)\n\n x = list(itertools.chain(*zip(xbins[:-1],xbins[1:])))\n y_median = list(itertools.chain(*zip(median,median)))\n y_lower = list(itertools.chain(*zip(lower,lower)))\n y_upper = list(itertools.chain(*zip(upper,upper)))\n \n return x, y_median, y_lower, y_upper\n\n\n\ndef plot_history(network_history,save=False,savefolder=None,use_logscale=False):\n \"\"\"\n Plot history of neural network's loss vs. epoch\n Recieves:\n network_history = array, saved metrics from neural network training\n save = optional, bool to save plot\n savefolder = optional, output folder to save to, if not in current dir\n Returns:\n line scatter plot of epoch vs loss\n \"\"\"\n plt.figure(figsize=(10,7))\n plt.xlabel('Epochs')\n plt.ylabel('Loss')\n if use_logscale:\n plt.yscale('log')\n plt.plot(network_history.history['loss'])\n plt.plot(network_history.history['val_loss'])\n plt.legend(['Training', 'Validation'])\n if save == True:\n plt.savefig(\"%sloss_vs_epochs.png\"%savefolder)\n\n plt.show()\n\ndef plot_history_from_list(loss,val,save=False,savefolder=None,logscale=False,ymin=None,ymax=None,title=None,variable=\"Energy\",pick_epoch=None,lr_start=None,lr_drop=None,lr_epoch=None,step=1):\n \n fig,ax = plt.subplots(figsize=(10,7))\n start=step\n end=len(loss)*step\n epochs = numpy.arange(start,end+step,step)\n ax.plot(epochs,loss,'b',label=\"Training\")\n ax.plot(epochs,val,'c',label=\"Validation\")\n \n #Edit Axis\n if logscale:\n ax.set_yscale('log')\n if ymin and ymax:\n pass\n elif ymin:\n ymax = max(max(loss),max(val))\n elif ymax:\n ymin = min(min(loss),min(val))\n else:\n ymax = max(max(loss),max(val))\n ymin = min(min(loss),min(val))\n ax.set_ylim(ymin,ymax)\n \n if pick_epoch is not None:\n ax.axvline(pick_epoch*step,linewidth=4, color='g',alpha=0.5,label=\"Chosen Model\")\n\n if lr_epoch is not None:\n epoch_drop = numpy.arange(0,end,lr_epoch*step)\n for lr_print in range(len(epoch_drop)):\n lrate = lr_start*(lr_drop**lr_print)\n ax.axvline(epoch_drop[lr_print],linewidth=1, color='r',linestyle=\"--\")\n ax.annotate(s='lrate='+str(\"{:.0e}\".format(lrate)),xy=(epoch_drop[lr_print]+step,ymax),rotation=90,verticalalignment='top')\n\n #Add labels\n if title:\n plt.title(title,fontsize=25)\n else:\n plt.title(\"Loss for %s CNN\"%variable,fontsize=25)\n \n plt.xlabel('Epochs',fontsize=20)\n if variable==\"Energy\":\n plt.ylabel(r'Loss = $\\frac{100}{n}\\sum_{i=1}^n \\vert \\frac{T_i - R_i}{T_i} \\vert$',fontsize=20)\n elif variable==\"Cosine Zenith\":\n plt.ylabel(r'Loss = $\\frac{1}{n}\\sum_{i=1}^n ( T_i - R_i )^2$',fontsize=20)\n else:\n plt.ylabel('Loss',fontsize=20)\n plt.legend(loc=\"center right\",fontsize=20)\n \n\n\n if save == True:\n plt.savefig(\"%sloss_vs_epochs.png\"%savefolder,bbox_inches='tight') \n plt.close()\n\n\ndef plot_history_from_list_split(energy_loss,val_energy_loss,zenith_loss,val_zenith_loss,save=True,savefolder=None,logscale=False,ymin=None,ymax=None,title=None):\n \n plt.figure(figsize=(10,7))\n plt.plot(energy_loss,'b',label=\"Energy Training\")\n plt.plot(val_energy_loss,'c',label=\"Energy Validation\")\n plt.plot(zenith_loss,'r',label=\"Zenith Training\")\n plt.plot(val_zenith_loss,'m',label=\"Zenith Validation\")\n \n #Edit Axis\n if logscale:\n plt.yscale('log')\n if ymin and ymax:\n plt.ylim(ymin,ymax)\n elif ymin:\n plt.ylim(ymin,max(max(loss),max(val)))\n elif ymax:\n plt.ylim(min(min(loss),min(val)),ymax)\n \n #Add labels\n if title:\n plt.title(title,fontsize=25)\n else:\n plt.title(\"Training and Validation Loss after %s Epochs\"%len(energy_loss),fontsize=25)\n plt.xlabel('Epochs',fontsize=20)\n plt.ylabel('Loss',fontsize=20)\n plt.legend(fontsize=20)\n \n if save == True:\n plt.savefig(\"%sloss_vs_epochs_split.png\"%savefolder)\n plt.close()\n\ndef plot_distributions_CCNC(truth_all_labels,truth,reco,save=False,savefolder=None):\n \"\"\"\n Plot testing set distribution, with CC and NC distinguished\n Recieves:\n truth_all_labels = array, Y_test truth labels that have ALL values in them (need CC vs NC info)\n truth = array, Y_test truth labels\n reco = array, neural network prediction output\n save = optional, bool to save plot\n savefolder = optional, output folder to save to, if not in current dir\n Returns:\n 1D histogram of reco - true with sepearated CC and NC distinction\n \"\"\"\n CC_mask = truth_all_labels[:,11] ==1\n NC_mask = truth_all_labels[:,11] ==0\n num_CC = sum(CC_mask)\n num_NC = sum(NC_mask)\n print(\"CC events: %i, NC events: %i, Percent NC: %.2f\"%(num_CC,num_NC,float(num_NC/(num_CC+num_NC))*100.))\n\n plt.figure(figsize=(10,7))\n plt.title(\"True Energy Distribution\",fontsize=25)\n plt.hist(truth[CC_mask], bins=100,color='b',alpha=0.5,label=\"CC\");\n plt.hist(truth[NC_mask], bins=100,color='g',alpha=0.5,label=\"NC\");\n plt.xlabel(\"Energy (GeV)\",fontsize=20)\n plt.legend(fontsize=10)\n if save:\n plt.savefig(\"%sTrueEnergyDistribution_CCNC.png\"%savefolder)\n\n plt.figure(figsize=(10,7))\n plt.title(\"NN Energy Distribution\",fontsize=25)\n plt.hist(reco[CC_mask], bins=100,color='b', alpha=0.5, label=\"CC\");\n plt.hist(reco[NC_mask], bins=100,color='g', alpha=0.5, label=\"NC\");\n plt.xlabel(\"Energy (GeV)\",fontsize=20)\n plt.legend(fontsize=10)\n if save:\n plt.savefig(\"%sNNEnergyDistribution_CCNC.png\"%savefolder)\n plt.close()\n\ndef plot_distributions(truth,reco=None,save=False,savefolder=None,old_reco=None,weights=None,variable=\"Energy\",units=\"(GeV)\",reco_name=\"Retro\", minval=None, maxval=None,bins=100,cnn_name=\"CNN\",ylog=False,xlog=False,old_reco_weights=None,title=None,xline=None,xline_label=None,flavor=None,sample=None):\n \"\"\"\n Plot testing set distribution\n Recieves:\n truth = array, Y_test truth labels\n reco = array, neural network prediction output\n save = optional, bool to save plot\n savefolder = optional, output folder to save to, if not in current dir\n variable = string, variable name\n units = string, units for variable\n Returns:\n 1D histogram of variable's absolute distribution for truth and for reco overlaid\n \"\"\"\n\n if maxval is None:\n if reco is not None:\n if old_reco is None:\n maxval = numpy.max([numpy.max(truth),numpy.max(reco)])\n else:\n maxval = numpy.max([numpy.max([numpy.max(truth),numpy.max(reco)]),numpy.max(old_reco)])\n else:\n maxval = numpy.max(truth)\n if minval is None:\n if reco is not None:\n if old_reco is None:\n minval = numpy.min([numpy.min(truth),numpy.min(reco)])\n else:\n minval = numpy.min([numpy.min([numpy.min(truth),numpy.min(reco)]),numpy.min(old_reco)])\n else:\n minval = numpy.min(truth)\n print(\"Using\", minval, maxval)\n \n plt.figure(figsize=(10,7))\n outname = \"\"\n if weights is not None:\n if old_reco_weights is None:\n old_reco_weights = weights\n #name += \"Weighted\"\n weights_factor = 1e7\n if title is not None:\n plt.title(\"%s\"%(title),fontsize=25)\n else:\n name = \"%s Distribution\"%(variable)\n if flavor is not None:\n if flavor == \"NuMu\" or flavor == \"numu\":\n name += r' for $\\nu_\\mu$ '\n elif flavor == \"NuE\" or flavor == \"nue\":\n name += r' for $\\nu_e$ '\n elif flavor == \"NuTau\" or flavor == \"nutau\":\n name += r' for $\\nu_\\tau$ '\n elif flavor == \"Mu\" or flavor == \"mu\":\n name += r' for $\\mu$ '\n elif flavor == \"Nu\" or flavor == \"nu\":\n name += r' for $\\nu$ '\n else:\n name += flavor\n name += sample\n plt.title(name,fontsize=25)\n\n if xlog:\n if minval <=0:\n print(\"MINVAL AT OR BELOW ZERO, USING ZERO FOR LOG SCALE\")\n logmin = 0\n else:\n logmin = numpy.log(minval)\n number_bins = bins\n bins = 10**numpy.linspace(logmin, numpy.log10(maxval),number_bins)\n plt.xscale('log')\n plt.hist(truth, bins=bins,color='g',alpha=0.5,range=[minval,maxval],weights=weights,label=\"Truth\");\n maskT = numpy.logical_and(truth > minval, truth < maxval)\n print(\"Truth Total: %i, Events in Plot: %i, Overflow: %i\"%(len(truth),sum(maskT),len(truth)-sum(maskT)))\n if weights is not None:\n print(\"WEIGHTED Truth Total: %.2f, Events in Plot: %.2f, Overflow: %.2f\"%(sum(weights)*weights_factor,sum(weights[maskT])*weights_factor,(sum(weights)-sum(weights[maskT]))*weights_factor))\n plt.ylabel(\"weighted event count\")\n else:\n plt.ylabel(\"event count\")\n outname += \"T\"\n \n if reco is not None:\n plt.hist(reco, bins=bins,color='b', alpha=0.5,range=[minval,maxval],weights=weights,label=cnn_name);\n outname += \"R\"\n maskR = numpy.logical_and(reco > minval, reco < maxval)\n print(\"Reco Total: %i, Events in Plot: %i, Overflow: %i\"%(len(reco),sum(maskR),len(reco)-sum(maskR)))\n if weights is not None:\n print(\"WEIGHTED Reco Total: %.2f, Events in Plot: %.2f, Overflow: %.2f\"%(sum(weights)*weights_factor,sum(weights[maskR])*weights_factor,(sum(weights)-sum(weights[maskR]))*weights_factor))\n if old_reco is not None:\n plt.hist(old_reco, bins=bins,color='orange', alpha=0.5,range=[minval,maxval],weights=old_reco_weights,label=reco_name);\n outname += \"OR\"\n maskOR = numpy.logical_and(old_reco > minval, old_reco < maxval)\n print(\"Old Reco Total: %i, Events in Plot: %i, Overflow: %i\"%(len(old_reco),sum(maskOR),len(old_reco)-sum(maskOR)))\n if weights is not None:\n print(\"WEIGHTED Old Reco Total: %.2f, Events in Plot: %.2f, Overflow: %.2f\"%(sum(old_reco_weights)*weights_factor,sum(old_reco_weights[maskOR])*weights_factor,(sum(old_reco_weights)-sum(old_reco_weights[maskOR]))*weights_factor))\n plt.xlabel(\"%s %s\"%(variable,units),fontsize=20)\n if ylog:\n plt.yscale(\"log\")\n if xline is not None:\n plt.axvline(xline,linewidth=3,color='k',linestyle=\"-\",label=\"%s\"%xline_label)\n if reco is not None or old_reco is not None or xline is not None:\n plt.legend(fontsize=20)\n\n if title is not None:\n outname += \"%s\"%title.replace(\" \", \"\")\n outname += \"%s\"%variable.replace(\" \",\"\")\n if flavor is not None:\n outname += \"%s\"%flavor.replace(\" \",\"\")\n if sample is not None:\n outname += \"%s\"%sample.replace(\" \",\"\")\n if save:\n plt.savefig(\"%s%sDistribution_%ito%i.png\"%(savefolder,outname,int(minval),int(maxval)),bbox_inches='tight')\n plt.close()\n\n\ndef plot_2D_prediction(truth, nn_reco, \\\n save=False,savefolder=None,weights=None,syst_set=\"\",\\\n bins=60,minval=None,maxval=None, switch_axis=False,\\\n cut_truth = False, axis_square =False,\n zmin = None, zmax=None,log=True,\n variable=\"Energy\", units = \"(GeV)\", epochs=None,\\\n flavor=\"NuMu\", sample=None,\\\n variable_type=\"True\", reco_name=\"CNN\",new_labels=None,\n new_units=None,save_name=None,no_contours=False,\n xline=None,yline=None):\n \"\"\"\n Plot testing set reconstruction vs truth\n Recieves:\n truth = array, Y_test truth\n nn_reco = array, neural network prediction output\n save = optional, bool to save plot\n savefolder = optional, output folder to save to, if not in current dir\n syst_set = string, name of the systematic set (for title and saving)\n bins = int, number of bins plot (will use for both the x and y direction)\n minval = float, minimum value to cut nn_reco results\n maxval = float, maximum value to cut nn_reco results\n cut_truth = bool, true if you want to make the value cut on truth rather than nn results\n axis_square = bool, cut axis to be square based on minval and maxval inputs\n variable = string, name of the variable you are plotting\n units = string, units for the variable you are plotting\n Returns:\n 2D plot of True vs Reco\n \"\"\"\n\n maxplotline = min([max(nn_reco),max(truth)])\n minplotline = max([min(nn_reco),min(truth)])\n \n truth = truth #[mask]\n nn_reco = nn_reco #[mask]\n \n #Cut axis\n if axis_square:\n xmin = minval\n ymin = minval\n xmax = maxval\n ymax = maxval\n else:\n xmin = min(truth)\n ymin = min(nn_reco)\n xmax = max(truth)\n ymax = max(nn_reco)\n if switch_axis:\n xmin, ymin = ymin, xmin\n xmax, ymax = ymax, xmax\n\n\n if weights is None:\n cmin = 1\n else:\n cmin = 1e-12\n if zmin is not None:\n cmin = zmin\n \n plt.figure(figsize=(10,7))\n if log:\n if switch_axis:\n cts,xbin,ybin,img = plt.hist2d(nn_reco, truth, bins=bins,range=[[xmin,xmax],[ymin,ymax]], cmap='viridis_r', norm=colors.LogNorm(), weights=weights, cmax=zmax, cmin=cmin)\n else:\n cts,xbin,ybin,img = plt.hist2d(truth, nn_reco, bins=bins,range=[[xmin,xmax],[ymin,ymax]],cmap='viridis_r', norm=colors.LogNorm(), weights=weights, cmax=zmax, cmin=cmin)\n else:\n if switch_axis:\n cts,xbin,ybin,img = plt.hist2d(nn_reco, truth, bins=bins,range=[[xmin,xmax],[ymin,ymax]],cmap='viridis_r', weights=weights, cmax=zmax, cmin=cmin)\n else:\n cts,xbin,ybin,img = plt.hist2d(truth, nn_reco, bins=bins,range=[[xmin,xmax],[ymin,ymax]],cmap='viridis_r', weights=weights, cmax=zmax, cmin=cmin)\n cbar = plt.colorbar()\n if weights is None:\n cbar.ax.set_ylabel('counts', rotation=90)\n else:\n cbar.ax.set_ylabel('Rate (Hz)', rotation=90)\n plt.xlabel(\"%s %s %s\"%(variable_type,variable,units),fontsize=20)\n plt.ylabel(\"%s Reconstructed %s %s\"%(reco_name,variable,units),fontsize=20)\n if switch_axis:\n plt.ylabel(\"%s %s %s\"%(variable_type,variable,units),fontsize=20)\n plt.xlabel(\"%s Reconstructed %s %s\"%(reco_name,variable,units),fontsize=20)\n if new_labels is not None:\n plt.ylabel(\"%s %s\"%(new_labels[0],new_units[0]),fontsize=20)\n plt.xlabel(\"%s %s\"%(new_labels[1],new_units[1]),fontsize=20)\n \n #NAMING\n title = \"%s vs %s for %s %s\"%(reco_name,variable_type,variable,syst_set)\n if flavor == \"NuMu\" or flavor == \"numu\":\n title += r' for $\\nu_\\mu$ '\n elif flavor == \"NuE\" or flavor == \"nue\":\n title += r' for $\\nu_e$ '\n elif flavor == \"NuTau\" or flavor == \"nutau\":\n title += r' for $\\nu_\\tau$ '\n elif flavor == \"Mu\" or flavor == \"mu\":\n title += r' for $\\mu$ '\n elif flavor == \"Nu\" or flavor == \"nu\":\n title += r' for $\\nu$ '\n else:\n title += flavor\n if sample is not None:\n title += sample\n #if weights is not None:\n # title += \" Weighted\"\n if epochs:\n title += \" at %i Epochs\"%epochs\n plt.suptitle(title,fontsize=25)\n #if cutting:\n # plt.title(\"%s, plotted %i, overflow %i\"%(name,len(truth),overflow),fontsize=20)\n \n #Plot 1:1 line\n if axis_square:\n plt.plot([minval,maxval],[minval,maxval],'k:',label=\"1:1\")\n else:\n plt.plot([minplotline,maxplotline],[minplotline,maxplotline],'k:',label=\"1:1\")\n \n if switch_axis:\n x, y, y_l, y_u = find_contours_2D(nn_reco,truth,xbin,weights=weights)\n else:\n x, y, y_l, y_u = find_contours_2D(truth,nn_reco,xbin,weights=weights)\n\n if not no_contours:\n plt.plot(x,y,color='r',label='Median')\n plt.plot(x,y_l,color='r',label='68% band',linestyle='dashed')\n plt.plot(x,y_u,color='r',linestyle='dashed')\n plt.legend(fontsize=20)\n if yline is not None:\n if type(yline) is list:\n for y_val in yline:\n plt.axhline(y_val,linewidth=3,color='red',label=\"Cut\")\n else:\n plt.axhline(yline,linewidth=3,color='red',label=\"Cut\")\n plt.legend(fontsize=20)\n if xline is not None:\n if type(xline) is list:\n for x_val in xline:\n plt.axvline(x_val,linewidth=3,color='magenta',linestyle=\"dashed\",label=\"Cut\")\n else:\n plt.axvline(xline,linewidth=3,color='magenta',linestyle=\"dashed\",label=\"Cut\")\n plt.legend(fontsize=20)\n\n reco_name = reco_name.replace(\" \",\"\")\n variable = variable.replace(\" \",\"\")\n variable_type = variable_type.replace(\" \",\"\")\n nocut_name = \"\"\n if weights is not None:\n nocut_name=\"Weighted\"\n if flavor is not None:\n nocut_name += \"%s\"%flavor.replace(\" \",\"\")\n if sample is not None:\n nocut_name += \"%s\"%sample\n if not axis_square:\n nocut_name +=\"_nolim\"\n if zmax:\n nocut_name += \"_zmax%.1e\"%zmax \n if zmin:\n nocut_name += \"_zmin%.1e\"%zmin \n if switch_axis:\n nocut_name +=\"_SwitchedAxis\"\n if save_name is not None:\n nocut_name += \"%s\"%save_name.replace(\" \",\"\")\n if save:\n plt.savefig(\"%s%s%sReco%s_2DHist%s%s.png\"%(savefolder,variable_type,reco_name,variable,syst_set,nocut_name),bbox_inches='tight')\n plt.close()\n\ndef plot_2D_prediction_fraction(truth, nn_reco, weights=None,\\\n save=False,savefolder=None,syst_set=\"\",\\\n bins=60,xminval=None,xmaxval=None,\\\n yminval=None,ymaxval=None,log=True,zmax=None,\\\n variable=\"Energy\", units = \"(GeV)\",reco_name=\"CNN\"):\n \"\"\"\n Plot testing set reconstruction vs truth\n Recieves:\n truth = array, Y_test truth\n nn_reco = array, neural network prediction output\n save = optional, bool to save plot\n savefolder = optional, output folder to save to, if not in current dir\n syst_set = string, name of the systematic set (for title and saving)\n bins = int, number of bins plot (will use for both the x and y direction)\n minval = float, minimum value to cut (truth - nn_reco)/truth fractional results\n maxval = float, maximum value to cut (truth - nn_reco)/truth fractional results\n variable = string, name of the variable you are plotting\n units = string, units for the variable you are plotting\n Returns:\n 2D plot of True vs (True - Reco)/True\n \"\"\"\n \n fractional_error = abs(truth - nn_reco)/ truth\n \n nolim=False\n if xminval is None:\n xminval = min(truth)\n if xmaxval is None:\n nolim = True\n xmaxval = max(truth)\n if yminval is None:\n yminval = min(fractional_error)\n if ymaxval is None:\n ymaxval = max(fractional_error)\n \n if weights is None:\n cmin = 1\n else:\n cmin = 1e-12\n \n plt.figure(figsize=(10,7))\n\n if log:\n cts,xbin,ybin,img = plt.hist2d(truth, fractional_error, bins=bins,range=[[xminval,xmaxval],[yminval,ymaxval]],cmap='viridis_r', norm=colors.LogNorm(), weights=weights, cmax=zmax, cmin=cmin)\n else:\n cts,xbin,ybin,img = plt.hist2d(truth, fractional_error, bins=bins,range=[[xminval,xmaxval],[yminval,ymaxval]],cmap='viridis_r', weights=weights, cmax=zmax, cmin=cmin)\n cbar = plt.colorbar()\n cbar.ax.set_ylabel('counts', rotation=90)\n \n plt.xlabel(\"True Neutrino %s %s\"%(variable,units),fontsize=20)\n plt.ylabel(r'Fractional Resolution: $\\frac{reconstruction - truth}{truth}$',fontsize=20)\n plt.title(\"%s Fractional Error vs. True %s %s\"%(reco_name,variable,syst_set),fontsize=25)\n \n x, y, y_l, y_u = find_contours_2D(truth,fractional_error,xbin,weights=weights)\n plt.plot(x,y,color='r',label='Median')\n plt.plot(x,y_l,color='r',label='68% band',linestyle='dashed')\n plt.plot(x,y_u,color='r',linestyle='dashed')\n plt.legend(fontsize=12)\n \n nocut_name = \"\"\n variable = variable.replace(\" \",\"\")\n if weights is not None:\n nocut_name=\"Weighted\"\n if nolim:\n nocut_name =\"_nolim\"\n if save:\n plt.savefig(\"%sTruth%sRecoFrac%s_2DHist%s%s.png\"%(savefolder,reco_name,variable,syst_set,nocut_name),bbox_inches='tight')\n\ndef plot_resolution_CCNC(truth_all_labels,truth,reco,save=False,savefolder=None,variable=\"Energy\", units = \"(GeV)\"):\n \"\"\"\n Plot testing set resolution of reconstruction - truth, with CC and NC distinguished\n Recieves:\n truth_all_labels = array, Y_test truth labels that have ALL values in them (need CC vs NC info)\n truth = array, Y_test truth labels\n reco = array, neural network prediction output\n save = optional, bool to save plot\n savefolder = optional, output folder to save to, if not in current dir\n Returns:\n 1D histogram of reco - true with sepearated CC and NC distinction\n \"\"\"\n CC_mask = truth_all_labels[:,11] ==1\n NC_mask = truth_all_labels[:,11] ==0\n num_CC = sum(CC_mask)\n num_NC = sum(NC_mask)\n print(\"CC events: %i, NC events: %i, Percent NC: %.2f\"%(num_CC,num_NC,float(num_NC/(num_CC+num_NC))*100.))\n\n resolution = reco - truth\n resolution_fraction = (reco - truth)/truth\n resolution = numpy.array(resolution)\n resolution_fraction = numpy.array(resolution_fraction)\n\n plt.figure(figsize=(10,7)) \n plt.title(\"%s Resolution\"%variable)\n plt.hist(resolution[CC_mask], bins=50,color='b', alpha=0.5, label=\"CC\");\n plt.hist(resolution[NC_mask], bins=50,color='g', alpha=0.5, label=\"NC\");\n plt.xlabel(\"NN reconstruction - truth (%s)\"%units)\n plt.legend()\n if save:\n plt.savefig(\"%s%sResolution_CCNC.png\"%(savefolder,variable))\n\n plt.figure(figsize=(10,7)) \n plt.title(\"Fractional %s Resolution\"%variable)\n plt.hist(resolution_fraction[CC_mask], bins=50,color='b', alpha=0.5, label=\"CC\");\n plt.hist(resolution_fraction[NC_mask], bins=50,color='g', alpha=0.5, label=\"NC\");\n plt.xlabel(\"(NN reconstruction - truth) / truth\")\n plt.legend()\n\n variable = variable.replace(\" \",\"\")\n if save:\n plt.savefig(\"%s%sResolutionFrac_CCNC.png\"%(savefolder,variable))\n plt.close()\n\ndef plot_single_resolution(truth,nn_reco,weights=None, \\\n bins=100, use_fraction=False,\\\n use_old_reco = False, old_reco=None,\\\n old_reco_truth=None,old_reco_weights=None,\\\n mintrue=None,maxtrue=None,\\\n minaxis=None,maxaxis=None,\\\n save=False,savefolder=None,\n flavor=\"NuMu\", sample=None,\n variable=\"Energy\", units = \"GeV\", epochs=None,\n reco_name=\"CNN\", old_reco_name=\"Retro\"):\n \"\"\"Plots resolution for dict of inputs, one of which will be a second reco\n Recieves:\n truth = array of truth or Y_test labels\n nn_reco = array of NN predicted reco or Y_test_predicted results\n bins = int value\n use_fraction = use fractional resolution instead of absolute, where (reco - truth)/truth\n use_old_reco = True if you want to compare to another reconstruction (like pegleg)\n old_reco = optional, pegleg array of labels\n mintrue = float, min true value if cut desired\n maxtrue = float, max true value if cut desired\n minaxis = float, min x axis cut\n maxaxis = float, max x axis cut\n Returns:\n 1D histogram of Reco - True (or fractional)\n Can have two distributions of NN Reco Resolution vs Pegleg Reco Resolution\n \"\"\"\n weights_reco = old_reco_weights\n if weights is not None:\n import wquantiles as wq\n if old_reco_weights is None:\n weights_reco = numpy.array(weights)\n\n fig, ax = plt.subplots(figsize=(10,7))\n\n ## Assume old_reco truth is the same as test sample, option to set it otherwise\n if old_reco_truth is None:\n truth2 = truth\n else:\n truth2 = old_reco_truth\n # NAN CUT FOR OLD RECO\n #if old_reco is not None:\n # not_nan = numpy.logical_not(numpy.isnan(old_reco))\n # if sum(not_nan) != len(not_nan):\n # print(\"CUTTING NAN VALUES FROM OLD RECO\")\n # old_reco = old_reco[not_nan]\n # truth2 = truth2[not_nan]\n # weights_reco = weights_reco[not_nan]\n #Check nan\n if old_reco is not None:\n is_nan = numpy.isnan(old_reco)\n assert sum(is_nan) == 0, \"OLD RECO HAS NAN\"\n is_nan = numpy.isnan(nn_reco)\n assert sum(is_nan) == 0, \"CNN RECO HAS NAN\"\n \n if use_fraction:\n nn_resolution = (nn_reco - truth)/truth\n if use_old_reco:\n old_reco_resolution = (old_reco - truth2)/truth2\n title = \"Fractional %s Resolution\"%variable\n xlabel = r'$\\frac{reconstruction - truth}{truth}$'\n else:\n nn_resolution = nn_reco - truth\n if use_old_reco:\n old_reco_resolution = old_reco - truth2\n title = \"%s Resolution\"%variable\n xlabel = \"reconstruction - truth %s\"%units\n if epochs:\n title += \" at %i Epochs\"%epochs\n if flavor == \"NuMu\" or flavor == \"numu\":\n title += r' for $\\nu_\\mu$ ' \n elif flavor == \"NuE\" or flavor == \"nue\":\n title += r' for $\\nu_e$ '\n elif flavor == \"NuTau\" or flavor == \"nutau\":\n title += r' for $\\nu_\\tau$ '\n elif flavor == \"Nu\" or flavor == \"nu\":\n title += r' for $\\nu$ '\n elif flavor == \"Mu\" or flavor == \"mu\":\n title += r' for $\\mu$ '\n else:\n title += flavor\n if sample is not None:\n title += sample\n\n #if weights is not None:\n # title += \" Weighted\"\n original_size = len(nn_resolution)\n \n \n #Get stats before axis cut!\n rms_nn = get_RMS(nn_resolution,weights)\n if weights is not None:\n r1 = wq.quantile(nn_resolution,weights,0.16)\n r2 = wq.quantile(nn_resolution,weights,0.84)\n median = wq.median(nn_resolution,weights)\n else:\n r1, r2 = numpy.percentile(nn_resolution, [16,84])\n median = numpy.median(nn_resolution)\n if use_old_reco:\n true_cut_size_reco=len(old_reco_resolution)\n rms_old_reco = get_RMS(old_reco_resolution,weights_reco)\n if weights is not None:\n r1_old_reco = wq.quantile(old_reco_resolution,weights_reco,0.16)\n r2_old_reco = wq.quantile(old_reco_resolution,weights_reco,0.84)\n median_old_reco = wq.median(old_reco_resolution,weights_reco)\n else:\n r1_old_reco, r2_old_reco = numpy.percentile(old_reco_resolution, [16,84])\n median_old_reco = numpy.median(old_reco_resolution)\n\n\n # Find cut for plot axis\n #print(minaxis,maxaxis)\n axis_cut = False\n if minaxis or maxaxis:\n axis_cut = True\n if not minaxis:\n minaxis = min(nn_resolution)\n if use_old_reco:\n if minaxis > min(old_reco_resolution):\n minaxis = min(old_reco_resolution)\n if not maxaxis:\n maxaxis = max(nn_resolution)\n if use_old_reco:\n if maxaxis < max(old_reco_resolution):\n maxaxis = max(old_reco_resolution)\n \n\n hist_nn, bins, p = ax.hist(nn_resolution, bins=bins, range=[minaxis,maxaxis], weights=weights, alpha=0.5, label=reco_name);\n weights_factor = 1 #1e7\n total_events = len(nn_resolution) #sum(weights)\n outside_range = numpy.logical_or(nn_resolution < minaxis, nn_resolution > maxaxis)\n overflow = sum(outside_range) #sum(weights[outside_range])\n\n #Statistics\n #weighted_avg_and_std(nn_resolution,weights)\n \n textstr = '\\n'.join((\n r'%s' % (reco_name),\n r'$\\mathrm{events}=%.0f$' % (total_events*weights_factor, ),\n r'$\\mathrm{median}=%.2f$' % (median, ),\n r'$\\mathrm{overflow}=%.0f$' % (overflow*weights_factor, ),\n #r'$\\mathrm{RMS}=%.2f$' % (rms_nn, ),\n r'$\\mathrm{1\\sigma}=%.2f,%.2f$' % (r1,r2 )))\n props = dict(boxstyle='round', facecolor='blue', alpha=0.3)\n ax.text(0.6, 0.95, textstr, transform=ax.transAxes, fontsize=20,\n verticalalignment='top', bbox=props)\n\n if use_old_reco:\n ax.hist(old_reco_resolution, bins=bins, range=[minaxis,maxaxis], weights=weights_reco, alpha=0.5, label=\"%s\"%old_reco_name);\n total_events_reco = len(old_reco_resolution) #sum(weights_reco) #len(weights_reco)\n outside_range_reco = numpy.logical_or(old_reco_resolution < minaxis, old_reco_resolution > maxaxis)\n overflow_reco = sum(outside_range_reco) #sum(weights_reco[outside_range_reco]) #sum(outside_range_reco)\n ax.legend(loc=\"upper left\",fontsize=20)\n textstr = '\\n'.join((\n '%s' % (old_reco_name),\n r'$\\mathrm{events}=%.0f$' % (total_events_reco*weights_factor, ),\n r'$\\mathrm{median}=%.2f$' % (median_old_reco, ),\n r'$\\mathrm{overflow}=%.0f$' % (overflow_reco*weights_factor, ),\n #r'$\\mathrm{RMS}=%.2f$' % (rms_old_reco, ),\n r'$\\mathrm{1\\sigma}=%.2f,%.2f$' % (r1_old_reco,r2_old_reco )))\n props = dict(boxstyle='round', facecolor='wheat', alpha=0.5)\n ax.text(0.6, 0.55, textstr, transform=ax.transAxes, fontsize=20,\n verticalalignment='top', bbox=props)\n\n #if axis_cut:\n ax.set_xlim(minaxis,maxaxis)\n ax.set_xlabel(xlabel,fontsize=20)\n ax.set_title(title,fontsize=25)\n\n old_reco_name = old_reco_name.replace(\" \",\"\")\n variable = variable.replace(\" \",\"\")\n savename = \"%sResolution\"%variable\n if weights is not None:\n savename+=\"Weighted\"\n if use_fraction:\n savename += \"Frac\"\n if use_old_reco:\n savename += \"_Compare%sReco\"%reco_name\n if axis_cut:\n savename += \"_xlim\"\n if maxtrue or mintrue:\n savename += \"_truthcut\"\n if save == True:\n plt.savefig(\"%s%s.png\"%(savefolder,savename),bbox_inches='tight')\n plt.close()\n\ndef plot_compare_resolution(truth,nn_reco,namelist, weights_dict=None, savefolder=None,\\\n save=False,bins=100,use_fraction=False, mask_dict=None,mask_index=None,\n minval=None,maxval=None,reco_name=\"CNN\",variable=\"Energy\",units=\"(GeV)\"):\n \"\"\"Plots resolution for dict of inputs\n Receives:\n truth = dict of truth or Y_test labels\n (contents = [key name, energy], shape = [number syst sets, number of events])\n nn_reco = dict of NN predicted or Y_test_predicted results\n (contents = [key name, energy], shape = [number syst sets, number of events])\n weights = dict of weights\n namelist = list of names for the dict, to use as pretty labels\n save_folder_name = string for output file\n save = bool where True saves and False does not save plot\n bins = int value\n use_fraction: bool, uses fractional resolution if True\n Returns:\n Histograms of resolutions for systematic sets, overlaid\n Prints statistics for all histograms into table\n \"\"\"\n\n if weights_dict is not None:\n import wquantiles as wq\n\n \n print(\"Resolution\")\n print('Name\\t Events\\t Overflow\\t Median\\t RMS\\t Percentiles\\t')\n plt.figure(figsize=(10,7)) \n if use_fraction:\n title = \"%s Fractional %s Resolution\"%(reco_name,variable)\n xlabel = \"(reconstruction - truth) / truth\"\n else:\n title = \"%s %s Resolution\"%(reco_name, variable)\n xlabel = \"reconstruction - truth %s\"%units\n \n find_minval = True\n find_maxval = True\n if minval is not None:\n find_minval = False\n if maxval is not None:\n find_maxval = False\n \n resolution = {} \n for index in range(0,len(namelist)):\n keyname = namelist[index]\n #if mask_dict is None:\n # mask = numpy.ones(len(truth[keyname]),dtype=bool)\n #else:\n # mask = mask_dict[keyname][mask_index]\n if use_fraction:\n resolution[keyname] = (nn_reco[keyname] - truth[keyname]) / truth[keyname]\n else:\n resolution[keyname] = nn_reco[keyname] - truth[keyname]\n \n if find_minval:\n if index == 0:\n minval = min(resolution[keyname])\n if minval > min(resolution[keyname]):\n minval = min(resolution[keyname])\n if find_maxval:\n if index == 0:\n maxval = max(resolution[keyname])\n if maxval < max(resolution[keyname]):\n maxval = max(resolution[keyname])\n \n #Get Statistics & Plot\n for index in range(0,len(namelist)):\n keyname = namelist[index]\n #if mask_dict is None:\n # mask = numpy.ones(len(truth[keyname]),dtype=bool)\n #else:\n # mask = mask_dict[keyname][mask_index]\n \n rms = get_RMS(resolution[keyname],weights_dict[keyname])\n if weights_dict is not None:\n r1 = wq.quantile(resolution[keyname],weights_dict[keyname],0.16)\n r2 = wq.quantile(resolution[keyname],weights_dict[keyname],0.84)\n median = wq.median(resolution[keyname],weights_dict[keyname])\n else:\n r1, r2 = numpy.percentile(resolution[keyname], [16,84])\n median = numpy.median(resolution[keyname])\n\n total_events = len(resolution[keyname]) #sum(weights_dict[keyname][mask])\n outside_range = numpy.logical_or(resolution[keyname] < minval, resolution[keyname] > maxval)\n #masked_weights=weights_dict[keyname]\n overflow = sum(outside_range) #sum(masked_weights[outside_range])\n weights_factor = 1 #1e7\n\n plt.hist(resolution[keyname], range=[minval,maxval],bins=bins, alpha=0.5, label=\"%s\"%namelist[index]);\n \n print(\"%s\\t %.0f\\t %.0f\\t %.2f\\t %.2f\\t %.2f, %.2f\\t\"%(namelist[index], \\\n total_events*weights_factor,\\\n overflow*weights_factor,\\\n median,\\\n rms,\\\n r1, r2))\n plt.title(title) \n plt.xlabel(xlabel)\n if use_fraction:\n plt.legend(fontsize=20)\n\n else:\n plt.legend(loc=9, bbox_to_anchor=(0.5, -0.1), ncol=2)\n \n basename=\"\"\n if weights_dict is not None:\n basename += \"Weighted\"\n if use_fraction:\n basename += \"Frac\"\n basename += \"%sResolution_CompareSets%s\"%(variable,reco_name)\n if save:\n plt.savefig(\"%s%s.png\"%(savefolder,basename))\n plt.close()\n\ndef plot_systematic_slices(truth_dict, nn_reco_dict, namelist,\n weights_dict = None, use_fraction=False,\n mask_dict=None, mask_index=None,title=None,\\\n use_old_reco = False, old_reco_dict=None,\n old_reco_weights_dict = None, old_reco_truth_dict = None,\\\n save=False,savefolder=None,cnn_name=\"CNN\",old_reco_name=\"Retro\"):\n \"\"\"Plots different arrays vs each other (systematic set arrays)\n Receives:\n truth_dict = dict of arrays with truth labels\n (contents = [key name, energy], shape = [number syst sets, number of events])\n nn_reco_dict = dict of arrays that has NN predicted reco results\n (contents = [key name, energy], shape = [number syst sets, number of events])\n namelist = list of names to be used for x_axis ticks\n use_fraction = use fractional resolution instead of absolute, where (reco - truth)/truth\n use_reco = True if you want to compare to another reconstruction (like pegleg)\n (contents = [key name, energy], shape = [number syst sets, number of events])\n old_reco = optional, dict of pegleg arrays with the labels\n Returns:\n \"scatter plot\" with systematic sets on x axis,\n y axis has median of resolution with error bars containing 68% of resolution\n \"\"\"\n if weights_dict is not None:\n if old_reco_weights_dict is None:\n old_reco_weights_dict = weights_dict\n if old_reco_dict is not None:\n if old_reco_truth_dict is None:\n old_reco_truth_dict = truth_dict\n \n number_sets = len(namelist)\n percentile_in_peak = 68.27\n\n left_tail_percentile = (100.-percentile_in_peak)/2\n right_tail_percentile = 100.-left_tail_percentile\n \n medians = numpy.zeros(number_sets)\n err_from = numpy.zeros(number_sets)\n err_to = numpy.zeros(number_sets)\n \n if use_old_reco:\n medians_reco = numpy.zeros(number_sets)\n err_from_reco = numpy.zeros(number_sets)\n err_to_reco = numpy.zeros(number_sets)\n \n resolution = {}\n for index in range(0,number_sets):\n keyname = namelist[index]\n #if mask_dict is None:\n # mask = numpy.ones(len(truth_dict[keyname]),dtype=bool)\n #else:\n # mask = mask_dict[keyname][mask_index]\n if use_fraction:\n resolution = (nn_reco_dict[keyname] - truth_dict[keyname])/truth_dict[keyname]\n else:\n resolution = (nn_reco_dict[keyname] - truth_dict[keyname])\n \n if weights_dict is not None:\n import wquantiles as wq\n lower_lim = wq.quantile(resolution,weights_dict[keyname],0.16)\n upper_lim = wq.quantile(resolution,weights_dict[keyname],0.84)\n median = wq.median(resolution,weights_dict[keyname])\n else:\n lower_lim = numpy.percentile(resolution, left_tail_percentile)\n upper_lim = numpy.percentile(resolution, right_tail_percentile)\n median = numpy.percentile(resolution, 50.)\n \n medians[index] = median\n err_from[index] = lower_lim\n err_to[index] = upper_lim\n \n if use_old_reco:\n if use_fraction:\n resolution_old_reco = ((old_reco_dict[keyname]-old_reco_truth_dict[keyname])/old_reco_truth_dict[keyname])\n else:\n resolution_old_reco = (old_reco_dict[keyname]-old_reco_truth_dict[keyname])\n \n if weights_dict is not None:\n lower_lim_reco = wq.quantile(resolution_old_reco,old_reco_weights_dict[keyname],0.16)\n upper_lim_reco = wq.quantile(resolution_old_reco,old_reco_weights_dict[keyname],0.84)\n median_reco = wq.median(resolution_old_reco,old_reco_weights_dict[keyname])\n else:\n lower_lim_reco = numpy.percentile(resolution_old_reco, left_tail_percentile)\n upper_lim_reco = numpy.percentile(resolution_old_reco, right_tail_percentile)\n median_reco = numpy.percentile(resolution_old_reco, 50.)\n \n medians_reco[index] = median_reco\n err_from_reco[index] = lower_lim_reco\n err_to_reco[index] = upper_lim_reco\n\n\n x_range = numpy.linspace(1,number_sets,number_sets)\n \n fig, ax = plt.subplots(figsize=(10,7))\n plt.errorbar(x_range, medians, yerr=[medians-err_from, err_to-medians], capsize=5.0, fmt='o',label=\"%s\"%cnn_name)\n if use_old_reco:\n plt.errorbar(x_range, medians_reco, yerr=[medians_reco-err_from_reco, err_to_reco-medians_reco], capsize=5.0,fmt='o',label=\"%s\"%old_reco_name)\n plt.legend(loc=\"upper right\")\n ax.plot([0,number_sets+1], [0,0], color='k')\n ax.set_xlim(0,number_sets+1)\n \n #rename axis\n my_xlabels = [item.get_text() for item in ax.get_xticklabels()]\n new_namelist = [\" \"] + namelist\n for index in range(0,number_sets+1):\n my_xlabels[index] = new_namelist[index]\n ax.set_xticklabels(my_xlabels)\n \n ax.set_xlabel(\"Systematic Set\")\n if use_fraction:\n ax.set_ylabel(r'Fractional Resolution: $\\frac{reconstruction - truth}{truth}$')\n else:\n ax.set_ylabel(\"Resolution: \\n reconstruction - truth (GeV)\")\n ax.set_title(\"%s\"%title)\n \n savename = \"\"\n if weights_dict is not None:\n savename += \"Weighted\"\n if use_fraction:\n savename += \"Frac\"\n if use_old_reco:\n savename += \"_CompareOldReco\"\n savename += \"SystematicResolutionCompare\"\n if save == True:\n plt.savefig(\"%s%s.png\"%(savefolder,savename))\n plt.close()\n\ndef plot_bin_slices(truth, nn_reco, energy_truth=None, weights=None,\\\n use_fraction = False, old_reco=None,old_reco_truth=None,\\\n reco_energy_truth=None,old_reco_weights=None,\\\n bins=10,min_val=0.,max_val=60., ylim = None,\\\n save=False,savefolder=None,vs_predict=False,\\\n flavor=\"NuMu\", sample=\"CC\",style=\"contours\",\\\n variable=\"Energy\",units=\"(GeV)\",xlog=False,\n xvariable=\"Energy\",xunits=\"(GeV)\",\\\n specific_bins = None,xline=None,xline_name=\"DeepCore\",\n epochs=None,reco_name=\"Retro\",cnn_name=\"CNN\"):\n \"\"\"Plots different variable slices vs each other (systematic set arrays)\n Receives:\n truth= array with truth labels for this one variable\n nn_reco = array that has NN predicted reco results for one variable (same size of truth)\n energy_truth = optional (will use if given), array that has true energy information (same size of truth)\n use_fraction = bool, use fractional resolution instead of absolute, where (reco - truth)/truth\n old_reco = optional (will use if given), array of pegleg labels for one variable\n bins = integer number of data points you want (range/bins = width)\n min_val = minimum value for variable to start cut at (default = 0.)\n max_val = maximum value for variable to end cut at (default = 60.)\n ylim = List with two entries of ymin and ymax for plot [min, max], leave as None for no ylim applied\n style= contours or errorbars\n Returns:\n Scatter plot with energy values on x axis (median of bin width)\n y axis has median of resolution with error bars containing 68% of resolution\n \"\"\"\n reco_weights = old_reco_weights\n if weights is not None:\n import wquantiles as wq\n if reco_weights is None:\n reco_weights = numpy.array(weights)\n\n nn_reco = numpy.array(nn_reco)\n truth = numpy.array(truth)\n ## Assume old_reco truth is the same as test sample, option to set it otherwise\n if old_reco_truth is None:\n truth2 = numpy.array(truth)\n else:\n truth2 = numpy.array(old_reco_truth)\n if reco_energy_truth is None:\n energy_truth2 = numpy.array(energy_truth)\n else:\n energy_truth2 = numpy.array(reco_energy_truth)\n #Check nan\n if old_reco is not None:\n is_nan = numpy.isnan(old_reco)\n assert sum(is_nan) == 0, \"OLD RECO HAS NAN\"\n is_nan = numpy.isnan(nn_reco)\n assert sum(is_nan) == 0, \"CNN RECO HAS NAN\"\n \n # NAN CUT FOR OLD RECO\n #if old_reco is not None:\n # not_nan = numpy.logical_not(numpy.isnan(old_reco))\n # if sum(not_nan) != len(not_nan):\n # print(\"CUTTING NAN VALUES FROM OLD RECO\")\n # old_reco = old_reco[not_nan]\n # truth2 = truth2[not_nan]\n # if reco_weights is not None:\n # reco_weights = reco_weights[not_nan]\n # if energy_truth is not None:\n # energy_truth2 = energy_truth2[not_nan]\n\n\n if use_fraction:\n resolution = ((nn_reco-truth)/truth) # in fraction\n else:\n resolution = (nn_reco-truth)\n resolution = numpy.array(resolution)\n percentile_in_peak = 68.27\n\n left_tail_percentile = (100.-percentile_in_peak)/2\n right_tail_percentile = 100.-left_tail_percentile\n\n if specific_bins is None:\n variable_ranges = numpy.linspace(min_val,max_val, num=bins+1)\n variable_centers = (variable_ranges[1:] + variable_ranges[:-1])/2.\n else:\n max_val = specific_bins[-1]\n min_val = specific_bins[0]\n variable_ranges = specific_bins\n variable_centers = []\n for i in range(len(specific_bins)-1):\n variable_centers.append(specific_bins[i] + ((specific_bins[i+1] - specific_bins[i])/2.))\n\n medians = numpy.zeros(len(variable_centers))\n err_from = numpy.zeros(len(variable_centers))\n err_to = numpy.zeros(len(variable_centers))\n\n if old_reco is not None:\n if use_fraction:\n resolution_reco = ((old_reco-truth2)/truth2)\n else:\n resolution_reco = (old_reco-truth2)\n resolution_reco = numpy.array(resolution_reco)\n medians_reco = numpy.zeros(len(variable_centers))\n err_from_reco = numpy.zeros(len(variable_centers))\n err_to_reco = numpy.zeros(len(variable_centers))\n\n for i in range(len(variable_ranges)-1):\n var_from = variable_ranges[i]\n var_to = variable_ranges[i+1]\n \n if vs_predict:\n x_axis_array = nn_reco\n x_axis_array2 = old_reco #nn_reco\n title=\"%s Resolution Dependence\"%(variable)\n else:\n if energy_truth is None:\n title=\"%s Resolution Dependence\"%(variable)\n x_axis_array = truth\n x_axis_array2 = truth2\n else:\n title=\"%s Resolution %s Dependence\"%(variable,xvariable)\n energy_truth = numpy.array(energy_truth)\n x_axis_array = energy_truth\n x_axis_array2 = energy_truth2\n \n cut = (x_axis_array >= var_from) & (x_axis_array < var_to)\n #print(\"Events in \", var_from, \" to \", var_to, sum(cut))\n if old_reco is not None:\n cut2 = (x_axis_array2 >= var_from) & (x_axis_array2 < var_to)\n #print(\"Events in \", var_from, \" to \", var_to, sum(cut2))\n\n if weights is not None:\n lower_lim = wq.quantile(resolution[cut],weights[cut],left_tail_percentile/100.)\n upper_lim = wq.quantile(resolution[cut],weights[cut], right_tail_percentile/100.)\n median = wq.median(resolution[cut],weights[cut])\n else:\n lower_lim = numpy.percentile(resolution[cut], left_tail_percentile/100.)\n upper_lim = numpy.percentile(resolution[cut], right_tail_percentile/100.)\n median = numpy.percentile(resolution[cut], 0.50)\n\n medians[i] = median\n err_from[i] = lower_lim\n err_to[i] = upper_lim\n\n if old_reco is not None:\n if reco_weights is not None:\n lower_lim_reco = wq.quantile(resolution_reco[cut2],reco_weights[cut2],left_tail_percentile/100.)\n upper_lim_reco = wq.quantile(resolution_reco[cut2],reco_weights[cut2],right_tail_percentile/100.)\n median_reco = wq.median(resolution_reco[cut2],reco_weights[cut2])\n else:\n lower_lim_reco = numpy.percentile(resolution_reco[cut2], left_tail_percentile/100.)\n upper_lim_reco = numpy.percentile(resolution_reco[cut2], right_tail_percentile/100.)\n median_reco = numpy.percentile(resolution_reco[cut2], 0.50)\n\n medians_reco[i] = median_reco\n err_from_reco[i] = lower_lim_reco\n err_to_reco[i] = upper_lim_reco\n\n plt.figure(figsize=(10,7))\n plt.plot([min_val,max_val], [0,0], color='k')\n if style == \"errorbars\":\n (_, caps, _) = plt.errorbar(variable_centers, medians, yerr=[medians-err_from, err_to-medians], xerr=[ variable_centers-variable_ranges[:-1], variable_ranges[1:]-variable_centers ], capsize=3.0, fmt='o',label=cnn_name)\n for cap in caps:\n cap.set_markeredgewidth(5)\n if old_reco is not None:\n (_, caps_reco, _) = plt.errorbar(variable_centers, medians_reco, yerr=[medians_reco-err_from_reco, err_to_reco-medians_reco], xerr=[ variable_centers-variable_ranges[:-1], variable_ranges[1:]-variable_centers ], capsize=3.0, fmt='o',label=\"%s\"%reco_name)\n for cap in caps_reco:\n cap.set_markeredgewidth(5)\n plt.legend(loc=\"upper center\")\n if vs_predict:\n plt.legend(loc=\"lower center\")\n if xline is not None:\n if type(xline) is list:\n for x_val in xline:\n plt.axvline(x_val,linewidth=3,color='k',linestyle=\"dashed\",label=\"%s\"%xline_name)\n else:\n plt.axvline(xline,linewidth=3,color='k',linestyle=\"dashed\",label=\"%s\"%xline_name)\n \n plt.legend(loc=\"upper center\")\n if vs_predict:\n plt.legend(loc=\"lower center\")\n else: #countours\n alpha=0.5\n lwid=3\n cmap = plt.get_cmap('Blues')\n colors = cmap(numpy.linspace(0, 1, 2 + 2))[2:]\n color=colors[0]\n cmap = plt.get_cmap('Oranges')\n rcolors = cmap(numpy.linspace(0, 1, 2 + 2))[2:]\n rcolor=rcolors[0] \n ax = plt.gca()\n ax.plot(variable_centers, medians,linestyle='-',label=\"%s median\"%(cnn_name), color=color, linewidth=lwid)\n ax.fill_between(variable_centers,medians, err_from,color=color, alpha=alpha)\n ax.fill_between(variable_centers,medians, err_to, color=color, alpha=alpha,label=cnn_name + ' 68%')\n if old_reco is not None:\n ax.plot(variable_centers,medians_reco, color=rcolor, linestyle='-', label=\"%s median\"%reco_name, linewidth=lwid)\n ax.fill_between(variable_centers,medians_reco,err_from_reco, color=rcolor, alpha=alpha)\n ax.fill_between(variable_centers,medians_reco,err_to_reco, color=rcolor,alpha=alpha,label=reco_name + ' 68%')\n if xline is not None:\n if type(xline) is list:\n for x_val in xline:\n plt.axvline(x_val,linewidth=3,color='k',linestyle=\"dashed\",label=\"%xline_name\"%s)\n else:\n plt.axvline(xline,linewidth=3,color='k',linestyle=\"dashed\",label=\"%xline_name\"%s)\n if vs_predict:\n plt.legend(loc=\"lower center\")\n else:\n plt.legend(loc=\"upper center\")\n plt.xlim(min_val,max_val)\n if ylim is not None:\n plt.ylim(ylim)\n if vs_predict:\n plt.xlabel(\"Reconstructed %s %s\"%(variable,units),fontsize=20)\n elif energy_truth is not None:\n plt.xlabel(\"%s %s\"%(xvariable,units),fontsize=20)\n else:\n plt.xlabel(\"True %s %s\"%(variable,units),fontsize=20)\n if use_fraction:\n plt.ylabel(r'Fractional Resolution: $\\frac{reconstruction - truth}{truth}$',fontsize=20)\n else:\n plt.ylabel(\"Resolution: \\n reconstruction - truth %s\"%units,fontsize=20)\n if xlog:\n plt.xscale('log')\n\n #if epochs:\n # title += \" at %i Epochs\"%epochs\n if flavor == \"NuMu\" or flavor == \"numu\":\n title += r' for $\\nu_\\mu$ ' \n elif flavor == \"NuE\" or flavor == \"nue\":\n title += r' for $\\nu_e$ '\n elif flavor == \"NuTau\" or flavor == \"nutau\":\n title += r' for $\\nu_\\tau$ '\n elif flavor == \"Mu\" or flavor == \"mu\":\n title += r' for $\\mu$ '\n elif flavor == \"Nu\" or flavor == \"nu\":\n title += r' for $\\nu$ '\n else:\n title += flavor\n title += sample\n plt.title(title,fontsize=25)\n\n reco_name = reco_name.replace(\" \",\"\")\n variable = variable.replace(\" \",\"\")\n savename = \"%s%sResolutionSlices\"%(variable,cnn_name)\n if vs_predict:\n savename +=\"VsPredict\"\n if use_fraction:\n savename += \"Frac\"\n if weights is not None:\n savename += \"Weighted\"\n if flavor is not None:\n savename += \"%s\"%flavor.replace(\" \",\"\")\n if energy_truth is not None:\n xvar_no_space = xvariable.replace(\" \",\"\")\n savename += \"_%sBinned\"%xvar_no_space\n if style == \"errorbars\":\n savename += \"ErrorBars\"\n if xlog:\n savename +=\"_xlog\"\n if old_reco is not None:\n savename += \"_Compare%sReco\"%reco_name\n if ylim is not None:\n savename += \"_ylim\"\n if save == True:\n plt.savefig(\"%s%s.png\"%(savefolder,savename),bbox_inches='tight')\n plt.close()\n\ndef plot_rms_slices(truth, nn_reco, energy_truth=None, use_fraction = False, \\\n old_reco=None,old_reco_truth=None, reco_energy_truth=None,\\\n bins=10,min_val=0.,max_val=60., ylim = None,weights=None,\\\n old_reco_weights=None,save=False,savefolder=None,\\\n variable=\"Energy\",units=\"(GeV)\",epochs=None,reco_name=\"Retro\",\n flavor=\"NuMu\", sample=\"CC\"):\n \"\"\"Plots different variable slices vs each other (systematic set arrays)\n Receives:\n truth= array with truth labels for this one variable\n nn_reco = array that has NN predicted reco results for one variable (same size of truth)\n energy_truth = optional (will use if given), array that has true energy information (same size of truth)\n use_fraction = bool, use fractional resolution instead of absolute, where (reco - truth)/truth\n old_reco = optional (will use if given), array of pegleg labels for one variable\n bins = integer number of data points you want (range/bins = width)\n min_val = minimum value for variable to start cut at (default = 0.)\n max_val = maximum value for variable to end cut at (default = 60.)\n ylim = List with two entries of ymin and ymax for plot [min, max], leave as None for no ylim applied\n Returns:\n Scatter plot with energy values on x axis (median of bin width)\n y axis has median of resolution with error bars containing 68% of resolution\n \"\"\"\n nn_reco = numpy.array(nn_reco)\n truth = numpy.array(truth)\n ## Assume old_reco truth is the same as test sample, option to set it otherwise\n if old_reco_truth is None:\n truth2 = numpy.array(truth)\n else:\n truth2 = numpy.array(old_reco_truth)\n if reco_energy_truth is None:\n energy_truth2 = numpy.array(energy_truth)\n else:\n energy_truth2 = numpy.array(reco_energy_truth)\n if weights is not None:\n if old_reco_weights is None:\n old_reco_weights = weights\n\n if use_fraction:\n resolution = ((nn_reco-truth)/truth) # in fraction\n else:\n resolution = (nn_reco-truth)\n resolution = numpy.array(resolution)\n\n variable_ranges = numpy.linspace(min_val,max_val, num=bins+1)\n variable_centers = (variable_ranges[1:] + variable_ranges[:-1])/2.\n\n rms_all = numpy.zeros(len(variable_centers))\n\n if old_reco is not None:\n if use_fraction:\n resolution_reco = ((old_reco-truth2)/truth2)\n else:\n resolution_reco = (old_reco-truth2)\n resolution_reco = numpy.array(resolution_reco)\n rms_reco_all = numpy.zeros(len(variable_centers))\n\n for i in range(len(variable_ranges)-1):\n var_from = variable_ranges[i]\n var_to = variable_ranges[i+1]\n \n title=\"\"\n #else:\n # title=\"Weighted \"\n if energy_truth is None:\n title+=\"%s RMS Resolution Dependence\"%(variable)\n cut = (truth >= var_from) & (truth < var_to)\n cut2 = (truth2 >= var_from) & (truth2 < var_to)\n else:\n #print(\"Using energy for x-axis. Make sure your min_val and max_val are in terms of energy!\")\n title+=\"%s RMS Resolution Energy Dependence\"%(variable)\n energy_truth = numpy.array(energy_truth)\n cut = (energy_truth >= var_from) & (energy_truth < var_to)\n cut2 = (energy_truth2 >= var_from) & (energy_truth2 < var_to)\n\n if weights is not None:\n weight_here = weights[cut]\n reco_weight_here = old_reco_weights[cut2]\n else:\n weight_here = None\n rms = get_RMS(resolution[cut],weight_here)\n rms_all[i] = rms\n \n if old_reco is not None:\n rms_reco = get_RMS(resolution_reco[cut2],reco_weight_here)\n rms_reco_all[i] = rms_reco\n\n #cnn_name = \"CNN\"\n cnn_name = \"Neural Network\"\n diff_width=abs(variable_ranges[1] - variable_ranges[0])\n plt.figure(figsize=(10,7))\n\n if old_reco is not None:\n rms_reco_all = numpy.append(rms_reco_all, rms_reco_all[-1])\n plt.step(variable_ranges, rms_reco_all, where='post', color=\"orange\",label=\"%s\"%reco_name)\n rms_all = numpy.append(rms_all,rms_all[-1])\n plt.step(variable_ranges, rms_all, where='post', color=\"blue\",label=cnn_name)\n plt.legend(fontsize=15)\n \n \n plt.ylim(bottom=0)\n plt.xlim(min_val,max_val)\n if type(ylim) is not None:\n plt.ylim(ylim)\n plt.xlabel(\"True %s %s\"%(variable,units),fontsize=20)\n if use_fraction:\n if weights is not None:\n plt.ylabel(r'Weighted RMS of Fractional Resoltion: $\\frac{reconstruction - truth}{truth}$',fontsize=20)\n else:\n plt.ylabel(r'RMS of Fractional Resoltion: $\\frac{reconstruction - truth}{truth}$',fontsize=20)\n else:\n if weights is not None:\n plt.ylabel(\"Weighted RMS of Resolution: \\n reconstruction - truth %s\"%units,fontsize=20)\n else:\n plt.ylabel(\"RMS of Resolution: \\n reconstruction - truth %s\"%units,fontsize=20)\n #if epochs:\n # title += \" at %i Epochs\"%epochs\n if flavor == \"NuMu\" or flavor == \"numu\":\n title += r' for $\\nu_\\mu$ ' \n elif flavor == \"NuE\" or flavor == \"nue\":\n title += r' for $\\nu_e$ '\n else:\n title += flavor\n title += sample\n plt.title(title,fontsize=25)\n \n #print(rms_all,rms_reco_all)\n \n reco_name = reco_name.replace(\" \",\"\")\n variable = variable.replace(\" \",\"\")\n savename = \"%sRMSSlices\"%variable\n if use_fraction:\n savename += \"Frac\"\n if weights is not None:\n savename += \"Weighted\"\n if energy_truth is not None:\n savename += \"_EnergyBinned\"\n plt.xlabel(\"True Energy (GeV)\",fontsize=20)\n if old_reco is not None:\n savename += \"_Compare%sReco\"%reco_name\n if type(ylim) is not None:\n savename += \"_ylim\"\n if save == True:\n plt.savefig(\"%s%s.png\"%(savefolder,savename),bbox_inches='tight')\n plt.close()\n\ndef imshow_plot(array,name,emin,emax,tmin,tmax,zlabel,savename):\n \n afig = plt.figure(figsize=(10,7))\n plt.imshow(array,origin='lower',extent=[emin,emax,tmin,tmax],aspect='auto')\n cbar = plt.colorbar()\n cbar.set_label(zlabel,rotation=90,fontsize=20)\n plt.set_cmap('viridis_r')\n cbar.ax.tick_params(labelsize=20) \n plt.xlabel(\"True Neutrino Energy (GeV)\",fontsize=20)\n plt.ylabel(\"True Track Length (m)\",fontsize=20)\n plt.title(\"%s for Track Length vs. Energy\"%name,fontsize=25)\n plt.xticks(fontsize=20)\n plt.yticks(fontsize=20)\n plt.savefig(savename,bbox_inches='tight')\n return afig\n \ndef plot_length_energy(truth, nn_reco, track_index=2,tfactor=200.,\\\n save=False,savefolder=None,use_fraction=False,\\\n ebins=10,tbins=10,emin=5.,emax=100.,tmin=0.,tmax=430.,\\\n cut_truth = False, axis_square =False, zmax=None,\n variable=\"Energy\", units = \"(GeV)\", epochs=None,reco_name=\"CNN\"):\n \n\n true_energy = truth[:,0]*emax\n true_track = truth[:,track_index]*tfactor\n #nn_reco = nn_reco[:,0]*emax\n \n #print(true_energy.shape,nn_reco.shape)\n if use_fraction:\n resolution = (nn_reco - true_energy)/true_energy\n title = \"Fractional %s Resolution\"%variable\n zlabel = \"(reco - truth) / truth\" \n else:\n resolution = nn_reco - true_energy\n title = \"%s Resolution\"%variable\n zlabel = \"reconstruction - truth (GeV)\"\n #print(nn_reco[:10],true_energy[:10]) \n \n percentile_in_peak = 68.27\n left_tail_percentile = (100.-percentile_in_peak)/2\n right_tail_percentile = 100.-left_tail_percentile\n \n \n energy_ranges = numpy.linspace(emin,emax, num=ebins+1)\n energy_centers = (energy_ranges[1:] + energy_ranges[:-1])/2.\n track_ranges = numpy.linspace(tmin,tmax, num=tbins+1)\n track_centers = (track_ranges[1:] + track_ranges[:-1])/2.\n\n medians = numpy.zeros((len(energy_centers),len(track_centers)))\n err_from = numpy.zeros((len(energy_centers),len(track_centers)))\n err_to = numpy.zeros((len(energy_centers),len(track_centers)))\n rms = numpy.zeros((len(energy_centers),len(track_centers)))\n \n #print(energy_ranges,track_ranges)\n for e in range(len(energy_ranges)-1):\n e_from = energy_ranges[e]\n e_to = energy_ranges[e+1]\n for t in range(len(track_ranges)-1):\n t_from = track_ranges[t]\n t_to = track_ranges[t+1]\n \n \n e_cut = (true_energy >= e_from) & (true_energy < e_to)\n t_cut = (true_track >= t_from) & (true_track < t_to)\n cut = e_cut & t_cut\n\n subset = resolution[cut]\n #print(subset)\n #print(e_from,e_to,t_from,t_to,true_energy[cut],true_track[cut])\n if sum(cut)==0:\n lower_lim = numpy.nan\n upper_lim = numpy.nan\n median = numpy.nan\n one_rms = numpy.nan\n else:\n lower_lim = numpy.percentile(subset, left_tail_percentile)\n upper_lim = numpy.percentile(subset, right_tail_percentile)\n median = numpy.percentile(subset, 50.)\n mean_array = numpy.ones_like(subset)*numpy.mean(subset)\n one_rms = numpy.sqrt( sum((mean_array - subset)**2)/len(subset) )\n #Invert saving because imshow does (M,N) where M is rows and N is columns\n medians[t,e] = median\n err_from[t,e] = lower_lim\n err_to[t,e] = upper_lim\n rms[t,e] = one_rms\n \n stat=[\"Median\", \"Lower 1 sigma\", \"Upper 1 sigma\", \"RMS\"]\n z_name = [zlabel, \"lower 1 sigma of \" + zlabel, \"upper 1 sigma of \" + zlabel, \"RMS of \" + zlabel ]\n \n savename = \"%sTrueEnergyTrackReco%s_2DHist_%s.png\"%(savefolder,reco_name,stat[0])\n imshow_plot(medians,stat[0],emin,emax,tmin,tmax,z_name[0],savename)\n \n savename=\"%sTrueEnergyTrackReco%s_2DHist_%s.png\"%(savefolder,reco_name,\"LowSigma\")\n imshow_plot(err_from,stat[1],emin,emax,tmin,tmax,z_name[1],savename)\n \n savename=\"%sTrueEnergyTrackReco%s_2DHist_%s.png\"%(savefolder,reco_name,\"HighSigma\")\n imshow_plot(err_to,stat[2],emin,emax,tmin,tmax,z_name[2],savename)\n \n savename=\"%sTrueEnergyTrackReco%s_2DHist_%s.png\"%(savefolder,reco_name,stat[3])\n imshow_plot(rms,stat[3],emin,emax,tmin,tmax,z_name[3],savename)\n", "sub_path": "data_check/PlottingFunctions.py", "file_name": "PlottingFunctions.py", "file_ext": "py", "file_size_in_byte": 66535, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "matplotlib.rc", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.ones_like", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.histogram", "line_number": 52, "usage_type": "call"}, {"api_name": "scipy.interpolate.UnivariateSpline", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.digitize", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 89, "usage_type": "call"}, {"api_name": "wquantiles.quantile", "line_number": 91, "usage_type": "call"}, {"api_name": "wquantiles.quantile", "line_number": 92, "usage_type": "call"}, {"api_name": "wquantiles.median", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 96, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 97, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 98, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 104, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 106, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 107, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 108, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 125, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 125, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yscale", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 173, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 173, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 175, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 175, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 177, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 177, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 179, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 179, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 181, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 181, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 184, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 189, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 190, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 190, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 195, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 195, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 196, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 196, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 197, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 197, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 198, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 198, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 199, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 199, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yscale", "line_number": 203, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 203, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 205, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 205, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 207, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 209, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 209, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 213, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 213, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 215, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 215, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 216, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 216, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 217, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 217, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 218, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 218, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 221, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 221, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 222, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 222, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 242, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 242, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 243, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 243, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 244, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 244, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 245, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 245, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 246, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 246, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 247, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 247, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 249, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 249, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 251, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 251, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 252, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 252, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 253, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 253, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 254, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 254, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 255, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 255, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 256, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 256, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 258, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 258, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 259, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 259, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 278, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 280, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 282, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 286, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 288, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 290, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 293, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 293, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 301, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 301, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 318, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 318, "usage_type": "name"}, {"api_name": "numpy.log", "line_number": 325, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 327, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 327, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xscale", "line_number": 328, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 328, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 329, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 329, "usage_type": "name"}, {"api_name": "numpy.logical_and", "line_number": 330, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 334, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 334, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 336, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 336, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 340, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 340, "usage_type": "name"}, {"api_name": "numpy.logical_and", "line_number": 342, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 347, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 347, "usage_type": "name"}, {"api_name": "numpy.logical_and", "line_number": 349, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 353, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 353, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yscale", "line_number": 355, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 355, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 357, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 357, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 359, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 359, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 369, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 369, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 370, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 370, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 431, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 431, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist2d", "line_number": 434, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 434, "usage_type": "name"}, {"api_name": "matplotlib.colors.LogNorm", "line_number": 434, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 434, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist2d", "line_number": 436, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 436, "usage_type": "name"}, {"api_name": "matplotlib.colors.LogNorm", "line_number": 436, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 436, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist2d", "line_number": 439, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 439, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist2d", "line_number": 441, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 441, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 442, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 442, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 447, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 447, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 448, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 448, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 450, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 450, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 451, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 451, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 453, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 453, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 454, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 454, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.suptitle", "line_number": 476, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 476, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 482, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 482, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 484, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 484, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 492, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 492, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 493, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 493, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 494, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 494, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 495, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 495, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axhline", "line_number": 499, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 499, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axhline", "line_number": 501, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 501, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 502, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 502, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 506, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 506, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 508, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 508, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 509, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 509, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 532, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 532, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 533, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 533, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 575, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 575, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist2d", "line_number": 578, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 578, "usage_type": "name"}, {"api_name": "matplotlib.colors.LogNorm", "line_number": 578, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 578, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist2d", "line_number": 580, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 580, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 581, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 581, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 584, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 584, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 585, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 585, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 586, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 586, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 589, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 589, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 590, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 590, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 591, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 591, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 592, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 592, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 601, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 601, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 623, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 624, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 626, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 626, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 627, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 627, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 628, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 628, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 629, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 629, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 630, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 630, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 631, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 631, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 633, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 633, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 635, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 635, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 636, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 636, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 637, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 637, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 638, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 638, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 639, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 639, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 640, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 640, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 644, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 644, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 645, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 645, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 677, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 679, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 679, "usage_type": "name"}, {"api_name": "numpy.isnan", "line_number": 696, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 698, "usage_type": "call"}, {"api_name": "wquantiles.quantile", "line_number": 738, "usage_type": "call"}, {"api_name": "wquantiles.quantile", "line_number": 739, "usage_type": "call"}, {"api_name": "wquantiles.median", "line_number": 740, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 742, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 743, "usage_type": "call"}, {"api_name": "wquantiles.quantile", "line_number": 748, "usage_type": "call"}, {"api_name": "wquantiles.quantile", "line_number": 749, "usage_type": "call"}, {"api_name": "wquantiles.median", "line_number": 750, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 752, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 753, "usage_type": "call"}, {"api_name": "numpy.logical_or", "line_number": 776, "usage_type": "call"}, {"api_name": "numpy.logical_or", "line_number": 796, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 829, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 829, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 830, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 830, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 858, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 858, "usage_type": "name"}, {"api_name": "wquantiles.quantile", "line_number": 906, "usage_type": "call"}, {"api_name": "wquantiles.quantile", "line_number": 907, "usage_type": "call"}, {"api_name": "wquantiles.median", "line_number": 908, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 910, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 911, "usage_type": "call"}, {"api_name": "numpy.logical_or", "line_number": 914, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.hist", "line_number": 919, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 919, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 927, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 927, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 928, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 928, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 930, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 930, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 933, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 933, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 942, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 942, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 943, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 943, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 979, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 980, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 981, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 984, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 985, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 986, "usage_type": "call"}, {"api_name": "wquantiles.quantile", "line_number": 1002, "usage_type": "call"}, {"api_name": "wquantiles.quantile", "line_number": 1003, "usage_type": "call"}, {"api_name": "wquantiles.median", "line_number": 1004, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 1006, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 1007, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 1008, "usage_type": "call"}, {"api_name": "wquantiles.quantile", "line_number": 1021, "usage_type": "call"}, {"api_name": "wquantiles.quantile", "line_number": 1022, "usage_type": "call"}, {"api_name": "wquantiles.median", "line_number": 1023, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 1025, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 1026, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 1027, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 1034, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 1036, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1036, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 1037, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1037, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 1039, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1039, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 1040, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1040, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1067, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1067, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 1068, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1068, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 1100, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1102, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1103, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1106, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1108, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1110, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1112, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 1115, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 1117, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1137, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 1144, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1154, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1155, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1156, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1163, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1164, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1165, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1166, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1183, "usage_type": "call"}, {"api_name": "wquantiles.quantile", "line_number": 1194, "usage_type": "call"}, {"api_name": "wquantiles.quantile", "line_number": 1195, "usage_type": "call"}, {"api_name": "wquantiles.median", "line_number": 1196, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 1198, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 1199, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 1200, "usage_type": "call"}, {"api_name": "wquantiles.quantile", "line_number": 1208, "usage_type": "call"}, {"api_name": "wquantiles.quantile", "line_number": 1209, "usage_type": "call"}, {"api_name": "wquantiles.median", "line_number": 1210, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 1212, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 1213, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 1214, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1220, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1220, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 1221, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1221, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 1223, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1223, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 1227, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1227, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 1230, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1230, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 1232, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1232, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 1236, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1236, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 1238, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1238, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 1240, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1240, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 1242, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1242, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.get_cmap", "line_number": 1246, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1246, "usage_type": "name"}, {"api_name": "matplotlib.colors", "line_number": 1247, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 1247, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 1248, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.get_cmap", "line_number": 1249, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1249, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 1250, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 1252, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1252, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 1263, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1263, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 1265, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1265, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 1267, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1267, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 1269, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1269, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 1270, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1270, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 1272, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1272, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 1274, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1274, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 1276, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1276, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 1278, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1278, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 1280, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1280, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 1282, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1282, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xscale", "line_number": 1284, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1284, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1301, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1301, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1326, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1326, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 1327, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1327, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 1350, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1351, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1354, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1356, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1358, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1360, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1369, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 1371, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1374, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1381, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1382, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1398, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1417, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1417, "usage_type": "name"}, {"api_name": "numpy.append", "line_number": 1420, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.step", "line_number": 1421, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1421, "usage_type": "name"}, {"api_name": "numpy.append", "line_number": 1422, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.step", "line_number": 1423, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1423, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 1424, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1424, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 1427, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1427, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 1428, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1428, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 1430, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1430, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 1431, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1431, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 1434, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1434, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 1436, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1436, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 1439, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1439, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 1441, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1441, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1451, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1451, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 1464, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1464, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1470, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1470, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 1471, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1471, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 1475, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1475, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 1476, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1476, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 1477, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1477, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.set_cmap", "line_number": 1479, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1479, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 1481, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1481, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 1482, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1482, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 1483, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1483, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 1484, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1484, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 1485, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1485, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 1486, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1486, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 1516, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 1518, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1521, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1522, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1523, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1524, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 1543, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 1544, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 1545, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 1546, "usage_type": "attribute"}, {"api_name": "numpy.percentile", "line_number": 1548, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 1549, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 1550, "usage_type": "call"}, {"api_name": "numpy.ones_like", "line_number": 1551, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 1551, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 1552, "usage_type": "call"}]}
+{"seq_id": "468295051", "text": "\"\"\" OpenCursor: context object for sqlite3 \"\"\"\n\nimport sqlite3\n\nDBNAME = \"game.db\"\n\ndef setDB(dbname):\n \"\"\" opencursor.setDB() sets the default DBNAME for OpenCursor objects \"\"\"\n global DBNAME\n DBNAME = dbname\n\n\nclass OpenCursor:\n \"\"\" Context object for sqlite3 \"\"\"\n\n def __init__(self, db=None, *args, **kwargs):\n \"\"\" db, args, kwargs passed to sqlite3.connect \"\"\"\n if db is None:\n db = DBNAME\n\n # by default, set check_same_thread to False (optimization for single-\n # threaded applications)\n kwargs['check_same_thread'] = kwargs.get('check_same_thread', False)\n\n self.conn = sqlite3.connect(db, *args, **kwargs)\n\n # setting row_factory to sqlite3.Row makes fetch statements return\n # objects that can be indexed by column name\n self.conn.row_factory = sqlite3.Row # access fetch results by col name\n\n self.cursor = self.conn.cursor()\n\n def __enter__(self):\n \"\"\" with OpenCursor as x returns self.cursor \"\"\"\n return self.cursor\n\n def __exit__(self, extype, exvalue, extraceback):\n \"\"\" commit changes upon exiting a with block if no errors raised \"\"\"\n if not extype:\n self.conn.commit()\n self.cursor.close()\n self.conn.close()\n", "sub_path": "FlaskApp/run/src/models/opencursor.py", "file_name": "opencursor.py", "file_ext": "py", "file_size_in_byte": 1284, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sqlite3.connect", "line_number": 25, "usage_type": "call"}, {"api_name": "sqlite3.Row", "line_number": 29, "usage_type": "attribute"}]}
+{"seq_id": "154034355", "text": "from scraper import main_scraper\r\nimport sqlite3\r\nfrom sqlite3 import Error\r\nimport json\r\nimport csv\r\nimport os\r\n\r\ndb_file = 'scraping.db'\r\n\r\n#Data Base Functions\r\ndef create_connection(db_file):\r\n \"\"\" create a database connection to the SQLite database\r\n specified by db_file\r\n :param db_file: database file\r\n :return: Connection object or None\r\n \"\"\"\r\n conn = None\r\n try:\r\n conn = sqlite3.connect(db_file)\r\n return conn\r\n except Error as e:\r\n print(e)\r\n\r\n return conn\r\n\r\n\r\ndef adding_task(task):\r\n conn=create_connection(db_file)\r\n\r\n if conn is not None:\r\n sql = ''' INSERT INTO tasks(task_id,task_name,task_role,task_urls) VALUES (?,?,?,?) '''\r\n\r\n try:\r\n cur = conn.cursor()\r\n cur.execute(sql,task)\r\n except Exception as e:\r\n print('Error in inserting task -->',e)\r\n else:\r\n print('Error in database (create task)')\r\n\r\n task_id = cur.lastrowid\r\n conn.commit()\r\n conn.close()\r\n return task_id\r\n\r\n\r\n##### Fetching tasks\r\ndef get_task():\r\n conn=create_connection(db_file)\r\n cur = conn.cursor()\r\n #cur.execute(\"SELECT player_name FROM players WHERE team_id=?\", (team_id,))\r\n\r\n cur.execute(\"SELECT * FROM tasks where task_role == 'scrape'\")\r\n\r\n\r\n rows = cur.fetchall()\r\n #print(rows)\r\n #print(len(rows))\r\n\r\n return rows\r\n\r\n#Data Base function to get Tasks to be monitored\r\ndef get_monitor_tasks():\r\n\r\n\r\n conn=create_connection(db_file)\r\n cur = conn.cursor()\r\n #cur.execute(\"SELECT player_name FROM players WHERE team_id=?\", (team_id,))\r\n\r\n cur.execute(\"SELECT * FROM tasks where task_role = 'monitor'\")\r\n\r\n rows = cur.fetchall()\r\n #print(rows)\r\n #print(len(rows))\r\n\r\n return rows\r\n\r\n\r\n# Enter fetch data into data base .\r\ndef enter_scraped_data(data):\r\n\r\n conn=create_connection(db_file)\r\n\r\n if conn is not None:\r\n\r\n if len(get_one_task_data(data[3])) == 0:\r\n\r\n sql = ''' INSERT INTO data_scraped(data_scraped_id,data_scraped_name,data_scraped_data,task_id) VALUES (?,?,?,?) '''\r\n\r\n try:\r\n cur = conn.cursor()\r\n cur.execute(sql,data)\r\n except Exception as e:\r\n print('Error in inserting Scraped Data --',e)\r\n\r\n else:\r\n\r\n sql = ''' UPDATE data_scraped\r\n SET data_scraped_data = ? WHERE task_id = ? '''\r\n\r\n try:\r\n d = (data[2],data[3]) #Updating data in data base ...\r\n cur = conn.cursor()\r\n cur.execute(sql,d)\r\n except Exception as e:\r\n print('Error in Updating Scraped Data --',e)\r\n\r\n\r\n else:\r\n print('Error in database (enter_scraped_data)')\r\n\r\n id = cur.lastrowid\r\n conn.commit()\r\n conn.close()\r\n return id\r\n\r\n#Database function to fetch data using task_id\r\ndef get_one_task_data(task_id):\r\n\r\n conn=create_connection(db_file)\r\n cur = conn.cursor()\r\n #cur.execute(\"SELECT player_name FROM players WHERE team_id=?\", (team_id,))\r\n try:\r\n\r\n cur.execute(\"SELECT * FROM data_scraped where task_id =?\", (task_id,))\r\n\r\n\r\n rows = cur.fetchall()\r\n except Exception as e:\r\n print('Error in get_one_task_data() --',e)\r\n #print(rows)\r\n #print(len(rows))\r\n\r\n return rows\r\n\r\n#DataBase functon to delete Tas from data base\r\n\r\ndef remove_task(task_id):\r\n print('Delete')\r\n conn=create_connection(db_file)\r\n cur = conn.cursor()\r\n #cur.execute(\"SELECT player_name FROM players WHERE team_id=?\", (team_id,))\r\n try:\r\n cur.execute(\"DELETE FROM data_scraped WHERE task_id =?\", (task_id,))\r\n cur.execute(\"DELETE FROM tasks WHERE task_id =?\", (task_id,))\r\n conn.commit()\r\n\r\n except Exception as e:\r\n print('Error in delete_task() --',e)\r\n #print(rows)\r\n #print(len(rows))\r\n\r\n######################## New Functions for New Table ################\r\n#Function to fetch monitor status\r\ndef get_monitor_status():\r\n\r\n conn=create_connection(db_file)\r\n cur = conn.cursor()\r\n #cur.execute(\"SELECT player_name FROM players WHERE team_id=?\", (team_id,))\r\n try:\r\n\r\n cur.execute(\"SELECT * FROM app_status where status_id =?\", (1,))\r\n rows = cur.fetchall()\r\n except Exception as e:\r\n print('Error in get_one_task_data() --',e)\r\n #print(rows)\r\n #print(len(rows))\r\n\r\n return rows[0][2]\r\n\r\n#Function to update monitor status in data base\r\ndef update_monitor_status(status):\r\n conn=create_connection(db_file)\r\n cur = conn.cursor()\r\n #cur.execute(\"SELECT player_name FROM players WHERE team_id=?\", (team_id,))\r\n try:\r\n sql=''' UPDATE app_status\r\n SET monitor = ? WHERE status_id = ? '''\r\n rows=cur.execute(sql, (status,1))\r\n\r\n conn.commit()\r\n print('Updated')\r\n\r\n\r\n\r\n except Exception as e:\r\n print('Error in updating app_status --',e)\r\n #print(rows)\r\n #print(len(rows))\r\n\r\n\r\n#DB Function to get Current Interval of Monitoring\r\ndef get_current_interval():\r\n\r\n conn=create_connection(db_file)\r\n cur = conn.cursor()\r\n #cur.execute(\"SELECT player_name FROM players WHERE team_id=?\", (team_id,))\r\n try:\r\n\r\n cur.execute(\"SELECT * FROM app_status where status_id =?\", (1,))\r\n rows = cur.fetchall()\r\n except Exception as e:\r\n print('Error in get_one_task_data() --',e)\r\n #print(rows)\r\n #print(len(rows))\r\n\r\n return rows[0][1]\r\n\r\n#Function to update Interval\r\ndef update_interval_time(interval):\r\n conn=create_connection(db_file)\r\n cur = conn.cursor()\r\n #cur.execute(\"SELECT player_name FROM players WHERE team_id=?\", (team_id,))\r\n try:\r\n sql=''' UPDATE app_status\r\n SET interval_time = ? WHERE status_id = ? '''\r\n rows=cur.execute(sql, (interval,1))\r\n conn.commit()\r\n print('Updated')\r\n\r\n\r\n except Exception as e:\r\n print('Error in updating Interval Time --',e)\r\n\r\n\r\n###CSV MAKING FUNCTION\r\ndef making_csv(task_name,task_data):\r\n d = task_data\r\n\r\n here = os.path.dirname(os.path.realpath(__file__))\r\n filepath = os.path.join(here,'csv-files', task_name+'.csv')\r\n #file = task_name+'.csv'\r\n csv_file = open(filepath, 'a', newline=\"\")\r\n writer = csv.writer(csv_file)\r\n for i in d:\r\n writer.writerow([i])\r\n for j in d[i]:\r\n writer.writerow([j,d[i][j]['name'],d[i][j]['price'],d[i][j]['stock']])\r\n\r\n csv_file.close()\r\n\r\n\r\n\r\n###################################################################\r\n#Function to check status of stocks\r\ndef check_stock(request,response):\r\n\r\n print('Checking Stock Availability ')\r\n\r\n m = ''\r\n s = '{x}'\r\n\r\n for i in request:\r\n for j in i['products']:\r\n print(j['stock'])\r\n m=m + s.format(x=j['stock']) + '\\n'\r\n print(response[i['site']][j['url']]['stock'])\r\n\r\n m=m + s.format(x=response[i['site']][j['url']]['stock']) + '\\n'\r\n for x in j['stock']:\r\n if x not in response[i['site']][j['url']]['stock']:\r\n print(x,' Out of stock ',response[i['site']][j['url']]['name'])\r\n m = m + s.format(x=(x,' Out of stock ',response[i['site']][j['url']]['name']+' '+j['url'])) +'\\n'\r\n return m\r\n\r\n#####################################################################################\r\n\"\"\"\r\ntask_urls = [\r\n\r\n { 'site':'adidas',\r\n 'products':[\r\n {\r\n 'url':'https://www.adidas.co.uk/solarboost-19-shoes/EF1413.html',\r\n 'stock':['6']\r\n },\r\n {\r\n 'url':'https://www.adidas.co.uk/copa-gloro-19.2-soft-ground-boots/F36080.html',\r\n 'stock':['6']\r\n },\r\n ]\r\n },\r\n\r\n]\r\n\r\n\r\nt3 = (None,'jogging','scrape',json.dumps(task_urls))\r\n\r\nprint('Creating Task')\r\n\r\ntry:\r\n create_task(t3)\r\nexcept Exception as e:\r\n\r\n print('Error adding task',e)\r\nprint('Ok')\r\n\r\n\r\n########## Now Code for , Fetching tasks from data base ##########\r\n\r\n\"\"\"\r\n\r\n#rows=get_task()\r\n\r\n#print(rows)\r\n#print(json.loads(rows[0][3]))\r\n\r\n\"\"\"\r\n\r\nprint('\\n')\r\nfor i in json.loads(rows[0][3]):\r\n print(i)\r\n print('******')\"\"\"\r\n\r\n#req = json.loads(rows[0][3])\r\n#response = main_scraper(json.loads(rows[0][3]))\r\n\r\n\r\n#task_id=rows[8][0]\r\n#data_name = rows[8][1]\r\n#data = response\r\n\r\n#d1 = (None,data_name,json.dumps(data),task_id)\r\n#enter_scraped_data(d1)\r\n\r\n######### Testing get_monitor_tasks() #################\r\n\r\n\r\n#rows=get_monitor_tasks()\r\n\r\n#print(type(json.loads(rows[0][3])))\r\n\r\n#print(delete_task(3))\r\n\r\n#print(len(get_one_task_data(5)))\r\n#################################\r\n\r\n#rows = update_monitor_status('start')\r\n\r\n#print(rows)\r\n\r\n#update_interval_time(5)\r\n\r\n\r\n\r\n\r\n\r\n", "sub_path": "worker.py", "file_name": "worker.py", "file_ext": "py", "file_size_in_byte": 8779, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sqlite3.connect", "line_number": 19, "usage_type": "call"}, {"api_name": "sqlite3.Error", "line_number": 21, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 230, "usage_type": "call"}, {"api_name": "os.path", "line_number": 230, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 230, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 231, "usage_type": "call"}, {"api_name": "os.path", "line_number": 231, "usage_type": "attribute"}, {"api_name": "csv.writer", "line_number": 234, "usage_type": "call"}]}
+{"seq_id": "124302447", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import (\n division, absolute_import, print_function, unicode_literals,\n)\nfrom builtins import * # noqa\nfrom future.builtins.disabled import * # noqa\n\nimport re\nfrom operator import methodcaller\nfrom collections import defaultdict\n# import itertools\n\nfrom zhlint.utils import (\n TextElement,\n count_newlines,\n count_offset,\n try_invoke\n)\n\n\nZH_CHARACTERS = (\n r'[\\u4e00-\\u9fff]'\n)\n\nZH_SYMBOLS = (\n r'['\n r'\\u3000-\\u303f'\n r'\\uff00-\\uff0f'\n r'\\uff1a-\\uff20'\n r'\\uff3b-\\uff40'\n r'\\uff5b-\\uff64'\n r'\\uffe0-\\uffee'\n r']'\n)\n\n\n# 1: a.\n# 2: whitespaces.\n# 3: b.\ndef single_space_patterns(\n a, b,\n a_join_b=True, b_join_a=True,\n a_non_preceding='', b_non_preceding=''):\n\n def join_non_preceding(non_preceding, text):\n if non_preceding:\n return r'(?:(??[\\\\]^_{}')\n\n ret = []\n\n # gerneral forms, ignore common shared characters: '@#%&+*-=|~'\n p1 = r'({0})(\\s*)'.format(\n r'|'.join(\n map(lambda punc: r'{0}+'.format(re.escape(punc)), PUNCTUATIONS),\n ),\n )\n\n patterns = [p1]\n ranges = parenthesis_ranges(element.content)\n\n for m in detect_by_patterns(\n patterns,\n element,\n ignore_matches=set(['......']),\n ):\n if match_in_ranges(m, ranges):\n continue\n if SpecialWordHelper.delimiter_in_word(element.content, m):\n continue\n if delimiter_in_email(element.content, m):\n continue\n if delimiter_in_simple_uri(element.content, m):\n continue\n if delimiter_in_latex_punctuation(element.content, m):\n continue\n\n if m.group(1) != '$$':\n ret.append(m)\n\n return ret\n\n\n# 1: chineses.\ndef detect_e202(element):\n if contains_chinese_characters(element.content):\n return False\n\n return detect_by_patterns(\n ['({0})'.format(ZH_SYMBOLS)],\n element,\n )\n\n\n# 1: wrong 「」.\ndef detect_e204(element):\n if not contains_chinese_characters(element.content):\n return False\n\n p = (\n r'('\n\n # ', \" is handled by E201.\n # r\"'\"\n # r'|'\n # r'\"'\n # r'|'\n\n # ‘’\n r'\\u2018|\\u2019'\n r'|'\n # “”\n r'\\u201c|\\u201d'\n\n r')'\n )\n return detect_by_patterns(\n [p],\n element,\n )\n\n\nclass SpecialWordHelper(object):\n\n WORD_PATTERN = {\n 'App': r'app',\n 'Android': r'android',\n 'iOS': r'ios',\n 'iPhone': r'iphone',\n 'App Store': r'app\\s?store',\n 'WiFi': r'wi-*fi',\n 'email': r'e-*mail',\n 'P.S.': r'P\\.*S\\.*',\n }\n\n WORD_MAX_LENGTH = None\n SENTENCE_DELIMITER_TO_WORD = None\n\n @classmethod\n def init(cls):\n if cls.WORD_MAX_LENGTH and cls.SENTENCE_DELIMITER_TO_WORD:\n return\n\n delimiters = [\n '!', ';', '.', '?',\n '\\uff01', '\\uff1b', '\\u3002', '\\uff1f',\n ]\n\n cls.SENTENCE_DELIMITER_TO_WORD = defaultdict(list)\n for delimiter in delimiters:\n for word in cls.WORD_PATTERN:\n if delimiter not in word:\n continue\n cls.SENTENCE_DELIMITER_TO_WORD[delimiter].append(word)\n\n cls.WORD_MAX_LENGTH = 0\n for word in cls.WORD_PATTERN:\n cls.WORD_MAX_LENGTH = max(cls.WORD_MAX_LENGTH, len(word))\n\n @classmethod\n def select_segment(cls, content, match):\n segment_begin = max(\n 0,\n match.end() - SpecialWordHelper.WORD_MAX_LENGTH,\n )\n segment_end = min(\n len(content) - 1,\n match.start() + SpecialWordHelper.WORD_MAX_LENGTH,\n )\n return content[segment_begin:segment_end]\n\n @classmethod\n def delimiter_in_word(cls, content, match):\n delimiter = match.group(1)\n if delimiter not in cls.SENTENCE_DELIMITER_TO_WORD:\n return False\n\n segment = cls.select_segment(content, match)\n for word in cls.SENTENCE_DELIMITER_TO_WORD.get(delimiter, []):\n if segment.find(word) >= 0:\n return True\n return False\n\n\nSpecialWordHelper.init()\n\n\n# 1: wrong special word.\ndef detect_e301(element):\n\n for correct_form, pattern in SpecialWordHelper.WORD_PATTERN.items():\n\n # (?/', apiviews.answer_view, name='ПОСМОТРЕТЬ ОТВЕТ'),\n path('answer/update//', apiviews.answer_update, name='ОБНОВИТЬ ОТВЕТ'),\n \n ############### опросы #################\n path('surveysApp/create/', apiviews.survey_create, name='СОЗДАТЬ ОПРОС'),\n path('surveysApp/update//', apiviews.survey_update, name='ОБНОВИТЬ ОПРОС'),\n path('surveysApp/view/', apiviews.survey_view, name='ПОСМОТРЕТЬ ОПРОС'),\n path('surveysApp/view/active/', apiviews.active_survey_view, name='ОБНОВИТЬ ОПРОС'),\n \n \n ############### вопросы #################\n path('question/create/', apiviews.question_create, name='СОЗДАТЬ ВОПРОС'),\n path('question/update//', apiviews.question_update, name='ОБНОВИТЬ ВООПРОС'),\n \n \n ############### выбор #################\n path('choice/create/', apiviews.choice_create, name='ВЫБРАТЬ ОТВЕТ'),\n path('choice/update//', apiviews.choice_update, name='ОБНОВИТЬ ОТВЕТ'),\n \n \n \n \n\n \n]\n\n", "sub_path": "surveysApp/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1556, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "surveysApp.apiviews.login", "line_number": 11, "usage_type": "attribute"}, {"api_name": "surveysApp.apiviews", "line_number": 11, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "surveysApp.apiviews.answer_create", "line_number": 14, "usage_type": "attribute"}, {"api_name": "surveysApp.apiviews", "line_number": 14, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 15, "usage_type": "call"}, {"api_name": "surveysApp.apiviews.answer_view", "line_number": 15, "usage_type": "attribute"}, {"api_name": "surveysApp.apiviews", "line_number": 15, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 16, "usage_type": "call"}, {"api_name": "surveysApp.apiviews.answer_update", "line_number": 16, "usage_type": "attribute"}, {"api_name": "surveysApp.apiviews", "line_number": 16, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 19, "usage_type": "call"}, {"api_name": "surveysApp.apiviews.survey_create", "line_number": 19, "usage_type": "attribute"}, {"api_name": "surveysApp.apiviews", "line_number": 19, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 20, "usage_type": "call"}, {"api_name": "surveysApp.apiviews.survey_update", "line_number": 20, "usage_type": "attribute"}, {"api_name": "surveysApp.apiviews", "line_number": 20, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 21, "usage_type": "call"}, {"api_name": "surveysApp.apiviews.survey_view", "line_number": 21, "usage_type": "attribute"}, {"api_name": "surveysApp.apiviews", "line_number": 21, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 22, "usage_type": "call"}, {"api_name": "surveysApp.apiviews.active_survey_view", "line_number": 22, "usage_type": "attribute"}, {"api_name": "surveysApp.apiviews", "line_number": 22, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 26, "usage_type": "call"}, {"api_name": "surveysApp.apiviews.question_create", "line_number": 26, "usage_type": "attribute"}, {"api_name": "surveysApp.apiviews", "line_number": 26, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 27, "usage_type": "call"}, {"api_name": "surveysApp.apiviews.question_update", "line_number": 27, "usage_type": "attribute"}, {"api_name": "surveysApp.apiviews", "line_number": 27, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 31, "usage_type": "call"}, {"api_name": "surveysApp.apiviews.choice_create", "line_number": 31, "usage_type": "attribute"}, {"api_name": "surveysApp.apiviews", "line_number": 31, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 32, "usage_type": "call"}, {"api_name": "surveysApp.apiviews.choice_update", "line_number": 32, "usage_type": "attribute"}, {"api_name": "surveysApp.apiviews", "line_number": 32, "usage_type": "name"}]}
+{"seq_id": "451285398", "text": "from flask import Flask # Constructor\napp = Flask(__name__) # Application\n\ncontacts = [\n {\n \"contact\": {\n \"id\": 1123231,\n \"name\": \"Name A\"\n }\n },\n {\n \"contact\": {\n \"id\": 1125261,\n \"name\": \"Name B\"\n }\n }\n ]\n\n@app.route('/') # Decortator - main route of the application\ndef microservices_api(): # Without parameters\n # Only for debugging purposes\n # a = 'Basic Flask web application'\n # b = int(a) # For debugging using the PIN\n\n # HTTP response JSON Object\n return {\n \"status\": \"200 OK\",\n \"msg\":\"Welcome to Microservices API in Flask with Python\"\n }\n\n\n@app.route('/contacts/') # Between /.../ in order to grant the navigation to this endpoint\ndef contacts_api():\n return {\n \"contacts\": contacts\n }\n\n@app.route('/contacts/')\ndef contacts_api_contact_endpoint(id):\n response = {\n \"status\": '404 Not Found',\n \"msg\": f\"There is no contact with id: {id} in our database\"\n }\n for index in range(0, len(contacts)):\n if contacts[index][\"contact\"][\"id\"] == id:\n response = {\n \"status\": \"200 OK\",\n \"msg\": contacts[index]\n }\n break\n return response\n\n# Remove if you use manager.py\nif __name__ == '__main__':\n app.run('0.0.0.0', 5000, debug=True) # Run web server, accept requests from any source by the port 5000 and run it in debug mode\n", "sub_path": "openweb-training/flask/app_fundamentals/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1565, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "flask.Flask", "line_number": 2, "usage_type": "call"}]}
+{"seq_id": "558287946", "text": "from django.db import models\nfrom rest_framework import serializers\nfrom rest_framework.utils import field_mapping\nfrom rest_framework.fields import OrderedDict, SkipField, empty, CharField, JSONField as BaseJSONField, IntegerField\nfrom rest_framework.relations import PrimaryKeyRelatedField\n\nfrom comment.models import Comment\nfrom post.models import Post\nfrom post.serializers import PostSerializer\nfrom user.mixins import AuthenticatedSerializerMixin\n\n\n# Property descriptor to be used on the model instance for above field, similar to in-built\n# class 'models.fields.related_descriptors.ForwardManyToOneDescriptor'\n#\nclass IntegerAsForeignFieldDescriptor:\n\tdef __init__(self, field):\n\t\tself.field = field\n\n\tdef __get__(self, instance, instance_type=None):\n\t\tif instance is None:\n\t\t\treturn None\n\n\t\tpk = getattr(instance, self.field.attname)\n\t\tif pk is None:\n\t\t\treturn None\n\n\t\t# use the saved related data, but only if it hasn't changed\n\t\tif hasattr(self, 'value') and pk == getattr(self.value, self.field.remote_field_name):\n\t\t\treturn self.value\n\n\t\tself.value = self.field.get_remote_obj(pk)\n\t\treturn self.value\n\n\tdef __set__(self, instance, value):\n\t\tif value is None and self.field.null is False:\n\t\t\traise ValueError(\n\t\t\t\t'Cannot assign None: \"%s.%s\" does not allow null values.' %\n\t\t\t\t(instance._meta.object_name, self.field.name)\n\t\t\t)\n\t\telif value is not None:\n\t\t\tif not isinstance(value, self.field.remote_model):\n\t\t\t\traise ValueError(\n\t\t\t\t\t'Cannot assign \"%r\": \"%s.%s\" must be a \"%s\" instance.' % (\n\t\t\t\t\t\tvalue,\n\t\t\t\t\t\tinstance._meta.object_name,\n\t\t\t\t\t\tself.field.name,\n\t\t\t\t\t\tself.field.remote_model._meta.object_name,\n\t\t\t\t\t)\n\t\t\t\t)\n\n\t\t\tpk = getattr(value, self.field.remote_field_name)\n\t\t\tsetattr(instance, self.field.attname, pk) # update the related field\n\t\t\tself.value = value # Also save the related instance to avoid fetching it again in 'get'\n\n\n# -----------------------------------------------\n# Custom Integer Field class to be used for 'soft' foreign object reference - Soft because there is no\n# actual referential integrity check/guarantee in DB, just in the code layer it ensures that the\n# foreign object used actually exists in DB, and no reverse check either\n#\n# This should be used for fields which are supposed to be foreign-key to other model\n# but are not defined such (no constrain in DB). For example, when the other model (table) is in foreign database,\n# like in case of Django's lack of support for cross-db foreign key\n#\n# Todo: Remove IntegerField dependency, could be derived from 'Field' directly allowing it to be used for non-int types\n#\nclass IntegerAsForeignField(models.IntegerField):\n\t# default_error_messages = {\n\t# \t'does_not_exist': _('Invalid pk \"{pk_value}\" - object does not exist.'),\n\t# \t'incorrect_type': _('Incorrect type. Expected pk value, received {data_type}.'),\n\t# }\n\n\tdef __init__(self, remote_model=None, remote_field_name='pk', **kwargs):\n\t\tself.remote_model = remote_model\n\t\tself.remote_field_name = remote_field_name\n\t\tsuper().__init__(**kwargs)\n\n\tdef contribute_to_class(self, cls, name, virtual_only=False):\n\t\tsuper().contribute_to_class(cls, name, virtual_only=virtual_only)\n\t\tsetattr(cls, self.name, IntegerAsForeignFieldDescriptor(self))\n\n\tdef check(self, **kwargs):\n\t\tassert (\n\t\tself.remote_model and isinstance(self.remote_model, type) and issubclass(self.remote_model, models.Model)), \\\n\t\t\t\"Invalid parameter 'remote_model': should be a valid Model class found '{0}' instead\".format(\n\t\t\t\tself.remote_model)\n\n\t\tassert issubclass(self.model, IntegerAsForeignFieldModelMixin), \\\n\t\t\t\"Model containing this field ('{0}=fieldfactory.IntegerAsForeignField(...)') must use 'IntegerAsForeignFieldModelMixin' as base\".format(\n\t\t\t\tself.name)\n\n\t\treturn super().check(**kwargs)\n\n\tdef get_attname(self):\n\t\treturn self.name if self.name.endswith('_id') else self.name + \"_id\"\n\n\tdef get_attval(self, obj_or_pk):\n\t\treturn obj_or_pk and (\n\t\tgetattr(obj_or_pk, self.remote_field_name) if isinstance(obj_or_pk, self.remote_model) else obj_or_pk)\n\n\tdef _raise_does_not_exist(self, value):\n\t\tpass\n\n\t# raise NotFound(_('{model} with {field}={value} does not exist').format(\n\t# \tmodel=self.remote_model.__name__,\n\t# \tfield=self.remote_field_name,\n\t# \tvalue=value\n\t# ))\n\n\tdef get_remote_obj(self, pk):\n\t\ttry:\n\t\t\treturn pk and self.remote_model.objects.get(**{self.remote_field_name: pk})\n\t\texcept self.remote_model.DoesNotExist:\n\t\t\tself._raise_does_not_exist(pk)\n\t\texcept (TypeError, ValueError):\n\t\t\traise\n\t\t# raise NotFound(_('Incorrect type. Expected pk value ({0}), received {1}.').format(\n\t\t# \tself.remote_field_name, type(pk).__name__))\n\n\n# Model mixin, use this if your model has any IntegerAsForeignField field\n#\nclass IntegerAsForeignFieldModelMixin:\n\tdef __init__(self, *args, **kwargs):\n\t\t# We override because we want to handle IntegerAsForeignField fields (which is actually IntegerField)\n\t\t# in same way as default ForeignField i.e accepting either of field.name ('fkobject=obj_or_pk')\n\t\t# or field.attname ('fkobject_id=obj_pk') as input\n\t\t#\n\t\t# Default implementation for IntegerField just looks for field.attname and ignores field.name\n\t\t# so here we just update kwargs with field.attname as well\n\t\tif kwargs:\n\t\t\tfor field in self._meta.fields:\n\t\t\t\tif isinstance(field, IntegerAsForeignField):\n\t\t\t\t\tif field.name in kwargs and not field.attname in kwargs:\n\t\t\t\t\t\tkwargs[field.attname] = field.get_attval(kwargs.pop(field.name))\n\n\t\tsuper().__init__(*args, **kwargs)\n\n\n# Serializer field class that would be used in conjunction with IntegerAsForeignField - is derived from ModelField\n# instead of IntegerField as we need the 'model_field' value which is only passed to ModelField instance (see\n# serializers.build_standard_field() method popping 'model_field' from kwargs for all but ModelField)\n#\nclass IntegerPrimaryKeySerializerField(serializers.ModelField):\n\tdef __init__(self, model_field, **kwargs):\n\t\t# our model field is actually an IntegerField which has min/max attribute, ignore it\n\t\tkwargs.pop('min_value', None)\n\t\tkwargs.pop('max_value', None)\n\t\tsuper().__init__(model_field, **kwargs)\n\n\tdef get_nested_serializer_class(self):\n\t\tparent = self.parent\n\n\t\t# For 'applied_depth' see utils.serializers.ModelSerializer\n\t\tcur_depth = getattr(parent, 'applied_depth', None) or getattr(parent.Meta, 'depth', None)\n\t\tif cur_depth and int(cur_depth) >= 1:\n\t\t\t# depth > 0, we need to provide serializer for expanding the field\n\t\t\t# Check if parent serializer specifies any custom serializer for this field via 'get_nested_field_serializer_class'\n\t\t\t# otherwise create and return the the default one\n\t\t\tget_nested_serializers = getattr(parent, 'get_nested_field_serializer_class', None)\n\t\t\tif get_nested_serializers and callable(get_nested_serializers):\n\t\t\t\tser = get_nested_serializers(self.field_name)\n\t\t\t\tif ser:\n\t\t\t\t\t# found one, create a copy with proper depth\n\t\t\t\t\tclass NestedSerializer(ser):\n\t\t\t\t\t\tclass Meta(ser.Meta):\n\t\t\t\t\t\t\tser.Meta.depth = cur_depth - 1\n\n\t\t\t\t\treturn NestedSerializer\n\n\t\t\t# No custom serializer, lets just the use the default one\n\t\t\tclass NestedSerializer(serializers.ModelSerializer):\n\t\t\t\tclass Meta:\n\t\t\t\t\tmodel = self.model_field.remote_model\n\t\t\t\t\tdepth = cur_depth - 1\n\n\t\t\treturn NestedSerializer\n\n\tdef to_internal_value(self, data):\n\t\tdata = super().to_internal_value(data)\n\t\treturn self.model_field.get_remote_obj(data)\n\n\tdef to_representation(self, value):\n\t\t# value is complete model instance and not this field value, see ModelField for details\n\n\t\t# handle depth > 0\n\t\tnested_serializer_class = self.get_nested_serializer_class()\n\t\tif nested_serializer_class:\n\t\t\treturn nested_serializer_class(getattr(value, self.field_name)).data\n\n\t\treturn self.model_field.value_from_object(value)\n\n\n# Add to standard mapping...\nserializers.ModelSerializer.serializer_field_mapping[IntegerAsForeignField] = IntegerPrimaryKeySerializerField\n\n\ndef serializer_skip_null_to_representation(self, instance):\n\t\"\"\"\n\tSame as base class implementation but skips null values.\n\n\tObject instance -> Dict of primitive datatypes.\n\t\"\"\"\n\tret = OrderedDict()\n\tfields = [field for field in self.fields.values()]\n\n\tdepth = self.applied_depth if hasattr(self, 'applied_depth') else getattr(self.Meta, 'depth', 0) if hasattr(self, 'Meta') else 0\n\tfor field in fields:\n\t\ttry:\n\t\t\tattribute = field.get_attribute(instance)\n\t\texcept SkipField:\n\t\t\tcontinue\n\n\t\t# Keeps only if it is non-null\n\t\tif attribute is not None:\n\t\t\tvalue = field.to_representation(attribute)\n\t\t\tif value is not None:\n\t\t\t\tif (depth == 0 or isinstance(value, int)) and isinstance(field, (\n\t\t\t\tPrimaryKeyRelatedField, IntegerPrimaryKeySerializerField)):\n\t\t\t\t\tret[field.field_name + '_id'] = value\n\t\t\t\telse:\n\t\t\t\t\tret[field.field_name] = value\n\t\t\t\t# else:\n\t\t\t\t# \tif isinstance(field, rest_framework.fields.IntegerField) or isinstance(field,\n\t\t\t\t# \t\t\trest_framework.fields.FloatField):\n\t\t\t\t# \t\tret[field.field_name] = 1234567890123456\n\t\t\t\t# \telse:\n\t\t\t\t# \t\tret[field.field_name] = 'THIS FIELD IS NULL'\n\treturn ret\n\n\nserializers.Serializer.to_representation = serializer_skip_null_to_representation\n\n\nclass CommentSerializer(AuthenticatedSerializerMixin, serializers.ModelSerializer):\n\n\tauthor = serializers.SerializerMethodField()\n\n\tdef get_author(self, obj):\n\t\tif obj.user:\n\t\t\treturn obj.user.get_author()\n\t\treturn 'Anonymous'\n\n\tclass Meta:\n\t\tmodel = Comment\n\t\tfields = ('id', 'post', 'author', 'desc', 'created_on', 'updated_on', 'comment')\n\t\tread_only_fields = ('user', 'post', 'created_on', 'updated_on')\n\t\tignore_depth_fields = ('user',)\n\n\t# maximum allowed depth for this serializer, relatively safe default of 1 (0?) :-) override if you need more\n\t# Note that drf has its own hard max set at 10 which can't be bypassed\n\tmax_depth = 1\n\n\t# List of fields that are NOT allowed to expand even on depth > 0\n\t# ignore_depth_fields = ('user',)\n\n\t# field_name vs. custom SerializerClass mapping\n\t# nested_field_serializers = {\n\t# \t# 'field_name': SerializerClass\n\t# }\n\n\t@classmethod\n\tdef _get_meta_or_class_property(cls, name, default=None):\n\t\t\"\"\"\n\t\t:param name: property name\n\t\t:return: Returns the Meta class property if defined, otherwise try on self class\n\t\t\"\"\"\n\t\tvalue = getattr(cls.Meta, name, None)\n\t\tif value is None:\n\t\t\tvalue = getattr(cls, name, default)\n\t\treturn value\n\n\t@classmethod\n\tdef get_nested_field_serializer_class(cls, field_name, default_field_class=None):\n\t\td = cls._get_meta_or_class_property('nested_field_serializers')\n\t\treturn d.get(field_name, default_field_class) if d else default_field_class\n\n\tdef __init__(self, *args, **kwargs):\n\t\tself.requested_fields = kwargs.pop('fields', None)\n\t\tself.requested_depth = kwargs.pop('depth', None)\n\t\tsuper().__init__(*args, **kwargs)\n\n\t\t# Enforce max_depth, in case of Meta.depth (set in code) complain immediately..\n\t\tmax_depth = self._get_meta_or_class_property('max_depth')\n\t\tdepth = getattr(self.Meta, 'depth', 0)\n\t\tassert max_depth is None or depth <= max_depth, \"depth={0} is more than max_depth={1}\".format(depth, max_depth)\n\n\t\t# ..but for requested_depth (set by client) silently cap it to max\n\t\tif self.requested_depth and self.requested_depth > max_depth:\n\t\t\tself.requested_depth = max_depth\n\n\t@property\n\tdef applied_depth(self):\n\t\tif self.requested_depth:\n\t\t\treturn self.requested_depth\n\t\treturn getattr(self.Meta, 'depth', 0)\n\n\tdef build_nested_field(self, field_name, relation_info, nested_depth):\n\t\t\"\"\"\n\t\tCreate nested fields for forward and reverse relationships\n\n\t\tWe override to specify our own Serializer classes for nested field if given (via 'nested_field_serializers'\n\t\t mapping or nested_field_serializers() method override)\n\n\t\tAlso checks if the requested nested field is marked as 'never-expand' fields in which case we simply\n\t\t route to non-depth path 'build_relational_field' (see super's method)\n\t\t\"\"\"\n\n\t\t# See if the field is allowed to expand at all, either because we hit the max-depth or because\n\t\t# this field itself is not allowed to - break right away it not\n\t\tmax_depth = self._get_meta_or_class_property('max_depth', default=1)\n\t\tassert nested_depth <= max_depth, \"depth={0} is more than max_depth={1}\".format(nested_depth, max_depth)\n\n\t\tif field_name in self._get_meta_or_class_property('ignore_depth_fields', default=[]):\n\t\t\treturn self.build_relational_field(field_name, relation_info)\n\n\t\t# Need to expand, see if there is a Serializer set explicitly for this field..\n\t\tfield_class = self.get_nested_field_serializer_class(field_name)\n\t\tif field_class:\n\t\t\t# We create a subclass on the fly as we don't want original's Meta.depth to be messed-up\n\t\t\tclass NestedSerializer(field_class):\n\t\t\t\tclass Meta(field_class.Meta):\n\t\t\t\t\tdepth = nested_depth - 1\n\n\t\t\treturn NestedSerializer, field_mapping.get_nested_relation_kwargs(relation_info)\n\n\t\t# Nothing special to do, just pass on to super\n\t\treturn super().build_nested_field(field_name, relation_info, nested_depth)\n\n\tdef get_fields(self):\n\t\t\"\"\"\n\t\tWe override to apply 'depth' requested by client just before the call to super (and reset it back before returning)\n\t\t:return:\n\t\t\"\"\"\n\t\tif self.requested_depth is not None:\n\t\t\tdefault_depth = getattr(self.Meta, 'depth', None)\n\t\t\tif default_depth != self.requested_depth:\n\t\t\t\tself.Meta.depth = self.requested_depth\n\t\t\t\tfields = super().get_fields()\n\t\t\t\tif default_depth is not None:\n\t\t\t\t\tself.Meta.depth = default_depth\n\t\t\t\telse:\n\t\t\t\t\tdelattr(self.Meta, 'depth')\n\t\t\t\treturn fields\n\t\treturn super().get_fields()\n", "sub_path": "comment/serializers.py", "file_name": "serializers.py", "file_ext": "py", "file_size_in_byte": 13330, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.db.models.IntegerField", "line_number": 68, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 68, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 85, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 85, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelField", "line_number": 145, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 145, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 173, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 173, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 196, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 196, "usage_type": "name"}, {"api_name": "rest_framework.fields.OrderedDict", "line_number": 205, "usage_type": "call"}, {"api_name": "rest_framework.fields.SkipField", "line_number": 212, "usage_type": "name"}, {"api_name": "rest_framework.relations.PrimaryKeyRelatedField", "line_number": 220, "usage_type": "name"}, {"api_name": "rest_framework.serializers.Serializer", "line_number": 233, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 233, "usage_type": "name"}, {"api_name": "user.mixins.AuthenticatedSerializerMixin", "line_number": 236, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 236, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 236, "usage_type": "name"}, {"api_name": "rest_framework.serializers.SerializerMethodField", "line_number": 238, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 238, "usage_type": "name"}, {"api_name": "comment.models.Comment", "line_number": 246, "usage_type": "name"}, {"api_name": "rest_framework.utils.field_mapping.get_nested_relation_kwargs", "line_number": 326, "usage_type": "call"}, {"api_name": "rest_framework.utils.field_mapping", "line_number": 326, "usage_type": "name"}]}
+{"seq_id": "528468227", "text": "# -*- coding: utf-8 -*-\nimport configparser\nimport argparse\nimport logging.handlers\nimport socket\nimport time\n\nimport urllib\nimport urllib.error\n\nfrom daemon import Daemon\nfrom unifi_protocol import create_broadcast_message, send_inform\nimport json\nfrom packtest import packet\nimport aquire\nfrom random import randint\nimport hashlib\nimport psutil\n\nDEFAULT_AUTHKEY = hashlib.md5(b'ubnt').digest().hex() # 'ba86f2bbe107c7c57eb5f2690775c712'\n\nhandler = logging.handlers.SysLogHandler(address='/dev/log')\nhandler.setFormatter(logging.Formatter('[unifi-gateway] : %(levelname)s : %(message)s'))\nlogger = logging.getLogger(__name__)\nlogger.setLevel(logging.DEBUG)\nlogger.addHandler(handler)\n\nch = logging.StreamHandler()\nch.setLevel(logging.DEBUG)\nlogger.addHandler(ch)\n\nproto_logger = logging.getLogger('unifi_protocol')\nproto_logger.setLevel(logging.DEBUG)\nproto_logger.addHandler(ch)\n\nCONFIG_FILE = 'conf/unifi-gateway.conf'\n\n\ndef print_bytearray(value, size=32):\n _data = [f'{b:02x}' for b in value]\n for i,d in enumerate(_data):\n print(f'{d} ', end='')\n\n if (i+1) % size == 0:\n print()\n print()\n\n\ndef mgmt_decode(items):\n return dict(map(lambda s : s.split('='), items))\n\n\nclass UnifiGateway(Daemon):\n\n def __init__(self, **kwargs):\n self.interval = 10\n self.config = configparser.RawConfigParser()\n self.config.read(CONFIG_FILE)\n self.authkey = self.config.get('gateway', 'key', fallback=DEFAULT_AUTHKEY)\n self.url = self.config.get('gateway', 'url', fallback='http://unifi:8080/inform')\n self.cfgversion = '?'\n self.is_adopted = False if self.authkey == DEFAULT_AUTHKEY else True\n self.encryption = self.config.get('gateway', 'encryption', fallback='CBC')\n\n Daemon.__init__(self, pidfile=self.config.get('global', 'pid_file'), **kwargs)\n\n def run(self):\n self.url = self.config.get('gateway', 'url')\n while True:\n logger.debug(f'run: Sending request to {self.url}')\n response = None\n try:\n _response = send_inform(self.create_inform(),\n url=self.config.get('gateway', 'url'),\n key=self.authkey,\n encryption=self.encryption)\n except urllib.error.HTTPError as E:\n logger.error(\"HTTP Error\")\n logger.error(E)\n time.sleep(self.interval)\n continue\n\n if _response:\n response = json.loads(str(_response))\n logger.debug(f'run: response from server: {response}')\n\n if response['_type'] == 'setparam':\n\n if 'mgmt_cfg' in response:\n mgmt_cfg = mgmt_decode(response['mgmt_cfg'].splitlines())\n if 'authkey' in mgmt_cfg:\n self.authkey = mgmt_cfg['authkey']\n self.config.set('gateway', 'is_adopted', True)\n\n if 'cfgversion' in mgmt_cfg:\n self.cfgversion = mgmt_cfg['cfgversion']\n\n if 'use_aes_gcm' in mgmt_cfg and mgmt_cfg['use_aes_gcm']:\n logger.debug('Setting encryption to GCM')\n self.encryption = 'GCM'\n self.config.set('gateway', 'encryption', 'GCM')\n\n for key, value in list(response.items()):\n if key == 'mgmt_cfg':\n self.authkey == value\n\n self._save_config()\n\n time.sleep(self.interval)\n\n def quit(self):\n pass\n\n def set_adopt(self, url, key):\n self.config.set('gateway', 'url', url)\n self.config.set('gateway', 'key', key)\n self._save_config()\n\n response = self._send_inform(create_inform(), url=self.config.get('gateway', 'url'))\n logger.debug('Receive {} from controller'.format(response))\n if response['_type'] == 'setparam':\n for key, value in list(response.items()):\n if key not in ['_type', 'server_time_in_utc', 'mgmt_cfg']:\n self.config.set('gateway', key, value)\n self.config.set('gateway', 'is_adopted', True)\n self._save_config()\n\n def _save_config(self):\n with open(CONFIG_FILE, 'w') as config_file:\n self.config.set('gateway', 'key', self.authkey)\n\n self.config.write(config_file)\n\n def _create_nic_stats(self, nic):\n netstats = aquire.network_statistics()\n return {\n 'enable': True,\n 'full_duplex': True,\n 'gateways': [\n ],\n 'ip': '158.38.145.81',\n 'latency': 4,\n 'mac': '80:2a:a8:cd:a9:54',\n 'name': 'eth0',\n 'nameservers': [\n ],\n 'netmask': '255.255.255.192',\n 'num_port': 1,\n 'drops': netstats[nic].rx_dropped,\n 'rx_bytes': netstats[nic].rx_bytes,\n 'rx_dropped': netstats[nic].rx_dropped,\n 'rx_errors': netstats[nic].rx_errors,\n 'rx_multicast': netstats[nic].rx_multicast,\n 'rx_packets': netstats[nic].rx_packets,\n 'speed': 1000,\n # 'speedtest_lastrun': int(time.time()),\n # 'speedtest_ping': randint(0, 2000),\n # 'speedtest_status': 'Idle',\n 'tx_bytes': netstats[nic].tx_bytes,\n 'tx_dropped': netstats[nic].tx_dropped,\n 'tx_errors': netstats[nic].tx_errors,\n 'tx_packets': netstats[nic].tx_packets,\n 'up': True,\n 'uptime': aquire.uptime(),\n 'xput_down': 0,\n 'xput_up': 0,\n }\n\n def _create_empty_nic_stats(self, nic):\n {\n 'enable': False,\n 'name': nic,\n 'up': False,\n }\n\n def _create_complete_inform(self):\n lan1_nic = 'eth0'\n if lan1_nic:\n lan1_stats = self._create_nic_stats(lan1_nic)\n else:\n lan1_stats = self._create_empty_stats(lan1_nic)\n\n pkg = {\n 'mac': 'f0:9f:c2:79:34:fd',\n 'ip': '158.38.145.81',\n 'netmask': '255.255.255.0',\n 'model': 'UGW4',\n 'model-display': 'UniFi-Gateway-4',\n 'version': '4.4.51.5287926',\n\n 'hostname': 'UBNT',\n 'inform_url': 'http://158.38.145.72:8080/inform',\n 'last_error': \"Unknown[11] (http://158.38.145.72:8080/inform)\",\n 'isolated': False,\n \"default\": not self.is_adopted,\n 'uptime': aquire.uptime(),\n 'cfgversion': self.cfgversion,\n 'led_enabled': True,\n 'discovery_response': not self.is_adopted,\n 'has_dpi': False,\n 'serial': self.config.get('gateway', 'lan_mac').replace(':', ''),\n\n 'bootrom_version': 'unknown',\n 'required_version': '4.0.0',\n\n 'system-stats': {\n 'cpu': '%s' % psutil.cpu_percent(),\n 'mem': '%s' % (100 - psutil.virtual_memory()[2]),\n 'uptime': '%s' % aquire.uptime()\n },\n\n\n 'config_network_wan': {'type': 'dhcp'},\n 'config_port_table': [\n {'ifname': 'eth0', 'name': 'lan'},\n {'ifname': 'eth1', 'name': 'lan'},\n {'ifname': 'eth2', 'name': 'lan'},\n {'ifname': 'eth3', 'name': 'wan'}\n ],\n 'has_eth1': True,\n 'has_porta': True,\n 'has_ssh_disable': True,\n 'time': int(time.time()),\n 'uplink': 'eth3',\n #'routes': [\n # { \n # 'nh': [{'intf': 'eth0', 'metric': '1/0', 't': 'S>*', 'via': '20.1.1.1'}],\n # 'pfx': '0.0.0.0/0'\n # }],\n 'if_table': [\n lan1_stats,\n {\n 'enable': False,\n 'full_duplex': True,\n 'speed': 100,\n 'name': 'eth1',\n 'up': True,\n },\n {\n 'enable': False,\n 'full_duplex': True,\n 'up': False\n },\n {\n 'enable': True,\n 'full_duplex': True,\n 'ip': '1.1.1.1/24',\n 'mac': '80:2a:a8:cd:a9:53',\n 'name': 'eth3',\n 'netmask': '255.255.255.0',\n 'num_port': 1,\n 'rx_bytes': 807912794876,\n 'rx_dropped': 2800,\n 'rx_errors': 0,\n 'rx_multicast': 412314,\n 'rx_packets': 700376545,\n 'speed': 1000,\n 'tx_bytes': 58901673253,\n 'tx_dropped': 0,\n 'tx_errors': 0,\n 'tx_packets': 347161831,\n 'up': True\n },\n \n ],\n #'network_table': [\n # {\n # 'address': '192.168.1.1/24',\n # 'addresses': [\n # '%s/24' % '1.1.1.1'\n # ],\n # 'autoneg': 'true',\n # 'duplex': 'full',\n # 'host_table': [\n # ],\n # 'l1up': 'true',\n # 'mac': '80:2a:a8:cd:a9:53',\n # 'mtu': '1500',\n # 'name': 'eth1',\n # 'speed': '1000',\n # 'stats': {\n # 'multicast': '412294',\n # 'rx_bps': '342',\n # 'rx_bytes': 52947224765,\n # 'rx_dropped': 2800,\n # 'rx_errors': 0,\n # 'rx_multicast': 412314,\n # 'rx_packets': 341232922,\n # 'tx_bps': '250',\n # 'tx_bytes': 792205417381,\n # 'tx_dropped': 0,\n # 'tx_errors': 0,\n # 'tx_packets': 590930778\n # },\n # 'up': 'true'\n # }\n # ],\n }\n return pkg\n\n def create_inform(self):\n return self._create_complete_inform()\n\n\ndef restart(args):\n UnifiGateway().restart()\n\n\ndef stop(args):\n UnifiGateway().stop()\n\n\ndef start(args):\n UnifiGateway().start()\n\n\ndef set_adopt(args):\n url, key = args.s, args.k\n UnifiGateway().set_adopt(url, key)\n\n\ndef run(args):\n UnifiGateway().run()\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n subparsers = parser.add_subparsers()\n\n parser_start = subparsers.add_parser('start', help='start unifi gateway daemon')\n parser_start.set_defaults(func=start)\n\n parser_run = subparsers.add_parser('run', help='Start unify gateway in the foreground')\n parser_run.set_defaults(func=run)\n\n parser_stop = subparsers.add_parser('stop', help='stop unifi gateway daemon')\n parser_stop.set_defaults(func=stop)\n\n parser_restart = subparsers.add_parser('restart', help='restart unifi gateway daemon')\n parser_restart.set_defaults(func=restart)\n\n parser_adopt = subparsers.add_parser('set-adopt', help='send the adoption request to the controller')\n parser_adopt.add_argument('-s', type=str, help='controller url', required=True)\n parser_adopt.add_argument('-k', type=str, help='key', required=True)\n parser_adopt.set_defaults(func=set_adopt)\n\n args = parser.parse_args()\n args.func(args)\n", "sub_path": "unifi_gateway.py", "file_name": "unifi_gateway.py", "file_ext": "py", "file_size_in_byte": 11708, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "hashlib.md5", "line_number": 20, "usage_type": "call"}, {"api_name": "logging.handlers.handlers.SysLogHandler", "line_number": 22, "usage_type": "call"}, {"api_name": "logging.handlers.handlers", "line_number": 22, "usage_type": "attribute"}, {"api_name": "logging.handlers", "line_number": 22, "usage_type": "name"}, {"api_name": "logging.handlers.Formatter", "line_number": 23, "usage_type": "call"}, {"api_name": "logging.handlers", "line_number": 23, "usage_type": "name"}, {"api_name": "logging.handlers.getLogger", "line_number": 24, "usage_type": "call"}, {"api_name": "logging.handlers", "line_number": 24, "usage_type": "name"}, {"api_name": "logging.handlers.DEBUG", "line_number": 25, "usage_type": "attribute"}, {"api_name": "logging.handlers", "line_number": 25, "usage_type": "name"}, {"api_name": "logging.handlers.StreamHandler", "line_number": 28, "usage_type": "call"}, {"api_name": "logging.handlers", "line_number": 28, "usage_type": "name"}, {"api_name": "logging.handlers.DEBUG", "line_number": 29, "usage_type": "attribute"}, {"api_name": "logging.handlers", "line_number": 29, "usage_type": "name"}, {"api_name": "logging.handlers.getLogger", "line_number": 32, "usage_type": "call"}, {"api_name": "logging.handlers", "line_number": 32, "usage_type": "name"}, {"api_name": "logging.handlers.DEBUG", "line_number": 33, "usage_type": "attribute"}, {"api_name": "logging.handlers", "line_number": 33, "usage_type": "name"}, {"api_name": "daemon.Daemon", "line_number": 53, "usage_type": "name"}, {"api_name": "configparser.RawConfigParser", "line_number": 57, "usage_type": "call"}, {"api_name": "daemon.Daemon.__init__", "line_number": 65, "usage_type": "call"}, {"api_name": "daemon.Daemon", "line_number": 65, "usage_type": "name"}, {"api_name": "unifi_protocol.send_inform", "line_number": 73, "usage_type": "call"}, {"api_name": "urllib.error", "line_number": 77, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 80, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 84, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 109, "usage_type": "call"}, {"api_name": "aquire.network_statistics", "line_number": 135, "usage_type": "call"}, {"api_name": "aquire.uptime", "line_number": 164, "usage_type": "call"}, {"api_name": "aquire.uptime", "line_number": 196, "usage_type": "call"}, {"api_name": "psutil.cpu_percent", "line_number": 207, "usage_type": "call"}, {"api_name": "psutil.virtual_memory", "line_number": 208, "usage_type": "call"}, {"api_name": "aquire.uptime", "line_number": 209, "usage_type": "call"}, {"api_name": "time.time", "line_number": 223, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 327, "usage_type": "call"}]}
+{"seq_id": "113504099", "text": "from django.conf import settings\nfrom django.shortcuts import render_to_response, get_object_or_404\nfrom django.template import RequestContext\nfrom django.http import HttpResponseRedirect, Http404, HttpResponse\nfrom django.db.models import Q\nfrom django.contrib.auth.models import User\nfrom django.contrib.admin.views.decorators import staff_member_required\nfrom django.utils.translation import ugettext\nfrom django.template.loader import render_to_string\n\nfrom account.utils import get_default_redirect\nfrom signup_codes.models import check_signup_code\nfrom signup_codes.forms import SignupForm, InviteUserForm\n\nfrom profiles.models import StudentProfile, MentorProfile, FieldOfInterest\nfrom forms import EventsForm\nfrom django.utils.translation import ugettext_lazy as _\n\nimport datetime\nimport profiles\nimport mail.utils\n\nif \"notification\" in settings.INSTALLED_APPS:\n from notification import models as notification\nelse:\n notification = None\n\n\n@staff_member_required\ndef admin_set_profile_visibility(request, username, visibility, template_name=\"mentorship_admin/admin_set_profile_visibility.html\", extra_context=None):\n if extra_context == None:\n extra_context = {}\n\n other_user = get_object_or_404(User,username=username)\n if visibility == 'activate':\n if other_user.is_active == False:\n other_user.is_active = True\n elif visibility == 'deactivate':\n if other_user.is_active == True:\n other_user.is_active = False\n elif visibility == 'status':\n pass\n else:\n raise Http404\n\n email = profiles.models.Profile.objects.get(user=other_user).email\n \n if other_user.is_active:\n subject = 'Contul tau a fost activat'\n else:\n subject = 'Contul tau a fost dezactivat'\n message = render_to_string('mentorship_admin/activate_message.txt', { 'active': other_user.is_active, 'username': username })\n mail.utils.mail(email, subject, message)\n\n if other_user.is_active == True:\n other_user_status = 'active'\n mail.utils.send_mail_confirm(other_user)\n else:\n other_user_status = 'inactive'\n \n\n other_user.save()\n\n return render_to_response(template_name, dict({\n 'user': other_user,\n 'status': other_user_status,\n }, **extra_context), context_instance=RequestContext(request))\n\n\n@staff_member_required\ndef admin_profiles(request, template_name=\"mentorship_admin/admin_search_profiles.html\", extra_context=None):\n if extra_context == None:\n extra_context = {}\n\n users = User.objects.all().order_by(\"-date_joined\").exclude(is_superuser=True).exclude(is_active=False)\n search_terms = request.GET.get('search', '')\n order = request.GET.get('order')\n selected_field_index = request.GET.get('field')\n if (selected_field_index == \"None\"):\n selected_field_index = None\n\n nr_fields = FieldOfInterest.objects.count()\n selected_field_of_interest = None\n if selected_field_index:\n if 1 <= int(selected_field_index) <= nr_fields:\n selected_field_of_interest = FieldOfInterest.objects.get(pk=selected_field_index).name\n\n if not order:\n order = 'name'\n if search_terms:\n if order == 'name':\n users = users.filter(\n Q(username__icontains=search_terms) |\n Q(profile__firstname__icontains=search_terms) |\n Q(profile__surname__icontains=search_terms)\n )\n elif order == 'faculty':\n users = users.filter(\n Q(profile__studentprofile__faculty__icontains=search_terms)\n )\n\n if selected_field_of_interest:\n users = users.filter(profile__fields_of_interest__field__name__iexact=selected_field_of_interest)\n # order by date\n if order == 'date':\n users = users.order_by(\"-date_joined\")\n # order by username\n elif order == 'name':\n users = users.order_by(\"username\")\n elif order == 'faculty':\n users = users.order_by('profile__studentprofile__faculty')\n elif order == 'students':\n student_list = [stud.pk for stud in StudentProfile.objects.all()]\n users = users.filter(profile__pk__in=student_list)\n elif order == 'mentors':\n mentor_list = [m.pk for m in MentorProfile.objects.all()]\n users = users.filter(profile__pk__in=mentor_list)\n\n fields_of_interest = FieldOfInterest.objects.all()\n\n return render_to_response(template_name, dict({\n 'users': users,\n 'order': order,\n 'search_terms': search_terms,\n 'fields_of_interest': fields_of_interest,\n 'field': selected_field_index,\n }, **extra_context), context_instance=RequestContext(request))\n\n\n@staff_member_required\ndef admin_invite_users(request, form_class = InviteUserForm,\n template_name=\"mentorship_admin/admin_invite_users.html\"):\n \"\"\"\n View that works inside the admin tab\n \"\"\"\n if request.method == \"POST\":\n form = form_class(request.POST)\n if form.is_valid():\n email = form.cleaned_data[\"email\"]\n form.send_signup_code()\n request.user.message_set.create(message=ugettext(\"An e-mail has been sent to %(email)s.\") % {\"email\": email})\n form = form_class() # reset\n else:\n form = form_class()\n return render_to_response(template_name, {\n \"title\": ugettext(\"Invite user\"),\n \"form\": form,\n }, context_instance = RequestContext(request))\n\ndef get_initial_form_values_from_session(request):\n ret_dict = {}\n \n ret_dict['name'] = request.session.get('events_selected_name', '')\n ret_dict['points'] = request.session.get('events_selected_points', '1')\n ret_dict['date'] = request.session.get('events_selected_date', datetime.datetime.now())\n ret_dict['location'] = request.session.get('events_selected_location', '')\n ret_dict['description'] = request.session.get('events_selected_description', '')\n \n return ret_dict\n\ndef set_form_session_values(request, data_dict):\n request.session['events_selected_name'] = data_dict.get('name', '')\n request.session['events_selected_points'] = data_dict.get('points', '')\n request.session['events_selected_date'] = data_dict.get('date', '')\n request.session['events_selected_location'] = data_dict.get('location', '')\n request.session['events_selected_description'] = data_dict.get('description', '')\n\ndef safely_delete_key(dict, key):\n if dict.has_key(key):\n del dict[key]\n\ndef cleanup_session_data(request):\n safely_delete_key(request.session, 'events_selected_name')\n safely_delete_key(request.session, 'events_selected_points')\n safely_delete_key(request.session, 'events_selected_date')\n safely_delete_key(request.session, 'events_selected_location')\n safely_delete_key(request.session, 'events_selected_description')\n \n\nfrom locations.models import EventLocation\n\n@staff_member_required\ndef admin_events(request, form_class = EventsForm,\n template_name=\"mentorship_admin/admin_events.html\"):\n \"\"\"\n View that controls the admin add events forms\n \"\"\"\n \n if request.method == 'POST' and request.is_ajax():\n set_form_session_values(request, request.POST)\n return HttpResponse(\"/locations/event_location/\")\n\n if request.method == 'POST':\n form = form_class(request.POST)\n if form.is_valid():\n event = form.save()\n \n (latitude, longitude) = request.session.get('events_selected_coordinates', '46.7667,23.6').split(\",\")\n event_location = EventLocation()\n event_location.event = event\n event_location.latitude = float(latitude)\n event_location.longitude = float(longitude)\n event_location.save()\n \n\n if notification:\n notification.send(User.objects.all(), \"profiles_new_event\", { \"event\": event }, queue=True)\n\n request.user.message_set.create(message=_(\"Event %(event)s has been created.\") % {'event':form.cleaned_data[\"name\"]})\n \n cleanup_session_data(request)\n form = form_class()\n else:\n form = form_class(initial=get_initial_form_values_from_session(request))\n\n return render_to_response(template_name, {\n \"form\": form,\n }, context_instance = RequestContext(request))\n\n", "sub_path": "src/mentorat/apps/mentorship_admin/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 8309, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.conf.settings.INSTALLED_APPS", "line_number": 23, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 23, "usage_type": "name"}, {"api_name": "notification.models", "line_number": 26, "usage_type": "name"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 34, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 34, "usage_type": "argument"}, {"api_name": "django.http.Http404", "line_number": 44, "usage_type": "name"}, {"api_name": "profiles.models.Profile.objects.get", "line_number": 46, "usage_type": "call"}, {"api_name": "profiles.models", "line_number": 46, "usage_type": "attribute"}, {"api_name": "django.template.loader.render_to_string", "line_number": 52, "usage_type": "call"}, {"api_name": "mail.utils.utils.mail", "line_number": 53, "usage_type": "call"}, {"api_name": "mail.utils.utils", "line_number": 53, "usage_type": "attribute"}, {"api_name": "mail.utils", "line_number": 53, "usage_type": "name"}, {"api_name": "mail.utils.utils.send_mail_confirm", "line_number": 57, "usage_type": "call"}, {"api_name": "mail.utils.utils", "line_number": 57, "usage_type": "attribute"}, {"api_name": "mail.utils", "line_number": 57, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 64, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 67, "usage_type": "call"}, {"api_name": "django.contrib.admin.views.decorators.staff_member_required", "line_number": 29, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.all", "line_number": 75, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 75, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 75, "usage_type": "name"}, {"api_name": "profiles.models.FieldOfInterest.objects.count", "line_number": 82, "usage_type": "call"}, {"api_name": "profiles.models.FieldOfInterest.objects", "line_number": 82, "usage_type": "attribute"}, {"api_name": "profiles.models.FieldOfInterest", "line_number": 82, "usage_type": "name"}, {"api_name": "profiles.models.FieldOfInterest.objects.get", "line_number": 86, "usage_type": "call"}, {"api_name": "profiles.models.FieldOfInterest.objects", "line_number": 86, "usage_type": "attribute"}, {"api_name": "profiles.models.FieldOfInterest", "line_number": 86, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 93, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 94, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 95, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 99, "usage_type": "call"}, {"api_name": "profiles.models.StudentProfile.objects.all", "line_number": 113, "usage_type": "call"}, {"api_name": "profiles.models.StudentProfile.objects", "line_number": 113, "usage_type": "attribute"}, {"api_name": "profiles.models.StudentProfile", "line_number": 113, "usage_type": "name"}, {"api_name": "profiles.models.MentorProfile.objects.all", "line_number": 116, "usage_type": "call"}, {"api_name": "profiles.models.MentorProfile.objects", "line_number": 116, "usage_type": "attribute"}, {"api_name": "profiles.models.MentorProfile", "line_number": 116, "usage_type": "name"}, {"api_name": "profiles.models.FieldOfInterest.objects.all", "line_number": 119, "usage_type": "call"}, {"api_name": "profiles.models.FieldOfInterest.objects", "line_number": 119, "usage_type": "attribute"}, {"api_name": "profiles.models.FieldOfInterest", "line_number": 119, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 121, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 127, "usage_type": "call"}, {"api_name": "django.contrib.admin.views.decorators.staff_member_required", "line_number": 70, "usage_type": "name"}, {"api_name": "signup_codes.forms.InviteUserForm", "line_number": 131, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext", "line_number": 141, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 145, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext", "line_number": 146, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 148, "usage_type": "call"}, {"api_name": "django.contrib.admin.views.decorators.staff_member_required", "line_number": 130, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 155, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 155, "usage_type": "attribute"}, {"api_name": "forms.EventsForm", "line_number": 183, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 191, "usage_type": "call"}, {"api_name": "locations.models.EventLocation", "line_number": 199, "usage_type": "call"}, {"api_name": "notification.models", "line_number": 206, "usage_type": "name"}, {"api_name": "notification.models.send", "line_number": 207, "usage_type": "call"}, {"api_name": "notification.models", "line_number": 207, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.all", "line_number": 207, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 207, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 207, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 209, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 216, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 218, "usage_type": "call"}, {"api_name": "django.contrib.admin.views.decorators.staff_member_required", "line_number": 182, "usage_type": "name"}]}
+{"seq_id": "529471437", "text": "import itertools as it\nfrom copy import deepcopy\nfrom functools import partial\nfrom typing import Dict, List, NamedTuple, Optional\n\nimport numpy as np\nimport pyro as p\nimport torch\nimport torch.distributions.constraints as constraints\nfrom pyro.contrib.autoname import scope\nfrom pyro.distributions import ( # pylint: disable=no-name-in-module\n Delta,\n NegativeBinomial,\n Normal,\n OneHotCategorical,\n RelaxedOneHotCategoricalStraightThrough,\n)\nfrom torch.distributions import transform_to\n\nfrom ....data import Data, Dataset\nfrom ....data.slide import DataSlide, Slide\nfrom ....data.utility.misc import make_dataloader\nfrom ....data.utility.misc import spot_size\nfrom ....logging import Progressbar, DEBUG, INFO, log\nfrom ....session import get, require\nfrom ....utility.core import center_crop\nfrom ....utility.state import (\n get_module,\n get_param,\n get_state_dict,\n load_state_dict,\n)\nfrom ....utility.tensor import checkpoint, isoftplus, sparseonehot, to_device\nfrom ..image import Image\n\n\nclass MetageneDefault(NamedTuple):\n r\"\"\"Metagene initialization template\"\"\"\n\n scale: float\n profile: Optional[torch.Tensor]\n\n\ndef _encode_metagene_name(n: str):\n return f\"!!metagene!{n}!!\"\n\n\nclass ST(Image):\n r\"\"\"Spatial Transcriptomics experiment\"\"\"\n\n @property\n def tag(self):\n return \"ST\"\n\n def __init__(\n self,\n *args,\n metagenes: Optional[List[MetageneDefault]] = None,\n **kwargs,\n ):\n super().__init__(*args, **kwargs)\n\n if metagenes is None:\n metagenes = [MetageneDefault(0.0, None)]\n\n if len(metagenes) == 0:\n raise ValueError(\"Needs at least one metagene\")\n\n self.__metagenes: Dict[str, MetageneDefault] = {}\n self.__metagene_queue: List[str] = []\n for metagene in metagenes:\n self.add_metagene(metagene)\n\n self.__init_scale = None\n self.__init_rate = None\n self.__init_logits = None\n\n @property\n def metagenes(self) -> Dict[str, MetageneDefault]:\n r\"\"\"Metagene initialization templates\"\"\"\n return deepcopy(self.__metagenes)\n\n def add_metagene(self, metagene: Optional[MetageneDefault] = None):\n r\"\"\"\n Adds a new metagene, optionally initialized from a\n :class:`MetageneDefault`.\n \"\"\"\n if metagene is None:\n metagene = MetageneDefault(0.0, None)\n\n if self.__metagene_queue != []:\n new_metagene = self.__metagene_queue.pop()\n else:\n new_metagene = f\"{len(self.__metagenes) + 1:d}\"\n assert new_metagene not in self.__metagenes\n\n log(INFO, \"Adding metagene: %s\", new_metagene)\n self.__metagenes.setdefault(new_metagene, metagene)\n\n return new_metagene\n\n def split_metagene(self, metagene: str):\n r\"\"\"Adds a new metagene by splitting an already existing metagene.\"\"\"\n new_metagene = self.add_metagene(self.metagenes[metagene])\n\n log(INFO, \"Copying metagene: %s -> %s\", metagene, new_metagene)\n\n name = _encode_metagene_name(metagene)\n new_name = _encode_metagene_name(new_metagene)\n\n state_dict = get_state_dict()\n\n for pname in [\n pname for pname in state_dict.params.keys() if name in pname\n ]:\n new_pname = pname.replace(name, new_name)\n log(DEBUG, \"Copying param: %s -> %s\", pname, new_pname)\n state_dict.params[new_pname] = deepcopy(state_dict.params[pname])\n\n for mname in [\n mname for mname in state_dict.modules.keys() if name in mname\n ]:\n new_mname = mname.replace(name, new_name)\n log(DEBUG, \"Copying module: %s -> %s\", mname, new_mname)\n state_dict.modules[new_mname] = deepcopy(state_dict.modules[mname])\n\n load_state_dict(state_dict)\n\n return new_metagene\n\n def remove_metagene(self, n, remove_params=False):\n r\"\"\"Removes a metagene\"\"\"\n if len(self.metagenes) == 1:\n raise RuntimeError(\"Cannot remove last metagene\")\n\n log(INFO, \"Removing metagene: %s\", n)\n\n try:\n self.__metagenes.pop(n)\n except KeyError as exc:\n raise ValueError(\n f\"Attempted to remove metagene {n}, which doesn't exist!\"\n ) from exc\n\n self.__metagene_queue.append(n)\n\n if remove_params:\n store = p.get_param_store()\n optim = get(\"optimizer\")\n pname = _encode_metagene_name(n)\n for x in [p for p in store.keys() if pname in p]:\n param = store[x].unconstrained()\n del store[x]\n if optim is not None:\n del optim.optim_objs[param]\n\n def __init_globals(self):\n dataloader = require(\"dataloader\")\n device = get(\"default_device\")\n\n dataloader = make_dataloader(\n Dataset(\n Data(\n slides={\n k: Slide(\n data=v.data,\n # pylint: disable=unnecessary-lambda\n # ^ Necessary for type checking to pass\n iterator=lambda x: DataSlide(x),\n )\n for k, v in dataloader.dataset.data.slides.items()\n if v.data.type == \"ST\"\n },\n design=dataloader.dataset.data.design,\n )\n ),\n num_workers=0,\n batch_size=100,\n )\n\n r2rp = transform_to(constraints.positive)\n\n scale = torch.zeros(1, requires_grad=True, device=device)\n rate = torch.zeros(\n len(dataloader.dataset.genes), requires_grad=True, device=device\n )\n logits = torch.zeros(\n len(dataloader.dataset.genes), requires_grad=True, device=device\n )\n\n optim = torch.optim.Adam((scale, rate, logits), lr=0.01)\n\n with Progressbar(it.count(1), leave=False, position=0) as iterator:\n running_rmse = None\n for epoch in iterator:\n previous_rmse = running_rmse\n for x in (\n torch.cat(x[\"ST\"][\"data\"]).to(device) for x in dataloader\n ):\n distr = NegativeBinomial(\n r2rp(scale) * r2rp(rate), logits=logits\n )\n rmse = (\n ((distr.mean - x) ** 2)\n .mean(1)\n .sqrt()\n .mean()\n .detach()\n .cpu()\n )\n try:\n running_rmse = running_rmse + 1e-2 * (\n rmse - running_rmse\n )\n except TypeError:\n running_rmse = rmse\n iterator.set_description(\n \"Initializing global coefficients, please wait...\"\n + f\" (RMSE: {running_rmse:.3f})\"\n )\n optim.zero_grad()\n nll = -distr.log_prob(x).sum()\n nll.backward()\n optim.step()\n if (epoch > 100) and (previous_rmse - running_rmse < 1e-4):\n break\n\n self.__init_scale = r2rp(scale).detach().cpu()\n self.__init_rate = r2rp(rate).detach().cpu()\n self.__init_logits = logits.detach().cpu()\n\n def __init_scale_baseline(self):\n if self.__init_scale is None:\n self.__init_globals()\n return self.__init_scale\n\n def __init_rate_baseline(self):\n if self.__init_rate is None:\n self.__init_globals()\n return self.__init_rate\n\n def __init_logits_baseline(self):\n if self.__init_logits is None:\n self.__init_globals()\n return self.__init_logits\n\n def _get_scale_decoder(self, in_channels):\n # pylint: disable=no-self-use\n def _create_scale_decoder():\n dataset = require(\"dataloader\").dataset\n decoder = torch.nn.Sequential(\n torch.nn.Conv2d(in_channels, in_channels, kernel_size=1),\n torch.nn.BatchNorm2d(in_channels, momentum=0.05),\n torch.nn.LeakyReLU(0.2, inplace=True),\n torch.nn.Conv2d(in_channels, 1, kernel_size=1),\n torch.nn.Softplus(),\n )\n torch.nn.init.normal_(decoder[-2].weight, std=1e-5)\n decoder[-2].bias.data[...] = isoftplus(\n self.__init_scale_baseline() / spot_size(dataset)[\"ST\"]\n )\n return decoder\n\n return get_module(\"scale\", _create_scale_decoder, checkpoint=True)\n\n def _create_metagene_decoder(self, in_channels, n):\n decoder = torch.nn.Sequential(\n torch.nn.Conv2d(in_channels, 1, kernel_size=1)\n )\n torch.nn.init.constant_(decoder[-1].bias, self.__metagenes[n][0])\n return decoder\n\n def model(self, x, zs):\n # pylint: disable=too-many-locals\n def _compute_rim(decoded):\n shared_representation = get_module(\n \"metagene_shared\",\n lambda: torch.nn.Sequential(\n torch.nn.Conv2d(\n decoded.shape[1], decoded.shape[1], kernel_size=1\n ),\n torch.nn.BatchNorm2d(decoded.shape[1], momentum=0.05),\n torch.nn.LeakyReLU(0.2, inplace=True),\n ),\n )(decoded)\n rim = torch.cat(\n [\n get_module(\n f\"decoder_{_encode_metagene_name(n)}\",\n partial(\n self._create_metagene_decoder, decoded.shape[1], n\n ),\n )(shared_representation)\n for n in self.metagenes\n ],\n dim=1,\n )\n rim = torch.nn.functional.softmax(rim, dim=1)\n return rim\n\n num_genes = x[\"data\"][0].shape[1]\n decoded = self._decode(zs)\n label = center_crop(x[\"label\"], [None, *decoded.shape[-2:]])\n\n rim = checkpoint(_compute_rim, decoded)\n rim = center_crop(rim, [None, None, *label.shape[-2:]])\n rim = p.sample(\"rim\", Delta(rim))\n\n scale = p.sample(\n \"scale\",\n Delta(\n center_crop(\n self._get_scale_decoder(decoded.shape[1])(decoded),\n [None, None, *label.shape[-2:]],\n )\n ),\n )\n rim = scale * rim\n\n with p.poutine.scale(scale=len(x[\"data\"]) / self.n):\n rate_mg_prior = Normal(\n 0.0,\n 1e-8\n + get_param(\n \"rate_mg_prior_sd\",\n lambda: torch.ones(num_genes),\n constraint=constraints.positive,\n ),\n )\n rate_mg = torch.stack(\n [\n p.sample(_encode_metagene_name(n), rate_mg_prior)\n for n in self.metagenes\n ]\n )\n rate_mg = p.sample(\"rate_mg\", Delta(rate_mg))\n\n rate_g_effects_baseline = get_param(\n \"rate_g_effects_baseline\",\n lambda: self.__init_rate_baseline().log(),\n lr_multiplier=5.0,\n )\n logits_g_effects_baseline = get_param(\n \"logits_g_effects_baseline\",\n # pylint: disable=unnecessary-lambda\n self.__init_logits_baseline,\n lr_multiplier=5.0,\n )\n rate_g_effects_prior = Normal(\n 0.0,\n 1e-8\n + get_param(\n \"rate_g_effects_prior_sd\",\n lambda: torch.ones(num_genes),\n constraint=constraints.positive,\n ),\n )\n rate_g_effects = p.sample(\"rate_g_effects\", rate_g_effects_prior)\n rate_g_effects = torch.cat(\n [rate_g_effects_baseline.unsqueeze(0), rate_g_effects]\n )\n logits_g_effects_prior = Normal(\n 0.0,\n 1e-8\n + get_param(\n \"logits_g_effects_prior_sd\",\n lambda: torch.ones(num_genes),\n constraint=constraints.positive,\n ),\n )\n logits_g_effects = p.sample(\n \"logits_g_effects\", logits_g_effects_prior,\n )\n logits_g_effects = torch.cat(\n [logits_g_effects_baseline.unsqueeze(0), logits_g_effects]\n )\n\n effects = []\n for covariate, vals in require(\"covariates\"):\n effect = p.sample(\n f\"effect-{covariate}\",\n OneHotCategorical(\n to_device(torch.ones(len(vals))) / len(vals)\n ),\n )\n effects.append(effect)\n effects = torch.cat(\n [to_device(torch.ones(x[\"effects\"].shape[0], 1)), *effects,], 1,\n ).float()\n\n logits_g = effects @ logits_g_effects\n rate_g = effects @ rate_g_effects\n rate_mg = rate_g[:, None] + rate_mg\n\n with scope(prefix=self.tag):\n image_distr = self._sample_image(x, decoded)\n\n def _compute_sample_params(data, label, rim, rate_mg, logits_g):\n zero_count_idxs = 1 + torch.where(data.sum(1) == 0)[0]\n partial_idxs = np.unique(\n torch.cat([label[0], label[-1], label[:, 0], label[:, -1]])\n .cpu()\n .numpy()\n )\n partial_idxs = np.setdiff1d(\n partial_idxs, zero_count_idxs.cpu().numpy()\n )\n mask = np.invert(\n np.isin(label.cpu().numpy(), [0, *partial_idxs])\n )\n mask = torch.as_tensor(mask, device=label.device)\n\n if not mask.any():\n return (\n data[[]],\n torch.zeros(0, num_genes).to(rim),\n logits_g.expand(0, -1),\n )\n\n label = label[mask] - 1\n idxs, label = torch.unique(label, return_inverse=True)\n data = data[idxs]\n\n rim = rim[:, mask]\n labelonehot = sparseonehot(label)\n rim = torch.sparse.mm(labelonehot.t().float(), rim.t())\n\n rgs = rim @ rate_mg.exp()\n\n return data, rgs, logits_g.expand(len(rgs), -1)\n\n data, rgs, logits_g = zip(\n *it.starmap(\n _compute_sample_params,\n zip(x[\"data\"], label, rim, rate_mg, logits_g),\n )\n )\n\n expression_distr = NegativeBinomial(\n total_count=1e-8 + torch.cat(rgs), logits=torch.cat(logits_g),\n )\n p.sample(\"xsg\", expression_distr, obs=torch.cat(data))\n\n return image_distr, expression_distr\n\n def _sample_globals(self):\n dataset = require(\"dataloader\").dataset\n device = get(\"default_device\")\n num_genes = len(dataset.genes)\n\n p.sample(\n \"rate_g_effects\",\n Normal(\n get_param(\n \"rate_g_effects_mu\",\n lambda: torch.zeros(\n dataset.data.design.shape[0], num_genes, device=device\n ),\n ),\n 1e-8\n + get_param(\n \"rate_g_effects_sd\",\n lambda: 1e-2\n * torch.ones(\n dataset.data.design.shape[0], num_genes, device=device\n ),\n constraint=constraints.positive,\n ),\n ),\n infer={\"is_global\": True},\n )\n\n p.sample(\n \"logits_g_effects\",\n Normal(\n get_param(\n \"logits_g_effects_mu\",\n lambda: torch.zeros(\n dataset.data.design.shape[0], num_genes, device=device\n ),\n ),\n 1e-8\n + get_param(\n \"logits_g_effects_sd\",\n lambda: 1e-2\n * torch.ones(\n dataset.data.design.shape[0], num_genes, device=device\n ),\n constraint=constraints.positive,\n ),\n ),\n infer={\"is_global\": True},\n )\n\n # Sample metagene profiles\n def _sample_metagene(metagene, name):\n mu = get_param(\n f\"{_encode_metagene_name(name)}_mu\",\n # pylint: disable=unnecessary-lambda\n lambda: metagene.profile.float(),\n lr_multiplier=2.0,\n )\n sd = get_param(\n f\"{_encode_metagene_name(name)}_sd\",\n lambda: 1e-2\n * torch.ones_like(metagene.profile, device=device).float(),\n constraint=constraints.positive,\n lr_multiplier=2.0,\n )\n if len(self.__metagenes) < 2:\n mu = mu.detach()\n sd = sd.detach()\n p.sample(\n _encode_metagene_name(name),\n Normal(mu, 1e-8 + sd),\n infer={\"is_global\": True},\n )\n\n for name, metagene in self.metagenes.items():\n if metagene.profile is None:\n metagene = MetageneDefault(\n metagene.scale, torch.zeros(num_genes)\n )\n _sample_metagene(metagene, name)\n\n def guide(self, x):\n with p.poutine.scale(scale=len(x[\"data\"]) / self.n):\n self._sample_globals()\n for covariate, _ in require(\"covariates\"):\n is_observed = x[\"effects\"][covariate].values.any(1)\n effect_distr = RelaxedOneHotCategoricalStraightThrough(\n temperature=to_device(torch.as_tensor(0.1)),\n logits=torch.stack(\n [\n get_param(\n f\"effect-{covariate}-{sample}-logits\",\n torch.zeros(len(vals)),\n )\n for sample, vals in x[\"effects\"][covariate].iterrows()\n ]\n ),\n )\n with p.poutine.mask(mask=~to_device(torch.as_tensor(is_observed))):\n effect = p.sample(f\"effect-{covariate}-all\", effect_distr)\n effect[is_observed] = torch.as_tensor(\n x[\"effects\"][covariate].values[is_observed]\n ).to(effect)\n p.sample(f\"effect-{covariate}\", Delta(effect))\n return super().guide(x)\n", "sub_path": "xfuse/model/experiment/st/st.py", "file_name": "st.py", "file_ext": "py", "file_size_in_byte": 19032, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "typing.NamedTuple", "line_number": 37, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 41, "usage_type": "name"}, {"api_name": "torch.Tensor", "line_number": 41, "usage_type": "attribute"}, {"api_name": "image.Image", "line_number": 48, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 58, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 58, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 69, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 70, "usage_type": "name"}, {"api_name": "copy.deepcopy", "line_number": 81, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 79, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 83, "usage_type": "name"}, {"api_name": "logging.log", "line_number": 97, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 97, "usage_type": "argument"}, {"api_name": "logging.log", "line_number": 106, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 106, "usage_type": "argument"}, {"api_name": "utility.state.get_state_dict", "line_number": 111, "usage_type": "call"}, {"api_name": "logging.log", "line_number": 117, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 117, "usage_type": "argument"}, {"api_name": "copy.deepcopy", "line_number": 118, "usage_type": "call"}, {"api_name": "logging.log", "line_number": 124, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 124, "usage_type": "argument"}, {"api_name": "copy.deepcopy", "line_number": 125, "usage_type": "call"}, {"api_name": "utility.state.load_state_dict", "line_number": 127, "usage_type": "call"}, {"api_name": "logging.log", "line_number": 136, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 136, "usage_type": "argument"}, {"api_name": "pyro.get_param_store", "line_number": 148, "usage_type": "call"}, {"api_name": "session.get", "line_number": 149, "usage_type": "call"}, {"api_name": "session.require", "line_number": 158, "usage_type": "call"}, {"api_name": "session.get", "line_number": 159, "usage_type": "call"}, {"api_name": "data.utility.misc.make_dataloader", "line_number": 161, "usage_type": "call"}, {"api_name": "data.Dataset", "line_number": 162, "usage_type": "call"}, {"api_name": "data.Data", "line_number": 163, "usage_type": "call"}, {"api_name": "data.slide.Slide", "line_number": 165, "usage_type": "call"}, {"api_name": "data.slide.DataSlide", "line_number": 169, "usage_type": "call"}, {"api_name": "torch.distributions.transform_to", "line_number": 181, "usage_type": "call"}, {"api_name": "torch.distributions.constraints.positive", "line_number": 181, "usage_type": "attribute"}, {"api_name": "torch.distributions.constraints", "line_number": 181, "usage_type": "name"}, {"api_name": "torch.zeros", "line_number": 183, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 184, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 187, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 191, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 191, "usage_type": "attribute"}, {"api_name": "logging.Progressbar", "line_number": 193, "usage_type": "call"}, {"api_name": "itertools.count", "line_number": 193, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 198, "usage_type": "call"}, {"api_name": "pyro.distributions.NegativeBinomial", "line_number": 200, "usage_type": "call"}, {"api_name": "session.require", "line_number": 250, "usage_type": "call"}, {"api_name": "torch.nn.Sequential", "line_number": 251, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 251, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv2d", "line_number": 252, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 252, "usage_type": "attribute"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 253, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 253, "usage_type": "attribute"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 254, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 254, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv2d", "line_number": 255, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 255, "usage_type": "attribute"}, {"api_name": "torch.nn.Softplus", "line_number": 256, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 256, "usage_type": "attribute"}, {"api_name": "torch.nn.init.normal_", "line_number": 258, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 258, "usage_type": "attribute"}, {"api_name": "utility.tensor.isoftplus", "line_number": 259, "usage_type": "call"}, {"api_name": "data.utility.misc.spot_size", "line_number": 260, "usage_type": "call"}, {"api_name": "utility.state.get_module", "line_number": 264, "usage_type": "call"}, {"api_name": "torch.nn.Sequential", "line_number": 267, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 267, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv2d", "line_number": 268, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 268, "usage_type": "attribute"}, {"api_name": "torch.nn.init.constant_", "line_number": 270, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 270, "usage_type": "attribute"}, {"api_name": "utility.state.get_module", "line_number": 276, "usage_type": "call"}, {"api_name": "torch.nn.Sequential", "line_number": 278, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 278, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv2d", "line_number": 279, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 279, "usage_type": "attribute"}, {"api_name": "torch.nn.BatchNorm2d", "line_number": 282, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 282, "usage_type": "attribute"}, {"api_name": "torch.nn.LeakyReLU", "line_number": 283, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 283, "usage_type": "attribute"}, {"api_name": "torch.cat", "line_number": 286, "usage_type": "call"}, {"api_name": "utility.state.get_module", "line_number": 288, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 290, "usage_type": "call"}, {"api_name": "torch.nn.functional.softmax", "line_number": 298, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 298, "usage_type": "attribute"}, {"api_name": "utility.core.center_crop", "line_number": 303, "usage_type": "call"}, {"api_name": "utility.tensor.checkpoint", "line_number": 305, "usage_type": "call"}, {"api_name": "utility.core.center_crop", "line_number": 306, "usage_type": "call"}, {"api_name": "pyro.sample", "line_number": 307, "usage_type": "call"}, {"api_name": "pyro.distributions.Delta", "line_number": 307, "usage_type": "call"}, {"api_name": "pyro.sample", "line_number": 309, "usage_type": "call"}, {"api_name": "pyro.distributions.Delta", "line_number": 311, "usage_type": "call"}, {"api_name": "utility.core.center_crop", "line_number": 312, "usage_type": "call"}, {"api_name": "pyro.poutine.scale", "line_number": 320, "usage_type": "call"}, {"api_name": "pyro.poutine", "line_number": 320, "usage_type": "attribute"}, {"api_name": "pyro.distributions.Normal", "line_number": 321, "usage_type": "call"}, {"api_name": "utility.state.get_param", "line_number": 324, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 326, "usage_type": "call"}, {"api_name": "torch.distributions.constraints.positive", "line_number": 327, "usage_type": "attribute"}, {"api_name": "torch.distributions.constraints", "line_number": 327, "usage_type": "name"}, {"api_name": "torch.stack", "line_number": 330, "usage_type": "call"}, {"api_name": "pyro.sample", "line_number": 332, "usage_type": "call"}, {"api_name": "pyro.sample", "line_number": 336, "usage_type": "call"}, {"api_name": "pyro.distributions.Delta", "line_number": 336, "usage_type": "call"}, {"api_name": "utility.state.get_param", "line_number": 338, "usage_type": "call"}, {"api_name": "utility.state.get_param", "line_number": 343, "usage_type": "call"}, {"api_name": "pyro.distributions.Normal", "line_number": 349, "usage_type": "call"}, {"api_name": "utility.state.get_param", "line_number": 352, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 354, "usage_type": "call"}, {"api_name": "torch.distributions.constraints.positive", "line_number": 355, "usage_type": "attribute"}, {"api_name": "torch.distributions.constraints", "line_number": 355, "usage_type": "name"}, {"api_name": "pyro.sample", "line_number": 358, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 359, "usage_type": "call"}, {"api_name": "pyro.distributions.Normal", "line_number": 362, "usage_type": "call"}, {"api_name": "utility.state.get_param", "line_number": 365, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 367, "usage_type": "call"}, {"api_name": "torch.distributions.constraints.positive", "line_number": 368, "usage_type": "attribute"}, {"api_name": "torch.distributions.constraints", "line_number": 368, "usage_type": "name"}, {"api_name": "pyro.sample", "line_number": 371, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 374, "usage_type": "call"}, {"api_name": "session.require", "line_number": 379, "usage_type": "call"}, {"api_name": "pyro.sample", "line_number": 380, "usage_type": "call"}, {"api_name": "pyro.distributions.OneHotCategorical", "line_number": 382, "usage_type": "call"}, {"api_name": "utility.tensor.to_device", "line_number": 383, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 383, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 387, "usage_type": "call"}, {"api_name": "utility.tensor.to_device", "line_number": 388, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 388, "usage_type": "call"}, {"api_name": "pyro.contrib.autoname.scope", "line_number": 395, "usage_type": "call"}, {"api_name": "torch.where", "line_number": 399, "usage_type": "call"}, {"api_name": "data.sum", "line_number": 399, "usage_type": "call"}, {"api_name": "numpy.unique", "line_number": 400, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 401, "usage_type": "call"}, {"api_name": "numpy.setdiff1d", "line_number": 405, "usage_type": "call"}, {"api_name": "numpy.invert", "line_number": 408, "usage_type": "call"}, {"api_name": "numpy.isin", "line_number": 409, "usage_type": "call"}, {"api_name": "torch.as_tensor", "line_number": 411, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 416, "usage_type": "call"}, {"api_name": "torch.unique", "line_number": 421, "usage_type": "call"}, {"api_name": "utility.tensor.sparseonehot", "line_number": 425, "usage_type": "call"}, {"api_name": "torch.sparse.mm", "line_number": 426, "usage_type": "call"}, {"api_name": "torch.sparse", "line_number": 426, "usage_type": "attribute"}, {"api_name": "itertools.starmap", "line_number": 433, "usage_type": "call"}, {"api_name": "pyro.distributions.NegativeBinomial", "line_number": 439, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 440, "usage_type": "call"}, {"api_name": "pyro.sample", "line_number": 442, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 442, "usage_type": "call"}, {"api_name": "session.require", "line_number": 447, "usage_type": "call"}, {"api_name": "session.get", "line_number": 448, "usage_type": "call"}, {"api_name": "pyro.sample", "line_number": 451, "usage_type": "call"}, {"api_name": "pyro.distributions.Normal", "line_number": 453, "usage_type": "call"}, {"api_name": "utility.state.get_param", "line_number": 454, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 456, "usage_type": "call"}, {"api_name": "utility.state.get_param", "line_number": 461, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 464, "usage_type": "call"}, {"api_name": "torch.distributions.constraints.positive", "line_number": 467, "usage_type": "attribute"}, {"api_name": "torch.distributions.constraints", "line_number": 467, "usage_type": "name"}, {"api_name": "pyro.sample", "line_number": 473, "usage_type": "call"}, {"api_name": "pyro.distributions.Normal", "line_number": 475, "usage_type": "call"}, {"api_name": "utility.state.get_param", "line_number": 476, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 478, "usage_type": "call"}, {"api_name": "utility.state.get_param", "line_number": 483, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 486, "usage_type": "call"}, {"api_name": "torch.distributions.constraints.positive", "line_number": 489, "usage_type": "attribute"}, {"api_name": "torch.distributions.constraints", "line_number": 489, "usage_type": "name"}, {"api_name": "utility.state.get_param", "line_number": 497, "usage_type": "call"}, {"api_name": "utility.state.get_param", "line_number": 503, "usage_type": "call"}, {"api_name": "torch.ones_like", "line_number": 506, "usage_type": "call"}, {"api_name": "torch.distributions.constraints.positive", "line_number": 507, "usage_type": "attribute"}, {"api_name": "torch.distributions.constraints", "line_number": 507, "usage_type": "name"}, {"api_name": "pyro.sample", "line_number": 513, "usage_type": "call"}, {"api_name": "pyro.distributions.Normal", "line_number": 515, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 522, "usage_type": "call"}, {"api_name": "pyro.poutine.scale", "line_number": 527, "usage_type": "call"}, {"api_name": "pyro.poutine", "line_number": 527, "usage_type": "attribute"}, {"api_name": "session.require", "line_number": 529, "usage_type": "call"}, {"api_name": "pyro.distributions.RelaxedOneHotCategoricalStraightThrough", "line_number": 531, "usage_type": "call"}, {"api_name": "utility.tensor.to_device", "line_number": 532, "usage_type": "call"}, {"api_name": "torch.as_tensor", "line_number": 532, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 533, "usage_type": "call"}, {"api_name": "utility.state.get_param", "line_number": 535, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 537, "usage_type": "call"}, {"api_name": "pyro.poutine.mask", "line_number": 543, "usage_type": "call"}, {"api_name": "pyro.poutine", "line_number": 543, "usage_type": "attribute"}, {"api_name": "utility.tensor.to_device", "line_number": 543, "usage_type": "call"}, {"api_name": "torch.as_tensor", "line_number": 543, "usage_type": "call"}, {"api_name": "pyro.sample", "line_number": 544, "usage_type": "call"}, {"api_name": "torch.as_tensor", "line_number": 545, "usage_type": "call"}, {"api_name": "pyro.sample", "line_number": 548, "usage_type": "call"}, {"api_name": "pyro.distributions.Delta", "line_number": 548, "usage_type": "call"}]}
+{"seq_id": "516280401", "text": "import click\n\nfrom app import assemble\n\n\n@click.command()\n@click.option(\n \"--asm-file\",\n required=True,\n help=\"The source assembly file to be translated\",\n)\ndef cli(asm_file: str) -> None:\n click.echo(f\"Hacking <{asm_file}>\")\n assemble(asm_file)\n click.echo(\"Done!\")\n\n\nif __name__ == \"__main__\":\n cli()\n", "sub_path": "Nand2Tetris/assignment6/n2t-Assembler/assembler.py", "file_name": "assembler.py", "file_ext": "py", "file_size_in_byte": 324, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "click.echo", "line_number": 13, "usage_type": "call"}, {"api_name": "app.assemble", "line_number": 14, "usage_type": "call"}, {"api_name": "click.echo", "line_number": 15, "usage_type": "call"}, {"api_name": "click.command", "line_number": 6, "usage_type": "call"}, {"api_name": "click.option", "line_number": 7, "usage_type": "call"}]}
+{"seq_id": "258606162", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\nimport os\n\nfrom django.conf import settings\n\nfrom Crypto.Hash import SHA256\nfrom Crypto.PublicKey import RSA\nfrom Crypto import Random\n\n\nSECRET_KEY = settings.SECRET_KEY\nPATH_KEY_DIR = os.path.join(settings.MEDIA_ROOT, 'keys')\n\n\ndef sign(in_file):\n \"\"\"電子署名を行う\n\n :param in_file:\n :return:\n \"\"\"\n generate_keys()\n with open(get_private_key_path(), b'r') as f:\n rsa = RSA.importKey(f.read(), SECRET_KEY)\n digest = get_file_checksum(in_file)\n signature = rsa.sign(digest, '')[0]\n return str(signature)\n\n\ndef verify(in_file, signature):\n \"\"\"ファイルを検証する、署名にあっているかどうかをチェックする。\n\n :param in_file:\n :param signature:\n :return:\n \"\"\"\n with open(get_public_key_path(), b'r') as f:\n rsa = RSA.importKey(f.read(), SECRET_KEY)\n digest = get_file_checksum(in_file)\n is_verify = rsa.verify(digest, (long(signature),))\n return is_verify\n\n\ndef get_file_checksum(filename):\n \"\"\"電子署名用のダイジェスト情報を取得する\n\n :param filename:\n :return:\n \"\"\"\n h = SHA256.new()\n chunk_size = 8192\n with open(filename, b'rb') as f:\n while True:\n chunk = f.read(chunk_size)\n if len(chunk) == 0:\n break\n h.update(chunk)\n return h.digest()\n\n\ndef get_private_key_path():\n \"\"\"秘密鍵の格納場所を取得する。\n\n :return:\n \"\"\"\n if not os.path.exists(PATH_KEY_DIR):\n os.mkdir(PATH_KEY_DIR)\n return os.path.join(PATH_KEY_DIR, 'signature_private.pem')\n\n\ndef get_public_key_path():\n \"\"\"公開鍵の格納場所を取得する。\n\n :return:\n \"\"\"\n if not os.path.exists(PATH_KEY_DIR):\n os.mkdir(PATH_KEY_DIR)\n return os.path.join(PATH_KEY_DIR, 'signature_public.pem')\n\n\ndef generate_keys():\n path_private = get_private_key_path()\n path_public = get_public_key_path()\n # 既に作成済みの場合、再作成不要\n if os.path.exists(path_private) and os.path.exists(path_public):\n return\n random_generator = Random.new().read\n rsa = RSA.generate(1024, random_generator)\n # 秘密鍵作成\n private_pem = rsa.exportKey(format='PEM', passphrase=SECRET_KEY)\n with open(path_private, b'w') as f:\n f.write(private_pem)\n # 公開鍵作成\n public_pem = rsa.publickey().exportKey()\n with open(path_public, b'w') as f:\n f.write(public_pem)\n", "sub_path": "utils/signature.py", "file_name": "signature.py", "file_ext": "py", "file_size_in_byte": 2517, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.conf.settings.SECRET_KEY", "line_number": 12, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 12, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "attribute"}, {"api_name": "django.conf.settings.MEDIA_ROOT", "line_number": 13, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 13, "usage_type": "name"}, {"api_name": "Crypto.PublicKey.RSA.importKey", "line_number": 24, "usage_type": "call"}, {"api_name": "Crypto.PublicKey.RSA", "line_number": 24, "usage_type": "name"}, {"api_name": "Crypto.PublicKey.RSA.importKey", "line_number": 38, "usage_type": "call"}, {"api_name": "Crypto.PublicKey.RSA", "line_number": 38, "usage_type": "name"}, {"api_name": "Crypto.Hash.SHA256.new", "line_number": 50, "usage_type": "call"}, {"api_name": "Crypto.Hash.SHA256", "line_number": 50, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 68, "usage_type": "call"}, {"api_name": "os.path", "line_number": 68, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path", "line_number": 78, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "Crypto.Random.new", "line_number": 87, "usage_type": "call"}, {"api_name": "Crypto.Random", "line_number": 87, "usage_type": "name"}, {"api_name": "Crypto.PublicKey.RSA.generate", "line_number": 88, "usage_type": "call"}, {"api_name": "Crypto.PublicKey.RSA", "line_number": 88, "usage_type": "name"}]}
+{"seq_id": "82332560", "text": "import re\nimport json\n\nfrom bs4 import BeautifulSoup\nfrom decimal import Decimal\n\nfrom storescraper.product import Product\nfrom storescraper.store import Store\nfrom storescraper.utils import session_with_proxy, html_to_markdown\n\n\nclass Comandato(Store):\n @classmethod\n def categories(cls):\n return [\n 'Oven',\n 'Refrigerator',\n 'WashingMachine',\n 'AirConditioner',\n 'Television',\n 'Cell',\n 'Tablet',\n 'StereoSystem',\n 'CellAccesory',\n 'Headphones',\n 'Wearable',\n 'MemoryCard'\n ]\n\n @classmethod\n def discover_urls_for_category(cls, category, extra_args=None):\n category_paths = [\n ['ft=cocinas', 'Oven'],\n ['fq=C:/1000001/1000034/', 'Refrigerator'],\n ['fq=C:/1000001/1000039/', 'WashingMachine'],\n ['fq=C:/1000002/', 'AirConditioner'],\n ['fq=C:/1000007/1000008/1000057/', 'Television'],\n ['fq=C:/1000007/1000090/', 'Cell']\n ]\n\n session = session_with_proxy(extra_args)\n product_urls = []\n\n for category_path, local_category in category_paths:\n if local_category != category:\n continue\n\n page = 1\n\n while True:\n if page > 10:\n raise Exception('Page overflow')\n\n url = 'https://www.comandato.com/buscapagina?PS=24&' \\\n 'sl=5fd2e9cb-dc33-4655-95e2-fc62e15a859a&cc=4&' \\\n 'sm=0&{}&PageNumber={}'.format(\n category_path, page)\n\n print(url)\n\n soup = BeautifulSoup(session.get(url).text, 'html.parser')\n products = soup.findAll('div', 'producto')\n\n if not products:\n if page == 1:\n raise Exception('Empty url {}'.format(url))\n else:\n break\n\n for product in products:\n if product.find('h3').find('strong').text != 'LG':\n continue\n product_url = product.find('a')['href']\n product_urls.append(product_url)\n\n page += 1\n\n return product_urls\n\n @classmethod\n def products_for_url(cls, url, category=None, extra_args=None):\n print(url)\n session = session_with_proxy(extra_args)\n data = session.get(url).text\n soup = BeautifulSoup(data, 'html.parser')\n\n name = soup.find('div', 'productDescriptionShort').text\n sku = soup.find('div', 'skuReference').text\n stock = 0\n if soup.find('link', {'itemprop': 'availability'})['href'] == \\\n 'http://schema.org/InStock':\n stock = -1\n\n pricing_data = re.search(r'vtex.events.addData\\(([\\S\\s]+?)\\);',\n data).groups()[0]\n pricing_data = json.loads(pricing_data)\n\n tax = Decimal('1.12')\n offer_price = Decimal(pricing_data['productPriceFrom'])*tax\n normal_price = Decimal(pricing_data['productListPriceFrom'])*tax\n\n picture_urls = [a['zoom'] for a in\n soup.findAll('a', {'id': 'botaoZoom'})]\n\n description = html_to_markdown(\n str(soup.find('div', {'id': 'caracteristicas'})))\n\n p = Product(\n name,\n cls.__name__,\n category,\n url,\n url,\n sku,\n stock,\n normal_price,\n offer_price,\n 'USD',\n sku=sku,\n picture_urls=picture_urls,\n description=description,\n )\n\n return [p]\n", "sub_path": "storescraper/stores/comandato.py", "file_name": "comandato.py", "file_ext": "py", "file_size_in_byte": 3736, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "storescraper.store.Store", "line_number": 12, "usage_type": "name"}, {"api_name": "storescraper.utils.session_with_proxy", "line_number": 41, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 61, "usage_type": "call"}, {"api_name": "storescraper.utils.session_with_proxy", "line_number": 83, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 85, "usage_type": "call"}, {"api_name": "re.search", "line_number": 94, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 96, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 98, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 99, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 100, "usage_type": "call"}, {"api_name": "storescraper.utils.html_to_markdown", "line_number": 105, "usage_type": "call"}, {"api_name": "storescraper.product.Product", "line_number": 108, "usage_type": "call"}]}
+{"seq_id": "132393789", "text": "import time\nfrom selenium import webdriver\nfrom selenium.webdriver.support.wait import WebDriverWait\nfrom selenium.webdriver.support import expected_conditions as EC\nfrom selenium.webdriver.common.by import By\nimport datetime\nimport sys\n\n# 配置抢购时间\nset_time = datetime.datetime(2021,3,15,21,50)\n\noption = webdriver.ChromeOptions()\noption.add_argument('disable-infobars')\ndriver = webdriver.Chrome(chrome_options=option)\n\ndriver.get(\"https://www.taobao.com\")\ndriver.find_element_by_link_text(\"亲,请登录\").click()\ntime.sleep(1)\nWebDriverWait(driver, 60).until(lambda x:not x.find_elements(\"id\",\"login\"))\nprint('登陆完成')\n\n# 检查时间\nwhile(1):\n now = datetime.datetime.now()\n if now > set_time:\n break\n\ndriver.get(\"https://cart.taobao.com/cart.htm\")\ndriver.find_element_by_id(\"J_SelectAll1\").click()\n\nWebDriverWait(driver, 10).until(lambda x:x.find_element_by_link_text('结 算'))\nWebDriverWait(driver, 10).until(lambda x:'disabled' not in x.find_element_by_link_text('结 算').get_attribute('class'))\ndriver.find_element_by_link_text(\"结 算\").click()\n\nWebDriverWait(driver, 10).until(lambda x:x.find_element_by_link_text('提交订单'))\ndriver.find_element_by_link_text(\"提交订单\").click()", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1234, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "datetime.datetime", "line_number": 10, "usage_type": "call"}, {"api_name": "selenium.webdriver.ChromeOptions", "line_number": 12, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 12, "usage_type": "name"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 14, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 14, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 18, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.wait.WebDriverWait", "line_number": 19, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 24, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 24, "usage_type": "attribute"}, {"api_name": "selenium.webdriver.support.wait.WebDriverWait", "line_number": 31, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.wait.WebDriverWait", "line_number": 32, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.wait.WebDriverWait", "line_number": 35, "usage_type": "call"}]}
+{"seq_id": "517080876", "text": "from django.http import HttpResponse\nfrom django.template import loader\nfrom django.utils.translation import gettext_lazy as _\n\nfrom hier.utils import get_base_context, save_folder_id\nfrom trip.models import trip_summary\n\n#----------------------------------\n# Index\n#----------------------------------\ndef index(request):\n title = ''\n context = get_base_context(request, 0, 0, title, 'content_list')\n\n if request.user.is_authenticated:\n save_folder_id(request.user, 0)\n title = _('applications')\n hide_title = False\n else:\n title = context['site_header']\n hide_title = True\n\n context['title'] = title\n context['hide_title'] = hide_title\n context['trip_summary'] = trip_summary(request.user.id)\n template = loader.get_template('index.html')\n return HttpResponse(template.render(context, request))\n\n#----------------------------------\n# Feedback\n#----------------------------------\ndef feedback(request):\n context = get_base_context(request, 0, 0, _('feedback'))\n template = loader.get_template('feedback.html')\n return HttpResponse(template.render(context, request))\n\n", "sub_path": "rusel/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1138, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "hier.utils.get_base_context", "line_number": 13, "usage_type": "call"}, {"api_name": "hier.utils.save_folder_id", "line_number": 16, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 17, "usage_type": "call"}, {"api_name": "trip.models.trip_summary", "line_number": 25, "usage_type": "call"}, {"api_name": "django.template.loader.get_template", "line_number": 26, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 26, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 27, "usage_type": "call"}, {"api_name": "hier.utils.get_base_context", "line_number": 33, "usage_type": "call"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 33, "usage_type": "call"}, {"api_name": "django.template.loader.get_template", "line_number": 34, "usage_type": "call"}, {"api_name": "django.template.loader", "line_number": 34, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 35, "usage_type": "call"}]}
+{"seq_id": "220555597", "text": "#!/usr/bin/env python\n# vim: ai ts=4 sts=4 et sw=4\n\nfrom django.utils.translation import ugettext as _\nfrom django.http import HttpResponse\nfrom django.template.loader import render_to_string\nfrom datetime import datetime, date, timedelta\n\n\"\"\"Shared report methods. These are for domains\n that share forms (or have very similar forms),\n for example, mvp and grameen share the safe\n pregnancy forms.\"\"\"\n \n# custom hacks for the branching have been tagged \"ANNOYING\"\n# it's still better to share code here, I think.\n \ndef monitoring_report(request, case):\n context = { }\n # use the cases we've put together for this\n data_maps = case.get_all_data_maps()\n \n # allow a list of usernames whose submissions don't show up\n # in the report. \n # TODO: move this to the real blacklist now\n blacklist = [\"teddy\", \"admin\", \"demo_user\"]\n blacklist_columns = [\"meta_username\"]\n \n final_list_of_maps = {}\n # do a once-through pruning by the blacklist, and reclassifying by \n # both chw id and case id\n for id, map in data_maps.items():\n for form, list_of_forms in map.items():\n for form_instance in list_of_forms:\n # check blacklist and don't use this instance if it's in \n # the blacklist\n if _is_blacklisted(form_instance, blacklist, blacklist_columns):\n continue\n # make a new id that includes the \n new_id = \"%s-%s\" % (form_instance[\"meta_username\"], id) \n if new_id not in final_list_of_maps:\n # initialize the new id with an empty list for each form\n this_id_map = {}\n for form_all in map:\n this_id_map[form_all] = []\n final_list_of_maps[new_id] = this_id_map\n final_list_of_maps[new_id][form].append(form_instance)\n\n all_moms = []\n healthy_moms = []\n very_pregnant_moms = []\n moms_needing_followup = []\n moms_with_open_referrals = []\n closed_moms = []\n for id, map in final_list_of_maps.items():\n mom = Mother(case, id, map)\n if not mom.chw:\n # don't include submissions from non-users\n continue\n all_moms.append(mom)\n prev_list_size = len(moms_needing_followup) +\\\n len(very_pregnant_moms) +\\\n len(moms_with_open_referrals)\n \n # we don't want the closed cases showing up on any\n # other lists \n if mom.is_closed:\n closed_moms.append(mom)\n continue\n if mom.needs_followup:\n moms_needing_followup.append(mom)\n if mom.months_pregnant >= 7:\n very_pregnant_moms.append(mom)\n if mom.has_open_referral:\n moms_with_open_referrals.append(mom)\n new_list_size = len(moms_needing_followup) +\\\n len(very_pregnant_moms) +\\\n len(moms_with_open_referrals)\n if new_list_size == prev_list_size:\n # we didn't add her to any lists so put her in\n # healthy moms\n healthy_moms.append(mom) \n \n \n context[\"domain\"] = case.domain\n context[\"all_moms\"] = all_moms\n context[\"healthy_moms\"] = healthy_moms\n context[\"open_referrals\"] = moms_with_open_referrals\n context[\"very_pregnant\"] = very_pregnant_moms\n context[\"need_followup\"] = moms_needing_followup\n context[\"closed_moms\"] = closed_moms\n context[\"empty_data_holder\"] = \"\"\n return render_to_string(\"custom/shared/monitoring.html\", context)\n\n\ndef _is_blacklisted(data, blacklist, blacklist_columns):\n '''Checks a set of columns and values, and if any of the\n columns contains one of the values, returns true'''\n for column in blacklist_columns:\n if column in data and data[column] in blacklist:\n return True\n return False\n\nclass Mother(object):\n \n def __init__(self, case, id, data_map):\n self.case = case\n self.id = id\n self.data_map = data_map \n # calculate some properties and set them for easy access\n # these are totally hard coded to the forms. \n # most of these depend on registration and will not display\n # very well if there is no registration\n forms = case.form_identifiers\n # ANNOYING: mvp has 5 forms cause it includes old reg. \n # grameen does not\n if len(forms) == 5:\n [new_reg_forms, followup_forms, close_forms, referrals, old_reg_forms] =\\\n [data_map[form] for form in forms]\n elif len(forms) == 4:\n [new_reg_forms, followup_forms, close_forms, referrals] =\\\n [data_map[form] for form in forms]\n old_reg_forms = []\n else:\n raise Exception(\"Need 4 or 5 forms for this case report, but found: %s\" % len(forms))\n # check against the new and old registration form, in that order\n if new_reg_forms:\n reg_form_data = new_reg_forms[0]\n elif old_reg_forms:\n reg_form_data = old_reg_forms[0]\n else:\n reg_form_data = None\n \n # set the registration data, if present\n if reg_form_data: \n self.mother_name = reg_form_data[\"sampledata_mother_name\"]\n self.date_of_reg = reg_form_data[\"meta_timestart\"]\n \n # ANNOYING: another branch because Grameen uses \"weeks\"\n if \"sampledata_months_pregnant\" in reg_form_data:\n self.months_preg_at_reg = reg_form_data[\"sampledata_months_pregnant\"]\n else:\n weeks_pregnant = reg_form_data[\"sampledata_weeks_pregnant\"]\n self.months_preg_at_reg = weeks_pregnant / 4 \n \n if self.date_of_reg and self.months_preg_at_reg:\n days_pregnant_at_reg = self.months_preg_at_reg * 30\n self.months_pregnant = ((datetime.now() - self.date_of_reg) + \n timedelta(days=days_pregnant_at_reg)).days / 30\n else:\n self.months_pregnant = None\n # high risk factors\n hi_risk_cols = [\"sampledata_hi_risk_info_old\",\n \"sampledata_hi_risk_info_young\",\n \"sampledata_hi_risk_info\",\n \"sampledata_hi_risk_info_education\",\n \"sampledata_hi_risk_info_small\",\n \"sampledata_hi_risk_info_10_years\",\n \"sampledata_hi_risk_info_5_years\",\n \"sampledata_hi_risk_info_complications\",\n \"sampledata_hi_risk_info_blood_group\",\n \"sampledata_hi_risk_info_many\",\n \"sampledata_hi_risk_info_hepb\",\n \"sampledata_hi_risk_info_anemia\",\n \"sampledata_hi_risk_info_health\",\n \"sampledata_hi_risk_info_hiv\",\n \"sampledata_hi_risk_info_syphilis\"]\n hi_risk_values = []\n for col in hi_risk_cols:\n # ANNOYING: we have to check the existence of the column, since\n # not all these are in both forms. \n if col in reg_form_data and reg_form_data[col]:\n risk_factor = self._clean_and_translate(col, \"sampledata_hi_risk_info_\", \"\")\n hi_risk_values.append(risk_factor)\n self.high_risk_factors = \",\".join(hi_risk_values)\n \n # for the rest of the properties, just add them as top level properties\n # available in the model, unless they've already been set\n for col, value in reg_form_data.items():\n prop_name = col.replace(\"sampledata_\", \"\")\n if not hasattr(self, prop_name):\n setattr(self, prop_name, value)\n else:\n self.months_pregnant = None\n chw_col = \"meta_username\"\n # loop through all forms searching for this.\n self.chw = None\n for form in forms:\n if data_map[form]:\n for sub_map in data_map[form]:\n if sub_map[chw_col]:\n self.chw = sub_map[chw_col]\n break\n if self.chw:\n break\n \n # set followup data\n if followup_forms:\n self.has_followup = True\n self.date_of_last_followup = followup_forms[0][\"meta_timestart\"] \n \n # checklist items come from the most recent followup\n checklist_items = [\"safe_pregnancy_preg_actions_iron_folic\",\n \"safe_pregnancy_preg_actions_start_tt\",\n \"safe_pregnancy_preg_actions_finish_tt\",\n \"safe_pregnancy_preg_actions_start_ipt\",\n \"safe_pregnancy_preg_actions_finish_ipt\",\n \"safe_pregnancy_preg_actions_deworm\",\n \"safe_pregnancy_preg_actions_birth_plan\",\n \"safe_pregnancy_preg_actions_test_hiv\",\n \"safe_pregnancy_preg_actions_test_syphilis\",\n \"safe_pregnancy_preg_actions_test_bp\",\n \"safe_pregnancy_preg_actions_test_hb\"]\n \n incomplete_checklist_items = []\n for item in checklist_items:\n if followup_forms[0][item] != 1:\n incomplete_checklist_items.append(self._clean_and_translate(item, \"safe_pregnancy_preg_actions_\", \"\"))\n \n self.incomplete_checklist_items = \", \".join(incomplete_checklist_items)\n else:\n self.incomplete_checklist_items = _(\"No followup visits found.\")\n \n # Women Needing Followup \n # > 1 month since last Followup if 1-6 Months, \n # > 15 days since last Followup if 7-10 Months\n if followup_forms and self.date_of_last_followup:\n self.days_since_followup = (datetime.now() - self.date_of_last_followup).days\n # this is yucky but we don't want to include folks with no followups\n # and no new reg, since we're not checkign the old followups\n elif new_reg_forms:\n self.days_since_followup = (datetime.now() - self.date_of_reg).days\n else:\n # no reg or follow-ups. leave them out for now\n self.days_since_followup = None\n self.needs_followup = False\n if self.days_since_followup is not None:\n if self.days_since_followup > 30:\n self.needs_followup = True\n elif self.days_since_followup > 15\\\n and self.months_pregnant and self.months_pregnant >= 7:\n self.needs_followup = True\n else:\n self.needs_followup = False\n \n # referrals\n self.has_followup_referral = False\n if followup_forms:\n # referrals from followup\n for followup in followup_forms:\n if not self.has_followup_referral:\n if followup[\"safe_pregnancy_referred\"] == \"yes\":\n self.has_followup_referral = True\n self.most_recent_followup_referral_id = followup[\"safe_pregnancy_referral_id\"]\n self.followup_referred = followup\n elif followup[\"safe_pregnancy_referral_id\"] == self.most_recent_followup_referral_id:\n # older instance of the same referral. Use this as the instance \n # they were referred\n self.followup_referred = followup\n \n \n self.has_close_referral = False\n self.is_closed = False\n if close_forms:\n self.is_closed = True\n # add some fields that we display in the \"closed\" table\n most_recent_closure = close_forms[0]\n self.closed_outcome = most_recent_closure[\"safe_pregnancy_what_happened\"]\n self.mother_survived = most_recent_closure[\"safe_pregnancy_mother_survived\"]\n self.birth_location = most_recent_closure[\"safe_pregnancy_birth_location\"]\n self.children_registered = most_recent_closure[\"safe_pregnancy_infants_registered\"]\n self.children_survived = most_recent_closure[\"safe_pregnancy_infants_survived\"]\n # referrals from close forms\n for close in close_forms:\n if not self.has_close_referral:\n if close[\"safe_pregnancy_referred\"] == \"yes\":\n self.has_close_referral = True\n self.most_recent_close_referral_id = close[\"safe_pregnancy_referral_id\"]\n self.close_referred = close\n elif close[\"safe_pregnancy_referral_id\"] == self.most_recent_close_referral_id:\n # older instance of the same referral. Use this as the instance \n # they were referred\n self.close_referred = close\n \n self.has_referral = self.has_followup_referral or self.has_close_referral \n if self.has_referral:\n self.date_referred = datetime.min\n if self.has_followup_referral:\n self.date_referred = self.followup_referred[\"meta_timeend\"]\n self.most_recent_referral_id = self.most_recent_followup_referral_id \n followup_wins = True\n if self.has_close_referral:\n self.date_referred = max(self.date_referred, self.close_referred[\"meta_timeend\"])\n if self.date_referred == self.close_referred[\"meta_timeend\"]:\n followup_wins = False\n self.most_recent_referral_id = self.most_recent_close_referral_id \n \n # danger signs for the referral - pull from the visit that generated\n # it\n danger_signs = []\n if followup_wins:\n # These are the values that can be set in the followup form\n # to generate a referral\n followup_warnings = {\"safe_pregnancy_feeling\" : \"not_well\", \n \"safe_pregnancy_pain_from_vagina\": \"yes\",\n \"safe_pregnancy_headache_or_b_vision\": \"yes\",\n \"safe_pregnancy_dark_urine\": \"yes\",\n \"safe_pregnancy_swelling\": \"yes\",\n \"safe_pregnancy_unusual_pain\": \"yes\",\n \"safe_pregnancy_burn_urinate\": \"yes\",\n \"safe_pregnancy_baby_not_moving\": \"yes\",\n \"safe_pregnancy_fever\": \"yes\",\n \"safe_pregnancy_other_illness\" : \"yes\"}\n \n for key, value in followup_warnings.items():\n if self.followup_referred[key] == value:\n danger_signs.append(\"%s: %s\" % (self._clean_and_translate(key, \"safe_pregnancy_\", \"\"),\n self._clean_and_translate(value, \"\", \"\")))\n # add which illness\n if \"other illness: yes\" in danger_signs and self.followup_referred[\"safe_pregnancy_which_illness\"]:\n danger_signs.remove(\"other illness: yes\")\n danger_signs.append(\"other illness: yes (%s)\" % self.followup_referred[\"safe_pregnancy_which_illness\"])\n \n else: \n # Close form is referred:\n if self.close_referred[\"safe_pregnancy_mother_survived\"] == \"yes\":\n closure_warnings = {\n \"safe_pregnancy_why_not\" : 'skeptical_clinic',\n \"safe_pregnancy_why_not\" : 'busy',\n \"safe_pregnancy_why_not\" : 'transport',\n \"safe_pregnancy_why_not\" : 'other',\n \"safe_pregnancy_treatment_why_not\" : 'no_medicine',\n \"safe_pregnancy_treatment_why_not\" : 'too_many_patients',\n \"safe_pregnancy_treatment_why_not\" : 'no_doctor',\n \"safe_pregnancy_treatment_why_not\" : 'no_health_workers',\n \"safe_pregnancy_treatment_why_not\" : 'other'\n }\n for key, value in closure_warnings.items():\n if self.close_referred[key] == value:\n danger_signs.append(\"%s: %s\" % (self._clean_and_translate(key, \"safe_pregnancy_\", \"\"),\n self._clean_and_translate(value, \"\", \"\")))\n \n self.danger_signs = \", \".join(danger_signs)\n \n else:\n self.date_referred = None\n \n if referrals:\n # we found a referral\n self.date_of_last_referral = referrals[0][\"meta_timestart\"] \n if referrals[0][\"safe_pregnancy_why_not\"] != \"feeling_better\" and\\\n referrals[0][\"safe_pregnancy_treatment\"] != \"yes\":\n # the referral is open no matter what if the referral \n # is present but not completed \n self.has_open_referral = True\n elif self.date_referred is None or self.date_of_last_referral > self.date_referred:\n # they completed it. \n self.has_open_referral = False\n else:\n # they closed a referral but have a more recent one \n # that's still open\n self.has_open_referral = True\n else:\n # no information about a referral. if there was one it's open\n self.has_open_referral = self.has_referral\n\n\n def _clean_and_translate(self, column, prefix, suffix):\n '''Cleans a column by removing a prefix and suffix, if they\n are present, and converting underscores to spaces.\n Also does translation of the data, if possible.'''\n if column.startswith(prefix):\n column = column[len(prefix):]\n if column.endswith(suffix):\n column = column[0:len(column) - len(suffix)] \n return _(column).replace(\"_\", \" \")", "sub_path": "apps/reports/custom/shared.py", "file_name": "shared.py", "file_ext": "py", "file_size_in_byte": 18632, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.template.loader.render_to_string", "line_number": 92, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 147, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 147, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 148, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext", "line_number": 221, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 227, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 227, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 231, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 231, "usage_type": "name"}, {"api_name": "datetime.datetime.min", "line_number": 286, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 286, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext", "line_number": 375, "usage_type": "call"}]}
+{"seq_id": "130794796", "text": "from django.shortcuts import render, redirect, get_object_or_404\nfrom django.http import HttpResponse\nfrom .forms import FoodFitnessModel, FoodFitnessForm\nfrom django.contrib.auth.models import User\n\n\n# Create your views here.\n#function renders to the index page\ndef index(request):\n return render(request, \"foodFitnessApp/index.html\")\n\n#function adds a new user and renders to the add user page\ndef addUser(request):\n form = FoodFitnessForm(request.POST or None)\n context = {\n \"form\": form\n }\n return render(request, \"foodFitnessApp/addUser.html\", context)\n\n#my function that prints the post made and create a new user in the admin . once submitted it will post the httpresponse\n# listed\ndef info(request):\n print(request.POST)\n User.objects.create_user(request.POST[\"userName\"], \"\", \"\")\n\n return HttpResponse(\"you gonna be skinny\")\n", "sub_path": "foodFitnessProject/foodFitnessApp/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 865, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.shortcuts.render", "line_number": 10, "usage_type": "call"}, {"api_name": "forms.FoodFitnessForm", "line_number": 14, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 18, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.create_user", "line_number": 24, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 24, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 24, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 26, "usage_type": "call"}]}
+{"seq_id": "267228586", "text": "from keras.models import Sequential\nfrom keras.layers import Dense\nfrom keras.layers import LSTM\nfrom keras.layers import Embedding\nfrom keras.utils import to_categorical\nfrom numpy import array\nfrom pickle import dump\nfrom keras.preprocessing.text import Tokenizer\n\nclass Generator:\n \"\"\"\n Generiert ein Modell mit den angegebenen Parametern:\n *vocab_size* Größe des Vokabulars\n *seq* Sequenz (fürs Embedding)\n *seq_length* Sequenzlänge\n *epochs* Epochen fürs Training\n *LSTM_neurons1/LSTM_neurons2* Anzahl Knoten in der 1. bzw. 2. LSTM-Layer\n *Dense_neurons* Anzahl Knoten in Dense-Layer\n *activation1/activation2* activation functions für die Dense-Layer\n \"\"\"\n def __init__(self, vocab_size, seq, seq_length, epochs, LSTM_neurons1, LSTM_neurons2, Dense_neurons,\n activation1, activation2):\n self.parameters = {\"epochs\": epochs, \"LSTM_neurons1\" : LSTM_neurons1, \"LSTM_neurons2\" : LSTM_neurons2,\n \"Dense_neurons\" : Dense_neurons, \"activation1\" : activation1,\n \"activation2\" : activation2}\n self.model = Sequential()\n self.model.add(Embedding(vocab_size, seq, input_length=seq_length))\n self.model.add(LSTM(self.parameters[\"LSTM_neurons1\"], return_sequences=True))\n self.model.add(LSTM(self.parameters[\"LSTM_neurons2\"]))\n self. model.add(Dense(self.parameters[\"Dense_neurons\"], activation=self.parameters[\"activation1\"]))\n self. model.add(Dense(vocab_size, activation=self.parameters[\"activation2\"]))\n self.model.compile(loss=\"categorical_crossentropy\", optimizer=\"adam\", metrics=[\"accuracy\"])\n #self.model.summary()\n\n def __str__(self):\n return str(self.parameters.items())\n\n def create_model(self):\n return self.model\n\n\ndef load_doc(filename):\n # open the file as read only\n file = open(filename, 'r')\n # read all text\n text = file.read()\n # close the file\n file.close()\n return text\n\n\n\"\"\"\ninput_data = \"sequence/German_animaltales_sequences.txt\"\n\nwith open(input_data) as f:\n doc = load_doc(input_data)\n lines = doc.split('\\n')\n # print(stories)\n # letztes Element der Sequenzen ist '', deswegen :-1\n # integer encode sequences of words\n tokenizer = Tokenizer() # create Tokenizer for encoding\n tokenizer.fit_on_texts(lines)\n # train it on the data -> it finds all unique words and assigns each an integer\n sequences = tokenizer.texts_to_sequences(lines)\n # make a list of integer out of each list of words\n vocab_size = len(tokenizer.word_index) + 1\n sequences = array(sequences)\n X, y = sequences[:, :-1], sequences[:, -1]\n y = to_categorical(y, num_classes=vocab_size)\n seq_length = X.shape[1]\n\nneurons1 = 64\nneurons2 = 147\nneurons3 = 91\nactivation1 = \"elu\"\nactivation2 = \"softmax\"\nmodel1 = Generator(vocab_size, 50, seq_length, 85, neurons1, neurons2, neurons3, activation1, activation2)\nmodel = model1.create_model()\nmodel.summary()\nmodel_name = \"models/german_animaltales_model\"\nprint(\"model created.\")\nmodel.fit(X, y, batch_size=128, epochs=85)\nmodel.save(model_name+\".h5\")\ndump(tokenizer, open(\"tokenizer/german_animaltales_tokenizer.pkl\", \"wb\"))\n\"\"\"\n", "sub_path": "generierung/model_generator.py", "file_name": "model_generator.py", "file_ext": "py", "file_size_in_byte": 3218, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "keras.models.Sequential", "line_number": 26, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 27, "usage_type": "call"}, {"api_name": "keras.layers.LSTM", "line_number": 28, "usage_type": "call"}, {"api_name": "keras.layers.LSTM", "line_number": 29, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 30, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 31, "usage_type": "call"}]}
+{"seq_id": "133041796", "text": "\"\"\"\nEvaluation Metrics for Image Retrieval.\n\"\"\"\n\n# import numpy as np\nimport torch\nfrom .metric import EvalMetric\n\n__all__ = ['PointDetectionMeanResidual']\n\n\nclass PointDetectionMeanResidual(EvalMetric):\n \"\"\"\n Computes mean residual for point detection.\n\n Parameters\n ----------\n axis : int, default 1\n The axis that represents classes\n name : str, default 'accuracy'\n Name of this metric instance for display.\n output_names : list of str, or None, default None\n Name of predictions that should be used when updating with update_dict.\n By default include all predictions.\n label_names : list of str, or None, default None\n Name of labels that should be used when updating with update_dict.\n By default include all labels.\n \"\"\"\n def __init__(self,\n axis=1,\n name=\"pt_det_mean_res\",\n output_names=None,\n label_names=None):\n super(PointDetectionMeanResidual, self).__init__(\n name,\n axis=axis,\n output_names=output_names,\n label_names=label_names,\n has_global_stats=True)\n self.axis = axis\n\n def update_alt(self,\n homography,\n src_pts,\n dst_pts,\n src_confs,\n dst_confs,\n src_img_size,\n dst_img_size):\n \"\"\"\n Updates the internal evaluation result.\n\n Parameters\n ----------\n homography : torch.Tensor\n Homography (from source image to destination one).\n src_pts : torch.Tensor\n Detected points for the first (source) image.\n dst_pts : torch.Tensor\n Detected points for the second (destination) image.\n src_confs : torch.Tensor\n Confidences for detected points on the source image.\n dst_confs : torch.Tensor\n Confidences for detected points on the destination image.\n src_img_size : tuple of 2 int\n Size (H, W) of the source image.\n dst_img_size : tuple of 2 int\n Size (H, W) of the destination image.\n \"\"\"\n # from scipy.spatial.distance import pdist\n from scipy.optimize import linear_sum_assignment\n\n print(\"src_img_size={}\".format(src_img_size))\n print(\"dst_img_size={}\".format(dst_img_size))\n\n self.normalize_homography(homography)\n homography_inv = self.calc_homography_inv(homography)\n\n print(\"homography={}\".format(homography))\n print(\"homography_inv={}\".format(homography_inv))\n\n print(\"src_pts.shape={}\".format(src_pts.shape))\n print(\"dst_pts.shape={}\".format(dst_pts.shape))\n print(\"src_pts={}\".format(src_pts[:10, :].int()))\n print(\"dst_pts={}\".format(dst_pts[:10, :].int()))\n\n # with torch.no_grad():\n src_hmg_pts = self.calc_homogeneous_coords(src_pts.float())\n dst_hmg_pts = self.calc_homogeneous_coords(dst_pts.float())\n\n print(\"src_hmg_pts={}\".format(src_hmg_pts[:10, :].int()))\n print(\"dst_hmg_pts={}\".format(dst_hmg_pts[:10, :].int()))\n\n self.filter_inside_points(\n src_hmg_pts,\n src_confs,\n homography,\n dst_img_size)\n self.filter_inside_points(\n dst_hmg_pts,\n dst_confs,\n homography_inv,\n src_img_size)\n\n print(\"src_hmg_pts.shape={}\".format(src_hmg_pts.shape))\n print(\"dst_hmg_pts.shape={}\".format(dst_hmg_pts.shape))\n\n src_pts_count = src_hmg_pts.shape[0]\n dst_pts_count = dst_hmg_pts.shape[0]\n pts_count = min(src_pts_count, dst_pts_count, 100)\n assert (pts_count > 0)\n src_hmg_pts = self.filter_best_points(\n src_hmg_pts,\n src_confs,\n pts_count)\n dst_hmg_pts = self.filter_best_points(\n dst_hmg_pts,\n dst_confs,\n pts_count)\n\n preds_dst_hmg_pts = self.transform_points(\n src_hmg_pts,\n homography)\n\n cost = torch.pairwise_distance(x1=preds_dst_hmg_pts, x2=dst_hmg_pts).cpu().detach().numpy()\n row_ind, col_ind = linear_sum_assignment(cost)\n mean_resudual = cost[row_ind, col_ind].sum()\n mean_resudual *= (100.0 / dst_img_size[0])\n\n self.sum_metric += mean_resudual\n self.global_sum_metric += mean_resudual\n self.num_inst += 1\n self.global_num_inst += 1\n\n @staticmethod\n def normalize_homography(homography):\n homography /= homography[2, 2]\n\n @staticmethod\n def calc_homography_inv(homography):\n homography_inv = homography.inverse()\n PointDetectionMeanResidual.normalize_homography(homography_inv)\n return homography_inv\n\n @staticmethod\n def calc_homogeneous_coords(pts):\n hmg_pts = torch.cat((pts, torch.ones((pts.shape[0], 1), dtype=pts.dtype, device=pts.device)), dim=1)\n return hmg_pts\n\n @staticmethod\n def calc_cartesian_coords(hmg_pts):\n pts = hmg_pts[:, :2]\n return pts\n\n @staticmethod\n def transform_points(src_hmg_pts,\n homography):\n # print(\"transform_points -> src_hmg_pts.shape={}\".format(src_hmg_pts.shape))\n # print(\"transform_points -> homography.shape={}\".format(homography.shape))\n\n print(\"homography={}\".format(homography))\n print(\"transform_points -> src_hmg_pts={}\".format(src_hmg_pts[:10, :].int()))\n\n dst_hmg_pts = torch.matmul(src_hmg_pts, homography.t())\n\n print(\"transform_points -> dst_hmg_pts={}\".format(dst_hmg_pts[:10, :].int()))\n # print(\"transform_points -> dst_hmg_pts.shape={}\".format(dst_hmg_pts.shape))\n\n dst_hmg_pts /= dst_hmg_pts[:, 2:]\n return dst_hmg_pts\n\n @staticmethod\n def calc_inside_pts_mask(pts,\n img_size):\n eps = 1e-3\n border_size = 1.0\n border = border_size - eps\n mask = (pts[:, 0] >= border) & (pts[:, 0] < img_size[0] - border) &\\\n (pts[:, 1] >= border) & (pts[:, 1] < img_size[1] - border)\n return mask\n\n @staticmethod\n def filter_inside_points(src_hmg_pts,\n src_confs,\n homography_inv,\n dst_img_size):\n print(\"fip->src_hmg_pts.shape={}\".format(src_hmg_pts.shape))\n print(\"fip->src_hmg_pts={}\".format(src_hmg_pts[:10, :].int()))\n print(\"fip->src_confs.shape={}\".format(src_confs.shape))\n print(\"fip->src_confs={}\".format(src_confs[:10]))\n print(\"homography_inv={}\".format(homography_inv))\n\n dst_hmg_pts = PointDetectionMeanResidual.transform_points(src_hmg_pts, homography_inv)\n\n print(\"fip->dst_hmg_pts.shape={}\".format(dst_hmg_pts.shape))\n print(\"fip->dst_hmg_pts={}\".format(dst_hmg_pts[:10, :]))\n\n mask = PointDetectionMeanResidual.calc_inside_pts_mask(dst_hmg_pts, dst_img_size)\n\n print(\"fip->mask={}\".format(mask[:10]))\n print(\"fip->mask.sum()={}\".format(mask.sum()))\n\n src_hmg_pts[:] = src_hmg_pts[mask]\n src_confs[:] = src_confs[mask]\n\n @staticmethod\n def filter_best_points(hmg_pts,\n confs,\n max_count):\n inds = confs.argsort(descending=True)[:max_count]\n return hmg_pts[inds]\n", "sub_path": "pytorch/metrics/ret_metrics.py", "file_name": "ret_metrics.py", "file_ext": "py", "file_size_in_byte": 7384, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "metric.EvalMetric", "line_number": 12, "usage_type": "name"}, {"api_name": "torch.pairwise_distance", "line_number": 125, "usage_type": "call"}, {"api_name": "scipy.optimize.linear_sum_assignment", "line_number": 126, "usage_type": "call"}, {"api_name": "{'linear_sum_assignment': 'scipy.optimize.linear_sum_assignment'}.normalize_homography", "line_number": 142, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 147, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 147, "usage_type": "call"}, {"api_name": "torch.matmul", "line_number": 164, "usage_type": "call"}, {"api_name": "{'linear_sum_assignment': 'scipy.optimize.linear_sum_assignment'}.transform_points", "line_number": 193, "usage_type": "call"}, {"api_name": "{'linear_sum_assignment': 'scipy.optimize.linear_sum_assignment'}.calc_inside_pts_mask", "line_number": 198, "usage_type": "call"}]}
+{"seq_id": "162837443", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nimport uuid, apiai\n\nCLIENT_ACCESS_TOKEN = '315682749ad04e5d9f440ec082d19e8f'\n\ndef main():\n ai = apiai.ApiAI(CLIENT_ACCESS_TOKEN)\n\n request = ai.event_request(apiai.events.Event(\"hi\"))\n request.lang = 'en' # optional, default value equal 'en'\n request.session_id = str(uuid.uuid4())\n response = request.getresponse()\n\n print (response.read())\n\n\nif __name__ == '__main__':\n main()\n", "sub_path": "Files/ApiAi/sendEvent.py", "file_name": "sendEvent.py", "file_ext": "py", "file_size_in_byte": 448, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "apiai.ApiAI", "line_number": 9, "usage_type": "call"}, {"api_name": "apiai.events.Event", "line_number": 11, "usage_type": "call"}, {"api_name": "apiai.events", "line_number": 11, "usage_type": "attribute"}, {"api_name": "uuid.uuid4", "line_number": 13, "usage_type": "call"}]}
+{"seq_id": "151565922", "text": "import os\nimport shutil\nimport unittest\nimport unittest.mock\nfrom contextlib import suppress\n\nimport numpy as np\nfrom ConfigSpace.hyperparameters import UniformFloatHyperparameter\nfrom ConfigSpace.util import get_one_exchange_neighbourhood\n\nfrom smac.callbacks import IncorporateRunResultCallback\nfrom smac.configspace import ConfigurationSpace\nfrom smac.epm.random_epm import RandomEPM\nfrom smac.epm.random_forest.rf_mo import MultiObjectiveRandomForest\nfrom smac.epm.random_forest.rf_with_instances import RandomForestWithInstances\nfrom smac.epm.utils import get_rng\nfrom smac.facade.smac_ac_facade import SMAC4AC\nfrom smac.initial_design.default_configuration_design import DefaultConfiguration\nfrom smac.initial_design.factorial_design import FactorialInitialDesign\nfrom smac.initial_design.initial_design import InitialDesign\nfrom smac.initial_design.latin_hypercube_design import LHDesign\nfrom smac.initial_design.random_configuration_design import RandomConfigurations\nfrom smac.initial_design.sobol_design import SobolDesign\nfrom smac.intensification.hyperband import Hyperband\nfrom smac.intensification.intensification import Intensifier\nfrom smac.intensification.successive_halving import SuccessiveHalving\nfrom smac.optimizer.acquisition import EI, EIPS, LCB\nfrom smac.optimizer.configuration_chooser.random_chooser import (\n ChooserNoCoolDown,\n ChooserProb,\n)\nfrom smac.runhistory.runhistory import RunHistory\nfrom smac.runhistory.runhistory2epm import (\n RunHistory2EPM4Cost,\n RunHistory2EPM4EIPS,\n RunHistory2EPM4LogCost,\n)\nfrom smac.scenario.scenario import Scenario\nfrom smac.tae import StatusType\nfrom smac.tae.execute_func import ExecuteTAFuncDict\n\n__copyright__ = \"Copyright 2021, AutoML.org Freiburg-Hannover\"\n__license__ = \"3-clause BSD\"\n\n\nclass TestSMACFacade(unittest.TestCase):\n def setUp(self):\n self.cs = ConfigurationSpace()\n self.scenario_dict_default = {\n \"cs\": self.cs,\n \"run_obj\": \"quality\",\n \"output_dir\": \"\",\n \"limit_resources\": True,\n \"deterministic\": False,\n }\n self.scenario = Scenario(self.scenario_dict_default)\n self.sh_intensifier_kwargs = {\n \"n_seeds\": 1,\n \"initial_budget\": 1,\n \"eta\": 3,\n \"min_chall\": 1,\n \"max_budget\": 100,\n }\n self.output_dirs = []\n\n def tearDown(self):\n for i in range(20):\n with suppress(Exception):\n dirname = \"run_1\" + (\".OLD\" * i)\n shutil.rmtree(dirname)\n for output_dir in self.output_dirs:\n if output_dir:\n shutil.rmtree(output_dir, ignore_errors=True)\n\n ####################################################################################################################\n # Test that the objects are constructed correctly\n\n def test_pass_callable(self):\n # Check that SMAC accepts a callable as target algorithm and that it is\n # correctly wrapped with ExecuteTaFunc\n def target_algorithm(conf, inst):\n return 5\n\n smac = SMAC4AC(tae_runner=target_algorithm, scenario=self.scenario)\n self.assertIsInstance(smac.solver.tae_runner, ExecuteTAFuncDict)\n self.assertIs(smac.solver.tae_runner.ta, target_algorithm)\n\n def test_pass_invalid_tae_runner(self):\n self.assertRaisesRegex(\n TypeError,\n \"Argument 'tae_runner' is , but must be either None, a callable or an \"\n \"object implementing BaseRunner.\",\n SMAC4AC,\n tae_runner=1,\n scenario=self.scenario,\n )\n\n def test_pass_tae_runner_objective(self):\n self.assertRaisesRegex(\n ValueError,\n \"Objective for the target algorithm runner and the scenario must be the same, but are 'runtime' and \"\n \"'quality'\",\n SMAC4AC,\n tae_runner=lambda: 1,\n tae_runner_kwargs={\"run_obj\": \"runtime\"},\n scenario=self.scenario,\n )\n\n def test_construct_runhistory2epm(self):\n \"\"\"Check default setup up for consistency\"\"\"\n smbo = SMAC4AC(self.scenario)\n self.assertTrue(type(smbo.solver.epm_chooser.rh2EPM) == RunHistory2EPM4Cost)\n self.assertSetEqual(\n set(smbo.solver.epm_chooser.rh2EPM.success_states),\n {StatusType.SUCCESS, StatusType.CRASHED, StatusType.MEMOUT},\n )\n self.assertSetEqual(set(smbo.solver.epm_chooser.rh2EPM.impute_state), set())\n self.assertSetEqual(set(smbo.solver.epm_chooser.rh2EPM.consider_for_higher_budgets_state), set())\n\n for intensifier in (SuccessiveHalving, Hyperband):\n smbo = SMAC4AC(\n self.scenario,\n intensifier=intensifier,\n intensifier_kwargs=self.sh_intensifier_kwargs,\n )\n self.assertTrue(type(smbo.solver.epm_chooser.rh2EPM) == RunHistory2EPM4Cost)\n self.assertSetEqual(\n set(smbo.solver.epm_chooser.rh2EPM.success_states),\n {\n StatusType.SUCCESS,\n StatusType.CRASHED,\n StatusType.MEMOUT,\n StatusType.DONOTADVANCE,\n },\n )\n self.assertSetEqual(set(smbo.solver.epm_chooser.rh2EPM.impute_state), set())\n self.assertSetEqual(\n set(smbo.solver.epm_chooser.rh2EPM.consider_for_higher_budgets_state),\n set(\n [\n StatusType.DONOTADVANCE,\n StatusType.TIMEOUT,\n StatusType.CRASHED,\n StatusType.MEMOUT,\n ]\n ),\n )\n\n self.scenario.run_obj = \"runtime\"\n smbo = SMAC4AC(self.scenario)\n self.assertTrue(type(smbo.solver.epm_chooser.rh2EPM) == RunHistory2EPM4LogCost)\n self.assertSetEqual(\n set(smbo.solver.epm_chooser.rh2EPM.success_states),\n {\n StatusType.SUCCESS,\n },\n )\n self.assertSetEqual(\n set(smbo.solver.epm_chooser.rh2EPM.impute_state),\n {\n StatusType.CAPPED,\n },\n )\n self.assertSetEqual(set(smbo.solver.epm_chooser.rh2EPM.consider_for_higher_budgets_state), set())\n\n def test_construct_runhistory(self):\n\n smbo = SMAC4AC(self.scenario)\n self.assertIsInstance(smbo.solver.runhistory, RunHistory)\n self.assertFalse(smbo.solver.runhistory.overwrite_existing_runs)\n smbo = SMAC4AC(self.scenario, runhistory_kwargs={\"overwrite_existing_runs\": True})\n self.assertIsInstance(smbo.solver.runhistory, RunHistory)\n self.assertTrue(smbo.solver.runhistory.overwrite_existing_runs)\n smbo = SMAC4AC(self.scenario, runhistory=RunHistory)\n self.assertIsInstance(smbo.solver.runhistory, RunHistory)\n\n def test_construct_random_configuration_chooser(self):\n rng = np.random.RandomState(42)\n smbo = SMAC4AC(self.scenario)\n self.assertIsInstance(smbo.solver.epm_chooser.random_configuration_chooser, ChooserProb)\n self.assertIsNot(smbo.solver.epm_chooser.random_configuration_chooser, rng)\n smbo = SMAC4AC(self.scenario, rng=rng)\n self.assertIsInstance(smbo.solver.epm_chooser.random_configuration_chooser, ChooserProb)\n self.assertIs(smbo.solver.epm_chooser.random_configuration_chooser.rng, rng)\n smbo = SMAC4AC(self.scenario, random_configuration_chooser_kwargs={\"rng\": rng})\n self.assertIsInstance(smbo.solver.epm_chooser.random_configuration_chooser, ChooserProb)\n self.assertIs(smbo.solver.epm_chooser.random_configuration_chooser.rng, rng)\n smbo = SMAC4AC(self.scenario, random_configuration_chooser_kwargs={\"prob\": 0.1})\n self.assertIsInstance(smbo.solver.epm_chooser.random_configuration_chooser, ChooserProb)\n self.assertEqual(smbo.solver.epm_chooser.random_configuration_chooser.prob, 0.1)\n smbo = SMAC4AC(\n self.scenario,\n random_configuration_chooser=ChooserNoCoolDown,\n random_configuration_chooser_kwargs={\"modulus\": 10},\n )\n self.assertIsInstance(smbo.solver.epm_chooser.random_configuration_chooser, ChooserNoCoolDown)\n # Check for construction failure on wrong argument\n with self.assertRaisesRegex(Exception, \"got an unexpected keyword argument\"):\n SMAC4AC(self.scenario, random_configuration_chooser_kwargs={\"dummy\": 0.1})\n\n def test_construct_epm(self):\n rng = np.random.RandomState(42)\n smbo = SMAC4AC(self.scenario)\n self.assertIsInstance(smbo.solver.epm_chooser.model, RandomForestWithInstances)\n smbo = SMAC4AC(self.scenario, rng=rng)\n self.assertIsInstance(smbo.solver.epm_chooser.model, RandomForestWithInstances)\n self.assertEqual(smbo.solver.epm_chooser.model.seed, 1935803228)\n smbo = SMAC4AC(self.scenario, model_kwargs={\"seed\": 2})\n self.assertIsInstance(smbo.solver.epm_chooser.model, RandomForestWithInstances)\n self.assertEqual(smbo.solver.epm_chooser.model.seed, 2)\n smbo = SMAC4AC(self.scenario, model_kwargs={\"num_trees\": 20})\n self.assertIsInstance(smbo.solver.epm_chooser.model, RandomForestWithInstances)\n self.assertEqual(smbo.solver.epm_chooser.model.rf_opts.num_trees, 20)\n smbo = SMAC4AC(self.scenario, model=RandomEPM, model_kwargs={\"seed\": 2})\n self.assertIsInstance(smbo.solver.epm_chooser.model, RandomEPM)\n self.assertEqual(smbo.solver.epm_chooser.model.seed, 2)\n # Check for construction failure on wrong argument\n with self.assertRaisesRegex(Exception, \"got an unexpected keyword argument\"):\n SMAC4AC(self.scenario, model_kwargs={\"dummy\": 0.1})\n\n def test_construct_acquisition_function(self):\n rng = np.random.RandomState(42)\n smbo = SMAC4AC(self.scenario)\n self.assertIsInstance(smbo.solver.epm_chooser.acquisition_func, EI)\n smbo = SMAC4AC(self.scenario, rng=rng)\n self.assertIsInstance(smbo.solver.epm_chooser.acquisition_func.model, RandomForestWithInstances)\n self.assertEqual(smbo.solver.epm_chooser.acquisition_func.model.seed, 1935803228)\n smbo = SMAC4AC(self.scenario, acquisition_function_kwargs={\"par\": 17})\n self.assertIsInstance(smbo.solver.epm_chooser.acquisition_func, EI)\n self.assertEqual(smbo.solver.epm_chooser.acquisition_func.par, 17)\n smbo = SMAC4AC(\n self.scenario,\n acquisition_function=LCB,\n acquisition_function_kwargs={\"par\": 19},\n )\n self.assertIsInstance(smbo.solver.epm_chooser.acquisition_func, LCB)\n self.assertEqual(smbo.solver.epm_chooser.acquisition_func.par, 19)\n # Check for construction failure on wrong argument\n with self.assertRaisesRegex(Exception, \"got an unexpected keyword argument\"):\n SMAC4AC(self.scenario, acquisition_function_kwargs={\"dummy\": 0.1})\n\n def test_construct_intensifier(self):\n class DummyIntensifier(Intensifier):\n pass\n\n rng = np.random.RandomState(42)\n smbo = SMAC4AC(self.scenario)\n self.assertIsInstance(smbo.solver.intensifier, Intensifier)\n self.assertIsNot(smbo.solver.intensifier.rs, rng)\n smbo = SMAC4AC(self.scenario, rng=rng)\n self.assertIsInstance(smbo.solver.intensifier, Intensifier)\n self.assertIs(smbo.solver.intensifier.rs, rng)\n smbo = SMAC4AC(self.scenario, intensifier_kwargs={\"maxR\": 987})\n self.assertEqual(smbo.solver.intensifier.maxR, 987)\n smbo = SMAC4AC(\n self.scenario,\n intensifier=DummyIntensifier,\n intensifier_kwargs={\"maxR\": 987},\n )\n self.assertIsInstance(smbo.solver.intensifier, DummyIntensifier)\n self.assertEqual(smbo.solver.intensifier.maxR, 987)\n\n dummy_intensifier = DummyIntensifier(stats=None, traj_logger=None, rng=rng, instances=self.scenario.train_insts)\n smbo = SMAC4AC(self.scenario, intensifier=dummy_intensifier)\n self.assertEqual(smbo.solver.intensifier, dummy_intensifier)\n\n # Assert that minR, maxR and use_ta_time propagate from scenario to the default intensifier.\n for scenario_dict in [\n {},\n {\n \"minR\": self.scenario.minR + 1,\n \"maxR\": self.scenario.maxR + 1,\n \"use_ta_time\": not self.scenario.use_ta_time,\n },\n ]:\n for k, v in self.scenario_dict_default.items():\n if k not in scenario_dict:\n scenario_dict[k] = v\n scenario = Scenario(scenario_dict)\n smac = SMAC4AC(scenario=scenario)\n self.assertEqual(scenario.minR, smac.solver.intensifier.minR)\n self.assertEqual(scenario.maxR, smac.solver.intensifier.maxR)\n self.assertEqual(scenario.use_ta_time, smac.solver.intensifier.use_ta_time_bound)\n\n def test_construct_initial_design(self):\n\n rng = np.random.RandomState(42)\n smbo = SMAC4AC(self.scenario)\n self.assertIsInstance(smbo.solver.initial_design, DefaultConfiguration)\n self.assertIsNot(smbo.solver.intensifier.rs, rng)\n smbo = SMAC4AC(self.scenario, rng=rng)\n self.assertIsInstance(smbo.solver.intensifier, Intensifier)\n self.assertIs(smbo.solver.intensifier.rs, rng)\n smbo = SMAC4AC(self.scenario, intensifier_kwargs={\"maxR\": 987})\n self.assertEqual(smbo.solver.intensifier.maxR, 987)\n smbo = SMAC4AC(\n self.scenario,\n initial_design=InitialDesign,\n initial_design_kwargs={\"configs\": \"dummy\"},\n )\n self.assertIsInstance(smbo.solver.initial_design, InitialDesign)\n self.assertEqual(smbo.solver.initial_design.configs, \"dummy\")\n\n for initial_incumbent_string, expected_instance in (\n (\"DEFAULT\", DefaultConfiguration),\n (\"RANDOM\", RandomConfigurations),\n (\"LHD\", LHDesign),\n (\"FACTORIAL\", FactorialInitialDesign),\n (\"SOBOL\", SobolDesign),\n ):\n self.scenario.initial_incumbent = initial_incumbent_string\n smbo = SMAC4AC(self.scenario)\n self.assertIsInstance(smbo.solver.initial_design, expected_instance)\n\n def test_init_EIPS_as_arguments(self):\n for objective in [\"runtime\", \"quality\"]:\n self.scenario.run_obj = objective\n smbo = SMAC4AC(\n self.scenario,\n model=MultiObjectiveRandomForest,\n model_kwargs={\"target_names\": [\"a\", \"b\"], \"model_kwargs\": {\"seed\": 1}},\n acquisition_function=EIPS,\n runhistory2epm=RunHistory2EPM4EIPS,\n ).solver\n self.assertIsInstance(\n smbo.epm_chooser.model,\n MultiObjectiveRandomForest,\n )\n self.assertIsInstance(smbo.epm_chooser.acquisition_func, EIPS)\n self.assertIsInstance(\n smbo.epm_chooser.acquisition_func.model,\n MultiObjectiveRandomForest,\n )\n self.assertIsInstance(smbo.epm_chooser.rh2EPM, RunHistory2EPM4EIPS)\n\n with self.assertRaisesRegex(TypeError, \"the surrogate model must support multi-objective prediction!\"):\n SMAC4AC(self.scenario, acquisition_function=EIPS, runhistory2epm=RunHistory2EPM4EIPS)\n\n ####################################################################################################################\n # Other tests...\n\n @unittest.mock.patch.object(SMAC4AC, \"__init__\")\n def test_check_random_states(self, patch):\n patch.return_value = None\n smac = SMAC4AC()\n smac.logger = unittest.mock.MagicMock()\n\n # Check some properties\n # Check whether different seeds give different random states\n _, rng_1 = get_rng(1)\n _, rng_2 = get_rng(2)\n self.assertNotEqual(sum(rng_1.get_state()[1] - rng_2.get_state()[1]), 0)\n\n # Check whether no seeds gives different random states\n _, rng_1 = get_rng(logger=smac.logger)\n self.assertEqual(smac.logger.debug.call_count, 1)\n _, rng_2 = get_rng(logger=smac.logger)\n self.assertEqual(smac.logger.debug.call_count, 2)\n\n self.assertNotEqual(sum(rng_1.get_state()[1] - rng_2.get_state()[1]), 0)\n\n # Check whether the same int seeds give the same random states\n _, rng_1 = get_rng(1)\n _, rng_2 = get_rng(1)\n self.assertEqual(sum(rng_1.get_state()[1] - rng_2.get_state()[1]), 0)\n\n # Check all execution paths\n self.assertRaisesRegex(\n TypeError,\n \"Argument rng accepts only arguments of type None, int or np.random.RandomState, \"\n \"you provided .\",\n get_rng,\n rng=\"ABC\",\n )\n self.assertRaisesRegex(\n TypeError,\n \"Argument run_id accepts only arguments of type None, int, you provided .\",\n get_rng,\n run_id=\"ABC\",\n )\n\n run_id, rng_1 = get_rng(rng=None, run_id=None, logger=smac.logger)\n self.assertIsInstance(run_id, int)\n self.assertIsInstance(rng_1, np.random.RandomState)\n self.assertEqual(smac.logger.debug.call_count, 3)\n\n run_id, rng_1 = get_rng(rng=None, run_id=1, logger=smac.logger)\n self.assertEqual(run_id, 1)\n self.assertIsInstance(rng_1, np.random.RandomState)\n\n run_id, rng_1 = get_rng(rng=1, run_id=None, logger=smac.logger)\n self.assertEqual(run_id, 1)\n self.assertIsInstance(rng_1, np.random.RandomState)\n\n run_id, rng_1 = get_rng(rng=1, run_id=1337, logger=smac.logger)\n self.assertEqual(run_id, 1337)\n self.assertIsInstance(rng_1, np.random.RandomState)\n\n rs = np.random.RandomState(1)\n run_id, rng_1 = get_rng(rng=rs, run_id=None, logger=smac.logger)\n self.assertIsInstance(run_id, int)\n self.assertIs(rng_1, rs)\n\n run_id, rng_1 = get_rng(rng=rs, run_id=2505, logger=smac.logger)\n self.assertEqual(run_id, 2505)\n self.assertIs(rng_1, rs)\n\n @unittest.mock.patch(\"smac.optimizer.acquisition.maximizer.get_one_exchange_neighbourhood\")\n def test_check_deterministic_rosenbrock(self, patch):\n\n # Make SMAC a bit faster\n patch.side_effect = lambda configuration, seed: get_one_exchange_neighbourhood(\n configuration=configuration,\n stdev=0.05,\n num_neighbors=2,\n seed=seed,\n )\n\n def rosenbrock_2d(x):\n x1 = x[\"x1\"]\n x2 = x[\"x2\"]\n val = 100.0 * (x2 - x1**2.0) ** 2.0 + (1 - x1) ** 2.0\n return val\n\n def opt_rosenbrock():\n cs = ConfigurationSpace()\n\n cs.add_hyperparameter(UniformFloatHyperparameter(\"x1\", -5, 5, default_value=-3))\n cs.add_hyperparameter(UniformFloatHyperparameter(\"x2\", -5, 5, default_value=-4))\n\n scenario = Scenario(\n {\n \"run_obj\": \"quality\", # we optimize quality (alternatively runtime)\n \"runcount-limit\": 50, # maximum function evaluations\n \"cs\": cs, # configuration space\n \"deterministic\": True,\n \"limit_resources\": True,\n \"intensification_percentage\": 0.000000001,\n }\n )\n\n smac = SMAC4AC(\n scenario=scenario,\n rng=np.random.RandomState(42),\n tae_runner=rosenbrock_2d,\n )\n incumbent = smac.optimize()\n return incumbent, smac.scenario.output_dir\n\n i1, output_dir = opt_rosenbrock()\n self.output_dirs.append(output_dir)\n x1_1 = i1.get(\"x1\")\n x2_1 = i1.get(\"x2\")\n i2, output_dir = opt_rosenbrock()\n self.output_dirs.append(output_dir)\n x1_2 = i2.get(\"x1\")\n x2_2 = i2.get(\"x2\")\n self.assertAlmostEqual(x1_1, x1_2)\n self.assertAlmostEqual(x2_1, x2_2)\n\n def test_get_runhistory_and_trajectory_and_tae_runner(self):\n def func(x):\n return x**2\n\n smac = SMAC4AC(tae_runner=func, scenario=self.scenario)\n self.assertRaises(ValueError, smac.get_runhistory)\n self.assertRaises(ValueError, smac.get_trajectory)\n smac.trajectory = \"dummy\"\n self.assertEqual(smac.get_trajectory(), \"dummy\")\n smac.runhistory = \"dummy\"\n self.assertEqual(smac.get_runhistory(), \"dummy\")\n self.assertEqual(smac.get_tae_runner().ta, func)\n\n def test_output_structure(self):\n \"\"\"Test whether output-dir is moved correctly.\"\"\"\n test_scenario_dict = {\n \"output_dir\": \"tests/test_files/scenario_test/tmp_output\",\n \"run_obj\": \"quality\",\n \"cs\": ConfigurationSpace(),\n }\n scen1 = Scenario(test_scenario_dict)\n self.output_dirs.append(scen1.output_dir)\n smac = SMAC4AC(scenario=scen1, run_id=1)\n\n self.assertEqual(smac.output_dir, os.path.join(test_scenario_dict[\"output_dir\"], \"run_1\"))\n self.assertTrue(os.path.isdir(smac.output_dir))\n\n smac2 = SMAC4AC(scenario=scen1, run_id=1)\n self.assertTrue(os.path.isdir(smac2.output_dir + \".OLD\"))\n\n smac3 = SMAC4AC(scenario=scen1, run_id=1)\n self.assertTrue(os.path.isdir(smac3.output_dir + \".OLD.OLD\"))\n\n smac4 = SMAC4AC(scenario=scen1, run_id=2)\n self.assertEqual(smac4.output_dir, os.path.join(test_scenario_dict[\"output_dir\"], \"run_2\"))\n self.assertTrue(os.path.isdir(smac4.output_dir))\n self.assertFalse(os.path.isdir(smac4.output_dir + \".OLD.OLD.OLD\"))\n\n # clean up (at least whats not cleaned up by tearDown)\n shutil.rmtree(smac.output_dir + \".OLD.OLD\")\n shutil.rmtree(smac.output_dir + \".OLD\")\n # This is done by teardown!\n # shutil.rmtree(smac.output_dir)\n shutil.rmtree(smac4.output_dir)\n\n def test_no_output(self):\n \"\"\"Test whether a scenario with \"\" as output really does not create an\n output.\"\"\"\n test_scenario_dict = {\n \"output_dir\": \"\",\n \"run_obj\": \"quality\",\n \"cs\": ConfigurationSpace(),\n }\n scen1 = Scenario(test_scenario_dict)\n smac = SMAC4AC(scenario=scen1, run_id=1)\n self.assertFalse(os.path.isdir(smac.output_dir))\n\n def test_register_callback(self):\n smac = SMAC4AC(scenario=self.scenario, run_id=1)\n\n with self.assertRaisesRegex(ValueError, \"Cannot register callback of type \"):\n smac.register_callback(lambda: 1)\n\n with self.assertRaisesRegex(ValueError, \"Cannot register callback of type \"):\n smac.register_callback(IncorporateRunResultCallback)\n\n smac.register_callback(IncorporateRunResultCallback())\n self.assertEqual(len(smac.solver._callbacks[\"_incorporate_run_results\"]), 1)\n\n class SubClass(IncorporateRunResultCallback):\n pass\n\n smac.register_callback(SubClass())\n self.assertEqual(len(smac.solver._callbacks[\"_incorporate_run_results\"]), 2)\n\n def test_set_limit_resources_with_tae_func_dict(self):\n # To optimize, we pass the function to the SMAC-object\n def tmp(**kwargs):\n return 1\n\n scenario = Scenario(\n {\n \"cs\": self.cs,\n \"run_obj\": \"quality\",\n \"output_dir\": \"\",\n \"limit_resources\": True,\n }\n )\n smac = SMAC4AC(scenario=scenario, tae_runner=tmp, rng=1)\n self.assertTrue(smac.solver.tae_runner.use_pynisher)\n self.assertIsNone(smac.solver.tae_runner.memory_limit)\n\n scenario = Scenario(\n {\n \"cs\": self.cs,\n \"run_obj\": \"quality\",\n \"output_dir\": \"\",\n \"memory_limit\": 333,\n \"limit_resources\": True,\n }\n )\n smac = SMAC4AC(scenario=scenario, tae_runner=tmp, rng=1)\n self.assertTrue(smac.solver.tae_runner.use_pynisher)\n self.assertEqual(smac.solver.tae_runner.memory_limit, 333)\n\n scenario = Scenario(\n {\n \"cs\": self.cs,\n \"run_obj\": \"quality\",\n \"output_dir\": \"\",\n \"memory_limit\": 333,\n \"limit_resources\": False,\n }\n )\n smac = SMAC4AC(scenario=scenario, tae_runner=tmp, rng=1)\n self.assertFalse(smac.solver.tae_runner.use_pynisher)\n self.assertEqual(smac.solver.tae_runner.memory_limit, 333)\n", "sub_path": "tests/test_facade/test_smac_facade.py", "file_name": "test_smac_facade.py", "file_ext": "py", "file_size_in_byte": 24692, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "unittest.TestCase", "line_number": 46, "usage_type": "attribute"}, {"api_name": "smac.configspace.ConfigurationSpace", "line_number": 48, "usage_type": "call"}, {"api_name": "smac.scenario.scenario.Scenario", "line_number": 56, "usage_type": "call"}, {"api_name": "contextlib.suppress", "line_number": 68, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 70, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 73, "usage_type": "call"}, {"api_name": "smac.callbacks", "line_number": 84, "usage_type": "name"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 84, "usage_type": "call"}, {"api_name": "smac.tae.execute_func.ExecuteTAFuncDict", "line_number": 85, "usage_type": "argument"}, {"api_name": "smac.callbacks.solver", "line_number": 85, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 85, "usage_type": "name"}, {"api_name": "smac.callbacks.solver", "line_number": 86, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 86, "usage_type": "name"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 93, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 103, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 111, "usage_type": "call"}, {"api_name": "smac.runhistory.runhistory2epm.RunHistory2EPM4Cost", "line_number": 112, "usage_type": "name"}, {"api_name": "smac.tae.StatusType.SUCCESS", "line_number": 115, "usage_type": "attribute"}, {"api_name": "smac.tae.StatusType", "line_number": 115, "usage_type": "name"}, {"api_name": "smac.tae.StatusType.CRASHED", "line_number": 115, "usage_type": "attribute"}, {"api_name": "smac.tae.StatusType.MEMOUT", "line_number": 115, "usage_type": "attribute"}, {"api_name": "smac.intensification.successive_halving.SuccessiveHalving", "line_number": 120, "usage_type": "name"}, {"api_name": "smac.intensification.hyperband.Hyperband", "line_number": 120, "usage_type": "name"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 121, "usage_type": "call"}, {"api_name": "smac.runhistory.runhistory2epm.RunHistory2EPM4Cost", "line_number": 126, "usage_type": "name"}, {"api_name": "smac.tae.StatusType.SUCCESS", "line_number": 130, "usage_type": "attribute"}, {"api_name": "smac.tae.StatusType", "line_number": 130, "usage_type": "name"}, {"api_name": "smac.tae.StatusType.CRASHED", "line_number": 131, "usage_type": "attribute"}, {"api_name": "smac.tae.StatusType", "line_number": 131, "usage_type": "name"}, {"api_name": "smac.tae.StatusType.MEMOUT", "line_number": 132, "usage_type": "attribute"}, {"api_name": "smac.tae.StatusType", "line_number": 132, "usage_type": "name"}, {"api_name": "smac.tae.StatusType.DONOTADVANCE", "line_number": 133, "usage_type": "attribute"}, {"api_name": "smac.tae.StatusType", "line_number": 133, "usage_type": "name"}, {"api_name": "smac.tae.StatusType.DONOTADVANCE", "line_number": 141, "usage_type": "attribute"}, {"api_name": "smac.tae.StatusType", "line_number": 141, "usage_type": "name"}, {"api_name": "smac.tae.StatusType.TIMEOUT", "line_number": 142, "usage_type": "attribute"}, {"api_name": "smac.tae.StatusType", "line_number": 142, "usage_type": "name"}, {"api_name": "smac.tae.StatusType.CRASHED", "line_number": 143, "usage_type": "attribute"}, {"api_name": "smac.tae.StatusType", "line_number": 143, "usage_type": "name"}, {"api_name": "smac.tae.StatusType.MEMOUT", "line_number": 144, "usage_type": "attribute"}, {"api_name": "smac.tae.StatusType", "line_number": 144, "usage_type": "name"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 150, "usage_type": "call"}, {"api_name": "smac.runhistory.runhistory2epm.RunHistory2EPM4LogCost", "line_number": 151, "usage_type": "name"}, {"api_name": "smac.tae.StatusType.SUCCESS", "line_number": 155, "usage_type": "attribute"}, {"api_name": "smac.tae.StatusType", "line_number": 155, "usage_type": "name"}, {"api_name": "smac.tae.StatusType.CAPPED", "line_number": 161, "usage_type": "attribute"}, {"api_name": "smac.tae.StatusType", "line_number": 161, "usage_type": "name"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 168, "usage_type": "call"}, {"api_name": "smac.runhistory.runhistory.RunHistory", "line_number": 169, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 171, "usage_type": "call"}, {"api_name": "smac.runhistory.runhistory.RunHistory", "line_number": 172, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 174, "usage_type": "call"}, {"api_name": "smac.runhistory.runhistory.RunHistory", "line_number": 174, "usage_type": "name"}, {"api_name": "smac.runhistory.runhistory.RunHistory", "line_number": 175, "usage_type": "argument"}, {"api_name": "numpy.random.RandomState", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 178, "usage_type": "attribute"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 179, "usage_type": "call"}, {"api_name": "smac.optimizer.configuration_chooser.random_chooser.ChooserProb", "line_number": 180, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 182, "usage_type": "call"}, {"api_name": "smac.optimizer.configuration_chooser.random_chooser.ChooserProb", "line_number": 183, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 185, "usage_type": "call"}, {"api_name": "smac.optimizer.configuration_chooser.random_chooser.ChooserProb", "line_number": 186, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 188, "usage_type": "call"}, {"api_name": "smac.optimizer.configuration_chooser.random_chooser.ChooserProb", "line_number": 189, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 191, "usage_type": "call"}, {"api_name": "smac.optimizer.configuration_chooser.random_chooser.ChooserNoCoolDown", "line_number": 193, "usage_type": "name"}, {"api_name": "smac.optimizer.configuration_chooser.random_chooser.ChooserNoCoolDown", "line_number": 196, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 199, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 202, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 202, "usage_type": "attribute"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 203, "usage_type": "call"}, {"api_name": "smac.epm.random_forest.rf_with_instances.RandomForestWithInstances", "line_number": 204, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 205, "usage_type": "call"}, {"api_name": "smac.epm.random_forest.rf_with_instances.RandomForestWithInstances", "line_number": 206, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 208, "usage_type": "call"}, {"api_name": "smac.epm.random_forest.rf_with_instances.RandomForestWithInstances", "line_number": 209, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 211, "usage_type": "call"}, {"api_name": "smac.epm.random_forest.rf_with_instances.RandomForestWithInstances", "line_number": 212, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 214, "usage_type": "call"}, {"api_name": "smac.epm.random_epm.RandomEPM", "line_number": 214, "usage_type": "name"}, {"api_name": "smac.epm.random_epm.RandomEPM", "line_number": 215, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 219, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 222, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 222, "usage_type": "attribute"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 223, "usage_type": "call"}, {"api_name": "smac.optimizer.acquisition.EI", "line_number": 224, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 225, "usage_type": "call"}, {"api_name": "smac.epm.random_forest.rf_with_instances.RandomForestWithInstances", "line_number": 226, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 228, "usage_type": "call"}, {"api_name": "smac.optimizer.acquisition.EI", "line_number": 229, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 231, "usage_type": "call"}, {"api_name": "smac.optimizer.acquisition.LCB", "line_number": 233, "usage_type": "name"}, {"api_name": "smac.optimizer.acquisition.LCB", "line_number": 236, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 240, "usage_type": "call"}, {"api_name": "smac.intensification.intensification.Intensifier", "line_number": 243, "usage_type": "name"}, {"api_name": "numpy.random.RandomState", "line_number": 246, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 246, "usage_type": "attribute"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 247, "usage_type": "call"}, {"api_name": "smac.intensification.intensification.Intensifier", "line_number": 248, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 250, "usage_type": "call"}, {"api_name": "smac.intensification.intensification.Intensifier", "line_number": 251, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 253, "usage_type": "call"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 255, "usage_type": "call"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 264, "usage_type": "call"}, {"api_name": "smac.scenario.scenario.Scenario", "line_number": 279, "usage_type": "call"}, {"api_name": "smac.callbacks", "line_number": 280, "usage_type": "name"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 280, "usage_type": "call"}, {"api_name": "smac.callbacks.solver", "line_number": 281, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 281, "usage_type": "name"}, {"api_name": "smac.callbacks.solver", "line_number": 282, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 282, "usage_type": "name"}, {"api_name": "smac.callbacks.solver", "line_number": 283, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 283, "usage_type": "name"}, {"api_name": "numpy.random.RandomState", "line_number": 287, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 287, "usage_type": "attribute"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 288, "usage_type": "call"}, {"api_name": "smac.initial_design.default_configuration_design.DefaultConfiguration", "line_number": 289, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 291, "usage_type": "call"}, {"api_name": "smac.intensification.intensification.Intensifier", "line_number": 292, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 294, "usage_type": "call"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 296, "usage_type": "call"}, {"api_name": "smac.initial_design.initial_design.InitialDesign", "line_number": 298, "usage_type": "name"}, {"api_name": "smac.initial_design.initial_design.InitialDesign", "line_number": 301, "usage_type": "argument"}, {"api_name": "smac.initial_design.default_configuration_design.DefaultConfiguration", "line_number": 305, "usage_type": "name"}, {"api_name": "smac.initial_design.random_configuration_design.RandomConfigurations", "line_number": 306, "usage_type": "name"}, {"api_name": "smac.initial_design.latin_hypercube_design.LHDesign", "line_number": 307, "usage_type": "name"}, {"api_name": "smac.initial_design.factorial_design.FactorialInitialDesign", "line_number": 308, "usage_type": "name"}, {"api_name": "smac.initial_design.sobol_design.SobolDesign", "line_number": 309, "usage_type": "name"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 312, "usage_type": "call"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 318, "usage_type": "call"}, {"api_name": "smac.epm.random_forest.rf_mo.MultiObjectiveRandomForest", "line_number": 320, "usage_type": "name"}, {"api_name": "smac.optimizer.acquisition.EIPS", "line_number": 322, "usage_type": "name"}, {"api_name": "smac.runhistory.runhistory2epm.RunHistory2EPM4EIPS", "line_number": 323, "usage_type": "name"}, {"api_name": "smac.epm.random_forest.rf_mo.MultiObjectiveRandomForest", "line_number": 327, "usage_type": "argument"}, {"api_name": "smac.optimizer.acquisition.EIPS", "line_number": 329, "usage_type": "argument"}, {"api_name": "smac.epm.random_forest.rf_mo.MultiObjectiveRandomForest", "line_number": 332, "usage_type": "argument"}, {"api_name": "smac.runhistory.runhistory2epm.RunHistory2EPM4EIPS", "line_number": 334, "usage_type": "argument"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 337, "usage_type": "call"}, {"api_name": "smac.optimizer.acquisition.EIPS", "line_number": 337, "usage_type": "name"}, {"api_name": "smac.runhistory.runhistory2epm.RunHistory2EPM4EIPS", "line_number": 337, "usage_type": "name"}, {"api_name": "smac.callbacks", "line_number": 345, "usage_type": "name"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 345, "usage_type": "call"}, {"api_name": "smac.callbacks.logger", "line_number": 346, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 346, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 346, "usage_type": "call"}, {"api_name": "unittest.mock", "line_number": 346, "usage_type": "attribute"}, {"api_name": "smac.epm.utils.get_rng", "line_number": 350, "usage_type": "call"}, {"api_name": "smac.epm.utils.get_rng", "line_number": 351, "usage_type": "call"}, {"api_name": "smac.epm.utils.get_rng", "line_number": 355, "usage_type": "call"}, {"api_name": "smac.callbacks.logger", "line_number": 355, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 355, "usage_type": "name"}, {"api_name": "smac.callbacks.logger", "line_number": 356, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 356, "usage_type": "name"}, {"api_name": "smac.epm.utils.get_rng", "line_number": 357, "usage_type": "call"}, {"api_name": "smac.callbacks.logger", "line_number": 357, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 357, "usage_type": "name"}, {"api_name": "smac.callbacks.logger", "line_number": 358, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 358, "usage_type": "name"}, {"api_name": "smac.epm.utils.get_rng", "line_number": 363, "usage_type": "call"}, {"api_name": "smac.epm.utils.get_rng", "line_number": 364, "usage_type": "call"}, {"api_name": "smac.epm.utils.get_rng", "line_number": 372, "usage_type": "argument"}, {"api_name": "smac.epm.utils.get_rng", "line_number": 378, "usage_type": "argument"}, {"api_name": "smac.epm.utils.get_rng", "line_number": 382, "usage_type": "call"}, {"api_name": "smac.callbacks.logger", "line_number": 382, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 382, "usage_type": "name"}, {"api_name": "numpy.random", "line_number": 384, "usage_type": "attribute"}, {"api_name": "smac.callbacks.logger", "line_number": 385, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 385, "usage_type": "name"}, {"api_name": "smac.epm.utils.get_rng", "line_number": 387, "usage_type": "call"}, {"api_name": "smac.callbacks.logger", "line_number": 387, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 387, "usage_type": "name"}, {"api_name": "numpy.random", "line_number": 389, "usage_type": "attribute"}, {"api_name": "smac.epm.utils.get_rng", "line_number": 391, "usage_type": "call"}, {"api_name": "smac.callbacks.logger", "line_number": 391, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 391, "usage_type": "name"}, {"api_name": "numpy.random", "line_number": 393, "usage_type": "attribute"}, {"api_name": "smac.epm.utils.get_rng", "line_number": 395, "usage_type": "call"}, {"api_name": "smac.callbacks.logger", "line_number": 395, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 395, "usage_type": "name"}, {"api_name": "numpy.random", "line_number": 397, "usage_type": "attribute"}, {"api_name": "numpy.random.RandomState", "line_number": 399, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 399, "usage_type": "attribute"}, {"api_name": "smac.epm.utils.get_rng", "line_number": 400, "usage_type": "call"}, {"api_name": "smac.callbacks.logger", "line_number": 400, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 400, "usage_type": "name"}, {"api_name": "smac.epm.utils.get_rng", "line_number": 404, "usage_type": "call"}, {"api_name": "smac.callbacks.logger", "line_number": 404, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 404, "usage_type": "name"}, {"api_name": "unittest.mock.patch.object", "line_number": 342, "usage_type": "call"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 342, "usage_type": "argument"}, {"api_name": "unittest.mock", "line_number": 342, "usage_type": "attribute"}, {"api_name": "ConfigSpace.util.get_one_exchange_neighbourhood", "line_number": 412, "usage_type": "call"}, {"api_name": "smac.configspace.ConfigurationSpace", "line_number": 426, "usage_type": "call"}, {"api_name": "ConfigSpace.hyperparameters.UniformFloatHyperparameter", "line_number": 428, "usage_type": "call"}, {"api_name": "ConfigSpace.hyperparameters.UniformFloatHyperparameter", "line_number": 429, "usage_type": "call"}, {"api_name": "smac.scenario.scenario.Scenario", "line_number": 431, "usage_type": "call"}, {"api_name": "smac.callbacks", "line_number": 442, "usage_type": "name"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 442, "usage_type": "call"}, {"api_name": "numpy.random.RandomState", "line_number": 444, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 444, "usage_type": "attribute"}, {"api_name": "smac.callbacks.optimize", "line_number": 447, "usage_type": "call"}, {"api_name": "smac.callbacks", "line_number": 447, "usage_type": "name"}, {"api_name": "smac.callbacks.scenario", "line_number": 448, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 448, "usage_type": "name"}, {"api_name": "unittest.mock.patch", "line_number": 408, "usage_type": "call"}, {"api_name": "unittest.mock", "line_number": 408, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 465, "usage_type": "name"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 465, "usage_type": "call"}, {"api_name": "smac.callbacks.get_runhistory", "line_number": 466, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 466, "usage_type": "name"}, {"api_name": "smac.callbacks.get_trajectory", "line_number": 467, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 467, "usage_type": "name"}, {"api_name": "smac.callbacks.trajectory", "line_number": 468, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 468, "usage_type": "name"}, {"api_name": "smac.callbacks.get_trajectory", "line_number": 469, "usage_type": "call"}, {"api_name": "smac.callbacks", "line_number": 469, "usage_type": "name"}, {"api_name": "smac.callbacks.runhistory", "line_number": 470, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 470, "usage_type": "name"}, {"api_name": "smac.callbacks.get_runhistory", "line_number": 471, "usage_type": "call"}, {"api_name": "smac.callbacks", "line_number": 471, "usage_type": "name"}, {"api_name": "smac.callbacks.get_tae_runner", "line_number": 472, "usage_type": "call"}, {"api_name": "smac.callbacks", "line_number": 472, "usage_type": "name"}, {"api_name": "smac.configspace.ConfigurationSpace", "line_number": 479, "usage_type": "call"}, {"api_name": "smac.scenario.scenario.Scenario", "line_number": 481, "usage_type": "call"}, {"api_name": "smac.callbacks", "line_number": 483, "usage_type": "name"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 483, "usage_type": "call"}, {"api_name": "smac.callbacks.output_dir", "line_number": 485, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 485, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 485, "usage_type": "call"}, {"api_name": "os.path", "line_number": 485, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 486, "usage_type": "call"}, {"api_name": "os.path", "line_number": 486, "usage_type": "attribute"}, {"api_name": "smac.callbacks.output_dir", "line_number": 486, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 486, "usage_type": "name"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 488, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 489, "usage_type": "call"}, {"api_name": "os.path", "line_number": 489, "usage_type": "attribute"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 491, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 492, "usage_type": "call"}, {"api_name": "os.path", "line_number": 492, "usage_type": "attribute"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 494, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 495, "usage_type": "call"}, {"api_name": "os.path", "line_number": 495, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 496, "usage_type": "call"}, {"api_name": "os.path", "line_number": 496, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 497, "usage_type": "call"}, {"api_name": "os.path", "line_number": 497, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 500, "usage_type": "call"}, {"api_name": "smac.callbacks.output_dir", "line_number": 500, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 500, "usage_type": "name"}, {"api_name": "shutil.rmtree", "line_number": 501, "usage_type": "call"}, {"api_name": "smac.callbacks.output_dir", "line_number": 501, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 501, "usage_type": "name"}, {"api_name": "shutil.rmtree", "line_number": 504, "usage_type": "call"}, {"api_name": "smac.configspace.ConfigurationSpace", "line_number": 512, "usage_type": "call"}, {"api_name": "smac.scenario.scenario.Scenario", "line_number": 514, "usage_type": "call"}, {"api_name": "smac.callbacks", "line_number": 515, "usage_type": "name"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 515, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 516, "usage_type": "call"}, {"api_name": "os.path", "line_number": 516, "usage_type": "attribute"}, {"api_name": "smac.callbacks.output_dir", "line_number": 516, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 516, "usage_type": "name"}, {"api_name": "smac.callbacks", "line_number": 519, "usage_type": "name"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 519, "usage_type": "call"}, {"api_name": "smac.callbacks.register_callback", "line_number": 522, "usage_type": "call"}, {"api_name": "smac.callbacks", "line_number": 522, "usage_type": "name"}, {"api_name": "smac.callbacks.register_callback", "line_number": 525, "usage_type": "call"}, {"api_name": "smac.callbacks.IncorporateRunResultCallback", "line_number": 525, "usage_type": "argument"}, {"api_name": "smac.callbacks", "line_number": 525, "usage_type": "name"}, {"api_name": "smac.callbacks.register_callback", "line_number": 527, "usage_type": "call"}, {"api_name": "smac.callbacks", "line_number": 527, "usage_type": "name"}, {"api_name": "smac.callbacks.IncorporateRunResultCallback", "line_number": 527, "usage_type": "call"}, {"api_name": "smac.callbacks.solver", "line_number": 528, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 528, "usage_type": "name"}, {"api_name": "smac.callbacks.IncorporateRunResultCallback", "line_number": 530, "usage_type": "name"}, {"api_name": "smac.callbacks.register_callback", "line_number": 533, "usage_type": "call"}, {"api_name": "smac.callbacks", "line_number": 533, "usage_type": "name"}, {"api_name": "smac.callbacks.solver", "line_number": 534, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 534, "usage_type": "name"}, {"api_name": "smac.scenario.scenario.Scenario", "line_number": 541, "usage_type": "call"}, {"api_name": "smac.callbacks", "line_number": 549, "usage_type": "name"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 549, "usage_type": "call"}, {"api_name": "smac.callbacks.solver", "line_number": 550, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 550, "usage_type": "name"}, {"api_name": "smac.callbacks.solver", "line_number": 551, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 551, "usage_type": "name"}, {"api_name": "smac.scenario.scenario.Scenario", "line_number": 553, "usage_type": "call"}, {"api_name": "smac.callbacks", "line_number": 562, "usage_type": "name"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 562, "usage_type": "call"}, {"api_name": "smac.callbacks.solver", "line_number": 563, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 563, "usage_type": "name"}, {"api_name": "smac.callbacks.solver", "line_number": 564, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 564, "usage_type": "name"}, {"api_name": "smac.scenario.scenario.Scenario", "line_number": 566, "usage_type": "call"}, {"api_name": "smac.callbacks", "line_number": 575, "usage_type": "name"}, {"api_name": "smac.facade.smac_ac_facade.SMAC4AC", "line_number": 575, "usage_type": "call"}, {"api_name": "smac.callbacks.solver", "line_number": 576, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 576, "usage_type": "name"}, {"api_name": "smac.callbacks.solver", "line_number": 577, "usage_type": "attribute"}, {"api_name": "smac.callbacks", "line_number": 577, "usage_type": "name"}]}
+{"seq_id": "307559474", "text": "import pygame\r\nimport math\r\nfrom queue import PriorityQueue\r\n\r\nWIDTH = 400\r\nGRID = 25\r\nWIN = pygame.display.set_mode((WIDTH,WIDTH))\r\npygame.display.set_caption(\"PATH VISUALIZER\")\r\n\r\nCLOSED = RED = (255, 0, 0)\r\nOPEN = GREEN = (0, 255, 0)\r\nDEFAULT = WHITE = (255, 255, 255)\r\nBLOCKAGE = BLACK = (0, 0, 0)\r\nPATH = PURPLE = (128, 0, 128)\r\nSTART = ORANGE = (255, 165 ,0)\r\nGREY = (128, 128, 128)\r\nEND = TURQUOISE = (64, 224, 208)\r\n\r\nclass Square:\r\n\r\n def __init__(self, row, col):\r\n self.row = row\r\n self.col = col\r\n self.width = WIDTH // GRID\r\n self.x = row * self.width\r\n self.y = col * self.width\r\n self.state = DEFAULT\r\n self.neighbors = []\r\n\r\n def get_pos(self):\r\n return self.row, self.col\r\n \r\n def draw(self, win):\r\n pygame.draw.rect(win, self.state, (self.x, self.y, self.width, self.width))\r\n\r\n def create_neighbors(self, table):\r\n if self.row < GRID - 1:\r\n if table[self.row + 1][self.col].state != BLOCKAGE: #DOWN\r\n self.neighbors.append(table[self.row + 1][self.col])\r\n if self.col > 0 and table[self.row + 1][self.col -1].state != BLOCKAGE: #DOWN LEFT\r\n self.neighbors.append(table[self.row + 1][self.col - 1])\r\n if self.col < GRID - 1 and table[self.row + 1][self.col + 1].state != BLOCKAGE: #DOWN RIGHT\r\n self.neighbors.append(table[self.row + 1][self.col + 1])\r\n if self.row > 0:\r\n if table[self.row - 1][self.col].state != BLOCKAGE: #UP\r\n self.neighbors.append(table[self.row - 1][self.col])\r\n if self.col > 0 and table[self.row - 1][self.col - 1].state != BLOCKAGE: #UP LEFT\r\n self.neighbors.append(table[self.row - 1][self.col - 1])\r\n if self.col < GRID - 1 and table[self.row - 1][self.col + 1].state != BLOCKAGE: #UP RIGHT\r\n self.neighbors.append(table[self.row - 1][self.col + 1])\r\n if self.col < GRID - 1 and table[self.row][self.col + 1].state != BLOCKAGE: #RIGHT\r\n self.neighbors.append(table[self.row][self.col + 1])\r\n if self.col > 0 and table[self.row][self.col - 1].state != BLOCKAGE: #LEFT\r\n self.neighbors.append(table[self.row][self.col - 1])\r\n\r\ndef e(p1, p2):\r\n x1, y1 = p1.get_pos()\r\n x2, y2 = p2.get_pos()\r\n return math.floor(math.sqrt((x2 - x1)**2 + (y2 - y1)**2))\r\n\r\n\r\ndef create_path(came_from, current, draw):\r\n while current in came_from:\r\n current = came_from[current]\r\n current.state=PATH\r\n draw()\r\n\r\ndef algorithm(draw, grid, start, end):\r\n count = 0\r\n open_set=PriorityQueue()\r\n open_set.put((0,count,start))\r\n came_from = {}\r\n s = {sq:float('inf') for row in grid for sq in row}\r\n s[start]=0\r\n t = {sq: float('inf') for row in grid for sq in row}\r\n t[start] = e(start,end)\r\n open_set_hash = {start}\r\n while not open_set.empty():\r\n for event in pygame.event.get():\r\n if event.type == pygame.QUIT:\r\n pygame.quit()\r\n current = open_set.get()[2]\r\n open_set_hash.remove(current)\r\n if current == end:\r\n create_path(came_from,end,draw)\r\n end.state=END\r\n start.state=START\r\n return\r\n for neighbor in current.neighbors:\r\n temp_s = s[current] + 1\r\n if temp_s < s[neighbor]:\r\n came_from[neighbor]=current\r\n s[neighbor]=temp_s\r\n t[neighbor]=temp_s+e(neighbor,end)\r\n if neighbor not in open_set_hash:\r\n count += 1\r\n open_set.put((t[neighbor], count, neighbor))\r\n open_set_hash.add(neighbor)\r\n neighbor.state=OPEN\r\n draw()\r\n if current != start:\r\n current.state = CLOSED\r\n return\r\n \r\n\r\ndef make_grid():\r\n grid = []\r\n for row in range(GRID):\r\n grid.append([])\r\n for col in range(GRID):\r\n sq = Square(row, col)\r\n grid[row].append(sq)\r\n return grid\r\n\r\ndef draw_grid(win):\r\n gap = WIDTH//GRID\r\n for line in range(GRID):\r\n pygame.draw.line(win, GREY, (0, line*gap), (WIDTH, line*gap))\r\n pygame.draw.line(win, GREY, (line*gap, 0), (line*gap, WIDTH))\r\n\r\ndef draw(win, grid):\r\n win.fill(WHITE)\r\n for row in grid:\r\n for col in row:\r\n col.draw(win)\r\n draw_grid(win)\r\n pygame.display.update()\r\n\r\ndef get_clicked_pos(pos):\r\n gap = WIDTH // GRID\r\n y, x = pos\r\n row = y // gap\r\n col = x // gap\r\n return row, col\r\n\r\ndef main(win):\r\n grid = make_grid()\r\n start = end = None\r\n run = True\r\n while run:\r\n draw(win, grid)\r\n for event in pygame.event.get():\r\n if event.type == pygame.QUIT:\r\n run = False\r\n if pygame.mouse.get_pressed()[0]:\r\n pos = pygame.mouse.get_pos()\r\n row, col = get_clicked_pos(pos)\r\n sq = grid[row][col]\r\n if not start and sq != end:\r\n start = sq\r\n start.state = START\r\n elif not end and sq != start:\r\n end = sq\r\n end.state = END\r\n elif sq != end and sq != start:\r\n sq.state = BLOCKAGE\r\n elif pygame.mouse.get_pressed()[2]:\r\n pos = pygame.mouse.get_pos()\r\n row, col = get_clicked_pos(pos)\r\n sq = grid[row][col]\r\n sq.state=DEFAULT\r\n if sq == start:\r\n start = None\r\n if sq == end:\r\n end = None\r\n if event.type == pygame.KEYDOWN:\r\n if event.key == pygame.K_SPACE and start and end:\r\n for row in grid:\r\n for sq in row:\r\n sq.create_neighbors(grid)\r\n algorithm(lambda:draw(win,grid), grid, start, end)\r\n if event.key == pygame.K_c:\r\n start = end = None\r\n grid = make_grid()\r\nmain(WIN)", "sub_path": "myAstar.py", "file_name": "myAstar.py", "file_ext": "py", "file_size_in_byte": 6113, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pygame.display.set_mode", "line_number": 7, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 7, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 8, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 8, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 34, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 34, "usage_type": "attribute"}, {"api_name": "math.floor", "line_number": 59, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 59, "usage_type": "call"}, {"api_name": "queue.PriorityQueue", "line_number": 70, "usage_type": "call"}, {"api_name": "pygame.event.get", "line_number": 79, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 79, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 80, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 81, "usage_type": "call"}, {"api_name": "pygame.draw.line", "line_number": 118, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 118, "usage_type": "attribute"}, {"api_name": "pygame.draw.line", "line_number": 119, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 119, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 127, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 127, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 142, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 142, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 143, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pressed", "line_number": 145, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 145, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pos", "line_number": 146, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 146, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pressed", "line_number": 157, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 157, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pos", "line_number": 158, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 158, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 166, "usage_type": "attribute"}, {"api_name": "pygame.K_SPACE", "line_number": 167, "usage_type": "attribute"}, {"api_name": "pygame.K_c", "line_number": 172, "usage_type": "attribute"}]}
+{"seq_id": "551137886", "text": "from odoo import api, fields, models, _\nfrom datetime import datetime\nfrom odoo.exceptions import UserError\nimport logging\n_logger = logging.getLogger(__name__)\n\nclass RoyaltyFeetoInvoice(models.TransientModel):\n _name='royalty.fee.to.invoice'\n\n royalty_fee_to_invoice_ids = fields.One2many('royalty.fee.lines.to.invoice','royalty_fee_lines_id', string=\"Royalty Fee to Invoice\")\n is_royalty_fee = fields.Boolean(default=True)\n branch_id = fields.Many2one('branches.cost.center', string=\"Branch\")\n\n\n\n @api.model\n def default_get(self, vals):\n res = super(RoyaltyFeetoInvoice, self).default_get(vals)\n royalty_fee_obj = self.env['monthly.royalty.fees.line'].browse(self._context.get('active_id'))\n\n items = []\n for line in royalty_fee_obj:\n invoice_royalty_line = self.env['royalty.fee.lines.to.invoice'].create({\n 'royalty_fee_lines_id' : self.id,\n 'month': line.month.id,\n 'royalty_fee': line.royalty_fee\n })\n items.append(invoice_royalty_line.id)\n res.update({'royalty_fee_to_invoice_ids':[(6,0,items)]})\n return res\n\nclass RoyaltyFeeLinestoInvoice(models.TransientModel):\n _name = 'royalty.fee.lines.to.invoice'\n\n royalty_fee_lines_id = fields.Many2one('royalty.fee.to.invoice')\n month = fields.Many2one('royalty.fee.month', string=\"Month\")\n royalty_fee = fields.Float(string=\"Royalty Fee\")\n\n\n\n", "sub_path": "suds/.history/wizard/royalty_fee_lines_20190806160537.py", "file_name": "royalty_fee_lines_20190806160537.py", "file_ext": "py", "file_size_in_byte": 1442, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.getLogger", "line_number": 5, "usage_type": "call"}, {"api_name": "odoo.models.TransientModel", "line_number": 7, "usage_type": "attribute"}, {"api_name": "odoo.models", "line_number": 7, "usage_type": "name"}, {"api_name": "odoo.fields.One2many", "line_number": 10, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 10, "usage_type": "name"}, {"api_name": "odoo.fields.Boolean", "line_number": 11, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 11, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 12, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 12, "usage_type": "name"}, {"api_name": "odoo.api.model", "line_number": 16, "usage_type": "attribute"}, {"api_name": "odoo.api", "line_number": 16, "usage_type": "name"}, {"api_name": "odoo.models.TransientModel", "line_number": 32, "usage_type": "attribute"}, {"api_name": "odoo.models", "line_number": 32, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 35, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 35, "usage_type": "name"}, {"api_name": "odoo.fields.Many2one", "line_number": 36, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 36, "usage_type": "name"}, {"api_name": "odoo.fields.Float", "line_number": 37, "usage_type": "call"}, {"api_name": "odoo.fields", "line_number": 37, "usage_type": "name"}]}
+{"seq_id": "388573438", "text": "from gevent import monkey; monkey.patch_all()\nfrom utils import Singleton\nimport telebot\nimport copy\nimport json\nfrom app import MasterBot, Bot\n\n\nclass PollingProcessor(Singleton):\n tokens = {}\n mb = MasterBot({'token': open('token').read().replace('\\n', '')})\n\n def get_updates(self, silent=False):\n res = False\n for token in copy.copy(Bot.bots.keys()):\n updates = telebot.apihelper.get_updates(token, offset=self.tokens.get(token) or 0)\n for update in updates:\n if update['update_id'] > self.tokens.get(token):\n self.tokens[token] = update['update_id']\n res = True\n if not silent:\n self.mb.route_update(token, json.dumps(update))\n return res\n\n def start(self):\n while self.get_updates(silent=True):\n pass\n while True:\n self.get_updates()\n\n\nif __name__ == \"__main__\":\n PollingProcessor().start()\n", "sub_path": "polling_listener.py", "file_name": "polling_listener.py", "file_ext": "py", "file_size_in_byte": 986, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "gevent.monkey.patch_all", "line_number": 1, "usage_type": "call"}, {"api_name": "gevent.monkey", "line_number": 1, "usage_type": "name"}, {"api_name": "utils.Singleton", "line_number": 9, "usage_type": "name"}, {"api_name": "app.MasterBot", "line_number": 11, "usage_type": "call"}, {"api_name": "copy.copy", "line_number": 15, "usage_type": "call"}, {"api_name": "app.Bot.bots.keys", "line_number": 15, "usage_type": "call"}, {"api_name": "app.Bot.bots", "line_number": 15, "usage_type": "attribute"}, {"api_name": "app.Bot", "line_number": 15, "usage_type": "name"}, {"api_name": "telebot.apihelper.get_updates", "line_number": 16, "usage_type": "call"}, {"api_name": "telebot.apihelper", "line_number": 16, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 22, "usage_type": "call"}]}
+{"seq_id": "443639826", "text": "# coding: utf-8\n\n\"\"\" This file contains the PlotFrame class extending tkinter.Frame \"\"\"\n\nimport logging\nimport tkinter\nfrom tkinter import simpledialog\nfrom tkinter.ttk import Combobox, Button, Label, Entry\n\nfrom matplotlib import style\nfrom matplotlib.figure import Figure\nfrom matplotlib.backends.backend_tkagg import (FigureCanvasTkAgg, NavigationToolbar2Tk)\n\nfrom ranalysis.gui.dialog.plotdialog import PlotDialog\nfrom ranalysis.log.loghandler import logger\nfrom ranalysis.plot.graph import graph_from_plot_ids, graph_clear, graph_add_title\nfrom ranalysis.plot.plotcreator import PlotCreator\n\n\nclass PlotFrame(tkinter.Frame):\n \"\"\" Display buttons, list, combobox and matplotlib canvas \"\"\"\n\n __margin = 0\n\n def __init__(self, parent, plot_frame_id):\n \"\"\" PlotFrame constructor \"\"\"\n self.top = tkinter.Toplevel(parent)\n self.__parent = parent\n self.__plot_frame_id = plot_frame_id\n\n position_right = 50\n position_down = 250\n self.top.geometry(\"1280x600+{}+{}\".format(position_right + PlotFrame.__margin, position_down))\n self.top.title(plot_frame_id)\n PlotFrame.__margin += 50\n\n self.__style_combo = None\n self.__marker_combo = None\n self.__line_combo = None\n self.__plot_list = None\n self.__entry_line_width = None\n\n self.__canvas = None\n self.__graph = None\n self.__graph_frame = None\n\n self.__compare_button = None\n\n self.__marker_select_combo4 = tkinter.StringVar()\n self.__line_select_combo5 = tkinter.StringVar()\n\n self.initialize()\n\n def get_plot_frame_id(self):\n \"\"\" Return the plot frame id \"\"\"\n return self.__plot_frame_id\n\n def initialize(self):\n \"\"\" Initialize all the tkinter objects \"\"\"\n # frame\n list_frame = tkinter.Frame(self.top, borderwidth=2, relief=tkinter.GROOVE)\n list_frame.pack(side=tkinter.LEFT, fill=tkinter.Y, padx=5, pady=10)\n\n graph_frame = tkinter.Frame(self.top)\n graph_frame.pack(side=tkinter.RIGHT, fill=tkinter.Y, padx=5, pady=5)\n\n customize_frame = tkinter.Frame(graph_frame, borderwidth=2, relief=tkinter.GROOVE)\n customize_frame.pack(side=tkinter.RIGHT, fill=tkinter.Y, padx=5, pady=5)\n\n self.__graph_frame = tkinter.Frame(graph_frame)\n self.__graph_frame.pack(side=tkinter.LEFT, padx=5, pady=5, fill=tkinter.BOTH, expand=True)\n\n # button\n title_button = Button(customize_frame, text=\"Add title\", command=self.add_title, width=20)\n title_button.grid(row=1, column=1, columnspan=1, rowspan=1, padx=5, pady=5)\n style_button = Button(customize_frame, text=\"Update windows style\", command=self.style_plot, width=20)\n style_button.grid(row=4, column=1, columnspan=1, rowspan=1, padx=5, pady=5)\n marker_line_button = Button(customize_frame, text=\"Update plot style\", command=self.on_list_select, width=20)\n marker_line_button.grid(row=10, column=1, columnspan=2, rowspan=1, padx=5, pady=5)\n\n remove_button = Button(list_frame, text=\"Remove plot\", command=self.remove_plot, width=20)\n self.__compare_button = Button(list_frame, state=tkinter.DISABLED, text=\"Compare two plots\",\n command=self.compare_plots, width=20)\n\n # label\n label_style = Label(customize_frame, text=\"Windows theme: \", width=15)\n label_style.grid(row=2, column=1, rowspan=1, padx=5, pady=5, sticky=tkinter.W)\n label_style = Label(customize_frame, text=\"Marker style: \", width=15)\n label_style.grid(row=8, column=1, rowspan=1, padx=5, pady=5, sticky=tkinter.W)\n label_line = Label(customize_frame, text=\"Line style: \", width=15)\n label_line.grid(row=5, column=1, rowspan=1, padx=5, pady=5, sticky=tkinter.W)\n\n # list\n self.__plot_list = tkinter.Listbox(list_frame, selectmode=tkinter.MULTIPLE, exportselection=False)\n self.__plot_list.bind('<>', self.on_list_select)\n\n # combo\n style_select_combo3 = tkinter.StringVar()\n self.__style_combo = Combobox(customize_frame, textvariable=style_select_combo3, values=style.available,\n state='readonly', width=18)\n self.__style_combo.grid(row=3, column=1, rowspan=1, padx=5, pady=5)\n\n self.__marker_combo = Combobox(customize_frame, textvariable=self.__marker_select_combo4,\n state='readonly', width=18)\n self.__marker_combo['values'] = (\".\", \",\", \"o\", \"v\", \"^\", \"<\", \">\", \"1\", \"2\", \"3\", \"4\", \"8\", \"s\", \"p\", \"P\", \"*\",\n \"h\", \"H\", \"+\", \"x\", \"X\", \"D\", \"d\", \"|\", \"_\", \"None\")\n self.__marker_combo.current(0)\n self.__marker_combo.grid(row=9, column=1, rowspan=1, padx=5, pady=5)\n\n self.__line_combo = Combobox(customize_frame, textvariable=self.__line_select_combo5,\n state='readonly', width=18)\n self.__line_combo['values'] = (\"solid\", \"dotted\", \"dashed\", \"dashdot\")\n self.__line_combo.current(0)\n self.__line_combo.grid(row=6, column=1, rowspan=1, padx=5, pady=5)\n\n # entry\n self.__entry_line_width = Entry(customize_frame, width=21)\n self.__entry_line_width.insert(tkinter.END, '1.0')\n self.__entry_line_width.grid(row=7, column=1, columnspan=1, rowspan=1, padx=5, pady=5)\n\n # pack\n self.__plot_list.pack(padx=10, pady=10, fill=tkinter.Y, expand=1)\n remove_button.pack(padx=10, pady=5)\n self.__compare_button.pack(padx=10, pady=5)\n\n # plot\n self.create_plot()\n\n # quit\n self.top.protocol(\"WM_DELETE_WINDOW\", self.quit)\n\n def create_plot(self, name_style='ggplot'):\n \"\"\" Create a matplotlib plot in a tkinter environment \"\"\"\n if self.__canvas is not None:\n self.__canvas.get_tk_widget().destroy()\n if self.__graph_frame is not None:\n self.__graph_frame.destroy()\n\n self.__graph = None\n self.__canvas = None\n self.__graph_frame = None\n\n self.__graph_frame = tkinter.Frame(self.top, borderwidth=2, relief=tkinter.GROOVE)\n self.__graph_frame.pack(side=tkinter.RIGHT, padx=10, pady=10, fill=tkinter.BOTH, expand=True)\n\n style.use(name_style)\n figure = Figure()\n figure.subplots_adjust(left=0.08, right=0.84, bottom=0.125)\n self.__graph = figure.add_subplot(111)\n\n self.__canvas = FigureCanvasTkAgg(figure, master=self.__graph_frame)\n self.__canvas.draw()\n self.__canvas.get_tk_widget().pack(side=tkinter.TOP, fill=tkinter.BOTH, expand=1)\n\n toolbar = NavigationToolbar2Tk(self.__canvas, self.__graph_frame)\n toolbar.update()\n self.__canvas.get_tk_widget().pack(side=tkinter.TOP, fill=tkinter.BOTH, expand=1)\n\n def style_plot(self):\n \"\"\" Change the matplotlib plot style \"\"\"\n if self.__style_combo.get() != \"\":\n logger.log(logging.INFO, \"[PlotFrame] Style: \" + self.__style_combo.get())\n self.create_plot(self.__style_combo.get())\n\n def remove_plot(self):\n \"\"\" Remove selected plot from the list \"\"\"\n if self.__plot_list.size() > 0:\n logger.log(logging.INFO, \"[PlotFrame] Remove plot\")\n idxs = self.__plot_list.curselection()\n for idx in idxs:\n self.__plot_list.delete(idx)\n else:\n logger.log(logging.ERROR, \"[PlotFrame] No plot to remove\")\n\n def compare_plots(self):\n \"\"\" Compare two plots in a graph \"\"\"\n plot_ids = []\n for idx in self.__plot_list.curselection():\n plot_ids.append(self.__plot_list.get(idx))\n if len(plot_ids) == 2:\n plot1 = PlotCreator.get_instance().get_plot_from_stringify(plot_ids[0])\n plot2 = PlotCreator.get_instance().get_plot_from_stringify(plot_ids[1])\n if len(plot1.get_x()) == len(plot2.get_x()):\n PlotDialog(self.top, plot1, plot2, self.__marker_combo.get(), self.__line_combo.get(),\n float(self.__entry_line_width.get()))\n else:\n logger.log(logging.ERROR, \"[PlotFrame] Plots do not have the same x interval\")\n else:\n logger.log(logging.ERROR, \"[PlotFrame] You can compare only two plots\")\n\n def add_plot(self, plot):\n logger.log(logging.INFO, \"[PlotFrame] Adding plot \" + str(plot))\n self.__plot_list.insert(tkinter.END, plot)\n\n def clear_all_plots(self):\n \"\"\" Clear/reset the plot frame \"\"\"\n logger.log(logging.INFO, \"[PlotFrame] Clear all plots\")\n graph_clear(self.__graph)\n self.__plot_list.selection_clear(0, tkinter.END)\n self.__canvas.draw()\n\n def remove_all_plots(self):\n \"\"\" Remove all the plots from the plot list \"\"\"\n logger.log(logging.INFO, \"[PlotFrame] Remove all plots\")\n self.__plot_list.delete(0, tkinter.END)\n\n def on_list_select(self, evt=None):\n \"\"\" Display plots when plots are selected in the list (triggered by event on the list) \"\"\"\n graph_clear(self.__graph)\n plot_ids = []\n for idx in self.__plot_list.curselection():\n plot_ids.append(self.__plot_list.get(idx))\n\n if len(self.__plot_list.curselection()) == 2:\n self.__compare_button['state'] = 'normal'\n else:\n self.__compare_button['state'] = 'disabled'\n\n if len(plot_ids) > 0:\n graph_from_plot_ids(self.__graph, plot_ids, self.__marker_combo.get(), self.__line_combo.get(),\n float(self.__entry_line_width.get()))\n self.__canvas.draw()\n else:\n self.clear_all_plots()\n\n def add_title(self):\n \"\"\" Add title to the graph\"\"\"\n title = simpledialog.askstring(\"Graph Title\", \"What is the title of the Graph?\", parent=self.top)\n if title is not None and title != \"\":\n graph_add_title(self.__graph, title)\n self.__canvas.draw()\n\n def quit(self):\n \"\"\" Destroy the frame \"\"\"\n self.__parent.reset_plot_frame(self.__plot_frame_id)\n self.top.destroy()\n", "sub_path": "ranalysis/gui/frame/plotframe.py", "file_name": "plotframe.py", "file_ext": "py", "file_size_in_byte": 10168, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "tkinter.Frame", "line_number": 20, "usage_type": "attribute"}, {"api_name": "tkinter.Toplevel", "line_number": 27, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 49, "usage_type": "call"}, {"api_name": "tkinter.StringVar", "line_number": 50, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 61, "usage_type": "call"}, {"api_name": "tkinter.GROOVE", "line_number": 61, "usage_type": "attribute"}, {"api_name": "tkinter.LEFT", "line_number": 62, "usage_type": "attribute"}, {"api_name": "tkinter.Y", "line_number": 62, "usage_type": "attribute"}, {"api_name": "tkinter.Frame", "line_number": 64, "usage_type": "call"}, {"api_name": "tkinter.RIGHT", "line_number": 65, "usage_type": "attribute"}, {"api_name": "tkinter.Y", "line_number": 65, "usage_type": "attribute"}, {"api_name": "tkinter.Frame", "line_number": 67, "usage_type": "call"}, {"api_name": "tkinter.GROOVE", "line_number": 67, "usage_type": "attribute"}, {"api_name": "tkinter.RIGHT", "line_number": 68, "usage_type": "attribute"}, {"api_name": "tkinter.Y", "line_number": 68, "usage_type": "attribute"}, {"api_name": "tkinter.Frame", "line_number": 70, "usage_type": "call"}, {"api_name": "tkinter.LEFT", "line_number": 71, "usage_type": "attribute"}, {"api_name": "tkinter.BOTH", "line_number": 71, "usage_type": "attribute"}, {"api_name": "tkinter.ttk.Button", "line_number": 74, "usage_type": "call"}, {"api_name": "tkinter.ttk.Button", "line_number": 76, "usage_type": "call"}, {"api_name": "tkinter.ttk.Button", "line_number": 78, "usage_type": "call"}, {"api_name": "tkinter.ttk.Button", "line_number": 81, "usage_type": "call"}, {"api_name": "tkinter.ttk.Button", "line_number": 82, "usage_type": "call"}, {"api_name": "tkinter.DISABLED", "line_number": 82, "usage_type": "attribute"}, {"api_name": "tkinter.ttk.Label", "line_number": 86, "usage_type": "call"}, {"api_name": "tkinter.W", "line_number": 87, "usage_type": "attribute"}, {"api_name": "tkinter.ttk.Label", "line_number": 88, "usage_type": "call"}, {"api_name": "tkinter.W", "line_number": 89, "usage_type": "attribute"}, {"api_name": "tkinter.ttk.Label", "line_number": 90, "usage_type": "call"}, {"api_name": "tkinter.W", "line_number": 91, "usage_type": "attribute"}, {"api_name": "tkinter.Listbox", "line_number": 94, "usage_type": "call"}, {"api_name": "tkinter.MULTIPLE", "line_number": 94, "usage_type": "attribute"}, {"api_name": "tkinter.StringVar", "line_number": 98, "usage_type": "call"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.style.available", "line_number": 99, "usage_type": "attribute"}, {"api_name": "matplotlib.style", "line_number": 99, "usage_type": "name"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 103, "usage_type": "call"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 110, "usage_type": "call"}, {"api_name": "tkinter.ttk.Entry", "line_number": 117, "usage_type": "call"}, {"api_name": "tkinter.END", "line_number": 118, "usage_type": "attribute"}, {"api_name": "tkinter.Y", "line_number": 122, "usage_type": "attribute"}, {"api_name": "tkinter.Frame", "line_number": 143, "usage_type": "call"}, {"api_name": "tkinter.GROOVE", "line_number": 143, "usage_type": "attribute"}, {"api_name": "tkinter.RIGHT", "line_number": 144, "usage_type": "attribute"}, {"api_name": "tkinter.BOTH", "line_number": 144, "usage_type": "attribute"}, {"api_name": "matplotlib.style.use", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.style", "line_number": 146, "usage_type": "name"}, {"api_name": "matplotlib.figure.Figure", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.backends.backend_tkagg.FigureCanvasTkAgg", "line_number": 151, "usage_type": "call"}, {"api_name": "tkinter.TOP", "line_number": 153, "usage_type": "attribute"}, {"api_name": "tkinter.BOTH", "line_number": 153, "usage_type": "attribute"}, {"api_name": "matplotlib.backends.backend_tkagg.NavigationToolbar2Tk", "line_number": 155, "usage_type": "call"}, {"api_name": "tkinter.TOP", "line_number": 157, "usage_type": "attribute"}, {"api_name": "tkinter.BOTH", "line_number": 157, "usage_type": "attribute"}, {"api_name": "ranalysis.log.loghandler.logger.log", "line_number": 162, "usage_type": "call"}, {"api_name": "ranalysis.log.loghandler.logger", "line_number": 162, "usage_type": "name"}, {"api_name": "logging.INFO", "line_number": 162, "usage_type": "attribute"}, {"api_name": "ranalysis.log.loghandler.logger.log", "line_number": 168, "usage_type": "call"}, {"api_name": "ranalysis.log.loghandler.logger", "line_number": 168, "usage_type": "name"}, {"api_name": "logging.INFO", "line_number": 168, "usage_type": "attribute"}, {"api_name": "ranalysis.log.loghandler.logger.log", "line_number": 173, "usage_type": "call"}, {"api_name": "ranalysis.log.loghandler.logger", "line_number": 173, "usage_type": "name"}, {"api_name": "logging.ERROR", "line_number": 173, "usage_type": "attribute"}, {"api_name": "ranalysis.plot.plotcreator.PlotCreator.get_instance", "line_number": 181, "usage_type": "call"}, {"api_name": "ranalysis.plot.plotcreator.PlotCreator", "line_number": 181, "usage_type": "name"}, {"api_name": "ranalysis.plot.plotcreator.PlotCreator.get_instance", "line_number": 182, "usage_type": "call"}, {"api_name": "ranalysis.plot.plotcreator.PlotCreator", "line_number": 182, "usage_type": "name"}, {"api_name": "ranalysis.gui.dialog.plotdialog.PlotDialog", "line_number": 184, "usage_type": "call"}, {"api_name": "ranalysis.log.loghandler.logger.log", "line_number": 187, "usage_type": "call"}, {"api_name": "ranalysis.log.loghandler.logger", "line_number": 187, "usage_type": "name"}, {"api_name": "logging.ERROR", "line_number": 187, "usage_type": "attribute"}, {"api_name": "ranalysis.log.loghandler.logger.log", "line_number": 189, "usage_type": "call"}, {"api_name": "ranalysis.log.loghandler.logger", "line_number": 189, "usage_type": "name"}, {"api_name": "logging.ERROR", "line_number": 189, "usage_type": "attribute"}, {"api_name": "ranalysis.log.loghandler.logger.log", "line_number": 192, "usage_type": "call"}, {"api_name": "ranalysis.log.loghandler.logger", "line_number": 192, "usage_type": "name"}, {"api_name": "logging.INFO", "line_number": 192, "usage_type": "attribute"}, {"api_name": "tkinter.END", "line_number": 193, "usage_type": "attribute"}, {"api_name": "ranalysis.log.loghandler.logger.log", "line_number": 197, "usage_type": "call"}, {"api_name": "ranalysis.log.loghandler.logger", "line_number": 197, "usage_type": "name"}, {"api_name": "logging.INFO", "line_number": 197, "usage_type": "attribute"}, {"api_name": "ranalysis.plot.graph.graph_clear", "line_number": 198, "usage_type": "call"}, {"api_name": "tkinter.END", "line_number": 199, "usage_type": "attribute"}, {"api_name": "ranalysis.log.loghandler.logger.log", "line_number": 204, "usage_type": "call"}, {"api_name": "ranalysis.log.loghandler.logger", "line_number": 204, "usage_type": "name"}, {"api_name": "logging.INFO", "line_number": 204, "usage_type": "attribute"}, {"api_name": "tkinter.END", "line_number": 205, "usage_type": "attribute"}, {"api_name": "ranalysis.plot.graph.graph_clear", "line_number": 209, "usage_type": "call"}, {"api_name": "ranalysis.plot.graph.graph_from_plot_ids", "line_number": 220, "usage_type": "call"}, {"api_name": "tkinter.simpledialog.askstring", "line_number": 228, "usage_type": "call"}, {"api_name": "tkinter.simpledialog", "line_number": 228, "usage_type": "name"}, {"api_name": "ranalysis.plot.graph.graph_add_title", "line_number": 230, "usage_type": "call"}]}
+{"seq_id": "316897222", "text": "import gamelib\nimport random\nimport math\nimport warnings\nfrom sys import maxsize\nimport json\n\nimport random\n\n\n\"\"\"\nMost of the algo code you write will be in this file unless you create new\nmodules yourself. Start by modifying the 'on_turn' function.\n\nAdvanced strategy tips:\n\n - You can analyze action frames by modifying on_action_frame function\n\n - The GameState.map object can be manually manipulated to create hypothetical\n board states. Though, we recommended making a copy of the map to preserve\n the actual current map state.\n\"\"\"\n\nclass AlgoStrategy(gamelib.AlgoCore):\n def __init__(self):\n super().__init__()\n seed = random.randrange(maxsize)\n random.seed(seed)\n gamelib.debug_write('Random seed: {}'.format(seed))\n\n def on_game_start(self, config):\n \"\"\"\n Read in config and perform any initial setup here\n \"\"\"\n gamelib.debug_write('Configuring your custom algo strategy...')\n self.config = config\n global FILTER, ENCRYPTOR, DESTRUCTOR, PING, EMP, SCRAMBLER\n FILTER = config[\"unitInformation\"][0][\"shorthand\"]\n ENCRYPTOR = config[\"unitInformation\"][1][\"shorthand\"]\n DESTRUCTOR = config[\"unitInformation\"][2][\"shorthand\"]\n PING = config[\"unitInformation\"][3][\"shorthand\"]\n EMP = config[\"unitInformation\"][4][\"shorthand\"]\n SCRAMBLER = config[\"unitInformation\"][5][\"shorthand\"]\n # This is a good place to do initial setup\n self.scored_on_locations = []\n\n\n\n\n def on_turn(self, turn_state):\n \"\"\"\n This function is called every turn with the game state wrapper as\n an argument. The wrapper stores the state of the arena and has methods\n for querying its state, allocating your current resources as planned\n unit deployments, and transmitting your intended deployments to the\n game engine.\n \"\"\"\n game_state = gamelib.GameState(self.config, turn_state)\n gamelib.debug_write('Performing turn {} of your custom algo strategy'.format(game_state.turn_number))\n game_state.suppress_warnings(True) #Comment or remove this line to enable warnings.\n\n self.starter_strategy(game_state)\n\n game_state.submit_turn()\n\n\n \"\"\"\n NOTE: All the methods after this point are part of the sample starter-algo\n strategy and can safely be replaced for your custom algo.\n \"\"\"\n\n\n def starter_strategy(self, game_state):\n \"\"\"\n For defense we will use a spread out layout and some Scramblers early on.\n We will place destructors near locations the opponent managed to score on.\n For offense we will use long range EMPs if they place stationary units near the enemy's front.\n If there are no stationary units to attack in the front, we will send Pings to try and score quickly.\n \"\"\"\n # First, place basic defenses\n self.build_defences(game_state)\n # self.build_reactive_defense(game_state)\n self.stall_with_scramblers(game_state)\n #self.build_wall_from_left(game_state)\n\n if game_state.get_resource(game_state.BITS,0) > 7:\n empt_attack = True if random.randint(0,3) == 1 else False\n percent = .1 + (game_state.get_resource(game_state.BITS,0) - 10)*.08\n spawn_location_options = [[13, 0]]\n if(random.uniform(0,1) < percent):\n game_state.attempt_spawn(EMP if empt_attack else PING, spawn_location_options, 1000)\n\n def build_defences(self, game_state):\n encryptor_locations = [[2,13],[3,12],[4,11],[5,10],[6,9],[7,8],[8,7],[9,6],[10,5],[11,4],[12,3],[13,2],[14,3],[15,4],[16,5],[17,6],[18,7],[19,8],[20,9],[21,10],[23,12]]\n if(game_state.turn_number < 5):\n encryptor_locations += [[24,13],[25,13]]\n filter_locations = [[26,13],[27,13],[0,13],[1,13]]\n if(game_state.turn_number >5):\n filter_locations += [[24,13],[25,13]]\n game_state.attempt_spawn(ENCRYPTOR,encryptor_locations)\n game_state.attempt_spawn(FILTER,filter_locations)\n\n #game_state.attempt_spawn(FILTER, random.sample(filter_locations,4))\n\n # Place destructors that attack enemy units\n #simple heuristics. IDK if this is at all helpful...\n if game_state.turn_number >1:\n destructor__xtra_locations = [[17,7],[16,6],[15,5],[14,4],[13,3],[12,2],[11,3]]\n destructor_locations = [[22, 12], [20, 10], [19, 9], [18, 8],[17,7]]\n #add in\n if game_state.get_resource(game_state.CORES,0) > 20:\n destructor_locations += destructor__xtra_locations\n if game_state.get_resource(game_state.CORES,0) > 40:\n destructor_locations += [[10,4],[9,5],[8,6],[7,7],[6,8],[5,9],[4,10],[3,11],[2,12]]\n # attempt_spawn will try to spawn units if we have resources, and will check if a blocking unit is already there\n elif game_state.get_resource(game_state.CORES,0) > 10:\n len_dest = len(destructor__xtra_locations)\n choice1 = random.randint(0, len_dest-1)\n destructor_locations.append(destructor__xtra_locations[choice1])\n choice2 = -1\n while(True and len_dest > 1):\n choice2 = random.randint(0, len_dest-1)\n if(choice1 != choice2):\n destructor_locations.append(destructor__xtra_locations[choice2])\n break\n\n # destructor_locations += random.choice(destructor__xtra_locations,2)\n\n\n game_state.attempt_spawn(DESTRUCTOR, destructor_locations)\n\n filter_locations2 = [[1,12],[2,12],[24,12],[25,12],[26,12]]\n game_state.attempt_spawn(FILTER,filter_locations2)\n\n filter_locations3 = [[24,10],[25,11],[26,12],[23,9],[22,8],[21,7],[20,6]]\n game_state.attempt_spawn(FILTER,filter_locations3)\n \n if game_state.get_resource(game_state.CORES,0) > 20: \n\n add_filter_rand = []\n filter_locations4 = [[1,12],[2,11],[3,10],[4,9],[5,8],[6,7],[7,6],[8,5],[9,4],[10,3],[11,2],[12,1],[14,0],[18,4],[17,3],[16,2],[15,1]]\n len_dest = len(filter_locations4)\n choice1 = random.randint(0, len_dest-1)\n add_filter_rand.append(filter_locations4[choice1])\n choice2 = -1\n while(True and len_dest > 1):\n choice2 = random.randint(0, len_dest-1)\n if(choice1 != choice2):\n add_filter_rand.append(filter_locations4[choice2])\n break\n game_state.attempt_spawn(FILTER,add_filter_rand)\n\n\n\n #secondary_filter_locations = [x + 1 for x in filter_locations[1]]\n #game_state.attempt_spawn(FILTER, secondary_filter_locations)\n\n\n def build_reactive_defense(self, game_state):\n \"\"\"\n This function builds reactive defenses based on where the enemy scored on us from.\n We can track where the opponent scored by looking at events in action frames\n as shown in the on_action_frame function\n \"\"\"\n for location in self.scored_on_locations:\n # Build destructor one space above so that it doesn't block our own edge spawn locations\n build_location = [location[0], location[1]+1]\n game_state.attempt_spawn(DESTRUCTOR, build_location)\n\n def build_wall_from_left(self, game_state):\n filter_locations = [[i, 13] for i in range(28)]\n game_state.attempt_spawn(FILTER, filter_locations)\n\n # def check_left_hall_hole(self, game_state):\n # start_location = [15,1]\n # path = game_state.find_path_to_edge(start_location)\n # count = 0\n # for path_location in path:\n # count += 1\n # if(count > 3):\n # return False\n # if(path_location[0] <= 13):\n # return True\n # return False\n\n def stall_with_scramblers(self, game_state):\n \"\"\"\n Send out Scramblers at random locations to defend our base from enemy moving units.\n \"\"\"\n # We can spawn moving units on our edges so a list of all our edge locations\n # friendly_edges = game_state.game_map.get_edge_locations(game_state.game_map.BOTTOM_LEFT) + game_state.game_map.get_edge_locations(game_state.game_map.BOTTOM_RIGHT)\n\n # # Remove locations that are blocked by our own firewalls\n # # since we can't deploy units there.\n # deploy_locations = self.filter_blocked_locations(friendly_edges, game_state)\n\n deploy_locations = [[19,5]]\n num_scramblers = 1\n opp_bits = game_state.get_resource(game_state.BITS,1)\n if(opp_bits >= 7 and opp_bits < 9):\n num_scramblers = 2\n elif( opp_bits >= 9 and opp_bits < 14):\n num_scramblers = 4\n elif(opp_bits >= 14):\n num_scramblers = 6\n game_state.attempt_spawn(SCRAMBLER, deploy_locations,num_scramblers)\n\n\n\n\n def emp_line_strategy(self, game_state):\n \"\"\"\n Build a line of the cheapest stationary unit so our EMP's can attack from long range.\n \"\"\"\n # First let's figure out the cheapest unit\n # We could just check the game rules, but this demonstrates how to use the GameUnit class\n #stationary_units = [FILTER, DESTRUCTOR, ENCRYPTOR]\n #cheapest_unit = FILTER\n # for unit in stationary_units:\n # unit_class = gamelib.GameUnit(unit, game_state.config)\n # if unit_class.cost < gamelib.GameUnit(cheapest_unit, game_state.config).cost:\n # cheapest_unit = unit\n\n # Now let's build out a line of stationary units. This will prevent our EMPs from running into the enemy base.\n # Instead they will stay at the perfect distance to attack the front two rows of the enemy base.\n for x in range(27, 5, -1):\n game_state.attempt_spawn(FILTER, [x, 11])\n\n # Now spawn EMPs next to the line\n # By asking attempt_spawn to spawn 1000 units, it will essentially spawn as many as we have resources for\n game_state.attempt_spawn(EMP, [24, 10], 1000)\n\n\n #builds a defensive destruct-filter block\n def build_spot_defense(self,game_state,location):\n filter_deploy_locations = [[location[0]-1,location[1]],[location[0],location[1]+1],[location[0]+1,location[1]]]\n game_state.attempt_spawn(FILTER, filter_deploy_locations)\n destructor_deploy_locations = location\n game_state.attempt_spawn(DESTRUCTOR,destructor_deploy_locations)\n\n def least_damage_spawn_location(self, game_state, location_options):\n \"\"\"\n This function will help us guess which location is the safest to spawn moving units from.\n It gets the path the unit will take then checks locations on that path to\n estimate the path's damage risk.\n \"\"\"\n damages = []\n # Get the damage estimate each path will take\n for location in location_options:\n path = game_state.find_path_to_edge(location)\n damage = 0\n for path_location in path:\n # Get number of enemy destructors that can attack the final location and multiply by destructor damage\n damage += len(game_state.get_attackers(path_location, 0)) * gamelib.GameUnit(DESTRUCTOR, game_state.config).damage\n damages.append(damage)\n\n # Now just return the location that takes the least damage\n return location_options[damages.index(min(damages))]\n\n def detect_enemy_unit(self, game_state, unit_type=None, valid_x = None, valid_y = None):\n total_units = 0\n for location in game_state.game_map:\n if game_state.contains_stationary_unit(location):\n for unit in game_state.game_map[location]:\n if unit.player_index == 1 and (unit_type is None or unit.unit_type == unit_type) and (valid_x is None or location[0] in valid_x) and (valid_y is None or location[1] in valid_y):\n total_units += 1\n return total_units\n\n def filter_blocked_locations(self, locations, game_state):\n filtered = []\n for location in locations:\n if not game_state.contains_stationary_unit(location):\n filtered.append(location)\n return filtered\n\n def on_action_frame(self, turn_string):\n \"\"\"\n This is the action frame of the game. This function could be called\n hundreds of times per turn and could slow the algo down so avoid putting slow code here.\n Processing the action frames is complicated so we only suggest it if you have time and experience.\n Full doc on format of a game frame at: https://docs.c1games.com/json-docs.html\n \"\"\"\n # Let's record at what position we get scored on\n state = json.loads(turn_string)\n events = state[\"events\"]\n breaches = events[\"breach\"]\n for breach in breaches:\n location = breach[0]\n unit_owner_self = True if breach[4] == 1 else False\n # When parsing the frame data directly,\n # 1 is integer for yourself, 2 is opponent (StarterKit code uses 0, 1 as player_index instead)\n if not unit_owner_self:\n gamelib.debug_write(\"Got scored on at: {}\".format(location))\n self.scored_on_locations.append(location)\n gamelib.debug_write(\"All locations: {}\".format(self.scored_on_locations))\n\n\nif __name__ == \"__main__\":\n algo = AlgoStrategy()\n algo.start()\n", "sub_path": "algo_strategy.py", "file_name": "algo_strategy.py", "file_ext": "py", "file_size_in_byte": 13600, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "gamelib.AlgoCore", "line_number": 24, "usage_type": "attribute"}, {"api_name": "random.randrange", "line_number": 27, "usage_type": "call"}, {"api_name": "sys.maxsize", "line_number": 27, "usage_type": "argument"}, {"api_name": "random.seed", "line_number": 28, "usage_type": "call"}, {"api_name": "gamelib.debug_write", "line_number": 29, "usage_type": "call"}, {"api_name": "gamelib.debug_write", "line_number": 35, "usage_type": "call"}, {"api_name": "gamelib.GameState", "line_number": 58, "usage_type": "call"}, {"api_name": "gamelib.debug_write", "line_number": 59, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 87, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 90, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 118, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 122, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 143, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 147, "usage_type": "call"}, {"api_name": "gamelib.GameUnit", "line_number": 254, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 284, "usage_type": "call"}, {"api_name": "gamelib.debug_write", "line_number": 293, "usage_type": "call"}, {"api_name": "gamelib.debug_write", "line_number": 295, "usage_type": "call"}]}
+{"seq_id": "213045562", "text": "#! /usr/bin/python\n\nimport sys\nimport urllib2\nimport urllib\nimport thread\nimport pygst\npygst.require(\"0.10\")\nimport gst\nimport json\nimport random\nimport cookielib\nimport keybinder\nfrom PyQt5 import QtCore,QtWidgets,QtGui\n\napp = QtWidgets.QApplication(sys.argv)\ncookie = cookielib.CookieJar()\n\nclass ClickableLable(QtWidgets.QLabel):\n captchaClicked = QtCore.pyqtSignal()\n def mouseReleaseEvent(self,event):\n self.captchaClicked.emit()\n\nclass LoginClass(QtWidgets.QDialog):\n def __init__(self):\n QtWidgets.QWidget.__init__(self)\n\n usernameLabel = QtWidgets.QLabel('username:')\n passwdLabel = QtWidgets.QLabel('password:')\n idcodeLabel = QtWidgets.QLabel('idcode: ')\n self.usernameText= QtWidgets.QLineEdit()\n self.usernameText.setFixedSize(150,24)\n self.passwdText = QtWidgets.QLineEdit()\n self.passwdText.setFixedSize(150,24)\n self.passwdText.setEchoMode(QtWidgets.QLineEdit.Password)\n self.idcodeText = QtWidgets.QLineEdit()\n self.idcodeText.setFixedSize(60,24)\n self.idcodePic = ClickableLable()\n self.idcodePic.setFixedSize(70,24)\n self.idcodePic.setScaledContents(True)\n self.idcodePic.captchaClicked.connect(self.getCaptcha)\n self.feedBack = QtWidgets.QLabel()\n\n button = QtWidgets.QPushButton('submit')\n button.clicked.connect(self.onSubmitClick)\n\n layout = QtWidgets.QVBoxLayout()\n contentLayout = QtWidgets.QHBoxLayout()\n labelLayout = QtWidgets.QVBoxLayout()\n editLayout = QtWidgets.QVBoxLayout()\n captchaLayout = QtWidgets.QHBoxLayout()\n submitLayout = QtWidgets.QHBoxLayout()\n\n spacer = QtWidgets.QLabel()\n spacer.setSizePolicy(QtWidgets.QSizePolicy.Expanding,QtWidgets.QSizePolicy.Expanding)\n\n labelLayout.addWidget(usernameLabel)\n labelLayout.addWidget(passwdLabel)\n labelLayout.addWidget(idcodeLabel)\n captchaLayout.addWidget(self.idcodeText)\n captchaLayout.addWidget(spacer)\n captchaLayout.addWidget(self.idcodePic)\n editLayout.addWidget(self.usernameText)\n editLayout.addWidget(self.passwdText)\n editLayout.addLayout(captchaLayout)\n contentLayout.addLayout(labelLayout)\n contentLayout.addLayout(editLayout)\n submitLayout.addWidget(self.feedBack)\n submitLayout.addWidget(spacer)\n submitLayout.addWidget(button)\n\n layout.addLayout(contentLayout)\n layout.addLayout(submitLayout)\n self.setLayout(layout)\n self.getCaptcha()\n\n def closeEvent(self,event):\n self.reject()\n def getCaptchaInThread(self):\n thread.start_new_thread(self.getCaptcha,())\n\n def getCaptcha(self):\n image,self.captcha_id = HttpRequest().getCaptchaRequest()\n self.idcodePic.setPixmap(image)\n\n def onSubmitClick(self):\n url = 'http://douban.fm/j/login'\n body = {'task': 'sync_channel_list', 'source': 'radio', 'captcha_solution': self.idcodeText.text(), 'alias': self.usernameText.text(), 'form_password': self.passwdText.text(), 'captcha_id': self.captcha_id }\n content = HttpRequest().loginRequest(url,body)\n if json.loads(content)['r'] == 0:\n self.accept()\n else:\n self.feedBack.setText(json.loads(content)['err_msg'])\n self.getCaptcha()\n\nclass MainWindow(QtWidgets.QWidget):\n def __init__(self):\n QtWidgets.QWidget.__init__(self)\n self.player = Player()\n self.length = 1\n self.like = 0\n layout = QtWidgets.QHBoxLayout()\n mainLayout = QtWidgets.QVBoxLayout()\n self.processBar = QtWidgets.QProgressBar()\n self.processBar.setValue(0)\n self.processBar.setTextVisible(False)\n self.processBar.setFixedHeight(5)\n self.pic = QtWidgets.QLabel()\n self.pic.setFixedSize(150,150)\n self.pic.setScaledContents(True)\n self.pic.setPixmap(QtGui.QPixmap('./dbfm.png'))\n buttonLayout = QtWidgets.QVBoxLayout()\n self.playButton = QtWidgets.QPushButton(\"(&P)pause\")\n nextButton = QtWidgets.QPushButton(\"(&N)next\")\n self.likeButton =QtWidgets.QPushButton('(&L)like')\n hateButton = QtWidgets.QPushButton('(&H)hate')\n self.playButton.clicked.connect(self.onPlayButtonClick)\n nextButton.clicked.connect(self.onNextButtonClick)\n self.likeButton.clicked.connect(self.likes)\n hateButton.clicked.connect(self.hate)\n buttonLayout.addWidget(self.playButton)\n buttonLayout.addWidget(nextButton)\n buttonLayout.addWidget(self.likeButton)\n buttonLayout.addWidget(hateButton)\n icon = QtGui.QIcon('./dbfm.png')\n self.setWindowIcon(icon)\n self.trayIcon = QtWidgets.QSystemTrayIcon(self)\n self.trayIcon.activated.connect(self.showMainWindow)\n trayIconMenu = QtWidgets.QMenu(self)\n minimizeAction = QtWidgets.QAction('minisize', self,triggered=self.hide)\n restoreAction = QtWidgets.QAction('restore', self,triggered=self.showNormal)\n quitAction = QtWidgets.QAction('exit', self,triggered=QtWidgets.qApp.quit)\n self.likeAction = QtWidgets.QAction('like',self,triggered=self.likes)\n self.playAction = QtWidgets.QAction('pause',self,triggered=self.onPlayButtonClick)\n nextAction = QtWidgets.QAction('next',self,triggered=self.onNextButtonClick)\n\n trayIconMenu.addAction(self.playAction)\n trayIconMenu.addAction(nextAction)\n trayIconMenu.addAction(self.likeAction)\n trayIconMenu.addSeparator()\n trayIconMenu.addAction(minimizeAction)\n trayIconMenu.addAction(restoreAction)\n trayIconMenu.addSeparator()\n trayIconMenu.addAction(quitAction)\n self.trayIcon.setContextMenu(trayIconMenu)\n self.trayIcon.setIcon(icon)\n self.trayIcon.setToolTip('douban FM')\n self.trayIcon.show()\n\n keybinder.bind('n',self.onNextButtonClick)\n keybinder.bind('r',self.justLike)\n keybinder.bind('b',self.hate)\n keybinder.bind('p',self.onPlayButtonClick)\n keybinder.bind('t',self.showNormal)\n\n layout.addWidget(self.pic)\n layout.addLayout(buttonLayout)\n mainLayout.addLayout(layout)\n mainLayout.addWidget(self.processBar)\n timer = QtCore.QTimer(self)\n timer.timeout.connect(self.test)\n timer.start(1000)\n self.setLayout(mainLayout)\n self.play()\n\n def showMainWindow(self,mode):\n if mode == 2:\n if self.isVisible() == True:\n self.setVisible(False)\n else:\n self.setVisible(True)\n self.move((QtWidgets.QApplication.desktop().width()-main.width())/2,(QtWidgets.QApplication.desktop().height()-main.height())/2)\n\n def closeEvent(self,event):\n self.setVisible(False)\n event.ignore()\n\n def getRandom(self):\n r = str(hex(int(random.random()*1E17)))\n return r[2:12]\n\n def play(self):\n firstPlayUrl = \"http://douban.fm/j/mine/playlist?type=n&sid=&pt=0.0&channel=0&pb=64&from=mainsite&r=%s\" % self.getRandom()\n self.getMusic(firstPlayUrl)\n\n def onPlayButtonClick(self):\n if self.playButton.text() == '(&P)play':\n self.playButton.setText('(&P)pause')\n self.playAction.setText('pause')\n self.player.onPlay()\n else:\n self.playButton.setText('(&P)play')\n self.playAction.setText('play')\n self.player.onPause()\n\n def likes(self):\n if self.likeButton.text() == '(&L)like':\n self.likeButton.setText('(&L)dislike')\n self.likeAction.setText('dislike')\n likeUrl = \"http://douban.fm/j/mine/playlist?type=r&sid=%s&pt=%s&channel=0&pb=64&from=mainsite&r=%s\" % (str(self.sid), str(self.player.getPosition()), self.getRandom())\n else:\n self.likeButton.setText('(&L)like')\n self.likeAction.setText('like')\n likeUrl = \"http://douban.fm/j/mine/playlist?type=u&sid=%s&pt=%s&channel=0&pb=64&from=mainsite&r=%s\" % (str(self.sid), str(self.player.getPosition()), self.getRandom())\n HttpRequest().getRequest(likeUrl)\n\n def justLike(self):\n self.likeButton.setText('(&L)dislike')\n self.likeAction.setText('dislike')\n likeUrl = \"http://douban.fm/j/mine/playlist?type=r&sid=%s&pt=%s&channel=0&pb=64&from=mainsite&r=%s\" % (str(self.sid), str(self.player.getPosition()), self.getRandom())\n HttpRequest().getRequest(likeUrl)\n\n def hate(self):\n hateUrl = \"http://douban.fm/j/mine/playlist?type=b&sid=%s&pt=%s&channel=0&pb=64&from=mainsite&r=%s\" % (str(self.sid), str(self.player.getPosition()), self.getRandom())\n self.getMusic(hateUrl)\n\n def onNextButtonClick(self):\n nextUrl = \"http://douban.fm/j/mine/playlist?type=s&sid=%s&pt=%s&channel=0&pb=64&from=mainsite&r=%s\" % (str(self.sid), str(self.player.getPosition()), self.getRandom())\n self.getMusic(nextUrl)\n\n def getMusic(self,url):\n self.player.onStop()\n title,artist,url,picture,self.sid,self.length,self.like = HttpRequest().analisysMusic(url)\n if self.like == 0:\n self.likeButton.setText('(&L)like')\n self.likeAction.setText('like')\n else:\n self.likeButton.setText('(&L)dislike')\n self.likeAction.setText('dislike')\n self.setPic(picture)\n self.setTitle(artist,title)\n self.player.loadMusic(url)\n self.player.onPlay()\n\n\n def setTitle(self,artist,title):\n self.setWindowTitle(artist + '-' + title)\n self.trayIcon.setToolTip('douban FM\\n' + artist + '-' + title)\n self.trayIcon.showMessage('douban FM',artist + '-' + title)\n\n def setPic(self,url):\n self.pic.setPixmap(HttpRequest().getImageRequest(url))\n\n def setPicInThread(self,url):\n thread.start_new_thread(self.setPic,(url,))\n\n def test(self):\n self.player.queryPosition()\n self.processBar.setValue(self.player.getPosition()/self.length*100)\n if self.player.getPosition() >= self.length:\n self.player.onStop()\n nextUrl = \"http://douban.fm/j/mine/playlist?type=p&sid=%s&pt=%s&channel=0&pb=64&from=mainsite&r=%s\" % (str(self.sid), str(self.length) , self.getRandom())\n self.getMusic(nextUrl)\n\n\nclass Player(QtCore.QObject):\n def __init__(self):\n QtCore.QObject.__init__(self)\n self.player = gst.element_factory_make('playbin','player')\n self.player.set_state(gst.STATE_READY)\n self.position = 0.0\n\n def onPlay(self):\n self.player.set_state(gst.STATE_PLAYING)\n\n def onPause(self):\n self.player.set_state(gst.STATE_PAUSED)\n\n def onStop(self):\n self.player.set_state(gst.STATE_READY)\n\n def loadMusic(self,url):\n self.player.set_property('uri',url)\n\n def queryPosition(self):\n excuteState,currentState,lastState = self.player.get_state()\n if currentState == gst.STATE_PLAYING:\n currentPosition,timeFormat = self.player.query_position()\n currentPosition = currentPosition/1000000000\n self.position = float('%.1f' % currentPosition)\n\n def getPosition(self):\n return self.position\n\nclass HttpRequest():\n def loginRequest(self,url,body):\n opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cookie))\n opener.addheaders = [('Content-type','application/x-www-form-urlencoded')]\n urllib2.install_opener(opener)\n req = urllib2.Request(url,urllib.urlencode(body))\n res = urllib2.urlopen(req)\n return res.read()\n\n def getRequest(self,url):\n opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cookie))\n urllib2.install_opener(opener)\n res = urllib2.urlopen(url)\n return res.read()\n\n def getCaptchaRequest(self):\n url = 'http://douban.fm/j/new_captcha'\n content = self.getRequest(url)\n captcha_id = str(content).replace('\"','')\n url = 'http://douban.fm/misc/captcha?size=m&id=%s' % captcha_id\n return self.getImageRequest(url),captcha_id\n\n def getImageRequest(self,url):\n content = self.getRequest(url)\n file = open('./temp.jpg','w')\n file.write(content)\n file.close()\n image = QtGui.QPixmap('./temp.jpg')\n return image\n\n def analisysMusic(self,url):\n content = self.getRequest(url)\n jsonmap = json.loads(content)\n select = random.randint(0,4)\n title = jsonmap['song'][select]['title']\n artist= jsonmap['song'][select]['artist']\n url = jsonmap['song'][select]['url']\n picture= jsonmap['song'][select]['picture']\n sid = jsonmap['song'][select]['sid']\n length = jsonmap['song'][select]['length']\n like = jsonmap['song'][select]['like']\n return title,artist,url,picture,sid,length,like\n\n\n\nlogintest = LoginClass()\nif logintest.exec_() == 1:\n main = MainWindow()\n main.show()\n main.move((QtWidgets.QApplication.desktop().width()-main.width())/2,(QtWidgets.QApplication.desktop().height()-main.height())/2)\n sys.exit(app.exec_())\n", "sub_path": "doubanqt5.py", "file_name": "doubanqt5.py", "file_ext": "py", "file_size_in_byte": 13174, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "pygst.require", "line_number": 8, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 16, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 16, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 16, "usage_type": "attribute"}, {"api_name": "cookielib.CookieJar", "line_number": 17, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 19, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 19, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 20, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 20, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QDialog", "line_number": 24, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 24, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QWidget.__init__", "line_number": 26, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 26, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 26, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 28, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 28, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 29, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 29, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 30, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 30, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 31, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 31, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 33, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 33, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 35, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 35, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLineEdit", "line_number": 36, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 36, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 42, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 42, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 44, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 44, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 47, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 47, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 48, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 48, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 49, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 49, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 50, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 50, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 51, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 51, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 52, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 52, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 54, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 54, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QSizePolicy", "line_number": 55, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 55, "usage_type": "name"}, {"api_name": "thread.start_new_thread", "line_number": 80, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 90, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 93, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 96, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 96, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QWidget.__init__", "line_number": 98, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 98, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 98, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QHBoxLayout", "line_number": 102, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 102, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 103, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 103, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QProgressBar", "line_number": 104, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 104, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QLabel", "line_number": 108, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 108, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 111, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 111, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QVBoxLayout", "line_number": 112, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 112, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 113, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 113, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 114, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 114, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 115, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 115, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QPushButton", "line_number": 116, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 116, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 125, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 125, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QSystemTrayIcon", "line_number": 127, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 127, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMenu", "line_number": 129, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 129, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QAction", "line_number": 130, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 130, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QAction", "line_number": 131, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 131, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QAction", "line_number": 132, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 132, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.qApp", "line_number": 132, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QAction", "line_number": 133, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 133, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QAction", "line_number": 134, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 134, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QAction", "line_number": 135, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 135, "usage_type": "name"}, {"api_name": "keybinder.bind", "line_number": 150, "usage_type": "call"}, {"api_name": "keybinder.bind", "line_number": 151, "usage_type": "call"}, {"api_name": "keybinder.bind", "line_number": 152, "usage_type": "call"}, {"api_name": "keybinder.bind", "line_number": 153, "usage_type": "call"}, {"api_name": "keybinder.bind", "line_number": 154, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QTimer", "line_number": 160, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 160, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QApplication.desktop", "line_number": 172, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 172, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 172, "usage_type": "name"}, {"api_name": "random.random", "line_number": 179, "usage_type": "call"}, {"api_name": "thread.start_new_thread", "line_number": 245, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QObject", "line_number": 256, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 256, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QObject.__init__", "line_number": 258, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QObject", "line_number": 258, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 258, "usage_type": "name"}, {"api_name": "gst.element_factory_make", "line_number": 259, "usage_type": "call"}, {"api_name": "gst.STATE_READY", "line_number": 260, "usage_type": "attribute"}, {"api_name": "gst.STATE_PLAYING", "line_number": 264, "usage_type": "attribute"}, {"api_name": "gst.STATE_PAUSED", "line_number": 267, "usage_type": "attribute"}, {"api_name": "gst.STATE_READY", "line_number": 270, "usage_type": "attribute"}, {"api_name": "gst.STATE_PLAYING", "line_number": 277, "usage_type": "attribute"}, {"api_name": "urllib2.build_opener", "line_number": 287, "usage_type": "call"}, {"api_name": "urllib2.HTTPCookieProcessor", "line_number": 287, "usage_type": "call"}, {"api_name": "urllib2.install_opener", "line_number": 289, "usage_type": "call"}, {"api_name": "urllib2.Request", "line_number": 290, "usage_type": "call"}, {"api_name": "urllib.urlencode", "line_number": 290, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 291, "usage_type": "call"}, {"api_name": "urllib2.build_opener", "line_number": 295, "usage_type": "call"}, {"api_name": "urllib2.HTTPCookieProcessor", "line_number": 295, "usage_type": "call"}, {"api_name": "urllib2.install_opener", "line_number": 296, "usage_type": "call"}, {"api_name": "urllib2.urlopen", "line_number": 297, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QPixmap", "line_number": 312, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 312, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 317, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 318, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication.desktop", "line_number": 334, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 334, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 334, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 335, "usage_type": "call"}]}
+{"seq_id": "271079737", "text": "import os\nimport cv2\nimport shutil\nimport argparse\nimport logging\nfrom flip_utils import flip_samples\n\n\ndef _parse_args():\n parser = argparse.ArgumentParser()\n\n parser.add_argument(\"--categoried-base-path\", type=str,\n default=\"/ssd4/zhangyiyang/data/AR/categoried/0428\")\n parser.add_argument(\"--videos-dir-name\", type=str, default=\"videos\")\n parser.add_argument(\"--ffmpeg\", action=\"store_true\", default=False)\n parser.add_argument(\"--fps\", type=int, default=10)\n parser.add_argument(\"--tmp-video\", type=str,\n default=\"./test-categoried.avi\")\n parser.add_argument(\"--category-file-path\", type=str,\n default=\"/ssd4/zhangyiyang/data/AR/label/category.txt\")\n\n # to labels 相关\n parser.add_argument(\"--to-frames-dir-name\", type=str, default=\"frames\")\n parser.add_argument(\"--to-labels-file-name\",\n type=str, default=\"to_label.txt\")\n parser.add_argument(\"--to-labels-file-append\", action=\"store_true\")\n parser.add_argument(\"--global-to-labels-dir\", type=str,\n default=\"/ssd4/zhangyiyang/data/AR/label/summary\")\n parser.add_argument(\"--img-format\", type=str, default=\"{:05d}.jpg\")\n\n # flip\n parser.add_argument(\"--flip\", action=\"store_true\")\n parser.add_argument(\"--flip-frames-dir-name\", type=str,\n default=\"flip_frames\")\n parser.add_argument(\"--flip-to-labels-file-name\",\n type=str, default=\"flip_to_label.txt\")\n\n return parser.parse_args()\n\n\ndef _convert_fps(source_fps, target_fps):\n assert source_fps >= target_fps, \"source fps must larger than target fps\"\n\n interval = int(source_fps * 1.0 / target_fps)\n return [i*interval for i in range(int(source_fps/interval))]\n\n\ndef _handle_single_video(video_path, # 输入视频绝对路径\n cur_idx, # 当前视频对应编号\n to_label_file, # TSM标签文件 writer,用于构建 to_label.txt\n category_id, # 当前视频代表的行为类别编号,\n ffmpeg, # 是否使用 ffmpeg 来提取帧\n cur_frames_dir, # frames 保存路径\n img_format,\n fps,\n tmp_video,\n ):\n logging.info(\"start handling {}\".format(video_path))\n if not os.path.exists(video_path):\n logging.warn(\"{} doesn't exist.\".format(video_path))\n return\n\n # ffmpeg 提取帧\n if ffmpeg:\n # {:06d}.jpg -> %06d.jpg\n fmt = img_format.replace(\"{\", \"\") \\\n .replace(\"}\", \"\").replace(\":\", \"%\")\n raw_cmd = [\n 'ffmpeg',\n '-i', '{}',\n '-r', '{}',\n '-threads 1 -vf scale=-1:256 -q:v 0',\n '\"{}/{}\"'\n ]\n cmd = (\" \".join(raw_cmd))\\\n .format(video_path, fps, cur_frames_dir, fmt)\n os.system(cmd)\n to_label_file.write(cur_frames_dir +\n \" \" + str(len(os.listdir(cur_frames_dir))) +\n \" \" + str(category_id) + \"\\n\")\n return\n\n # opencv 提取帧\n\n # 转换输入视频格式\n cmd = \"ffmpeg -i {} -q:v 6 {}\".format(video_path, tmp_video)\n if os.path.exists(tmp_video):\n os.remove(tmp_video)\n os.system(cmd)\n cap = cv2.VideoCapture(tmp_video)\n source_fps = int(cap.get(cv2.CAP_PROP_FPS))\n\n # 有一个问题需要处理:\n # 输入视频文件的fps和我们需要的fps是不同的\n # 一般来说,视频的fps大于我们需要的fps\n # 所以,需要在视频的fps中选择我们需要的若干帧图像\n # 下面这个函数就是选择的帧的编号\n ids = _convert_fps(source_fps, fps)\n\n # 分别读取每一帧,然后分别处理\n id = -1\n file_name_id = 0\n frame_cnt = 0\n while True:\n flag, frame = cap.read()\n if not flag:\n break\n frame_cnt += 1\n id += 1\n if id >= source_fps:\n id = 0\n\n if id in ids:\n # 如果当前帧需要保存,则要保存到目标文件夹中\n file_name_id += 1\n img_name = img_format.format(file_name_id)\n cv2.imwrite(os.path.join(cur_frames_dir, img_name), frame)\n to_label_file.write(cur_frames_dir + \" \" + str(file_name_id) +\n \" \" + str(category_id) + \"\\n\")\n cap.release()\n if os.path.exists(tmp_video):\n os.remove(tmp_video)\n\n\ndef _get_start_id(cur_dir):\n max_id = 0\n for file_name in os.listdir(cur_dir):\n try:\n idx = int(file_name)\n max_id = max(idx, max_id)\n except:\n pass\n return max_id + 1\n\n\ndef main(args):\n # 1. 构建相关文件夹路径,并验证\n src_videos_dir = os.path.join(\n args.categoried_base_path,\n args.videos_dir_name,\n )\n assert os.path.exists(src_videos_dir)\n to_frames_dir = os.path.join(\n args.categoried_base_path,\n args.to_frames_dir_name,\n )\n if not os.path.exists(to_frames_dir):\n os.makedirs(to_frames_dir)\n cur_idx = _get_start_id(to_frames_dir)\n\n # 2. 获取类别信息\n with open(args.category_file_path, \"r\") as f:\n categories = f.readlines()\n categories = [c.replace(\"\\n\", \"\") for c in categories]\n category_to_id = {c: idx for idx, c in enumerate(categories)}\n\n # 3. 获取 to labels 文件,并以此遍历每个子文件夹\n to_labels_file_path = os.path.join(\n args.categoried_base_path,\n args.to_labels_file_name,\n )\n if args.to_labels_file_append:\n to_file = open(to_labels_file_path, \"a\")\n else:\n to_file = open(to_labels_file_path, \"w\")\n for category in categories:\n # 依次遍历每类视频\n video_dir_path = os.path.join(src_videos_dir, category)\n if not os.path.isdir(video_dir_path):\n continue\n for video_name in os.listdir(video_dir_path):\n cur_frames_dir = os.path.join(to_frames_dir, str(cur_idx))\n if not os.path.exists(cur_frames_dir):\n os.mkdir(cur_frames_dir)\n _handle_single_video(\n os.path.join(video_dir_path, video_name), # 输入视频绝对路径\n cur_idx, # 当前视频对应编号\n to_file, # TSM标签文件 writer,用于构建 to_label.txt\n category_to_id[category], # 当前视频代表的行为类别编号,\n args.ffmpeg, # 是否使用 ffmpeg 来提取帧\n cur_frames_dir, # frames 保存路径\n args.img_format,\n args.fps,\n args.tmp_video,\n )\n cur_idx += 1\n to_file.close()\n\n # 3. flip\n if args.flip:\n flip_to_frames_dir = os.path.join(\n args.categoried_base_path,\n args.flip_frames_dir_name,\n )\n if not os.path.exists(flip_to_frames_dir):\n os.makedirs(flip_to_frames_dir)\n flip_samples(\n to_labels_file_path,\n os.path.join(args.categoried_base_path,\n args.flip_to_labels_file_name),\n flip_to_frames_dir\n )\n\n # 4. 复制 to label 文件到to-labels-file-path\n shutil.copy(to_labels_file_path,\n os.path.join(args.global_to_labels_dir,\n os.path.basename(args.categoried_base_path)\n + \"_categoried_to_labels.txt\"))\n if args.flip:\n shutil.copy(to_labels_file_path,\n os.path.join(args.global_to_labels_dir,\n os.path.basename(args.categoried_base_path)\n + \"_categoried_flip_to_labels.txt\"))\n\n\nif __name__ == '__main__':\n main(_parse_args())\n", "sub_path": "src/categoried_videos.py", "file_name": "categoried_videos.py", "file_ext": "py", "file_size_in_byte": 7853, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 10, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "logging.warn", "line_number": 60, "usage_type": "call"}, {"api_name": "os.system", "line_number": 77, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path", "line_number": 87, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 88, "usage_type": "call"}, {"api_name": "os.system", "line_number": 89, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 90, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_FPS", "line_number": 91, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 117, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 117, "usage_type": "call"}, {"api_name": "os.path", "line_number": 117, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 121, "usage_type": "call"}, {"api_name": "os.path", "line_number": 121, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 122, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 127, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path", "line_number": 138, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 142, "usage_type": "call"}, {"api_name": "os.path", "line_number": 142, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 143, "usage_type": "call"}, {"api_name": "os.path", "line_number": 143, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 147, "usage_type": "call"}, {"api_name": "os.path", "line_number": 147, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 148, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 158, "usage_type": "call"}, {"api_name": "os.path", "line_number": 158, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 168, "usage_type": "call"}, {"api_name": "os.path", "line_number": 168, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 169, "usage_type": "call"}, {"api_name": "os.path", "line_number": 169, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 171, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 172, "usage_type": "call"}, {"api_name": "os.path", "line_number": 172, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 173, "usage_type": "call"}, {"api_name": "os.path", "line_number": 173, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 174, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 176, "usage_type": "call"}, {"api_name": "os.path", "line_number": 176, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 191, "usage_type": "call"}, {"api_name": "os.path", "line_number": 191, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 195, "usage_type": "call"}, {"api_name": "os.path", "line_number": 195, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 196, "usage_type": "call"}, {"api_name": "flip_utils.flip_samples", "line_number": 197, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 199, "usage_type": "call"}, {"api_name": "os.path", "line_number": 199, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 205, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 206, "usage_type": "call"}, {"api_name": "os.path", "line_number": 206, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 207, "usage_type": "call"}, {"api_name": "os.path", "line_number": 207, "usage_type": "attribute"}, {"api_name": "shutil.copy", "line_number": 210, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 211, "usage_type": "call"}, {"api_name": "os.path", "line_number": 211, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 212, "usage_type": "call"}, {"api_name": "os.path", "line_number": 212, "usage_type": "attribute"}]}
+{"seq_id": "207448590", "text": "#!/usr/bin/env python\n\n\"\"\"Utilities for common tasks needed to use hotjs framework.\n\"\"\"\n\nimport optparse\nimport subprocess\nimport logging\nimport sys\nimport os.path\nimport zipfile\nimport re\nimport shutil\nimport fileinput\nimport mimetypes\nfrom os.path import join, splitext, split, exists\nfrom shutil import copyfile\nfrom datetime import datetime\nimport base64\nimport json\n\nif sys.version_info[0]==3:\n from urllib.request import urlretrieve\nelse :\n from urllib import urlretrieve\n\n\nbasedir = os.path.dirname(os.path.dirname(os.path.realpath(__file__)))\ncurdir = os.path.abspath('.')\n\nprojects_path = join(basedir,'bin/projects')\n\ndef removeDupes(seq):\n # Not order preserving\n keys = {}\n for e in seq:\n keys[e.rstrip()] = 1\n return keys.keys()\n \ndef makeProjectPaths(add):\n lines = open(projects_path,'r').readlines()\n if len(add):\n lines.append(add)\n newlines = filter(lambda x: exists(join(basedir,x.rstrip())) and len(x.rstrip()),lines)\n newlines = removeDupes(newlines)\n \n f = open(projects_path,'w')\n f.write('\\n'.join(newlines))\n f.close()\n \ndef escapeSpace(s):\n return s.replace(\" \",\"\\\\ \")\n \ndef quoteSpace(s):\n return s.replace(\" \",\"' '\")\n\ndef create(name):\n \n path = os.path.join(curdir, name)\n \n name = os.path.basename(path)\n \n if exists(path):\n logging.error('Directory already exists: %s',path)\n sys.exit(1) \n \n proj = os.path.relpath(path,basedir)\n \n shutil.copytree(os.path.join(basedir,'apps/template'),path)\n \n for root, dirs, files in os.walk(path):\n for fname in files:\n newname = fname.replace('__name__',name)\n if fname.find(\"__name__\")!=-1:\n os.rename(os.path.join(path,fname),os.path.join(path,newname))\n for line in fileinput.FileInput(os.path.join(path,newname),inplace=1):\n line = line.replace('{name}',name)\n print(line.rstrip())\n \n print ('Created %s' % path)\n \n \n if proj!='.':\n makeProjectPaths(os.path.relpath(path,basedir))\n\ndef build(name,options):\n pass\n\ndef update():\n hotjs_path = join(basedir,'hotjs/')\n hotjslist_path = join(hotjs_path,'files')\n hotjsbin_path = join(hotjs_path,'hotjs-bin.js')\n\n print( \"updating hotjsbin: \" + hotjsbin_path )\n f = open(hotjsbin_path,'w')\n \n hotjslist = open(hotjslist_path,'r').readlines()\n for line in hotjslist:\n name = line.strip()\n if( len(name) > 0 ):\n jspath = os.path.join(hotjs_path, name)\n if( exists(jspath) ):\n print( \"adding \" + jspath )\n f.write('\\n// ------- ' + name + ' ------------- \\n\\n')\n jscontents = open(jspath, 'r').readlines()\n f.write(''.join(jscontents))\n else:\n print( jspath + \" not found.\" );\n f.close()\n \n print( \"hotjsbin updated.\" )\n \ndef main():\n \"\"\"The entrypoint for this script.\"\"\"\n \n usage = \"\"\"usage: %prog [command] [options]\nCommands:\n init Check lime dependecies and setup if needed\n create [path/name] Setup new project [name]\n build [name] Compile project to single Javascript file\"\"\"\n parser = optparse.OptionParser(usage)\n \n parser.add_option(\"-o\", \"--output\", dest=\"output\", action=\"store\", type=\"string\",\n help=\"Output file for build result\")\n \n (options, args) = parser.parse_args()\n \n if not (len(args) == 2 or (len(args)==1 and ['init','update'].count(args[0])==1 )) :\n parser.error('incorrect number of arguments')\n \n print( \"welcome to hotjs.\" )\n \n if args[0]=='init' or args[0]=='update':\n update()\n \n elif args[0]=='create':\n create(args[1])\n \n else:\n logging.error('No such command: %s',args[0])\n exit(1)\n \nif __name__ == '__main__':\n main()\n \n ", "sub_path": "bin/hotjs.py", "file_name": "hotjs.py", "file_ext": "py", "file_size_in_byte": 3930, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sys.version_info", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.path.dirname", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 28, "usage_type": "name"}, {"api_name": "os.path.path.realpath", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.path.abspath", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 29, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 59, "usage_type": "name"}, {"api_name": "os.path.path.basename", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 61, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 61, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 63, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 64, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path.path.relpath", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 67, "usage_type": "name"}, {"api_name": "shutil.copytree", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 69, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 69, "usage_type": "name"}, {"api_name": "os.path.walk", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path", "line_number": 71, "usage_type": "name"}, {"api_name": "os.path.rename", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path", "line_number": 75, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 75, "usage_type": "attribute"}, {"api_name": "fileinput.FileInput", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 76, "usage_type": "name"}, {"api_name": "os.path.path.relpath", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 84, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 91, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 101, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 101, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 102, "usage_type": "call"}, {"api_name": "optparse.OptionParser", "line_number": 121, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 140, "usage_type": "call"}]}
+{"seq_id": "222758203", "text": "import json\r\nimport urllib.request\r\nimport mmap\r\nimport struct\r\n\r\n# Takes data from get_all_current_maps() and removes duplicates.\r\n# Going to be for listing assignable maps.\r\ndef get_unique_map_names(requested_maps):\r\n\tall_maps = {}\r\n\t\r\n\tfor map_data in requested_maps:\r\n\t\tmap_name = map_data.get(\"name\")\r\n\t\tif map_name in all_maps.keys():\r\n\t\t\tpass\r\n\t\telse:\r\n\t\t\tall_maps[map_name] = 0\r\n\t\t\r\n\treturn all_maps\r\n\r\n# Reads GW2's API for a dictionary of all current map IDs and names.\r\ndef get_all_current_maps():\r\n\trequest = urllib.request.Request('https://api.guildwars2.com/v1/map_names.json')\r\n\twith urllib.request.urlopen(request) as response:\r\n\t\trequested_maps = json.loads(response.readall().decode('utf-8'))\r\n\r\n\treturn requested_maps\r\n\r\ndef get_coordinates():\r\n\tdata = mmap.mmap(0, 20, \"MumbleLink\", mmap.ACCESS_READ)\r\n\tprint(data.readline())\r\n\tcoord = struct.unpack('IL3f', data)[2:5]\r\n\tdata.close()\r\n\t\r\n\tprint(coord)\r\n\r\ndef main():\r\n\t#all_maps_dict = get_all_current_maps()\r\n\t#all_unique_maps = get_unique_map_names(all_maps_dict)\r\n\tget_coordinates()\r\n\r\nmain()", "sub_path": "GW2 Soundtrack.py", "file_name": "GW2 Soundtrack.py", "file_ext": "py", "file_size_in_byte": 1065, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "urllib.request.request.Request", "line_number": 22, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 22, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 22, "usage_type": "name"}, {"api_name": "urllib.request.request.urlopen", "line_number": 23, "usage_type": "call"}, {"api_name": "urllib.request.request", "line_number": 23, "usage_type": "attribute"}, {"api_name": "urllib.request", "line_number": 23, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 24, "usage_type": "call"}, {"api_name": "mmap.mmap", "line_number": 29, "usage_type": "call"}, {"api_name": "mmap.ACCESS_READ", "line_number": 29, "usage_type": "attribute"}, {"api_name": "struct.unpack", "line_number": 31, "usage_type": "call"}]}
+{"seq_id": "652131644", "text": "import time\nimport json\nimport random\nfrom collections import OrderedDict\nimport string\nfrom kafka import KafkaProducer\n\nTOPIC='custom'\nplantNames=['EMEA-LU-J650-IP21-LU5-PIMS','EMEA-LU-J650-IP21-LU5-PIMS2']\nhiphen=\"-\"\ndot=\".\"\nqualityNames=['good','bad']\n\ndef fuzzyGenerator(numrange=None):\n if numrange==None:\n numrange=100\n for _ in range(numrange):\n jsonList={}\n correlationIdDict=OrderedDict()\n bodyDict=OrderedDict()\n correlationId=(''.join(random.choice(string.ascii_lowercase+string.digits+hiphen) for _ in range(48)))\n correlationIdDict['correlationId']=correlationId\n \n bodyDict['plantId']=random.choice(plantNames) \n bodyDict['timestampCalculation']=int(time.time())\n bodyDict['tagValue']=random.choice(string.digits) \n bodyDict['timestamp']=int(time.time())\n bodyDict['tagName']=(''.join(random.choice(string.ascii_uppercase+string.digits+dot) for _ in range(10)))\n bodyDict['quality']=random.choice(qualityNames)\n \n jsonList['userProperties']=correlationIdDict\n jsonList['body']=bodyDict\n yield jsonList\n \ndef publishMessage(producer_instance, topic_name, key, value):\n producer_instance.send(topic_name, value)\n producer_instance.flush()\n\nif __name__ == '__main__':\n producer=KafkaProducer(bootstrap_servers=['sandbox.hortonworks.com:6667'],value_serializer=lambda v: json.dumps(v).encode('utf-8'))\n for rec in fuzzyGenerator():\n publishMessage(producer, TOPIC, 'json', rec)\n producer.close(timeout=None)\n", "sub_path": "KafkaProducer.py", "file_name": "KafkaProducer.py", "file_ext": "py", "file_size_in_byte": 1576, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "collections.OrderedDict", "line_number": 19, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 20, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 21, "usage_type": "call"}, {"api_name": "string.ascii_lowercase", "line_number": 21, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 21, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 24, "usage_type": "call"}, {"api_name": "time.time", "line_number": 25, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 26, "usage_type": "call"}, {"api_name": "string.digits", "line_number": 26, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 27, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 28, "usage_type": "call"}, {"api_name": "string.ascii_uppercase", "line_number": 28, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 28, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 29, "usage_type": "call"}, {"api_name": "kafka.KafkaProducer", "line_number": 40, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 40, "usage_type": "call"}]}
+{"seq_id": "611391840", "text": "# encoding=utf-8\nimport multiprocessing as mp\nimport cv2\nimport time\nfrom imutils.video import FPS\n\nimport threading\nimport queue\n'''2018-05-21 Yonv1943'''\n'''2018-07-02 setattr(), run_multi_camera()'''\n\n\ndef queue_img_put(q, name, pwd, ip, channel=1):\n cap = cv2.VideoCapture(\"rtsp://%s:%s@%s//Streaming/Channels/%d?tcp\" % (name, pwd, ip, channel))\n time.sleep(1)\n\n while True:\n is_opened, frame = cap.read()\n q.put(frame) if is_opened else None\n q.get() if q.qsize() > 1 else None\n\n\ndef queue_img_get(q, window_name):\n cv2.namedWindow(window_name, flags=cv2.WINDOW_FREERATIO)\n while True:\n frame = q.get()\n cv2.imshow(window_name, frame)\n cv2.waitKey(1)\n\n\ndef run(): # single camera\n user_name, user_pwd, camera_ip = \"admin\", \"nowvciji121226\", \"192.168.1.64\"\n\n Q = queue.Queue(maxsize=2) # 构造一个不限制大小的的队列\n _WORKER_THREAD_NUM = 2 # 设置线程的个数\n\n processes = [threading.Thread(target=queue_img_put, args=(Q, user_name, user_pwd, camera_ip)),\n threading.Thread(target=queue_img_get, args=(Q, camera_ip))]\n\n [setattr(process, \"daemon\", True) for process in processes] # process.daemon = True\n [process.start() for process in processes]\n [process.join() for process in processes]\n\n\n\n\ndef run_multi_camera():\n user_name, user_pwd = \"admin\", \"huarui2019.\"\n user_name3, user_pwd3, camera_ip3 = \"admin\", \"nowvciji121226\", \"192.168.1.64\"\n camera_ip_l = [\n \"192.168.1.44\",\n \"192.168.1.43\",\n ]\n\n mp.set_start_method(method='spawn') # init\n\n queues = [mp.Queue(maxsize=2) for _ in camera_ip_l]\n\n processes = []\n for queue, camera_ip in zip(queues, camera_ip_l):\n processes.append(mp.Process(target=queue_img_put, args=(queue, user_name, user_pwd, camera_ip)))\n processes.append(mp.Process(target=queue_img_get, args=(queue, camera_ip)))\n\n processes.append(mp.Process(target=queue_img_put, args=(queue, user_name3, user_pwd3, camera_ip3)))\n processes.append(mp.Process(target=queue_img_get, args=(queue, camera_ip3)))\t\n [setattr(process, \"daemon\", True) for process in processes] # process.daemon = True\n [process.start() for process in processes]\n [process.join() for process in processes]\n\n\nif __name__ == '__main__':\n run()\n #run_multi_camera()\n", "sub_path": "PythonScript/PythonScript/camera_rtsp_desktop_single.py", "file_name": "camera_rtsp_desktop_single.py", "file_ext": "py", "file_size_in_byte": 2344, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "cv2.VideoCapture", "line_number": 14, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.namedWindow", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.WINDOW_FREERATIO", "line_number": 24, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 27, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 28, "usage_type": "call"}, {"api_name": "queue.Queue", "line_number": 34, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 37, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 38, "usage_type": "call"}, {"api_name": "multiprocessing.set_start_method", "line_number": 55, "usage_type": "call"}, {"api_name": "multiprocessing.Queue", "line_number": 57, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 61, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 62, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 64, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 65, "usage_type": "call"}]}
+{"seq_id": "601402094", "text": "import os\nimport logging\n\nlogger = logging.getLogger(__name__)\n\nimport numpy as np\nimport scipy.interpolate as intp\nimport scipy.optimize as opt\nimport astropy.io.fits as fits\nimport matplotlib.pyplot as plt\n\nfrom .imageproc import table_to_array\nfrom ..utils.onedarray import pairwise, smooth\nfrom ..utils.regression import get_clip_mean\n\ndef sum_extract(infilename, mskfilename, outfilename, channels, apertureset_lst,\n upper_limit=5, lower_limit=5, figure=None):\n \"\"\"Extract 1-D spectra from an individual image.\n \n Args:\n infilename (str): Name of the input image.\n outfilename (str): Name of the output image.\n channels (list): List of channels as strings.\n apertureset_lst (dict): Dict of :class:`~gamse.echelle.trace.ApertureSet`\n instances for different channels.\n upper_limit (float): Upper limit of the extracted aperture.\n lower_limit (float): Lower limit of the extracted aperture.\n figure (:class:`matplotlib.figure.Figure`): Figure to display the\n extracted 1d spectra.\n \"\"\"\n data, head = fits.getdata(infilename, header=True)\n h, w = data.shape\n\n # read data mask\n mask_table = fits.getdata(mskfilename)\n if mask_table.size==0:\n mask = np.zeros_like(data, dtype=np.int16)\n else:\n mask = table_to_array(mask_table, data.shape)\n data_mask = (np.int16(mask) & 4) > 0\n\n xx, yy = np.meshgrid(np.arange(w),np.arange(h))\n\n # seperate each type of mask\n #cov_mask = (mdata & 1)>0\n #bad_mask = (mdata & 2)>0\n #sat_mask = (mdata & 4)>0\n \n # define a numpy structured array\n types = [\n ('aperture', np.int32),\n ('channel', '|1S'),\n ('points', np.int32),\n ('flux', '(%d,)float32'%w),\n ('mask', '(%d,)int16'%w),\n ]\n tmp = list(zip(*types))\n eche_spec = np.dtype({'names':tmp[0], 'formats':tmp[1]})\n\n spec = []\n\n newx = np.arange(w)\n\n # find integration limits\n info_lst = []\n for channel in channels:\n for aper, aperloc in apertureset_lst[channel].items():\n center = aperloc.get_center()\n info_lst.append((center, channel, aper))\n # sort the info_lst\n newinfo_lst = sorted(info_lst, key=lambda item: item[0])\n\n # find the middle bounds for every adjacent apertures\n lower_bounds = {}\n upper_bounds = {}\n prev_channel = None\n prev_aper = None\n prev_position = None\n for item in newinfo_lst:\n channel = item[1]\n aper = item[2]\n position = apertureset_lst[channel][aper].position(newx)\n if prev_position is not None:\n mid = (position + prev_position)/2.\n lower_bounds[(channel, aper)] = mid\n upper_bounds[(prev_channel, prev_aper)] = mid\n prev_position = position\n prev_channel = channel\n prev_aper = aper\n\n for channel in channels:\n for aper, aper_loc in apertureset_lst[channel].items():\n position = aper_loc.position(newx)\n # determine the lower and upper limits\n lower_line = position - lower_limit\n upper_line = position + upper_limit\n key = (channel, aper)\n if key in lower_bounds:\n lower_line = np.maximum(lower_line, lower_bounds[key])\n if key in upper_bounds:\n upper_line = np.minimum(upper_line, upper_bounds[key])\n lower_line = np.maximum(lower_line, np.zeros(w)-0.5)\n upper_line = np.minimum(upper_line, np.zeros(w)+h-1-0.5)\n lower_ints = np.int32(np.round(lower_line))\n upper_ints = np.int32(np.round(upper_line))\n m1 = yy > lower_ints\n m2 = yy < upper_ints\n mask = m1*m2\n mask = np.float32(mask)\n # determine the weight in the boundary\n mask[lower_ints, newx] = 1-(lower_line+0.5)%1\n mask[upper_ints, newx] = (upper_line+0.5)%1\n\n # determine the upper and lower row of summing\n r1 = int(lower_line.min())\n r2 = int(upper_line.max())+1\n mask = mask[r1:r2]\n\n # summing the data and mask\n fluxdata = (data[r1:r2,]*mask).sum(axis=0)\n sat_flux = (data_mask[r1:r2,]*mask).sum(axis=0)>0\n\n fluxmask = np.int16(sat_flux*4)\n item = np.array((aper, channel, fluxdata.size, fluxdata, fluxmask),\n dtype=eche_spec)\n spec.append(item)\n\n # update header. Put coefficients of aperture locations into header.\n leading_string = 'HIERARCH EDRS TRACE CHANNEL %s APERTURE %d'%(\n channel, aper)\n for ic, c in enumerate(aper_loc.position.coef):\n head[leading_string + ' COEFF %d'%ic] = c\n\n\n spec = np.array(spec, dtype=eche_spec)\n\n pri_hdu = fits.PrimaryHDU(header=head)\n tbl_hdu = fits.BinTableHDU(spec)\n hdu_lst = fits.HDUList([pri_hdu, tbl_hdu])\n hdu_lst.writeto(outfilename, overwrite=True)\n logger.info('Write 1D spectra file \"%s\"'%outfilename)\n\ndef extract_aperset(data, mask, apertureset, lower_limit=5, upper_limit=5, variance=False):\n \"\"\"Extract 1-D spectra from the input image data following the input\n :class:`~gamse.echelle.trace.ApertureSet`.\n\n Args:\n data (:class:`numpy.ndarray`): Input data image.\n mask (:class:`numpy.ndarray`): Input mask.\n apertureset (:class:`~gamse.echelle.trace.ApertureSet`): Input\n :class:`~gamse.echelle.trace.ApertureSet` instance.\n lower_limit (float): Lower limit of the extracted aperture.\n upper_limit (float): Upper limit of the extracted aperture.\n variance (bool) : If a variance array is processed the weights \n need to be squared\n\n Returns:\n dict: A dict of 1-d spectra with the aperture numbers as keys, and a\n dict of (\"flux_sum\", \"flux_mean\", \"mask_sat\") as values.\n \n \"\"\"\n h, w = data.shape\n\n # find saturation mask and bad pixel mask\n sat_mask = (mask&4 > 0)\n bad_mask = (mask&2 > 0)\n gap_mask = (mask&1 > 0)\n\n yy, xx = np.mgrid[:h:,:w:]\n spectra1d = {}\n for aper, aper_loc in sorted(apertureset.items()):\n domain = aper_loc.position.domain\n d1, d2 = int(domain[0]), int(domain[1])+1\n newx = np.arange(d1, d2)\n position = aper_loc.position(newx)\n lower_line = position - lower_limit\n upper_line = position + upper_limit\n lower_line = np.maximum(lower_line, -0.5)\n lower_line = np.minimum(lower_line, h-1-0.5)\n upper_line = np.maximum(upper_line, -0.5)\n upper_line = np.minimum(upper_line, h-1-0.5)\n lower_ints = np.int32(np.round(lower_line))\n upper_ints = np.int32(np.round(upper_line))\n m1 = yy[:,d1:d2] > lower_ints\n m2 = yy[:,d1:d2] < upper_ints\n newmask = np.zeros_like(data, dtype=np.bool)\n newmask[:,d1:d2] = m1*m2\n newmask = np.float32(newmask)\n # determine the weight in the boundary\n if variance:\n newmask[lower_ints, newx] = (1-(lower_line+0.5)%1)**2\n newmask[upper_ints, newx] = ((upper_line+0.5)%1)**2\n else:\n newmask[lower_ints, newx] = 1-(lower_line+0.5)%1\n newmask[upper_ints, newx] = (upper_line+0.5)%1\n # filter the bad, saturated, and gap pixels\n newmask = newmask*(~sat_mask)\n newmask = newmask*(~bad_mask)\n newmask = newmask*(~gap_mask)\n\n ## determine the upper and lower row of summing\n r1 = int(lower_line.min())\n r2 = int(upper_line.max())+1\n\n # summing the data and mask\n weight_sum = newmask[r1:r2].sum(axis=0)\n # summing the flux\n fluxsum = (data[r1:r2]*newmask[r1:r2]).sum(axis=0)\n # calculate mean flux\n # filter the zero values\n _m = weight_sum>0\n fluxmean = np.zeros_like(fluxsum)\n fluxmean[_m] = fluxsum[_m]/weight_sum[_m]\n\n spectra1d[aper] = {\n 'flux_sum': fluxsum,\n 'flux_mean': fluxmean,\n 'mask': ~_m,\n 'nsum': weight_sum,\n }\n\n\n # summing the masks\n fluxsat = (sat_mask[r1:r2]*newmask[r1:r2]).sum(axis=0)>0\n spectra1d[aper]['mask_sat'] = fluxsat\n\n return spectra1d\n\ndef get_mean_profile(nodex_lst, nodey_lst, profx_lst):\n \"\"\"Calculate the mean profiles for a series of (*x*, *y*) data.\n\n Args:\n nodex_lst (:class:`numpy.ndarray`): Input *x* data.\n nodey_lst (:class:`numpy.ndarray`): Input *y* data with the same length\n as **nodex_lst**.\n profx_lst (:class:`numpy.ndarray`): X-coordinates of the mean profile.\n\n Returns:\n A tuple containing:\n **profile** (:class:`numpy.ndarray`): Mean profile.\n **profile_std** (:class:`numpy.ndarray`): Standard deviations of\n mean profile.\n\n\n \"\"\"\n # find middle points\n mid_profx_lst = (profx_lst + np.roll(profx_lst, -1))/2\n mid_profx_lst = np.insert(mid_profx_lst, 0,\n profx_lst[0]-(profx_lst[1]-profx_lst[0])/2)\n mid_profx_lst[-1] = profx_lst[-1] + (profx_lst[-1] - profx_lst[-2])/2\n\n # calculate mean profile\n mean_x_lst, mean_y_lst, std_y_lst = [], [], []\n for y1, y2 in pairwise(mid_profx_lst):\n mask = (nodex_lst > y1)*(nodex_lst < y2)\n if mask.sum() > 0:\n meany, std, _ = get_clip_mean(nodey_lst[mask], maxiter=20)\n #xcenter = (nodex_lst[mask]*nodey_lst[mask]).sum()/nodey_lst[mask].sum()\n #mean_x_lst.append(xcenter)\n mean_x_lst.append((y1+y2)/2)\n mean_y_lst.append(meany)\n std_y_lst.append(std)\n\n # convert to numpy arrays\n mean_x_lst = np.array(mean_x_lst)\n mean_y_lst = np.array(mean_y_lst)\n std_y_lst = np.array(std_y_lst)\n\n # fill the missing values with cubic interpolation\n if mean_x_lst.size < profx_lst.size:\n f1 = intp.InterpolatedUnivariateSpline(mean_x_lst, mean_y_lst, k=3)\n mean_y_lst = f1(profx_lst)\n f2 = intp.InterpolatedUnivariateSpline(mean_x_lst, std_y_lst, k=3)\n std_y_lst = f2(profx_lst)\n\n return mean_y_lst, std_y_lst\n\ndef optimal_extract(data, mask, apertureset):\n \"\"\"Optimal extraction.\n\n Args:\n data (:class:`ndarray`):\n mask (:class:`ndarray`):\n apertureset ():\n\n Returns:\n \"\"\"\n\n daper = 10\n aper1_lst = np.arange(min(apertureset), max(apertureset), daper)\n aper2_lst = aper1_lst + daper\n if len(apertureset)%daper < daper/2:\n aper1_lst = np.delete(aper1_lst, -1)\n aper2_lst = np.delete(aper2_lst, -1)\n else:\n pass\n aper2_lst[-1] = max(apertureset)+1\n\n profx_lst = np.arange(-10, 10+1e-3, 0.5)\n h, w = data.shape\n\n for loop in range(2):\n apercen_lst = []\n profilesamp_lst = []\n for iregion, (aper1, aper2) in enumerate(zip(aper1_lst, aper2_lst)):\n print(aper1, aper2)\n apernode_x_lst = []\n apernode_y_lst = []\n for iaper, aper in enumerate(np.arange(aper1, aper2)):\n aperloc = apertureset[aper]\n ycen_lst = aperloc.position(np.arange(w))\n\n node_x_lst = []\n node_y_lst = []\n if loop > 0:\n profile = allprofile_lst[aper]\n interf = intp.InterpolatedUnivariateSpline(\n profx_lst, profile, k=3, ext=1)\n\n for x in np.arange(w//2-200, w//2+200):\n ycen = ycen_lst[x]\n yceni = np.int(np.round(ycen))\n yrows = np.arange(yceni-10, yceni+10+1)\n flux = data[yrows, x]\n negative_mask = flux<0\n if negative_mask.sum()>0.5*flux.size:\n continue\n\n if loop == 0:\n A = flux.sum()\n else:\n para = [flux.sum(),ycen]\n mask = np.ones_like(flux, dtype=np.bool)\n for ite in range(10):\n result = opt.least_squares(errfunc, para,\n bounds=((-np.inf,ycen-2),(np.inf,ycen+2)),\n args=(flux[mask], interf, yrows[mask]))\n newpara = result['x']\n pro = fitfunc(newpara, interf, yrows)\n res = flux - pro\n std = res[mask].std()\n new_mask = (res < 3*std)*(res > -3*std)\n if new_mask.sum() == mask.sum():\n break\n mask = new_mask\n para = newpara\n A, ycen = newpara\n\n if A<0:\n continue\n normflux = flux/A\n for v in yrows-ycen:\n node_x_lst.append(v)\n apernode_x_lst.append(v)\n for v in normflux:\n node_y_lst.append(v)\n apernode_y_lst.append(v)\n ### loop for x pixel ends here\n\n # now calculate the mean profile\n apernode_x_lst = np.array(apernode_x_lst)\n apernode_y_lst = np.array(apernode_y_lst)\n profile, profile_std = get_mean_profile(\n apernode_x_lst, apernode_y_lst, profx_lst)\n\n # smooth the profile\n profile = smooth(profile, points=5, deg=3)\n profile_std = smooth(profile_std, points=5, deg=3)\n\n # calculate the typical S/N of this profile\n snr_lst = profile/profile_std\n ic = profx_lst.size//2\n snr = snr_lst[ic-1:ic+2].mean()\n\n # calculate the center of mass\n profile_cen = (profx_lst*profile).sum()/profile.sum()\n\n # align the profile to the center of mass with cubic interpolataion\n func = intp.InterpolatedUnivariateSpline(\n profx_lst-profile_cen, profile, k=3, ext=3)\n newprofile = func(profx_lst)\n\n # append the results\n profilesamp_lst.append(newprofile)\n apercen_lst.append((aper1+aper2)/2)\n\n # plot figure for this aper group\n fig1 = plt.figure(dpi=150, figsize=(12,8))\n ax1 = fig1.gca()\n ax1.scatter(apernode_x_lst, apernode_y_lst, c='gray', s=1, alpha=0.1)\n ax1.plot(profx_lst, profile, '-', lw=1, c='C1')\n ax1.plot(profx_lst, profile+profile_std, '--', lw=0.5, c='C1')\n ax1.plot(profx_lst, profile-profile_std, '--', lw=0.5, c='C1')\n ax1.set_xlim(profx_lst[0]-1, profx_lst[-1]+1)\n ax1.set_ylim(-0.02, 0.13)\n ax1.axvline(x=0, color='k', ls='--', lw=1)\n ax1.axhline(y=0, color='k', ls='--', lw=1)\n ax1.grid(True, ls=':', color='k')\n fig1.savefig('img2/apergroup_%02d_%02d_loop%d.png'%(aper1, aper2, loop))\n plt.close(fig1)\n\n ### loop for aper region ends here\n # build interp functions for all apertures\n profilesamp_lst = np.array(profilesamp_lst)\n\n # get profiles for all apertures\n allprofile_lst = {}\n for aper in sorted(apertureset):\n profile = []\n for col in np.arange(profx_lst.size):\n func = intp.InterpolatedUnivariateSpline(\n apercen_lst, profilesamp_lst[:,col], k=3, ext=0)\n profile.append(func(aper))\n allprofile_lst[aper] = np.array(profile)\n\n # plot interpolated profiles\n fig = plt.figure(dpi=150)\n ax = fig.gca()\n for aper in sorted(apertureset):\n ax.plot(profx_lst, allprofile_lst[aper], alpha=0.6, lw=0.5)\n ax.set_xlim(-11,11)\n ax.set_ylim(-0.01, 0.12)\n ax.grid(True, color='k', ls=':', lw=0.5)\n fig.savefig('img2/intp_profiles_loop%d.png'%loop)\n plt.close(fig)\n # profile loop ends here\n\n ######################################################################\n flux_opt_lst = {}\n flux_sum_lst = {}\n #for aper, aperloc in sorted(aperset.items()):\n for aper in [10, 63]:\n aperloc = aperset[aper]\n print(aper)\n ycen_lst = aperloc.position(np.arange(w))\n profile = allprofile_lst[aper]\n interf = intp.InterpolatedUnivariateSpline(\n profx_lst, profile, k=3, ext=1)\n flux_opt_lst[aper] = []\n flux_sum_lst[aper] = []\n newycen_lst = []\n for x in np.arange(w):\n ycen = ycen_lst[x]\n yceni = np.int(np.round(ycen))\n yrows = np.arange(yceni-10, yceni+10+1)\n flux = data[yrows, x]\n para = [flux.sum(),ycen]\n mask = np.ones_like(flux, dtype=np.bool)\n for ite in range(10):\n result = opt.least_squares(errfunc, para,\n bounds=((-np.inf,ycen-2),(np.inf,ycen+2)),\n args=(flux[mask], interf, yrows[mask]))\n newpara = result['x']\n pro = fitfunc(newpara, interf, yrows)\n res = flux - pro\n std = res[mask].std()\n new_mask = (res < 3*std)*(res > -3*std)\n if new_mask.sum() == mask.sum():\n break\n mask = new_mask\n para = newpara\n print(x, newpara, mask.size-mask.sum(), ite)\n newycen_lst.append(newpara[1])\n s_lst = 1/(np.maximum((flux+240),0)+1.0**2)\n normpro = pro/pro.sum()\n fopt = ((s_lst*normpro*flux)[mask].sum())/((s_lst*normpro**2)[mask].sum())\n flux_opt_lst[aper].append(fopt)\n flux_sum_lst[aper].append(flux.sum())\n\n ################################################\n if aper==63:\n if x%30==0:\n fig1 = plt.figure(figsize=(18,10), dpi=150)\n irow = int((x%30)/6)\n icol = (x%30)%6\n _x = 0.04 + icol*0.16\n _y = 0.05 + (4-irow)*0.19\n ax = fig1.add_axes([_x, _y, 0.14, 0.17])\n ax.plot(yrows-para[1], pro, 'o-', color='w',\n markeredgecolor='C0',ms=4, lw=0.8)\n ax.plot(yrows[mask]-para[1], pro[mask], 'o', color='C0',\n ms=4)\n ax.plot(yrows-para[1], pro+1*std, '--', color='C0', lw=0.5)\n ax.plot(yrows-para[1], pro-1*std, '--', color='C0', lw=0.5)\n ax.plot(yrows[mask]-para[1], flux[mask], '-', color='C1', lw=0.8)\n x1, x2 = ax.get_xlim()\n y1, y2 = ax.get_ylim()\n ax.plot(yrows-para[1], flux, '--', color='C1', lw=0.8)\n ax.text(0.95*x1+0.05*x2, 0.1*y1+0.9*y2, 'X=%d'%x, fontsize=9)\n ax.text(0.35*x1+0.65*x2, 0.1*y1+0.9*y2, '%7g'%(flux.sum()), fontsize=9, color='C0')\n ax.text(0.35*x1+0.65*x2, 0.2*y1+0.8*y2, '%7g'%fopt, fontsize=9, color='C1')\n ax.set_xlim(x1, x2)\n ax.set_ylim(y1, y2)\n ax.axhline(y=0, c='k', ls='--', lw=0.5)\n ax.axvline(x=0, c='k', ls='--', lw=0.5)\n for tick in ax.xaxis.get_major_ticks():\n tick.label1.set_fontsize(7)\n for tick in ax.yaxis.get_major_ticks():\n tick.label1.set_fontsize(7)\n if x%30 == 29 or x == w-1:\n fig1.savefig('img4/fitting-%02d-%04d.png'%(aper,x))\n plt.close(fig1)\n ################################################\n\n\n\n flux_opt_lst[aper] = np.array(flux_opt_lst[aper])\n flux_sum_lst[aper] = np.array(flux_sum_lst[aper])\n\n '''\n fig = plt.figure(dpi=150, figsize=(15,10))\n ax = fig.gca()\n ax.plot(flux_opt_lst[aper], ls='-', lw=0.5, color='C1')\n ax.set_xlim(0, w-1)\n fig.savefig('img4/flux_%02d.png'%aper)\n plt.close(fig)\n '''\n newycen_lst = np.array(newycen_lst)\n fig2 = plt.figure(dpi=150, figsize=(12,8))\n ax1 = fig2.add_subplot(211)\n ax2 = fig2.add_subplot(212)\n ax1.plot(ycen_lst, color='C0', ls='-')\n ax1.plot(newycen_lst, color='C1', ls='-')\n ax2.plot(newycen_lst-ycen_lst, color='C1', ls='-')\n fig2.savefig('img4/comp_center_aper%02d.png'%aper)\n plt.close(fig2)\n\n types = [\n ('aperture', np.int16),\n ('order', np.int16),\n ('points', np.int16),\n ('wavelength', (np.float64, w)),\n ('flux_sum', (np.float32, w)),\n ('flux_opt', (np.float32, w)),\n ]\n names, formats = list(zip(*types))\n spectype = np.dtype({'names': names, 'formats': formats})\n spec = []\n for aper in sorted(flux_opt_lst):\n flux_sum = flux_sum_lst[aper]\n flux_opt = flux_opt_lst[aper]\n n = flux_sum.size\n spec.append((aper, 0, n,\n np.zeros(n, dtype=np.float64),\n flux_sum, flux_opt))\n spec = np.array(spec, dtype=spectype)\n\n", "sub_path": "gamse/echelle/extract.py", "file_name": "extract.py", "file_ext": "py", "file_size_in_byte": 21290, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.getLogger", "line_number": 4, "usage_type": "call"}, {"api_name": "astropy.io.fits.getdata", "line_number": 31, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 31, "usage_type": "name"}, {"api_name": "astropy.io.fits.getdata", "line_number": 35, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 35, "usage_type": "name"}, {"api_name": "numpy.zeros_like", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 37, "usage_type": "attribute"}, {"api_name": "imageproc.table_to_array", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 51, "usage_type": "attribute"}, {"api_name": "numpy.int32", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.dtype", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.int16", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 124, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 135, "usage_type": "call"}, {"api_name": "astropy.io.fits.PrimaryHDU", "line_number": 137, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 137, "usage_type": "name"}, {"api_name": "astropy.io.fits.BinTableHDU", "line_number": 138, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 138, "usage_type": "name"}, {"api_name": "astropy.io.fits.HDUList", "line_number": 139, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 139, "usage_type": "name"}, {"api_name": "numpy.mgrid", "line_number": 169, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 181, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 182, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 182, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.bool", "line_number": 186, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 188, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 212, "usage_type": "call"}, {"api_name": "numpy.roll", "line_number": 247, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 248, "usage_type": "call"}, {"api_name": "utils.onedarray.pairwise", "line_number": 254, "usage_type": "call"}, {"api_name": "utils.regression.get_clip_mean", "line_number": 257, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 265, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 266, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 267, "usage_type": "call"}, {"api_name": "scipy.interpolate.InterpolatedUnivariateSpline", "line_number": 271, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 271, "usage_type": "name"}, {"api_name": "scipy.interpolate.InterpolatedUnivariateSpline", "line_number": 273, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 273, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 290, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 293, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 294, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 299, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 309, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 311, "usage_type": "call"}, {"api_name": "scipy.interpolate.InterpolatedUnivariateSpline", "line_number": 317, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 317, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 320, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 322, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 322, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 323, "usage_type": "call"}, {"api_name": "numpy.ones_like", "line_number": 333, "usage_type": "call"}, {"api_name": "numpy.bool", "line_number": 333, "usage_type": "attribute"}, {"api_name": "scipy.optimize.least_squares", "line_number": 335, "usage_type": "call"}, {"api_name": "scipy.optimize", "line_number": 335, "usage_type": "name"}, {"api_name": "numpy.inf", "line_number": 336, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 361, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 362, "usage_type": "call"}, {"api_name": "utils.onedarray.smooth", "line_number": 367, "usage_type": "call"}, {"api_name": "utils.onedarray.smooth", "line_number": 368, "usage_type": "call"}, {"api_name": "scipy.interpolate.InterpolatedUnivariateSpline", "line_number": 379, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 379, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 388, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 388, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 400, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 400, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 404, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 410, "usage_type": "call"}, {"api_name": "scipy.interpolate.InterpolatedUnivariateSpline", "line_number": 411, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 411, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 414, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 417, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 417, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 425, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 425, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 435, "usage_type": "call"}, {"api_name": "scipy.interpolate.InterpolatedUnivariateSpline", "line_number": 437, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 437, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 442, "usage_type": "call"}, {"api_name": "numpy.int", "line_number": 444, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 444, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 445, "usage_type": "call"}, {"api_name": "numpy.ones_like", "line_number": 448, "usage_type": "call"}, {"api_name": "numpy.bool", "line_number": 448, "usage_type": "attribute"}, {"api_name": "scipy.optimize.least_squares", "line_number": 450, "usage_type": "call"}, {"api_name": "scipy.optimize", "line_number": 450, "usage_type": "name"}, {"api_name": "numpy.inf", "line_number": 451, "usage_type": "attribute"}, {"api_name": "numpy.maximum", "line_number": 464, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 473, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 473, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 502, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 502, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 507, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 508, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 518, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 519, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 519, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 526, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 526, "usage_type": "name"}, {"api_name": "numpy.int16", "line_number": 529, "usage_type": "attribute"}, {"api_name": "numpy.int16", "line_number": 530, "usage_type": "attribute"}, {"api_name": "numpy.int16", "line_number": 531, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 532, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 533, "usage_type": "attribute"}, {"api_name": "numpy.float32", "line_number": 534, "usage_type": "attribute"}, {"api_name": "numpy.dtype", "line_number": 537, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 544, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 544, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 546, "usage_type": "call"}]}
+{"seq_id": "365953381", "text": "import numpy as np\nfrom datetime import datetime\nfrom keras.models import Model, Input\nfrom keras.layers import LSTM, Embedding, Dense, TimeDistributed, Dropout, Bidirectional\n\n\nclass NeuralNetwork(object):\n\n def __init__(self, num_words, num_entities, X_train, Y_train, X_validation, Y_validation, X_test, Y_test):\n self.num_words = num_words\n self.num_entities = num_entities\n self.X_train = X_train\n self.Y_train = Y_train\n self.X_validation = X_validation\n self.Y_validation = Y_validation\n self.X_test = X_test\n self.Y_test = Y_test\n\n def train(self):\n input = Input(shape=(120,))\n model = Embedding(input_dim=self.num_words, output_dim=50, input_length=120)(input)\n model = Dropout(0.1)(model)\n model = Bidirectional(LSTM(units=100, return_sequences=True, recurrent_dropout=0.1))(model)\n out = TimeDistributed(Dense(self.num_entities, activation=\"softmax\"))(model)\n\n model = Model(input, out)\n\n model.compile(optimizer=\"rmsprop\", loss=\"categorical_crossentropy\", metrics=[\"accuracy\"])\n\n history = model.fit(x=self.X_train, y=np.array(self.Y_train), batch_size=64, epochs=10,\n validation_data=(self.X_validation, self.Y_validation))\n\n model.save(\"../models/ner_\" + str(datetime.utcnow().microsecond))\n\n test_eval = model.evaluate(self.X_test, self.Y_test, verbose=0)\n print('Test loss:', test_eval[0])\n print('Test accuracy:', test_eval[1])\n\n return model, history\n", "sub_path": "root/model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 1550, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "keras.models.Input", "line_number": 20, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 21, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 22, "usage_type": "call"}, {"api_name": "keras.layers.Bidirectional", "line_number": 23, "usage_type": "call"}, {"api_name": "keras.layers.LSTM", "line_number": 23, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 24, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 24, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 33, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 33, "usage_type": "name"}]}
+{"seq_id": "46351689", "text": "from django.conf.urls import url\nfrom .views import (DoctorLogin,\n EnterpriseHeadquarters,\n ClinicsList,\n EmergencyAttentionList,\n EmergencyDoctor,\n SpecialistDoctor,\n MedicalHistoryList,\n MedicalHistoryDetail,\n PatientView,\n PatientUpdateToken,\n PatientByTokenList,\n DoctorUpdateLocationView,\n MedicalHistoryUpdate,\n PatientVerifyView,\n PatientAppointments,\n PatientHistoryView,\n ArtifactMeasurementView,\n MeasurementWeight,\n MeasurementBodyTemperature,\n BloodPressure,\n PatientUpdate,\n CallDoctorView,\n MedicalHistorySpecialistList,\n CallActivate,\n ArtifactMeasurementTool,\n PatientRegisterView,\n #MedicalHistoryRegister,\n MedicalHistoryListByEmergDoctor,\n MedicalHistoryUpdating,\n )\n\nurlpatterns = [\n url(r'^doctorlogin/$', DoctorLogin.as_view(), name='doctorlogin'),\n url(r'^headquarters/(?P\\d+)/$', EnterpriseHeadquarters.as_view(), name='headquarters'),\n\n url(r'^enterprise/$', ClinicsList.as_view(), name='enterprise'),\n url(r'^emergency_attention/(?P\\d+)/$', EmergencyAttentionList.as_view(), name='emergency_attention'),\n\n # list of emergency people by local\n url(r'^specialist_doctor/headquarters/(?P\\d+)/emergency_attention/(?P\\d+)/$', SpecialistDoctor.as_view(), name='specialist_doctor'),\n url(r'^emergency_doctor/(?P\\d+)/$', EmergencyDoctor.as_view(), name='emergency_doctor'),\n\n # medical history by emergencista within doctor_id\n url(r'^emergency_history/doctor/(?P\\d+)/location/(?P\\d+)/emergency_doctor/(?P\\d+)/$',\n MedicalHistoryList.as_view(), name='emergency_history'),\n\n\n # medical history ONLY by emergencista\n url(r'^emergency_history/emergency_doctor/(?P\\d+)/location/(?P\\d+)/$',\n MedicalHistoryListByEmergDoctor.as_view(), name='emergency_doctor_history'),\n\n # medical history by specialist\n url(r'^emergency_history/doctor/(?P\\d+)/location/(?P\\d+)/$',\n MedicalHistorySpecialistList.as_view(), name='emergency_history'),\n\n\n url(r'^medical_history_detail/(?P\\d+)/$', MedicalHistoryDetail.as_view(), name='medical_history_detail'),\n\n # updating medical history detail\n url(r'^medical_history_updating/$', MedicalHistoryUpdating.as_view(), name='medical_history_detail'),\n\n\n # patient\n url(r'^patient_verify/$', PatientView.as_view(), name='patient_verify'),\n url(r'^patient_register/$', PatientRegisterView.as_view(), name='patient_register'),\n\n url(r'^patient_update_token/(?P[0-9]+)/$', PatientUpdateToken.as_view(), name='patient'),\n url(r'^patient_token/(?P[\\w\\-]+)/$', PatientByTokenList.as_view(), name='patient'),\n\n\n # update doctor headquarters\n url(r'^update_doctor_headquarters/(?P\\d+)/doctor/(?P\\d+)/$',\n DoctorUpdateLocationView.as_view(), name='patient'),\n url(r'^medical_history_update/(?P\\d+)/$', MedicalHistoryUpdate.as_view(), name='medical_history_update'),\n\n # ocupacional\n url(r'^patientverify/$', PatientVerifyView.as_view(), name='patient_verify'),\n url(r'^patient_appointments/(?P\\d+)/$', PatientAppointments.as_view(), name='patient_appointments'),\n\n # patient history\n url(r'^patienthistory/(?P\\d+)/$', PatientHistoryView.as_view(), name='patient_history'),\n\n #artifacts measures\n url('^artifact_measurement/(?P[\\w\\-]+)/$', ArtifactMeasurementView.as_view(), name='artifact_measurement'),\n url('^artifact_measurement_tool/$', ArtifactMeasurementTool.as_view(), name='artifact_measurement'),\n\n url('^artifact_weight/$', MeasurementWeight.as_view(), name='artifact_measurement'),\n url('^artifact_body_temperature/$', MeasurementBodyTemperature.as_view(), name='artifact_measurement'),\n url('^artifact_blood_pressure/$', BloodPressure.as_view(), name='artifact_measurement'),\n\n\n url(r'^patient_update/(?P\\d+)/$', PatientUpdate.as_view(), name='patient_update'),\n\n url(r'^call_doctor/(?P[\\w\\-]+)/$', CallDoctorView.as_view(), name='call_doctor'),\n\n # only one character , 1 or 0\n url(r'^call_activate/doctor_id/(?P\\d+)/status/(?P[\\w.]+)/$', CallActivate.as_view(), name='call_activate'),\n\n #medical history register\n #url(r'^medical_history_register/$', MedicalHistoryRegister.as_view(), name='medical_history_register'),\n\n\n\n\n]", "sub_path": "midoc/api/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 4994, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "django.conf.urls.url", "line_number": 34, "usage_type": "call"}, {"api_name": "views.DoctorLogin.as_view", "line_number": 34, "usage_type": "call"}, {"api_name": "views.DoctorLogin", "line_number": 34, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 35, "usage_type": "call"}, {"api_name": "views.EnterpriseHeadquarters.as_view", "line_number": 35, "usage_type": "call"}, {"api_name": "views.EnterpriseHeadquarters", "line_number": 35, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 37, "usage_type": "call"}, {"api_name": "views.ClinicsList.as_view", "line_number": 37, "usage_type": "call"}, {"api_name": "views.ClinicsList", "line_number": 37, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 38, "usage_type": "call"}, {"api_name": "views.EmergencyAttentionList.as_view", "line_number": 38, "usage_type": "call"}, {"api_name": "views.EmergencyAttentionList", "line_number": 38, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 41, "usage_type": "call"}, {"api_name": "views.SpecialistDoctor.as_view", "line_number": 41, "usage_type": "call"}, {"api_name": "views.SpecialistDoctor", "line_number": 41, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 42, "usage_type": "call"}, {"api_name": "views.EmergencyDoctor.as_view", "line_number": 42, "usage_type": "call"}, {"api_name": "views.EmergencyDoctor", "line_number": 42, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 45, "usage_type": "call"}, {"api_name": "views.MedicalHistoryList.as_view", "line_number": 46, "usage_type": "call"}, {"api_name": "views.MedicalHistoryList", "line_number": 46, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 50, "usage_type": "call"}, {"api_name": "views.MedicalHistoryListByEmergDoctor.as_view", "line_number": 51, "usage_type": "call"}, {"api_name": "views.MedicalHistoryListByEmergDoctor", "line_number": 51, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 54, "usage_type": "call"}, {"api_name": "views.MedicalHistorySpecialistList.as_view", "line_number": 55, "usage_type": "call"}, {"api_name": "views.MedicalHistorySpecialistList", "line_number": 55, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 58, "usage_type": "call"}, {"api_name": "views.MedicalHistoryDetail.as_view", "line_number": 58, "usage_type": "call"}, {"api_name": "views.MedicalHistoryDetail", "line_number": 58, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 61, "usage_type": "call"}, {"api_name": "views.MedicalHistoryUpdating.as_view", "line_number": 61, "usage_type": "call"}, {"api_name": "views.MedicalHistoryUpdating", "line_number": 61, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 65, "usage_type": "call"}, {"api_name": "views.PatientView.as_view", "line_number": 65, "usage_type": "call"}, {"api_name": "views.PatientView", "line_number": 65, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 66, "usage_type": "call"}, {"api_name": "views.PatientRegisterView.as_view", "line_number": 66, "usage_type": "call"}, {"api_name": "views.PatientRegisterView", "line_number": 66, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 68, "usage_type": "call"}, {"api_name": "views.PatientUpdateToken.as_view", "line_number": 68, "usage_type": "call"}, {"api_name": "views.PatientUpdateToken", "line_number": 68, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 69, "usage_type": "call"}, {"api_name": "views.PatientByTokenList.as_view", "line_number": 69, "usage_type": "call"}, {"api_name": "views.PatientByTokenList", "line_number": 69, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 73, "usage_type": "call"}, {"api_name": "views.DoctorUpdateLocationView.as_view", "line_number": 74, "usage_type": "call"}, {"api_name": "views.DoctorUpdateLocationView", "line_number": 74, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 75, "usage_type": "call"}, {"api_name": "views.MedicalHistoryUpdate.as_view", "line_number": 75, "usage_type": "call"}, {"api_name": "views.MedicalHistoryUpdate", "line_number": 75, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 78, "usage_type": "call"}, {"api_name": "views.PatientVerifyView.as_view", "line_number": 78, "usage_type": "call"}, {"api_name": "views.PatientVerifyView", "line_number": 78, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 79, "usage_type": "call"}, {"api_name": "views.PatientAppointments.as_view", "line_number": 79, "usage_type": "call"}, {"api_name": "views.PatientAppointments", "line_number": 79, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 82, "usage_type": "call"}, {"api_name": "views.PatientHistoryView.as_view", "line_number": 82, "usage_type": "call"}, {"api_name": "views.PatientHistoryView", "line_number": 82, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 85, "usage_type": "call"}, {"api_name": "views.ArtifactMeasurementView.as_view", "line_number": 85, "usage_type": "call"}, {"api_name": "views.ArtifactMeasurementView", "line_number": 85, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 86, "usage_type": "call"}, {"api_name": "views.ArtifactMeasurementTool.as_view", "line_number": 86, "usage_type": "call"}, {"api_name": "views.ArtifactMeasurementTool", "line_number": 86, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 88, "usage_type": "call"}, {"api_name": "views.MeasurementWeight.as_view", "line_number": 88, "usage_type": "call"}, {"api_name": "views.MeasurementWeight", "line_number": 88, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 89, "usage_type": "call"}, {"api_name": "views.MeasurementBodyTemperature.as_view", "line_number": 89, "usage_type": "call"}, {"api_name": "views.MeasurementBodyTemperature", "line_number": 89, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 90, "usage_type": "call"}, {"api_name": "views.BloodPressure.as_view", "line_number": 90, "usage_type": "call"}, {"api_name": "views.BloodPressure", "line_number": 90, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 93, "usage_type": "call"}, {"api_name": "views.PatientUpdate.as_view", "line_number": 93, "usage_type": "call"}, {"api_name": "views.PatientUpdate", "line_number": 93, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 95, "usage_type": "call"}, {"api_name": "views.CallDoctorView.as_view", "line_number": 95, "usage_type": "call"}, {"api_name": "views.CallDoctorView", "line_number": 95, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 98, "usage_type": "call"}, {"api_name": "views.CallActivate.as_view", "line_number": 98, "usage_type": "call"}, {"api_name": "views.CallActivate", "line_number": 98, "usage_type": "name"}]}
+{"seq_id": "566561077", "text": "#creating a list of synonyms and antonyms\r\nfrom nltk.corpus import wordnet\r\nsynonyms = []\r\nantonyms = []\r\n\r\nfor word in wordnet.synsets(\"funny\"):\r\n for w in word.lemmas():\r\n synonyms.append(w.name())\r\n if w.antonyms():\r\n antonyms.append(w.antonyms()[0].name())\r\n \r\nprint(set(synonyms))\r\nprint(set(antonyms))", "sub_path": "syn_ant.py", "file_name": "syn_ant.py", "file_ext": "py", "file_size_in_byte": 334, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "nltk.corpus.wordnet.synsets", "line_number": 6, "usage_type": "call"}, {"api_name": "nltk.corpus.wordnet", "line_number": 6, "usage_type": "name"}]}
+{"seq_id": "85851267", "text": "import torch\nfrom torch import optim\nimport torch.nn as nn\nfrom torch.nn.modules.activation import ReLU\nimport torchvision\nimport torchvision.transforms as transform\nimport torch.nn.functional as F\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport math\nfrom db import get_imgs,process_image,process_image2\nfrom sklearn.model_selection import train_test_split\n\n# Imagen: 2835 x 3543 todas \n#224 \nclass CNN(nn.Module):\n def __init__(self):\n super(CNN, self).__init__()\n self.conv1 = nn.Sequential(\n nn.Conv2d(in_channels=3, out_channels=8, kernel_size=5, stride=1, padding=1),\n nn.ReLU(),\n nn.MaxPool2d(kernel_size=4,stride=2)\n )\n self.conv2 = nn.Sequential(\n nn.Conv2d(in_channels=8, out_channels=16, kernel_size=5, stride=1, padding=1),\n nn.ReLU(),\n nn.MaxPool2d(kernel_size=4,stride=2)\n )\n self.conv3 = nn.Sequential(\n nn.Conv2d(in_channels=16, out_channels=32, kernel_size=5, stride=1, padding=1),\n nn.ReLU(),\n nn.MaxPool2d(kernel_size=4,stride=2)\n )\n self.conv4 = nn.Sequential(\n nn.Conv2d(in_channels=32, out_channels=64, kernel_size=5, stride=1, padding=1),\n nn.ReLU(),\n nn.MaxPool2d(kernel_size=4,stride=2)\n )\n self.conv5 = nn.Sequential(\n nn.Conv2d(in_channels=64, out_channels=128, kernel_size=5, stride=1, padding=1),\n nn.ReLU(),\n nn.MaxPool2d(kernel_size=4,stride=2)\n )\n self.fc = nn.Sequential(\n nn.Linear(in_features=128*3*4, out_features=768),\n nn.ReLU(),\n nn.Linear(in_features=768, out_features=384),\n nn.ReLU(),\n nn.Linear(in_features=384, out_features=192),\n nn.ReLU(),\n nn.Linear(in_features=192, out_features=96),\n nn.ReLU(),\n nn.Linear(in_features=96, out_features=48),\n nn.ReLU(),\n nn.Linear(in_features=48, out_features=6),\n #nn.Sigmoid(),\n nn.Softmax(),\n )\n \n\n\n # self.conv2 = nn.Conv2d(in_channels=64, out_channels=64*2, kernel_size=4, stride=3, padding=1)\n # self.conv3 = nn.Conv2d(in_channels=64*2, out_channels=64*2*2, kernel_size=4, stride=2, padding=2)\n # self.fc = nn.Linear(in_features=64*2*2*158*198, out_features=6)\n\n # self.layer1 = nn.Sequential(\n # nn.Conv2d(in_channels=1, out_channels=16, kernel_size=3, stride=1, padding=2),\n # nn.ReLU(),\n # nn.MaxPool2d(kernel_size=2, stride=2))\n # self.layer2 = nn.Sequential(\n # nn.Conv2d(16, 32, kernel_size=5, stride=1, padding=2),\n # nn.ReLU(),\n # nn.MaxPool2d(kernel_size=2, stride=2))\n # self.fc = nn.Linear(7*7*32, num_classes)\n \n\n def forward(self, image):\n out = self.conv1(image)\n #print(out.shape)\n \n out = self.conv2(out)\n #print(out.shape)\n\n out = self.conv3(out)\n #print(out.shape)\n\n out = self.conv4(out)\n #print(out.shape)\n\n out = self.conv5(out)\n #print(out.shape)\n\n \n\n out = out.reshape(out.size(0), -1)\n out = self.fc(out)\n #print(\"3\", out.shape)\n \n # out = out.view(out.size(0), -1)\n return out\n\n\nbatch_size = 8\n\ndevice = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu')\n#print(device)\n\nimgs =get_imgs()\nimg_train,img_test = train_test_split(imgs, test_size=0.3, random_state=42,shuffle=True)\n\ntrain_loader = torch.utils.data.DataLoader(dataset=img_train, batch_size=batch_size, shuffle=True)\n#test_loader = torch.utils.data.DataLoader(dataset=imgs, batch_size=batch_size, shuffle=False)\n\ndef test(model,filenames):\n with torch.no_grad(): \n goods = 0\n bads = 0\n for filename in filenames:\n image,labels = process_image2(filename)\n image = image.unsqueeze(0) #[3][2000][3000]\n image = image.to(device)\n labels = labels.unsqueeze(0)\n labels = labels.to(device)\n output = model(image)\n output_label = torch.argmax(output)\n if(output_label == labels):\n goods +=1\n else:\n bads +=1\n print(\"goods are: \",goods)\n print(\"bads are: \",bads)\n \ndef train(model, optimizer, loss_fn, num_epochs):\n \n loss_vals = []\n running_loss =0.0\n # train the model\n\n list_loss= []\n list_time = []\n j=0\n\n for epoch in range(num_epochs):\n filenames = next(iter(train_loader))\n i =0 \n for i,filenames in enumerate(train_loader):\n for filename in filenames:\n image,labels = process_image2(filename)\n image = image.unsqueeze(0) #[3][2000][3000]\n image =image.to(device)\n labels = labels.unsqueeze(0)\n labels = labels.to(device)\n #print(labels.shape)\n # forward \n output = model(image)\n #print(output)\n #print(labels )\n print(output)\n print(labels)\n loss = loss_fn(output, labels)\n # change the params\n optimizer.zero_grad()\n loss.backward()\n optimizer.step() \n list_loss.append(loss.item())\n list_time.append(j)\n j+=1\n if (i+1) % 1 == 0:\n print ('Epoch [{}/{}], Loss: {:.4f}' \n .format(epoch+1, num_epochs,loss.item()))\n \n print('Finished Training Trainset')\n return list_loss\n \nlearning_rate = 0.0001\nepochs = 40\ncnn = CNN().to(device)\nloss = nn.CrossEntropyLoss()\noptimizer = torch.optim.Adam(params=cnn.parameters(), lr=learning_rate)\n\n#print(data.shape)\n#print(cnn(data))\ntrain(cnn,optimizer,loss,epochs)\ntest(cnn,img_test)\n\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 6060, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "torch.nn.Module", "line_number": 17, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 17, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 21, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 23, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 25, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 27, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 28, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 30, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 31, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 32, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 33, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 36, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 37, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 38, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 40, "usage_type": "name"}, {"api_name": "torch.nn.Conv2d", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 41, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 42, "usage_type": "name"}, {"api_name": "torch.nn.MaxPool2d", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 43, "usage_type": "name"}, {"api_name": "torch.nn.Sequential", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 45, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 46, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 47, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 48, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 49, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 50, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 51, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 52, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 53, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 54, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 54, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 55, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 56, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 56, "usage_type": "name"}, {"api_name": "torch.nn.Softmax", "line_number": 58, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 58, "usage_type": "name"}, {"api_name": "torch.device", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 106, "usage_type": "attribute"}, {"api_name": "db.get_imgs", "line_number": 109, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 112, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 112, "usage_type": "attribute"}, {"api_name": "torch.no_grad", "line_number": 116, "usage_type": "call"}, {"api_name": "db.process_image2", "line_number": 120, "usage_type": "call"}, {"api_name": "torch.argmax", "line_number": 126, "usage_type": "call"}, {"api_name": "db.process_image2", "line_number": 149, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 179, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 179, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 180, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 180, "usage_type": "attribute"}]}
+{"seq_id": "195934723", "text": "#разделим выборку на обучающую и тестовую\r\nimport argparse\r\nimport os\r\nimport torch\r\nfrom torch.utils import data\r\nfrom torchvision import transforms\r\nimport torch.nn as nn\r\nimport torch.optim as optim\r\nimport numpy as np\r\nimport itertools\r\nfrom train import *\r\nfrom utils import *\r\n\r\nparser = argparse.ArgumentParser()\r\nparser.add_argument(\"--dataset_photo\", type=str, help=\"name of the dataset with photo\")\r\nparser.add_argument(\"--dataset_style\", type=str, help=\"name of the dataset with style photo\")\r\nparser.add_argument(\"--num_epochs\", default=50, type=int, help=\"number of epochs of training\")\r\nopt = parser.parse_args()\r\nseparate_data(opt.dataset_photo, opt.dataset_style)\r\n\r\n\r\n#создаем каталог для сохранения моделей и изображений\r\nos.makedirs(\"saved_models/\", exist_ok=True)\r\nos.makedirs(\"images/\", exist_ok=True)\r\n\r\nparams = {\r\n 'batch_size':1,\r\n 'input_size':256,\r\n 'crop_size':256,\r\n 'fliplr':True,\r\n #model params\r\n 'num_pool':50,\r\n 'decay_epoch':100,\r\n 'ngf':32, #number of generator filters\r\n 'ndf':64, #number of discriminator filters\r\n 'num_resnet':6, #number of resnet blocks\r\n 'lrG':0.0002, #learning rate for generator\r\n 'lrD':0.0002, #learning rate for discriminator\r\n 'beta1':0.5 , #beta1 for Adam optimizer\r\n 'beta2':0.999 , #beta2 for Adam optimizer\r\n 'lambdaA':10 , #lambdaA for cycle loss\r\n 'lambdaB':10 , #lambdaB for cycle loss\r\n}\r\n\r\n#Загрузим наши изображения\r\ntransform = transforms.Compose([\r\n transforms.Resize(size=(params['input_size'], params['input_size'])), \r\n transforms.ToTensor(),\r\n transforms.Normalize(mean=(0.5, 0.5, 0.5), std=(0.5, 0.5, 0.5))\r\n])\r\n\r\ndir_human_train = '../working/photo/train' #'/../photo/train'\r\ndir_human_test = '../working/photo/test' #'/../photo/test'\r\ndir_simpsons_train = '../working/style/train' #'/../style/train'\r\ndir_simpsons_test = '../working/style/test/' #'/../style/test/'\r\nbatch_size = 1\r\ntrain_data_A = Mydataset(dir_human_train, transform)\r\ntrain_data_loader_A = torch.utils.data.DataLoader(\r\n train_data_A, batch_size=batch_size, shuffle=True, num_workers=batch_size)\r\n\r\ntest_data_A = Mydataset(dir_human_test, transform)\r\ntest_data_loader_A = torch.utils.data.DataLoader(\r\n test_data_A, batch_size=batch_size, shuffle=True, num_workers=batch_size)\r\n\r\ntrain_data_B = Mydataset(dir_simpsons_train, transform)\r\ntrain_data_loader_B = torch.utils.data.DataLoader(\r\n train_data_B, batch_size=batch_size, shuffle=True, num_workers=batch_size)\r\n\r\ntest_data_B = Mydataset(dir_simpsons_test, transform)\r\ntest_data_loader_B = torch.utils.data.DataLoader(\r\n test_data_B, batch_size=batch_size, shuffle=True, num_workers=batch_size)\r\n\r\n\r\n\r\n# Get specific test images\r\ntest_real_A_data = train_data_A.__getitem__(0).unsqueeze(0) # Convert to 4d tensor (BxNxHxW)\r\ntest_real_B_data = train_data_B.__getitem__(1).unsqueeze(0)\r\n\r\n#Build Model \r\nG_A = Generator(3, params['ngf'], 3, params['num_resnet']).cuda() # input_dim, num_filter, output_dim, num_resnet\r\nG_B = Generator(3, params['ngf'], 3, params['num_resnet']).cuda()\r\n\r\nD_A = Discriminator(3, params['ndf'], 1).cuda() # input_dim, num_filter, output_dim\r\nD_B = Discriminator(3, params['ndf'], 1).cuda()\r\n\r\nG_A.normal_weight_init(mean=0.0, std=0.02)\r\nG_B.normal_weight_init(mean=0.0, std=0.02)\r\nD_A.normal_weight_init(mean=0.0, std=0.02)\r\nD_B.normal_weight_init(mean=0.0, std=0.02)\r\n\r\n\r\nG_optimizer = torch.optim.Adam(itertools.chain(G_A.parameters(), G_B.parameters()), lr=params['lrG'], betas=(params['beta1'], params['beta2']))\r\nD_A_optimizer = torch.optim.Adam(D_A.parameters(), lr=params['lrD'], betas=(params['beta1'], params['beta2']))\r\nD_B_optimizer = torch.optim.Adam(D_B.parameters(), lr=params['lrD'], betas=(params['beta1'], params['beta2']))\r\n\r\nMSE_Loss = torch.nn.MSELoss().cuda()\r\nL1_Loss = torch.nn.L1Loss().cuda()\r\n\r\n# # Training GAN\r\nD_A_avg_losses = []\r\nD_B_avg_losses = []\r\nG_A_avg_losses = []\r\nG_B_avg_losses = []\r\ncycle_A_avg_losses = []\r\ncycle_B_avg_losses = []\r\n\r\n# Generated image pool\r\nfake_A_pool = ImagePool(params['num_pool'])\r\nfake_B_pool = ImagePool(params['num_pool'])\r\n\r\nstep = 0\r\ndevice = torch.device(\"cuda:0\" if torch.cuda.is_available() else \"cpu\")\r\nfor epoch in range(opt.num_epochs):\r\n D_A_losses = []\r\n D_B_losses = []\r\n G_A_losses = []\r\n G_B_losses = []\r\n cycle_A_losses = []\r\n cycle_B_losses = []\r\n\r\n # Learing rate decay \r\n if(epoch + 1) > params['decay_epoch']:\r\n D_A_optimizer.param_groups[0]['lr'] -= params['lrD'] / (opt.num_epochs -params['decay_epoch'])\r\n D_B_optimizer.param_groups[0]['lr'] -= params['lrD'] / (opt.num_epochs - params['decay_epoch'])\r\n G_optimizer.param_groups[0]['lr'] -= params['lrD'] / (opt.num_epochs - params['decay_epoch'])\r\n# training \r\n for i, (real_A, real_B) in enumerate(zip(train_data_loader_A, train_data_loader_B)):\r\n\r\n # input image data\r\n real_A = real_A.to(device)\r\n real_B = real_B.to(device)\r\n\r\n # -------------------------- train generator G --------------------------\r\n # A --> B\r\n fake_B = G_A(real_A)\r\n D_B_fake_decision = D_B(fake_B)\r\n G_A_loss = MSE_Loss(D_B_fake_decision, torch.ones(D_B_fake_decision.size()).cuda())\r\n\r\n # forward cycle loss\r\n recon_A = G_B(fake_B)\r\n cycle_A_loss = L1_Loss(recon_A, real_A) * params['lambdaA']\r\n\r\n # B --> A\r\n fake_A = G_B(real_B)\r\n D_A_fake_decision = D_A(fake_A)\r\n G_B_loss = MSE_Loss(D_A_fake_decision, torch.ones(D_A_fake_decision.size()).cuda())\r\n\r\n # backward cycle loss\r\n recon_B = G_A(fake_A)\r\n cycle_B_loss = L1_Loss(recon_B, real_B) * params['lambdaB']\r\n\r\n # Back propagation\r\n G_loss = G_A_loss + G_B_loss + cycle_A_loss + cycle_B_loss\r\n G_optimizer.zero_grad()\r\n G_loss.backward()\r\n G_optimizer.step()\r\n\r\n\r\n # -------------------------- train discriminator D_A --------------------------\r\n D_A_real_decision = D_A(real_A)\r\n D_A_real_loss = MSE_Loss(D_A_real_decision, torch.ones(D_A_real_decision.size()).cuda())\r\n\r\n fake_A = fake_A_pool.query(fake_A)\r\n\r\n D_A_fake_decision = D_A(fake_A)\r\n D_A_fake_loss = MSE_Loss(D_A_fake_decision, torch.zeros(D_A_fake_decision.size()).cuda())\r\n\r\n # Back propagation\r\n D_A_loss = (D_A_real_loss + D_A_fake_loss) * 0.5\r\n D_A_optimizer.zero_grad()\r\n D_A_loss.backward()\r\n D_A_optimizer.step()\r\n\r\n # -------------------------- train discriminator D_B --------------------------\r\n D_B_real_decision = D_B(real_B)\r\n D_B_real_loss = MSE_Loss(D_B_real_decision, torch.ones(D_B_fake_decision.size()).cuda())\r\n\r\n fake_B = fake_B_pool.query(fake_B)\r\n\r\n D_B_fake_decision = D_B(fake_B)\r\n D_B_fake_loss = MSE_Loss(D_B_fake_decision, torch.zeros(D_B_fake_decision.size()).cuda())\r\n\r\n # Back propagation\r\n D_B_loss = (D_B_real_loss + D_B_fake_loss) * 0.5\r\n D_B_optimizer.zero_grad()\r\n D_B_loss.backward()\r\n D_B_optimizer.step()\r\n\r\n # ------------------------ Print -----------------------------\r\n # loss values\r\n D_A_losses.append(D_A_loss.item())\r\n D_B_losses.append(D_B_loss.item())\r\n G_A_losses.append(G_A_loss.item())\r\n G_B_losses.append(G_B_loss.item())\r\n cycle_A_losses.append(cycle_A_loss.item())\r\n cycle_B_losses.append(cycle_B_loss.item())\r\n\r\n if i%100 == 0:\r\n print('Epoch [%d/%d], Step [%d/%d], D_A_loss: %.4f, D_B_loss: %.4f, G_A_loss: %.4f, G_B_loss: %.4f'\r\n % (epoch+1, opt.num_epochs, i+1, len(train_data_loader_A), D_A_loss.item(), D_B_loss.item(), G_A_loss.item(), G_B_loss.item()))\r\n\r\n step += 1\r\n\r\n D_A_avg_loss = torch.mean(torch.FloatTensor(D_A_losses))\r\n D_B_avg_loss = torch.mean(torch.FloatTensor(D_B_losses))\r\n G_A_avg_loss = torch.mean(torch.FloatTensor(G_A_losses))\r\n G_B_avg_loss = torch.mean(torch.FloatTensor(G_B_losses))\r\n cycle_A_avg_loss = torch.mean(torch.FloatTensor(cycle_A_losses))\r\n cycle_B_avg_loss = torch.mean(torch.FloatTensor(cycle_B_losses))\r\n\r\n # avg loss values for plot\r\n D_A_avg_losses.append(D_A_avg_loss.item())\r\n D_B_avg_losses.append(D_B_avg_loss.item())\r\n G_A_avg_losses.append(G_A_avg_loss.item())\r\n G_B_avg_losses.append(G_B_avg_loss.item())\r\n cycle_A_avg_losses.append(cycle_A_avg_loss.item())\r\n cycle_B_avg_losses.append(cycle_B_avg_loss.item())\r\n\r\n # Show result for test image\r\n test_real_A = test_real_A_data.cuda()\r\n test_fake_B = G_A(test_real_A)\r\n test_recon_A = G_B(test_fake_B)\r\n\r\n test_real_B = test_real_B_data.cuda()\r\n test_fake_A = G_B(test_real_B)\r\n test_recon_B = G_A(test_fake_A)\r\n\r\n plot_train_result([test_real_A, test_real_B], [test_fake_B, test_fake_A], [test_recon_A, test_recon_B],\r\n epoch, save=True)\r\n \r\n # Save model checkpoints\r\n torch.save(G_A.state_dict(), \"saved_models/G_AB_%d.pth\" % (epoch + 1))\r\n torch.save(G_B.state_dict(), \"saved_models/G_BA_%d.pth\" % (epoch + 1))\r\n torch.save(D_A.state_dict(), \"saved_models/D_A_%d.pth\" % (epoch + 1))\r\n torch.save(D_B.state_dict(), \"saved_models/D_B_%d.pth\" % (epoch + 1))\r\n\r\nall_losses = pd.DataFrame()\r\nall_losses['D_A_avg_losses'] = D_A_avg_losses\r\nall_losses['D_B_avg_losses'] = D_B_avg_losses\r\nall_losses['G_A_avg_losses'] = G_A_avg_losses\r\nall_losses['G_B_avg_losses'] = G_B_avg_losses\r\nall_losses['cycle_A_avg_losses'] = cycle_A_avg_losses\r\nall_losses['cycle_B_avg_losses'] = cycle_B_avg_losses\r\nall_losses.to_csv('avg_losses',index=False)\r\n\r\n", "sub_path": "cyclegan_kaggle.py", "file_name": "cyclegan_kaggle.py", "file_ext": "py", "file_size_in_byte": 9726, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 14, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 23, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 24, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 46, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 46, "usage_type": "name"}, {"api_name": "torchvision.transforms.Resize", "line_number": 47, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 47, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 48, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 48, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 49, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 49, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 58, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 58, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 62, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 62, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 66, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 70, "usage_type": "attribute"}, {"api_name": "torch.optim.Adam", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 92, "usage_type": "attribute"}, {"api_name": "itertools.chain", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 93, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 93, "usage_type": "attribute"}, {"api_name": "torch.optim.Adam", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 94, "usage_type": "attribute"}, {"api_name": "torch.nn.MSELoss", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 96, "usage_type": "attribute"}, {"api_name": "torch.nn.L1Loss", "line_number": 97, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 97, "usage_type": "attribute"}, {"api_name": "torch.device", "line_number": 112, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 112, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 112, "usage_type": "attribute"}, {"api_name": "torch.ones", "line_number": 137, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 146, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 161, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 166, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 176, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 181, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 204, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 204, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 205, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 205, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 206, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 206, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 207, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 207, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 208, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 208, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 209, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 209, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 232, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 233, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 234, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 235, "usage_type": "call"}]}
+{"seq_id": "171308787", "text": "import os\nimport random\nimport logging\nimport torch\nimport re\nimport json\nimport locale\nimport unidecode\nimport string\nimport numpy as np\nimport pandas as pd\nimport seaborn as sns\nfrom IPython.display import Markdown, display\nfrom operator import itemgetter\nfrom datetime import datetime, timedelta\nfrom random import random, randint, choice, seed\nfrom torch.nn import CrossEntropyLoss\nfrom torch.utils.data import DataLoader, RandomSampler, SequentialSampler, TensorDataset\nfrom tqdm import tqdm\nfrom faker import Faker\nfrom transformers import get_linear_schedule_with_warmup\nfrom transformers import AutoConfig, AutoModelForTokenClassification, AutoTokenizer\nfrom sklearn.metrics import precision_score as sk_precision_score, recall_score as sk_recall_score, \\\n f1_score as sk_f1_score, confusion_matrix as sk_confusion_matrix\n\nfrom utils import convert_to_features, switch_entity, find_sub_list, get_text_and_labels\nfrom focal_loss import FocalLoss\n\n\nclass Ner:\n\n def __init__(self, _inputs, log_level=logging.INFO):\n\n self.logger = logging.getLogger(__name__)\n self.log_level = log_level\n\n self.inputs = _inputs\n self.pad_token_label_id = CrossEntropyLoss().ignore_index\n self.set_seed(_inputs[\"seed\"])\n self.list_entities = _inputs[\"entities\"]\n self.underlines = {\n ent: '#%02x%02x%02x' % (int(sns.color_palette('pastel', len(self.list_entities))[i][0] * 255),\n int(sns.color_palette('pastel', len(self.list_entities))[i][1] * 255),\n int(sns.color_palette('pastel', len(self.list_entities))[i][2] * 255))\n for i, ent in enumerate(self.list_entities)}\n self.list_regex = _inputs[\"regex\"]\n self.max_seq_length = _inputs[\"max_seq_length\"]\n self.per_gpu_batch_size = _inputs[\"per_gpu_batch_size\"]\n self.model_path = _inputs[\"model_path\"]\n self.tokenizer_path = _inputs[\"tokenizer_path\"]\n self.labels_format = _inputs[\"labels_format\"]\n\n self.device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\n\n # Initialisation des paramètres\n self.adam_epsilon, self.learning_rate, self.max_steps, self.gradient_accumulation_steps, self.num_train_epochs,\\\n self.max_grad_norm, self.warmup_steps, self.weight_decay, self.white_space_token, self.loss_function, \\\n self.output_dir = [None] * 11\n\n def evaluate_model(self, corpus):\n \"\"\"\n Evaluation du modèle pour les entités précisées.\n :param corpus: DataFrame du corpus à utiliser pour l'évaluation\n \"\"\"\n\n # Loading labels\n labels, labels_weight = self.load_labels(None)\n # Loading model and tokenizer\n model, tokenizer = self.load_model_and_tokenizer(labels)\n # Evaluation\n eval_dataset = self.load_and_cache_texts(corpus, tokenizer, labels)\n\n # Save config and logs\n self.save_config_and_logs()\n\n model.to(self.device)\n result, _ = self.run_predict_and_eval(eval_dataset, model, tokenizer, labels, self.model_path)\n\n def evaluate_and_display_results(self, eval_loss, real_labels, predicted_labels, labels, no_saving, model_file):\n \"\"\"\n Evalue les performances d'un modèle et sauvegarde les résultats dans le dossier du modèle.\n :param eval_loss: valeur moyenne de la fonction de perte\n :param real_labels: liste des labels correspondant aux tokens du texte\n :param predicted_labels: liste des labels prédits par l'algorithme\n :param labels: liste des différents labels possibles\n :param no_saving: booléen précisant si les résultats de l'évaluation doivent etre enregistrés ou non\n :param model_file: chemin vers le modèle évalué\n return: résultats de l'évaluation sous forme de dictionnaire\n \"\"\"\n # Computes metrics\n results = self.get_scores(real_labels, predicted_labels, labels, eval_loss)\n # Displays results and saves them to a file\n # for key in sorted(results.keys()):\n # self.logger.info(\" %s = %s\", key, str(results[key]))\n self.logger.info(\"1. results by entity\\n\")\n for ent in self.list_entities:\n end_of_line = \"\\n\" if ent == self.list_entities[-1] else \"\"\n self.logger.info(\"\\t%s : %s%s\", ent, str(results[ent]), end_of_line)\n self.logger.info(\"2. global results\\n\")\n other_keys = set(results.keys()) - set(self.list_entities) - {\"confusion_matrix\"}\n for key in other_keys:\n end_of_line = \"\\n\" if key == list(other_keys)[-1] else \"\"\n self.logger.info(\"\\t%s = %s%s\", key, str(results[key]), end_of_line)\n self.logger.info(\"3. confusion matrix\\n\")\n self.logger.info(\"\\t%s\\n\", str(results[\"confusion_matrix\"]))\n # Saves results\n if not no_saving:\n output_eval_file = model_file.replace('.pt', '_eval_results.txt')\n with open(output_eval_file, \"w\") as writer:\n for key in sorted(results.keys()):\n writer.write(\"{} = {}\\n\".format(key, str(results[key])))\n\n return results\n\n def extract_info_from_batch(self, tokenizer, batch, _output_probabilities, label_map, threshold=None):\n \"\"\"\n Extraction des différentes informations contenues dans un batch de données.\n :param tokenizer: tokenizer du modèle\n :param batch: batch de données\n :param _output_probabilities: probabilités des différentes classes données par l'algorithme\n :param label_map: dictionnaire de correspondance entre les libellés des labels et leurs identifiants\n :param threshold: seuils associés à chaque classe. Si la probabilité de sortie dépasse ce seuil, on considère\n que l'algorithme l'a prédite meme si ce n'est pas la probabilité maximale.\n :return:\n \"\"\"\n\n token_2_ignore = [tokenizer.sep_token, tokenizer.cls_token, tokenizer.pad_token]\n token_ids_2_ignore = [tokenizer.sep_token_id, tokenizer.cls_token_id, tokenizer.pad_token_id]\n\n # Extract texts and predicted labels\n text_tokens = [[tokenizer.convert_ids_to_tokens(int(x)) for x in y] for y in batch[0]]\n labels_probabilities = _output_probabilities.detach().to(\"cpu\").numpy()\n predicted_labels_ids = np.argmax(labels_probabilities, axis=2)\n\n # Using manual threshold\n if threshold is not None:\n for i, row in enumerate(labels_probabilities):\n for j, token in enumerate(row):\n if any([x >= threshold[ind] for ind, x in enumerate(token)][1:]) and np.argmax(token) == 0:\n _rescaled_tokens = [x if (ind != 0) and (x >= threshold[ind]) else -1000 for ind, x in\n enumerate(token)]\n predicted_labels_ids[i][j] = np.argmax(_rescaled_tokens)\n\n predicted_labels = [[label_map[x] for x in y] for y in predicted_labels_ids]\n # Delete functional tokens\n labels_probabilities = [[\", \".join([str(z) for z in y]) for y in x] for x in labels_probabilities]\n _joined = [[(x, y, z) for x, y, z in zip(text_tokens[i], predicted_labels[i], labels_probabilities[i]) if\n x not in token_2_ignore]\n for i in range(len(text_tokens))]\n _valid_examples = [i for i, x in enumerate(_joined) if len(x) > 0]\n _joined = [list(zip(*_joined[i])) for i in _valid_examples]\n text_tokens = [list(x[0]) for x in _joined]\n predicted_labels = [list(x[1]) for x in _joined]\n labels_probabilities = [list(x[2]) for x in _joined]\n # Extract real labels\n real_labels = [[label_map[int(x)] for x in y if x != self.pad_token_label_id] for y in batch[3]]\n real_labels = [x for i, x in enumerate(real_labels) if i in _valid_examples]\n # Extract file names\n file_tokens = [[tokenizer.convert_ids_to_tokens(int(x)) for x in y if x not in token_ids_2_ignore] for y in\n batch[4]]\n files = [\"\".join([x.replace('▁', ' ') for x in y]).strip() for y in file_tokens]\n files = [x for i, x in enumerate(files) if i in _valid_examples]\n # Extract text part\n text_parts = [int(x) for x in batch[5]]\n text_parts = [x for i, x in enumerate(text_parts) if i in _valid_examples]\n\n return files, text_parts, text_tokens, real_labels, predicted_labels, labels_probabilities\n\n def find_regex_entities(self, corpus):\n \"\"\"\n Détection des entités repérées par des expressions régulières et remplacement des tags correspondant par le\n label adéquat.\n param corpus: corpus de textes\n return : corpus avec nouveaux labels\n \"\"\"\n\n func_dic = {\"TIME\": self.regex_time, \"PHONE\": self.regex_phone, \"IMMAT\": self.regex_immat,\n \"EMAIL\": self.regex_email}\n\n for regex in self.list_regex:\n corpus = corpus.apply(lambda x: func_dic[regex](x), axis=1)\n\n return corpus\n\n @staticmethod\n def get_pos_class_freq(train_df):\n \"\"\"\n Calcule le vecteur de poids pour la dernière couche du réseau, après l'encodeur. Le poids de chaque classe de\n sortie est inversement proportionnel à la fréquence de la classe dans le dataset d'entrainement.\n :param train_df: DataFrame du corpus d'entrainement\n :return: dictionnaire associant un poids à chaque classe\n \"\"\"\n\n count_df = pd.Series([y for x in train_df.labels.values for y in x]).value_counts().reset_index()\n return {e[0]: e[1] for e in count_df[['index', 0]].values}\n\n def get_scores(self, real_labels, predicted_labels, labels, eval_loss):\n \"\"\"\n Calcul des performances du modèle (f1, rappel et précision) au global et pour chaque entité.\n :param real_labels: liste des labels correspondant aux tokens du texte\n :param predicted_labels: liste des labels prédits par l'algorithme\n :param labels: liste des différents labels possibles\n :param eval_loss: valeur moyenne de la fonction de perte\n :return: dictionnaire des performances de l'algorithme.\n \"\"\"\n\n _s_labels = list(sorted(labels))\n _flat_real_labels = [x for y in real_labels for x in y]\n _flat_predicted_labels = [x for y in predicted_labels for x in y]\n _flat_real_labels_type_only = [x.split(\"-\")[-1] for y in real_labels for x in y]\n _flat_predicted_labels_type_only = [x.split(\"-\")[-1] for y in predicted_labels for x in y]\n _labels_type_only = list(set([x.split(\"-\")[-1] for x in labels if x != 'O']))\n cm = sk_confusion_matrix(_flat_real_labels, _flat_predicted_labels, labels=_s_labels)\n cm = np.concatenate((np.transpose(np.array([[''] + _s_labels])), np.concatenate((np.array([_s_labels]),\n cm), axis=0)), axis=1)\n\n results = {\n \"loss\": eval_loss,\n \"precision (entity type only)\": sk_precision_score(_flat_real_labels_type_only,\n _flat_predicted_labels_type_only,\n labels=_labels_type_only, average='micro',\n zero_division=0),\n \"precision (BIO labels)\": sk_precision_score(_flat_real_labels, _flat_predicted_labels,\n labels=[x for x in labels if x != \"O\"], average='micro',\n zero_division=0),\n \"recall (entity type only)\": sk_recall_score(_flat_real_labels_type_only, _flat_predicted_labels_type_only,\n labels=_labels_type_only, average='micro', zero_division=0),\n \"recall (BIO labels)\": sk_recall_score(_flat_real_labels, _flat_predicted_labels,\n labels=[x for x in labels if x != \"O\"], average='micro',\n zero_division=0),\n \"f1 (entity type only)\": sk_f1_score(_flat_real_labels_type_only, _flat_predicted_labels_type_only,\n labels=_labels_type_only, average='micro', zero_division=0),\n \"f1 (BIO labels)\": sk_f1_score(_flat_real_labels, _flat_predicted_labels,\n labels=[x for x in labels if x != \"O\"], average='micro', zero_division=0),\n \"confusion_matrix\": cm\n }\n\n for ent in self.list_entities:\n _preds = [1 if x == ent else 0 for x in _flat_predicted_labels_type_only]\n _reals = [1 if x == ent else 0 for x in _flat_real_labels_type_only]\n results[ent] = f\"precision: {sk_precision_score(_reals, _preds, zero_division=0)}, \" \\\n f\"recall: {sk_recall_score(_reals, _preds, zero_division=0)}\"\n\n return results\n\n def get_corpus_stats(self, corpus):\n \"\"\"\n Ajoute aux logs les caractéristiques du corpus traité.\n :param corpus: DataFrame du corpus de textes\n \"\"\"\n _global = f\"{len(corpus)} textes dans le corpus, soit {sum([len(x) for x in corpus.text.tolist()])} tokens.\\n\"\n _per_entity = \"Nombre d'entités:\\n\"\n for ent in self.list_entities:\n _per_entity += f\"\\t- {ent} : {[x for y in corpus.labels.tolist() for x in y].count(ent)}\\n\"\n self.logger.info(\"%s\\n%s\", _global, _per_entity)\n\n def load_model_and_tokenizer(self, labels):\n \"\"\"\n Chargement du modèle et du tokenizer associé.\n :param labels: liste des différents labels possibles\n :return: modèle et tokenizer\n \"\"\"\n\n if self.model_path.endswith(\".pt\"):\n if self.device.type == \"cpu\":\n model = torch.load(self.model_path, map_location=torch.device('cpu'))\n else:\n model = torch.load(self.model_path)\n\n tokenizer_path = os.path.join(self.tokenizer_path, model.__class__.__name__)\n\n tokenizer = AutoTokenizer.from_pretrained(tokenizer_path)\n self.white_space_token = tokenizer.tokenize(\"le\")[0].replace(\"le\", \"\")\n else:\n config_file = os.path.join(self.model_path, \"config.json\")\n\n config = AutoConfig.from_pretrained(config_file)\n config.num_labels = len(labels)\n\n model = AutoModelForTokenClassification.from_config(config)\n\n tokenizer = AutoTokenizer.from_pretrained(self.model_path)\n self.white_space_token = tokenizer.tokenize(\"le\")[0].replace(\"le\", \"\")\n\n return model, tokenizer\n\n def load_labels(self, train_df):\n \"\"\"\n Génère les labels en fonction du format choisi ainsi que leurs poids en fonction de leur fréquence\n d'apparition dans le corpus.\n :param train_df: corpus de textes d'entrainement\n :return: liste des labels et poids correspondants\n \"\"\"\n if self.labels_format == \"BIO\":\n labels = [\"O\"] + [y for z in [[f\"B-{x}\", f\"I-{x}\"] for x in self.list_entities] for y in z]\n else:\n labels = [\"O\"] + self.list_entities\n\n # Les poids des différents labels sont calculés à partir de leur fréquence d'apparition.\n if (train_df is None) or (len(train_df) == 0):\n labels_weights = [1 for _ in labels]\n # Si l'on veut uniquement faire des prédictions, on peut se contenter d'un vecteur de poids constant\n else:\n freqs = self.get_pos_class_freq(train_df)\n labels_weights = np.array([freqs.get(key, None) for key in labels], dtype=np.float64)\n labels_weights = [np.nanmax(labels_weights) / x if not np.isnan([x]) else np.nanmax(labels_weights) for x in\n labels_weights]\n labels_weights = [np.log(x) if x != 1 else x for x in labels_weights]\n labels_weights = torch.tensor(labels_weights).float()\n labels_weights = labels_weights.to(device=self.device)\n\n return labels, labels_weights\n\n def load_and_cache_texts(self, corpus_df, tokenizer, labels):\n \"\"\"\n Charge les différents textes du corpus dans un TensorDataset.\n :param corpus_df: DataFrame du corpus de textes\n :param tokenizer: tokeinzer associé au modèle prédictif\n :param labels: liste des différents labels possibles\n :return:TensorDataset du corpus\n \"\"\"\n\n tokenizer_special_tokens = {\"cls_token\": tokenizer.cls_token, \"cls_token_segment_id\": 0,\n \"sep_token\": tokenizer.sep_token, \"sep_token_extra\": False,\n \"pad_on_left\": False, \"cls_token_at_end\": False,\n \"pad_token\": tokenizer.convert_tokens_to_ids([tokenizer.pad_token])[0],\n \"pad_token_segment_id\": 0, \"pad_token_label_id\": self.pad_token_label_id,\n \"sequence_a_segment_id\": 0, \"mask_padding_with_zero\": True}\n\n features = convert_to_features(corpus_df, labels, self.max_seq_length, tokenizer, tokenizer_special_tokens)\n\n # Convert to Tensors and build dataset\n all_text_token_ids = torch.tensor([f.text_token_ids for f in features], dtype=torch.long)\n all_text_mask = torch.tensor([f.text_mask for f in features], dtype=torch.long)\n all_segment_ids = torch.tensor([f.segment_ids for f in features], dtype=torch.long)\n all_label_ids = torch.tensor([f.label_ids for f in features], dtype=torch.long)\n all_file_token_ids = torch.tensor([f.file_token_ids for f in features], dtype=torch.long)\n all_text_parts_ids = torch.tensor([f.text_part_index for f in features], dtype=torch.long)\n\n dataset = TensorDataset(all_text_token_ids, all_text_mask, all_segment_ids, all_label_ids, all_file_token_ids,\n all_text_parts_ids)\n return dataset\n\n def loss_with_weights(self, labels, attention_mask, preds, labels_weights):\n \"\"\"\n Calcule la fonction de perte (Focal loss ou Cross Entropy loss) en prenant en compte les poids associés à chaque\n catégorie.\n :param labels: labels associés à chaque token\n :param attention_mask: masque d'attention\n :param preds: prédictions de l'algorithme pour chaque token\n :param labels_weights: poids associées à chaque classe\n :return: perte\n \"\"\"\n\n loss = None\n if labels is not None:\n if self.loss_function == \"FocalLoss\":\n loss_fct = FocalLoss(alpha=labels_weights, gamma=2)\n else:\n loss_fct = CrossEntropyLoss(labels_weights)\n\n # Only keep active parts of the loss\n if attention_mask is not None:\n active_loss = attention_mask.view(-1) == 1\n active_logits = preds.view(-1, len(labels_weights))\n active_labels = torch.where(\n active_loss, labels.view(-1), torch.tensor(loss_fct.ignore_index).type_as(labels)\n )\n loss = loss_fct(active_logits, active_labels)\n else:\n loss = loss_fct(preds.view(-1, len(labels_weights)), labels.view(-1))\n\n return loss\n\n def parse_input_json(self):\n \"\"\"\n Récupère les paramètres du fichier d'entrée au format json et les assigne à des variables de classe.\n \"\"\"\n\n try:\n # Paramètres de configuration liés à l'entraînement (facultatifs pour de la prédiction ou de l'évaluation)\n self.adam_epsilon = self.inputs[\"adam_epsilon\"]\n self.learning_rate = self.inputs[\"learning_rate\"]\n self.max_steps = self.inputs[\"max_steps\"]\n self.gradient_accumulation_steps = self.inputs[\"gradient_accumulation_steps\"]\n self.num_train_epochs = self.inputs[\"num_train_epochs\"]\n self.max_grad_norm = self.inputs[\"max_grad_norm\"]\n self.warmup_steps = self.inputs[\"warmup_steps\"]\n self.weight_decay = self.inputs[\"weight_decay\"]\n self.loss_function = self.inputs[\"loss_function\"]\n except (Exception,):\n _mandatory_parameters = [\"adam_epsilon\", \"learning_rate\", \"max_seq_length\", \"max_steps\",\n \"gradient_accum, _steps\", \"num_train_epochs\", \"max_grad_norm\",\n \"per_gpu_batch_size\", \"warmup_steps\", \"weight_decay\",\n \"loss_function\", \"output_dir\"]\n _missing_ones = [x for x in _mandatory_parameters if x not in self.inputs.keys()]\n self.logger.error(f\"Missing training parameter(s): {_missing_ones}\")\n\n def predict_with_model(self, corpus, threshold):\n \"\"\"\n Détecte les entités nommées voulues dans un corpus donné.\n :param corpus: DataFrame de corpus de textes\n :param threshold: seuils de détection manuels. Si la probabilité d'une entité dépasse ce seuil, on prédit\n cette entité meme si elle ne correspond pas à la probabilité maximale.\n return: DataFrame du corpus enrichi des annotations\n \"\"\"\n\n # Loading labels\n labels, labels_weight = self.load_labels(None)\n # Loading model and tokenizer\n model, tokenizer = self.load_model_and_tokenizer(labels)\n # Evaluation\n predict_dataset = self.load_and_cache_texts(corpus, tokenizer, labels)\n\n model.to(self.device)\n _, processed_corpus = self.run_predict_and_eval(predict_dataset, model, tokenizer, labels, None,\n no_evaluation=True, threshold=threshold)\n\n return processed_corpus\n\n @staticmethod\n def regex_immat(row):\n \"\"\"\n Finds immats in texts using REGEX rules.\n :param row: DataFrame row\n :return: pd.Series with file, text and label columns\n \"\"\"\n\n # Loads text\n raw_ppel = row[\"raw_text\"]\n\n # REGEX immat patterns and exceptions\n regex_pattern = r\"[\\s\\\"\\''\\(\\,\\.][a-zA-Z]{2}[\\s\\.-]?[0-9]{3}[\\s\\.-]?[a-zA-Z]{2}[\\s\\\"\\''\\)\\,\\.]\"\n exceptions = ['de', 'et', 'le', 'go']\n\n # Finds immat patterns\n plaque = []\n for _immat in re.finditer(regex_pattern, raw_ppel):\n s = _immat.start()\n e = _immat.end()\n if not ((raw_ppel[s + 1:s + 3] in exceptions) and (raw_ppel[e - 3:e - 1] in exceptions)):\n plaque.append(raw_ppel[s + 1:e - 1])\n\n # Creates labels\n splitted_text = row[\"text\"]\n if \"predicted_labels\" in row.keys():\n bio_tags = row[\"predicted_labels\"]\n else:\n bio_tags = [\"O\" for _ in row[\"labels\"]]\n plaque = [x.split(\" \") for x in plaque]\n _ppel = splitted_text.copy()\n for _immat in plaque:\n ind = find_sub_list(_immat, _ppel)\n if ind is None:\n ind = find_sub_list(_immat, _ppel, strict=False)\n if ind is None:\n print(f\"entity {_immat} not found in text\")\n continue\n for i, _tag in zip(ind, _immat):\n bio_tags[i] = 'IMMAT'\n _ppel = [None for _ in _ppel[:ind[0] + len(_immat)]] + _ppel[min(len(_ppel), ind[0] + len(_immat)):]\n\n return pd.Series(\n {\"file\": row[\"file\"], \"raw_text\": row[\"raw_text\"], \"text\": splitted_text, \"labels\": row[\"labels\"],\n \"predicted_labels\": bio_tags})\n\n @staticmethod\n def regex_email(row):\n \"\"\"\n Finds e-mails in texts using REGEX rules.\n :param row: DataFrame row\n :return: pd.Series with file, text and label columns\n \"\"\"\n\n # Loads text\n raw_ppel = row[\"raw_text\"]\n\n # REGEX time patterns\n regex_pattern = r\"([a-zA-Z0-9_.+-]+@[a-zA-Z0-9-]+\\.[a-zA-Z0-9-.]+)\"\n\n # Finds e-mail patterns\n emails = []\n for _mail in re.finditer(regex_pattern, raw_ppel):\n s = _mail.start()\n e = _mail.end()\n if raw_ppel[e - 1] == '.':\n emails.append(raw_ppel[s:e - 1])\n else:\n emails.append(raw_ppel[s:e])\n\n # Creates labels\n splitted_text = row[\"text\"]\n if \"predicted_labels\" in row.keys():\n bio_tags = row[\"predicted_labels\"]\n else:\n bio_tags = [\"O\" for _ in row[\"labels\"]]\n emails = [x.split(\" \") for x in emails]\n _ppel = splitted_text.copy()\n for _mail in emails:\n ind = find_sub_list(_mail, _ppel, strict=False)\n if ind is None:\n print(f\"entity {_mail} not found in text\")\n continue\n for i, _tag in zip(ind, _mail):\n bio_tags[i] = 'EMAIL'\n _ppel = [None for _ in _ppel[:ind[0] + len(_mail)]] + _ppel[min(len(_ppel), ind[0] + len(_mail)):]\n\n return pd.Series(\n {\"file\": row[\"file\"], \"raw_text\": row[\"raw_text\"], \"text\": splitted_text, \"labels\": row[\"labels\"],\n \"predicted_labels\": bio_tags})\n\n @staticmethod\n def regex_phone(row):\n \"\"\"\n Finds phone numbers in texts using REGEX rules.\n :param row: DataFrame row\n :return: pd.Series with file, text and label columns\n \"\"\"\n\n # Loads text\n raw_ppel = row[\"raw_text\"]\n # REGEX time patterns\n regex_pattern = [\n r\"[\\s\\\"\\''\\(\\,\\.]0[0-9][\\s\\.-]?([0-9]{2}[\\s\\.-]?){3}[0-9]{2}[\\s\\\"\\''\\)\\,\\.]\",\n r\"[\\s\\\"\\''\\(\\,\\.]\\+[0-9]{1,4}[\\s\\.-]?[0-9][\\s\\.-]?([0-9]{2}[\\s\\.-]?){3}[0-9]{2}[\\s\\\"\\''\\)\\,\\.]\",\n r\"[\\s\\\"\\''\\(\\,\\.][0-9]{4}[\\s\\.-][0-9]{3}[\\s\\.-][0-9]{3}[\\s\\\"\\''\\)\\,\\.]\"\n ]\n\n # Finds phone number patterns\n phones = []\n for pattern in regex_pattern:\n for _phone in re.finditer(pattern, raw_ppel):\n s = _phone.start() + 1\n e = _phone.end() - 1\n phones.append((s, raw_ppel[s:e].strip()))\n phones.sort(key=itemgetter(0))\n phones = [x[1] for x in phones]\n\n # Creates labels\n splitted_text = row[\"text\"]\n if \"predicted_labels\" in row.keys():\n bio_tags = row[\"predicted_labels\"]\n else:\n bio_tags = [\"O\" for _ in row[\"labels\"]]\n phones = [x.split(\" \") for x in phones]\n _ppel = splitted_text.copy()\n for _phone in phones:\n ind = find_sub_list(_phone, _ppel)\n if ind is None:\n ind = find_sub_list(_phone, _ppel, strict=False)\n if ind is None:\n print(f\"entity {_phone} not found in text\")\n continue\n for i, _tag in zip(ind, _phone):\n bio_tags[i] = 'PHONE'\n _ppel = [None for _ in _ppel[:ind[0] + len(_phone)]] + _ppel[min(len(_ppel), ind[0] + len(_phone)):]\n\n return pd.Series(\n {\"file\": row[\"file\"], \"raw_text\": row[\"raw_text\"], \"text\": splitted_text, \"labels\": row[\"labels\"],\n \"predicted_labels\": bio_tags})\n\n @staticmethod\n def regex_time(row):\n \"\"\"\n Finds times in texts using REGEX rules.\n :param row: DataFrame row\n :return: pd.Series with file, text and label columns\n \"\"\"\n\n # Loads text\n raw_ppel = row[\"raw_text\"]\n\n # REGEX time patterns\n regex_pattern = [r\"[0-9][0-9]?[\\:][0-9][0-9]?\", r\"[0-9][0-9]?[Hh][0-9]?[0-9]?\",\n r\"[0-9][0-9]?\\s[hH][eE][uU][rR][eE][s]?\\s[0-9]?[0-9]?\",\n r\"[0-9][0-9]?\\s[Hh]\\s[0-9]?[0-9]?\"]\n\n # Finds time patterns\n times = []\n for pattern in regex_pattern:\n for _time in re.finditer(pattern, raw_ppel):\n s = _time.start()\n e = _time.end()\n times.append((s, raw_ppel[s:e].strip()))\n times.sort(key=itemgetter(0))\n times = [x[1] for x in times]\n\n # Creates labels\n splitted_text = row[\"text\"]\n if \"predicted_labels\" in row.keys():\n bio_tags = row[\"predicted_labels\"]\n else:\n bio_tags = [\"O\" for _ in row[\"labels\"]]\n times = [x.split(\" \") for x in times]\n _ppel = splitted_text.copy()\n for _time in times:\n ind = find_sub_list(_time, _ppel)\n if ind is None:\n ind = find_sub_list(_time, _ppel, strict=False)\n if ind is None:\n print(f\"entity {_time} not found in text\")\n continue\n for i, _tag in zip(ind, _time):\n bio_tags[i] = 'TIME'\n _ppel = [None for _ in _ppel[:ind[0] + len(_time)]] + _ppel[min(len(_ppel), ind[0] + len(_time)):]\n\n return pd.Series(\n {\"file\": row[\"file\"], \"raw_text\": row[\"raw_text\"], \"text\": splitted_text, \"labels\": row[\"labels\"],\n \"predicted_labels\": bio_tags})\n\n def run_predict_and_eval(self, dataset, model, tokenizer, labels, save_folder, no_evaluation=False, no_saving=False,\n threshold=None):\n \"\"\"\n\n :param dataset:\n :param model:\n :param tokenizer:\n :param labels:\n :param save_folder:\n :param no_evaluation:\n :param no_saving:\n :param threshold:\n :return:\n \"\"\"\n\n batch_size = self.per_gpu_batch_size\n sampler = SequentialSampler(dataset)\n dataloader = DataLoader(dataset, sampler=sampler, batch_size=batch_size)\n label_map = {i: label for i, label in enumerate(labels)}\n if threshold is not None:\n threshold = {ind: threshold[ent] if ent in threshold.keys() else 1000 for ind, ent in label_map.items()}\n\n processed_corpus = pd.DataFrame()\n eval_loss = 0.0\n nb_eval_steps = 0\n model.to(self.device)\n model.eval()\n for batch in tqdm(dataloader, desc=\"Evaluating\", position=0, leave=True):\n batch = tuple(t.to(self.device) for t in batch)\n with torch.no_grad():\n _inputs = {\"input_ids\": batch[0], \"attention_mask\": batch[1], \"labels\": batch[3],\n \"token_type_ids\": None}\n outputs = model(**_inputs)\n tmp_eval_loss, _output_probabilities = outputs[:2]\n eval_loss += tmp_eval_loss.item()\n\n nb_eval_steps += 1\n\n files, text_parts, text_tokens, real_labels, predicted_labels, labels_probabilities = \\\n self.extract_info_from_batch(tokenizer, batch, _output_probabilities, label_map, threshold)\n\n processed_corpus = pd.concat([processed_corpus, pd.DataFrame({\"file\": files, \"text_part\": text_parts,\n \"text\": text_tokens,\n \"labels\": real_labels,\n \"predicted_labels\": predicted_labels,\n \"labels_probabilities\":\n labels_probabilities})])\n\n # Evaluate results\n if (not no_evaluation) & (len(processed_corpus) > 0):\n eval_loss = eval_loss / nb_eval_steps if nb_eval_steps else 0\n results = self.evaluate_and_display_results(eval_loss, processed_corpus[\"labels\"].tolist(),\n processed_corpus[\"predicted_labels\"].tolist(),\n labels, no_saving, save_folder)\n else:\n results = None\n\n return results, processed_corpus.reset_index(drop=True)\n\n def run_training(self, train_dataset, test_dataset, model, tokenizer, labels, labels_weights):\n \"\"\"\n Train a transformer model.\n :param train_dataset:\n :param test_dataset:\n :param model:\n :param tokenizer:\n :param labels:\n :param labels_weights:\n :return:\n \"\"\"\n\n train_batch_size = self.per_gpu_batch_size\n train_sampler = RandomSampler(train_dataset)\n train_dataloader = DataLoader(train_dataset, sampler=train_sampler, batch_size=train_batch_size)\n\n t_total = len(train_dataloader) // self.gradient_accumulation_steps * self.num_train_epochs\n\n # Initializing optimizer\n optimizer, scheduler = self.set_scheduler_and_optimizer(model, t_total)\n\n global_step = 0\n tr_loss, logging_loss = 0.0, 0.0\n model.zero_grad()\n model.to(self.device)\n\n for step in range(int(self.num_train_epochs)):\n self.logger.info(f\"############ EPOCH : {step + 1} / {self.num_train_epochs} ############\\n\")\n epoch_iterator = tqdm(train_dataloader, desc=\"Iteration\", position=0, leave=True)\n for batch in epoch_iterator:\n # Ce n'est pas ici qu'a lieu le training, on passe simplement le modèle en mode entrainement\n model.train()\n batch = tuple(t.to(self.device) for t in batch)\n _inputs = {\"input_ids\": batch[0], \"attention_mask\": batch[1], \"labels\": batch[3],\n \"token_type_ids\": None}\n\n # Appelle la fonction forward de la classe RobertaForTokenClassification\n outputs = model(**_inputs)\n loss = self.loss_with_weights(batch[3], batch[1], outputs[1], labels_weights)\n loss.backward()\n tr_loss += loss.item()\n\n if (step + 1) % self.gradient_accumulation_steps == 0:\n torch.nn.utils.clip_grad_norm_(model.parameters(), self.max_grad_norm)\n optimizer.step()\n scheduler.step() # Update learning rate schedule\n model.zero_grad()\n global_step += 1\n\n if 0 < self.max_steps < global_step:\n epoch_iterator.close()\n break\n # On évalue les performances du modèle à la fin de chaque epoch\n self.run_predict_and_eval(test_dataset, model, tokenizer, labels, None, no_saving=True)\n\n # Sauvegarge du modèle final et suppression des checkpoints\n save_path = os.path.join(self.output_dir, f\"{model.__class__.__name__}.pt\")\n self.save_model(save_path, model)\n\n return save_path\n\n @staticmethod\n def save_model(save_path, model):\n\n torch.save(model, save_path)\n\n def save_config_and_logs(self):\n\n # Export du fichier log et json\n self.output_dir = os.path.join(self.inputs[\"output_dir\"], f\"{datetime.now().strftime('%m_%d_%Y_%H%M%S')}\")\n os.mkdir(self.output_dir)\n _log_file = os.path.join(self.output_dir, \"log.txt\")\n logging.basicConfig(filename=_log_file, level=self.log_level,\n format='%(asctime)s %(name)s %(levelname)s:%(message)s')\n _json_file = os.path.join(self.output_dir, \"config.json\")\n with open(_json_file, \"w\") as json_file:\n json.dump(self.inputs, json_file)\n\n def set_scheduler_and_optimizer(self, model, t_total):\n\n # Linear warmup and decay\n no_decay = [\"bias\", \"LayerNorm.weight\"]\n optimizer_grouped_parameters = [\n {\"params\": [p for n, p in model.named_parameters() if not any(nd in n for nd in no_decay)],\n \"weight_decay\": self.weight_decay},\n {\"params\": [p for n, p in model.named_parameters() if any(nd in n for nd in no_decay)], \"weight_decay\": 0.0}\n ]\n optimizer = torch.optim.AdamW(optimizer_grouped_parameters, lr=self.learning_rate, eps=self.adam_epsilon)\n scheduler = get_linear_schedule_with_warmup(optimizer, num_warmup_steps=self.warmup_steps,\n num_training_steps=t_total)\n\n return optimizer, scheduler\n\n @staticmethod\n def set_seed(seed_num):\n\n seed(seed_num)\n np.random.seed(seed_num)\n torch.manual_seed(seed_num)\n\n def train_model_on_corpus(self, train_corpus, test_corpus):\n \"\"\"\n Entrainement d'un modèle de reconnaissance d'entités nommées.\n :param train_corpus: DataFrame du corpus de textes d'entrainement\n :param test_corpus: DataFrame du corpus de textes de test\n \"\"\"\n\n self.parse_input_json()\n\n # Loading labels\n labels, labels_weight = self.load_labels(train_corpus)\n # Loading model and tokenizer\n model, tokenizer = self.load_model_and_tokenizer(labels)\n # Loading training and eval datasets\n train_dataset = self.load_and_cache_texts(train_corpus, tokenizer, labels)\n test_dataset = self.load_and_cache_texts(test_corpus, tokenizer, labels)\n\n # Save config and logs\n self.save_config_and_logs()\n\n # Train model\n self.run_training(train_dataset, test_dataset, model, tokenizer, labels, labels_weight)\n\n # Show examples\n _, processed_corpus = \\\n self.run_predict_and_eval(test_dataset, model, tokenizer, labels, None, no_evaluation=True)\n show_legend(self.list_entities)\n show_annotations(processed_corpus, self.list_entities, self.white_space_token)\n\n\nclass Pseudo:\n\n def __init__(self, _names_path, _address_path, _car_path, societies_path, labels_column, labels_format,\n white_space_token):\n\n self.names_path = _names_path\n self.address_path = _address_path\n self.car_path = _car_path\n self.societies_path = societies_path\n self.labels_col = labels_column\n self.labels_format = labels_format\n self.fake = Faker('fr_FR')\n self.white_space_token = white_space_token\n Faker.seed()\n\n self.address, self.names, self.zip, self.cars, self.societies, self.train_df, self.dev_df, self.test_df = \\\n [None] * 8\n\n def chain_of_replacements_other_entity(self, corpus, list_entities):\n \"\"\"\n Remplace toutes les entités de la liste donnée par d'autres entités factices du meme type dans le corpus.\n :param corpus: DataFrame du corpus de textes\n :param list_entities: liste des entités à remplacer\n return: corpus avec entités remplacées\n \"\"\"\n\n self.address = pd.read_csv(self.address_path)\n self.names = pd.read_csv(self.names_path)\n self.zip = self.address['postcode'].unique().tolist()\n self.cars = pd.read_csv(self.car_path)\n self.societies = pd.read_csv(self.societies_path)\n\n if \"REF_NUM\" in list_entities:\n print(\"REF_NUM - \", end='')\n corpus = corpus.apply(lambda x: self.replace_refnum(x), axis=1)\n if \"LOC\" in list_entities:\n print(\"LOC - \", end='')\n corpus = corpus.apply(lambda x: self.replace_loc(x), axis=1)\n if \"PERSON\" in list_entities:\n print(\"PERSON - \", end='')\n corpus = corpus.apply(lambda x: self.replace_person(x), axis=1)\n if \"ORGANIZATION\" in list_entities:\n print(\"ORGANIZATION - \", end='')\n corpus = corpus.apply(lambda x: self.replace_organization(x), axis=1)\n if \"WEBSITE\" in list_entities:\n print(\"WEBSITE - \", end='')\n corpus = corpus.apply(lambda x: self.replace_website(x), axis=1)\n if \"ADDRESS\" in list_entities:\n print(\"ADDRESS - \", end='')\n corpus = corpus.apply(lambda x: self.replace_address_zip_gpe(x), axis=1)\n if \"EMAIL\" in list_entities:\n print(\"EMAIL - \", end='')\n corpus = corpus.apply(lambda x: self.replace_email(x), axis=1)\n if \"PHONE\" in list_entities:\n print(\"PHONE - \", end='')\n corpus = corpus.apply(lambda x: self.replace_phone(x), axis=1)\n if \"IMMAT\" in list_entities:\n print(\"IMMAT - \", end='')\n corpus = corpus.apply(lambda x: self.replace_immat(x), axis=1)\n if \"MONEY\" in list_entities:\n print(\"MONEY - \", end='')\n corpus = corpus.apply(lambda x: self.replace_money(x), axis=1)\n if \"DATE\" in list_entities:\n print(\"DATE - \", end='')\n corpus = corpus.apply(lambda x: self.replace_date(x), axis=1)\n if \"TIME\" in list_entities:\n print(\"TIME - \", end='')\n corpus = corpus.apply(lambda x: self.replace_time(x), axis=1)\n if \"CAR\" in list_entities:\n print(\"CAR\", end='')\n corpus = corpus.apply(lambda x: self.replace_car(x), axis=1)\n print('\\n')\n\n return corpus\n\n def chain_of_replacements_tags(self, corpus, list_entities):\n \"\"\"\n Remplace toutes les entités de la liste donnée par un tag au format ''.\n :param corpus: DataFrame du corpus de textes\n :param list_entities: liste des entités à remplacer\n return: corpus avec entités remplacées\n \"\"\"\n\n for entity in list_entities:\n corpus = corpus.apply(lambda x: self.replace_by_tag(x, entity), axis=1)\n\n return corpus\n\n def concat_entities(self, found_list):\n \"\"\"\n Concaténation des entités ayant été partagées en plusieurs tokens successifs en une seule et meme entrée.\n :param found_list: liste des entités, chaque élément étant de type (index, label, label)\n :return: liste des entités concaténées\n \"\"\"\n\n clean_list = []\n if found_list:\n full_entity = found_list[0][1]\n for i in range(1, len(found_list)):\n if self.labels_format == \"BIO\":\n _is_same_entity = (found_list[i][0] == found_list[i - 1][0] + 1) & (found_list[i][2] == 'I')\n else:\n _is_same_entity = (found_list[i][0] == found_list[i - 1][0] + 1)\n if _is_same_entity:\n full_entity += ' ' + found_list[i][1]\n else:\n clean_list.append(full_entity)\n full_entity = found_list[i][1]\n clean_list.append(full_entity)\n else:\n clean_list = []\n\n return clean_list\n\n def create_csv_name(self, _name):\n \"\"\"\n Création d'un nom propre factice à partir d'un csv.\n :param _name: nom à remplacer\n :return: nom généré\n \"\"\"\n\n prefixes_h = ['', 'M ', 'M. ', 'Mr ', 'Mr. ', 'Monsieur ']\n weights_h = [0.1, 0.8, 0.4, 0.2, 0.1, 0.5]\n weights_h = [x / sum(weights_h) for x in weights_h]\n\n prefixes_f = ['', 'Mme ', 'Melle ', 'Madame ']\n weights_f = [0.1, 0.8, 0.2, 0.3]\n weights_f = [x / sum(weights_f) for x in weights_f]\n\n if any([x.strip() in _name.split(\" \") for x in prefixes_f]):\n _prefix = np.random.choice(prefixes_f, p=weights_f)\n elif any([x.strip() in _name.split(\" \") for x in prefixes_h]):\n _prefix = np.random.choice(prefixes_h, p=weights_h)\n else:\n _prefix = \"\"\n\n _firstname = np.random.choice(self.names[\"prenoms\"].tolist(), p=self.names[\"prenoms_poids\"].tolist())\n _lastname = np.random.choice(self.names[\"noms\"].tolist(), p=self.names[\"noms_poids\"].tolist()).capitalize()\n\n if len(_prefix) > 0:\n if random() < 0.5:\n replacement_name = f\"{_prefix} {_firstname} {_lastname}\"\n else:\n replacement_name = f\"{_prefix} {_lastname}\"\n else:\n if random() < 0.7:\n replacement_name = f\"{_firstname} {_lastname}\"\n else:\n replacement_name = f\"{_firstname}\"\n\n return replacement_name\n\n def create_faker_name(self, _name):\n \"\"\"\n Création d'un nom propre factice à partir par la librairie Faker.\n :param _name: nom à remplacer\n :return: nom généré\n \"\"\"\n\n prefixes_h = ['', 'M ', 'M. ', 'Mr ', 'Mr. ', 'Monsieur ']\n weights_h = [0.1, 0.8, 0.4, 0.2, 0.1, 0.5]\n weights_h = [x / sum(weights_h) for x in weights_h]\n male_name_generator = [self.fake.name_male, self.fake.last_name_male]\n prefixes_f = ['', 'Mme ', 'Melle ', 'Madame ']\n weights_f = [0.1, 0.8, 0.2, 0.3]\n weights_f = [x / sum(weights_f) for x in weights_f]\n female_name_generator = [self.fake.name_female, self.fake.last_name_female]\n\n _need_prefix = any([x.strip() in _name.split(\" \") for x in prefixes_f + prefixes_h])\n if random() < 0.5:\n # male name\n if _need_prefix:\n _prefix = np.random.choice(prefixes_h, p=weights_h)\n _fake_name = np.random.choice(male_name_generator, p=[0.6, 0.4])()\n else:\n _prefix = ''\n _fake_name = np.random.choice(male_name_generator, p=[1.0, 0.0])()\n replacement_name = _prefix + _fake_name\n else:\n # female name\n if _need_prefix:\n _prefix = np.random.choice(prefixes_f, p=weights_f)\n _fake_name = np.random.choice(female_name_generator, p=[0.6, 0.4])()\n else:\n _prefix = ''\n _fake_name = np.random.choice(female_name_generator, p=[1.0, 0.0])()\n replacement_name = _prefix + _fake_name\n\n return replacement_name\n\n @staticmethod\n def match_case(new_entity, old_entity):\n \"\"\"\n Fait correspondre le style de l'entité de remplacement avec celui de l'entité originale afin de respecter la\n casse.\n :param new_entity: entité de remplacement\n :param old_entity: entité originale\n :return: entité de remplacement formatée\n \"\"\"\n entity_letters = [x for x in old_entity if x not in ' ' + string.digits]\n if len(entity_letters) == 0:\n return new_entity\n # Make the fake name lower or upper depending on the text style\n if sum(1 for c in entity_letters if c.isupper()) / len(entity_letters) > 0.8:\n new_entity = new_entity.upper()\n if sum(1 for c in entity_letters if c.islower()) / len(entity_letters) > 0.999:\n new_entity = new_entity.lower()\n return new_entity\n\n @staticmethod\n def match_ponctuation(new_entity, old_entity):\n \"\"\"\n Ajoute un point ou une virgule à la fin de l'entité s'il y en avait un à l'origine.\n :param new_entity: entité de remplacement\n :param old_entity: entité originale\n :return: entité de remplacement formatée\n \"\"\"\n\n # looking for possible endings to reproduce\n endings = ['.
', ',
', ',', '.']\n for _ending in endings:\n if old_entity[-len(_ending):] == _ending:\n new_entity += _ending\n return new_entity\n\n @staticmethod\n def oversample_corpus(corpus, test_corpus, generated_ppel, list_ppel, distinct_ppel=0, method=\"first\",\n entities=None, ratio=0.5):\n \"\"\"\n Loads all ppel .txt and .ann files and stores information into a dataframe.\n :param corpus:\n :param test_corpus:\n :param generated_ppel: number of artificial ppel to generate\n :param list_ppel:\n :param distinct_ppel: number of ppel used for generating new ones.\n :param method:\n :param entities:\n :param ratio:\n \"\"\"\n if entities is None:\n entities = []\n _error_message = \"The parameter 'method' must be equal to 'first' (default value), 'random' or 'balanced'\"\n assert method in [\"first\", \"random\", \"balanced\"], _error_message\n if not list_ppel:\n if distinct_ppel == 0:\n corpus = corpus.sample(generated_ppel, replace=True)\n else:\n assert distinct_ppel <= len(\n corpus), \"The number of texts must be smaller or equal to the corpus length\"\n if method == \"first\":\n test_corpus = corpus.iloc[distinct_ppel:].copy()\n # corpus = corpus.iloc[:distinct_ppel].sample(generated_ppel, replace=True)\n corpus = corpus.iloc[:distinct_ppel]\n corpus = corpus[corpus[\"label\"].apply(lambda x: \"PERSON\" in \" \".join(x))].sample(\n generated_ppel, replace=True)\n\n elif method == \"random\":\n test_corpus = corpus.sample(len(corpus) - distinct_ppel)\n corpus = corpus[~corpus[\"file\"].isin(test_corpus[\"file\"].unique())].\\\n sample(generated_ppel, replace=True)\n elif method == \"balanced\":\n assert 0 < ratio < 1, \"The ratio must be in ]0, 1[.\"\n assert len(entities) > 0, \"If the method 'balanced' is selected, you must define the entities.\"\n # test_corpus = corpus.sample(len(corpus) - distinct_ppel)\n # _train_corpus = corpus[~corpus[\"file\"].isin(test_corpus[\"file\"].unique())]\n _train_corpus = corpus.drop_duplicates([\"file\"])\n _train_positive = _train_corpus[\n _train_corpus[\"label\"].apply(lambda x: any([y.split(\"-\")[-1] in entities for y in x]))]\n _train_negative = _train_corpus[\n _train_corpus[\"label\"].apply(lambda x: not any([y.split(\"-\")[-1] in entities for y in x]))]\n n_pos = int(ratio * generated_ppel)\n corpus = pd.concat([_train_positive.sample(n_pos, replace=True),\n _train_negative.sample(generated_ppel - n_pos, replace=True)], axis=0).sample(\n frac=1)\n else:\n corpus = corpus[corpus.file in list_ppel].sample(generated_ppel, replace=True)\n\n return corpus, test_corpus\n\n def replace_by_tag(self, x, entity):\n \"\"\"\n Remplace une entité détectée par un tag au format ''.\n :param x: ligne du DataFrame à traiter\n :param entity: entité à remplacer\n :return: ligne modifiée\n \"\"\"\n\n x[\"text_anonyme\"] = x[\"text\"].copy()\n entities = [(ind, x[\"text\"][ind], entity.split('-')[0]) for ind, ent in enumerate(x[\"labels\"]) if\n ent.split('-')[-1] == entity]\n clean_entities = self.concat_entities(entities)\n replacement_tags = {y: f\"<{entity}_{i + 1}>\" for i, y in enumerate(list(dict.fromkeys(clean_entities)))}\n\n for _ent in clean_entities:\n x[\"text_anonyme\"], x[\"labels\"] = switch_entity(x, _ent, replacement_tags[_ent], entity, split_first=False)\n\n return x\n\n def replace_person(self, x, method=\"dataset\"):\n \"\"\"\n Remplace les noms propres d'un texte par d'autres noms factices.\n :param x: ligne du DataFrame à traiter\n :param method: méthode de remplacement (dataset ou Faker)\n :return: ligne modifiée\n \"\"\"\n person_names = [(ind, x.text[ind], entity.split('-')[0]) for ind, entity in enumerate(x[self.labels_col]) if\n entity.split('-')[-1] == \"PERSON\"]\n clean_names = self.concat_entities(person_names)\n\n pers_dic = {}\n for _name in clean_names:\n replacement_name = None\n # If we already met this entity, we use the same substitution again\n if _name in pers_dic.keys():\n replacement_name = pers_dic[_name]\n # If this is the first time we meet this entity, we create a substitution\n else:\n # Creating a fake replacement name\n if method == \"dataset\":\n replacement_name = self.create_csv_name(_name)\n if method == \"Faker\":\n replacement_name = self.create_faker_name(_name)\n replacement_name = self.match_case(replacement_name, _name)\n pers_dic[_name] = replacement_name\n replacement_name = self.match_ponctuation(replacement_name, _name)\n x.text, x[self.labels_col] = self.switch_entity(x, _name, replacement_name, 'PERSON')\n\n return x\n\n def replace_organization(self, x):\n \"\"\"\n Remplace les sociétés d'un texte par d'autres sociétés factices.\n :param x: ligne du DataFrame à traiter\n :return: ligne modifiée\n \"\"\"\n gafam = ['Google', 'Amazon', 'Facebook', 'Apple', 'Microsoft']\n banks = ['BNP PARIBAS', 'Boursorama', 'Crédit agricole', \"Caisse d'épargne\", \"Société générale\",\n 'Crédit Mutuel', 'Banque populaire', 'Banque postale']\n malls = ['Agora', 'Atac', 'Auchan', 'Carrefour', 'Carrefour Market', 'Casino', 'Coop', 'Cora', 'Costco',\n 'Douka Be',\n 'E. Leclerc', 'Entrepot Produits Frais', 'Intermarche', 'Leader Price', 'Leclerc', 'Simply Market',\n 'Spar', 'Super U',\n 'Supermarche Match']\n security = ['Gendarmerie', 'Commissariat', 'Police municipale', 'Police']\n internet = ['OVH', 'Gandhi', '1&1', 'GoDaddy']\n telecom = ['Orange', 'SFR', 'Bouygues Telecom', 'Free', 'Sosh']\n prefixes = ['société', 'sté', 'SARL', 'sci']\n\n all_lists = [gafam, banks, malls, security, internet, telecom]\n found_org = [(ind, x.text[ind], entity.split('-')[0]) for ind, entity in enumerate(x[self.labels_col]) if\n entity.split('-')[-1] == \"ORGANIZATION\"]\n clean_org = self.concat_entities(found_org)\n\n org_dic = {}\n if len(clean_org) > 0:\n replacement_org = None\n for org in clean_org:\n # If we already met this entity, we use the same substitution again\n if org in org_dic.keys():\n replacement_org = org_dic[org]\n # If this is the first time we meet this entity, we create a substitution\n else:\n # Remplacement du préfixe s'il existe\n _prefix = [x for x in prefixes if x in org.split(\" \")]\n if len(_prefix) > 0:\n _prefix = choice(prefixes)\n else:\n _prefix = \"\"\n # Remplacement du nom de la société\n _found_something = False\n # S'il s'agit d'une société fréquemment citée, on la remplace par une autre du meme type par soucis\n # de cohérence\n for list_org in all_lists:\n _l_splitted = [[unidecode.unidecode(y).lower() for y in x.split(' ')] for x in list_org]\n _o_splitted = [unidecode.unidecode(y).lower() for y in org.split(' ')]\n known_pattern = set(\n list_org[ind] for ind, x in enumerate(_l_splitted) if\n find_sub_list(x, _o_splitted) is not None)\n if len(known_pattern) > 0:\n _found_something = True\n org_dic[list(known_pattern)[0]] = choice(list_org)\n replacement_org = org_dic[list(known_pattern)[0]]\n break\n # S'il s'agit d'une société moins connue, on la remplace par une société sélectionnée aléatoirement\n # dans la base Infogreffe\n if not _found_something:\n replacement_org = self.societies.sample(1).iloc[0][\"Dénomination\"].strip().capitalize()\n replacement_org = _prefix + \" \" + replacement_org\n replacement_org = self.match_case(replacement_org, org).strip()\n org_dic[org] = replacement_org\n replacement_org = self.match_ponctuation(replacement_org, org)\n # Modification du texte\n x.text, x[self.labels_col] = self.switch_entity(x, org, replacement_org, 'ORGANIZATION')\n return x\n\n def replace_loc(self, x):\n \"\"\"\n Remplace les emplacements d'un texte par d'autres emplacements factices.\n :param x: ligne du DataFrame à traiter\n :return: ligne modifiée\n \"\"\"\n\n _places = [\"de la place\", \"du supermarché\", \"du centre commercial\", \"de l'hopital\", \"du cinéma\", \"de la mairie\",\n \"du centre\", \"du marché\", \"de la bilbiothèque\", \"du magasin\", \"du parc\"]\n found_loc = [(ind, x.text[ind], entity.split('-')[0]) for ind, entity in enumerate(x[self.labels_col]) if\n entity.split('-')[-1] == \"LOC\"]\n clean_loc = self.concat_entities(found_loc)\n\n for _loc in clean_loc:\n if 'parking' in unidecode.unidecode(_loc).lower():\n _city = self.address['city'].sample(1).iloc[0]\n _place = choice(_places)\n replacement_loc = choice([f\"parking {_place} de {_city}\", f\"parking de {_city}\"])\n elif 'gare' in unidecode.unidecode(_loc).lower():\n _city = self.address['city'].sample(1).iloc[0]\n replacement_loc = f\"gare de {_city}\"\n elif 'parc' in unidecode.unidecode(_loc).lower():\n _city = self.address['city'].sample(1).iloc[0]\n replacement_loc = f\"parc de {_city}\"\n elif 'gendarmerie' in unidecode.unidecode(_loc).lower():\n _city = self.address['city'].sample(1).iloc[0]\n replacement_loc = f\"gendarmerie de {_city}\"\n elif 'commissariat' in unidecode.unidecode(_loc).lower():\n _city = self.address['city'].sample(1).iloc[0]\n replacement_loc = f\"commissariat de {_city}\"\n elif ('rond-point' in unidecode.unidecode(_loc).lower()) or (\n 'rond point' in unidecode.unidecode(_loc).lower()):\n _city = self.address['city'].sample(1).iloc[0]\n _place = choice(_places)\n replacement_loc = choice([f\"rond-point {_place} de {_city}\", f\"rond-point de {_city}\",\n f\"rond point {_place} de {_city}\", f\"rond point de {_city}\"])\n elif ('station service' in unidecode.unidecode(_loc).lower()) or (\n 'station essence' in unidecode.unidecode(_loc).lower()):\n _city = self.address['city'].sample(1).iloc[0]\n _place = choice(_places)\n replacement_loc = choice([f\"station service {_place} de {_city}\", f\"station service de {_city}\",\n f\"station essence {_place} de {_city}\", f\"station essence de {_city}\"])\n elif 'restaurant' in unidecode.unidecode(_loc).lower():\n _nom = choice([self.fake.last_name_male, self.fake.last_name_female])()\n replacement_loc = choice([f\"restaurant Chez {_nom}\", f\"restaurant le {self.fake.word()}\"])\n elif 'hotel' in unidecode.unidecode(_loc).lower():\n replacement_loc = f\"restaurant le {self.fake.word()}\"\n elif 'residence' in unidecode.unidecode(_loc).lower():\n _nom = choice([self.fake.last_name_male, self.fake.last_name_female])()\n replacement_loc = choice([f\"résidence Le {_nom}\"])\n else:\n replacement_loc = _loc\n\n replacement_loc = self.match_case(replacement_loc, _loc)\n replacement_loc = self.match_ponctuation(replacement_loc, _loc)\n x.text, x[self.labels_col] = self.switch_entity(x, _loc, replacement_loc, 'LOC')\n\n return x\n\n def replace_date(self, x):\n \"\"\"\n Remplace les dates d'un texte par d'autres dates factices.\n :param x: ligne du DataFrame à traiter\n :return: ligne modifiée\n \"\"\"\n\n locale.setlocale(locale.LC_ALL, 'fr_FR.utf-8')\n # date_formats_with_days = [\"%d/%m/%Y\", \"%d/%m/%y\", \"%d %b %Y\", \"%d-%m-%Y\", \"%d-%m-%y\", \"%d.%m.%Y\", \"%d.%m.%y\",\n # \"%d %b\", \"%A %d/%m/%Y\", \"%A %d/%m/%y\", \"%A %d %b %Y\", \"%A %d %b\", \"%-d/%-m/%Y\",\n # \"%-d/%-m/%y\", \"%-d %b %Y\", \"%-d-%-m-%Y\", \"%-d-%-m-%y\", \"%-d.%-m.%Y\", \"%-d.%-m.%y\",\n # \"%-d %b\", \"%A %-d/%-m/%Y\", \"%A %-d/%-m/%y\", \"%A %-d %b %Y\", \"%A %-d %b\"]\n date_formats_without_days = [\"%d/%m/%Y\", \"%d/%m/%y\", \"%d %b %Y\", \"%d-%m-%Y\", \"%d-%m-%y\", \"%d.%m.%Y\", \"%d.%m.%y\",\n \"%d %b\", \"%-d/%-m/%Y\", \"%-d/%-m/%y\", \"%-d %b %Y\", \"%-d-%-m-%Y\", \"%-d-%-m-%y\",\n \"%-d.%-m.%Y\", \"%-d.%-m.%y\", \"%-d %b\"]\n weights = [1, 1, 0.8, 0.3, 0.3, 0.1, 0.1, 0.2, 1, 1, 0.8, 0.3, 0.3, 0.1, 0.1, 0.2]\n weights = [x / sum(weights) for x in weights]\n month_list = ['janvier', 'fevrier', 'mars', 'avril', 'mai', 'juin', 'juillet', 'aout', 'septembre', 'octobre',\n 'novembre', 'decembre']\n\n found_dates = [(ind, x.text[ind], entity.split('-')[0]) for ind, entity in enumerate(x[self.labels_col]) if\n entity.split('-')[-1] == \"DATE\"]\n clean_dates = self.concat_entities(found_dates)\n\n if len(found_dates) > 0:\n for _date in clean_dates:\n # month only\n if unidecode.unidecode(_date.split(' ')[0]) in month_list:\n replacement_date = datetime.now() - timedelta(days=randint(0, 365))\n replacement_date = replacement_date.strftime(\"%b %Y\")\n replacement_date = self.match_case(replacement_date, _date)\n replacement_date = self.match_ponctuation(replacement_date, _date)\n else:\n replacement_date = datetime.now() - timedelta(days=randint(0, 365))\n replacement_date = replacement_date.strftime(np.random.choice(date_formats_without_days, p=weights))\n replacement_date = self.match_case(replacement_date, _date)\n replacement_date = self.match_ponctuation(replacement_date, _date)\n x.text, x[self.labels_col] = self.switch_entity(x, _date, replacement_date, 'DATE')\n\n return x\n\n def replace_time(self, x):\n \"\"\"\n Remplace les heures d'un texte par d'autres heures factices.\n :param x: ligne du DataFrame à traiter\n :return: ligne modifiée\n \"\"\"\n\n locale.setlocale(locale.LC_ALL, 'fr_FR.utf-8')\n time_formats = [\"%Hh%M\", \"%HH%M\", \"%-Hh%-M\", \"%-HH%-M\", \"%H:%M\", \"%H heures %M\", \"%-H heures %-M\", \"%H h %M\",\n \"%H H %M\",\n \"%-H h %-M\", \"%-H H %-M\"]\n weights = [1, 0.1, 0.1, 0.05, 0.05, 0.3, 0.05, 0.02, 0.02, 0.02, 0.02]\n weights = [x / sum(weights) for x in weights]\n\n found_times = [(ind, x.text[ind], entity.split('-')[0]) for ind, entity in enumerate(x[self.labels_col]) if\n entity.split('-')[-1] == \"TIME\"]\n clean_times = self.concat_entities(found_times)\n\n if len(clean_times) > 0:\n former_text = ' '.join(x.text)\n free_format = [True for _ in clean_times]\n if len(clean_times) > 1:\n for i in range(len(clean_times) - 1):\n _link = former_text[former_text.index(clean_times[i]) + len(clean_times[i]):former_text.index(\n clean_times[i + 1])]\n _link = unidecode.unidecode(_link).strip()\n if (_link == 'et') or (_link == 'a'):\n free_format[i + 1] = False\n _previous_time, _previous_format = [None] * 2\n for ind, _time in enumerate(clean_times):\n if free_format[ind]:\n replacement_time = datetime.now().replace(hour=randint(0, 23), minute=randint(0, 59))\n _previous_time = replacement_time\n _previous_format = np.random.choice(time_formats, p=weights)\n else:\n replacement_time = _previous_time + timedelta(minutes=randint(0, 180))\n replacement_time = replacement_time.strftime(_previous_format)\n replacement_time = self.match_case(replacement_time, _time)\n replacement_time = self.match_ponctuation(replacement_time, _time)\n x.text, x[self.labels_col] = self.switch_entity(x, _time, replacement_time, 'TIME')\n\n return x\n\n def replace_address_zip_gpe(self, x):\n \"\"\"\n Remplace les adresses, code postaux et entités géopolitiques d'un texte par d'autres factices.\n :param x: ligne du DataFrame à traiter\n :return: ligne modifiée\n \"\"\"\n\n found_zip = [(ind, x.text[ind], entity.split('-')[0]) for ind, entity in enumerate(x[self.labels_col]) if\n entity.split('-')[-1] == \"ZIP\"]\n found_addresses = [(ind, x.text[ind], entity.split('-')[0]) for ind, entity in enumerate(x[self.labels_col]) if\n entity.split('-')[-1] == \"ADDRESS\"]\n found_gpe = [(ind, x.text[ind], entity.split('-')[0]) for ind, entity in enumerate(x[self.labels_col]) if\n entity.split('-')[-1] == \"GPE\"]\n\n # If there is at least one person name, replace a random one. Else return None.\n clean_addresses = self.concat_entities(found_addresses)\n clean_zip = self.concat_entities(found_zip)\n clean_zip = [str(x) for x in clean_zip]\n clean_gpe = self.concat_entities(found_gpe)\n\n for _address in clean_addresses:\n with_num = sum(c.isdigit() for c in _address)\n if with_num > 0:\n _random_place = self.address.sample(1).iloc[0]\n replacement_address = \"{} {}\".format(randint(1, 60), _random_place.street)\n else:\n ind = min(find_sub_list(_address.split(' '), x.text)) - 1\n word_before = x.text[ind] if ind > 0 else None\n if word_before == 'la':\n _compat_types = ['Résidence', 'Rue', 'Cite', 'Place', 'Route', 'Residence', 'Ruelle', 'Traverse',\n 'Ferme', 'Voie']\n compat_df = self.address[self.address['street_type'].isin(_compat_types)]\n elif word_before == 'le':\n _compat_types = ['Lieu', 'Chemin', 'Lotissement', 'Square', 'Sentier', 'Domaine', 'Hameau', 'Clos',\n 'Quai', 'Chez', 'Passage', 'Boulevard', 'Lieu-dit', 'Cour']\n compat_df = self.address[self.address['street_type'].isin(_compat_types)]\n elif word_before == 'l':\n _compat_types = ['Allée', 'Impasse', 'Avenue', 'Allee']\n compat_df = self.address[self.address['street_type'].isin(_compat_types)]\n else:\n compat_df = self.address\n\n _random_place = compat_df.sample(1).iloc[0]\n replacement_address = \"{}\".format(_random_place.street)\n\n replacement_zip = str(_random_place.postcode)\n replacement_gpe = _random_place.city\n replacement_address = self.match_case(replacement_address, _address)\n replacement_address = self.match_ponctuation(replacement_address, _address)\n x.text, x[self.labels_col] = self.switch_entity(x, _address, replacement_address, 'ADDRESS')\n\n if len(clean_zip):\n x.text, x[self.labels_col] = self.switch_entity(x, clean_zip[0], replacement_zip, 'ZIP')\n del clean_zip[0]\n if len(clean_gpe):\n replacement_gpe = self.match_case(replacement_gpe, clean_gpe[0])\n replacement_gpe = self.match_ponctuation(replacement_gpe, clean_gpe[0])\n x.text, x[self.labels_col] = self.switch_entity(x, clean_gpe[0], replacement_gpe, 'GPE')\n del clean_gpe[0]\n\n for _zip in clean_zip:\n _random_place = self.address.sample(1).iloc[0]\n replacement_zip = str(_random_place.postcode)\n replacement_gpe = _random_place.city\n x.text, x[self.labels_col] = self.switch_entity(x, _zip, replacement_zip, 'ZIP')\n\n if len(clean_gpe):\n replacement_gpe = self.match_case(replacement_gpe, clean_gpe[0])\n replacement_gpe = self.match_ponctuation(replacement_gpe, clean_gpe[0])\n x.text, x[self.labels_col] = self.switch_entity(x, clean_gpe[0], replacement_gpe, 'GPE')\n del clean_gpe[0]\n\n gpe_dict = {}\n for _gpe in clean_gpe:\n # If we already met this entity, we use the same substitution again\n if _gpe in gpe_dict.keys():\n replacement_gpe = gpe_dict[_gpe]\n # If this is the first time we meet this entity, we create a substitution\n else:\n _random_place = self.address.sample(1).iloc[0]\n replacement_gpe = _random_place.city\n replacement_gpe = self.match_case(replacement_gpe, _gpe)\n replacement_gpe = self.match_ponctuation(replacement_gpe, _gpe)\n gpe_dict[_gpe] = replacement_gpe\n x.text, x[self.labels_col] = self.switch_entity(x, _gpe, replacement_gpe, 'GPE')\n\n return x\n\n def replace_email(self, x):\n \"\"\"\n Remplace les e-mails d'un texte par d'autres e-mails factices.\n :param x: ligne du DataFrame à traiter\n :return: ligne modifiée\n \"\"\"\n\n generators = [self.fake.ascii_company_email, self.fake.ascii_email, self.fake.ascii_free_email,\n self.fake.ascii_safe_email]\n weights = [0.4, 0.5, 1, 0.1]\n weights = [x / sum(weights) for x in weights]\n\n found_mails = [(ind, x.text[ind], entity.split('-')[0]) for ind, entity in enumerate(x[self.labels_col]) if\n entity.split('-')[-1] == \"EMAIL\"]\n clean_emails = self.concat_entities(found_mails)\n\n mail_dic = {}\n for _mail in clean_emails:\n # If we already met this entity, we use the same substitution again\n if _mail in mail_dic.keys():\n replacement_mail = mail_dic[_mail]\n # If this is the first time we meet this entity, we create a substitution\n else:\n replacement_mail = np.random.choice(generators, p=weights)()\n replacement_mail = self.match_case(replacement_mail, _mail)\n replacement_mail = self.match_ponctuation(replacement_mail, _mail)\n mail_dic[_mail] = replacement_mail\n x.text, x[self.labels_col] = self.switch_entity(x, _mail, replacement_mail, 'EMAIL')\n\n return x\n\n def replace_phone(self, x):\n \"\"\"\n Remplace les numéros de téléphone d'un texte par d'autres numéros factices.\n :param x: ligne du DataFrame à traiter\n :return: ligne modifiée\n \"\"\"\n\n found_phones = [(ind, x.text[ind], entity.split('-')[0]) for ind, entity in enumerate(x[self.labels_col]) if\n entity.split('-')[-1] == \"PHONE\"]\n clean_phones = self.concat_entities(found_phones)\n\n for _phone in clean_phones:\n replacement_phone = self.fake.phone_number()\n replacement_phone = self.match_case(replacement_phone, _phone)\n replacement_phone = self.match_ponctuation(replacement_phone, _phone)\n x.text, x[self.labels_col] = self.switch_entity(x, _phone, replacement_phone, 'PHONE')\n\n return x\n\n def replace_immat(self, x):\n \"\"\"\n Remplace les immatriculations d'un texte par d'autres immatriculations factices.\n :param x: ligne du DataFrame à traiter\n :return: ligne modifiée\n \"\"\"\n\n links = ['-', '.', '', ' ']\n weights = [1, 0.1, 0.1, 0.1]\n weights = [x / sum(weights) for x in weights]\n\n found_immat = [(ind, x.text[ind], entity.split('-')[0]) for ind, entity in enumerate(x[self.labels_col]) if\n entity.split('-')[-1] == \"IMMAT\"]\n clean_immat = self.concat_entities(found_immat)\n\n for _immat in clean_immat:\n _link = np.random.choice(links, p=weights)\n replacement_immat = [choice(string.ascii_uppercase) for _ in range(2)] + [_link] + [choice(string.digits)\n for _ in range(3)] + [\n _link] + [choice(string.ascii_uppercase) for _ in range(2)]\n replacement_immat = ''.join(replacement_immat)\n x.text, x[self.labels_col] = self.switch_entity(x, _immat, replacement_immat, 'IMMAT')\n\n return x\n\n def replace_money(self, x):\n \"\"\"\n Remplace les sommes d'argent d'un texte par d'autres sommes factices.\n :param x: ligne du DataFrame à traiter\n :return: ligne modifiée\n \"\"\"\n\n currency = [' €', '€', ' euros', 'euros', ' Euros']\n weights = [1, 0.5, 0.5, 0.01, 0.02]\n weights = [x / sum(weights) for x in weights]\n\n found_money = [(ind, x.text[ind], entity.split('-')[0]) for ind, entity in enumerate(x[self.labels_col]) if\n entity.split('-')[-1] == \"MONEY\"]\n clean_money = self.concat_entities(found_money)\n\n for _money in clean_money:\n alpha = random()\n if alpha < 0.5:\n _amount = \"{0:.2f}\".format(abs(np.random.normal(20, 100)))\n else:\n _amount = str(int(abs(np.random.normal(20, 100))))\n _currency = np.random.choice(currency, p=weights)\n replacement_money = _amount + np.random.choice(currency, p=weights)\n replacement_money = ''.join(replacement_money)\n replacement_money = self.match_ponctuation(replacement_money, _money)\n x.text, x[self.labels_col] = self.switch_entity(x, _money, replacement_money, 'MONEY')\n\n return x\n\n def replace_website(self, x):\n \"\"\"\n Remplace les sites internet d'un texte par d'autres sites factices.\n :param x: ligne du DataFrame à traiter\n :return: ligne modifiée\n \"\"\"\n\n websites = ['Amazon', 'Facebook', 'Google', 'OVH', 'Orange', 'Paypal', 'Le bon coin', \"Bon Coin\", \"Boncoin\",\n 'Leboncoin', 'Microsoft', 'Instagram', 'Booking.com', 'Airbnb']\n domain_name = r\"(\\.[a-zA-Z]{2,3}\\Z)\"\n weights = [1 for _ in websites]\n weights = [x / sum(weights) for x in weights]\n\n found_website = [(ind, x.text[ind], entity.split('-')[0]) for ind, entity in enumerate(x[self.labels_col]) if\n entity.split('-')[-1] == \"WEBSITE\"]\n clean_website = self.concat_entities(found_website)\n\n web_dict = {}\n for _website in clean_website:\n # If we already met this entity, we use the same substitution again\n if _website in web_dict.keys():\n replacement_web = web_dict[_website]\n # If this is the first time we meet this entity, we create a substitution\n else:\n if (_website.lower().startswith(\"http\")) or (_website.lower().startswith(\"http\")) or bool(\n re.search(domain_name, _website)):\n replacement_web = self.fake.url()\n if random() < 0.9:\n replacement_web = replacement_web.replace('https://', '').replace('http://', '').replace('/',\n '')\n if random() < 0.7:\n replacement_web = replacement_web.replace('www.', '')\n else:\n replacement_web = np.random.choice(websites, p=weights)\n replacement_web = self.match_case(replacement_web, _website)\n replacement_web = self.match_ponctuation(replacement_web, _website)\n web_dict[_website] = replacement_web\n x.text, x[self.labels_col] = self.switch_entity(x, _website, replacement_web, 'WEBSITE')\n\n return x\n\n def replace_car(self, x):\n \"\"\"\n Remplace les véhicules d'un texte par d'autres véhicules factices.\n :param x: ligne du DataFrame à traiter\n :return: ligne modifiée\n \"\"\"\n\n colors_f = ['grise', 'blanche', 'noire', 'rouge', 'bleue', 'verte', 'beige']\n colors_h = ['gris', 'blanc', 'noir', 'rouge', 'bleu', 'vert', 'beige']\n\n found_car = [(ind, x.text[ind], entity.split('-')[0]) for ind, entity in enumerate(x[self.labels_col]) if\n entity.split('-')[-1] == \"CAR\"]\n clean_car = self.concat_entities(found_car)\n\n for _car in clean_car:\n if np.any([x in unidecode.unidecode(_car).lower() for x in ['voiture', 'vehicule']]):\n if 'vehicule' in unidecode.unidecode(_car).lower():\n all_cars = self.cars[self.cars['genre'] == 'H']\n random_car = all_cars.sample(1).iloc[0]\n _rcm = random_car['marque']\n _rct = random_car['modele']\n colf = choice(colors_f)\n colh = choice(colors_h)\n formats = [f\"véhicule {_rcm} {_rct}\", f\"véhicule {_rcm} {_rct} {colh}\",\n f\"véhicule {_rcm} {_rct} de couleur {colf}\",\n f\"véhicule ({_rcm} {_rct})\", f\"{_rcm} {_rct}\", f\"{_rct}\"]\n else:\n all_cars = self.cars[self.cars['genre'] == 'F']\n random_car = all_cars.sample(1).iloc[0]\n _rcm = random_car['marque']\n _rct = random_car['modele']\n colf = choice(colors_f)\n formats = [f\"voiture {_rcm} {_rct} {colf}\", f\"voiture {_rcm} {_rct}\", f\"{_rcm} {_rct}\", f\"{_rct}\"]\n else:\n ind = None\n try:\n ind = find_sub_list(_car.split(' '), x.text)[0] - 1\n except (Exception, ):\n print(f\"entity of type {_car} not found\")\n if x.text[ind] in ['un', 'mon', 'son', 'le']:\n all_cars = self.cars[self.cars['genre'] == 'H']\n random_car = all_cars.sample(1).iloc[0]\n _rcm = random_car['marque']\n _rct = random_car['modele']\n colf = choice(colors_f)\n colh = choice(colors_h)\n formats = [f\"véhicule {_rcm} {_rct}\", f\"véhicule {_rcm} {_rct} {colh}\",\n f\"véhicule {_rcm} {_rct} de couleur {colf}\",\n f\"véhicule ({_rcm} {_rct})\", f\"{_rcm} {_rct}\", f\"{_rct}\"]\n else:\n all_cars = self.cars[self.cars['genre'] == 'F']\n random_car = all_cars.sample(1).iloc[0]\n _rcm = random_car['marque']\n _rct = random_car['modele']\n colf = choice(colors_f)\n formats = [f\"voiture {_rcm} {_rct} {colf}\", f\"voiture {_rcm} {_rct}\", f\"{_rcm} {_rct}\", f\"{_rct}\"]\n\n replacement_car = choice(formats)\n replacement_car = self.match_case(replacement_car, _car)\n replacement_car = self.match_ponctuation(replacement_car, _car)\n x.text, x[self.labels_col] = self.switch_entity(x, _car, replacement_car, 'CAR')\n\n return x\n\n def replace_refnum(self, x):\n \"\"\"\n Remplace les numéros de référence d'un texte par d'autres numéros factices.\n :param x: ligne du DataFrame à traiter\n :return: ligne modifiée\n \"\"\"\n\n categories = ['rapport', 'contravention', 'constatation', 'commande', 'facture', 'courante', 'numero', 'iban',\n 'reference', 'bancaire', 'verbal', 'adresse', 'contrat', 'suivi', 'carte', 'cheque', 'code',\n 'colis', 'compte', 'dossier', 'puce', 'siret', 'siren', 'proces', 'ipv6', 'avis',\n 'produit', 'serie', 'bic', 'imei', 'ip', 'main', 'n°', 'ref', 'de', 'vol', ':']\n found_ref = [(ind, x.text[ind], entity.split('-')[0]) for ind, entity in enumerate(x[self.labels_col]) if\n entity.split('-')[-1] == \"REF_NUM\"]\n clean_ref = self.concat_entities(found_ref)\n\n for _ref in clean_ref:\n ref_print = _ref\n clean_ref = _ref\n for cat in categories:\n ind = '°'.join([unidecode.unidecode(x).lower() for x in clean_ref.split('°')]).find(cat)\n if ind != -1:\n clean_ref = clean_ref[:ind] + clean_ref[ind + len(cat):]\n ind_or = '°'.join([unidecode.unidecode(x).lower() for x in _ref.split('°')]).find(cat)\n ref_print = ref_print[:ind_or] + ''.join([' ' for _ in cat]) + ref_print[ind_or + len(cat):]\n clean_ref = clean_ref.strip()\n _temp = []\n for _char in clean_ref:\n if _char in string.ascii_lowercase:\n _temp.append(choice(string.ascii_lowercase))\n elif _char in string.ascii_uppercase:\n _temp.append(choice(string.ascii_uppercase))\n elif _char in string.digits:\n _temp.append(choice(string.digits))\n elif _char in [' ', '-', '(', ')', '.', ':']:\n _temp.append(_char)\n else:\n _temp.append(choice(string.punctuation))\n _temp = ''.join(_temp)\n\n ind = _ref.find(clean_ref)\n replacement_ref = _ref[:ind] + _temp + _ref[ind + len(_temp):]\n\n replacement_ref = self.match_case(replacement_ref, _ref)\n replacement_ref = self.match_ponctuation(replacement_ref, _ref)\n x.text, x[self.labels_col] = self.switch_entity(x, _ref, replacement_ref, 'REF_NUM')\n\n return x\n\n def switch_entity(self, x, old_entity, new_entity, entity_type):\n \"\"\"\n Remplace une entité par une autre dans un texte.\n :param x: ligne du DataFrame à traiter\n :param old_entity: entité à remplacer\n :param new_entity: nouvelle entité à insérer\n :param entity_type: type de l'entité à remplacer (sans B- ou I-)\n :return: texte et labels modifiés\n \"\"\"\n\n def _find_sub_list(sl, ll, start=0):\n \"\"\"\n Finds a sublist in another list and returns the starting and ending indices.\n :param sl: sublist\n :param ll: list\n :param start : starting index (all elements before that will be ignored by the search)\n :return: starting and ending positions of the sublist in the list\n \"\"\"\n sll = len(sl)\n ll = [None for _ in range(start)] + ll[start:]\n for ind in (i for i, e in enumerate(ll) if e == sl[0]):\n if ll[ind:ind + sll] == sl:\n return ind, ind + sll - 1\n return None, None\n\n new_entity = new_entity.replace(\" \", self.white_space_token)\n if old_entity.endswith(self.white_space_token):\n new_entity += self.white_space_token\n found_right_entity = False\n char_start = 0\n max_pos = len(x[self.labels_col]) - [x.split('-')[-1] for x in x[self.labels_col]][::-1].index(entity_type) - 1\n i_s, i_e = [None] * 2\n while not found_right_entity:\n i_s, i_e = _find_sub_list(old_entity.split(' '), x.text, char_start)\n found_label = None if x[self.labels_col][i_s] == 'O' else x[self.labels_col][i_s].split(\"-\")[-1]\n\n # we check whether we could find the entity in the text and if the corresponding labels are the right ones.\n # This prevents replacing a word or a sequence contained inside other entities.\n if (i_s is not None) & (found_label == entity_type):\n found_right_entity = True\n else:\n char_start = i_e + 1\n if char_start > max_pos:\n break\n assert found_right_entity, \"the identity we want to replace was not found\"\n\n x.text = x.text[:i_s] + new_entity.split(' ') + x.text[i_e + 1:]\n entity_labels = ['I-' + entity_type for _ in new_entity.split(' ')]\n if len(entity_labels) > 1:\n entity_labels[0] = 'B-' + entity_type\n x[self.labels_col] = x[self.labels_col][:i_s] + entity_labels + x[self.labels_col][i_e + 1:]\n\n return x.text, x[self.labels_col]\n\n\ndef detect_entities(_inputs, corpus, threshold=None):\n \"\"\"\n Détecte les entités nommées sélectionnées dans le corpus donné en argument.\n :param _inputs: paramètres d'entrainement du modèle\n :param corpus: corpus à annoter\n :param threshold: seuils de détection manuels. Si la probabilité d'une catégorie dépasse ce seuil, on prédit cette\n catégorie meme si elle ne correspond pas à la probabilité maximale.\n :return: corpus avec prédictions sur la nature des entités\n \"\"\"\n # Initialisation de la classe de pseudonymisation et entrainement du modèle.\n ner = Ner(_inputs)\n corpus_with_labels = ner.predict_with_model(corpus, threshold)\n\n return corpus_with_labels\n\n\ndef evaluate_model(_inputs, corpus):\n \"\"\"\n Evalue les performances du modèle indiqué dans _inputs sur un corpus donné.\n :param _inputs: paramètres du modèle à charger\n :param corpus: corpus sur lequel évaluer le modèle\n :return: statistiques d'évaluation du modèle\n \"\"\"\n # Initialisation de la classe de pseudonymisation et entrainement du modèle.\n ner = Ner(_inputs)\n ner.evaluate_model(corpus)\n\n\ndef load_custom_corpus(list_of_texts, labels_format):\n \"\"\"\n Convertit une liste de textes en un DataFrame au format attendu par l'algorithme de détection d'entités.\n :param list_of_texts: liste des textes à analyser\n :param labels_format: type de labels (BIO ou type d'entité uniquement)\n :return: DataFrame des textes\n \"\"\"\n df = pd.DataFrame()\n texts = []\n texts_raw = []\n files = []\n\n for i, raw_text in enumerate(list_of_texts):\n annotations = pd.DataFrame(columns=[\"start_pos\", \"end_pos\", \"type\", \"word\"])\n # Getting labels\n text, _ = get_text_and_labels(raw_text, annotations, labels_format)\n texts.append(text)\n texts_raw.append(raw_text)\n files.append(i)\n\n df['file'] = files\n df['raw_text'] = texts_raw\n df['text'] = texts\n\n return df\n\n\ndef load_doccano_corpus(list_of_corpus_path, labels_format):\n \"\"\"\n Convertit un ou des corpus doccano au format jsonl en DataFrame(s) au format attendu par l'algorithme de détection\n d'entités.\n :param list_of_corpus_path: liste des chemins des corpus à charger\n :param labels_format: type de labels (BIO ou type d'entité uniquement)\n :return: liste des DataFrame correspondant aux corpus\n \"\"\"\n corpus = {}\n for i, corpus_path in enumerate(list_of_corpus_path):\n\n _df = pd.DataFrame()\n texts = []\n texts_raw = []\n labels = []\n files = []\n df = pd.read_json(path_or_buf=corpus_path, lines=True)\n for _, row in df.iterrows():\n if len(row[\"label\"]) > 0:\n annotations = pd.concat([pd.DataFrame([_ann + [row[\"text\"][_ann[0]: _ann[1]]]],\n columns=[\"start_pos\", \"end_pos\", \"type\", \"word\"]) for _ann in\n row[\"label\"]], ignore_index=True).sort_values(['start_pos'])\n else:\n annotations = pd.DataFrame(columns=[\"start_pos\", \"end_pos\", \"type\", \"word\"])\n # Getting labels\n text, label = get_text_and_labels(row[\"text\"], annotations, labels_format)\n texts.append(text)\n texts_raw.append(row[\"text\"])\n labels.append(label)\n files.append(row[\"id\"])\n\n _df['file'] = files\n _df['raw_text'] = texts_raw\n _df['text'] = texts\n _df['labels'] = labels\n corpus[i] = _df\n\n return list(corpus.values())\n\n\ndef replace_entities(corpus, entities, paths, labels_column, labels_format, white_space_token):\n \"\"\"\n Remplace les entités précisées dans un corpus donné.\n :param corpus: DataFrame du corpus de textes\n :param entities: liste des entités à remplacer\n :param paths: chemins des fichiers csv utilisés pour générer des entités factices\n :param labels_column: nom de la colonne contenant les labels\n :param labels_format: type de labels (BIO ou type d'entité uniquement)\n :return: corpus avec entités remplacées\n \"\"\"\n\n pseudo = Pseudo(paths[\"noms\"], paths[\"adresses\"], paths[\"vehicules\"], paths[\"societes\"], labels_column,\n labels_format, white_space_token)\n corpus = pseudo.chain_of_replacements_other_entity(corpus, entities)\n\n return corpus\n\n\ndef show_annotations(corpus_df, entities, white_space_token, i=None):\n \"\"\"\n Affiche des exemples de textes annotés dans un notebook.\n :param corpus_df: corpus de textes avec annotations\n :param entities: entités à afficher\n :param white_space_token: token du tokenizer correspondant à un espace (différents entre flauBERT et camemBERT)\n :param i: numéro du texte à afficher. Si non précisé, la fonction affiche 3 textes au hasard.\n \"\"\"\n\n underlines = {ent: '#%02x%02x%02x' % (int(sns.color_palette('tab20', len(entities))[i][0] * 255),\n int(sns.color_palette('tab20', len(entities))[i][1] * 255),\n int(sns.color_palette('tab20', len(entities))[i][2] * 255))\n for i, ent in enumerate(entities)}\n\n if i is not None:\n assert i <= len(corpus_df), \"Le numéro de texte choisi ne doit pas excéder la taille du DataFrame.\"\n display(Markdown(f\"# Texte n°{i}\"))\n _sample = corpus_df.iloc[i:i + 1]\n else:\n display(Markdown(f\"# Exemples de textes\"))\n _sample = corpus_df[corpus_df[f\"labels\"].apply(lambda x: True if len(list(set(x))) > 1 else False)].sample(3)\n\n for _, _example in _sample.iterrows():\n text = \"\"\n _previous_label = \"O\"\n open_tag = 0\n for _word, _label in zip(_example[\"text\"], _example[f\"predicted_labels\"]):\n _word = _word.replace(white_space_token, ' ')\n if _label.split('-')[-1] != _previous_label:\n if _label != \"O\":\n if open_tag == 1:\n text += ''\n open_tag = 1\n entity = _label.split(\"-\")[-1]\n text += f''\n else:\n open_tag = 0\n text += ''\n text += f'{_word}'\n _previous_label = _label.split('-')[-1]\n if open_tag == 1:\n text += ''\n text += '
------------------------------------------
'\n\n display(Markdown(text))\n\n\ndef show_legend(entities):\n \"\"\"\n Affiche dans un notebook la légende de couleurs des différentes entités (à appeler avec la fonction show_annotations\n ci-dessous).\n :param entities: liste des entités à afficher\n \"\"\"\n underlines = {ent: '#%02x%02x%02x' % (int(sns.color_palette('tab20', len(entities))[i][0] * 255),\n int(sns.color_palette('tab20', len(entities))[i][1] * 255),\n int(sns.color_palette('tab20', len(entities))[i][2] * 255))\n for i, ent in enumerate(entities)}\n legend = \"\"\n for ent in entities:\n legend += f'{ent} '\n display(Markdown(legend))\n\n\ndef train_and_evaluate_model(_inputs, train_corpus, test_corpus):\n \"\"\"\n Entraine un modèle Transformers selon les paramètres précisés dans _inputs sur le corpus train_corpus et évalue ses\n performances sur le corpus test_corpus.\n :param _inputs: paramètres d'entrainement du modèle\n :param train_corpus: corpus d'entrainement\n :param test_corpus: corpus de test\n :return: statistiques d'évaluation du modèle\n \"\"\"\n # Initialisation de la classe de pseudonymisation et entrainement du modèle.\n ner = Ner(_inputs)\n ner.get_corpus_stats(train_corpus)\n ner.train_model_on_corpus(train_corpus, test_corpus)\n\n\ndef write_doccano_format(corpus, white_space_token, output_file):\n \"\"\"\n Exporte un DataFrame annoté au format doccano.\n :param corpus: DataFrame du corpus à exporter\n :param white_space_token: token du tokenizer correspondant à un espace (différents entre flauBERT et camemBERT)\n :param output_file: chemin du fichier de sortie\n \"\"\"\n\n _n_start = 0\n _n_end = 0\n jsonl = []\n _full_json = {\"id\": [], \"text\": '', \"label\": []}\n _previous_part = 0\n for _, row in corpus.iterrows():\n labels = []\n _prev = \"O\"\n for i, _lab in enumerate(row[\"predicted_labels\"]):\n if _lab != _prev:\n if _lab != \"O\":\n if len(row[\"text\"][i]) == 0:\n _prefix = 0\n else:\n _prefix = 1 if row[\"text\"][i][0] == white_space_token else 0\n _n_start = len(\"\".join(row[\"text\"][:i]).replace(white_space_token, ' ')) + _prefix\n else:\n _n_end = len(\"\".join(row[\"text\"][:i]).replace(white_space_token, ' '))\n labels.append([_n_start, _n_end, _prev])\n _prev = _lab\n if _n_end < _n_start:\n labels.append([_n_start, len(\"\".join(row[\"text\"]).replace(white_space_token, ' ')), _prev])\n\n text_joined = \"\".join(row[\"text\"]).replace(white_space_token, ' ')\n id_clean = row[\"file\"].replace(white_space_token, ' ')\n _json = {\"id\": id_clean, \"text\": text_joined, \"label\": labels}\n # Merging parts of the same text together\n if row[\"text_part\"] <= _previous_part:\n jsonl.append(_full_json)\n _full_json = {\"id\": _json[\"id\"], \"text\": _json[\"text\"], \"label\": _json[\"label\"]}\n else:\n _full_json = {\"id\": _json[\"id\"],\n \"text\": _full_json[\"text\"] + _json[\"text\"],\n \"label\": _full_json[\"label\"] + _json[\"label\"]}\n _previous_part = row[\"text_part\"]\n jsonl.append(_full_json)\n\n with open(output_file, 'w', encoding='utf-8') as f:\n for file in jsonl:\n f.write(json.dumps(file, ensure_ascii=False) + '\\n')\n", "sub_path": "pseudo.py", "file_name": "pseudo.py", "file_ext": "py", "file_size_in_byte": 94984, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.INFO", "line_number": 32, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 38, "usage_type": "call"}, {"api_name": "seaborn.color_palette", "line_number": 42, "usage_type": "call"}, {"api_name": "seaborn.color_palette", "line_number": 43, "usage_type": "call"}, {"api_name": "seaborn.color_palette", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 53, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.argmax", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 139, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 142, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 194, "usage_type": "call"}, {"api_name": "sklearn.metrics.confusion_matrix", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 214, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 214, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 214, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_score", "line_number": 219, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_score", "line_number": 223, "usage_type": "call"}, {"api_name": "sklearn.metrics.recall_score", "line_number": 226, "usage_type": "call"}, {"api_name": "sklearn.metrics.recall_score", "line_number": 228, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 231, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 233, "usage_type": "call"}, {"api_name": "sklearn.metrics.precision_score", "line_number": 241, "usage_type": "call"}, {"api_name": "sklearn.metrics.recall_score", "line_number": 242, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 266, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 266, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 268, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 270, "usage_type": "call"}, {"api_name": "os.path", "line_number": 270, "usage_type": "attribute"}, {"api_name": "transformers.AutoTokenizer.from_pretrained", "line_number": 272, "usage_type": "call"}, {"api_name": "transformers.AutoTokenizer", "line_number": 272, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 275, "usage_type": "call"}, {"api_name": "os.path", "line_number": 275, "usage_type": "attribute"}, {"api_name": "transformers.AutoConfig.from_pretrained", "line_number": 277, "usage_type": "call"}, {"api_name": "transformers.AutoConfig", "line_number": 277, "usage_type": "name"}, {"api_name": "transformers.AutoModelForTokenClassification.from_config", "line_number": 280, "usage_type": "call"}, {"api_name": "transformers.AutoModelForTokenClassification", "line_number": 280, "usage_type": "name"}, {"api_name": "transformers.AutoTokenizer.from_pretrained", "line_number": 282, "usage_type": "call"}, {"api_name": "transformers.AutoTokenizer", "line_number": 282, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 305, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 305, "usage_type": "attribute"}, {"api_name": "numpy.isnan", "line_number": 306, "usage_type": "call"}, {"api_name": "numpy.nanmax", "line_number": 306, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 308, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 309, "usage_type": "call"}, {"api_name": "utils.convert_to_features", "line_number": 330, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 333, "usage_type": "call"}, {"api_name": "torch.long", "line_number": 333, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 334, "usage_type": "call"}, {"api_name": "torch.long", "line_number": 334, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 335, "usage_type": "call"}, {"api_name": "torch.long", "line_number": 335, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 336, "usage_type": "call"}, {"api_name": "torch.long", "line_number": 336, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 337, "usage_type": "call"}, {"api_name": "torch.long", "line_number": 337, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 338, "usage_type": "call"}, {"api_name": "torch.long", "line_number": 338, "usage_type": "attribute"}, {"api_name": "torch.utils.data.TensorDataset", "line_number": 340, "usage_type": "call"}, {"api_name": "focal_loss.FocalLoss", "line_number": 358, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 360, "usage_type": "call"}, {"api_name": "torch.where", "line_number": 366, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 367, "usage_type": "call"}, {"api_name": "re.finditer", "line_number": 438, "usage_type": "call"}, {"api_name": "utils.find_sub_list", "line_number": 453, "usage_type": "call"}, {"api_name": "utils.find_sub_list", "line_number": 455, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 463, "usage_type": "call"}, {"api_name": "re.finditer", "line_number": 483, "usage_type": "call"}, {"api_name": "utils.find_sub_list", "line_number": 500, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 508, "usage_type": "call"}, {"api_name": "re.finditer", "line_number": 532, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 536, "usage_type": "call"}, {"api_name": "utils.find_sub_list", "line_number": 548, "usage_type": "call"}, {"api_name": "utils.find_sub_list", "line_number": 550, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 558, "usage_type": "call"}, {"api_name": "re.finditer", "line_number": 581, "usage_type": "call"}, {"api_name": "operator.itemgetter", "line_number": 585, "usage_type": "call"}, {"api_name": "utils.find_sub_list", "line_number": 597, "usage_type": "call"}, {"api_name": "utils.find_sub_list", "line_number": 599, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 607, "usage_type": "call"}, {"api_name": "torch.utils.data.SequentialSampler", "line_number": 627, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 628, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 633, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 638, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 640, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 652, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 652, "usage_type": "call"}, {"api_name": "torch.utils.data.RandomSampler", "line_number": 683, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 684, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 698, "usage_type": "call"}, {"api_name": "torch.nn.utils.clip_grad_norm_", "line_number": 713, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 713, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 726, "usage_type": "call"}, {"api_name": "os.path", "line_number": 726, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 734, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 739, "usage_type": "call"}, {"api_name": "os.path", "line_number": 739, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 739, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 739, "usage_type": "name"}, {"api_name": "os.mkdir", "line_number": 740, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 741, "usage_type": "call"}, {"api_name": "os.path", "line_number": 741, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 742, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 744, "usage_type": "call"}, {"api_name": "os.path", "line_number": 744, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 746, "usage_type": "call"}, {"api_name": "torch.optim.AdamW", "line_number": 757, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 757, "usage_type": "attribute"}, {"api_name": "transformers.get_linear_schedule_with_warmup", "line_number": 758, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 766, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 767, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 767, "usage_type": "attribute"}, {"api_name": "torch.manual_seed", "line_number": 768, "usage_type": "call"}, {"api_name": "faker.Faker", "line_number": 811, "usage_type": "call"}, {"api_name": "faker.Faker.seed", "line_number": 813, "usage_type": "call"}, {"api_name": "faker.Faker", "line_number": 813, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 826, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 827, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 829, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 830, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 930, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 930, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 932, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 932, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 936, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 936, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 937, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 937, "usage_type": "attribute"}, {"api_name": "random.random", "line_number": 940, "usage_type": "call"}, {"api_name": "random.random", "line_number": 945, "usage_type": "call"}, {"api_name": "random.random", "line_number": 969, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 972, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 972, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 973, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 973, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 976, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 976, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 981, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 981, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 982, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 982, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 985, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 985, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 999, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 1071, "usage_type": "call"}, {"api_name": "utils.switch_entity", "line_number": 1094, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 1165, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 1173, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 1174, "usage_type": "call"}, {"api_name": "utils.find_sub_list", "line_number": 1177, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 1180, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 1209, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 1211, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 1212, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 1213, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 1216, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 1219, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 1222, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 1225, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 1226, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 1228, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 1229, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 1231, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 1232, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 1234, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 1235, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 1237, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 1238, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 1239, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 1240, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 1242, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 1243, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 1244, "usage_type": "call"}, {"api_name": "locale.setlocale", "line_number": 1261, "usage_type": "call"}, {"api_name": "locale.LC_ALL", "line_number": 1261, "usage_type": "attribute"}, {"api_name": "unidecode.unidecode", "line_number": 1281, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 1282, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1282, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 1282, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 1282, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 1287, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1287, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 1287, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 1287, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 1288, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1288, "usage_type": "attribute"}, {"api_name": "locale.setlocale", "line_number": 1302, "usage_type": "call"}, {"api_name": "locale.LC_ALL", "line_number": 1302, "usage_type": "attribute"}, {"api_name": "unidecode.unidecode", "line_number": 1320, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 1326, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1326, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 1326, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 1328, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1328, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 1330, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 1330, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 1362, "usage_type": "call"}, {"api_name": "utils.find_sub_list", "line_number": 1364, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 1449, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1449, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 1492, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1492, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 1493, "usage_type": "call"}, {"api_name": "string.ascii_uppercase", "line_number": 1493, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 1493, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 1495, "usage_type": "call"}, {"api_name": "string.ascii_uppercase", "line_number": 1495, "usage_type": "attribute"}, {"api_name": "random.random", "line_number": 1517, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 1519, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1519, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 1521, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1521, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 1522, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1522, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 1523, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1523, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 1555, "usage_type": "call"}, {"api_name": "random.random", "line_number": 1557, "usage_type": "call"}, {"api_name": "random.random", "line_number": 1560, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 1563, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1563, "usage_type": "attribute"}, {"api_name": "numpy.any", "line_number": 1586, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 1586, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 1587, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 1592, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 1593, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 1602, "usage_type": "call"}, {"api_name": "utils.find_sub_list", "line_number": 1607, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 1615, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 1616, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 1625, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 1628, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 1654, "usage_type": "call"}, {"api_name": "unidecode.unidecode", "line_number": 1657, "usage_type": "call"}, {"api_name": "string.ascii_lowercase", "line_number": 1662, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 1663, "usage_type": "call"}, {"api_name": "string.ascii_lowercase", "line_number": 1663, "usage_type": "attribute"}, {"api_name": "string.ascii_uppercase", "line_number": 1664, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 1665, "usage_type": "call"}, {"api_name": "string.ascii_uppercase", "line_number": 1665, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 1666, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 1667, "usage_type": "call"}, {"api_name": "string.digits", "line_number": 1667, "usage_type": "attribute"}, {"api_name": "random.choice", "line_number": 1671, "usage_type": "call"}, {"api_name": "string.punctuation", "line_number": 1671, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 1773, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1779, "usage_type": "call"}, {"api_name": "utils.get_text_and_labels", "line_number": 1781, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1804, "usage_type": "call"}, {"api_name": "pandas.read_json", "line_number": 1809, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 1812, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1812, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1816, "usage_type": "call"}, {"api_name": "utils.get_text_and_labels", "line_number": 1818, "usage_type": "call"}, {"api_name": "seaborn.color_palette", "line_number": 1860, "usage_type": "call"}, {"api_name": "seaborn.color_palette", "line_number": 1861, "usage_type": "call"}, {"api_name": "seaborn.color_palette", "line_number": 1862, "usage_type": "call"}, {"api_name": "IPython.display.display", "line_number": 1867, "usage_type": "call"}, {"api_name": "IPython.display.Markdown", "line_number": 1867, "usage_type": "call"}, {"api_name": "IPython.display.display", "line_number": 1870, "usage_type": "call"}, {"api_name": "IPython.display.Markdown", "line_number": 1870, "usage_type": "call"}, {"api_name": "IPython.display.display", "line_number": 1895, "usage_type": "call"}, {"api_name": "IPython.display.Markdown", "line_number": 1895, "usage_type": "call"}, {"api_name": "seaborn.color_palette", "line_number": 1904, "usage_type": "call"}, {"api_name": "seaborn.color_palette", "line_number": 1905, "usage_type": "call"}, {"api_name": "seaborn.color_palette", "line_number": 1906, "usage_type": "call"}, {"api_name": "IPython.display.display", "line_number": 1911, "usage_type": "call"}, {"api_name": "IPython.display.Markdown", "line_number": 1911, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 1976, "usage_type": "call"}]}
+{"seq_id": "623216590", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Aug 19 15:01:32 2017\n\n@author: cpkmanchee\n\nNotes:\n\n- classes of Pulse, Fiber, and FiberGain poses all the parameters required for th einput of most functions\n- functions should not change class object parameters; instead they should return a value which can be used to \nchange the object's parameters in the primary script \n\"\"\"\n\nimport numpy as np\nimport pickle\n\nfrom beamtools.constants import h, c, pi, mu0, eps0\nfrom beamtools.common import Func, gaussian, normalize, rk4, DataObj\n\nfrom tqdm import tqdm\nfrom copy import deepcopy\n\n\nclass Pulse:\n '''\n Defines a Pulse object\n .time = time array (s)\n .freq = corresponding angular freq array (rad/s)\n .At = time domain Field, units sqrt(power) i.e. |At|**2 = power\n .lambda0 = central wavelength of pulse\n\n .getAf() = returns freq domain Field\n\n Note: At should be used as the primary field. Af should only be reference. \n Any time field is modified it should be stored as At. Then use getAf() to get current freq domain field.\n\n '''\n\n T_BIT_DEFAULT = 12 #default time resolution, 2^12\n T_WIN_DEFAULT = 20E-12 #default window size, 20ps\n\n def __init__(self, lambda0=1.030E-6, ig=False):\n self.time = None\n self.freq = None\n self.At = None\n self.lambda0 = lambda0\n self.freq_dep = None\n\n if ig:\n self.initializeGrid()\n\n def initializeGrid(self, t_bit_res=T_BIT_DEFAULT, t_window=T_WIN_DEFAULT):\n nt = 2**t_bit_res #number of time steps, power of 2 for FFT\n dtau = 2*t_window/nt #time step size\n\n self.time = dtau*np.arange(-nt//2, nt//2) #time array\n self.freq = 2*np.pi*np.fft.fftfreq(nt,dtau) #frequency array\n self.nt = nt\n self.dt = dtau\n\n def getAf(self):\n '''Return Af, spectral field.\n '''\n return ((self.dt*self.nt)/(np.sqrt(2*np.pi)))*np.fft.fft(self.At)\n\n def getPt(self,t0=False):\n '''Return Power, time domain.\n t0 - True = center pulse about maximum\n '''\n pt = np.abs(self.At)**2\n if t0:\n pt = np.interp(self.time,self.time-self.time[np.argmax(pt)],pt)\n return pt\n\n def getPf(self):\n '''Return Pf, power spectral density.\n '''\n return np.abs(self.getAf())**2\n\n def phase(self):\n '''Yeild phase of temporal field'''\n return np.angle(self.At)\n\n def chirp(self):\n '''return chirp actually inst freq.\n (-)time derivative of phase\n Need to manually calc differences due to np.gradient causing spikes.\n Spikes occur when phase angle flips from +/-. Gradient gives wrong 'quadrant'.\n Linear approximation which is OK for even spaced x, which it is (self.time).\n '''\n ph=self.phase()\n dt=self.dt\n\n ch=np.array([np.arctan2(np.sin(ph[i+2]-ph[i]),np.cos(ph[i+2]-ph[i]))/(2*dt) for i in range(ph[1:-1].size)])\n ch = np.insert(ch, 0,(ph[1] - ph[0])/dt)\n ch = np.insert(ch,-1,(ph[-1]-ph[-2])/dt)\n\n return -ch\n\n def shift_t0(self):\n '''shift_t0 to maximum position'''\n self.At = np.interp(self.time,self.time-self.time[np.argmax(self.getPt())],self.At)\n\n def tl_Pt(self):\n '''calculate transform limited pulse\n '''\n return np.abs(np.fft.fftshift(np.fft.ifft((self.getPf)**(1/2))))**2\n\n def copyPulse(self, new_At=None):\n '''Duplicates pulse, outputs new pulse instance.\n Can set new At at same time by sending new_At. If not sent, new_pulse.At is same\n '''\n new_pulse = deepcopy(self)\n\n if new_At is not None:\n new_pulse.At = new_At\n\n return new_pulse\n\n\nclass Fiber:\n '''Defines a Fiber object.\n .length = length of fiber (m)\n .alpha = loss coefficient (m^-1), +alpha means loss\n .beta = dispersion parameters, 2nd 3rd 4th order. array\n .gamma = nonlinear parameter, (W*m)^-1\n \n can be used for simple gain fiber by using alpha (-alpha = gain-loss)\n\n .core_d = core diameter\n .clad_d = cladding diameter\n\n .z is the z-axis array for the fiber\n\n grid_type specifies whether the z-grid is defined by the grid spacing ('abs' or absolute),\n or number of points ('rel' or relative)\n z_grid is either the grid spacing (abs) or number of grid points (rel)\n\n '''\n Z_STP_DEFAULT = 0.003 #default grid size, in m, 3mm\n Z_NUM_DEFAULT = 256 #default number of grid points, 300\n\n CORE_D_DEFAULT = 6E-6 #default core diameter, 6um\n CLAD_D_DEFAULT = 125E-6 #default clad diameter, 125um\n\n def __init__(self, length=0, grid_type='rel', z_grid=None, alpha=0, beta=np.array([0,0,0]), gamma=0):\n\n self.length = length\n self.alpha = alpha\n self.beta = beta\n self.gamma = gamma\n\n self.core_d = self.CORE_D_DEFAULT\n self.clad_d = self.CLAD_D_DEFAULT\n\n self.initializeGrid(self.length, grid_type, z_grid)\n\n def initializeGrid(self, length, grid_type='rel', z_grid=None):\n '''\n -sets up the z-axis array for the fiber\n -can be called and re-called at any time (even after creation)\n -must provide fiber length, self.length is redefined when initializeGrid is called\n '''\n\n self.length = length\n\n if grid_type.lower() == 'abs':\n #grid type is 'absolute', z_grid is grid spacing\n if z_grid == None:\n z_grid = self.Z_STP_DEFAULT \n\n nz = np.max([1,np.round(self.length/np.abs(z_grid))]) + 1\n self.z = np.linspace(0, self.length, nz) #position array\n\n else:\n # grid type is 'relative', z_grid is number of grid points\n if z_grid == None or z_grid < 1:\n z_grid = self.Z_NUM_DEFAULT\n\n self.z = np.linspace(0, self.length, z_grid) #position array\n\n def copyFiber(self, length=None):\n '''Duplicates Fiber, outputs new fiber instance.\n Can set new length at same time. If not sent, new_fiber.length is same.\n '''\n new_fiber = deepcopy(self)\n\n if length is not None:\n new_fiber.length = length\n\n return new_fiber\n\n\nclass FiberGain:\n '''\n Defines a gain Fiber object with gain parameters\n .length = length of fiber (m)\n .alpha = loss coefficient (m^-1)\n .beta = dispersion parameters, 2nd 3rd 4th order. array\n .gamma = nonlinear parameter, (W*m)^-1\\\n .gain = fiber gain coefficient (m^-1), same units as alpha, can be z-array or constant\n \n .core_d = core diameter\n .clad_d = cladding diameter\n\n .sigma_x are 2x2 arrays. col 0 = wavelength, col 1 = sigma, row 0 = pump, row 1= signal\n .tau is excited state lifetime\n .N is core dopant density\n .z is the z-axis array for the fiber\n\n Note: N*sigma_a_pump = pump abs coefficient (often quoted by manufacturers in dB/m)\n\n grid_type specifies whether the z-grid is defined by the grid spacing ('abs' or absolute),\n or number of points ('rel' or relative)\n z_grid is either the grid spacing (abs) or number of grid points (rel)\n\n '''\n\n Z_STP_DEFAULT = 0.003 #default grid size, in m\n Z_NUM_DEFAULT = 256 #default number of grid points\n\n CORE_D_DEFAULT = 6E-6 #default core diameter, 6um\n CLAD_D_DEFAULT = 125E-6 #default clad diameter, 125um\n\n def __init__(self, length=0, alpha=0, beta=np.array([0,0,0]), gamma=0, gain=0, grid_type='abs', z_grid=None):\n\n self.length = length\n self.alpha = alpha\n self.beta = beta\n self.gamma = gamma\n self.gain = gain\n\n self.sigma_a = np.zeros(2)\n self.sigma_e = np.zeros(2)\n \n self.lambdas = np.zeros(2)\n\n self.tau = 770E-6\n self.N = 1.891669E25 #See Nov. 24 Book 8 Page 72\n\n self.core_d = self.CORE_D_DEFAULT\n self.clad_d = self.CLAD_D_DEFAULT\n\n self.z = np.arange(1)\n self.initializeGrid(self.length, grid_type, z_grid)\n\n\n def initializeGrid(self, length, grid_type='abs', z_grid=None):\n '''\n -sets up the z-axis array for the fiber\n -can be called and re-called at any time (even after creation)\n -must provide fiber length, self.length is redefined when initializeGrid is called\n -also initializes gain array\n '''\n self.length = length\n old_z = self.z\n\n if grid_type.lower() == 'abs':\n #grid type is 'absolute', z_grid is grid spacing\n if z_grid == None:\n z_grid = self.Z_STP_DEFAULT \n\n nz = np.max([1,np.round(self.length/np.abs(z_grid))]) + 1\n self.z = np.linspace(0, self.length, nz) #position array\n\n else:\n # grid type is 'relative', z_grid is number of grid points\n if z_grid == None or z_grid < 1:\n z_grid = self.Z_NUM_DEFAULT\n\n self.z = np.linspace(0, self.length, z_grid) #position array\n\n self.updateGain(old_z)\n\n\n def updateGain(self, old_z):\n '''Update gain array\n '''\n if np.size(self.gain) is 1:\n self.gain = self.gain*np.ones(np.size(self.z))\n elif np.size(self.gain) is np.size(old_z):\n self.gain = np.interp(self.z,old_z,self.gain)\n elif np.size(self.gain) < np.size(old_z):\n self.gain = np.interp(self.z,old_z[:np.size(self.gain)],self.gain)\n elif np.size(self.gain) > np.size(old_z):\n self.gain = np.interp(self.z,old_z,self.gain[:np.size(old_z)])\n else:\n self.gain = np.zeros(np.size(self.z))\n\n\n def copyFiber(self, length=None):\n '''Duplicates Fiber, outputs new fiber instance.\n Can set new length at same time. If not sent, new_fiber.length is same.\n '''\n new_fiber = deepcopy(self)\n\n if length is not None:\n new_fiber.length = length\n\n return new_fiber\n \n\ndef save_obj(obj, filename):\n with open(filename, 'wb') as output:\n pickle.dump(obj, output, -1)\n\n\ndef load_obj(filename):\n with open(filename, 'rb') as input:\n obj = pickle.load(input)\n\n return obj\n\n\ndef check_input(inputData, requiredType, *inputNum):\n \n if len(inputNum)==1:\n number = inputNum[0]\n else:\n number = '#'\n\n if not(isinstance(inputData, eval(requiredType))):\n errMsg = 'Input ' + str(number) + ' is type ' + str(type(inputData)) + '\\nRequired:' + ' \\'' + str(requiredType) + '\\'\\n' \n else:\n errMsg = -1\n \n return(errMsg)\n\n\ndef rms_width(x,F):\n \n if isinstance(x, np.ndarray):\n pass\n else:\n x = np.asarray(x)\n \n if isinstance(F, np.ndarray):\n pass\n else:\n F = np.asarray(F)\n \n dx = np.gradient(x)\n \n #Normalization integration\n areaF=0\n for i in range(len(x)):\n areaF += dx[i]*F[i]\n\n #Average value\n mu=0\n for i in range(len(x)):\n mu += x[i]*F[i]*dx[i]/areaF\n\n #Varience (sd = sqrt(var))\n var = 0\n for i in range(len(x)):\n var += dx[i]*F[i]*(x[i]-mu)**2/areaF\n \n #returns avg and rms width\n return(mu, np.sqrt(var))\n \n\ndef propagate_fiber (pulse, fiber, autodz=False):\n '''This function will propagate the input field along the length of...\n a fibre with the given properties...\n\n # Pulse propagation via Nonlinear Schrodinger Equation (NLSE)\n # dA/dz = -ib2/2 (d^2A/dtau^2) + b3/6 (d^3 A/dtau^3) -aplha/2 + ig|A|^2*A \n # --> A is field A = sqrt(P0)*u\n\n Requires a Pulse class object and Fiber class object. Fiber can also be FiberGain class\n\n Inputs:\n pulse = Pulse class object\n fiber = Fiber class object (Fiber or FiberGain)\n autodz = optional to automatically calc z_grid. \n False - no change to fiber z_grid\n True - autocalculates, sets to medium resolution (35 points per Lref)\n N - integer, autocalculates sets res to N points per Lref\n \n Outputs:\n outputField = time domain output field, At\n \n Warning: setting autodz = True will modify fiber object!!!\n autodz uses calc_zgrid to calculate dz based on the input pulse and fiber\n Should not be used for gain fiber!!!, since gain calc depends on dz as well\n ''' \n if autodz == False:\n pass\n else:\n if autodz is True:\n res = 'med'\n else:\n try:\n res = autodz//1\n except TypeError:\n res = 'med'\n\n dz = calc_zgrid(fiber,pulse,res)\n fiber.initializeGrid(fiber.length, 'abs', dz)\n \n #Pulse inputs\n #nt = pulse.nt\n #tau = pulse.time\n #dtau = pulse.dt\n omega = pulse.freq\n\n #fiber inputs\n #nz = np.size(fiber.z)\n dz = np.diff(fiber.z) #position step size\n ndz= np.size(dz) #ndz*dz[i] = fiber length (constant dz)\n\n #compile losses(alpha) and gain appropriately, result should have same dim as fiber.z\n if type(fiber) is Fiber:\n #Fiber does not have inherent gain parameter, thus gain is set to 0\n gain = np.zeros(np.shape(fiber.z))\n elif type(fiber) is FiberGain:\n #FiberGain has gain parameter\n #if fiber.gain is const, this creates a const arrray, if .gain is an array this is simply X 1\n gain = np.ones(np.shape(fiber.z))*fiber.gain\n else:\n #Don't know when this would apply\n gain = np.zeros(np.shape(fiber.z))\n\n #combined loss and gain, will be array same dim as fiber.z\n #fiber.alpha could be const. or array, result is same dimensionally\n alpha = (fiber.alpha - gain)\n\n try:\n if np.shape(pulse.freq_dep) == np.shape(pulse.freq):\n alpha = np.outer(alpha,pulse.freq_dep)\n except:\n print('freq_dep failed')\n\n #Define Dispersion operator: D = G + B, G = gain/loss, B = dispersion\n G = -alpha/2 + 0j*alpha\n B = 0\n \n #Dispersion components\n for i in range(len(fiber.beta)):\n B += (1j*fiber.beta[i]*omega**(i+2)/np.math.factorial(i+2))\n \n #Nonlinear operator, constant\n N = 1j*fiber.gamma\n \n #Main propagation loop\n At = pulse.At*np.exp(-np.abs(pulse.At)**2*N*dz[0]/2)\n for i in tqdm(range(ndz),desc='Progagate Fiber',leave=False):\n \n D = G[i] + B\n \n Af = np.fft.fft(At)\n Af = Af*np.exp(D*dz[i])\n At = np.fft.ifft(Af)\n At = At*np.exp(N*dz[i]*np.abs(At)**2)\n\n #Final Propagation steps\n Af = np.fft.fft(At)\n Af = Af*np.exp(D[-1]*dz[-1])\n At = np.fft.ifft(Af)\n outputField = At*np.exp(np.abs(At)**2*N*dz[-1]/2)\n \n return outputField\n \n\ndef calc_zgrid(fiber, pulse, res='med'):\n '''Autocalculation of zgrid of Fiber.\n fiber = class Fiber() or FiberGain() instance\n pulse = class Pulse() instance\n res = resolution, either numeric or string\n '''\n if isinstance(res, str):\n if res.lower() == 'low':\n n = 10\n elif res.lower() == 'med':\n n = 35\n elif res.lower() == 'high':\n n = 100\n else:\n n = 35\n \n elif isinstance(res, int) or isinstance(res, float):\n n = res//1\n \n if n<=0:\n n = 1\n\n _, t0 = rms_width(pulse.time, np.abs(pulse.At))\n p0 = (np.abs(pulse.At)**2).max()\n \n ld = t0**2/(np.abs(fiber.beta[0]))\n ln = 1/(p0*fiber.gamma)\n l_ref = 1/((1/ld)+(1/ln))\n \n return l_ref/n\n \n \ndef calc_gain(fiber, Pp, Ps, \n pump_scheme='core', \n pump_dir='forward', \n method='simple', \n min_err=1E-4):\n '''\n Calculate steady state gain over fiber\n Output z-array of gain\n fiber.sigma_x are 2x2 arrays. col 0 = wavelength, col 1 = sigma, row 0 = pump, row 1= signal\n '''\n\n s_ap = fiber.sigma_a[0]\n s_as = fiber.sigma_a[1]\n\n s_ep = fiber.sigma_e[0]\n s_es = fiber.sigma_e[1]\n\n v_p = c/fiber.lambdas[0]\n v_s = c/fiber.lambdas[1]\n\n b_p = (s_ap + s_ep)/(h*v_p)\n b_s = (s_as + s_es)/(h*v_s)\n a_p = s_ap/(h*v_p)\n a_s = s_as/(h*v_s)\n\n tau_se = fiber.tau\n dv_ase = (53E-9)*(v_s**2/c)\n\n MFA = np.pi*(fiber.core_d/2)**2\n G_s = MFA/(np.pi*(fiber.core_d/2)**2)\n Is = Ps/(np.pi*(fiber.core_d/2)**2)\n\n #define pump overlap, core pumped or clad pumped\n if pump_scheme.lower() in {'clad', 'cladding'}:\n G_p = MFA/(np.pi*(fiber.clad_d/2)**2)\n Ip = Pp/(np.pi*(fiber.clad_d/2)**2)\n else:\n G_p = MFA/(np.pi*(fiber.core_d/2)**2)\n Ip = Pp/(np.pi*(fiber.core_d/2)**2)\n\n\n g = np.zeros(np.shape(fiber.z))\n N=fiber.N\n dz = np.diff(fiber.z)\n\n\n if method.lower() in {'simple', 's'}:\n #simple integration for intensities and n\n for i in tqdm(range(np.size(dz)),desc='Calculate Gain',leave=False):\n\n n = (a_p*Ip + a_s*Is)/(b_p*Ip + b_s*Is + 1/tau_se)\n \n Ip = Ip*np.exp(-G_p*(s_ap*N*(1-n) - s_ep*N*n)*dz[i])\n Is = Is*np.exp(-G_s*(s_as*N*(1-n) - s_es*N*n)*dz[i])\n\n g[i] = (s_es*N*n - s_as*N*(1-n))\n\n\n elif method.lower() in {'rk4', 'r'}:\n #iterative rk4 method, significantly longer than 'simple'\n #likely only necessary for double clad fiber\n I0p = Ip\n I0s = Is\n Ip = I0p*np.ones(np.shape(fiber.z))\n Isig = I0s*np.ones(np.shape(fiber.z))\n Iasef = np.zeros(np.shape(fiber.z))\n Iaseb = np.zeros(np.shape(fiber.z))\n \n dIp = lambda z, I, n: -G_p*(s_ap*N*(1-n.at(z)) - s_ep*N*n.at(z))*I \n dIsig = lambda z, I, n: -G_s*(s_as*N*(1-n.at(z)) - s_es*N*n.at(z))*I \n dIase = lambda z, I, n: -G_s*(s_as*N*(1-n.at(z)) - s_es*N*n.at(z))*I + n.at(z)*h*v_s*N*s_es*dv_ase/MFA\n\n n = Func()\n n.ind = fiber.z\n n.val = np.zeros(np.shape(n.ind))\n\n loop_num = 0\n gain_err = 1\n c_gain = 0\n \n max_loops = 500\n\n while (loop_num < max_loops and gain_err > min_err):\n\n p_gain = c_gain\n \n for i in tqdm(range(np.size(fiber.z)),desc='Calculate Gain',leave=False):\n Is = Isig + Iasef + Iaseb\n n.val[i] = (a_p*Ip[i] + a_s*Is[i])/(b_p*Ip[i] + b_s*Is[i] + 1/tau_se)\n \n if i < np.size(fiber.z)-1:\n zf = np.array([fiber.z[i],fiber.z[i+1]])\n zb = np.array([fiber.z[-i-2],fiber.z[-i-1]])\n \n if pump_dir.lower().startswith('b'):\n Ip[-i-2],_ = np.flipud(rk4(dIp, np.flipud(zb), Ip[-i-1], [n], True))\n else:\n _,Ip[i+1] = rk4(dIp, zf, Ip[i], [n])\n _,Isig[i+1] = rk4(dIsig, zf, Isig[i], [n])\n _,Iasef[i+1] = rk4(dIase, zf, Iasef[i], [n])\n Iaseb[-i-2],_ = np.flipud(rk4(dIase, np.flipud(zb), Iaseb[-i-1], [n], True))\n\n g = (s_es*N*n.val - s_as*N*(1-n.val))\n\n c_gain = g.sum()\n gain_err = np.abs((c_gain-p_gain)/c_gain)\n\n loop_num = loop_num + 1\n \n\n #print(gain_err, loop_num)\n \n\n else:\n g.fill(1)\n print('Unknown method. Gain set to 1')\n \n return g\n\n\ndef grating_pair(pulse, L, N, AOI, loss=0, return_coef=False):\n '''\n Simulate grating pair, double pass!\n pulse = input pulse object\n L = grating separation (m), use (-) L for stretcher, (+) L for compressor geometry\n N = lns/mm of gratings\n AOI = angle of incidence (deg)\n loss = %loss of INTENSITY (not field)\n\n theta = diffraction angle (assumed -1 order, as is standard)\n d = groove spacing\n\n returns time-domain output field\n\n '''\n m = 1\n g = AOI*np.pi/180 #convert AOI into rad\n d = 1E-3/N #gives grove spacing in m\n\n Af = np.fft.fft(pulse.At)\n w0 = 2*np.pi*c/pulse.lambda0\n omega = pulse.freq\n theta = np.arcsin(m*2*np.pi*c/(w0*d) - np.sin(g)) \n\n phi2 = (-m**2*2*4*(np.pi**2)*L*c/(d**2*w0**3))*(1/np.cos(theta)**3)\n phi3 = (-3*phi2/w0)*(1+(2*np.pi*c*m*np.sin(theta)/(w0*d*np.cos(theta)**2)))\n phi4 = ((2*phi3)**2/(3*phi2)) + phi2*(2*np.pi*c*m/(w0**2*d*np.cos(theta)**2))**2\n\n output_At = np.sqrt(1-loss)*np.fft.ifft(Af*np.exp(1j*(phi2*(omega)**2/2 + phi3*(omega)**3/6 + phi4*(omega)**4/24)))\n \n if return_coef:\n return output_At, np.array([phi2,phi3,phi4])\n else:\n return output_At\n\n\ndef power_tap(pulse, tap, loss=0):\n '''\n Simulate splitter or tap\n tap is 'output', 'signal' is to cavity. Just semantics though\n signal pulse is (1-tap)\n\n pulse = input pulse\n tap = tap ratio, ex. 1 == 1%, 50 = %50\n loss = % loss\n\n tap and loss ratios are of INTENSITY, not field\n\n Note: tap and signal are 'dephased', differ by factor of i. This is how these work in real life\n \n '''\n\n At = np.sqrt(1-loss)*pulse.At\n output_tap = 1j*np.sqrt(tap/100)*At\n output_signal = np.sqrt(1-tap/100)*At\n\n return output_signal, output_tap\n\n\ndef coupler_2x2(pulse1, pulse2, tap, loss=0):\n '''Simulates splitter/coupler\n requires 2 pulses, outputs 2 pulses.\n set either pulse to None for 'splitter' behaviour\n\n B(pulse2)-----[=======]-----SignalA, tapB\n [==2x2==]\n A(pulse1)-----[=======]-----SignalB, tapA\n\n pulse1 goes to output_sig with (1-tap)\n pulse1 goes to output_tap with tap\n pulse2 goes to output_tap with tap\n pulse2 goes to output_sig with (1-tap)\n\n tap can be concidered output coupler value\n '''\n if pulse1 is None:\n At1 = 0\n else:\n At1 = np.sqrt(1-loss)*pulse1.At\n\n if pulse2 is None:\n At2 = 0\n else:\n At2 = np.sqrt(1-loss)*pulse2.At\n\n output_signal = np.sqrt(1-tap/100)*At1 + 1j*np.sqrt(tap/100)*At2\n output_tap = np.sqrt(1-tap/100)*At2 + 1j*np.sqrt(tap/100)*At1\n\n return output_signal, output_tap\n\n\ndef optical_filter(pulse, filter_type, lambda0=None, bandwidth=2E-9, loss=0, order=1):\n '''\n Simulate filter, bandpass, longpass, shortpass\n default bandwidth is 2nm\n\n pulse.lambda0 = central wavelength of PULSE\n lambda0 = central wavelength of FILTER\n w0 is central freq (ang) of FILTER\n '''\n \n Af = np.fft.fft(pulse.At)\n \n if lambda0 == None:\n lambda0 = pulse.lambda0\n\n w = pulse.freq + 2*np.pi*c/pulse.lambda0\n w0 = 2*np.pi*c/lambda0\n\n if filter_type.lower() == 'lpf':\n '''\n long-pass, pass low freq\n w0-w is (+) for ww0 (pass region)\n '''\n filter_profile = 0.5*(np.sign(w-w0) + 1)\n\n elif filter_type.lower() == 'bpf':\n '''\n bandpass\n '''\n dw = w0*(bandwidth/lambda0)\n bw=dw*(np.log(2)/2)\n #filter_profile = (0.5*(np.sign(w0-w+dw/2) + 1))*(0.5 * (np.sign(w-w0+dw/2) + 1))\n filter_profile = gaussian(w,bw,1,w0,sg=order)\n\n else:\n '''\n if no filter is specified, only losses are applied (filter is == 1 for all freq)\n '''\n filter_profile = np.ones(np.shape(w))\n \n output_At = np.sqrt(1-loss)*np.fft.ifft(Af*filter_profile)\n\n return output_At\n\n\ndef saturable_abs(pulse,sat_int,spot_size,mod_depth=1,loss=0):\n ''' Simulate saturable absorber.\n sat_int = saturation intensity, J/m**2\n ***NOTE energy density, NOT intensity***\n spot_size = beam diameter\n mod_depth = modulation depth, ratio e.g. 1% -> 0.01\n loss = non saturable losses\n\n small signal -> refl ~ 1-loss-mod_depth\n high signal --> refl ~ 1-loss\n '''\n intensity = pulse.dt*np.abs(pulse.At)**2/(np.pi*(spot_size/2)**2)\n outputField = np.sqrt(1-loss)*pulse.At*np.sqrt((1-mod_depth/(1+intensity/sat_int)))\n\n return outputField\n\n\ndef interp_gain_freq(pulse,cs):\n '''\n Interpolate crosssection data onto frequency array of pulse\n pulse = Pulse() object\n cs = crosssection data object\n '''\n p_wl = 2*pi*c/(pulse.freq + 2*pi*c/pulse.lambda0)\n cs_scale = (cs.emission - cs.absorption)/(np.max(cs.emission - cs.absorption))\n \n return np.interp(p_wl,cs.wavelength,cs_scale)\n\n\ndef shift_t0(pulse):\n '''Shift T0 to center of pulse\n '''\n t0 = pulse.time[np.argmax(pulse.getPt())]\n return np.interp(pulse.time,pulse.time-t0,pulse.At)\n\n\n\n\n\n", "sub_path": "beamtools/ultrafast_pulse_propagation.py", "file_name": "ultrafast_pulse_propagation.py", "file_ext": "py", "file_size_in_byte": 24152, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "numpy.arange", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 57, "usage_type": "attribute"}, {"api_name": "numpy.fft.fftfreq", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 57, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 64, "usage_type": "attribute"}, {"api_name": "numpy.fft.fft", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 64, "usage_type": "attribute"}, {"api_name": "numpy.abs", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.angle", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.arctan2", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.fft.fftshift", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 107, "usage_type": "attribute"}, {"api_name": "numpy.fft.ifft", "line_number": 107, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 146, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 172, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 180, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 225, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 233, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 234, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 236, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 244, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 263, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 263, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 263, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 264, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 271, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 279, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 280, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 280, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 281, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 282, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 283, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 284, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 284, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 285, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 286, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 286, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 288, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 288, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 295, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 305, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 310, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 332, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 335, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 337, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 340, "usage_type": "call"}, {"api_name": "numpy.gradient", "line_number": 342, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 360, "usage_type": "call"}, {"api_name": "numpy.diff", "line_number": 410, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 411, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 416, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 416, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 420, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 420, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 423, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 423, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 430, "usage_type": "call"}, {"api_name": "numpy.outer", "line_number": 431, "usage_type": "call"}, {"api_name": "numpy.math.factorial", "line_number": 441, "usage_type": "call"}, {"api_name": "numpy.math", "line_number": 441, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 447, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 447, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 448, "usage_type": "call"}, {"api_name": "numpy.fft.fft", "line_number": 452, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 452, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 453, "usage_type": "call"}, {"api_name": "numpy.fft.ifft", "line_number": 454, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 454, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 455, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 455, "usage_type": "call"}, {"api_name": "numpy.fft.fft", "line_number": 458, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 458, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 459, "usage_type": "call"}, {"api_name": "numpy.fft.ifft", "line_number": 460, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 460, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 461, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 461, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 488, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 489, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 491, "usage_type": "call"}, {"api_name": "beamtools.constants.c", "line_number": 515, "usage_type": "name"}, {"api_name": "beamtools.constants.c", "line_number": 516, "usage_type": "name"}, {"api_name": "beamtools.constants.h", "line_number": 518, "usage_type": "name"}, {"api_name": "beamtools.constants.h", "line_number": 519, "usage_type": "name"}, {"api_name": "beamtools.constants.h", "line_number": 520, "usage_type": "name"}, {"api_name": "beamtools.constants.h", "line_number": 521, "usage_type": "name"}, {"api_name": "beamtools.constants.c", "line_number": 524, "usage_type": "name"}, {"api_name": "numpy.pi", "line_number": 526, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 527, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 528, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 532, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 533, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 535, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 536, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 539, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 539, "usage_type": "call"}, {"api_name": "numpy.diff", "line_number": 541, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 546, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 546, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 550, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 551, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 561, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 561, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 562, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 562, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 563, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 563, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 564, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 564, "usage_type": "call"}, {"api_name": "beamtools.constants.h", "line_number": 568, "usage_type": "name"}, {"api_name": "beamtools.common.Func", "line_number": 570, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 572, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 572, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 584, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 584, "usage_type": "call"}, {"api_name": "numpy.size", "line_number": 588, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 589, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 590, "usage_type": "call"}, {"api_name": "numpy.flipud", "line_number": 593, "usage_type": "call"}, {"api_name": "beamtools.common.rk4", "line_number": 593, "usage_type": "call"}, {"api_name": "beamtools.common.rk4", "line_number": 595, "usage_type": "call"}, {"api_name": "beamtools.common.rk4", "line_number": 596, "usage_type": "call"}, {"api_name": "beamtools.common.rk4", "line_number": 597, "usage_type": "call"}, {"api_name": "numpy.flipud", "line_number": 598, "usage_type": "call"}, {"api_name": "beamtools.common.rk4", "line_number": 598, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 603, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 634, "usage_type": "attribute"}, {"api_name": "numpy.fft.fft", "line_number": 637, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 637, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 638, "usage_type": "attribute"}, {"api_name": "beamtools.constants.c", "line_number": 638, "usage_type": "name"}, {"api_name": "numpy.arcsin", "line_number": 640, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 640, "usage_type": "attribute"}, {"api_name": "beamtools.constants.c", "line_number": 640, "usage_type": "name"}, {"api_name": "numpy.sin", "line_number": 640, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 642, "usage_type": "attribute"}, {"api_name": "beamtools.constants.c", "line_number": 642, "usage_type": "name"}, {"api_name": "numpy.cos", "line_number": 642, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 643, "usage_type": "attribute"}, {"api_name": "beamtools.constants.c", "line_number": 643, "usage_type": "name"}, {"api_name": "numpy.sin", "line_number": 643, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 643, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 644, "usage_type": "attribute"}, {"api_name": "beamtools.constants.c", "line_number": 644, "usage_type": "name"}, {"api_name": "numpy.cos", "line_number": 644, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 646, "usage_type": "call"}, {"api_name": "numpy.fft.ifft", "line_number": 646, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 646, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 646, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 649, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 670, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 671, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 672, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 696, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 701, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 703, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 704, "usage_type": "call"}, {"api_name": "numpy.fft.fft", "line_number": 719, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 719, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 724, "usage_type": "attribute"}, {"api_name": "beamtools.constants.c", "line_number": 724, "usage_type": "name"}, {"api_name": "numpy.pi", "line_number": 725, "usage_type": "attribute"}, {"api_name": "beamtools.constants.c", "line_number": 725, "usage_type": "name"}, {"api_name": "numpy.sign", "line_number": 732, "usage_type": "call"}, {"api_name": "numpy.sign", "line_number": 739, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 746, "usage_type": "call"}, {"api_name": "beamtools.common.gaussian", "line_number": 748, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 754, "usage_type": "call"}, {"api_name": "numpy.shape", "line_number": 754, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 756, "usage_type": "call"}, {"api_name": "numpy.fft.ifft", "line_number": 756, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 756, "usage_type": "attribute"}, {"api_name": "numpy.abs", "line_number": 772, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 772, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 773, "usage_type": "call"}, {"api_name": "beamtools.constants.pi", "line_number": 784, "usage_type": "name"}, {"api_name": "beamtools.constants.c", "line_number": 784, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 785, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 787, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 793, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 794, "usage_type": "call"}]}
+{"seq_id": "472445280", "text": "#!/usr/bin/env python\n\nimport os, sys, glob, subprocess, textwrap\n\ntry:\n import setuptools\n assert int(setuptools.__version__.split(\".\", 1)[0]) >= 19\nexcept (ImportError, AssertionError):\n msg = 'Error: Aegea failed to install because your version of setuptools is too old ({}; 19 is required). Run \"make install_venv\" to install aegea in its own virtualenv, or upgrade your pip and setuptools to their latest versions.' # noqa\n exit(textwrap.fill(msg.format(setuptools.__version__)))\n\ntry:\n # Git version extraction logic designed to be compatible with both semver and PEP 440\n version = subprocess.check_output([\"git\", \"describe\", \"--tags\", \"--match\", \"v*.*.*\"]).decode()\n version = version.strip(\"v\\n\").replace(\"-\", \"+\", 1).replace(\"-\", \".\")\nexcept:\n version = \"0.0.0\"\n\nsetuptools.setup(\n name=\"aegea\",\n version=version,\n url=\"https://github.com/kislyuk/aegea\",\n license=open(\"LICENSE.md\").readline().strip(),\n author=\"Andrey Kislyuk\",\n author_email=\"kislyuk@gmail.com\",\n description=\"Amazon Web Services Operator Interface\",\n long_description=open(\"README.rst\").read(),\n install_requires=[\n \"boto3 >= 1.4.7\",\n \"botocore >= 1.8.0\",\n \"argcomplete >= 1.8.2, < 2\",\n \"paramiko >= 2.1.1, < 3\",\n \"requests >= 2.12.4, < 3\",\n \"tweak >= 0.4.0, < 1\",\n \"keymaker >= 0.3.3, < 1\",\n \"pyyaml >= 3.11, < 4\",\n \"python-dateutil >= 2.1, <2.7.0\",\n \"babel >= 2.3.4, < 3\",\n \"ipwhois >= 0.13.0, < 1\",\n \"uritemplate >= 3.0.0, < 4\",\n \"awscli >= 1.2.9\"\n ],\n extras_require={\n ':python_version == \"2.7\"': [\n \"enum34 >= 1.1.6, < 2\",\n \"ipaddress >= 1.0.17, < 2\",\n \"subprocess32 >= 3.2.7, < 4\"\n ]\n },\n tests_require=[\n \"coverage\",\n \"flake8\"\n ],\n packages=setuptools.find_packages(exclude=[\"test\"]),\n scripts=glob.glob(\"scripts/*\"),\n platforms=[\"MacOS X\", \"Posix\"],\n test_suite=\"test\",\n include_package_data=True\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 2020, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "setuptools.__version__.split", "line_number": 7, "usage_type": "call"}, {"api_name": "setuptools.__version__", "line_number": 7, "usage_type": "attribute"}, {"api_name": "textwrap.fill", "line_number": 10, "usage_type": "call"}, {"api_name": "setuptools.__version__", "line_number": 10, "usage_type": "attribute"}, {"api_name": "subprocess.check_output", "line_number": 14, "usage_type": "call"}, {"api_name": "setuptools.setup", "line_number": 19, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 54, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 55, "usage_type": "call"}]}
+{"seq_id": "65477082", "text": "# -*- coding: utf-8 -*-\n# @Time : 2019/9/16 14:17\nimport unittest\nimport base64\nimport requests\nfrom data import yamlutil\n\n\nclass Testcheckquery(unittest.TestCase):\n '''查黑服务接口'''\n def setUp(self):\n print(\"测试查黑服务接口开始\")\n\n # 查黑服务\n def testtwitter(self):\n e = yamlutil.getdata('checkquery.yaml')\n case = e[0]\n url = case.get('url')\n wd = case.get('word')\n word = base64.b64encode(wd.encode('utf-8'))\n data = case.get('data') + word.decode()\n print(data)\n response = requests.post(url=url, data=data, headers={'Connection': 'close'})\n print (response.text)\n self.assertEqual(response.status_code, 200, msg=\"测试通过\")\n\n def tearDown(self):\n print(\"测试查黑服务接口结束\")\n\nif __name__=='__main__':\n suite = unittest.TestSuite()\n suite.addTest(Testcheckquery(\"testtwitter\"))\n runner = unittest.TextTestRunner()\n runner.run(suite)", "sub_path": "Browser/case/checkquery.py", "file_name": "checkquery.py", "file_ext": "py", "file_size_in_byte": 990, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "unittest.TestCase", "line_number": 9, "usage_type": "attribute"}, {"api_name": "data.yamlutil.getdata", "line_number": 16, "usage_type": "call"}, {"api_name": "data.yamlutil", "line_number": 16, "usage_type": "name"}, {"api_name": "base64.b64encode", "line_number": 20, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 23, "usage_type": "call"}, {"api_name": "unittest.TestSuite", "line_number": 31, "usage_type": "call"}, {"api_name": "unittest.TextTestRunner", "line_number": 33, "usage_type": "call"}]}
+{"seq_id": "22435774", "text": "from PyQt5.QtWidgets import QMainWindow, QWidget, QLabel, QSizePolicy\nfrom PyQt5.QtCore import Qt, QMetaObject, QTimer, QRect\nfrom constant.path import Path\nfrom constant.default_string import DefaultString\nfrom utils.write_log import writeExecutionSteps, writeExceptionToFile\nfrom views.custom_style.label import RedBackgroundBottom, RedBackgroundTop, ScreenTitle, NormalText, ConnectionErrorText\nfrom constant.request_api_result import RequestAPIResult\nimport time\nfrom firebase import firebase\n\nclass TakeKeyView():\n \"\"\" Take key to the locker view\n \"\"\"\n \n TAG = 'TakeKeyView'\n\n def __init__(self, user, controller, main):\n self.__controller = controller\n self.__main = main\n self.__user = user\n\n writeExecutionSteps(self.TAG)\n\n self.central_widget = QWidget(main)\n self.central_widget.setFixedSize(self.__main.width, self.__main.height)\n\n self.firebase = firebase.FirebaseApplication(\"https://test-60f7d-default-rtdb.firebaseio.com/\", None)\n self.result = self.firebase.put('/Customer/' + self.__user.user_id , 'door_status', 'Open')\n\n red_background_top = RedBackgroundTop(Path.PATH_ICON + '/logo.png', self.central_widget)\n\n self.screen_title = ScreenTitle(self.central_widget)\n self.screen_title.setXY(0, int(self.__main.height * 0.15))\n self.screen_title.setHeight(int(self.__main.height * 0.3))\n self.screen_title.setTextSize(100)\n\n self.content = NormalText(self.central_widget)\n self.content.setXY(0, int(self.__main.height * 0.7))\n\n self.connection_error = ConnectionErrorText(self.central_widget)\n\n red_background_bottom = RedBackgroundBottom(self.central_widget)\n\n self.retranslateUI()\n QMetaObject.connectSlotsByName(self.central_widget)\n main.setCentralWidget(self.central_widget)\n \n self.__timer = QTimer()\n self.__timer.timeout.connect(self.__closeDoor)\n self.__timer.setSingleShot(True)\n self.__timer.start(2000)\n\n def retranslateUI(self):\n self.__default_string = DefaultString.getDefaultString()\n self.screen_title.setText(self.__default_string.TAKE_KEY)\n self.content.setText(self.__default_string.DOOR_OPENING)\n\n def connectionError(self):\n self.connection_error.setText(self.__default_string.NOT_CONNECT_INTERNET)\n\n def fail(self):\n self.connection_error.setText(self.__default_string.TRY_AGAIN)\n\n def stopComponentsRunning(self):\n self.__timer.stop()\n self.__controller.closeControlDoor()\n\n def backToWelcome(self):\n try:\n self.stopComponentsRunning()\n except Exception:\n writeExceptionToFile()\n finally:\n self.__main.onTransferScreen(screen=\"startScreenWelcome\")\n\n def __closeDoor(self):\n self.__count_timer_close_door = 15\n self.__timer_close_door = QTimer()\n self.__timer_close_door.timeout.connect(self.__updateCloseDoorTimeText)\n self.__timer_close_door.start(1000)\n\n def __updateCloseDoorTimeText(self):\n self.__count_timer_close_door -= 1\n\n second = self.__default_string.SECONDS\n if self.__count_timer_close_door == 1:\n second = self.__default_string.SECOND\n\n if self.__count_timer_close_door == 0:\n self.__timer_close_door.stop() \n self.content.setText(self.__default_string.SHUT_DOOR)\n \n self.__main.onTransferScreen(screen=\"startScreenThankYou\")\n \n else:\n self.content.setText(self.__default_string.DOOR_WILL_CLOSE \n + ' ' \n + str(self.__count_timer_close_door)\n + ' '\n + second)\n", "sub_path": "key-vending-copy1/app/views/take_key_view.py", "file_name": "take_key_view.py", "file_ext": "py", "file_size_in_byte": 3721, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "utils.write_log.writeExecutionSteps", "line_number": 22, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 24, "usage_type": "call"}, {"api_name": "firebase.firebase.FirebaseApplication", "line_number": 27, "usage_type": "call"}, {"api_name": "firebase.firebase", "line_number": 27, "usage_type": "name"}, {"api_name": "views.custom_style.label.RedBackgroundTop", "line_number": 30, "usage_type": "call"}, {"api_name": "constant.path.Path.PATH_ICON", "line_number": 30, "usage_type": "attribute"}, {"api_name": "constant.path.Path", "line_number": 30, "usage_type": "name"}, {"api_name": "views.custom_style.label.ScreenTitle", "line_number": 32, "usage_type": "call"}, {"api_name": "views.custom_style.label.NormalText", "line_number": 37, "usage_type": "call"}, {"api_name": "views.custom_style.label.ConnectionErrorText", "line_number": 40, "usage_type": "call"}, {"api_name": "views.custom_style.label.RedBackgroundBottom", "line_number": 42, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QMetaObject.connectSlotsByName", "line_number": 45, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QMetaObject", "line_number": 45, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QTimer", "line_number": 48, "usage_type": "call"}, {"api_name": "constant.default_string.DefaultString.getDefaultString", "line_number": 54, "usage_type": "call"}, {"api_name": "constant.default_string.DefaultString", "line_number": 54, "usage_type": "name"}, {"api_name": "utils.write_log.writeExceptionToFile", "line_number": 72, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QTimer", "line_number": 78, "usage_type": "call"}]}
+{"seq_id": "356382209", "text": "from collections.abc import Sized\nfrom datetime import datetime\nimport re\nfrom typing import NamedTuple, Set, Iterable, Dict, TypeVar, Callable, List, Optional\nimport logging\nfrom functools import lru_cache\n\nimport pytz\n\nfrom .normalise import normalise_url\n\nDate = datetime\nclass Visit(NamedTuple):\n dt: datetime\n tag: Optional[str] = None\n context: Optional[str] = None\n\nUrl = str\nclass Entry(NamedTuple):\n url: Url\n visits: Set[Visit]\n # TODO compare urls?\n\nFilter = Callable[[Url], bool]\n\ndef make_filter(thing) -> Filter:\n if isinstance(thing, str):\n rc = re.compile(thing)\n def filter_(u: str) -> bool:\n return rc.search(u) is not None\n return filter_\n else: # must be predicate\n return thing\n\n# TODO do i really need to inherit this??\nclass History(Sized):\n FILTERS: List[Filter] = [\n make_filter(f) for f in\n [\n r'^chrome-devtools://',\n r'^chrome-extension://',\n r'^chrome-error://',\n r'^chrome-native://',\n r'^chrome-search://',\n\n r'chrome://newtab',\n r'chrome://apps',\n r'chrome://history',\n\n r'^about:',\n r'^blob:',\n r'^view-source:',\n\n r'^content:',\n\n # TODO maybe file:// too?\n # chrome-search:\n ]\n ]\n\n @classmethod\n def add_filter(cls, filterish):\n cls.FILTERS.append(make_filter(filterish))\n\n def __init__(self):\n self.urls: Dict[Url, Entry] = {}\n\n @classmethod\n def from_urls(cls, urls: Dict[Url, Entry], filters: List[Filter] = None) -> 'History':\n hist = cls()\n hist.urls = urls\n return hist\n\n # TODO mm. maybe history should get filters from some global config?\n # wonder how okay is it to set class attribute..\n\n @classmethod\n def filtered(cls, url: Url) -> bool:\n for f in cls.FILTERS:\n if f(url):\n return True\n return False\n\n def register(self, url: Url, v: Visit) -> None:\n if History.filtered(url):\n return\n if v.dt.tzinfo is None:\n # TODO log that?...\n pass\n # TODO replace dt i\n\n # TODO hmm some filters make sense before stripping off protocol...\n # TODO is it a good place to normalise?\n url = normalise_url(url)\n\n e = self.urls.get(url, None)\n if e is None:\n e = Entry(url=url, visits=set())\n e.visits.add(v)\n self.urls[url] = e\n\n def __len__(self) -> int:\n return len(self.urls)\n\n def __getitem__(self, url: Url) -> Entry:\n return self.urls[url]\n\n def items(self):\n return self.urls.items()\n\n def __repr__(self):\n return 'History{' + repr(self.urls) + '}'\n\ndef simple_history(urls: List[Url], tag: str) -> History:\n h = History()\n for u in urls:\n ts = datetime.utcnow().replace(tzinfo=pytz.utc)\n visit = Visit(\n ts,\n tag=tag,\n )\n h.register(u, visit)\n return h\n\n# f is value merger function\n_K = TypeVar(\"_K\")\n_V = TypeVar(\"_V\")\n\ndef merge_dicts(f: Callable[[_V, _V], _V], dicts: Iterable[Dict[_K, _V]]):\n res: Dict[_K, _V] = {}\n for d in dicts:\n for k, v in d.items():\n if k not in res:\n res[k] = v\n else:\n res[k] = f(res[k], v)\n return res\n\ndef entry_merger(a: Entry, b: Entry):\n a.visits.update(b.visits)\n return a\n\ndef merge_histories(hists: Iterable[History]) -> History:\n return History.from_urls(merge_dicts(entry_merger, [h.urls for h in hists]))\n\ndef get_logger():\n return logging.getLogger(\"WereYouHere\")\n\n\n# kinda singleton\n@lru_cache()\ndef get_tmpdir():\n import tempfile\n tdir = tempfile.TemporaryDirectory(suffix=\"wereyouhere\")\n return tdir\n", "sub_path": "wereyouhere/common.py", "file_name": "common.py", "file_ext": "py", "file_size_in_byte": 3835, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "datetime.datetime", "line_number": 12, "usage_type": "name"}, {"api_name": "typing.NamedTuple", "line_number": 13, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 14, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 15, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 16, "usage_type": "name"}, {"api_name": "typing.NamedTuple", "line_number": 19, "usage_type": "name"}, {"api_name": "typing.Set", "line_number": 21, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 24, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 28, "usage_type": "call"}, {"api_name": "collections.abc.Sized", "line_number": 36, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 37, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 66, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 69, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 69, "usage_type": "name"}, {"api_name": "normalise.normalise_url", "line_number": 94, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 114, "usage_type": "name"}, {"api_name": "datetime.datetime.utcnow", "line_number": 117, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 117, "usage_type": "name"}, {"api_name": "pytz.utc", "line_number": 117, "usage_type": "attribute"}, {"api_name": "typing.TypeVar", "line_number": 126, "usage_type": "call"}, {"api_name": "typing.TypeVar", "line_number": 127, "usage_type": "call"}, {"api_name": "typing.Callable", "line_number": 129, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 129, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 129, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 130, "usage_type": "name"}, {"api_name": "typing.Iterable", "line_number": 143, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 147, "usage_type": "call"}, {"api_name": "tempfile.TemporaryDirectory", "line_number": 154, "usage_type": "call"}, {"api_name": "functools.lru_cache", "line_number": 151, "usage_type": "call"}]}
+{"seq_id": "617465511", "text": "import sqlite3\r\n\r\nclass BancoDeDados:\r\n\t\"\"\"Classe que representa o banco de dados (database) da aplicação\"\"\"\r\n\r\n\tdef __init__(self, nome='banco.db'):\r\n\t\tself.nome, self.conexao = nome, None\r\n\r\n\tdef conecta(self):\r\n\t\t\"\"\"Conecta passando o nome do arquivo\"\"\"\r\n\t\tself.conexao = sqlite3.connect(self.nome)\r\n\r\n\tdef desconecta(self):\r\n\t\t\"\"\"Desconecta do banco\"\"\"\r\n\t\ttry:\r\n\t\t\tself.conexao.close()\r\n\t\texcept AttributeError:\r\n\t\t\tpass\r\n\r\n\tdef criar_tabelas(self):\r\n\t\t\"\"\"Cria as tabelas do banco\"\"\"\r\n\t\ttry:\r\n\t\t\tcursor = self.conexao.cursor()\r\n\r\n\t\t\tcursor.execute(\"\"\"\r\n\t\t\tCREATE TABLE IF NOT EXISTS clientes (\r\n\t\t\t\t\tid INTEGER NOT NULL PRIMARY KEY AUTOINCREMENT,\r\n\t\t\t\t\tnome TEXT NOT NULL,\r\n\t\t\t\t\tcpf VARCHAR(11) UNIQUE NOT NULL,\r\n\t\t\t\t\temail TEXT NOT NULL\r\n\t\t\t);\r\n\t\t\t\"\"\")\r\n\r\n\t\texcept AttributeError:\r\n\t\t\tprint('Faça a conexão do banco antes de criar as tabelas.')\r\n\r\n\tdef inserir_cliente(self, nome, cpf, email):\r\n\t\t\"\"\"Insere cliente no banco\"\"\"\r\n\t\ttry:\r\n\t\t\tcursor = self.conexao.cursor()\r\n\r\n\t\t\ttry:\r\n\t\t\t\tcursor.execute(\"\"\"\r\n\t\t\t\t\tINSERT INTO clientes (nome, cpf, email) VALUES (?,?,?)\r\n\t\t\t\t\"\"\", (nome, cpf, email))\r\n\t\t\texcept sqlite3.IntegrityError:\r\n\t\t\t\tprint('O cpf %s já existe!' % cpf)\r\n\r\n\t\t\tself.conexao.commit()\r\n\r\n\t\texcept AttributeError:\r\n\t\t\tprint('Faça a conexão do banco antes de inserir clientes.')\r\n\r\n\tdef buscar_cliente(self, cpf):\r\n\t\t\"\"\"Busca um cliente pelo cpf\"\"\"\r\n\t\ttry:\r\n\t\t\tcursor = self.conexao.cursor()\r\n\r\n\t\t\t# Tentando EAFP!!\r\n\t\t\ttry:\r\n\t\t\t\tcursor.execute(\"\"\"SELECT nome FROM clientes WHERE cpf =\"\"\" + cpf)\r\n\t\t\t\tvar_nome = cursor.fetchone()\r\n\t\t\t\tif var_nome is None:\r\n\t\t\t\t\tprint('Cliente não encontrado. Verifique se o CPF foi inserido corretamente!')\r\n\t\t\t\telse:\r\n\t\t\t\t\tprint('Cliente %s encontrado.' % var_nome)\r\n\t\t\texcept:\r\n\t\t\t\tpass\r\n\t\texcept AttributeError:\r\n\t\t\tprint('Faça a conexão do banco antes de buscar clientes.')\r\n\r\n\tdef remover_cliente(self, cpf):\r\n\t\t\"\"\"Removendo cliente a partir do CPF\"\"\"\r\n\t\t# EAFP\r\n\t\ttry:\r\n\t\t\tcursor = self.conexao.cursor()\r\n\t\t\t# EAFP\r\n\t\t\ttry:\r\n\t\t\t\tcursor.execute(\"\"\"SELECT nome FROM clientes WHERE cpf =\"\"\" + cpf)\r\n\t\t\t\tvar_nome = cursor.fetchone()\r\n\t\t\t\t# LBYL\r\n\t\t\t\tif var_nome is None:\r\n\t\t\t\t\tprint('Cliente não encontrado. Verifique se o CPF foi inserido corretamente!')\r\n\t\t\t\t\t# break\r\n\t\t\t\telse:\r\n\t\t\t\t\tcursor.execute(\"\"\"DELETE FROM clientes WHERE cpf =\"%s\" \"\"\" % cpf)\r\n\t\t\t\t\tprint('Cliente %s foi removido com suscesso!' % var_nome)\r\n\r\n\t\t\texcept:\r\n\t\t\t\tpass\r\n\r\n\t\t\tself.conexao.commit()\r\n\t\texcept AttributeError:\r\n\t\t\tprint('Faça a conexão do banco antes de buscar clientes.')\r\n\r\n\tdef buscar_email(self, email):\r\n\t\t\"\"\"Buscando cliente por email e retornando verdadeiro (True) ou falso (False)\"\"\"\r\n\t\t# EAFP\r\n\t\ttry:\r\n\t\t\tcursor = self.conexao.cursor()\r\n\t\t\t# EAFP\r\n\t\t\ttry:\r\n\t\t\t\tcursor.execute(\"\"\"SELECT nome FROM clientes WHERE email=\"%s\" \"\"\" % email)\r\n\t\t\t\tvar_nome = cursor.fetchone()\r\n\t\t\t\t# LBYL\r\n\t\t\t\tif var_nome is None:\r\n\t\t\t\t\tprint('False! E-mail não encontrado!')\r\n\t\t\t\telse:\r\n\t\t\t\t\tprint('True! E-mail referente ao cliente %s!' % var_nome)\r\n\t\t\texcept:\r\n\t\t\t\tpass\r\n\r\n\t\texcept AttributeError:\r\n\t\t\tprint('Faça a conexão do banco antes de buscar clientes.')\r\n", "sub_path": "ProjetoSQlite_BrunoCayres/database.py", "file_name": "database.py", "file_ext": "py", "file_size_in_byte": 3104, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "sqlite3.connect", "line_number": 11, "usage_type": "call"}, {"api_name": "sqlite3.IntegrityError", "line_number": 46, "usage_type": "attribute"}]}
+{"seq_id": "36171611", "text": "# https://mannekentech.com/2017/01/02/playing-with-a-cassandra-cluster-via-docker/\nimport logging\nfrom cassandra import ConsistencyLevel\nfrom cassandra.cluster import Cluster\nfrom cassandra.query import SimpleStatement\n\nlog = logging.getLogger()\nlog.setLevel('INFO')\nhandler = logging.StreamHandler()\nhandler.setFormatter(logging.Formatter(\"%(asctime)s [%(levelname)s] %(name)s: %(message)s\"))\nlog.addHandler(handler)\n\n\nKEYSPACE = \"mykeyspace\"\n\ndef createKeySpace():\n cluster = Cluster(contact_points=['127.0.0.1'],port=9142)\n session = cluster.connect()\n\n log.info(\"Creating keyspace...\")\n try:\n session.execute(\"\"\"\n CREATE KEYSPACE %s\n WITH replication = { 'class': 'SimpleStrategy', 'replication_factor': '2' }\n \"\"\" % KEYSPACE)\n\n log.info(\"setting keyspace...\")\n session.set_keyspace(KEYSPACE)\n\n log.info(\"creating table...\")\n session.execute(\"\"\"\n CREATE TABLE mytable (\n mykey text,\n col1 text,\n col2 text,\n PRIMARY KEY (mykey, col1)\n )\n \"\"\")\n except Exception as e:\n log.error(\"Unable to create keyspace\")\n log.error(e)\n\ncreateKeySpace();\n\n\ndef insertData(number):\n cluster = Cluster(contact_points=['127.0.0.1'],port=9142)\n session = cluster.connect()\n\n log.info(\"setting keyspace...\")\n session.set_keyspace(KEYSPACE)\n\n prepared = session.prepare(\"\"\"\n INSERT INTO mytable (mykey, col1, col2)\n VALUES (?, ?, ?)\n \"\"\")\n\n for i in range(number):\n if(i%100 == 0):\n log.info(\"inserting row %d\" % i)\n session.execute(prepared.bind((\"rec_key_%d\" % i, 'aaa', 'bbb')))\n\ninsertData(1000)\n\n\ndef readRows():\n cluster = Cluster(contact_points=['127.0.0.1'],port=9142)\n session = cluster.connect()\n\n log.info(\"setting keyspace...\")\n session.set_keyspace(KEYSPACE)\n\n rows = session.execute(\"SELECT * FROM mytable\")\n log.info(\"key\\tcol1\\tcol2\")\n log.info(\"---------\\t----\\t----\")\n\n res = []\n\n count=0\n for row in rows:\n if(count%100==0):\n log.info('\\t'.join(row))\n count=count+1;\n res.append(row)\n\n log.info(\"Total\")\n log.info(\"-----\")\n log.info(\"rows %d\" %(count))\n\n return res\n\n\nrows = readRows()\n", "sub_path": "scratchpad/cass/cass.py", "file_name": "cass.py", "file_ext": "py", "file_size_in_byte": 2304, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 9, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 10, "usage_type": "call"}, {"api_name": "cassandra.cluster.Cluster", "line_number": 17, "usage_type": "call"}, {"api_name": "cassandra.cluster.Cluster", "line_number": 47, "usage_type": "call"}, {"api_name": "cassandra.cluster.Cluster", "line_number": 67, "usage_type": "call"}]}
+{"seq_id": "157301760", "text": "from collections import OrderedDict\n\nfrom django.utils.translation import ugettext as _\n\nfrom drf_extra_fields.fields import Base64FieldMixin\nfrom drf_yasg import openapi\nfrom drf_yasg.inspectors import (\n CamelCaseJSONFilter,\n FieldInspector,\n NotHandled,\n ViewInspector,\n)\nfrom drf_yasg.utils import filter_none, get_serializer_ref_name\nfrom rest_framework import serializers\n\n\nclass FileFieldInspector(CamelCaseJSONFilter):\n def get_schema(self, serializer):\n if self.method not in ViewInspector.body_methods:\n return NotHandled\n\n # only do this if there are base64 mixin fields\n if any(\n isinstance(field, Base64FieldMixin) for field in serializer.fields.values()\n ):\n return self.probe_field_inspectors(serializer, openapi.Schema, True)\n\n return NotHandled\n\n def field_to_swagger_object(\n self, field, swagger_object_type, use_references, **kwargs\n ):\n if isinstance(field, serializers.Serializer):\n return self._serializer_to_swagger_object(\n field, swagger_object_type, use_references, **kwargs\n )\n\n if not isinstance(field, Base64FieldMixin):\n return NotHandled\n\n SwaggerType, ChildSwaggerType = self._get_partial_types(\n field, swagger_object_type, use_references, **kwargs\n )\n\n type_b64 = SwaggerType(\n type=openapi.TYPE_STRING,\n format=openapi.FORMAT_BASE64,\n description=_(\"Base64 encoded binary content.\"),\n )\n type_uri = SwaggerType(\n type=openapi.TYPE_STRING,\n read_only=True,\n format=openapi.FORMAT_URI,\n description=_(\"Download URL of the binary content.\"),\n )\n\n if swagger_object_type == openapi.Schema:\n # on writes, it's always b64\n if self.method in ViewInspector.body_methods:\n return type_b64\n\n # if not representing in base64, it's a link\n return type_uri if not field.represent_in_base64 else type_b64\n\n return NotHandled\n\n def _serializer_to_swagger_object(\n self, serializer, swagger_object_type, use_references, **kwargs\n ):\n if self.method not in ViewInspector.body_methods:\n return NotHandled\n\n if not any(\n isinstance(field, Base64FieldMixin) for field in serializer.fields.values()\n ):\n return NotHandled\n\n SwaggerType, ChildSwaggerType = self._get_partial_types(\n serializer, swagger_object_type, use_references, **kwargs\n )\n\n ref_name = get_serializer_ref_name(serializer)\n ref_name = f\"{ref_name}Data\" if ref_name else None\n\n def make_schema_definition():\n properties = OrderedDict()\n required = []\n for property_name, child in serializer.fields.items():\n prop_kwargs = {\"read_only\": bool(child.read_only) or None}\n prop_kwargs = filter_none(prop_kwargs)\n\n child_schema = self.probe_field_inspectors(\n child, ChildSwaggerType, use_references, **prop_kwargs\n )\n properties[property_name] = child_schema\n\n if child.required and not getattr(child_schema, \"read_only\", False):\n required.append(property_name)\n\n result = SwaggerType(\n type=openapi.TYPE_OBJECT,\n properties=properties,\n required=required or None,\n )\n if not ref_name and \"title\" in result:\n # on an inline model, the title is derived from the field name\n # but is visually displayed like the model name, which is confusing\n # it is better to just remove title from inline models\n del result.title\n\n # Provide an option to add manual paremeters to a schema\n # for example, to add examples\n # self.add_manual_fields(serializer, result)\n return self.process_result(result, None, None)\n\n if not ref_name or not use_references:\n return make_schema_definition()\n\n definitions = self.components.with_scope(openapi.SCHEMA_DEFINITIONS)\n definitions.setdefault(ref_name, make_schema_definition)\n return openapi.SchemaRef(definitions, ref_name)\n", "sub_path": "vng_api_common/inspectors/files.py", "file_name": "files.py", "file_ext": "py", "file_size_in_byte": 4386, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "drf_yasg.inspectors.CamelCaseJSONFilter", "line_number": 17, "usage_type": "name"}, {"api_name": "drf_yasg.inspectors.ViewInspector.body_methods", "line_number": 19, "usage_type": "attribute"}, {"api_name": "drf_yasg.inspectors.ViewInspector", "line_number": 19, "usage_type": "name"}, {"api_name": "drf_yasg.inspectors.NotHandled", "line_number": 20, "usage_type": "name"}, {"api_name": "drf_extra_fields.fields.Base64FieldMixin", "line_number": 24, "usage_type": "argument"}, {"api_name": "drf_yasg.openapi.Schema", "line_number": 26, "usage_type": "attribute"}, {"api_name": "drf_yasg.openapi", "line_number": 26, "usage_type": "name"}, {"api_name": "drf_yasg.inspectors.NotHandled", "line_number": 28, "usage_type": "name"}, {"api_name": "rest_framework.serializers.Serializer", "line_number": 33, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 33, "usage_type": "name"}, {"api_name": "drf_extra_fields.fields.Base64FieldMixin", "line_number": 38, "usage_type": "argument"}, {"api_name": "drf_yasg.inspectors.NotHandled", "line_number": 39, "usage_type": "name"}, {"api_name": "drf_yasg.openapi.TYPE_STRING", "line_number": 46, "usage_type": "attribute"}, {"api_name": "drf_yasg.openapi", "line_number": 46, "usage_type": "name"}, {"api_name": "drf_yasg.openapi.FORMAT_BASE64", "line_number": 47, "usage_type": "attribute"}, {"api_name": "drf_yasg.openapi", "line_number": 47, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext", "line_number": 48, "usage_type": "call"}, {"api_name": "drf_yasg.openapi.TYPE_STRING", "line_number": 51, "usage_type": "attribute"}, {"api_name": "drf_yasg.openapi", "line_number": 51, "usage_type": "name"}, {"api_name": "drf_yasg.openapi.FORMAT_URI", "line_number": 53, "usage_type": "attribute"}, {"api_name": "drf_yasg.openapi", "line_number": 53, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext", "line_number": 54, "usage_type": "call"}, {"api_name": "drf_yasg.openapi.Schema", "line_number": 57, "usage_type": "attribute"}, {"api_name": "drf_yasg.openapi", "line_number": 57, "usage_type": "name"}, {"api_name": "drf_yasg.inspectors.ViewInspector.body_methods", "line_number": 59, "usage_type": "attribute"}, {"api_name": "drf_yasg.inspectors.ViewInspector", "line_number": 59, "usage_type": "name"}, {"api_name": "drf_yasg.inspectors.NotHandled", "line_number": 65, "usage_type": "name"}, {"api_name": "drf_yasg.inspectors.ViewInspector.body_methods", "line_number": 70, "usage_type": "attribute"}, {"api_name": "drf_yasg.inspectors.ViewInspector", "line_number": 70, "usage_type": "name"}, {"api_name": "drf_yasg.inspectors.NotHandled", "line_number": 71, "usage_type": "name"}, {"api_name": "drf_extra_fields.fields.Base64FieldMixin", "line_number": 74, "usage_type": "argument"}, {"api_name": "drf_yasg.inspectors.NotHandled", "line_number": 76, "usage_type": "name"}, {"api_name": "drf_yasg.utils.get_serializer_ref_name", "line_number": 82, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 86, "usage_type": "call"}, {"api_name": "drf_yasg.utils.filter_none", "line_number": 90, "usage_type": "call"}, {"api_name": "drf_yasg.openapi.TYPE_OBJECT", "line_number": 101, "usage_type": "attribute"}, {"api_name": "drf_yasg.openapi", "line_number": 101, "usage_type": "name"}, {"api_name": "drf_yasg.openapi.SCHEMA_DEFINITIONS", "line_number": 119, "usage_type": "attribute"}, {"api_name": "drf_yasg.openapi", "line_number": 119, "usage_type": "name"}, {"api_name": "drf_yasg.openapi.SchemaRef", "line_number": 121, "usage_type": "call"}, {"api_name": "drf_yasg.openapi", "line_number": 121, "usage_type": "name"}]}
+{"seq_id": "207819658", "text": "from os import path\nimport threading\nimport requests\nimport time\n\nfrom lxml import html\nfrom utils.text import millify\nfrom utils.text import trim_string\nfrom utils.logger import logger\n\n\nclass MotdUpdater(threading.Thread):\n\n def __init__(self, server, scoreboard_type):\n self.server = server\n\n self.scoreboard_type = scoreboard_type\n self.time_interval = 5 * 60\n self.motd = self.load_motd()\n\n threading.Thread.__init__(self)\n\n def run(self):\n while True:\n self.server.write_all_players()\n try:\n motd_payload = self.get_configuration()\n except requests.exceptions.RequestException:\n continue\n\n motd = self.render_motd(self.motd)\n motd_payload['ServerMOTD'] = motd.encode(\"iso-8859-1\", \"ignore\")\n\n try:\n self.submit_motd(motd_payload)\n except requests.exceptions.RequestException:\n continue\n\n time.sleep(self.time_interval)\n\n def submit_motd(self, payload):\n motd_url = \"http://\" + self.server.address + \\\n \"/ServerAdmin/settings/welcome\"\n\n logger.debug(\"Updating MOTD ({})\".format(self.server.name))\n try:\n self.server.session.post(motd_url, data=payload)\n self.server.save_settings()\n except requests.exceptions.RequestException:\n logger.warning(\"Couldn't submit motd (RequestException) to {}\"\n .format(self.server.name))\n raise\n\n def load_motd(self):\n if not path.exists(self.server.name + \".motd\"):\n logger.warning(\"No motd file for \" + self.server.name)\n return \"\"\n\n motd_f = open(self.server.name + \".motd\")\n motd = motd_f.read()\n motd_f.close()\n return motd\n\n def render_motd(self, src_motd):\n # Wouldn't this be better to do with something like fuzzy?\n if self.scoreboard_type in ['kills', 'Kills', 'kill', 'Kill']:\n scores = self.server.database.top_kills()\n elif self.scoreboard_type in ['Dosh','dosh']:\n scores = self.server.database.top_dosh()\n else:\n logger.error(\"Bad configuration, scoreboard_type. \"\n \"Options are: dosh, kills ({})\"\n .format(self.server.name))\n return\n\n for player in scores:\n name = player[0].replace(\"<\", \"<\")\n name = trim_string(name, 12)\n score = player[1]\n\n src_motd = src_motd.replace(\"%PLR\", name, 1)\n src_motd = src_motd.replace(\"%SCR\", millify(score), 1)\n\n if \"%SRV_K\" in src_motd:\n server_kills = self.server.database.server_kills()\n src_motd = src_motd.replace(\"%SRV_K\", millify(server_kills), 1)\n\n if \"%SRV_D\" in src_motd:\n server_dosh = self.server.database.server_dosh()\n src_motd = src_motd.replace(\"%SRV_D\", millify(server_dosh), 1)\n\n return src_motd\n\n def get_configuration(self):\n motd_url = \"http://\" + self.server.address + \\\n \"/ServerAdmin/settings/welcome\"\n\n try:\n motd_response = self.server.session.get(motd_url, timeout=2)\n except requests.exceptions.RequestException as e:\n logger.debug(\"Couldn't get motd config(RequestException)\")\n raise\n\n motd_tree = html.fromstring(motd_response.content)\n\n banner_link = motd_tree.xpath('//input[@name=\"BannerLink\"]/@value')[0]\n web_link = motd_tree.xpath('//input[@name=\"WebLink\"]/@value')[0]\n\n return {\n 'BannerLink': banner_link,\n 'ClanMotto': '',\n 'ClanMottoColor': '#FF0000',\n 'ServerMOTDColor': '#FF0000',\n 'WebLink': web_link,\n 'WebLinkColor': '#FF0000',\n 'liveAdjust': '1',\n 'action': 'save'\n }\n", "sub_path": "magicked_admin/server/managers/motd_updater.py", "file_name": "motd_updater.py", "file_ext": "py", "file_size_in_byte": 3959, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "threading.Thread", "line_number": 12, "usage_type": "attribute"}, {"api_name": "threading.Thread.__init__", "line_number": 21, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 21, "usage_type": "attribute"}, {"api_name": "requests.exceptions", "line_number": 28, "usage_type": "attribute"}, {"api_name": "requests.exceptions", "line_number": 36, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 39, "usage_type": "call"}, {"api_name": "utils.logger.logger.debug", "line_number": 45, "usage_type": "call"}, {"api_name": "utils.logger.logger", "line_number": 45, "usage_type": "name"}, {"api_name": "requests.exceptions", "line_number": 49, "usage_type": "attribute"}, {"api_name": "utils.logger.logger.warning", "line_number": 50, "usage_type": "call"}, {"api_name": "utils.logger.logger", "line_number": 50, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "name"}, {"api_name": "utils.logger.logger.warning", "line_number": 56, "usage_type": "call"}, {"api_name": "utils.logger.logger", "line_number": 56, "usage_type": "name"}, {"api_name": "utils.logger.logger.error", "line_number": 71, "usage_type": "call"}, {"api_name": "utils.logger.logger", "line_number": 71, "usage_type": "name"}, {"api_name": "utils.text.trim_string", "line_number": 78, "usage_type": "call"}, {"api_name": "utils.text.millify", "line_number": 82, "usage_type": "call"}, {"api_name": "utils.text.millify", "line_number": 86, "usage_type": "call"}, {"api_name": "utils.text.millify", "line_number": 90, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 100, "usage_type": "attribute"}, {"api_name": "utils.logger.logger.debug", "line_number": 101, "usage_type": "call"}, {"api_name": "utils.logger.logger", "line_number": 101, "usage_type": "name"}, {"api_name": "lxml.html.fromstring", "line_number": 104, "usage_type": "call"}, {"api_name": "lxml.html", "line_number": 104, "usage_type": "name"}]}
+{"seq_id": "22968575", "text": "# -*- coding: utf-8 -*-\n# @Author: chunyang.xu\n# @Date: 2022-05-20 18:59:04\n# @Last Modified by: longfengpili\n# @Last Modified time: 2023-06-07 15:10:10\n\nimport sys\nimport openpyxl\n\nimport logging\nelogger = logging.getLogger(__name__)\n\n\nclass ReadDataFromExcel(object):\n\n def __init__(self, filepath: str, data_only: bool = True):\n \"\"\"[summary]\n \n [description]\n \n Arguments:\n filepath {str} -- [文件地址]\n \n Keyword Arguments:\n data_only {bool} -- [是否读取公式结果] (default: {True})\n \"\"\"\n \n self.filepath = filepath\n self.data_only = data_only\n\n def open_excel(self):\n try:\n book = openpyxl.load_workbook(self.filepath, data_only=self.data_only)\n return book\n except Exception as e:\n elogger.error(f\"{e}\")\n sys.exit(0)\n\n def get_sheets(self):\n book = self.open_excel()\n sheetnames = book.sheetnames\n sheets = dict([(sheetname, book[sheetname]) for sheetname in sheetnames])\n return sheets\n\n def open_sheet(self, sheetname: str):\n book = self.open_excel()\n sheet = book[sheetname]\n return sheet\n\n def get_sheetvalues_by_rows(self, sheetname: str):\n \"\"\"[summary]\n \n [以行的模式获取sheetvalues]\n \n Arguments:\n sheetname {str} -- [sheetname]\n \n Returns:\n [list] -- [row value list]\n \"\"\"\n\n sheet = self.open_sheet(sheetname)\n srow_values = [row for row in sheet.iter_rows(values_only=True)]\n return srow_values\n\n def get_sheetvalues_by_cols(self, sheetname: str):\n \"\"\"[summary]\n \n [以列的模式获取sheetvalues]\n \n Arguments:\n sheetname {str} -- [sheetname]\n \n Keyword Arguments:\n col_num {int} -- [获取的列数] (default: {None})\n \n Returns:\n [list] -- [col value list]\n \"\"\"\n\n sheet = self.open_sheet(sheetname)\n scol_values = [col for col in sheet.iter_cols(values_only=True)]\n return scol_values\n\n def get_sheet_values_by_header(self, sheetname, headers, header_row=0):\n \"\"\"[summary]\n\n [根据列名选择]\n\n Arguments:\n sheetname {[str]} -- [表格名]\n headers {[list]} -- [要选择的列名]\n\n Keyword Arguments:\n header_row {number} -- [header所在的row] (default: {0})\n\n Returns:\n [type] -- [description]\n \"\"\"\n \n srow_values = self.get_sheetvalues_by_rows(sheetname)\n srow_headers = srow_values[header_row]\n headers_index = [srow_headers.index(header) for header in headers]\n\n scol_values = self.get_sheetvalues_by_cols(sheetname)\n headers_values = [scol_values[hid] for hid in headers_index]\n\n headers_values = list(zip(*headers_values))\n return headers_values\n\n\nif __name__ == '__main__':\n filepath = './test.xlsx'\n rexcel = ReadDataFromExcel(filepath)\n srow_values = rexcel.get_sheetvalues_by_rows('first')\n print(srow_values)\n\n scol_values = rexcel.get_sheetvalues_by_rows('second')\n print(scol_values)", "sub_path": "excel_api/read_excel_new.py", "file_name": "read_excel_new.py", "file_ext": "py", "file_size_in_byte": 3255, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "openpyxl.load_workbook", "line_number": 33, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 37, "usage_type": "call"}]}
+{"seq_id": "366829104", "text": "from PIL import Image\nimport os\nfrom os import path\nimport random\nimport torch\nimport torch.nn.functional as F\nimport torchvision.transforms.functional as TF\nimport io\nimport json\nimport argparse\nimport time\n\n\nTEST_IMAGE = path.join(path.dirname(__file__), \"images\", \"donut_noisy.png\")\nOUTPUT_DIR = path.relpath(path.join(path.dirname(__file__), \"..\", \"..\",\n \"tmp\", \"search_qtable\"))\n\n\nCHECKBOARD_KERNEL = torch.tensor([\n [1, -1, 1, -1, 1, -1, 1, -1],\n [-1, 1, -1, 1, -1, 1, -1, 1],\n [1, -1, 1, -1, 1, -1, 1, -1],\n [-1, 1, -1, 1, -1, 1, -1, 1],\n [1, -1, 1, -1, 1, -1, 1, -1],\n [-1, 1, -1, 1, -1, 1, -1, 1],\n [1, -1, 1, -1, 1, -1, 1, -1],\n [-1, 1, -1, 1, -1, 1, -1, 1],\n], dtype=torch.float32).reshape(1, 1, 8, 8)\n\n# v/h could be reversed\nVERTICAL_LINE_KERNEL = torch.tensor([\n [0, 0, 0.5, -1, 0.5, 0, 0, 0],\n [0, 0, 0.5, -1, 0.5, 0, 0, 0],\n [0, 0, 0.5, -1, 0.5, 0, 0, 0],\n [0, 0, 0.5, -1, 0.5, 0, 0, 0],\n [0, 0, 0.5, -1, 0.5, 0, 0, 0],\n [0, 0, 0.5, -1, 0.5, 0, 0, 0],\n [0, 0, 0.5, -1, 0.5, 0, 0, 0],\n [0, 0, 0.5, -1, 0.5, 0, 0, 0],\n], dtype=torch.float32).reshape(1, 1, 8, 8)\nHORIZONTAL_LINE_KERNEL = torch.tensor([\n [0, 0, 0, 0, 0, 0, 0, 0],\n [0, 0, 0, 0, 0, 0, 0, 0],\n [0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5],\n [-1, -1, -1, -1, -1, -1, -1, -1],\n [0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5],\n [0, 0, 0, 0, 0, 0, 0, 0],\n [0, 0, 0, 0, 0, 0, 0, 0],\n [0, 0, 0, 0, 0, 0, 0, 0],\n], dtype=torch.float32).reshape(1, 1, 8, 8)\n\n\nZIGZAG_SCAN_INDEX = [\n 0, 1, 8, 16, 9, 2, 3, 10,\n 17, 24, 32, 25, 18, 11, 4,\n 5, 12, 19, 26, 33, 40, 48,\n 41, 34, 27, 20, 13, 6, 7,\n 14, 21, 28, 35, 42, 49, 56,\n 57, 50, 43, 36, 29, 22, 15,\n 23, 30, 37, 44, 51, 58, 59,\n 52, 45, 38, 31, 39, 46, 53,\n 60, 61, 54, 47, 55, 62, 63]\n\n\nFIXED_UV_QTABLE = [\n 7, 7, 10, 19, 32, 32, 32, 32,\n 7, 8, 10, 26, 32, 32, 32, 32,\n 10, 10, 22, 32, 32, 32, 32, 32,\n 19, 26, 32, 32, 32, 32, 32, 32,\n 32, 32, 32, 32, 32, 32, 32, 32,\n 32, 32, 32, 32, 32, 32, 32, 32,\n 32, 32, 32, 32, 32, 32, 32, 32,\n 32, 32, 32, 32, 32, 32, 32, 32,\n]\n\n\ndef calc_score(im, qtable, uv_qtable, kernels, criterion):\n with io.BytesIO() as buff:\n im.save(buff, format=\"jpeg\",\n qtables={0: qtable, 1: uv_qtable}, subsampling=\"4:2:0\")\n buff.seek(0)\n im = Image.open(buff)\n x = TF.to_tensor(TF.to_grayscale(im)).unsqueeze(0)\n score = 0\n for kernel in kernels:\n if criterion == \"max\":\n score += F.adaptive_max_pool2d(\n F.conv2d(x, weight=kernel, stride=1, padding=0),\n (16, 16)\n ).mean().item()\n else:\n # score += F.conv2d(x, weight=kernel, stride=1, padding=0).abs().mean().item()\n score += F.conv2d(x, weight=kernel, stride=1, padding=0).mean().item()\n return score\n\n\ndef save_best(output_dir, name, im, qtable, uv_qtable):\n im.save(path.join(output_dir, f\"{name}.jpg\"), format=\"jpeg\",\n qtables={0: qtable, 1: uv_qtable}, subsampling=\"4:2:0\")\n with open(path.join(output_dir, f\"{name}.json\"), mode=\"w\") as f:\n f.write(json.dumps({0: qtable, 1: uv_qtable}))\n\n\ndef change_qtable(qtable, n, protect_index):\n qtable = qtable[:]\n for _ in range(n):\n i = random.randint(protect_index + 1, 64 - 1)\n qtable[ZIGZAG_SCAN_INDEX[i]] = random.randint(0, 255)\n return qtable\n\n\ndef main():\n parser = argparse.ArgumentParser(formatter_class=argparse.ArgumentDefaultsHelpFormatter)\n parser.add_argument(\"--input\", \"-i\", type=str, default=TEST_IMAGE,\n help=\"test image file\")\n parser.add_argument(\"--output-dir\", \"-o\", type=str, default=OUTPUT_DIR,\n help=\"output directory\")\n parser.add_argument(\"--target\", type=str, default=\"checkboard\",\n choices=[\"vline\", \"hline\", \"vhline\", \"checkboard\"],\n help=\"target patterns\")\n parser.add_argument(\"--protect-index\", type=int, default=10,\n help=\"protect qtable index to reduce block noise. 6, 10, 15 are reasonable values.\")\n parser.add_argument(\"--protect-value\", type=int, default=10,\n help=\"qtable value for protect indexes. 1-255: 1=quality 100, 255=quality 0\")\n parser.add_argument(\"--max-epoch\", type=int, default=80000,\n help=\"max epoch\")\n parser.add_argument(\"--criterion\", type=str, choices=[\"mean\", \"max\"], default=\"max\",\n help=\"criterion\")\n parser.add_argument(\"--fixed-uv\", action=\"store_true\", \n help=\"use fixed qtable for uv\")\n parser.add_argument(\"--fixed-uv-value\", type=int, default=40,\n help=\"qtable value for fixed uv qtable\")\n\n args = parser.parse_args()\n runtime_name = f\"{args.target}_{int(time.time())}\"\n if args.target == \"vline\":\n kernels = [VERTICAL_LINE_KERNEL]\n elif args.target == \"hline\":\n kernels = [HORIZONTAL_LINE_KERNEL]\n elif args.target == \"vhline\":\n kernels = [HORIZONTAL_LINE_KERNEL, VERTICAL_LINE_KERNEL]\n elif args.target == \"checkboard\":\n kernels = [CHECKBOARD_KERNEL]\n else:\n raise NotImplementedError()\n num_changes = [1, 2, 3, 4, 8, 16]\n for i in range(15, 64 - 1):\n FIXED_UV_QTABLE[ZIGZAG_SCAN_INDEX[i]] = args.fixed_uv_value\n\n # init\n test_image = Image.open(args.input)\n test_image = test_image.convert(\"RGB\")\n os.makedirs(args.output_dir, exist_ok=True)\n qtable = [args.protect_value for _ in range(64)]\n uv_qtable = FIXED_UV_QTABLE if args.fixed_uv else qtable\n best_score = calc_score(test_image, qtable, uv_qtable, kernels, args.criterion)\n\n # search best(worst) qtable\n print(f\"Start output_dir={args.output_dir}, name={runtime_name}\")\n for epoch in range(args.max_epoch):\n n = random.choice(num_changes)\n new_qtable = change_qtable(qtable, n=n, protect_index=args.protect_index)\n uv_qtable = FIXED_UV_QTABLE if args.fixed_uv else new_qtable\n new_score = calc_score(test_image, new_qtable, uv_qtable, kernels, args.criterion)\n if new_score > best_score:\n qtable[:] = new_qtable[:]\n best_score = new_score\n save_best(args.output_dir, runtime_name, test_image, new_qtable, uv_qtable)\n print(f\"update score epoch={epoch}: {best_score}\")\n print(\"done\")\n\n\nif __name__ == \"__main__\":\n main()\n", "sub_path": "playground/jpeg_qtable/search_qtable.py", "file_name": "search_qtable.py", "file_ext": "py", "file_size_in_byte": 6521, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.relpath", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 28, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 40, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 50, "usage_type": "attribute"}, {"api_name": "io.BytesIO", "line_number": 78, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 82, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 82, "usage_type": "name"}, {"api_name": "torchvision.transforms.functional.to_tensor", "line_number": 83, "usage_type": "call"}, {"api_name": "torchvision.transforms.functional", "line_number": 83, "usage_type": "name"}, {"api_name": "torchvision.transforms.functional.to_grayscale", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.nn.functional.adaptive_max_pool2d", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 87, "usage_type": "name"}, {"api_name": "torch.nn.functional.conv2d", "line_number": 88, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 88, "usage_type": "name"}, {"api_name": "torch.nn.functional.conv2d", "line_number": 93, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 93, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 98, "usage_type": "call"}, {"api_name": "os.path", "line_number": 98, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path", "line_number": 100, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 101, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 107, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 108, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 113, "usage_type": "call"}, {"api_name": "argparse.ArgumentDefaultsHelpFormatter", "line_number": 113, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 135, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 151, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 151, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 153, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 161, "usage_type": "call"}]}
+{"seq_id": "1327307", "text": "#! /usr/bin/python\n\nfrom landlab import Component\nimport numpy as np\nimport datetime as datetime\nfrom scipy.sparse import csr_matrix\nfrom scipy.sparse import linalg\nimport six\n\n\nclass Glacier(Component):\n\n\tdef __init__(self,grid,dictionary,**kwds):\n\t\t'''\n\t\tDefine physical parameters (here assuming EISMINT-1 values)\n\t\t'''\n\t\tself.n_GLEN = 3 # Glen's flow law exponent\n\t\tself.A_GLEN = 7.5738e-17 #6.05904e-18; Monthly #7.5738e-17 Cuffey & Paterson (4th ed) Glen's law parameter in Pa^{-3} yr^{-1} units (same as A_GLEN=2.4e-24 Pa^{-3} s^{-1})\n\n\t\tself.m_SLIDE = 2 # Sliding law exponent\n\t\tself.C_SLIDE = 0 # 1.0e-08; # 1.0e-06; # Sliding coefficient in Pa, metre,(Year units)\n\n\t\tself.RHO = 900 # Density (SI units)\n\t\tself.g = 9.80 # Gravity (SI units, rho*g has units of Pa)\n\t\tself.K_eps = 1.0e-12\n\t\tself.OMEGA = 1.5 # 1.6\n\t\tsuper(Glacier,self).__init__(grid)\n\t\tself.initialize(dictionary)\n\n\tdef initialize(self,dictionary,**kwds):\n\t\t'''\n\t\tInitialize values for calculation:\n\n\t\tS: ice surface ice_elevation\n\t\tB: bed elevation\n\t\tb_dot: mass of ice added or subtracted from each cell\n\t\tdt: time interval\n\t\tt_STOP: ending time for modeling\n\t\tt: starting time for modeling\n\t\tdx: node spacing\n\t\tnx: number of columns of nodes\n\t\tny: number of rows of nodes\n\t\tN: number of nodes\n\t\t'''\n\t\tself.S = kwds.pop('S',dictionary['S'])\n\t\tself.B = kwds.pop('B',dictionary['B'])\n\t\tself.b_dot = kwds.pop('b_dot',dictionary['b_dot'])\n\t\tself.dt = kwds.pop('dt',dictionary['dt'])\n\t\tself.t_STOP = kwds.pop('t_STOP',dictionary['t_STOP'])\n\t\tself.t = kwds.pop('t',dictionary['t'])\n\t\tself.dx = kwds.pop('dx',dictionary['dx'])\n\t\tself.nx = kwds.pop('nx',dictionary['nx'])\n\t\tself.ny = kwds.pop('ny',dictionary['ny'])\n\t\tself.N = self.nx * self.ny\n\t\tself.setupIndexArrays()\n\n\tdef step_update(self):\n\t\t'''\n\t\tcalculate S (ice surface elevation) for each timestep\n\n\t\tH_max: maximum ice thickness\n\t\tS_max: maximum ice surface elevation\n\t\t'''\n\t\tself.S, self.t = self.step()\n\t\tSB = self.S - self.B\n\t\tself.H_max = np.max(SB)\n\t\tself.k_H_max = np.argmax(SB)\n\t\tself.S_max = np.max(self.S)\n\t\tself.k_S_max = np.argmax(self.S)\n\n\tdef recursive_steps(self):\n\t\t'''\n\t\tIterate over each time step to update the ice surface elevation\n\t\t'''\n\t\twhile 1:\n\t\t\tself.step_update()\n\t\t\t# print self.S[0:5]\n\t\t\tself.ALPHA_I = 100*np.sum(self.S > self.B)/float(self.N)\n\n\t\t\tsix.print_('BKS: At t={:8.2f} yr ALPHA_I={:.2f}% and maxima are: H({:d}) = {:f} \\\n\t\t\tS({:d})={:f}\\n'.format(self.t, self.ALPHA_I, self.k_H_max, self.H_max, self.k_S_max, self.S_max))\n\n\t\t\t### Stop iterating until the final timestep\n\t\t\tif self.t > self.t_STOP:\n\t\t\t\tI = np.zeros(self.N)\n\t\t\t\tI[self.S > self.B] = 1\n\n\t\t\t\t# S_map = self.S.reshape(self.ny,self.nx)\n\t\t\t\t# B_map = self.B.reshape(self.ny,self.nx)\n\t\t\t\t# I_map = I.reshape(self.ny,self.nx)\n\n\t\t\t\t### Note: the difference between python and matlab in matrix orders\n\t\t\t\tS_map = self.S.reshape(self.nx,self.ny).T\n\t\t\t\tB_map = self.B.reshape(self.nx,self.ny).T\n\t\t\t\tI_map = I.reshape(self.nx,self.ny).T\n\t\t\t\tH_map = S_map - B_map\n\t\t\t\tself.grid['node']['ice_elevation'] = S_map\n\t\t\t\tself.grid['node']['B_map'] = B_map\n\t\t\t\tself.grid['node']['I_map'] = I_map\n\t\t\t\tself.grid['node']['ice_thickness'] = H_map\n\t\t\t\tnow = datetime.datetime.now().strftime('%H:%M:%S')\n\t\t\t\tfile_str = 'S_map.txt'\n\t\t\t\tsix.print_('main(): Output stored in file \"{:s}\" at time {:s} \\n'.format(file_str,now))\n\t\t\t\tbreak\n\n\tdef step(self):\n\t\t'''\n\t\tFor each timestep, a sparse linear system (Ax = C) need to be solved to update ice surface elevation\n\t\t'''\n\n\t\t### update diffusivity for each timestep\n\t\tself.diffusion_update()\n\t\tD_sum = self.D_IC_jc + self.D_IP_jc + self.D_ic_JC + self.D_ic_JP\n\n\t\trow = np.int64([[self.ic_jc],[self.ic_jc],[self.ic_jc],[self.ic_jc],[self.ic_jc]]).flatten()\n\t\tcol = np.int64([[self.im_jc],[self.ip_jc],[self.ic_jm],[self.ic_jp],[self.ic_jc]]).flatten()\n\t\tval = np.array([[-self.OMEGA * self.D_IC_jc],[-self.OMEGA * self.D_IP_jc],[-self.OMEGA * self.D_ic_JC],[-self.OMEGA * self.D_ic_JP],[1/self.dt + self.OMEGA * D_sum]]).flatten()\n\t\tC = (1 - self.OMEGA) * ((self.D_IC_jc * self.S[self.im_jc]) + self.D_IP_jc * self.S[self.ip_jc] + self.D_ic_JC * self.S[self.ic_jm] + self.D_ic_JP * \\\n\t\t\tself.S[self.ic_jp]) + (1/self.dt - (1 - self.OMEGA) * D_sum) * self.S[self.ic_jc] + self.b_dot\n\t\tC = C.flatten()\n\n\t\t### construct a sparse matrix A\n\t\tA = csr_matrix( (val,(row,col)), shape=(self.N, self.N))\n\t\t# print 'solving'\n\t\tS_out = linalg.spsolve(A,C)\n\t\t# print 'solved'\n\n\t\t### ice thickness couldn't be negative, ice surface elevation should not be less than bed elevation\n\t\tS_out[S_out < self.B] = self.B[S_out < self.B]\n\n\t\tt_n = self.t + self.dt\n\t\treturn S_out, t_n\n\n\tdef diffusion_update(self):\n\t\t'''\n\t\tcalculate diffusivity for each timestep\n\t\t'''\n\t\tA_tilde = 2 * self.A_GLEN * (self.RHO * self.g) ** self.n_GLEN/(self.n_GLEN + 2)/(self.dx ** 2)\n\t\tC_tilde = self.C_SLIDE * (self.RHO * self.g)**self.m_SLIDE/(self.dx**2)\n\t\tnm_half = (self.n_GLEN - 1) / 2.0 ### @\n\t\tnpl = self.n_GLEN + 1\n\t\tmm_half = (self.m_SLIDE - 1) / 2.0 ### @\n\t\tml = self.m_SLIDE\n\n\t\tSB = self.S - self.B\n\t\tSB[SB<0] = 0\n\t\tH = SB\n\n\t\tH_IC_jc = 0.5*(H[self.ic_jc] + H[self.im_jc])\n\t\tH_ic_JC = 0.5*(H[self.ic_jc] + H[self.ic_jm])\n\n\t\tH_IC_jc_up = H[self.im_jc]\n\t\tH_ic_JC_up = H[self.ic_jm]\n\n\t\tix = (self.S[self.ic_jc]>self.S[self.im_jc]).reshape(-1)\n\t\tH_IC_jc_up[self.S[self.ic_jc]>self.S[self.im_jc]] = H[self.ic_jc[ix]].reshape(-1)\n\n\t\tix = (self.S[self.ic_jc]>self.S[self.ic_jm]).reshape(-1)\n\t\tH_ic_JC_up[self.S[self.ic_jc]>self.S[self.ic_jm]] = H[self.ic_jc[ix]].reshape(-1)\n\n\t\tdS_dx_IC_jc = (self.S[self.ic_jc] - self.S[self.im_jc])/self.dx\n\t\tdS_dy_IC_jc = (self.S[self.ic_jp] + self.S[self.im_jp] - self.S[self.ic_jm] - self.S[self.im_jm])/(4*self.dx)\n\t\tdS_dx_ic_JC = (self.S[self.ip_jc] + self.S[self.ip_jm] - self.S[self.im_jc] - self.S[self.im_jm])/(4*self.dx)\n\t\tdS_dy_ic_JC = (self.S[self.ic_jc] - self.S[self.ic_jm])/self.dx\n\n\t\tS2_IC_jc = np.square(dS_dx_IC_jc) + np.square(dS_dy_IC_jc) + self.K_eps\n\t\tS2_ic_JC = np.square(dS_dx_ic_JC) + np.square(dS_dy_ic_JC) + self.K_eps\n\n\t\tif C_tilde == 0: ### No sliding case\n\t\t\tself.D_IC_jc = A_tilde*H_IC_jc_up*np.power(H_IC_jc,npl)*np.power(S2_IC_jc,nm_half)\n\t\t\tself.D_ic_JC = A_tilde*H_ic_JC_up*np.power(H_ic_JC,npl)*np.power(S2_ic_JC,nm_half)\n\t\telif C_tilde > 0: ### Sliding case\n\t\t\tself.D_IC_jc = A_tilde*H_IC_jc_up*np.power(H_IC_jc,npl)*np.power(S2_IC_jc,nm_half) \\\n\t\t\t\t\t+ C_tilde*H_IC_jc_up*np.power(H_IC_jc,ml)*np.power(S2_IC_jc,mm_half)\n\t\t\tself.D_ic_JC = A_tilde*H_ic_JC_up*np.power(H_ic_JC,npl)*np.power(S2_ic_JC,nm_half) \\\n\t\t\t\t\t+ C_tilde*H_ic_JC_up*np.power(H_ic_JC,ml)*np.power(S2_ic_JC,mm_half)\n\t\telse:\n\t\t\tsix.print_('diffusion(): C_tilde is undefined or incorrectly defined')\n\n\t\tself.D_IP_jc = self.D_IC_jc[self.ip_jc]\n\t\tself.D_ic_JP = self.D_ic_JC[self.ic_jp]\n\n\tdef setupIndexArrays(self):\n\n\t\tic = np.arange(self.nx)\n\t\tip = np.append(np.array([np.arange(1,self.nx)]),self.nx - 1)\n\t\tim = np.append(0,np.array([np.arange(self.nx - 1)]))\n\n\t\tjc = np.arange(self.ny)\n\t\tjp = np.append(0,np.array([np.arange(self.ny - 1)]))\n\t\tjm = np.append(np.array([np.arange(1,self.ny)]),self.ny - 1)\n\n\t\tic_jc = np.arange(1, self.N + 1).reshape(self.nx, self.ny)\n\n\t\tself.ip_jc = self.setupArrays(ip,jc,ic_jc) - 1\n\t\tself.im_jc = self.setupArrays(im,jc,ic_jc) - 1\n\t\tself.ic_jp = self.setupArrays(ic,jp,ic_jc) - 1\n\t\tself.ic_jm = self.setupArrays(ic,jm,ic_jc) - 1\n\n\t\tself.im_jm = self.setupArrays(im,jm,ic_jc) - 1\n\t\tself.ip_jm = self.setupArrays(ip,jm,ic_jc) - 1\n\t\tself.im_jp = self.setupArrays(im,jp,ic_jc) - 1\n\t\tself.ip_jp = self.setupArrays(ip,jp,ic_jc) - 1\n\n\t\tself.ic_jc = ic_jc.reshape(-1) - 1\n\n\n\tdef setupArrays(self,a, b, ic_jc):\n\t\tx,y = np.meshgrid(b,a)\n\t\tarray = []\n\t\tfor l in zip(y.ravel(),x.ravel()):\n\t\t\tarray.append(ic_jc[l])\n\t\tarray = np.array(array)\n\t\treturn array\n", "sub_path": "landlab/components/glacier_thin_ice_model/glacier.py", "file_name": "glacier.py", "file_ext": "py", "file_size_in_byte": 7765, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "32", "api": [{"api_name": "landlab.Component", "line_number": 11, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 78, "usage_type": "call"}, {"api_name": "six.print_", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 85, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 101, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 101, "usage_type": "attribute"}, {"api_name": "six.print_", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 117, "usage_type": "call"}, {"api_name": "scipy.sparse.csr_matrix", "line_number": 123, "usage_type": "call"}, {"api_name": "scipy.sparse.linalg.spsolve", "line_number": 125, "usage_type": "call"}, {"api_name": "scipy.sparse.linalg", "line_number": 125, "usage_type": "name"}, {"api_name": "numpy.square", "line_number": 166, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 170, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 175, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 176, "usage_type": "call"}, {"api_name": "six.print_", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 185, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 186, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 187, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 189, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 190, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 191, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 191, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 191, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 193, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 209, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 213, "usage_type": "call"}]}
+{"seq_id": "396170895", "text": "import streamlit as st\nimport pandas as pd\nimport numpy as np\nimport sys\nimport re\nimport itertools\nimport nltk\nimport re\nimport base64\nimport os\nfrom nltk.corpus import stopwords\nfrom nltk.tokenize import word_tokenize, sent_tokenize\nfrom sklearn.feature_extraction.text import TfidfVectorizer\nfrom collections import defaultdict\nfrom string import punctuation\nimport gensim\nfrom gensim.utils import simple_preprocess\nfrom gensim.parsing.preprocessing import strip_non_alphanum, stem_text, preprocess_string, strip_tags, strip_punctuation\nfrom PIL import Image\nfrom textwrap import wrap\nimport plotly\nimport plotly.graph_objects as go\nnltk.download('stopwords', quiet=True)\nnltk.download('wordnet', quiet=True)\nnltk.download('punkt', quiet=True)\nnltk.download('averaged_perceptron_tagger', quiet=True)\n\n\n\n\n# Imports from the oats package.\nsys.path.append(\"../oats\")\nimport oats\nfrom oats.utils.utils import load_from_pickle, save_to_pickle, flatten\nfrom oats.biology.dataset import Dataset\nfrom oats.annotation.ontology import Ontology\n\n\n# Imports from within this package.\nfrom token_similarities import TokenSimilarities, LossLogger\nimport query_handlers as qh\n\n\n\n\n\n\n\n\n\n\n\n\n\n# This section imports all the paths and dictionaries that are in the config file, that would change\n# between different datasets or uses of the streamlit application, that way everything needs to be \n# altered when using a different dataset, models, or ontologies can be changed in that configutation\n# file and nothing else in any other section in this main script needs to be changed at all.\nimport paths_and_models\nDATASET_PATH = paths_and_models.DATASET_PATH\nONTOLOGY_NAMES = paths_and_models.ONTOLOGY_NAMES\nONTOLOGY_OBO_PATHS = paths_and_models.ONTOLOGY_OBO_PATHS\nONTOLOGY_PICKLE_PATHS = paths_and_models.ONTOLOGY_PICKLE_PATHS\nSPECIES_STRINGS_IN_DATA = paths_and_models.SPECIES_STRINGS_IN_DATA\nSPECIES_STRINGS_TO_DISPLAY = paths_and_models.SPECIES_STRINGS_TO_DISPLAY\nTO_SPECIES_DISPLAY_NAME = paths_and_models.TO_SPECIES_DISPLAY_NAME\nGET_APPROACHES = paths_and_models.get_methods\nWORD2VEC_MODEL_PATH = paths_and_models.WORD2VEC_MODEL_PATH\nWORD_EMBEDDINGS_PICKLE_PATH = paths_and_models.WORD_EMBEDDINGS_PICKLE_PATH\nWORD_EMBEDDINGS_PREPROCESSING_FUNCTION = paths_and_models.WORD_EMBEDDINGS_PREPROCESSING_FUNCTION\n# End section.\n\n\n\n\n\n\n\n\n\n\n\n# Path to the documentation and dataset of genes, phenotype descriptions and annotations to be used. \nDOC_PARAGRAPH_PATH = \"documentation/paragraph.txt\"\nDOC_TABLE_PATH = \"documentation/table.txt\"\nREF_TEXT_PATH = \"documentation/references.txt\"\nCONTACT_PATH = \"documentation/contact.txt\"\nACKKNOWLEDGEMENTS_PATH = \"documentation/acknowledgement.txt\"\n\n\n\n\n\n# Constants that help define how the tables look and how the text wraps within the table cells.\nTABLE_HEADER_COLOR = \"#808080\"\nTABLE_HEIGHT = 1500\nHEADER_HEIGHT = 30\nRESULT_COLUMN_WIDTH = 55\nMAX_LINES_IN_RESULT_COLUMN = 100\nDESCRIPTION_COLUMN_WIDTH = 90\nNEWLINE_TOKEN = \"[NEWLINE]\"\n\n# What color the alternating rows should be, and what column will indicate when they should alternate?\nTABLE_ROWS_COLOR_1 = \"#FFFFFF\" # light gray\nTABLE_ROWS_COLOR_2 = \"#F1F2F6\" # slighly darker gray\nALTERNATE_ROW_COLOR_BASED_ON_COLUMN_KEY = \"gene\"\n\n\n\n\n\n\n\n\n\n# Constants that help define how the columns appear in the plottly tables. \n# The first value is the universal string key used throughout the script, so leave that alone.\n# The second is how the column is titled in the presented tables so that can be changed just here and the change will take effect throughout.\n# The third is the relative width of the column to all the other columns. Leave the rank column as 1 (the smallest), and change all othere\n# with respect to that column.\nCOLUMN_SETTINGS = [\n\t(\"rank\", \"Rank\", 0.9),\n\t(\"score\", \"Score\", 1),\n\t(\"result\", \"Result\", 1),\n\t(\"keywords\", \"Query Keywords\", 8),\n\t(\"sentences\", \"Query Sentences\", 8),\n\t(\"terms\", \"Ontology Terms\", 6),\n\t(\"species\", \"Species\", 2),\n\t(\"gene\", \"Gene\", 2),\n\t(\"model\", \"Gene Model\", 2),\n\t(\"phenotype\", \"Phenotype Description\", 12),\n\t(\"query_sentence\", \"Query\", 5),\n\t(\"matching_sentence\", \"Matches\", 12),\n\t(\"matching_sentence_truncated\", \"Matches (Truncated)\", 12),\n\t(\"query_term_id\", \"Query Term ID\", 2),\n\t(\"query_term_name\", \"Query Term Name\", 4),\n\t(\"annotated_term_id\", \"Annotated Term ID\", 2),\n\t(\"annotated_term_name\", \"Annotated Term Name\", 4),\n\t(\"internal_id\", \"Internal ID\", 1),\n]\n\nCOLUMN_NAMES = {x[0]:x[1] for x in COLUMN_SETTINGS}\nCOLUMN_NAMES_TO_OUTPUT_COLUMN_NAME = {v:v.replace(\"\",\"\") for k,v in COLUMN_NAMES.items()}\nCOLUMN_WIDTHS = {x[0]:x[2] for x in COLUMN_SETTINGS}\n\n\n\n\n\n\n\n\n# How should keywords and phrases be cleaned and handled as far as preprocessing or stemming goes?\nKEYWORD_DELIM = \"[DELIM]\"\nKEYWORD_PREPROCESSING_FILTERS = [lambda x: x.lower(), strip_non_alphanum, strip_tags, strip_punctuation, stem_text]\nPREPROCESSING_FOR_KEYWORD_SEARCH_FUNCTION = lambda x: \"{}{}{}\".format(KEYWORD_DELIM, KEYWORD_DELIM.join([token for token in preprocess_string(x, KEYWORD_PREPROCESSING_FILTERS)]), KEYWORD_DELIM)\n\n\n\n\n\n\n\n\n\n\n# Initial configuration and the header image at the top of the page.\nst.set_page_config(page_title=\"QuOATS\", layout=\"wide\", initial_sidebar_state=\"expanded\")\nPATH_TO_LOGO_PNG = \"images/header.png\"\nst.image(Image.open(PATH_TO_LOGO_PNG), caption=None, width=500, output_format=\"png\")\nst.markdown(\"### A tool for **Qu**erying phenotype descriptions with **O**ntology **A**nnotations and **T**ext **S**imilarity\")\n\n\nhide_streamlit_style = \"\"\"\n\n\"\"\"\nst.markdown(hide_streamlit_style, unsafe_allow_html=True) \n\n\n\n\n\n\n\n\n\n\n\n\n\nwith open(DOC_PARAGRAPH_PATH,\"r\") as f:\n\tdoc_paragraph = f.read()\n\nwith open(DOC_TABLE_PATH,\"r\") as f:\n\ttable_string = f.read()\n\tlines = table_string.split(\"\\n\")\n\tlines = [\"|{}|\".format(line.replace(\"\\t\",\"|\")) for line in lines]\n\tlines = [line.replace('\"','') for line in lines]\n\theader_separation_line = \"-\".join([\"|\"]*lines[0].count(\"|\"))\n\tall_lines = []\n\tall_lines.append(lines[0])\n\tall_lines.append(header_separation_line)\n\tall_lines.extend(lines[1:])\n\tdoc_table = \"\\n\".join(all_lines)\n\ndoc_expander = st.beta_expander(label=\"Show/Hide Documentation\", expanded=False)\nwith doc_expander:\n\tst.markdown(\"# Overview\")\n\tst.markdown(doc_paragraph)\n\tst.markdown(\"# Details\")\n\tst.markdown(doc_table)\n\n\n\n\n\n\n# Setting some of the color schemes and formatting of the page. \n# E7FD8E light green\n# 90918A middle gray\n# 952A53 eggplant\n# B3DE98 green\n# FF8B00 orange\n# E4F084 yellow\nst.markdown(\n\t\"\"\"\n\t