lines
listlengths 1
444
| raw_lines
listlengths 1
444
| label
listlengths 1
444
| type
listlengths 1
444
|
|---|---|---|---|
[
"def FUNC_0(self):...\n",
"super(CLASS_0, self).setUp()\n",
"VAR_7.environ.pop('SWARMING_LOAD_TEST', None)\n",
"self.root_dir = tempfile.mkdtemp(prefix='bot_main')\n",
"self.old_cwd = VAR_7.getcwd()\n",
"VAR_7.chdir(self.root_dir)\n",
"VAR_7.mkdir('logs')\n",
"self.server = xsrf_client.XsrfRemote('https://localhost:1/')\n",
"self.attributes = {'dimensions': {'foo': ['bar'], 'id': ['localhost'],\n 'pool': ['default']}, 'state': {'cost_usd_hour': 3600.0}, 'version': '123'}\n",
"self.mock(zip_package, 'generate_version', lambda : '123')\n",
"self.bot = bot.Bot(self.server, self.attributes, 'https://localhost:1/',\n 'version1', self.root_dir, self.fail)\n",
"self.mock(self.bot, 'post_error', self.fail)\n",
"self.mock(self.bot, 'restart', self.fail)\n",
"self.mock(subprocess42, 'call', self.fail)\n",
"self.mock(time, 'time', lambda : 100.0)\n",
"VAR_4 = VAR_7.path.join(test_env_bot_code.BOT_DIR, 'config', 'config.json')\n",
"VAR_36 = json.load(VAR_32)\n",
"self.mock(bot_main, 'get_config', lambda : VAR_36)\n",
"self.mock(bot_main, 'THIS_FILE', VAR_7.path.join(test_env_bot_code.BOT_DIR,\n 'swarming_bot.zip'))\n"
] |
[
"def setUp(self):...\n",
"super(TestBotMain, self).setUp()\n",
"os.environ.pop('SWARMING_LOAD_TEST', None)\n",
"self.root_dir = tempfile.mkdtemp(prefix='bot_main')\n",
"self.old_cwd = os.getcwd()\n",
"os.chdir(self.root_dir)\n",
"os.mkdir('logs')\n",
"self.server = xsrf_client.XsrfRemote('https://localhost:1/')\n",
"self.attributes = {'dimensions': {'foo': ['bar'], 'id': ['localhost'],\n 'pool': ['default']}, 'state': {'cost_usd_hour': 3600.0}, 'version': '123'}\n",
"self.mock(zip_package, 'generate_version', lambda : '123')\n",
"self.bot = bot.Bot(self.server, self.attributes, 'https://localhost:1/',\n 'version1', self.root_dir, self.fail)\n",
"self.mock(self.bot, 'post_error', self.fail)\n",
"self.mock(self.bot, 'restart', self.fail)\n",
"self.mock(subprocess42, 'call', self.fail)\n",
"self.mock(time, 'time', lambda : 100.0)\n",
"config_path = os.path.join(test_env_bot_code.BOT_DIR, 'config', 'config.json')\n",
"config = json.load(f)\n",
"self.mock(bot_main, 'get_config', lambda : config)\n",
"self.mock(bot_main, 'THIS_FILE', os.path.join(test_env_bot_code.BOT_DIR,\n 'swarming_bot.zip'))\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
5,
5,
0,
5,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"Expr'"
] |
[
"def FUNC_0(self):...\n",
"VAR_1 = 'sh'\n",
"self.assertEqual(escape_path_argument('/home/usr/a-file', VAR_1),\n '/home/usr/a-file')\n",
"self.assertEqual(escape_path_argument('/home/usr/a-dir/', VAR_1),\n '/home/usr/a-dir/')\n",
"self.assertEqual(escape_path_argument('/home/us r/a-file with spaces.bla',\n VAR_1), '/home/us\\\\ r/a-file\\\\ with\\\\ spaces.bla')\n",
"self.assertEqual(escape_path_argument('/home/us r/a-dir with spaces/x/',\n VAR_1), '/home/us\\\\ r/a-dir\\\\ with\\\\ spaces/x/')\n",
"self.assertEqual(escape_path_argument(\n 'relative something/with cherries and/pickles.delicious', VAR_1),\n 'relative\\\\ something/with\\\\ cherries\\\\ and/pickles.delicious')\n"
] |
[
"def test_escape_path_argument_sh(self):...\n",
"_type = 'sh'\n",
"self.assertEqual(escape_path_argument('/home/usr/a-file', _type),\n '/home/usr/a-file')\n",
"self.assertEqual(escape_path_argument('/home/usr/a-dir/', _type),\n '/home/usr/a-dir/')\n",
"self.assertEqual(escape_path_argument('/home/us r/a-file with spaces.bla',\n _type), '/home/us\\\\ r/a-file\\\\ with\\\\ spaces.bla')\n",
"self.assertEqual(escape_path_argument('/home/us r/a-dir with spaces/x/',\n _type), '/home/us\\\\ r/a-dir\\\\ with\\\\ spaces/x/')\n",
"self.assertEqual(escape_path_argument(\n 'relative something/with cherries and/pickles.delicious', _type),\n 'relative\\\\ something/with\\\\ cherries\\\\ and/pickles.delicious')\n"
] |
[
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_1(self, VAR_3, VAR_2):...\n",
"self.response_handlers[VAR_3] = VAR_2\n"
] |
[
"def set_response_handler(self, reqid, fun):...\n",
"self.response_handlers[reqid] = fun\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Assign'"
] |
[
"def FUNC_3(self, VAR_3, VAR_11=None):...\n",
"if super(CLASS_2, self).allowed(VAR_3, VAR_11):\n",
"self.verbose_name = _('Import Key Pair')\n",
"return True\n"
] |
[
"def allowed(self, request, keypair=None):...\n",
"if super(ImportKeyPair, self).allowed(request, keypair):\n",
"self.verbose_name = _('Import Key Pair')\n",
"return True\n"
] |
[
0,
0,
0,
0
] |
[
"FunctionDef'",
"Condition",
"Assign'",
"Return'"
] |
[
"\"\"\"\nUnit tests for the Deis api app.\n\nRun the tests with \"./manage.py test api\\\"\n\"\"\"\n",
"from __future__ import unicode_literals\n",
"import json\n",
"import mock\n",
"import requests\n",
"from django.contrib.auth.models import User\n",
"from django.test import TransactionTestCase\n",
"from django_fsm import TransitionNotAllowed\n",
"from api.models import Container, App\n",
"def FUNC_0(*VAR_0, **VAR_1):...\n",
"VAR_2 = requests.Response()\n",
"VAR_2.status_code = 200\n",
"VAR_2._content_consumed = True\n",
"return VAR_2\n"
] |
[
"\"\"\"\nUnit tests for the Deis api app.\n\nRun the tests with \"./manage.py test api\\\"\n\"\"\"\n",
"from __future__ import unicode_literals\n",
"import json\n",
"import mock\n",
"import requests\n",
"from django.contrib.auth.models import User\n",
"from django.test import TransactionTestCase\n",
"from django_fsm import TransitionNotAllowed\n",
"from api.models import Container, App\n",
"def mock_import_repository_task(*args, **kwargs):...\n",
"resp = requests.Response()\n",
"resp.status_code = 200\n",
"resp._content_consumed = True\n",
"return resp\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"Expr'",
"ImportFrom'",
"Import'",
"Import'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_58(self):...\n",
"VAR_2 = {'SEVERITY': {}, 'CONFIDENCE': {}}\n",
"self.check_example('nosec.py', VAR_2)\n"
] |
[
"def test_nosec(self):...\n",
"expect = {'SEVERITY': {}, 'CONFIDENCE': {}}\n",
"self.check_example('nosec.py', expect)\n"
] |
[
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Expr'"
] |
[
"def FUNC_5(self):...\n",
"return self.display_name\n"
] |
[
"def get_display_name(self):...\n",
"return self.display_name\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Return'"
] |
[
"def __init__(self, VAR_2, VAR_3):...\n",
"\"\"\"docstring\"\"\"\n",
"super(CLASS_6, self).__init__(VAR_2)\n",
"self.hostname = VAR_3\n",
"self.message = 'Component %s crashed on remote host %s' % (VAR_2, VAR_3)\n"
] |
[
"def __init__(self, comp_name, hostname):...\n",
"\"\"\"docstring\"\"\"\n",
"super(RemoteCrashEvent, self).__init__(comp_name)\n",
"self.hostname = hostname\n",
"self.message = 'Component %s crashed on remote host %s' % (comp_name, hostname)\n"
] |
[
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Expr'",
"Assign'",
"Assign'"
] |
[
"def FUNC_2(VAR_2):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_1 = 'string'.format(VAR_2)\n",
"return FUNC_0(VAR_1)\n"
] |
[
"def get_top_author(top_num):...\n",
"\"\"\"docstring\"\"\"\n",
"cmd = (\n \"\"\"SELECT authors.name,author_result.num\n FROM authors JOIN\n (SELECT SUM(article_result.num) as num,\n article_result.author\n from (SELECT articles.title, articles.author,\n SUM(log.views) AS num\n FROM articles\n INNER JOIN (\n SELECT path, count(path) AS views\n FROM log GROUP BY log.path\n ) AS log ON log.path = '/article/'\n || articles.slug\n GROUP BY articles.title, articles.author)\n AS article_result\n GROUP BY article_result.author) as author_result\n ON authors.id = author_result.author\n ORDER BY num DESC LIMIT {}\"\"\"\n .format(top_num))\n",
"return execute_query(cmd)\n"
] |
[
0,
0,
4,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"Return'"
] |
[
"def FUNC_19(self, VAR_25, VAR_27=''):...\n",
"VAR_20 = request_mock(VAR_25=path, VAR_27=body, VAR_28={'Content-Type': [\n 'application/json']}, method='DELETE')\n",
"return self._render(VAR_20)\n"
] |
[
"def delete(self, path, body=''):...\n",
"request = request_mock(path=path, body=body, headers={'Content-Type': [\n 'application/json']}, method='DELETE')\n",
"return self._render(request)\n"
] |
[
0,
5,
0
] |
[
"FunctionDef'",
"Assign'",
"Return'"
] |
[
"def FUNC_6(VAR_12, VAR_2, VAR_5):...\n",
"VAR_26 = VAR_12.request.body\n",
"VAR_2 = FUNC_1(VAR_2)\n",
"def FUNC_8(VAR_4):...\n",
"if not os.path.exists('D:' + VAR_4):\n",
"return VAR_4\n",
"VAR_28 = 1\n",
"while VAR_28 < 101:\n",
"VAR_32 = re.sub('\\\\.(.*?)$', ' (%d).\\\\1' % VAR_28, VAR_4)\n",
"return ''\n",
"if not os.path.exists('D:' + VAR_32):\n",
"return VAR_32\n",
"VAR_28 = VAR_28 + 1\n"
] |
[
"def save_file_async(alter_ego, target_path, file_name):...\n",
"upload_data = alter_ego.request.body\n",
"target_path = decode_hexed_b64_to_str(target_path)\n",
"def get_non_duplicate_path(file_path):...\n",
"if not os.path.exists('D:' + file_path):\n",
"return file_path\n",
"duplicate = 1\n",
"while duplicate < 101:\n",
"new_path = re.sub('\\\\.(.*?)$', ' (%d).\\\\1' % duplicate, file_path)\n",
"return ''\n",
"if not os.path.exists('D:' + new_path):\n",
"return new_path\n",
"duplicate = duplicate + 1\n"
] |
[
0,
0,
0,
0,
4,
4,
4,
4,
4,
4,
4,
4,
4
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"FunctionDef'",
"Condition",
"Return'",
"Assign'",
"Condition",
"Assign'",
"Return'",
"Condition",
"Return'",
"Assign'"
] |
[
"def FUNC_40(self, VAR_26):...\n",
"\"\"\"docstring\"\"\"\n",
"for filename in VAR_26:\n",
"if self.config['need_sudo'] or self.config['become_root']:\n",
"VAR_4 = 'Error collecting additional data from master: %s' % VAR_68\n",
"VAR_62 = self.retrieve_file(filename)\n",
"self.make_archive_readable(filename)\n",
"self.log_error('Unable to retrieve file %s' % filename)\n",
"self.log_error(VAR_4)\n",
"if VAR_62:\n",
"self.log_debug('Failed to make file %s readable: %s' % (filename, VAR_67))\n",
"self.remove_file(filename)\n",
"self.log_error('Unable to retrieve file %s' % filename)\n"
] |
[
"def collect_extra_cmd(self, filenames):...\n",
"\"\"\"docstring\"\"\"\n",
"for filename in filenames:\n",
"if self.config['need_sudo'] or self.config['become_root']:\n",
"msg = 'Error collecting additional data from master: %s' % e\n",
"ret = self.retrieve_file(filename)\n",
"self.make_archive_readable(filename)\n",
"self.log_error('Unable to retrieve file %s' % filename)\n",
"self.log_error(msg)\n",
"if ret:\n",
"self.log_debug('Failed to make file %s readable: %s' % (filename, err))\n",
"self.remove_file(filename)\n",
"self.log_error('Unable to retrieve file %s' % filename)\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"For",
"Condition",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Condition",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_2(self):...\n",
"return self.HandleGet()\n"
] |
[
"def HandlePost(self):...\n",
"return self.HandleGet()\n"
] |
[
0,
5
] |
[
"FunctionDef'",
"Return'"
] |
[
"def FUNC_11(VAR_24):...\n",
"for user, VAR_20, id1, id2 in VAR_23.findall(VAR_24):\n",
"VAR_21 = id1 + id2\n",
"FUNC_8(VAR_20, VAR_21, user)\n"
] |
[
"def atfu(urls):...\n",
"for user, domain, id1, id2 in r_di.findall(urls):\n",
"id_ = id1 + id2\n",
"add_target(domain, id_, user)\n"
] |
[
0,
0,
0,
0
] |
[
"FunctionDef'",
"For",
"Assign'",
"Expr'"
] |
[
"def FUNC_27(self, VAR_14, VAR_30, VAR_29):...\n",
"VAR_53 = \"%s '%s' '%s'\" % (VAR_5, VAR_30, VAR_14)\n",
"VAR_12 = \"ssh %s 'bash -s' < %s\" % (VAR_29, VAR_53)\n",
"FUNC_7(self.session, VAR_12)\n"
] |
[
"def start_remote_clone_session(self, comp_name, session_name, hostname):...\n",
"remote_cmd = \"%s '%s' '%s'\" % (SCRIPT_CLONE_PATH, session_name, comp_name)\n",
"cmd = \"ssh %s 'bash -s' < %s\" % (hostname, remote_cmd)\n",
"send_main_session_command(self.session, cmd)\n"
] |
[
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_2(VAR_0):...\n",
"VAR_2 = []\n",
"VAR_3 = FUNC_4(VAR_0)\n",
"VAR_4 = FUNC_3(VAR_0)\n",
"for item in VAR_3:\n",
"if VAR_4.get(item.name):\n",
"return VAR_2\n",
"VAR_9 = VAR_4[item.name]\n",
"VAR_10 = {'item_group': item.item_group, 'item': item.name, 'item_name':\n item.item_name}\n",
"if VAR_9.days_since_last_order > cint(VAR_0['days']):\n",
"VAR_2.append(VAR_10)\n",
"VAR_10 = {'territory': VAR_9.territory, 'item_group': VAR_9.item_group,\n 'item': VAR_9.name, 'item_name': VAR_9.item_name, 'customer': VAR_9.\n customer, 'last_order_date': VAR_9.last_order_date, 'qty': VAR_9.qty,\n 'days_since_last_order': VAR_9.days_since_last_order}\n",
"VAR_2.append(VAR_10)\n"
] |
[
"def get_data(filters):...\n",
"data = []\n",
"items = get_items(filters)\n",
"sales_invoice_data = get_sales_details(filters)\n",
"for item in items:\n",
"if sales_invoice_data.get(item.name):\n",
"return data\n",
"item_obj = sales_invoice_data[item.name]\n",
"row = {'item_group': item.item_group, 'item': item.name, 'item_name': item.\n item_name}\n",
"if item_obj.days_since_last_order > cint(filters['days']):\n",
"data.append(row)\n",
"row = {'territory': item_obj.territory, 'item_group': item_obj.item_group,\n 'item': item_obj.name, 'item_name': item_obj.item_name, 'customer':\n item_obj.customer, 'last_order_date': item_obj.last_order_date, 'qty':\n item_obj.qty, 'days_since_last_order': item_obj.days_since_last_order}\n",
"data.append(row)\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"For",
"Condition",
"Return'",
"Assign'",
"Assign'",
"Condition",
"Expr'",
"Assign'",
"Expr'"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self.dimensions.get('id', ['unknown'])[0]\n"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self.dimensions.get('id', ['unknown'])[0]\n"
] |
[
0,
0,
0
] |
[
"Condition",
"Docstring",
"Return'"
] |
[
"def __eq__(self, VAR_35):...\n",
"return self.__class__ == VAR_35.__class__ and self.inner_type == VAR_35.inner_type\n"
] |
[
"def __eq__(self, other):...\n",
"return self.__class__ == other.__class__ and self.inner_type == other.inner_type\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Return'"
] |
[
"@app.route('/auth/logout')...\n",
"logout_user()\n",
"return redirect(url_for('index'))\n"
] |
[
"@app.route('/auth/logout')...\n",
"logout_user()\n",
"return redirect(url_for('index'))\n"
] |
[
0,
0,
0
] |
[
"Condition",
"Expr'",
"Return'"
] |
[
"@VAR_0.route('/')...\n",
"VAR_1 = ['स्वर्गवर्गः', 'व्योमवर्गः', 'दिग्वर्गः', 'कालवर्गः', 'धीवर्गः',\n 'शब्दादिवर्गः', 'नाट्यवर्गः', 'पातालभोगिवर्गः', 'नरकवर्गः', 'वारिवर्गः',\n 'भूमिवर्गः', 'पुरवर्गः', 'शैलवर्गः', 'वनौषधिवर्गः', 'सिंहादिवर्गः',\n 'मनुष्यवर्गः', 'ब्रह्मवर्गः', 'क्षत्रियवर्गः', 'वैश्यवर्गः',\n 'शूद्रवर्गः', 'विशेष्यनिघ्नवर्गः', 'सङ्कीर्णवर्गः', 'विशेष्यनिघ्नवर्गः',\n 'सङ्कीर्णवर्गः', 'नानार्थवर्गः', 'अव्ययवर्गः']\n",
"return render_template('index.html', VAR_1=all_vargas)\n"
] |
[
"@app.route('/')...\n",
"all_vargas = ['स्वर्गवर्गः', 'व्योमवर्गः', 'दिग्वर्गः', 'कालवर्गः',\n 'धीवर्गः', 'शब्दादिवर्गः', 'नाट्यवर्गः', 'पातालभोगिवर्गः', 'नरकवर्गः',\n 'वारिवर्गः', 'भूमिवर्गः', 'पुरवर्गः', 'शैलवर्गः', 'वनौषधिवर्गः',\n 'सिंहादिवर्गः', 'मनुष्यवर्गः', 'ब्रह्मवर्गः', 'क्षत्रियवर्गः',\n 'वैश्यवर्गः', 'शूद्रवर्गः', 'विशेष्यनिघ्नवर्गः', 'सङ्कीर्णवर्गः',\n 'विशेष्यनिघ्नवर्गः', 'सङ्कीर्णवर्गः', 'नानार्थवर्गः', 'अव्ययवर्गः']\n",
"return render_template('index.html', all_vargas=all_vargas)\n"
] |
[
0,
0,
0
] |
[
"Condition",
"Assign'",
"Return'"
] |
[
"@staticmethod...\n",
"return FUNC_1(CLASS_0._TalkToHandlerAsync('', VAR_7, 'GET', VAR_10))\n"
] |
[
"@staticmethod...\n",
"return JsonFromFuture(BaseRequest._TalkToHandlerAsync('', handler, 'GET',\n timeout))\n"
] |
[
0,
0
] |
[
"Condition",
"Return'"
] |
[
"@memoized_property...\n",
"\"\"\"docstring\"\"\"\n",
"return self._zinc_factory._compiler_interface(self._products)\n"
] |
[
"@memoized_property...\n",
"\"\"\"docstring\"\"\"\n",
"return self._zinc_factory._compiler_interface(self._products)\n"
] |
[
0,
0,
0
] |
[
"Condition",
"Docstring",
"Return'"
] |
[
"import discord, yaml, vroll, pgsql, re\n",
"import texttable as tt\n",
"from discord.ext import commands\n",
"VAR_0 = yaml.safe_load(open('config.yaml'))\n",
"VAR_1 = VAR_0['token']\n",
"VAR_2 = VAR_0['chan_whitelist']\n",
"VAR_3 = VAR_0['pg_connection']\n",
"VAR_4 = VAR_0['role_whitelist']\n",
"VAR_5 = VAR_0['permission_error_message']\n",
"VAR_6 = VAR_0['quest_tiers']\n",
"pgsql.create_tables(VAR_3)\n",
"VAR_7 = \"\"\"\nVishnu, a multipurpose D&D bot.\n\"\"\"\n",
"VAR_8 = commands.Bot(command_prefix='!', VAR_7=description)\n",
"\"\"\"\nRole whitelisting function\n\"\"\"\n",
"def FUNC_0(VAR_9):...\n",
"for x in VAR_4:\n",
"if x in [y.id for y in VAR_9.message.author.roles]:\n",
"\"\"\"\nQUEST-RELATED COMMANDS\n\"\"\"\n",
"return True\n",
"return False\n",
"@VAR_8.command()...\n",
"\"\"\"docstring\"\"\"\n",
"if FUNC_0(VAR_9):\n",
"if VAR_10 in VAR_6:\n",
"await VAR_9.send(VAR_5)\n",
"if len(VAR_11) < 100:\n",
"await VAR_9.send(\n 'Error: The quest tier you specified is invalid. The valid quest tiers are: '\n + ', '.join(VAR_6) + '. You specified: ' + VAR_10)\n",
"@VAR_8.command()...\n",
"VAR_29 = ' '.join(VAR_11)\n",
"await VAR_9.send(\n 'Error: Your description is too long. The maximum allowed characters is 100, you had '\n + str(len(VAR_11)))\n",
"\"\"\"docstring\"\"\"\n",
"VAR_30 = str(VAR_9.author)\n",
"if FUNC_0(VAR_9):\n",
"pgsql.import_quest_data(VAR_3, VAR_10, VAR_29, VAR_30)\n",
"pgsql.delete_quest(VAR_3, VAR_12)\n",
"await VAR_9.send(VAR_5)\n",
"print('Tier {} quest added by {}. Description: {}'.format(VAR_10, str(VAR_9\n .author), VAR_29))\n",
"await VAR_9.send('Quest with ID ' + VAR_12 + ' deleted.')\n",
"@VAR_8.command()...\n",
"await VAR_9.send('Tier {} quest added by {}. Description: {}'.format(VAR_10,\n str(VAR_9.author), VAR_29))\n",
"\"\"\"docstring\"\"\"\n",
"if FUNC_0(VAR_9):\n",
"pgsql.complete_quest(VAR_3, VAR_12, True)\n",
"await VAR_9.send(VAR_5)\n",
"@VAR_8.command()...\n",
"\"\"\"docstring\"\"\"\n",
"if FUNC_0(VAR_9):\n",
"pgsql.complete_quest(VAR_3, VAR_12, False)\n",
"await VAR_9.send(\"You don't have permission to use this command\")\n",
"@VAR_8.command()...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_14 = ' '.join(map(str, VAR_13))\n",
"VAR_15 = 'id=([\\\\d])'\n",
"VAR_16 = 'tier=([^\\\\s]+)'\n",
"VAR_17 = 'creator=([^\\\\s]+)'\n",
"VAR_18 = ''\n",
"VAR_19 = ''\n",
"VAR_20 = ''\n",
"if re.search(VAR_15, VAR_14) is not None:\n",
"VAR_25 = re.search(VAR_15, VAR_14).group(1)\n",
"if re.search(VAR_16, VAR_14) is not None:\n",
"VAR_18 = 'AND id = {}'.format(VAR_25)\n",
"VAR_26 = re.search(VAR_16, VAR_14).group(1)\n",
"if re.search(VAR_17, VAR_14) is not None:\n",
"VAR_19 = \"AND tier = '{}'\".format(VAR_26)\n",
"VAR_27 = re.search(VAR_17, VAR_14).group(1)\n",
"VAR_21 = 'string'.format(VAR_18, VAR_19, VAR_20)\n",
"VAR_20 = \"AND creator = '{}'\".format(VAR_27)\n",
"VAR_22 = pgsql.retrieve_quest_data(VAR_3, VAR_21)\n",
"VAR_23 = tt.Texttable()\n",
"VAR_24 = ['ID', 'TIER', 'CREATOR', 'DESCRIPTION']\n",
"VAR_23.header(VAR_24)\n",
"for x in range(0, len(VAR_22), 5):\n",
"for row in VAR_22[x:x + 5]:\n",
"\"\"\"\nDICE COMMANDS\n\"\"\"\n",
"VAR_23.add_row(row)\n",
"VAR_28 = VAR_23.draw()\n",
"@VAR_8.command()...\n",
"print(len(VAR_22))\n",
"\"\"\"docstring\"\"\"\n",
"await VAR_9.send('```' + VAR_28 + '```')\n",
"if VAR_2 is None:\n",
"VAR_23.reset()\n",
"for x in VAR_13:\n",
"if VAR_9.channel.id in VAR_2:\n",
"print('!roll command recieved in channel ID ' + str(VAR_9.channel.id))\n",
"@VAR_8.event...\n",
"for x in VAR_13:\n",
"await VAR_9.send(vroll.roll(x))\n",
"print('{0.user} connected to server'.format(VAR_8))\n",
"print('!roll command recieved in channel ID ' + str(VAR_9.channel.id) +\n ' by user ' + str(VAR_9.author))\n",
"print('Whitelisted channel IDs are: ' + str(VAR_2))\n",
"await VAR_9.send(vroll.roll(x))\n",
"print('Whitelisted role IDs are: ' + str(VAR_4))\n",
"VAR_8.run(VAR_1)\n"
] |
[
"import discord, yaml, vroll, pgsql, re\n",
"import texttable as tt\n",
"from discord.ext import commands\n",
"config = yaml.safe_load(open('config.yaml'))\n",
"token = config['token']\n",
"chan_whitelist = config['chan_whitelist']\n",
"pg_connection = config['pg_connection']\n",
"role_whitelist = config['role_whitelist']\n",
"permission_error_message = config['permission_error_message']\n",
"quest_tier_whitelist = config['quest_tiers']\n",
"pgsql.create_tables(pg_connection)\n",
"description = \"\"\"\nVishnu, a multipurpose D&D bot.\n\"\"\"\n",
"bot = commands.Bot(command_prefix='!', description=description)\n",
"\"\"\"\nRole whitelisting function\n\"\"\"\n",
"def whitelist_check(ctx):...\n",
"for x in role_whitelist:\n",
"if x in [y.id for y in ctx.message.author.roles]:\n",
"\"\"\"\nQUEST-RELATED COMMANDS\n\"\"\"\n",
"return True\n",
"return False\n",
"@bot.command()...\n",
"\"\"\"docstring\"\"\"\n",
"if whitelist_check(ctx):\n",
"if quest_tier in quest_tier_whitelist:\n",
"await ctx.send(permission_error_message)\n",
"if len(desc) < 100:\n",
"await ctx.send(\n 'Error: The quest tier you specified is invalid. The valid quest tiers are: '\n + ', '.join(quest_tier_whitelist) + '. You specified: ' + quest_tier)\n",
"@bot.command()...\n",
"quest_desc = ' '.join(desc)\n",
"await ctx.send(\n 'Error: Your description is too long. The maximum allowed characters is 100, you had '\n + str(len(desc)))\n",
"\"\"\"docstring\"\"\"\n",
"creator = str(ctx.author)\n",
"if whitelist_check(ctx):\n",
"pgsql.import_quest_data(pg_connection, quest_tier, quest_desc, creator)\n",
"pgsql.delete_quest(pg_connection, quest_id)\n",
"await ctx.send(permission_error_message)\n",
"print('Tier {} quest added by {}. Description: {}'.format(quest_tier, str(\n ctx.author), quest_desc))\n",
"await ctx.send('Quest with ID ' + quest_id + ' deleted.')\n",
"@bot.command()...\n",
"await ctx.send('Tier {} quest added by {}. Description: {}'.format(\n quest_tier, str(ctx.author), quest_desc))\n",
"\"\"\"docstring\"\"\"\n",
"if whitelist_check(ctx):\n",
"pgsql.complete_quest(pg_connection, quest_id, True)\n",
"await ctx.send(permission_error_message)\n",
"@bot.command()...\n",
"\"\"\"docstring\"\"\"\n",
"if whitelist_check(ctx):\n",
"pgsql.complete_quest(pg_connection, quest_id, False)\n",
"await ctx.send(\"You don't have permission to use this command\")\n",
"@bot.command()...\n",
"\"\"\"docstring\"\"\"\n",
"command = ' '.join(map(str, args))\n",
"idsearch = 'id=([\\\\d])'\n",
"tiersearch = 'tier=([^\\\\s]+)'\n",
"creatorsearch = 'creator=([^\\\\s]+)'\n",
"idformat = ''\n",
"tierformat = ''\n",
"creatorformat = ''\n",
"if re.search(idsearch, command) is not None:\n",
"idmatch = re.search(idsearch, command).group(1)\n",
"if re.search(tiersearch, command) is not None:\n",
"idformat = 'AND id = {}'.format(idmatch)\n",
"tiermatch = re.search(tiersearch, command).group(1)\n",
"if re.search(creatorsearch, command) is not None:\n",
"tierformat = \"AND tier = '{}'\".format(tiermatch)\n",
"creatormatch = re.search(creatorsearch, command).group(1)\n",
"query = (\n \"\"\"\n SELECT id, tier, creator, description FROM quests\n WHERE completed = 'f'\n {}\n {}\n {};\n \"\"\"\n .format(idformat, tierformat, creatorformat))\n",
"creatorformat = \"AND creator = '{}'\".format(creatormatch)\n",
"query_return = pgsql.retrieve_quest_data(pg_connection, query)\n",
"tab = tt.Texttable()\n",
"headings = ['ID', 'TIER', 'CREATOR', 'DESCRIPTION']\n",
"tab.header(headings)\n",
"for x in range(0, len(query_return), 5):\n",
"for row in query_return[x:x + 5]:\n",
"\"\"\"\nDICE COMMANDS\n\"\"\"\n",
"tab.add_row(row)\n",
"s = tab.draw()\n",
"@bot.command()...\n",
"print(len(query_return))\n",
"\"\"\"docstring\"\"\"\n",
"await ctx.send('```' + s + '```')\n",
"if chan_whitelist is None:\n",
"tab.reset()\n",
"for x in args:\n",
"if ctx.channel.id in chan_whitelist:\n",
"print('!roll command recieved in channel ID ' + str(ctx.channel.id))\n",
"@bot.event...\n",
"for x in args:\n",
"await ctx.send(vroll.roll(x))\n",
"print('{0.user} connected to server'.format(bot))\n",
"print('!roll command recieved in channel ID ' + str(ctx.channel.id) +\n ' by user ' + str(ctx.author))\n",
"print('Whitelisted channel IDs are: ' + str(chan_whitelist))\n",
"await ctx.send(vroll.roll(x))\n",
"print('Whitelisted role IDs are: ' + str(role_whitelist))\n",
"bot.run(token)\n"
] |
[
4,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
4,
0,
4,
0,
0,
0,
0,
0,
0,
0,
0,
0,
4,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"Import'",
"Import'",
"ImportFrom'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"FunctionDef'",
"For",
"For",
"Expr'",
"Return'",
"Return'",
"Condition",
"Docstring",
"Condition",
"Condition",
"Expr'",
"Condition",
"Expr'",
"Condition",
"Assign'",
"Expr'",
"Docstring",
"Assign'",
"Condition",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Condition",
"Expr'",
"Docstring",
"Condition",
"Expr'",
"Expr'",
"Condition",
"Docstring",
"Condition",
"Expr'",
"Expr'",
"Condition",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"For",
"For",
"Expr'",
"Expr'",
"Assign'",
"Condition",
"Expr'",
"Docstring",
"Expr'",
"Condition",
"Expr'",
"For",
"Condition",
"Expr'",
"Condition",
"For",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_4(self, VAR_6, VAR_12, VAR_13):...\n",
"VAR_14 = []\n",
"VAR_14.append('string'.format(VAR_6))\n",
"VAR_14.append('string'.format(VAR_6, VAR_12, VAR_13))\n",
"for query in VAR_14:\n",
"self.cur.execute(query)\n",
"self.conn.commit()\n"
] |
[
"def update_attendance(self, username, event_id, attending):...\n",
"sql = []\n",
"sql.append(\n \"\"\"INSERT INTO users (username)\n VALUES ('{0}')\n ON DUPLICATE KEY UPDATE username = '{0}';\n \"\"\"\n .format(username))\n",
"sql.append(\n \"\"\"INSERT INTO user_event (username, event_id, attending)\n VALUES ('{0}', '{1}', '{2}')\n ON DUPLICATE KEY UPDATE attending = '{2}';\n \"\"\"\n .format(username, event_id, attending))\n",
"for query in sql:\n",
"self.cur.execute(query)\n",
"self.conn.commit()\n"
] |
[
0,
0,
0,
4,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Expr'",
"Expr'",
"For",
"Expr'",
"Expr'"
] |
[
"def FUNC_4(VAR_0):...\n",
"VAR_8 = {'disabled': 0, 'is_stock_item': 1}\n",
"if VAR_0.get('item_group'):\n",
"VAR_8.update({'item_group': VAR_0['item_group']})\n",
"if VAR_0.get('item'):\n",
"VAR_8.update({'name': VAR_0['item']})\n",
"VAR_3 = frappe.get_all('Item', fields=['name', 'item_group', 'item_name'],\n VAR_0=filters_dict, order_by='name')\n",
"return VAR_3\n"
] |
[
"def get_items(filters):...\n",
"filters_dict = {'disabled': 0, 'is_stock_item': 1}\n",
"if filters.get('item_group'):\n",
"filters_dict.update({'item_group': filters['item_group']})\n",
"if filters.get('item'):\n",
"filters_dict.update({'name': filters['item']})\n",
"items = frappe.get_all('Item', fields=['name', 'item_group', 'item_name'],\n filters=filters_dict, order_by='name')\n",
"return items\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Condition",
"Expr'",
"Condition",
"Expr'",
"Assign'",
"Return'"
] |
[
"def FUNC_2(self, VAR_5=None):...\n",
"VAR_25 = []\n",
"if self.automaton:\n",
"VAR_25.append(('automaton', '=', self.automaton.id))\n",
"VAR_25.append(('automaton', '=', self._get_model_automaton().id))\n",
"return VAR_25\n"
] |
[
"def _get_state_domain(self, domain=None):...\n",
"result = []\n",
"if self.automaton:\n",
"result.append(('automaton', '=', self.automaton.id))\n",
"result.append(('automaton', '=', self._get_model_automaton().id))\n",
"return result\n"
] |
[
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Condition",
"Expr'",
"Expr'",
"Return'"
] |
[
"def FUNC_37(self, VAR_14, VAR_18):...\n",
"VAR_31 = self.get_related_model(VAR_14)\n",
"return self.session.query(VAR_31).get(VAR_18)\n"
] |
[
"def get_related_obj(self, col_name, value):...\n",
"rel_model = self.get_related_model(col_name)\n",
"return self.session.query(rel_model).get(value)\n"
] |
[
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Return'"
] |
[
"import argparse\n",
"import json\n",
"import logging\n",
"import os\n",
"import subprocess\n",
"import sys\n",
"import tempfile\n",
"from pystache import context\n",
"from os_apply_config import collect_config\n",
"from os_apply_config import config_exception as exc\n",
"from os_apply_config import renderers\n",
"from os_apply_config import value_types\n",
"from os_apply_config import version\n",
"VAR_0 = '/usr/libexec/os-apply-config/templates'\n",
"def FUNC_0():...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_30 = os.environ.get('OS_CONFIG_APPLIER_TEMPLATES', None)\n",
"if VAR_30 is None:\n",
"VAR_30 = '/opt/stack/os-apply-config/templates'\n",
"return VAR_30\n",
"if not os.path.isdir(VAR_30):\n",
"VAR_30 = '/opt/stack/os-config-applier/templates'\n",
"if os.path.isdir(VAR_30) and not os.path.isdir(VAR_0):\n",
"logging.warning(\n 'Template directory %s is deprecated. The recommended location for template files is %s'\n , VAR_30, VAR_0)\n",
"VAR_30 = VAR_0\n"
] |
[
"import argparse\n",
"import json\n",
"import logging\n",
"import os\n",
"import subprocess\n",
"import sys\n",
"import tempfile\n",
"from pystache import context\n",
"from os_apply_config import collect_config\n",
"from os_apply_config import config_exception as exc\n",
"from os_apply_config import renderers\n",
"from os_apply_config import value_types\n",
"from os_apply_config import version\n",
"DEFAULT_TEMPLATES_DIR = '/usr/libexec/os-apply-config/templates'\n",
"def templates_dir():...\n",
"\"\"\"docstring\"\"\"\n",
"templates_dir = os.environ.get('OS_CONFIG_APPLIER_TEMPLATES', None)\n",
"if templates_dir is None:\n",
"templates_dir = '/opt/stack/os-apply-config/templates'\n",
"return templates_dir\n",
"if not os.path.isdir(templates_dir):\n",
"templates_dir = '/opt/stack/os-config-applier/templates'\n",
"if os.path.isdir(templates_dir) and not os.path.isdir(DEFAULT_TEMPLATES_DIR):\n",
"logging.warning(\n 'Template directory %s is deprecated. The recommended location for template files is %s'\n , templates_dir, DEFAULT_TEMPLATES_DIR)\n",
"templates_dir = DEFAULT_TEMPLATES_DIR\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"Import'",
"Import'",
"Import'",
"Import'",
"Import'",
"Import'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Assign'",
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Assign'",
"Return'",
"Condition",
"Assign'",
"Condition",
"Expr'",
"Assign'"
] |
[
"@property...\n",
"return super(CLASS_0, self).content_security_policy + '; sandbox allow-scripts'\n"
] |
[
"@property...\n",
"return super(FilesHandler, self\n ).content_security_policy + '; sandbox allow-scripts'\n"
] |
[
0,
0
] |
[
"Condition",
"Return'"
] |
[
"def FUNC_2(VAR_1):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_7 = {}\n",
"for VAR_24, deps in VAR_1.items():\n",
"VAR_16 = [d.check.name for d in deps]\n",
"VAR_8 = set()\n",
"VAR_7[VAR_24.check.name] += VAR_16\n",
"VAR_7[VAR_24.check.name] = VAR_16\n",
"VAR_9 = list(itertools.zip_longest(VAR_7.keys(), [], fillvalue=None))\n",
"VAR_10 = []\n",
"while VAR_9:\n",
"VAR_17, VAR_18 = VAR_9.pop()\n",
"while VAR_10 and VAR_10[-1] != VAR_18:\n",
"VAR_10.pop()\n",
"VAR_19 = reversed(VAR_7[VAR_17])\n",
"VAR_10.append(VAR_17)\n",
"for n in VAR_19:\n",
"if n in VAR_10:\n",
"VAR_8.add(VAR_17)\n",
"VAR_25 = '->'.join(VAR_10 + [n])\n",
"if n not in VAR_8:\n",
"VAR_9.append((n, VAR_17))\n"
] |
[
"def validate_deps(graph):...\n",
"\"\"\"docstring\"\"\"\n",
"test_graph = {}\n",
"for case, deps in graph.items():\n",
"test_deps = [d.check.name for d in deps]\n",
"visited = set()\n",
"test_graph[case.check.name] += test_deps\n",
"test_graph[case.check.name] = test_deps\n",
"unvisited = list(itertools.zip_longest(test_graph.keys(), [], fillvalue=None))\n",
"path = []\n",
"while unvisited:\n",
"node, parent = unvisited.pop()\n",
"while path and path[-1] != parent:\n",
"path.pop()\n",
"adjacent = reversed(test_graph[node])\n",
"path.append(node)\n",
"for n in adjacent:\n",
"if n in path:\n",
"visited.add(node)\n",
"cycle_str = '->'.join(path + [n])\n",
"if n not in visited:\n",
"unvisited.append((n, node))\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"For",
"Assign'",
"Assign'",
"AugAssign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Expr'",
"Assign'",
"Expr'",
"For",
"Condition",
"Expr'",
"Assign'",
"Condition",
"Expr'"
] |
[
"@api_post...\n",
"VAR_7 = VAR_3.pop('submit_id', None)\n",
"VAR_0.submit(VAR_7=submit_id, config=body)\n",
"return JsonResponse({'status': True, 'submit_id': VAR_7}, encoder=JsonSerialize\n )\n"
] |
[
"@api_post...\n",
"submit_id = body.pop('submit_id', None)\n",
"submit_manager.submit(submit_id=submit_id, config=body)\n",
"return JsonResponse({'status': True, 'submit_id': submit_id}, encoder=\n JsonSerialize)\n"
] |
[
0,
0,
0,
0
] |
[
"Condition",
"Assign'",
"Expr'",
"Return'"
] |
[
"\"\"\"Django models for the redirects app.\"\"\"\n",
"import logging\n",
"import re\n",
"from django.db import models\n",
"from django.utils.translation import ugettext\n",
"from django.utils.translation import ugettext_lazy as _\n",
"from readthedocs.core.resolver import resolve_path\n",
"from readthedocs.projects.models import Project\n",
"from .managers import RedirectManager\n",
"VAR_0 = logging.getLogger(__name__)\n",
"VAR_1 = (301, _('301 - Permanent Redirect')), (302, _(\n '302 - Temporary Redirect'))\n",
"VAR_2 = (True, _('Active')), (False, _('Inactive'))\n",
"VAR_3 = ('prefix', _('Prefix Redirect')), ('page', _('Page Redirect')), (\n 'exact', _('Exact Redirect')), ('sphinx_html', _('Sphinx HTMLDir -> HTML')\n ), ('sphinx_htmldir', _('Sphinx HTML -> HTMLDir'))\n",
"VAR_4 = _(\n 'Absolute path, excluding the domain. Example: <b>/docs/</b> or <b>/install.html</b>'\n )\n",
"VAR_5 = _('Absolute or relative URL. Example: <b>/tutorial/install.html</b>')\n",
"VAR_6 = _('The type of redirect you wish to use.')\n",
"\"\"\"A HTTP redirect associated with a Project.\"\"\"\n",
"VAR_7 = models.ForeignKey(Project, VAR_20=_('Project'), related_name=\n 'redirects')\n",
"VAR_8 = models.CharField(_('Redirect Type'), max_length=255, choices=\n TYPE_CHOICES, help_text=redirect_type_helptext)\n",
"VAR_9 = models.CharField(_('From URL'), max_length=255, db_index=True,\n help_text=from_url_helptext, blank=True)\n",
"VAR_10 = models.CharField(_('To URL'), max_length=255, db_index=True,\n help_text=to_url_helptext, blank=True)\n",
"VAR_11 = models.SmallIntegerField(_('HTTP Status'), choices=\n HTTP_STATUS_CHOICES, default=301)\n",
"VAR_12 = models.BooleanField(choices=STATUS_CHOICES, default=True)\n",
"VAR_13 = models.DateTimeField(auto_now_add=True)\n",
"VAR_14 = models.DateTimeField(auto_now=True)\n",
"VAR_15 = RedirectManager()\n",
"VAR_20 = _('redirect')\n",
"VAR_21 = _('redirects')\n",
"VAR_22 = '-update_dt',\n",
"def __str__(self):...\n",
"VAR_23 = '{type}: {from_to_url}'\n",
"if self.redirect_type in ['prefix', 'page', 'exact']:\n",
"return VAR_23.format(type=self.get_redirect_type_display(), from_to_url=\n self.get_from_to_url_display())\n",
"return ugettext('Redirect: {}'.format(self.get_redirect_type_display()))\n"
] |
[
"\"\"\"Django models for the redirects app.\"\"\"\n",
"import logging\n",
"import re\n",
"from django.db import models\n",
"from django.utils.translation import ugettext\n",
"from django.utils.translation import ugettext_lazy as _\n",
"from readthedocs.core.resolver import resolve_path\n",
"from readthedocs.projects.models import Project\n",
"from .managers import RedirectManager\n",
"log = logging.getLogger(__name__)\n",
"HTTP_STATUS_CHOICES = (301, _('301 - Permanent Redirect')), (302, _(\n '302 - Temporary Redirect'))\n",
"STATUS_CHOICES = (True, _('Active')), (False, _('Inactive'))\n",
"TYPE_CHOICES = ('prefix', _('Prefix Redirect')), ('page', _('Page Redirect')\n ), ('exact', _('Exact Redirect')), ('sphinx_html', _(\n 'Sphinx HTMLDir -> HTML')), ('sphinx_htmldir', _('Sphinx HTML -> HTMLDir'))\n",
"from_url_helptext = _(\n 'Absolute path, excluding the domain. Example: <b>/docs/</b> or <b>/install.html</b>'\n )\n",
"to_url_helptext = _(\n 'Absolute or relative URL. Example: <b>/tutorial/install.html</b>')\n",
"redirect_type_helptext = _('The type of redirect you wish to use.')\n",
"\"\"\"A HTTP redirect associated with a Project.\"\"\"\n",
"project = models.ForeignKey(Project, verbose_name=_('Project'),\n related_name='redirects')\n",
"redirect_type = models.CharField(_('Redirect Type'), max_length=255,\n choices=TYPE_CHOICES, help_text=redirect_type_helptext)\n",
"from_url = models.CharField(_('From URL'), max_length=255, db_index=True,\n help_text=from_url_helptext, blank=True)\n",
"to_url = models.CharField(_('To URL'), max_length=255, db_index=True,\n help_text=to_url_helptext, blank=True)\n",
"http_status = models.SmallIntegerField(_('HTTP Status'), choices=\n HTTP_STATUS_CHOICES, default=301)\n",
"status = models.BooleanField(choices=STATUS_CHOICES, default=True)\n",
"create_dt = models.DateTimeField(auto_now_add=True)\n",
"update_dt = models.DateTimeField(auto_now=True)\n",
"objects = RedirectManager()\n",
"verbose_name = _('redirect')\n",
"verbose_name_plural = _('redirects')\n",
"ordering = '-update_dt',\n",
"def __str__(self):...\n",
"redirect_text = '{type}: {from_to_url}'\n",
"if self.redirect_type in ['prefix', 'page', 'exact']:\n",
"return redirect_text.format(type=self.get_redirect_type_display(),\n from_to_url=self.get_from_to_url_display())\n",
"return ugettext('Redirect: {}'.format(self.get_redirect_type_display()))\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"Expr'",
"Import'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"FunctionDef'",
"Assign'",
"Condition",
"Return'",
"Return'"
] |
[
"def FUNC_21(self, VAR_15):...\n",
"\"\"\"docstring\"\"\"\n",
"for key, value in self.iscsi_ips.items():\n",
"if value['nsp'] == VAR_15:\n",
"return key\n"
] |
[
"def _get_ip_using_nsp(self, nsp):...\n",
"\"\"\"docstring\"\"\"\n",
"for key, value in self.iscsi_ips.items():\n",
"if value['nsp'] == nsp:\n",
"return key\n"
] |
[
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"For",
"Condition",
"Return'"
] |
[
"def FUNC_38(self, VAR_16, VAR_11=None, VAR_22=None, VAR_23=False, VAR_24=False...\n",
"from frappe.utils.formatters import format_value\n",
"VAR_25 = self.meta.get_field(VAR_16)\n",
"if not VAR_25 and VAR_16 in default_fields:\n",
"from frappe.model.meta import get_default_df\n",
"VAR_46 = self.get(VAR_16)\n",
"VAR_25 = get_default_df(VAR_16)\n",
"if VAR_24:\n",
"VAR_46 = _(VAR_46)\n",
"if VAR_23 and isinstance(VAR_46, (int, float)):\n",
"VAR_46 = abs(self.get(VAR_16))\n",
"if not VAR_11:\n",
"VAR_11 = getattr(self, 'parent_doc', None) or self\n",
"return format_value(VAR_46, VAR_25=df, VAR_11=doc, VAR_22=currency)\n"
] |
[
"def get_formatted(self, fieldname, doc=None, currency=None, absolute_value=...\n",
"from frappe.utils.formatters import format_value\n",
"df = self.meta.get_field(fieldname)\n",
"if not df and fieldname in default_fields:\n",
"from frappe.model.meta import get_default_df\n",
"val = self.get(fieldname)\n",
"df = get_default_df(fieldname)\n",
"if translated:\n",
"val = _(val)\n",
"if absolute_value and isinstance(val, (int, float)):\n",
"val = abs(self.get(fieldname))\n",
"if not doc:\n",
"doc = getattr(self, 'parent_doc', None) or self\n",
"return format_value(val, df=df, doc=doc, currency=currency)\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"Condition",
"ImportFrom'",
"Assign'",
"Condition",
"ImportFrom'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Return'"
] |
[
"def FUNC_16(self):...\n",
"VAR_6 = None\n",
"VAR_6 = tempfile.NamedTemporaryFile(delete=False)\n",
"if VAR_6:\n",
"VAR_6.close()\n",
"os.remove(VAR_6.name)\n",
"os.chmod(VAR_6.name, stat.S_IREAD)\n",
"self._mox.StubOutWithMock(url_helper, 'UrlOpen')\n",
"url_helper.UrlOpen(mox.IgnoreArg(), method='GET').AndReturn('data')\n",
"self._mox.ReplayAll()\n",
"self.assertFalse(url_helper.DownloadFile(VAR_6.name, 'http://www.fakeurl.com'))\n",
"self._mox.VerifyAll()\n"
] |
[
"def testDownloadFileSavingErrors(self):...\n",
"file_readonly = None\n",
"file_readonly = tempfile.NamedTemporaryFile(delete=False)\n",
"if file_readonly:\n",
"file_readonly.close()\n",
"os.remove(file_readonly.name)\n",
"os.chmod(file_readonly.name, stat.S_IREAD)\n",
"self._mox.StubOutWithMock(url_helper, 'UrlOpen')\n",
"url_helper.UrlOpen(mox.IgnoreArg(), method='GET').AndReturn('data')\n",
"self._mox.ReplayAll()\n",
"self.assertFalse(url_helper.DownloadFile(file_readonly.name,\n 'http://www.fakeurl.com'))\n",
"self._mox.VerifyAll()\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Condition",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_2(self):...\n",
"VAR_6 = self.client.get('/api/apps')\n",
"self.assertEqual(VAR_6.status_code, 200)\n",
"self.assertEqual(len(VAR_6.data['results']), 2)\n",
"VAR_8 = VAR_6.data['results'][0]['id']\n",
"self.assertTrue(self.client.login(VAR_1='autotest-2', VAR_2='password'))\n",
"VAR_6 = self.client.get('/api/apps')\n",
"self.assertEqual(len(VAR_6.data['results']), 1)\n",
"for model in ['builds', 'config', 'containers', 'limits', 'releases']:\n",
"VAR_6 = self.client.get('/api/apps/{}/{}/'.format(VAR_8, model))\n",
"self.assertTrue(self.client.login(VAR_1='autotest-1', VAR_2='password'))\n",
"self.assertEqual(VAR_6.data['detail'], 'Not found')\n",
"VAR_5 = '/api/apps/{}/perms'.format(VAR_8)\n",
"VAR_7 = {'username': 'autotest-2'}\n",
"VAR_6 = self.client.post(VAR_5, json.dumps(VAR_7), content_type=\n 'application/json')\n",
"self.assertEqual(VAR_6.status_code, 201)\n",
"self.assertTrue(self.client.login(VAR_1='autotest-2', VAR_2='password'))\n",
"VAR_6 = self.client.get('/api/apps')\n",
"self.assertEqual(VAR_6.status_code, 200)\n",
"self.assertEqual(len(VAR_6.data['results']), 2)\n",
"for model in ['builds', 'containers', 'releases']:\n",
"VAR_6 = self.client.get('/api/apps/{}/{}/'.format(VAR_8, model))\n",
"self.assertEqual(len(VAR_6.data['results']), 0)\n"
] |
[
"def test_create(self):...\n",
"response = self.client.get('/api/apps')\n",
"self.assertEqual(response.status_code, 200)\n",
"self.assertEqual(len(response.data['results']), 2)\n",
"app_id = response.data['results'][0]['id']\n",
"self.assertTrue(self.client.login(username='autotest-2', password='password'))\n",
"response = self.client.get('/api/apps')\n",
"self.assertEqual(len(response.data['results']), 1)\n",
"for model in ['builds', 'config', 'containers', 'limits', 'releases']:\n",
"response = self.client.get('/api/apps/{}/{}/'.format(app_id, model))\n",
"self.assertTrue(self.client.login(username='autotest-1', password='password'))\n",
"self.assertEqual(response.data['detail'], 'Not found')\n",
"url = '/api/apps/{}/perms'.format(app_id)\n",
"body = {'username': 'autotest-2'}\n",
"response = self.client.post(url, json.dumps(body), content_type=\n 'application/json')\n",
"self.assertEqual(response.status_code, 201)\n",
"self.assertTrue(self.client.login(username='autotest-2', password='password'))\n",
"response = self.client.get('/api/apps')\n",
"self.assertEqual(response.status_code, 200)\n",
"self.assertEqual(len(response.data['results']), 2)\n",
"for model in ['builds', 'containers', 'releases']:\n",
"response = self.client.get('/api/apps/{}/{}/'.format(app_id, model))\n",
"self.assertEqual(len(response.data['results']), 0)\n"
] |
[
0,
5,
0,
0,
0,
5,
0,
0,
0,
5,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"For",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"For",
"Assign'",
"Expr'"
] |
[
"def FUNC_22(self, VAR_6):...\n",
"VAR_49 = self.nodes.get(VAR_6['name'])\n",
"VAR_50 = []\n",
"VAR_51 = []\n",
"dep_resolve(VAR_49, VAR_50, VAR_51)\n",
"VAR_50.remove(VAR_49)\n",
"return VAR_50\n"
] |
[
"def get_dep_list(self, comp):...\n",
"node = self.nodes.get(comp['name'])\n",
"res = []\n",
"unres = []\n",
"dep_resolve(node, res, unres)\n",
"res.remove(node)\n",
"return res\n"
] |
[
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Return'"
] |
[
"def FUNC_21(self, VAR_5):...\n",
"\"\"\"docstring\"\"\"\n",
"for VAR_36 in self.config['nodes']:\n",
"return False\n",
"VAR_36 = fnmatch.translate(VAR_36)\n",
"VAR_1 = 'Error comparing %s to provided node regex %s: %s'\n",
"if re.match(VAR_36, VAR_5):\n",
"self.log_debug(VAR_1 % (VAR_5, VAR_36, err))\n",
"return True\n"
] |
[
"def compare_node_to_regex(self, node):...\n",
"\"\"\"docstring\"\"\"\n",
"for regex in self.config['nodes']:\n",
"return False\n",
"regex = fnmatch.translate(regex)\n",
"msg = 'Error comparing %s to provided node regex %s: %s'\n",
"if re.match(regex, node):\n",
"self.log_debug(msg % (node, regex, err))\n",
"return True\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"For",
"Return'",
"Assign'",
"Assign'",
"Condition",
"Expr'",
"Return'"
] |
[
"def __init__(self, VAR_1):...\n",
"self.database = lite.connect(VAR_0)\n"
] |
[
"def __init__(self, bot):...\n",
"self.database = lite.connect(DATABASE_PATH)\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Assign'"
] |
[
"def FUNC_4(self):...\n",
"self.response.set_status(403)\n",
"return\n"
] |
[
"def post(self):...\n",
"self.response.set_status(403)\n",
"return\n"
] |
[
0,
5,
5
] |
[
"FunctionDef'",
"Expr'",
"Return'"
] |
[
"def FUNC_23():...\n",
"return threading.current_thread().getName() == 'MainThread'\n"
] |
[
"def is_main_thread():...\n",
"return threading.current_thread().getName() == 'MainThread'\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Return'"
] |
[
"def FUNC_6(self, VAR_23):...\n",
"VAR_23 = FUNC_1(VAR_23)\n",
"if not VAR_23:\n",
"return self.error()\n",
"VAR_15 = Subreddit._by_name(VAR_23)\n",
"return VAR_23\n",
"return self.error(errors.SUBREDDIT_EXISTS)\n"
] |
[
"def run(self, name):...\n",
"name = chksrname(name)\n",
"if not name:\n",
"return self.error()\n",
"a = Subreddit._by_name(name)\n",
"return name\n",
"return self.error(errors.SUBREDDIT_EXISTS)\n"
] |
[
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Condition",
"Return'",
"Assign'",
"Return'",
"Return'"
] |
[
"def FUNC_6(self):...\n",
"for VAR_6 in self.configurations_for_servers:\n",
"VAR_16 = {'status': self.STOPPED, 'config': CONF.get('lsp-server', VAR_6),\n 'instance': None}\n",
"if VAR_6 not in self.clients:\n",
"self.clients[VAR_6] = VAR_16\n",
"VAR_0.debug(self.clients[VAR_6]['config'] != VAR_16['config'])\n",
"self.register_queue[VAR_6] = []\n",
"VAR_19 = self.clients[VAR_6]['config']\n",
"VAR_20 = VAR_16['config']\n",
"VAR_21 = ['cmd', 'args', 'host', 'port', 'external']\n",
"VAR_22 = any([(VAR_19[x] != VAR_20[x]) for x in VAR_21])\n",
"if VAR_22:\n",
"if self.clients[VAR_6]['status'] == self.STOPPED:\n",
"if self.clients[VAR_6]['status'] == self.RUNNING:\n",
"self.clients[VAR_6] = VAR_16\n",
"if self.clients[VAR_6]['status'] == self.RUNNING:\n",
"VAR_23 = self.clients[VAR_6]['instance']\n",
"self.close_client(VAR_6)\n",
"VAR_23.send_plugin_configurations(VAR_20['configurations'])\n",
"self.clients[VAR_6] = VAR_16\n",
"self.start_lsp_client(VAR_6)\n"
] |
[
"def update_server_list(self):...\n",
"for language in self.configurations_for_servers:\n",
"config = {'status': self.STOPPED, 'config': CONF.get('lsp-server', language\n ), 'instance': None}\n",
"if language not in self.clients:\n",
"self.clients[language] = config\n",
"logger.debug(self.clients[language]['config'] != config['config'])\n",
"self.register_queue[language] = []\n",
"current_config = self.clients[language]['config']\n",
"new_config = config['config']\n",
"restart_diff = ['cmd', 'args', 'host', 'port', 'external']\n",
"restart = any([(current_config[x] != new_config[x]) for x in restart_diff])\n",
"if restart:\n",
"if self.clients[language]['status'] == self.STOPPED:\n",
"if self.clients[language]['status'] == self.RUNNING:\n",
"self.clients[language] = config\n",
"if self.clients[language]['status'] == self.RUNNING:\n",
"client = self.clients[language]['instance']\n",
"self.close_client(language)\n",
"client.send_plugin_configurations(new_config['configurations'])\n",
"self.clients[language] = config\n",
"self.start_lsp_client(language)\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"For",
"Assign'",
"Condition",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Condition",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Expr'"
] |
[
"def FUNC_14(self, VAR_19):...\n",
"\"\"\"docstring\"\"\"\n",
"return any(VAR_19 in VAR_46 for VAR_46 in [' '.join(VAR_46) for VAR_46 in\n self.lines])\n"
] |
[
"def contains(self, command):...\n",
"\"\"\"docstring\"\"\"\n",
"return any(command in line for line in [' '.join(line) for line in self.lines])\n"
] |
[
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Return'"
] |
[
"def __init__(self, VAR_7, VAR_20=True, *VAR_15, **VAR_16):...\n",
"CLASS_6.__init__(self, VAR_7, *VAR_15, VAR_20=redirect, **kw)\n"
] |
[
"def __init__(self, param, redirect=True, *a, **kw):...\n",
"VMeetup.__init__(self, param, *a, redirect=redirect, **kw)\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Expr'"
] |
[
"def FUNC_0(VAR_2):...\n",
"\"\"\"docstring\"\"\"\n",
"return os.path.getsize(VAR_2)\n"
] |
[
"def get_file_size(path):...\n",
"\"\"\"docstring\"\"\"\n",
"return os.path.getsize(path)\n"
] |
[
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Return'"
] |
[
"def FUNC_11(self):...\n",
"self.driver._eql_execute = self.mox.CreateMock(self.driver._eql_execute)\n",
"VAR_1 = {'name': self.volume_name}\n",
"self.stubs.Set(self.driver, '_get_iscsi_properties', self.\n _fake_get_iscsi_properties)\n",
"self.driver._eql_execute('volume', 'select', VAR_1['name'], 'access',\n 'create', 'initiator', self.connector['initiator'], 'authmethod chap',\n 'username', self.configuration.eqlx_chap_login)\n",
"self.mox.ReplayAll()\n",
"VAR_8 = self.driver.initialize_connection(VAR_1, self.connector)\n",
"self.assertEqual(VAR_8['data'], self._fake_get_iscsi_properties(VAR_1))\n"
] |
[
"def test_initialize_connection(self):...\n",
"self.driver._eql_execute = self.mox.CreateMock(self.driver._eql_execute)\n",
"volume = {'name': self.volume_name}\n",
"self.stubs.Set(self.driver, '_get_iscsi_properties', self.\n _fake_get_iscsi_properties)\n",
"self.driver._eql_execute('volume', 'select', volume['name'], 'access',\n 'create', 'initiator', self.connector['initiator'], 'authmethod chap',\n 'username', self.configuration.eqlx_chap_login)\n",
"self.mox.ReplayAll()\n",
"iscsi_properties = self.driver.initialize_connection(volume, self.connector)\n",
"self.assertEqual(iscsi_properties['data'], self._fake_get_iscsi_properties(\n volume))\n"
] |
[
0,
0,
0,
0,
2,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Expr'"
] |
[
"@classmethod...\n",
"VAR_8 = VAR_12.routes or [VAR_12.navbar_tab_url]\n",
"return [webapp2.Route(r, VAR_12) for r in VAR_8]\n"
] |
[
"@classmethod...\n",
"routes = cls.routes or [cls.navbar_tab_url]\n",
"return [webapp2.Route(r, cls) for r in routes]\n"
] |
[
0,
0,
0
] |
[
"Condition",
"Assign'",
"Return'"
] |
[
"def FUNC_6(self):...\n",
"self.assertTrue({'name': 'DocField'} in DatabaseQuery('DocType').execute(\n filters={'name': 'DocField'}))\n"
] |
[
"def test_filters_4(self):...\n",
"self.assertTrue({'name': 'DocField'} in DatabaseQuery('DocType').execute(\n filters={'name': 'DocField'}))\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Expr'"
] |
[
"def FUNC_19(VAR_10, VAR_3, VAR_7):...\n",
"VAR_21 = VAR_3 + '-' + VAR_7\n",
"VAR_12 = 0\n",
"VAR_28 = VAR_21, VAR_12\n",
"VAR_18 = 'INSERT INTO {} VALUES (?, ?)'.format(CFG('options_table_name'))\n",
"VAR_10.execute(VAR_18, VAR_28)\n"
] |
[
"def insertOption(c, poll_name, option):...\n",
"key = poll_name + '-' + option\n",
"count = 0\n",
"params = key, count\n",
"req = 'INSERT INTO {} VALUES (?, ?)'.format(CFG('options_table_name'))\n",
"c.execute(req, params)\n"
] |
[
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_21(self, VAR_17, VAR_16):...\n",
"\"\"\"docstring\"\"\"\n",
"self._check_volume(VAR_16)\n",
"VAR_0.warn(_('Volume %s is not found!, it may have been deleted'), VAR_16[\n 'name'])\n",
"VAR_0.error(_('Failed to ensure export of volume %s'), VAR_16['name'])\n"
] |
[
"def ensure_export(self, context, volume):...\n",
"\"\"\"docstring\"\"\"\n",
"self._check_volume(volume)\n",
"LOG.warn(_('Volume %s is not found!, it may have been deleted'), volume['name']\n )\n",
"LOG.error(_('Failed to ensure export of volume %s'), volume['name'])\n"
] |
[
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_27(self, VAR_43, VAR_44):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_63 = self.getfile(VAR_43)\n",
"if VAR_63 == False:\n",
"VAR_64 = self.getfile(VAR_44)\n",
"if VAR_64 != False:\n",
"self.get_path(os.path.dirname(VAR_43)).remove(VAR_63)\n",
"VAR_63[VAR_1] = os.path.basename(VAR_44)\n",
"self.get_path(os.path.dirname(VAR_44)).append(VAR_63)\n",
"return\n"
] |
[
"def rename(self, oldpath, newpath):...\n",
"\"\"\"docstring\"\"\"\n",
"old = self.getfile(oldpath)\n",
"if old == False:\n",
"new = self.getfile(newpath)\n",
"if new != False:\n",
"self.get_path(os.path.dirname(oldpath)).remove(old)\n",
"old[A_NAME] = os.path.basename(newpath)\n",
"self.get_path(os.path.dirname(newpath)).append(old)\n",
"return\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Expr'",
"Assign'",
"Expr'",
"Return'"
] |
[
"def FUNC_17(self):...\n",
"self.assertFalse(zf._allowZip64)\n"
] |
[
"def test_open_zipFalse(self):...\n",
"self.assertFalse(zf._allowZip64)\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Expr'"
] |
[
"@VAR_0.route('/user_view')...\n",
"VAR_7 = get_db()\n",
"VAR_8 = VAR_2.split('.')[0]\n",
"VAR_9 = VAR_2.split('.')[1]\n",
"VAR_10 = 'string'.format(VAR_9)\n",
"VAR_11 = VAR_7.execute(VAR_10, (VAR_8,)).fetchall()\n",
"return render_template('admin/userview.html', VAR_11=users, VAR_2='{}.{}'.\n format(sortBy, sortOrder))\n"
] |
[
"@bp.route('/user_view')...\n",
"db = get_db()\n",
"sortBy = sort.split('.')[0]\n",
"sortOrder = sort.split('.')[1]\n",
"query = (\n 'SELECT * FROM user AS u LEFT OUTER JOIN (SELECT uid, count(uid) AS follower FROM follows GROUP BY uid) AS f ON u.id = f.uid ORDER BY ? {}'\n .format(sortOrder))\n",
"users = db.execute(query, (sortBy,)).fetchall()\n",
"return render_template('admin/userview.html', users=users, sort='{}.{}'.\n format(sortBy, sortOrder))\n"
] |
[
0,
0,
0,
0,
4,
4,
0
] |
[
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_0(self, VAR_0, VAR_1, VAR_2, VAR_3):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_6 = VAR_3.pop('kbsite_name', None)\n",
"if not VAR_6:\n",
"VAR_6 = 'default'\n",
"VAR_0.kbsite = models.KegbotSite.objects.get(name=kbsite_name)\n",
"return None\n"
] |
[
"def process_view(self, request, view_func, view_args, view_kwargs):...\n",
"\"\"\"docstring\"\"\"\n",
"kbsite_name = view_kwargs.pop('kbsite_name', None)\n",
"if not kbsite_name:\n",
"kbsite_name = 'default'\n",
"request.kbsite = models.KegbotSite.objects.get(name=kbsite_name)\n",
"return None\n"
] |
[
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Return'"
] |
[
"def __init__(self, VAR_1, VAR_7, VAR_8, VAR_9, VAR_10=None, VAR_4=None):...\n",
"GenericRequest.__init__(self, VAR_1, VAR_4)\n",
"self.url = '%stasks/%s/submit' % (self.base_url, VAR_7[1])\n",
"self.task = VAR_7\n",
"self.submission_format = VAR_8\n",
"self.filenames = VAR_9\n",
"self.data = {}\n",
"if VAR_10 is None:\n",
"for filename in VAR_9:\n",
"if VAR_10 is not None:\n",
"VAR_23 = filename_to_language(filename)\n",
"self.data = {'language': VAR_10}\n",
"if VAR_23 is not None:\n",
"VAR_10 = VAR_23.name\n"
] |
[
"def __init__(self, browser, task, submission_format, filenames, language=...\n",
"GenericRequest.__init__(self, browser, base_url)\n",
"self.url = '%stasks/%s/submit' % (self.base_url, task[1])\n",
"self.task = task\n",
"self.submission_format = submission_format\n",
"self.filenames = filenames\n",
"self.data = {}\n",
"if language is None:\n",
"for filename in filenames:\n",
"if language is not None:\n",
"lang = filename_to_language(filename)\n",
"self.data = {'language': language}\n",
"if lang is not None:\n",
"language = lang.name\n"
] |
[
0,
0,
5,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"For",
"Condition",
"Assign'",
"Assign'",
"Condition",
"Assign'"
] |
[
"def FUNC_22(VAR_37):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_64 = {}\n",
"for VAR_95, info in iteritems(VAR_37):\n",
"if not VAR_95.nostandalone:\n",
"return VAR_64\n",
"VAR_64[VAR_95] = info\n"
] |
[
"def clean_before_output(kw_matches):...\n",
"\"\"\"docstring\"\"\"\n",
"filtered_kw_matches = {}\n",
"for kw_match, info in iteritems(kw_matches):\n",
"if not kw_match.nostandalone:\n",
"return filtered_kw_matches\n",
"filtered_kw_matches[kw_match] = info\n"
] |
[
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"For",
"Condition",
"Return'",
"Assign'"
] |
[
"def FUNC_6(self, VAR_7):...\n",
"return self._job_text[VAR_7]\n"
] |
[
"def copy_text(self, job_id):...\n",
"return self._job_text[job_id]\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Return'"
] |
[
"def FUNC_20(VAR_8):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_8 = os.path.normpath(VAR_8)\n",
"VAR_33 = re.search('{[^{]', VAR_8)\n",
"VAR_34 = os.path.dirname(VAR_8[:VAR_33.start()]\n ) if VAR_33 else os.path.dirname(VAR_8)\n",
"if not VAR_34:\n",
"VAR_34 = '.'\n",
"VAR_35 = [VAR_29.group('name') for VAR_29 in VAR_5.finditer(VAR_8)]\n",
"VAR_36 = namedtuple('Wildcards', VAR_35)\n",
"VAR_11 = VAR_36(*[list() for VAR_40 in VAR_35])\n",
"VAR_8 = re.compile(FUNC_8(VAR_8))\n",
"for dirpath, dirnames, filenames in os.walk(VAR_34):\n",
"for VAR_0 in chain(filenames, dirnames):\n",
"return VAR_11\n",
"if dirpath != '.':\n",
"VAR_0 = os.path.join(dirpath, VAR_0)\n",
"VAR_29 = re.match(VAR_8, VAR_0)\n",
"if VAR_29:\n",
"for VAR_40, VAR_16 in VAR_29.groupdict().items():\n",
"getattr(VAR_11, VAR_40).append(VAR_16)\n"
] |
[
"def glob_wildcards(pattern):...\n",
"\"\"\"docstring\"\"\"\n",
"pattern = os.path.normpath(pattern)\n",
"first_wildcard = re.search('{[^{]', pattern)\n",
"dirname = os.path.dirname(pattern[:first_wildcard.start()]\n ) if first_wildcard else os.path.dirname(pattern)\n",
"if not dirname:\n",
"dirname = '.'\n",
"names = [match.group('name') for match in _wildcard_regex.finditer(pattern)]\n",
"Wildcards = namedtuple('Wildcards', names)\n",
"wildcards = Wildcards(*[list() for name in names])\n",
"pattern = re.compile(regex(pattern))\n",
"for dirpath, dirnames, filenames in os.walk(dirname):\n",
"for f in chain(filenames, dirnames):\n",
"return wildcards\n",
"if dirpath != '.':\n",
"f = os.path.join(dirpath, f)\n",
"match = re.match(pattern, f)\n",
"if match:\n",
"for name, value in match.groupdict().items():\n",
"getattr(wildcards, name).append(value)\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"For",
"For",
"Return'",
"Condition",
"Assign'",
"Assign'",
"Condition",
"For",
"Expr'"
] |
[
"def __init__(self, VAR_7, VAR_67=(), *VAR_15, **VAR_16):...\n",
"CLASS_0.__init__(self, VAR_7, *VAR_15, **kw)\n",
"self.options = VAR_67\n"
] |
[
"def __init__(self, param, options=(), *a, **kw):...\n",
"Validator.__init__(self, param, *a, **kw)\n",
"self.options = options\n"
] |
[
0,
0,
0
] |
[
"FunctionDef'",
"Expr'",
"Assign'"
] |
[
"def __init__(self, VAR_9, VAR_12):...\n",
"self._splitpath = fs.split_path(VAR_9)\n",
"self._lensplitpath = len(self._splitpath)\n",
"self._store = {}\n",
"self._ds = VAR_12\n"
] |
[
"def __init__(self, targetpath, client):...\n",
"self._splitpath = fs.split_path(targetpath)\n",
"self._lensplitpath = len(self._splitpath)\n",
"self._store = {}\n",
"self._ds = client\n"
] |
[
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Assign'"
] |
[
"def FUNC_2(self, VAR_0, *VAR_2, **VAR_3):...\n",
"VAR_11 = JustURLForm(VAR_0.POST or None)\n",
"if VAR_11.is_valid():\n",
"VAR_18 = VAR_11.cleaned_data['input_url']\n",
"return render(VAR_0, 'home.html', {'form': VAR_11})\n",
"VAR_21 = VAR_11.cleaned_data['short_url']\n",
"VAR_19 = VAR_11.cleaned_data['category']\n",
"if JustURL.objects.filter(short_url__contains=short_url).exists():\n",
"VAR_29 = 'Token is already in use'\n",
"VAR_20 = JustURL.objects.create(input_url=url, VAR_21=\n f'{request.get_host()}/{short_url}', VAR_19=category)\n",
"return render(VAR_0, 'custom-short-url.html', {'form': JustURLForm,\n 'message': VAR_29})\n",
"VAR_20.save()\n",
"if VAR_0.user.is_superuser:\n",
"return redirect(reverse('url-detail-view', VAR_3={'pk': created.pk}))\n",
"return redirect(reverse('success-url-view', VAR_3={'pk': created.pk}))\n"
] |
[
"def post(self, request, *args, **kwargs):...\n",
"form = JustURLForm(request.POST or None)\n",
"if form.is_valid():\n",
"url = form.cleaned_data['input_url']\n",
"return render(request, 'home.html', {'form': form})\n",
"short_url = form.cleaned_data['short_url']\n",
"category = form.cleaned_data['category']\n",
"if JustURL.objects.filter(short_url__contains=short_url).exists():\n",
"message = 'Token is already in use'\n",
"created = JustURL.objects.create(input_url=url, short_url=\n f'{request.get_host()}/{short_url}', category=category)\n",
"return render(request, 'custom-short-url.html', {'form': JustURLForm,\n 'message': message})\n",
"created.save()\n",
"if request.user.is_superuser:\n",
"return redirect(reverse('url-detail-view', kwargs={'pk': created.pk}))\n",
"return redirect(reverse('success-url-view', kwargs={'pk': created.pk}))\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Condition",
"Assign'",
"Return'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Return'",
"Expr'",
"Condition",
"Return'",
"Return'"
] |
[
"def __init__(self, VAR_6, VAR_7=None, VAR_8=None, VAR_9=None):...\n",
"self._conn = VAR_6\n",
"self._url_prefix = VAR_7 or ''\n",
"self._default_headers = VAR_8 or {}\n"
] |
[
"def __init__(self, connection, url_prefix=None, default_headers=None,...\n",
"self._conn = connection\n",
"self._url_prefix = url_prefix or ''\n",
"self._default_headers = default_headers or {}\n"
] |
[
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'"
] |
[
"def FUNC_5(self):...\n",
"VAR_5 = [{'application': self.TEST_APP, 'stack': self.TEST_STACK,\n 'credentials': self.bindings['GCE_CREDENTIALS'], 'loadBalancers': [self\n .__lb_name], 'targetSize': 1, 'capacity': {'min': 1, 'max': 1,\n 'desired': 1}, 'zone': self.TEST_ZONE, 'network': 'default',\n 'instanceMetadata': {'load-balancer-names': self.__lb_name},\n 'availabilityZones': {self.TEST_REGION: [self.TEST_ZONE]},\n 'cloudProvider': 'gce', 'source': {'account': self.bindings[\n 'GCE_CREDENTIALS'], 'region': self.TEST_REGION, 'zone': self.TEST_ZONE,\n 'serverGroupName': self.__server_group_name, 'asgName': self.\n __server_group_name}, 'instanceType': 'f1-micro', 'image': self.\n bindings['TEST_GCE_IMAGE_NAME'], 'initialNumReplicas': 1,\n 'loadBalancers': [self.__lb_name], 'type': 'cloneServerGroup',\n 'account': self.bindings['GCE_CREDENTIALS'], 'user': 'integration-tests'}]\n",
"VAR_6 = gcp.GceContractBuilder(self.gce_observer)\n",
"VAR_6.new_clause_builder('Server Group Cloned', retryable_for_secs=90\n ).list_resources('managed-instance-groups').contains_path_value(\n 'baseInstanceName', self.__cloned_server_group_name)\n",
"VAR_7 = self.agent.make_json_payload_from_kwargs(VAR_5=job, description=\n 'Server Group Test - clone server group', application=self.TEST_APP)\n",
"return st.OperationContract(self.new_post_operation(title=\n 'clone_server_group', data=payload, path=self.__path), contract=builder\n .build())\n"
] |
[
"def clone_server_group(self):...\n",
"job = [{'application': self.TEST_APP, 'stack': self.TEST_STACK,\n 'credentials': self.bindings['GCE_CREDENTIALS'], 'loadBalancers': [self\n .__lb_name], 'targetSize': 1, 'capacity': {'min': 1, 'max': 1,\n 'desired': 1}, 'zone': self.TEST_ZONE, 'network': 'default',\n 'instanceMetadata': {'load-balancer-names': self.__lb_name},\n 'availabilityZones': {self.TEST_REGION: [self.TEST_ZONE]},\n 'cloudProvider': 'gce', 'source': {'account': self.bindings[\n 'GCE_CREDENTIALS'], 'region': self.TEST_REGION, 'zone': self.TEST_ZONE,\n 'serverGroupName': self.__server_group_name, 'asgName': self.\n __server_group_name}, 'instanceType': 'f1-micro', 'image': self.\n bindings['TEST_GCE_IMAGE_NAME'], 'initialNumReplicas': 1,\n 'loadBalancers': [self.__lb_name], 'type': 'cloneServerGroup',\n 'account': self.bindings['GCE_CREDENTIALS'], 'user': 'integration-tests'}]\n",
"builder = gcp.GceContractBuilder(self.gce_observer)\n",
"builder.new_clause_builder('Server Group Cloned', retryable_for_secs=90\n ).list_resources('managed-instance-groups').contains_path_value(\n 'baseInstanceName', self.__cloned_server_group_name)\n",
"payload = self.agent.make_json_payload_from_kwargs(job=job, description=\n 'Server Group Test - clone server group', application=self.TEST_APP)\n",
"return st.OperationContract(self.new_post_operation(title=\n 'clone_server_group', data=payload, path=self.__path), contract=builder\n .build())\n"
] |
[
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Return'"
] |
[
"@staticmethod...\n",
"\"\"\"docstring\"\"\"\n",
"return User.get_all_users(VAR_12)\n"
] |
[
"@staticmethod...\n",
"\"\"\"docstring\"\"\"\n",
"return User.get_all_users(query)\n"
] |
[
0,
0,
0
] |
[
"Condition",
"Docstring",
"Return'"
] |
[
"def FUNC_17(self):...\n",
"return self.get('__islocal')\n"
] |
[
"def is_new(self):...\n",
"return self.get('__islocal')\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Return'"
] |
[
"def FUNC_5(VAR_18, VAR_20, VAR_21):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_36 = VAR_3 if VAR_18 else VAR_2\n",
"return len(VAR_18) >= VAR_1 or VAR_20 - VAR_21 > VAR_36\n"
] |
[
"def should_post_update(stdout, now, last_packet):...\n",
"\"\"\"docstring\"\"\"\n",
"packet_interval = MIN_PACKET_INTERNAL if stdout else MAX_PACKET_INTERVAL\n",
"return len(stdout) >= MAX_CHUNK_SIZE or now - last_packet > packet_interval\n"
] |
[
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"Return'"
] |
[
"def __init__(self, VAR_6):...\n",
"self.logger = VAR_6.bind(context=type(self).__name__)\n"
] |
[
"def __init__(self, logger):...\n",
"self.logger = logger.bind(context=type(self).__name__)\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Assign'"
] |
[
"def FUNC_0(VAR_2, VAR_3, VAR_4, VAR_5=None):...\n",
"VAR_13 = VAR_2\n",
"if isinstance(VAR_13, str):\n",
"VAR_13 = [VAR_13]\n",
"for istr in VAR_13:\n",
"if not VAR_5:\n",
"exec(istr)\n",
"if isinstance(VAR_2, str):\n",
"if sys.version_info[0] == 2:\n",
"return\n",
"VAR_5 = \"Failed '%s'.\" % VAR_2\n",
"VAR_5 = 'Unable to do any of %s.' % VAR_13\n",
"VAR_19 = VAR_3\n",
"VAR_19 = VAR_4\n"
] |
[
"def check_import(imports, py2pkgs, py3pkgs, message=None):...\n",
"import_group = imports\n",
"if isinstance(import_group, str):\n",
"import_group = [import_group]\n",
"for istr in import_group:\n",
"if not message:\n",
"exec(istr)\n",
"if isinstance(imports, str):\n",
"if sys.version_info[0] == 2:\n",
"return\n",
"message = \"Failed '%s'.\" % imports\n",
"message = 'Unable to do any of %s.' % import_group\n",
"pkgs = py2pkgs\n",
"pkgs = py3pkgs\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Condition",
"Assign'",
"For",
"Condition",
"Expr'",
"Condition",
"Condition",
"Return'",
"Assign'",
"Assign'",
"Assign'",
"Assign'"
] |
[
"def FUNC_3(self, VAR_5):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_5 = VAR_5[:]\n",
"for request in VAR_5:\n",
"self.assertEqual(tuple, request.__class__)\n",
"self.assertEqual([], self._requests)\n",
"self.assertIn(len(request), (3, 4))\n",
"self._requests = VAR_5\n"
] |
[
"def expected_requests(self, requests):...\n",
"\"\"\"docstring\"\"\"\n",
"requests = requests[:]\n",
"for request in requests:\n",
"self.assertEqual(tuple, request.__class__)\n",
"self.assertEqual([], self._requests)\n",
"self.assertIn(len(request), (3, 4))\n",
"self._requests = requests\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"For",
"Expr'",
"Expr'",
"Expr'",
"Assign'"
] |
[
"def FUNC_19(VAR_5):...\n",
"VAR_8 = {'Authorization': 'token ' + os.environ['GITHUB_TOKEN']}\n",
"VAR_9 = os.environ['BOT_USERNAME'], os.environ['BOT_PASSWORD']\n",
"VAR_42 = VAR_5.get('fork_fullname')\n",
"for VAR_47, new_file in VAR_5['results'].items():\n",
"VAR_10 = 'https://api.github.com/repos/{}/contents/{}'\n",
"VAR_10 = VAR_10.format(VAR_42, VAR_47)\n",
"VAR_56 = {'ref': VAR_5['new_branch']}\n",
"VAR_11 = requests.get(VAR_10, VAR_56=params, VAR_8=headers, VAR_9=auth)\n",
"VAR_57 = VAR_11.json().get('sha')\n",
"VAR_56['path'] = VAR_47\n",
"VAR_58 = base64.b64encode(new_file.encode()).decode('utf-8')\n",
"VAR_40 = {'path': VAR_47, 'message': 'Fix pep8 errors in {}'.format(VAR_47),\n 'content': VAR_58, 'sha': VAR_57, 'branch': VAR_5.get('new_branch')}\n",
"VAR_11 = requests.put(VAR_10, json=request_json, VAR_8=headers, VAR_9=auth)\n"
] |
[
"def commit(data):...\n",
"headers = {'Authorization': 'token ' + os.environ['GITHUB_TOKEN']}\n",
"auth = os.environ['BOT_USERNAME'], os.environ['BOT_PASSWORD']\n",
"fullname = data.get('fork_fullname')\n",
"for file, new_file in data['results'].items():\n",
"url = 'https://api.github.com/repos/{}/contents/{}'\n",
"url = url.format(fullname, file)\n",
"params = {'ref': data['new_branch']}\n",
"r = requests.get(url, params=params, headers=headers, auth=auth)\n",
"sha_blob = r.json().get('sha')\n",
"params['path'] = file\n",
"content_code = base64.b64encode(new_file.encode()).decode('utf-8')\n",
"request_json = {'path': file, 'message': 'Fix pep8 errors in {}'.format(\n file), 'content': content_code, 'sha': sha_blob, 'branch': data.get(\n 'new_branch')}\n",
"r = requests.put(url, json=request_json, headers=headers, auth=auth)\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"For",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'"
] |
[
"from osv import osv\n",
"from tools.translate import _\n",
"import time\n",
"VAR_0 = 'pos.open.statement'\n",
"VAR_1 = 'Open Statements'\n",
"def FUNC_0(self, VAR_2, VAR_3, VAR_4, VAR_5):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_6 = []\n",
"VAR_7 = self.pool.get('ir.model.data')\n",
"VAR_8 = self.pool.get('res.users').browse(VAR_2, VAR_3, VAR_3).company_id.id\n",
"VAR_9 = self.pool.get('account.bank.statement')\n",
"VAR_10 = self.pool.get('ir.sequence')\n",
"VAR_11 = self.pool.get('account.journal')\n",
"VAR_2.execute(\n 'select DISTINCT journal_id from pos_journal_users where user_id=%d order by journal_id'\n % VAR_3)\n",
"VAR_12 = map(lambda x1: x1[0], VAR_2.fetchall())\n",
"VAR_2.execute('string' % ','.join(map(lambda x: \"'\" + str(x) + \"'\", VAR_12)))\n",
"VAR_13 = map(lambda x1: x1[0], VAR_2.fetchall())\n",
"for journal in VAR_11.browse(VAR_2, VAR_3, VAR_13):\n",
"VAR_4 = VAR_9.search(VAR_2, VAR_3, [('state', '!=', 'confirm'), ('user_id',\n '=', VAR_3), ('journal_id', '=', journal.id)])\n",
"VAR_14 = self.pool.get('ir.model.data')\n",
"if len(VAR_4):\n",
"VAR_15 = VAR_14._get_id(VAR_2, VAR_3, 'account', 'view_bank_statement_tree')\n",
"VAR_17 = ''\n",
"VAR_16 = VAR_14._get_id(VAR_2, VAR_3, 'account', 'view_bank_statement_form2')\n",
"if journal.sequence_id:\n",
"if VAR_15:\n",
"VAR_17 = VAR_10.get_id(VAR_2, VAR_3, journal.sequence_id.id)\n",
"VAR_17 = VAR_10.get(VAR_2, VAR_3, 'account.bank.statement')\n",
"VAR_15 = VAR_14.browse(VAR_2, VAR_3, VAR_15, VAR_5=context).res_id\n",
"if VAR_16:\n",
"VAR_18 = VAR_9.create(VAR_2, VAR_3, {'journal_id': journal.id, 'company_id':\n VAR_8, 'user_id': VAR_3, 'state': 'open', 'name': VAR_17,\n 'starting_details_ids': VAR_9._get_cash_close_box_lines(VAR_2, VAR_3, [])})\n",
"VAR_16 = VAR_14.browse(VAR_2, VAR_3, VAR_16, VAR_5=context).res_id\n",
"return {'domain': \"[('state','=','open')]\", 'name': 'Open Statement',\n 'view_type': 'form', 'view_mode': 'tree,form', 'res_model':\n 'account.bank.statement', 'views': [(VAR_15, 'tree'), (VAR_16, 'form')],\n 'type': 'ir.actions.act_window'}\n",
"VAR_9.button_open(VAR_2, VAR_3, [VAR_18], VAR_5)\n"
] |
[
"from osv import osv\n",
"from tools.translate import _\n",
"import time\n",
"_name = 'pos.open.statement'\n",
"_description = 'Open Statements'\n",
"def open_statement(self, cr, uid, ids, context):...\n",
"\"\"\"docstring\"\"\"\n",
"list_statement = []\n",
"mod_obj = self.pool.get('ir.model.data')\n",
"company_id = self.pool.get('res.users').browse(cr, uid, uid).company_id.id\n",
"statement_obj = self.pool.get('account.bank.statement')\n",
"sequence_obj = self.pool.get('ir.sequence')\n",
"journal_obj = self.pool.get('account.journal')\n",
"cr.execute(\n 'select DISTINCT journal_id from pos_journal_users where user_id=%d order by journal_id'\n % uid)\n",
"j_ids = map(lambda x1: x1[0], cr.fetchall())\n",
"cr.execute(\n \"\"\" select id from account_journal\n where auto_cash='True' and type='cash'\n and id in (%s)\"\"\"\n % ','.join(map(lambda x: \"'\" + str(x) + \"'\", j_ids)))\n",
"journal_ids = map(lambda x1: x1[0], cr.fetchall())\n",
"for journal in journal_obj.browse(cr, uid, journal_ids):\n",
"ids = statement_obj.search(cr, uid, [('state', '!=', 'confirm'), ('user_id',\n '=', uid), ('journal_id', '=', journal.id)])\n",
"data_obj = self.pool.get('ir.model.data')\n",
"if len(ids):\n",
"id2 = data_obj._get_id(cr, uid, 'account', 'view_bank_statement_tree')\n",
"number = ''\n",
"id3 = data_obj._get_id(cr, uid, 'account', 'view_bank_statement_form2')\n",
"if journal.sequence_id:\n",
"if id2:\n",
"number = sequence_obj.get_id(cr, uid, journal.sequence_id.id)\n",
"number = sequence_obj.get(cr, uid, 'account.bank.statement')\n",
"id2 = data_obj.browse(cr, uid, id2, context=context).res_id\n",
"if id3:\n",
"statement_id = statement_obj.create(cr, uid, {'journal_id': journal.id,\n 'company_id': company_id, 'user_id': uid, 'state': 'open', 'name':\n number, 'starting_details_ids': statement_obj._get_cash_close_box_lines\n (cr, uid, [])})\n",
"id3 = data_obj.browse(cr, uid, id3, context=context).res_id\n",
"return {'domain': \"[('state','=','open')]\", 'name': 'Open Statement',\n 'view_type': 'form', 'view_mode': 'tree,form', 'res_model':\n 'account.bank.statement', 'views': [(id2, 'tree'), (id3, 'form')],\n 'type': 'ir.actions.act_window'}\n",
"statement_obj.button_open(cr, uid, [statement_id], context)\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
4,
4,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"ImportFrom'",
"ImportFrom'",
"Import'",
"Assign'",
"Assign'",
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Assign'",
"For",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Return'",
"Expr'"
] |
[
"def FUNC_2(VAR_7=None, VAR_8=0, VAR_9=0):...\n",
"\"\"\"docstring\"\"\"\n",
"if VAR_7 is None:\n",
"return True\n",
"return False\n",
"VAR_21 = re.match('^([^;]+)(; length=([0-9]+))?$', VAR_7, re.IGNORECASE)\n",
"VAR_22 = parse_http_date(VAR_21.group(1))\n",
"VAR_23 = VAR_21.group(3)\n",
"if VAR_23 and int(VAR_23) != VAR_9:\n",
"if int(VAR_8) > VAR_22:\n"
] |
[
"def was_modified_since(header=None, mtime=0, size=0):...\n",
"\"\"\"docstring\"\"\"\n",
"if header is None:\n",
"return True\n",
"return False\n",
"matches = re.match('^([^;]+)(; length=([0-9]+))?$', header, re.IGNORECASE)\n",
"header_mtime = parse_http_date(matches.group(1))\n",
"header_len = matches.group(3)\n",
"if header_len and int(header_len) != size:\n",
"if int(mtime) > header_mtime:\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Condition",
"Return'",
"Return'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Condition"
] |
[
"def FUNC_17(self, VAR_2, VAR_3, VAR_4, *VAR_5):...\n",
"self.write_data({'type': 'market_transaction_complete', 'event': VAR_5[0]})\n"
] |
[
"def on_market_transaction_complete(self, subject, changetype, objectID, *args):...\n",
"self.write_data({'type': 'market_transaction_complete', 'event': args[0]})\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Expr'"
] |
[
"def FUNC_2(self, VAR_3, VAR_4):...\n",
"if not VAR_4.is_course_staff:\n",
"VAR_8 = VAR_4.module\n",
"return True\n",
"return self.has_object_permission(VAR_3, VAR_4, VAR_8)\n"
] |
[
"def has_permission(self, request, view):...\n",
"if not view.is_course_staff:\n",
"module = view.module\n",
"return True\n",
"return self.has_object_permission(request, view, module)\n"
] |
[
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Condition",
"Assign'",
"Return'",
"Return'"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return set(VAR_28 for VAR_28 in self.input if not VAR_28.exists and not \n VAR_28 in self.subworkflow_input)\n"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return set(f for f in self.input if not f.exists and not f in self.\n subworkflow_input)\n"
] |
[
0,
0,
0
] |
[
"Condition",
"Docstring",
"Return'"
] |
[
"def FUNC_36(VAR_11):...\n",
"for VAR_51, VAR_58 in VAR_11.items():\n",
"if isinstance(VAR_58, str) or not isinstance(VAR_58, Iterable):\n",
"VAR_58 = [VAR_58]\n",
"yield [(VAR_51, VAR_16) for VAR_16 in VAR_58]\n"
] |
[
"def flatten(wildcards):...\n",
"for wildcard, values in wildcards.items():\n",
"if isinstance(values, str) or not isinstance(values, Iterable):\n",
"values = [values]\n",
"yield [(wildcard, value) for value in values]\n"
] |
[
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"For",
"Condition",
"Assign'",
"Expr'"
] |
[
"def FUNC_2(VAR_1, VAR_2, VAR_3='', VAR_4=None):...\n",
"\"\"\"docstring\"\"\"\n",
"return FUNC_4(VAR_1, VAR_2['oauth_token'], VAR_6=resp['oauth_token_secret'],\n VAR_4=extra_data, VAR_3=token_type)\n"
] |
[
"def oauth1_token_setter(remote, resp, token_type='', extra_data=None):...\n",
"\"\"\"docstring\"\"\"\n",
"return token_setter(remote, resp['oauth_token'], secret=resp[\n 'oauth_token_secret'], extra_data=extra_data, token_type=token_type)\n"
] |
[
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Return'"
] |
[
"def FUNC_5(self):...\n",
"VAR_28 = self.state.client_table()\n",
"self.local_scheduler_id_to_ip_map = {}\n",
"for local_scheduler_info in VAR_28:\n",
"VAR_36 = local_scheduler_info.get('DBClientID') or local_scheduler_info[\n 'ClientID']\n",
"VAR_44 = (local_scheduler_info.get('AuxAddress') or local_scheduler_info[\n 'NodeManagerAddress']).split(':')[0]\n",
"self.local_scheduler_id_to_ip_map[VAR_36] = VAR_44\n"
] |
[
"def update_local_scheduler_map(self):...\n",
"local_schedulers = self.state.client_table()\n",
"self.local_scheduler_id_to_ip_map = {}\n",
"for local_scheduler_info in local_schedulers:\n",
"client_id = local_scheduler_info.get('DBClientID') or local_scheduler_info[\n 'ClientID']\n",
"ip_address = (local_scheduler_info.get('AuxAddress') or\n local_scheduler_info['NodeManagerAddress']).split(':')[0]\n",
"self.local_scheduler_id_to_ip_map[client_id] = ip_address\n"
] |
[
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"For",
"Assign'",
"Assign'",
"Assign'"
] |
[
"@property...\n",
"return 'saml'\n"
] |
[
"@property...\n",
"return 'saml'\n"
] |
[
0,
0
] |
[
"Condition",
"Return'"
] |
[
"def FUNC_1(VAR_2, VAR_3, VAR_4, VAR_5=False):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_7 = []\n",
"VAR_8 = False\n",
"VAR_9 = os.path.join(VAR_2, VAR_4)\n",
"if VAR_5:\n",
"VAR_13 = file_to_read.read()\n",
"for line in file_to_read:\n",
"if VAR_3 in VAR_13:\n",
"if VAR_3 in line:\n",
"if VAR_8:\n",
"if re.search('^import .*models.*|import .*models.*', VAR_13):\n",
"VAR_8 = True\n",
"VAR_7.append(line)\n",
"for line in VAR_7:\n",
"print(\"sed -i 's/([^ ]+)\\\\.{0}([^ \\\\n]+)/\\\\1\\\\2.{0}/g' {1}\".format(VAR_3,\n VAR_4))\n",
"if re.search('^from .*models.*|from .*models.*', VAR_13):\n",
"if 'from' not in line:\n",
"file_to_write.write(line)\n",
"print(\"sed -i 's/([^ ]+)\\\\.{0} import (\\\\w+)/\\\\1.\\\\2 import {0}/g' {1}\".\n format(VAR_3, VAR_4))\n",
"VAR_7.append(re.sub('(?P<pre>[^ ]+)\\\\.{}(?P<post>[^ \\\\n]+)'.format(VAR_3),\n '\\\\g<pre>\\\\g<post>.{}'.format(VAR_3), line))\n",
"VAR_7.append(re.sub('(?P<pre>[^ ]+)\\\\.{} import (?P<post>\\\\w+)'.format(\n VAR_3), '\\\\g<pre>.\\\\g<post> import {}'.format(VAR_3), line))\n"
] |
[
"def change_import_path(directory, module, src_file, dry_run=False):...\n",
"\"\"\"docstring\"\"\"\n",
"new_lines = []\n",
"module_present = False\n",
"current_file = os.path.join(directory, src_file)\n",
"if dry_run:\n",
"containt = file_to_read.read()\n",
"for line in file_to_read:\n",
"if module in containt:\n",
"if module in line:\n",
"if module_present:\n",
"if re.search('^import .*models.*|import .*models.*', containt):\n",
"module_present = True\n",
"new_lines.append(line)\n",
"for line in new_lines:\n",
"print(\"sed -i 's/([^ ]+)\\\\.{0}([^ \\\\n]+)/\\\\1\\\\2.{0}/g' {1}\".format(module,\n src_file))\n",
"if re.search('^from .*models.*|from .*models.*', containt):\n",
"if 'from' not in line:\n",
"file_to_write.write(line)\n",
"print(\"sed -i 's/([^ ]+)\\\\.{0} import (\\\\w+)/\\\\1.\\\\2 import {0}/g' {1}\".\n format(module, src_file))\n",
"new_lines.append(re.sub('(?P<pre>[^ ]+)\\\\.{}(?P<post>[^ \\\\n]+)'.format(\n module), '\\\\g<pre>\\\\g<post>.{}'.format(module), line))\n",
"new_lines.append(re.sub('(?P<pre>[^ ]+)\\\\.{} import (?P<post>\\\\w+)'.format(\n module), '\\\\g<pre>.\\\\g<post> import {}'.format(module), line))\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"For",
"Condition",
"Condition",
"Condition",
"Condition",
"Assign'",
"Expr'",
"For",
"Expr'",
"Condition",
"Condition",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_31():...\n",
"VAR_24 = vim.eval('tagfiles()')\n",
"VAR_25 = VAR_0.getcwd()\n",
"return [VAR_0.path.join(VAR_25, x) for x in VAR_24]\n"
] |
[
"def GetTagFiles():...\n",
"tag_files = vim.eval('tagfiles()')\n",
"current_working_directory = os.getcwd()\n",
"return [os.path.join(current_working_directory, x) for x in tag_files]\n"
] |
[
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Return'"
] |
[
"import json\n",
"from django.contrib.postgres import lookups\n",
"from django.contrib.postgres.forms import SimpleArrayField\n",
"from django.contrib.postgres.validators import ArrayMaxLengthValidator\n",
"from django.core import checks, exceptions\n",
"from django.db.models import Field, IntegerField, Transform\n",
"from django.db.models.lookups import Exact, In\n",
"from django.utils.translation import gettext_lazy as _\n",
"from ..utils import prefix_validation_error\n",
"from .mixins import CheckFieldDefaultMixin\n",
"from .utils import AttributeSetter\n",
"__all__ = ['ArrayField']\n",
"VAR_0 = False\n",
"VAR_1 = {'item_invalid': _('Item %(nth)s in the array did not validate:'),\n 'nested_array_mismatch': _('Nested arrays must have the same length.')}\n",
"VAR_2 = 'list', '[]'\n",
"def __init__(self, VAR_3, VAR_4=None, **VAR_5):...\n",
"self.base_field = VAR_3\n",
"self.size = VAR_4\n",
"if self.size:\n",
"self.default_validators = [*self.default_validators,\n ArrayMaxLengthValidator(self.size)]\n",
"if hasattr(self.base_field, 'from_db_value'):\n",
"self.from_db_value = self._from_db_value\n",
"super().__init__(**kwargs)\n",
"@property...\n",
"return self.__dict__['model']\n",
"@VAR_6.setter...\n",
"self.__dict__['model'] = VAR_6\n",
"self.base_field.model = VAR_6\n",
"def FUNC_1(self, **VAR_5):...\n",
"VAR_22 = super().check(**kwargs)\n",
"if self.base_field.remote_field:\n",
"VAR_22.append(checks.Error(\n 'Base field for array cannot be a related field.', VAR_13=self, id=\n 'postgres.E002'))\n",
"VAR_31 = self.base_field.check()\n",
"return VAR_22\n",
"if VAR_31:\n",
"VAR_32 = '\\n '.join('%s (%s)' % (error.msg, error.id) for error in VAR_31)\n",
"VAR_22.append(checks.Error(\"\"\"Base field for array has errors:\n %s\"\"\" %\n VAR_32, VAR_13=self, id='postgres.E001'))\n"
] |
[
"import json\n",
"from django.contrib.postgres import lookups\n",
"from django.contrib.postgres.forms import SimpleArrayField\n",
"from django.contrib.postgres.validators import ArrayMaxLengthValidator\n",
"from django.core import checks, exceptions\n",
"from django.db.models import Field, IntegerField, Transform\n",
"from django.db.models.lookups import Exact, In\n",
"from django.utils.translation import gettext_lazy as _\n",
"from ..utils import prefix_validation_error\n",
"from .mixins import CheckFieldDefaultMixin\n",
"from .utils import AttributeSetter\n",
"__all__ = ['ArrayField']\n",
"empty_strings_allowed = False\n",
"default_error_messages = {'item_invalid': _(\n 'Item %(nth)s in the array did not validate:'), 'nested_array_mismatch':\n _('Nested arrays must have the same length.')}\n",
"_default_hint = 'list', '[]'\n",
"def __init__(self, base_field, size=None, **kwargs):...\n",
"self.base_field = base_field\n",
"self.size = size\n",
"if self.size:\n",
"self.default_validators = [*self.default_validators,\n ArrayMaxLengthValidator(self.size)]\n",
"if hasattr(self.base_field, 'from_db_value'):\n",
"self.from_db_value = self._from_db_value\n",
"super().__init__(**kwargs)\n",
"@property...\n",
"return self.__dict__['model']\n",
"@model.setter...\n",
"self.__dict__['model'] = model\n",
"self.base_field.model = model\n",
"def check(self, **kwargs):...\n",
"errors = super().check(**kwargs)\n",
"if self.base_field.remote_field:\n",
"errors.append(checks.Error(\n 'Base field for array cannot be a related field.', obj=self, id=\n 'postgres.E002'))\n",
"base_errors = self.base_field.check()\n",
"return errors\n",
"if base_errors:\n",
"messages = '\\n '.join('%s (%s)' % (error.msg, error.id) for error in\n base_errors)\n",
"errors.append(checks.Error(\"\"\"Base field for array has errors:\n %s\"\"\" %\n messages, obj=self, id='postgres.E001'))\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"Import'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"FunctionDef'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Expr'",
"Condition",
"Return'",
"Condition",
"Assign'",
"Assign'",
"FunctionDef'",
"Assign'",
"Condition",
"Expr'",
"Assign'",
"Return'",
"Condition",
"Assign'",
"Expr'"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self._icon\n"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self._icon\n"
] |
[
0,
0,
0
] |
[
"Condition",
"Docstring",
"Return'"
] |
[
"def __init__(self, VAR_20, VAR_21):...\n",
"self.message = VAR_21\n",
"self.code = VAR_20\n"
] |
[
"def __init__(self, code, message):...\n",
"self.message = message\n",
"self.code = code\n"
] |
[
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'"
] |
[
"def FUNC_6(self, VAR_40, VAR_41):...\n",
"VAR_77, VAR_47, VAR_78 = VAR_40.partition(VAR_41)\n",
"if VAR_47 != VAR_41:\n",
"if VAR_78.startswith('^'):\n",
"return int(VAR_40), 0\n",
"VAR_78 = VAR_78[1:].replace('{', '').replace('}', '')\n",
"VAR_78 = ''.join(map(lambda x: self.supunmap.get(x, x), VAR_78))\n",
"if not VAR_78:\n",
"VAR_78 = 1\n",
"VAR_78 = int(VAR_78)\n",
"if not VAR_77:\n",
"VAR_77 = 1\n",
"VAR_77 = int(VAR_77)\n",
"return VAR_77, VAR_78\n"
] |
[
"def _parse_coeff(self, cstr, var):...\n",
"coefficient, _, exponent = cstr.partition(var)\n",
"if _ != var:\n",
"if exponent.startswith('^'):\n",
"return int(cstr), 0\n",
"exponent = exponent[1:].replace('{', '').replace('}', '')\n",
"exponent = ''.join(map(lambda x: self.supunmap.get(x, x), exponent))\n",
"if not exponent:\n",
"exponent = 1\n",
"exponent = int(exponent)\n",
"if not coefficient:\n",
"coefficient = 1\n",
"coefficient = int(coefficient)\n",
"return coefficient, exponent\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Condition",
"Condition",
"Return'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_1(self, VAR_3, VAR_5, VAR_6):...\n",
"VAR_13 = VAR_5.enrollment_audience\n",
"VAR_14 = VAR_6.userprofile.is_external\n",
"VAR_15 = VAR_5.ENROLLMENT_AUDIENCE\n",
"if VAR_13 == VAR_15.INTERNAL_USERS and VAR_14:\n",
"self.error_msg(_('This course is only for internal students.'))\n",
"if VAR_13 == VAR_15.EXTERNAL_USERS and not VAR_14:\n",
"return False\n",
"self.error_msg(_('This course is only for external students.'))\n",
"return True\n",
"return False\n"
] |
[
"def enrollment_audience_check(self, request, course, user):...\n",
"audience = course.enrollment_audience\n",
"external = user.userprofile.is_external\n",
"EA = course.ENROLLMENT_AUDIENCE\n",
"if audience == EA.INTERNAL_USERS and external:\n",
"self.error_msg(_('This course is only for internal students.'))\n",
"if audience == EA.EXTERNAL_USERS and not external:\n",
"return False\n",
"self.error_msg(_('This course is only for external students.'))\n",
"return True\n",
"return False\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Expr'",
"Condition",
"Return'",
"Expr'",
"Return'",
"Return'"
] |
[
"from __future__ import unicode_literals, print_function, division\n",
"import logging\n",
"import argparse\n",
"from jinja2.loaders import FileSystemLoader\n",
"from veil.frontend.template import *\n",
"from veil.frontend.cli import *\n",
"from veil.environment import *\n",
"from veil.environment.setting import *\n",
"from .tornado import *\n",
"from .locale import *\n",
"from .routing import *\n",
"from .static_file import *\n",
"from .xsrf import *\n",
"from .web_installer import load_website_config\n",
"VAR_0 = logging.getLogger(__name__)\n",
"VAR_1 = {}\n",
"def FUNC_0(VAR_2, VAR_3):...\n",
"VAR_1.setdefault(VAR_2.lower(), []).append(VAR_3)\n",
"@script('up')...\n",
"VAR_8 = argparse.ArgumentParser('Website')\n",
"VAR_8.add_argument('purpose', help='which website to bring up')\n",
"VAR_8.add_argument('--dependency', type=str, help=\n 'where @periodic_job is defined', nargs='+', dest='dependencies')\n",
"VAR_9 = VAR_8.parse_args(VAR_4)\n",
"for dependency in VAR_9.dependencies:\n",
"__import__(dependency)\n",
"FUNC_3(VAR_9.purpose)\n",
"def FUNC_2(VAR_5, **VAR_6):...\n",
"VAR_7 = load_website_config(VAR_5)\n",
"VAR_10 = FUNC_4(VAR_5, **kwargs)\n",
"VAR_11 = start_test_http_server(VAR_10, host=config.host, port=config.port)\n",
"VAR_11.purpose = VAR_5\n",
"return VAR_11\n"
] |
[
"from __future__ import unicode_literals, print_function, division\n",
"import logging\n",
"import argparse\n",
"from jinja2.loaders import FileSystemLoader\n",
"from veil.frontend.template import *\n",
"from veil.frontend.cli import *\n",
"from veil.environment import *\n",
"from veil.environment.setting import *\n",
"from .tornado import *\n",
"from .locale import *\n",
"from .routing import *\n",
"from .static_file import *\n",
"from .xsrf import *\n",
"from .web_installer import load_website_config\n",
"LOGGER = logging.getLogger(__name__)\n",
"additional_context_managers = {}\n",
"def register_website_context_manager(website, context_manager):...\n",
"additional_context_managers.setdefault(website.lower(), []).append(\n context_manager)\n",
"@script('up')...\n",
"argument_parser = argparse.ArgumentParser('Website')\n",
"argument_parser.add_argument('purpose', help='which website to bring up')\n",
"argument_parser.add_argument('--dependency', type=str, help=\n 'where @periodic_job is defined', nargs='+', dest='dependencies')\n",
"args = argument_parser.parse_args(argv)\n",
"for dependency in args.dependencies:\n",
"__import__(dependency)\n",
"start_website(args.purpose)\n",
"def start_test_website(purpose, **kwargs):...\n",
"config = load_website_config(purpose)\n",
"http_handler = create_website_http_handler(purpose, **kwargs)\n",
"http_server = start_test_http_server(http_handler, host=config.host, port=\n config.port)\n",
"http_server.purpose = purpose\n",
"return http_server\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
5,
0,
0,
0
] |
[
"ImportFrom'",
"Import'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Assign'",
"Assign'",
"FunctionDef'",
"Expr'",
"Condition",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"For",
"Expr'",
"Expr'",
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_1(VAR_6, VAR_7, VAR_8):...\n",
"VAR_8.debug('Running component check for %s' % VAR_6['name'])\n",
"VAR_33 = len(VAR_6['cmd']) > 1 and 'check' in VAR_6['cmd'][1]\n",
"VAR_9 = FUNC_6(VAR_7, VAR_6['name'])\n",
"if VAR_9:\n",
"VAR_56 = FUNC_2(VAR_9)\n",
"VAR_8.debug('%s window is not running. Running custom check' % VAR_6['name'])\n",
"VAR_8.debug('Found window pid: %s' % VAR_56)\n",
"if VAR_33 and FUNC_0(VAR_6):\n",
"VAR_57 = []\n",
"VAR_8.debug('Component was not started by Hyperion, but the check succeeded')\n",
"VAR_8.debug(\n 'Window not running and no check command is available or it failed: returning false'\n )\n",
"for entry in VAR_56:\n",
"return CLASS_0.STARTED_BY_HAND\n",
"return CLASS_0.STOPPED\n",
"VAR_57.extend(Process(entry).children(recursive=True))\n",
"VAR_58 = [p.pid for p in VAR_57]\n",
"VAR_8.debug('Window is running %s child processes' % len(VAR_58))\n",
"if len(VAR_58) < 3:\n",
"VAR_8.debug(\n 'Main window process has finished. Running custom check if available')\n",
"if VAR_33 and FUNC_0(VAR_6):\n",
"if VAR_33 and FUNC_0(VAR_6):\n",
"VAR_8.debug('Check succeeded')\n",
"if not VAR_33:\n",
"VAR_8.debug('Process terminated but check was successful')\n",
"VAR_8.debug('Check failed or no check available: returning false')\n",
"return CLASS_0.RUNNING\n",
"VAR_8.debug(\n 'No custom check specified and got sufficient pid amount: returning true')\n",
"VAR_8.debug('Check failed: returning false')\n",
"return CLASS_0.STOPPED_BUT_SUCCESSFUL\n",
"return CLASS_0.STOPPED\n",
"return CLASS_0.RUNNING\n",
"return CLASS_0.STOPPED\n"
] |
[
"def check_component(comp, session, logger):...\n",
"logger.debug('Running component check for %s' % comp['name'])\n",
"check_available = len(comp['cmd']) > 1 and 'check' in comp['cmd'][1]\n",
"window = find_window(session, comp['name'])\n",
"if window:\n",
"pid = get_window_pid(window)\n",
"logger.debug('%s window is not running. Running custom check' % comp['name'])\n",
"logger.debug('Found window pid: %s' % pid)\n",
"if check_available and run_component_check(comp):\n",
"procs = []\n",
"logger.debug('Component was not started by Hyperion, but the check succeeded')\n",
"logger.debug(\n 'Window not running and no check command is available or it failed: returning false'\n )\n",
"for entry in pid:\n",
"return CheckState.STARTED_BY_HAND\n",
"return CheckState.STOPPED\n",
"procs.extend(Process(entry).children(recursive=True))\n",
"pids = [p.pid for p in procs]\n",
"logger.debug('Window is running %s child processes' % len(pids))\n",
"if len(pids) < 3:\n",
"logger.debug(\n 'Main window process has finished. Running custom check if available')\n",
"if check_available and run_component_check(comp):\n",
"if check_available and run_component_check(comp):\n",
"logger.debug('Check succeeded')\n",
"if not check_available:\n",
"logger.debug('Process terminated but check was successful')\n",
"logger.debug('Check failed or no check available: returning false')\n",
"return CheckState.RUNNING\n",
"logger.debug(\n 'No custom check specified and got sufficient pid amount: returning true')\n",
"logger.debug('Check failed: returning false')\n",
"return CheckState.STOPPED_BUT_SUCCESSFUL\n",
"return CheckState.STOPPED\n",
"return CheckState.RUNNING\n",
"return CheckState.STOPPED\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
7,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Expr'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Expr'",
"Expr'",
"Condition",
"Assign'",
"Expr'",
"Expr'",
"For",
"Return'",
"Return'",
"Expr'",
"Assign'",
"Expr'",
"Condition",
"Expr'",
"Condition",
"Condition",
"Expr'",
"Condition",
"Expr'",
"Expr'",
"Return'",
"Expr'",
"Expr'",
"Return'",
"Return'",
"Return'",
"Return'"
] |
[
"__author__ = 'Johannes Köster'\n",
"__copyright__ = 'Copyright 2015, Johannes Köster'\n",
"__email__ = 'koester@jimmy.harvard.edu'\n",
"__license__ = 'MIT'\n",
"import os\n",
"import sys\n",
"import base64\n",
"import json\n",
"from collections import defaultdict\n",
"from itertools import chain\n",
"from functools import partial\n",
"from operator import attrgetter\n",
"from snakemake.io import IOFile, Wildcards, Resources, _IOFile\n",
"from snakemake.utils import format, listfiles\n",
"from snakemake.exceptions import RuleException, ProtectedOutputException\n",
"from snakemake.exceptions import UnexpectedOutputException\n",
"from snakemake.logging import logger\n",
"def FUNC_0(VAR_0, VAR_1):...\n",
"return chain(*map(attrgetter(VAR_1), VAR_0))\n"
] |
[
"__author__ = 'Johannes Köster'\n",
"__copyright__ = 'Copyright 2015, Johannes Köster'\n",
"__email__ = 'koester@jimmy.harvard.edu'\n",
"__license__ = 'MIT'\n",
"import os\n",
"import sys\n",
"import base64\n",
"import json\n",
"from collections import defaultdict\n",
"from itertools import chain\n",
"from functools import partial\n",
"from operator import attrgetter\n",
"from snakemake.io import IOFile, Wildcards, Resources, _IOFile\n",
"from snakemake.utils import format, listfiles\n",
"from snakemake.exceptions import RuleException, ProtectedOutputException\n",
"from snakemake.exceptions import UnexpectedOutputException\n",
"from snakemake.logging import logger\n",
"def jobfiles(jobs, type):...\n",
"return chain(*map(attrgetter(type), jobs))\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
7,
0,
0,
0,
0,
0,
0
] |
[
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Import'",
"Import'",
"Import'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"FunctionDef'",
"Return'"
] |
[
"def FUNC_7(self, VAR_3):...\n",
"if VAR_3.name in self.host_names:\n",
"self.hosts.remove(VAR_3)\n",
"self._hosts.remove(VAR_3.name)\n",
"VAR_3.remove_group(self)\n",
"self.clear_hosts_cache()\n"
] |
[
"def remove_host(self, host):...\n",
"if host.name in self.host_names:\n",
"self.hosts.remove(host)\n",
"self._hosts.remove(host.name)\n",
"host.remove_group(self)\n",
"self.clear_hosts_cache()\n"
] |
[
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Condition",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"@classmethod...\n",
"return '-Dfile.encoding=UTF-8', '-Dzinc.analysis.cache.limit=1000', '-Djava.awt.headless=true', '-Xmx2g'\n"
] |
[
"@classmethod...\n",
"return '-Dfile.encoding=UTF-8', '-Dzinc.analysis.cache.limit=1000', '-Djava.awt.headless=true', '-Xmx2g'\n"
] |
[
0,
0
] |
[
"Condition",
"Return'"
] |
[
"def FUNC_0(self):...\n",
"self.client = Client()\n",
"self.user = User(username='testUser')\n",
"self.user.set_password('testPassword')\n",
"self.user.save()\n",
"self.grader = User(username='grader', is_staff=True)\n",
"self.grader.set_password('graderPassword')\n",
"self.grader.save()\n",
"self.superuser = User(username='staff', is_staff=False, is_superuser=True)\n",
"self.superuser.set_password('staffPassword')\n",
"self.superuser.save()\n",
"self.course = Course.objects.create(name='test course', code='123456',\n VAR_5='Course-Url')\n",
"self.today = timezone.now()\n",
"self.tomorrow = self.today + timedelta(days=1)\n",
"self.two_days_from_now = self.tomorrow + timedelta(days=1)\n",
"self.yesterday = self.today - timedelta(days=1)\n",
"self.past_course_instance = CourseInstance.objects.create(instance_name=\n 'Fall 2011 day 0', starting_time=self.yesterday, ending_time=self.today,\n course=self.course, VAR_5='T-00.1000_d0')\n",
"self.current_course_instance = CourseInstance.objects.create(instance_name=\n 'Fall 2011 day 1', starting_time=self.today, ending_time=self.tomorrow,\n course=self.course, VAR_5='T-00.1000_d1')\n",
"self.future_course_instance = CourseInstance.objects.create(instance_name=\n 'Fall 2011 day 2', starting_time=self.tomorrow, ending_time=self.\n two_days_from_now, course=self.course, VAR_5='T-00.1000_d2')\n",
"self.hidden_course_instance = CourseInstance.objects.create(instance_name=\n 'Secret super course', starting_time=self.tomorrow, ending_time=self.\n two_days_from_now, course=self.course, VAR_5='T-00.1000_hidden',\n visible_to_students=False)\n",
"self.course_module = CourseModule.objects.create(name='test module', VAR_5=\n 'test-module', points_to_pass=10, course_instance=self.\n current_course_instance, opening_time=self.today, closing_time=self.\n tomorrow)\n",
"self.course_module_with_late_submissions_allowed = CourseModule.objects.create(\n name='test module', VAR_5='test-module-late', points_to_pass=50,\n course_instance=self.current_course_instance, opening_time=self.today,\n closing_time=self.tomorrow, late_submissions_allowed=True,\n late_submission_deadline=self.two_days_from_now,\n late_submission_penalty=0.2)\n",
"self.learning_object_category = LearningObjectCategory.objects.create(name=\n 'test category', course_instance=self.current_course_instance,\n points_to_pass=5)\n",
"self.learning_object = LearningObject.objects.create(name=\n 'test learning object', course_module=self.course_module, category=self\n .learning_object_category, VAR_5='l1')\n",
"self.broken_learning_object = LearningObject.objects.create(name=\n 'test learning object', course_module=self.\n course_module_with_late_submissions_allowed, category=self.\n learning_object_category, VAR_5='l2')\n",
"self.base_exercise = BaseExercise.objects.create(name='test exercise',\n course_module=self.course_module, category=self.\n learning_object_category, service_url='http://localhost/', VAR_5='b1')\n",
"self.submission = Submission.objects.create(exercise=self.base_exercise,\n grader=self.grader.userprofile)\n",
"self.submission.submitters.add(self.user.userprofile)\n",
"self.course_hook = CourseHook.objects.create(hook_url='test_hook_url',\n course_instance=self.current_course_instance)\n"
] |
[
"def setUp(self):...\n",
"self.client = Client()\n",
"self.user = User(username='testUser')\n",
"self.user.set_password('testPassword')\n",
"self.user.save()\n",
"self.grader = User(username='grader', is_staff=True)\n",
"self.grader.set_password('graderPassword')\n",
"self.grader.save()\n",
"self.superuser = User(username='staff', is_staff=False, is_superuser=True)\n",
"self.superuser.set_password('staffPassword')\n",
"self.superuser.save()\n",
"self.course = Course.objects.create(name='test course', code='123456', url=\n 'Course-Url')\n",
"self.today = timezone.now()\n",
"self.tomorrow = self.today + timedelta(days=1)\n",
"self.two_days_from_now = self.tomorrow + timedelta(days=1)\n",
"self.yesterday = self.today - timedelta(days=1)\n",
"self.past_course_instance = CourseInstance.objects.create(instance_name=\n 'Fall 2011 day 0', starting_time=self.yesterday, ending_time=self.today,\n course=self.course, url='T-00.1000_d0')\n",
"self.current_course_instance = CourseInstance.objects.create(instance_name=\n 'Fall 2011 day 1', starting_time=self.today, ending_time=self.tomorrow,\n course=self.course, url='T-00.1000_d1')\n",
"self.future_course_instance = CourseInstance.objects.create(instance_name=\n 'Fall 2011 day 2', starting_time=self.tomorrow, ending_time=self.\n two_days_from_now, course=self.course, url='T-00.1000_d2')\n",
"self.hidden_course_instance = CourseInstance.objects.create(instance_name=\n 'Secret super course', starting_time=self.tomorrow, ending_time=self.\n two_days_from_now, course=self.course, url='T-00.1000_hidden',\n visible_to_students=False)\n",
"self.course_module = CourseModule.objects.create(name='test module', url=\n 'test-module', points_to_pass=10, course_instance=self.\n current_course_instance, opening_time=self.today, closing_time=self.\n tomorrow)\n",
"self.course_module_with_late_submissions_allowed = CourseModule.objects.create(\n name='test module', url='test-module-late', points_to_pass=50,\n course_instance=self.current_course_instance, opening_time=self.today,\n closing_time=self.tomorrow, late_submissions_allowed=True,\n late_submission_deadline=self.two_days_from_now,\n late_submission_penalty=0.2)\n",
"self.learning_object_category = LearningObjectCategory.objects.create(name=\n 'test category', course_instance=self.current_course_instance,\n points_to_pass=5)\n",
"self.learning_object = LearningObject.objects.create(name=\n 'test learning object', course_module=self.course_module, category=self\n .learning_object_category, url='l1')\n",
"self.broken_learning_object = LearningObject.objects.create(name=\n 'test learning object', course_module=self.\n course_module_with_late_submissions_allowed, category=self.\n learning_object_category, url='l2')\n",
"self.base_exercise = BaseExercise.objects.create(name='test exercise',\n course_module=self.course_module, category=self.\n learning_object_category, service_url='http://localhost/', url='b1')\n",
"self.submission = Submission.objects.create(exercise=self.base_exercise,\n grader=self.grader.userprofile)\n",
"self.submission.submitters.add(self.user.userprofile)\n",
"self.course_hook = CourseHook.objects.create(hook_url='test_hook_url',\n course_instance=self.current_course_instance)\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'"
] |
[
"def FUNC_8(self, VAR_9, VAR_12, VAR_13, VAR_6, VAR_4):...\n",
"VAR_21 = generate_password_hash(VAR_4)\n",
"VAR_14 = (\n \"INSERT INTO `testdb`.`user` (`ID`, `FName`, `LName`, `Email`, `Pass`, `Role`) VALUES ('\"\n + VAR_9 + \"', '\" + VAR_12 + \"', '\" + VAR_13 + \"', '\" + VAR_6 + \"', '\" +\n VAR_21 + \"', '0')\")\n",
"self.cursor.execute(VAR_14)\n",
"self.connection.commit()\n"
] |
[
"def create_basic_user(self, userID, fName, lName, email, password):...\n",
"password2 = generate_password_hash(password)\n",
"query = (\n \"INSERT INTO `testdb`.`user` (`ID`, `FName`, `LName`, `Email`, `Pass`, `Role`) VALUES ('\"\n + userID + \"', '\" + fName + \"', '\" + lName + \"', '\" + email + \"', '\" +\n password2 + \"', '0')\")\n",
"self.cursor.execute(query)\n",
"self.connection.commit()\n"
] |
[
0,
0,
4,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Expr'"
] |
[
"def FUNC_28(VAR_13):...\n",
"self.assertEqual('/request', VAR_13.path)\n",
"VAR_12.append('applicable')\n",
"return VAR_5\n"
] |
[
"def applicable(request):...\n",
"self.assertEqual('/request', request.path)\n",
"calls.append('applicable')\n",
"return ident\n"
] |
[
0,
0,
0,
0
] |
[
"FunctionDef'",
"Expr'",
"Expr'",
"Return'"
] |
[
"def FUNC_3(self, VAR_1):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_17 = CLASS_3(VAR_1)\n",
"VAR_30 = CLASS_4(VAR_17)\n",
"VAR_31 = VAR_30.parse()\n",
"if VAR_31:\n",
"VAR_31.execute(self.builtins)\n",
"VAR_31.wait()\n"
] |
[
"def execute(self, raw):...\n",
"\"\"\"docstring\"\"\"\n",
"tokens = Tokenizer(raw)\n",
"parser = Parser(tokens)\n",
"root = parser.parse()\n",
"if root:\n",
"root.execute(self.builtins)\n",
"root.wait()\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Expr'",
"Expr'"
] |
[
"def FUNC_3(self):...\n",
"for VAR_5 in self.groups:\n",
"self.add_group(VAR_5)\n"
] |
[
"def populate_ancestors(self):...\n",
"for group in self.groups:\n",
"self.add_group(group)\n"
] |
[
0,
0,
0
] |
[
"FunctionDef'",
"For",
"Expr'"
] |
[
"def FUNC_9(self, VAR_9, VAR_11=None):...\n",
"VAR_1 = get_and_check_project(VAR_9, VAR_11, ('change_project',))\n",
"VAR_27 = [VAR_35 for filesList in map(lambda key: VAR_9.FILES.getlist(key),\n [keys for keys in VAR_9.FILES]) for VAR_35 in filesList]\n",
"VAR_28 = models.Task.create_from_images(VAR_27, VAR_1)\n",
"if VAR_28 is not None:\n",
"return Response({'id': VAR_28.id}, status=status.HTTP_201_CREATED)\n"
] |
[
"def create(self, request, project_pk=None):...\n",
"project = get_and_check_project(request, project_pk, ('change_project',))\n",
"files = [file for filesList in map(lambda key: request.FILES.getlist(key),\n [keys for keys in request.FILES]) for file in filesList]\n",
"task = models.Task.create_from_images(files, project)\n",
"if task is not None:\n",
"return Response({'id': task.id}, status=status.HTTP_201_CREATED)\n"
] |
[
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Return'"
] |
[
"def FUNC_28(VAR_24):...\n",
"VAR_33.append(VAR_24)\n",
"return 23\n"
] |
[
"def exec_python(args):...\n",
"calls.append(args)\n",
"return 23\n"
] |
[
0,
0,
0
] |
[
"FunctionDef'",
"Expr'",
"Return'"
] |
[
"@Throttle(VAR_4)...\n",
"\"\"\"docstring\"\"\"\n",
"import pyatmo\n",
"self.station_data = pyatmo.WeatherStationData(self.auth)\n",
"if self.station is not None:\n",
"self.data = self.station_data.lastData(VAR_19=self.station, exclude=3600)\n",
"self.data = self.station_data.lastData(exclude=3600)\n"
] |
[
"@Throttle(MIN_TIME_BETWEEN_UPDATES)...\n",
"\"\"\"docstring\"\"\"\n",
"import pyatmo\n",
"self.station_data = pyatmo.WeatherStationData(self.auth)\n",
"if self.station is not None:\n",
"self.data = self.station_data.lastData(station=self.station, exclude=3600)\n",
"self.data = self.station_data.lastData(exclude=3600)\n"
] |
[
0,
0,
0,
0,
0,
0,
0
] |
[
"Condition",
"Docstring",
"Import'",
"Assign'",
"Condition",
"Assign'",
"Assign'"
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.