lines listlengths 1 444 | raw_lines listlengths 1 444 | label listlengths 1 444 | type listlengths 1 444 |
|---|---|---|---|
[
"async def FUNC_1(self, VAR_11):...\n",
"return await self.set_props_by_key('content', VAR_11)\n"
] | [
"async def set_content(self, content):...\n",
"return await self.set_props_by_key('content', content)\n"
] | [
0,
0
] | [
"AsyncFunctionDef'",
"Return'"
] |
[
"def FUNC_9(self, *VAR_15, **VAR_75):...\n",
""
] | [
"def newfn(self, *a, **env):...\n",
""
] | [
0,
0
] | [
"FunctionDef'",
"Condition"
] |
[
"def FUNC_23(self):...\n",
"self.assertEqual(3, VAR_23(pyodbc.version.split('.')))\n"
] | [
"def test_version(self):...\n",
"self.assertEqual(3, len(pyodbc.version.split('.')))\n"
] | [
0,
0
] | [
"FunctionDef'",
"Expr'"
] |
[
"async def FUNC_6(VAR_6):...\n",
"import csv\n",
"VAR_10 = {'Authorization': f'Bearer {VAR_1}'}\n",
"VAR_11 = csv.reader(open('jogging_dataset.csv'), delimiter=',')\n",
"for row in VAR_11:\n",
"VAR_7 = {'date': row[0], 'location': row[1], 'distance': int(row[2]),\n 'time': int(row[3])}\n",
"VAR_8 = a... | [
"async def test_positive_load_dataset(test_cli):...\n",
"import csv\n",
"headers = {'Authorization': f'Bearer {access_token}'}\n",
"dsreader = csv.reader(open('jogging_dataset.csv'), delimiter=',')\n",
"for row in dsreader:\n",
"data = {'date': row[0], 'location': row[1], 'distance': int(row[2]), 'time':\... | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"AsyncFunctionDef'",
"Import'",
"Assign'",
"Assign'",
"For",
"Assign'",
"Assign'",
"Assert'"
] |
[
"def FUNC_2(VAR_2):...\n",
"VAR_2.call_to_terminate_proxy()\n"
] | [
"def after_all(context):...\n",
"context.call_to_terminate_proxy()\n"
] | [
0,
5
] | [
"FunctionDef'",
"Expr'"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self._console_outstream\n"
] | [
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self._console_outstream\n"
] | [
0,
0,
0
] | [
"Condition",
"Docstring",
"Return'"
] |
[
"@staticmethod...\n",
"if not VAR_1:\n",
"if not VAR_5:\n",
"VAR_5 = CLASS_0.get_report(VAR_1)['analysis']\n",
"VAR_9 = {'data': [], 'status': True}\n",
"for VAR_17 in VAR_5.get('behavior', {}).get('generic', []):\n",
"VAR_9['data'].append({'process_name': VAR_17['process_name'], 'pid': VAR_17\n ['pi... | [
"@staticmethod...\n",
"if not task_id:\n",
"if not report:\n",
"report = AnalysisController.get_report(task_id)['analysis']\n",
"data = {'data': [], 'status': True}\n",
"for process in report.get('behavior', {}).get('generic', []):\n",
"data['data'].append({'process_name': process['process_name'], 'pid'... | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Condition",
"Condition",
"Assign'",
"Assign'",
"For",
"Expr'",
"Assign'",
"Return'"
] |
[
"def __eq__(self, VAR_17):...\n",
"return self.name == VAR_17.name\n"
] | [
"def __eq__(self, other):...\n",
"return self.name == other.name\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_20(self, VAR_10=None, VAR_11=None, VAR_1=None):...\n",
"VAR_10 = self.add_rule(VAR_10, VAR_11, VAR_1)\n",
"VAR_13 = self.get_rule(VAR_10)\n",
"def FUNC_39(VAR_101):...\n",
"if VAR_101.input:\n",
"VAR_13.set_input(*VAR_101.input[0], **ruleinfo.input[1])\n",
"if VAR_101.output:\n",
"VAR_13.set... | [
"def rule(self, name=None, lineno=None, snakefile=None):...\n",
"name = self.add_rule(name, lineno, snakefile)\n",
"rule = self.get_rule(name)\n",
"def decorate(ruleinfo):...\n",
"if ruleinfo.input:\n",
"rule.set_input(*ruleinfo.input[0], **ruleinfo.input[1])\n",
"if ruleinfo.output:\n",
"rule.set_out... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"FunctionDef'",
"Condition",
"Expr'",
"Condition",
"Expr'",
"Condition",
"Expr'",
"Condition",
"Condition",
"Condition",
"Assign'",
"Assign'",
"Condition",
"Condition",
"Condition",
"Condition",
"Condition",
"Assign'",
"Assign'",
"Con... |
[
"def FUNC_0(VAR_2, VAR_3):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_5 = dict(external_id=resp.get('orcid'), external_method='orcid')\n",
"return VAR_5\n"
] | [
"def account_info(remote, resp):...\n",
"\"\"\"docstring\"\"\"\n",
"account_info = dict(external_id=resp.get('orcid'), external_method='orcid')\n",
"return account_info\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Return'"
] |
[
"def FUNC_3(VAR_2):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_4 = {'id': 0, 'name': 0, 'size': 0, 'age': 0, 'passes': 0, 'verified': 0}\n",
"VAR_1.execute('SELECT * FROM files WHERE name=\"' + VAR_2 + '\";')\n",
"VAR_5 = VAR_1.fetchall()\n",
"if VAR_1.rowcount >= 1:\n",
"for row in VAR_5:\n",
"return VAR... | [
"def db_get_file_details(path):...\n",
"\"\"\"docstring\"\"\"\n",
"status = {'id': 0, 'name': 0, 'size': 0, 'age': 0, 'passes': 0, 'verified': 0}\n",
"cur.execute('SELECT * FROM files WHERE name=\"' + path + '\";')\n",
"result = cur.fetchall()\n",
"if cur.rowcount >= 1:\n",
"for row in result:\n",
"re... | [
0,
0,
4,
4,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'",
"Assign'",
"Condition",
"For",
"Return'",
"Assign'"
] |
[
"def FUNC_3(self, VAR_10, VAR_11):...\n",
"VAR_19 = VAR_11.icinga_lines\n",
"VAR_24 = CLASS_2(self.output_path(VAR_10))\n",
"VAR_24.write_lines(VAR_19)\n"
] | [
"def write_output(self, file_name, yaml_icinga):...\n",
"lines = yaml_icinga.icinga_lines\n",
"output_writer = OutputWriter(self.output_path(file_name))\n",
"output_writer.write_lines(lines)\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_2(VAR_2):...\n",
"VAR_34 = re.compile('UNIQUE constraint failed: (.*)$')\n",
"VAR_35 = VAR_34.match(str(VAR_2))\n",
"VAR_36 = {'blog.local_path': 'string', 'blog.url':\n \"\"\"\nThe URL for this blog is the same as another blog in this system.\nURLs for blogs must be unique.\n\"\"\"\n }[VAR_35... | [
"def field_error(e):...\n",
"_ = re.compile('UNIQUE constraint failed: (.*)$')\n",
"m = _.match(str(e))\n",
"error = {'blog.local_path':\n \"\"\"\nThe file path for this blog is the same as another blog in this system.\nFile paths must be unique.\n\"\"\"\n , 'blog.url':\n \"\"\"\nThe URL for this blo... | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_15(self, VAR_10, VAR_8):...\n",
"VAR_29, VAR_28 = VAR_10.compile(self.lhs)\n",
"return 'CASE WHEN %(lhs)s IS NULL THEN NULL ELSE coalesce(array_length(%(lhs)s, 1), 0) END' % {\n 'lhs': VAR_29}, VAR_28\n"
] | [
"def as_sql(self, compiler, connection):...\n",
"lhs, params = compiler.compile(self.lhs)\n",
"return 'CASE WHEN %(lhs)s IS NULL THEN NULL ELSE coalesce(array_length(%(lhs)s, 1), 0) END' % {\n 'lhs': lhs}, params\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Return'"
] |
[
"def FUNC_1():...\n",
"VAR_1 = [{'fieldname': 'territory', 'fieldtype': 'Link', 'label': _(\n 'Territory'), 'options': 'Territory', 'width': 100}, {'fieldname':\n 'item_group', 'fieldtype': 'Link', 'label': _('Item Group'), 'options':\n 'Item Group', 'width': 150}, {'fieldname': 'item_name', 'fieldtype':... | [
"def get_columns():...\n",
"columns = [{'fieldname': 'territory', 'fieldtype': 'Link', 'label': _(\n 'Territory'), 'options': 'Territory', 'width': 100}, {'fieldname':\n 'item_group', 'fieldtype': 'Link', 'label': _('Item Group'), 'options':\n 'Item Group', 'width': 150}, {'fieldname': 'item_name', 'fiel... | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Return'"
] |
[
"@VAR_0.route('/login/<remote_app>/')...\n",
"\"\"\"docstring\"\"\"\n",
"if VAR_1 not in oauth.remote_apps:\n",
"return abort(404)\n",
"VAR_2 = url_for('.authorized', VAR_1=remote_app, next=request.args.get(\n 'next') or request.referrer or None, _external=True)\n",
"return oauth.remote_apps[VAR_1].aut... | [
"@blueprint.route('/login/<remote_app>/')...\n",
"\"\"\"docstring\"\"\"\n",
"if remote_app not in oauth.remote_apps:\n",
"return abort(404)\n",
"callback_url = url_for('.authorized', remote_app=remote_app, next=request.\n args.get('next') or request.referrer or None, _external=True)\n",
"return oauth.r... | [
0,
0,
0,
0,
5,
5
] | [
"Condition",
"Docstring",
"Condition",
"Return'",
"Assign'",
"Return'"
] |
[
"import time\n",
"import mox\n",
"import paramiko\n",
"from cinder import context\n",
"from cinder import exception\n",
"from cinder.openstack.common import log as logging\n",
"from cinder.openstack.common import processutils\n",
"from cinder import test\n",
"from cinder.volume import configuration ... | [
"import time\n",
"import mox\n",
"import paramiko\n",
"from cinder import context\n",
"from cinder import exception\n",
"from cinder.openstack.common import log as logging\n",
"from cinder.openstack.common import processutils\n",
"from cinder import test\n",
"from cinder.volume import configuration ... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Import'",
"Import'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Assign'",
"FunctionDef'",
"Expr'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"... |
[
"from __future__ import absolute_import\n",
"import time\n",
"from binascii import hexlify\n",
"from twisted.web import resource, server\n",
"import Tribler.Core.Utilities.json_util as json\n",
"from Tribler.Core.Modules.restapi.util import fix_unicode_dict\n",
"from Tribler.Core.simpledefs import NTFY_... | [
"from __future__ import absolute_import\n",
"import time\n",
"from binascii import hexlify\n",
"from twisted.web import resource, server\n",
"import Tribler.Core.Utilities.json_util as json\n",
"from Tribler.Core.Modules.restapi.util import fix_unicode_dict\n",
"from Tribler.Core.simpledefs import NTFY_... | [
0,
0,
0,
0,
0,
0,
4,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
"ImportFrom'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Expr'",
"FunctionDef'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"E... |
[
"def FUNC_8(self, VAR_13):...\n",
"\"\"\"docstring\"\"\"\n"
] | [
"def p_expr(self, p):...\n",
"\"\"\"docstring\"\"\"\n"
] | [
0,
0
] | [
"FunctionDef'",
"Docstring"
] |
[
"def FUNC_19(VAR_18, VAR_20, VAR_10, VAR_26, VAR_27, VAR_28):...\n",
"VAR_48 = VAR_20, VAR_4[VAR_20]['name']\n",
"VAR_34 = FUNC_1(VAR_10)\n",
"VAR_49 = SubmitRequest(VAR_34, VAR_48, base_url=CWS_BASE_URL, VAR_26=\n submission_format, VAR_27=filenames, VAR_28=language)\n",
"VAR_49.execute()\n",
"VAR_29 ... | [
"def cws_submit(contest_id, task_id, user_id, submission_format, filenames,...\n",
"task = task_id, created_tasks[task_id]['name']\n",
"browser = get_cws_browser(user_id)\n",
"sr = SubmitRequest(browser, task, base_url=CWS_BASE_URL, submission_format=\n submission_format, filenames=filenames, language=lang... | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Condition",
"Return'"
] |
[
"def FUNC_2(self):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_7 = logging.getLogger(__name__)\n",
"VAR_7.debug('Killing process monitoring thread')\n",
"self.end = True\n"
] | [
"def kill(self):...\n",
"\"\"\"docstring\"\"\"\n",
"logger = logging.getLogger(__name__)\n",
"logger.debug('Killing process monitoring thread')\n",
"self.end = True\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'",
"Assign'"
] |
[
"def FUNC_11(self):...\n",
"return self.__cache_map.keys()\n"
] | [
"def keys(self):...\n",
"return self.__cache_map.keys()\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_6(self, VAR_46, VAR_47):...\n",
"return True\n"
] | [
"def run(self, vh, thing_name):...\n",
"return True\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_6():...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_12 = 1\n",
"if os.path.isfile(VAR_0):\n",
"VAR_22 = yaml.load(todays_tasks_entry)\n",
"click.echo(chalk.red(\n 'There are no goals set. Set a new goal by entering \"yoda goals new\"'))\n",
"VAR_27 = 0\n",
"VAR_28 = True\n",
"for entry in VAR_2... | [
"def complete_goal():...\n",
"\"\"\"docstring\"\"\"\n",
"not_valid_goal_number = 1\n",
"if os.path.isfile(GOALS_CONFIG_FILE_PATH):\n",
"contents = yaml.load(todays_tasks_entry)\n",
"click.echo(chalk.red(\n 'There are no goals set. Set a new goal by entering \"yoda goals new\"'))\n",
"i = 0\n",
"no_... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"For",
"AugAssign'",
"Condition",
"Condition",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"For",
"AugAssign'",
"Condition",
"Assign'",
"Expr'",
"Assi... |
[
"def FUNC_7(self, VAR_6, VAR_7=None, VAR_8='icon', VAR_9=None):...\n",
"if VAR_7 is None:\n",
"return self.icon\n",
"return self.aq_parent.display_icon(VAR_6, VAR_7, VAR_8, VAR_9)\n"
] | [
"def display_icon(self, REQUEST, meta_type=None, key='icon', zpt=None):...\n",
"if meta_type is None:\n",
"return self.icon\n",
"return self.aq_parent.display_icon(REQUEST, meta_type, key, zpt)\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Return'",
"Return'"
] |
[
"@handled_slot(bool)...\n",
"print(f\"{'Locking' if VAR_4 else 'Unlocking'} piezo etalon.\")\n",
"self.lock_piezo_etalon_action.setChecked(not VAR_4)\n",
"self.matisse.set_piezo_etalon_lock(VAR_4)\n",
"self.lock_piezo_etalon_action.setChecked(VAR_4)\n"
] | [
"@handled_slot(bool)...\n",
"print(f\"{'Locking' if checked else 'Unlocking'} piezo etalon.\")\n",
"self.lock_piezo_etalon_action.setChecked(not checked)\n",
"self.matisse.set_piezo_etalon_lock(checked)\n",
"self.lock_piezo_etalon_action.setChecked(checked)\n"
] | [
0,
0,
0,
0,
0
] | [
"Condition",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"@classmethod...\n",
"if VAR_3 is None:\n",
"VAR_3 = VAR_2.HOST, VAR_2.PORT\n",
"VAR_12 = CLASS_0.create_server(VAR_3, **kwargs)\n",
"return VAR_2(VAR_12, VAR_15=True, **kwargs)\n"
] | [
"@classmethod...\n",
"if addr is None:\n",
"addr = cls.HOST, cls.PORT\n",
"conn = DebugSessionConnection.create_server(addr, **kwargs)\n",
"return cls(conn, owned=True, **kwargs)\n"
] | [
0,
0,
0,
0,
0
] | [
"Condition",
"Condition",
"Assign'",
"Assign'",
"Return'"
] |
[
"@VAR_0.route('/saml2-ls', methods=['POST'])...\n",
"\"\"\"docstring\"\"\"\n",
"current_app.logger.debug('Logout service started')\n",
"VAR_14 = StateCache(VAR_3)\n",
"VAR_15 = IdentityCache(VAR_3)\n",
"VAR_16 = Saml2Client(current_app.saml2_config, state_cache=state,\n identity_cache=identity)\n",
"... | [
"@authn_views.route('/saml2-ls', methods=['POST'])...\n",
"\"\"\"docstring\"\"\"\n",
"current_app.logger.debug('Logout service started')\n",
"state = StateCache(session)\n",
"identity = IdentityCache(session)\n",
"client = Saml2Client(current_app.saml2_config, state_cache=state,\n identity_cache=identi... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"For",
"Expr'",
"For",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Condition",
"Condition",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Return'"... |
[
"def FUNC_6(self):...\n",
"if len(self.targets) > 0:\n",
"self.schedule(self.comment_loop)\n",
"if len(self.forums) == 0:\n",
"return\n",
"while len(self.forums) == 0:\n",
"self.schedule(self.scan_targets_loop)\n",
"self.counter_tick()\n",
"self.w.sleep(1)\n"
] | [
"def wait_loop(self):...\n",
"if len(self.targets) > 0:\n",
"self.schedule(self.comment_loop)\n",
"if len(self.forums) == 0:\n",
"return\n",
"while len(self.forums) == 0:\n",
"self.schedule(self.scan_targets_loop)\n",
"self.counter_tick()\n",
"self.w.sleep(1)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Expr'",
"For",
"Return'",
"For",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_0(self):...\n",
"if database.users.get_current_user():\n",
"database.render_template(self, 'items/new_item.html', {})\n",
"self.redirect('/')\n"
] | [
"def get(self):...\n",
"if database.users.get_current_user():\n",
"database.render_template(self, 'items/new_item.html', {})\n",
"self.redirect('/')\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Expr'",
"Expr'"
] |
[
"def FUNC_32(self, VAR_32, VAR_29):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_24 = self.getfile(VAR_32)\n",
"if VAR_24 == False:\n",
"return\n",
"if VAR_24[VAR_2] != VAR_13:\n",
"return\n",
"VAR_24[VAR_5] = VAR_29\n"
] | [
"def update_size(self, filename, size):...\n",
"\"\"\"docstring\"\"\"\n",
"f = self.getfile(filename)\n",
"if f == False:\n",
"return\n",
"if f[A_TYPE] != T_FILE:\n",
"return\n",
"f[A_SIZE] = size\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Return'",
"Condition",
"Return'",
"Assign'"
] |
[
"def FUNC_8(self, VAR_8=False):...\n",
"if not self.NativeFiletypeCompletionAvailable(\n",
"self._latest_completion_request = OmniCompletionRequest(self._omnicomp)\n",
"VAR_5 = {}\n",
"return self._latest_completion_request\n",
"self._AddExtraConfDataIfNeeded(VAR_5)\n",
"if VAR_8:\n",
"VAR_5['force_se... | [
"def CreateCompletionRequest(self, force_semantic=False):...\n",
"if not self.NativeFiletypeCompletionAvailable(\n",
"self._latest_completion_request = OmniCompletionRequest(self._omnicomp)\n",
"extra_data = {}\n",
"return self._latest_completion_request\n",
"self._AddExtraConfDataIfNeeded(extra_data)\n",... | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Assign'",
"Assign'",
"Return'",
"Expr'",
"Condition",
"Assign'",
"Assign'"
] |
[
"def FUNC_4(self, VAR_10):...\n",
"self.events.append(VAR_10)\n"
] | [
"def add_event(self, event):...\n",
"self.events.append(event)\n"
] | [
0,
0
] | [
"FunctionDef'",
"Expr'"
] |
[
"def FUNC_4(self, VAR_4):...\n",
"if not VAR_4:\n",
"return None if self.null else b''\n",
"if isinstance(VAR_4, bytes):\n",
"return VAR_4\n",
"return pickle.dumps(VAR_4)\n"
] | [
"def get_prep_value(self, value):...\n",
"if not value:\n",
"return None if self.null else b''\n",
"if isinstance(value, bytes):\n",
"return value\n",
"return pickle.dumps(value)\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Return'",
"Condition",
"Return'",
"Return'"
] |
[
"from django.http import HttpResponse\n",
"from django.template.response import TemplateResponse\n",
"from django.shortcuts import render, redirect\n",
"from pymongo import MongoClient\n",
"from django.core.mail import send_mail\n",
"import datetime\n",
"import random\n",
"import string\n",
"import ... | [
"from django.http import HttpResponse\n",
"from django.template.response import TemplateResponse\n",
"from django.shortcuts import render, redirect\n",
"from pymongo import MongoClient\n",
"from django.core.mail import send_mail\n",
"import datetime\n",
"import random\n",
"import string\n",
"import ... | [
0,
0,
5,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
5
] | [
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Import'",
"Import'",
"Import'",
"Import'",
"Assign'",
"Assign'",
"Assign'",
"Import'",
"Import'",
"Assign'",
"Import'",
"FunctionDef'",
"Return'"
] |
[
"def FUNC_7():...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_10['env'] = self.env\n",
"VAR_10['config'] = self.env.config\n",
"VAR_10['params'] = self.params\n",
"return VAR_10\n"
] | [
"def get_vars():...\n",
"\"\"\"docstring\"\"\"\n",
"template_vars['env'] = self.env\n",
"template_vars['config'] = self.env.config\n",
"template_vars['params'] = self.params\n",
"return template_vars\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_9(self, VAR_10, VAR_6=[], VAR_11=None, VAR_12=True):...\n",
"return super().command(VAR_10, VAR_6, VAR_11, VAR_12, '\\n', True)\n"
] | [
"def command(self, command, expected_output=[], error_message=None,...\n",
"return super().command(command, expected_output, error_message, log_event,\n '\\n', True)\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_31(VAR_45, VAR_42):...\n",
"if VAR_45.attrib.get('widget', False) == 'float_time':\n",
"VAR_45.set('widget', 'float')\n",
"for child in VAR_45:\n",
"FUNC_31(child, VAR_42)\n",
"return True\n"
] | [
"def _check_rec(eview, tm):...\n",
"if eview.attrib.get('widget', False) == 'float_time':\n",
"eview.set('widget', 'float')\n",
"for child in eview:\n",
"_check_rec(child, tm)\n",
"return True\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Expr'",
"For",
"Expr'",
"Return'"
] |
[
"def FUNC_0(self):...\n",
"if self.redirect_type in ['prefix', 'page', 'exact']:\n",
"VAR_9 = self.from_url\n",
"return ''\n",
"VAR_10 = self.to_url\n",
"if self.redirect_type == 'prefix':\n",
"VAR_10 = '/{lang}/{version}/'.format(lang=self.project.language, version=\n self.project.default_version)\n... | [
"def get_from_to_url_display(self):...\n",
"if self.redirect_type in ['prefix', 'page', 'exact']:\n",
"from_url = self.from_url\n",
"return ''\n",
"to_url = self.to_url\n",
"if self.redirect_type == 'prefix':\n",
"to_url = '/{lang}/{version}/'.format(lang=self.project.language, version=\n self.projec... | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Assign'",
"Return'",
"Assign'",
"Condition",
"Assign'",
"Return'"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self.binary('java')\n"
] | [
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self.binary('java')\n"
] | [
0,
0,
0
] | [
"Condition",
"Docstring",
"Return'"
] |
[
"def FUNC_3(self):...\n",
"auth.login(self.request, self.user)\n",
"VAR_5 = import_from_settings('OIDC_RENEW_ID_TOKEN_EXPIRY_SECONDS', 60 * 15)\n",
"self.request.session['oidc_id_token_expiration'] = time.time() + VAR_5\n",
"return HttpResponseRedirect(self.success_url)\n"
] | [
"def login_success(self):...\n",
"auth.login(self.request, self.user)\n",
"expiration_interval = import_from_settings('OIDC_RENEW_ID_TOKEN_EXPIRY_SECONDS'\n , 60 * 15)\n",
"self.request.session['oidc_id_token_expiration'] = time.time(\n ) + expiration_interval\n",
"return HttpResponseRedirect(self.suc... | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Expr'",
"Assign'",
"Assign'",
"Return'"
] |
[
"from time import sleep\n",
"import operator\n",
"from bs4 import BeautifulSoup\n",
"from requests import get\n",
"import constants\n",
"import logger\n",
"import re\n",
"import os\n",
"import pickle\n",
"import pysmash\n",
"from get_results import get_coalesced_tag\n",
"import datetime\n",
... | [
"from time import sleep\n",
"import operator\n",
"from bs4 import BeautifulSoup\n",
"from requests import get\n",
"import constants\n",
"import logger\n",
"import re\n",
"import os\n",
"import pickle\n",
"import pysmash\n",
"from get_results import get_coalesced_tag\n",
"import datetime\n",
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
4,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"ImportFrom'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"Import'",
"Import'",
"Import'",
"Import'",
"Import'",
"Import'",
"ImportFrom'",
"Import'",
"Assign'",
"Assign'",
"Assign'",
"FunctionDef'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Return'",
"Assign'",
"Condit... |
[
"@VAR_5.setter...\n",
"self._benchmark = IOFile(VAR_5, VAR_16=self)\n"
] | [
"@benchmark.setter...\n",
"self._benchmark = IOFile(benchmark, rule=self)\n"
] | [
0,
0
] | [
"Condition",
"Assign'"
] |
[
"def FUNC_12(VAR_1):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_3 = VAR_2.lexmatch.groupdict()\n",
"VAR_1.value = beamr.interpreters.VerbatimEnv(VAR_3['VBTM_HEAD'].strip(),\n VAR_3['VBTM_BODY'])\n",
"return VAR_1\n"
] | [
"def t_VERBATIM(t):...\n",
"\"\"\"docstring\"\"\"\n",
"gd = lexer.lexmatch.groupdict()\n",
"t.value = beamr.interpreters.VerbatimEnv(gd['VBTM_HEAD'].strip(), gd[\n 'VBTM_BODY'])\n",
"return t\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Return'"
] |
[
"def __init__(self, VAR_1, VAR_2, VAR_3):...\n",
"self.project_id = VAR_1\n",
"self.user_id = VAR_2\n",
"self.money = VAR_3\n",
"self.time = datetime.datetime.now().isoformat(' ')\n"
] | [
"def __init__(self, project_id, user_id, money):...\n",
"self.project_id = project_id\n",
"self.user_id = user_id\n",
"self.money = money\n",
"self.time = datetime.datetime.now().isoformat(' ')\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Assign'"
] |
[
"def FUNC_6(VAR_3):...\n",
"return {'commit_position': VAR_3.commit_position, 'pass_rate': VAR_3.\n pass_rate, 'task_id': VAR_3.task_id, 'build_number': VAR_3.build_number,\n 'git_hash': VAR_3.git_hash, 'try_job_url': VAR_3.try_job_url}\n"
] | [
"def _GetBasicData(point):...\n",
"return {'commit_position': point.commit_position, 'pass_rate': point.\n pass_rate, 'task_id': point.task_id, 'build_number': point.build_number,\n 'git_hash': point.git_hash, 'try_job_url': point.try_job_url}\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_22(self, VAR_11):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_35 = self.common._cli_run('showvlun -a -host %s' % VAR_11, None)\n",
"if VAR_35:\n",
"VAR_35 = VAR_35[1:]\n",
"for line in VAR_35:\n",
"VAR_30 = line.split(',')\n",
"if VAR_30 and len(VAR_30) > 4:\n",
"return VAR_30[4]\n"
] | [
"def _get_active_nsp(self, hostname):...\n",
"\"\"\"docstring\"\"\"\n",
"result = self.common._cli_run('showvlun -a -host %s' % hostname, None)\n",
"if result:\n",
"result = result[1:]\n",
"for line in result:\n",
"info = line.split(',')\n",
"if info and len(info) > 4:\n",
"return info[4]\n"
] | [
0,
0,
2,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Assign'",
"For",
"Assign'",
"Condition",
"Return'"
] |
[
"@staticmethod...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_23 = 'string'.format(VAR_10)\n",
"VAR_24 = db.engine.execute(VAR_23)\n",
"return VAR_24\n"
] | [
"@staticmethod...\n",
"\"\"\"docstring\"\"\"\n",
"query = (\n \"\"\"SELECT mapped_by as contributors from tasks where project_id = {0} and mapped_by is not null\n UNION\n SELECT validated_by from tasks where tasks.project_id = {0} and validated_by is not null\"\"\"\n .... | [
0,
0,
4,
4,
0
] | [
"Condition",
"Docstring",
"Assign'",
"Assign'",
"Return'"
] |
[
"@VAR_0.route('/<page_name>/save', methods=['POST'])...\n",
"VAR_7 = request.form.get('content')\n",
"VAR_3 = VAR_1.query('string' % VAR_2)\n",
"VAR_8 = VAR_3.namedresult()\n",
"if len(VAR_8) < 1:\n",
"VAR_1.insert('page', {'page_name': VAR_2})\n",
"VAR_3 = VAR_1.query(\"select id from page where page_n... | [
"@app.route('/<page_name>/save', methods=['POST'])...\n",
"content = request.form.get('content')\n",
"query = db.query(\n \"select page_content.content, page.id as page_id, page_content.id as content_id from page, page_content where page.id = page_content.page_id and page.page_name = '%s' order by page_conte... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"Return'"
] |
[
"def FUNC_1(VAR_1):...\n",
"VAR_3 = [_('Employee') + ':Link/Employee:150', _('Employee Name') + '::200',\n _('Department') + '::150']\n",
"for leave_type in VAR_1:\n",
"VAR_3.append(_(leave_type) + ' ' + _('Opening') + ':Float:160')\n",
"return VAR_3\n",
"VAR_3.append(_(leave_type) + ' ' + _('Taken') +... | [
"def get_columns(leave_types):...\n",
"columns = [_('Employee') + ':Link/Employee:150', _('Employee Name') +\n '::200', _('Department') + '::150']\n",
"for leave_type in leave_types:\n",
"columns.append(_(leave_type) + ' ' + _('Opening') + ':Float:160')\n",
"return columns\n",
"columns.append(_(leave_t... | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"For",
"Expr'",
"Return'",
"Expr'",
"Expr'"
] |
[
"async def FUNC_1(self) ->None:...\n",
"self.render('about.html', pagetitle='about')\n"
] | [
"async def get(self) ->None:...\n",
"self.render('about.html', pagetitle='about')\n"
] | [
0,
0
] | [
"AsyncFunctionDef'",
"Expr'"
] |
[
"def FUNC_10(self, VAR_10):...\n",
"VAR_33 = self.builder.get_object('composite_unlock_entry')\n",
"VAR_26 = self.builder.get_object('composite_unlock_select_button')\n",
"if VAR_33.get_text() is not '':\n",
"VAR_33.get_style_context().add_class('error')\n",
"self.database_manager = DatabaseManager(self.d... | [
"def on_composite_unlock_button_clicked(self, widget):...\n",
"composite_unlock_entry = self.builder.get_object('composite_unlock_entry')\n",
"composite_unlock_select_button = self.builder.get_object(\n 'composite_unlock_select_button')\n",
"if composite_unlock_entry.get_text() is not '':\n",
"composite_... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Condition",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_18(self, VAR_18):...\n",
"\"\"\"docstring\"\"\"\n",
"if isinstance(VAR_18, types.ListType):\n",
"VAR_18 = escape.json_encode(VAR_18)\n",
"RequestHandler.write(self, VAR_18)\n",
"RequestHandler.write(self, VAR_18)\n",
"self.set_header('Content-Type', 'application/json')\n"
] | [
"def write(self, chunk):...\n",
"\"\"\"docstring\"\"\"\n",
"if isinstance(chunk, types.ListType):\n",
"chunk = escape.json_encode(chunk)\n",
"RequestHandler.write(self, chunk)\n",
"RequestHandler.write(self, chunk)\n",
"self.set_header('Content-Type', 'application/json')\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Condition",
"Assign'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_1(VAR_0, VAR_1):...\n",
"return os.utime(VAR_0, VAR_1, follow_symlinks=os.utime not in os.\n supports_follow_symlinks)\n"
] | [
"def lutime(f, times):...\n",
"return os.utime(f, times, follow_symlinks=os.utime not in os.\n supports_follow_symlinks)\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_6(self):...\n",
"return shellutil.run('/sbin/chkconfig --del waagent', VAR_8=False)\n"
] | [
"def unregister_agent_service(self):...\n",
"return shellutil.run('/sbin/chkconfig --del waagent', chk_err=False)\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"import psycopg2\n",
"VAR_0 = 'forum'\n",
"def FUNC_0():...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_2 = psycopg2.connect(database=DBNAME)\n",
"VAR_3 = VAR_2.cursor()\n",
"VAR_3.execute('select content,time from posts order by time desc')\n",
"return VAR_3.fetchall()\n"
] | [
"import psycopg2\n",
"DBNAME = 'forum'\n",
"def get_posts():...\n",
"\"\"\"docstring\"\"\"\n",
"db = psycopg2.connect(database=DBNAME)\n",
"c = db.cursor()\n",
"c.execute('select content,time from posts order by time desc')\n",
"return c.fetchall()\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Import'",
"Assign'",
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Expr'",
"Return'"
] |
[
"\"\"\"string\"\"\"\n",
"import logging\n",
"import re\n",
"import ratelimit.mixins\n",
"from nacl.exceptions import CryptoError\n",
"from django.conf import settings\n",
"from django.core.exceptions import PermissionDenied\n",
"from django.http import HttpResponse\n",
"from django.shortcuts import ... | [
"\"\"\"\n\nView partials provide all the callisto-core front-end functionality.\nSubclass these partials with your own views if you are implementing\ncallisto-core. Many of the view partials only provide a subset of the\nfunctionality required for a full HTML view.\n\ndocs / reference:\n - https://docs.djangopro... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Expr'",
"Import'",
"Import'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Assign'",
"Assign'",
"Condition",
"Return'"
] |
[
"def FUNC_3(self):...\n",
"self.assertFalse({'name': 'DocType'} in DatabaseQuery('DocType').execute(\n filters=[['DocType', 'name', 'like', 'J%']]))\n"
] | [
"def test_filters_1(self):...\n",
"self.assertFalse({'name': 'DocType'} in DatabaseQuery('DocType').execute(\n filters=[['DocType', 'name', 'like', 'J%']]))\n"
] | [
0,
0
] | [
"FunctionDef'",
"Expr'"
] |
[
"from threading import Thread, Lock\n",
"import logging\n",
"import sys\n",
"import time\n",
"import hyperion.lib.util.config as config\n",
"from os import system\n",
"from subprocess import call\n",
"from psutil import Process, NoSuchProcess\n",
"VAR_0 = sys.version[0] == '2'\n",
"if VAR_0:\n",
... | [
"from threading import Thread, Lock\n",
"import logging\n",
"import sys\n",
"import time\n",
"import hyperion.lib.util.config as config\n",
"from os import system\n",
"from subprocess import call\n",
"from psutil import Process, NoSuchProcess\n",
"is_py2 = sys.version[0] == '2'\n",
"if is_py2:\n",... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"ImportFrom'",
"Import'",
"Import'",
"Import'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Assign'",
"Condition",
"Import'",
"Import'",
"Expr'",
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"FunctionDef'",
"Docstring",
"Expr'",
"FunctionDef'",
"Docstring... |
[
"def __init__(self, VAR_18, *VAR_15, **VAR_16):...\n",
"self.item = VAR_18\n",
"CLASS_0.__init__(self, VAR_18, *VAR_15, **kw)\n"
] | [
"def __init__(self, item, *a, **kw):...\n",
"self.item = item\n",
"Validator.__init__(self, item, *a, **kw)\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'"
] |
[
"def FUNC_4(VAR_7, VAR_8=None):...\n",
"\"\"\"docstring\"\"\"\n",
"def FUNC_7(VAR_17, *VAR_18, **VAR_19):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_20 = VAR_17.POST\n",
"VAR_21 = VAR_20.get('new_status')\n",
"if not VAR_21:\n",
"if not VAR_8 or not callable(VAR_8):\n",
"if not 'fields' in VAR_20:\n",
... | [
"def manageModelStatus(entity_logic, status_retriever=None):...\n",
"\"\"\"docstring\"\"\"\n",
"def manageModelsStatus(request, *args, **kwargs):...\n",
"\"\"\"docstring\"\"\"\n",
"post_dict = request.POST\n",
"new_status = post_dict.get('new_status')\n",
"if not new_status:\n",
"if not status_retriev... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
5
] | [
"FunctionDef'",
"Docstring",
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Condition",
"Condition",
"Condition",
"Return'",
"Expr'",
"Assign'",
"Assign'",
"For",
"Condition",
"Expr'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Return'",
"Return'"
] |
[
"def FUNC_7(self):...\n",
"for VAR_30 in VAR_7(len(self._['tags'])):\n",
"self._['tags'][VAR_30] = self._['tags'][VAR_30][:40]\n",
"VAR_10 = AnnTag.get_ann_tags(self.ann_id, self.sql_session)\n",
"if not VAR_4.match(self._['tags'][VAR_30]):\n",
"VAR_11 = set(self._['tags'])\n",
"VAR_12 = set(VAR_10)\n",... | [
"def parse_tag(self):...\n",
"for i in xrange(len(self._['tags'])):\n",
"self._['tags'][i] = self._['tags'][i][:40]\n",
"old_tags = AnnTag.get_ann_tags(self.ann_id, self.sql_session)\n",
"if not tag_re.match(self._['tags'][i]):\n",
"new_tag_set = set(self._['tags'])\n",
"old_tag_set = set(old_tags)\n",
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"For",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"For",
"Expr'",
"For",
"Expr'"
] |
[
"def FUNC_12(self, VAR_4, VAR_5):...\n",
"VAR_14 = ['8.9.10.11', '9.10.11.12']\n",
"VAR_15 = self.new_mocked_cluster(VAR_14, VAR_4)\n",
"self.assertEqual(VAR_5, VAR_15.health)\n"
] | [
"def _test_health(self, validate_fn, expected_health):...\n",
"conf_managers = ['8.9.10.11', '9.10.11.12']\n",
"api = self.new_mocked_cluster(conf_managers, validate_fn)\n",
"self.assertEqual(expected_health, api.health)\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_31():...\n",
"VAR_24 = vim.eval('tagfiles()')\n",
"VAR_25 = VAR_0.getcwd()\n",
"return [VAR_0.path.join(VAR_25, x) for x in VAR_24]\n"
] | [
"def GetTagFiles():...\n",
"tag_files = vim.eval('tagfiles()')\n",
"current_working_directory = os.getcwd()\n",
"return [os.path.join(current_working_directory, x) for x in tag_files]\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_11(VAR_0):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_20 = kijiji_api.KijijiApi()\n",
"VAR_20.login(VAR_0.username, VAR_0.password)\n",
"VAR_23 = VAR_20.get_all_ads()\n",
"[VAR_20.delete_ad(adId) for adName, adId in VAR_23]\n"
] | [
"def nuke(args):...\n",
"\"\"\"docstring\"\"\"\n",
"api = kijiji_api.KijijiApi()\n",
"api.login(args.username, args.password)\n",
"allAds = api.get_all_ads()\n",
"[api.delete_ad(adId) for adName, adId in allAds]\n"
] | [
0,
0,
0,
0,
0,
5
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'",
"Assign'",
"Expr'"
] |
[
"def FUNC_12(self):...\n",
"return self.vars.copy()\n"
] | [
"def get_vars(self):...\n",
"return self.vars.copy()\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_0(self, VAR_0, VAR_1, VAR_2):...\n",
"if '=' in VAR_0:\n",
"VAR_4, VAR_10 = VAR_0.split('=')\n",
"return []\n",
"VAR_3 = self.find_action(VAR_4, VAR_1, VAR_2)\n",
"if VAR_3.choices:\n",
"return [(VAR_4 + '=' + VAR_9) for VAR_9 in VAR_3.choices if VAR_9.\n startswith(VAR_10)]\n"
] | [
"def complete(self, token, parsed, parser):...\n",
"if '=' in token:\n",
"keyword, value_prefix = token.split('=')\n",
"return []\n",
"action = self.find_action(keyword, parsed, parser)\n",
"if action.choices:\n",
"return [(keyword + '=' + value) for value in action.choices if value.\n startswith(val... | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Assign'",
"Return'",
"Assign'",
"Condition",
"Return'"
] |
[
"def FUNC_6(self, VAR_22):...\n",
"if VAR_22:\n",
"VAR_107 = int(VAR_22, 36)\n",
"if self.redirect:\n",
"return Meetup._byID(VAR_107, True)\n",
"abort(404, 'page not found')\n",
"return None\n"
] | [
"def run(self, meetup_id36):...\n",
"if meetup_id36:\n",
"meetup_id = int(meetup_id36, 36)\n",
"if self.redirect:\n",
"return Meetup._byID(meetup_id, True)\n",
"abort(404, 'page not found')\n",
"return None\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Assign'",
"Condition",
"Return'",
"Expr'",
"Return'"
] |
[
"def FUNC_0(self):...\n",
"return hpcommon.HP3PARCommon(self.configuration)\n"
] | [
"def _init_common(self):...\n",
"return hpcommon.HP3PARCommon(self.configuration)\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_2(self):...\n",
"return \"\"\"Task: %s (ID %d)\nFile: %s\n\"\"\" % (self.task[1], self.task[0], repr\n (self.filenames)) + GenericRequest.specific_info(self)\n"
] | [
"def specific_info(self):...\n",
"return \"\"\"Task: %s (ID %d)\nFile: %s\n\"\"\" % (self.task[1], self.task[0], repr\n (self.filenames)) + GenericRequest.specific_info(self)\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def __lt__(self, VAR_11):...\n",
"return self.rule.__lt__(VAR_11.rule)\n"
] | [
"def __lt__(self, other):...\n",
"return self.rule.__lt__(other.rule)\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"@VAR_2.route('/edit_comment/<pic_name>', methods=['POST', 'GET'])...\n",
"VAR_5 = request.form['edit_comment']\n",
"return render_template('edit_page.html', VAR_5=edit_value, VAR_3=pic_name)\n"
] | [
"@app.route('/edit_comment/<pic_name>', methods=['POST', 'GET'])...\n",
"edit_value = request.form['edit_comment']\n",
"return render_template('edit_page.html', edit_value=edit_value, pic_name=\n pic_name)\n"
] | [
4,
4,
4
] | [
"Condition",
"Assign'",
"Return'"
] |
[
"def FUNC_4(VAR_1, VAR_5, VAR_6='', VAR_3='', VAR_4=None):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_16[VAR_17(VAR_1.name)] = VAR_5, VAR_6\n",
"if current_user.is_authenticated():\n",
"VAR_23 = current_user.get_id()\n",
"return None\n",
"VAR_24 = VAR_1.consumer_key\n",
"VAR_25 = RemoteToken.get(VAR_23, V... | [
"def token_setter(remote, token, secret='', token_type='', extra_data=None):...\n",
"\"\"\"docstring\"\"\"\n",
"session[token_session_key(remote.name)] = token, secret\n",
"if current_user.is_authenticated():\n",
"uid = current_user.get_id()\n",
"return None\n",
"cid = remote.consumer_key\n",
"t = Rem... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Assign'",
"Return'",
"Assign'",
"Assign'",
"Condition",
"Expr'",
"Assign'",
"Return'"
] |
[
"def FUNC_3():...\n",
"CLASS_0().cmdloop()\n"
] | [
"def invoke_cli():...\n",
"Shrub().cmdloop()\n"
] | [
0,
0
] | [
"FunctionDef'",
"Expr'"
] |
[
"def FUNC_8(self):...\n",
"VAR_3 = DatabaseQuery('DocField').execute(filters={'parent': 'DocType'},\n fields=['fieldname', 'fieldtype'], or_filters=[{'fieldtype': 'Table'},\n {'fieldtype': 'Select'}])\n",
"self.assertTrue({'fieldtype': 'Table', 'fieldname': 'fields'} in VAR_3)\n",
"self.assertTrue({'fie... | [
"def test_or_filters(self):...\n",
"data = DatabaseQuery('DocField').execute(filters={'parent': 'DocType'},\n fields=['fieldname', 'fieldtype'], or_filters=[{'fieldtype': 'Table'},\n {'fieldtype': 'Select'}])\n",
"self.assertTrue({'fieldtype': 'Table', 'fieldname': 'fields'} in data)\n",
"self.assertTru... | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_3(VAR_0, VAR_2, VAR_4, VAR_5, VAR_3):...\n",
"\"\"\"docstring\"\"\"\n",
"return FUNC_2(VAR_0, VAR_0, VAR_2, VAR_4, VAR_5, VAR_3)\n"
] | [
"def log_unexpected_keys(expected_keys, actual_keys, request, source, name):...\n",
"\"\"\"docstring\"\"\"\n",
"return log_unexpected_subset_keys(expected_keys, expected_keys, actual_keys,\n request, source, name)\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Return'"
] |
[
"def FUNC_2(self):...\n",
"VAR_7 = Pa11yCrawler('')\n",
"self.assertEqual(VAR_7.cmd, self._expected_command(VAR_7.pa11y_report_dir,\n VAR_7.start_urls))\n"
] | [
"def test_default(self):...\n",
"suite = Pa11yCrawler('')\n",
"self.assertEqual(suite.cmd, self._expected_command(suite.pa11y_report_dir,\n suite.start_urls))\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'"
] |
[
"@web.authenticated...\n",
"self.get(VAR_1, VAR_2=False)\n"
] | [
"@web.authenticated...\n",
"self.get(path, include_body=False)\n"
] | [
0,
5
] | [
"Condition",
"Expr'"
] |
[
"import unittest\n",
"from .base import BaseTestCase\n",
"\"\"\" Test List answers api \"\"\"\n",
"def FUNC_0(self):...\n",
"VAR_0 = self.client.get('/api/v1/questions/answers', headers={\n 'Authorization': 'JWT ' + self.token})\n",
"self.assertEqual(VAR_0.status_code, 200)\n",
"self.assertEqual(VAR_... | [
"import unittest\n",
"from .base import BaseTestCase\n",
"\"\"\" Test List answers api \"\"\"\n",
"def test_list_answers(self):...\n",
"response = self.client.get('/api/v1/questions/answers', headers={\n 'Authorization': 'JWT ' + self.token})\n",
"self.assertEqual(response.status_code, 200)\n",
"self... | [
0,
4,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Import'",
"ImportFrom'",
"Expr'",
"FunctionDef'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Condition",
"Expr'"
] |
[
"def FUNC_3(self):...\n",
"VAR_15 = self._server_popen.poll()\n",
"return VAR_15 is None\n"
] | [
"def _IsServerAlive(self):...\n",
"returncode = self._server_popen.poll()\n",
"return returncode is None\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Return'"
] |
[
"from __future__ import unicode_literals\n",
"from six import iteritems, string_types\n",
"import datetime\n",
"import frappe, sys\n",
"from frappe import _\n",
"from frappe.utils import cint, flt, now, cstr, strip_html, sanitize_html, sanitize_email, cast_fieldtype\n",
"from frappe.model import default... | [
"from __future__ import unicode_literals\n",
"from six import iteritems, string_types\n",
"import datetime\n",
"import frappe, sys\n",
"from frappe import _\n",
"from frappe.utils import cint, flt, now, cstr, strip_html, sanitize_html, sanitize_email, cast_fieldtype\n",
"from frappe.model import default... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"ImportFrom'",
"ImportFrom'",
"Import'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Assign'",
"FunctionDef'",
"Docstring",
"ImportFrom'",
"Condition",
"Assign'",
"Return'",
... |
[
"def __init__(self, VAR_0, VAR_1):...\n",
"self._doc = VAR_0\n",
"self._root = VAR_1\n"
] | [
"def __init__(self, compileddoc, startingpath):...\n",
"self._doc = compileddoc\n",
"self._root = startingpath\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'"
] |
[
"def FUNC_10(self):...\n",
"VAR_6 = ('\"normal_string\"', '\"string with spaces\"',\n '\"string with quotes\\\\\"a\"', '\"string with s-quotes\\'b\"', '\"bsn \\n A\"',\n '\"unrecognized \\\\q escape\"')\n",
"for string, result in zip(self.test_strings, VAR_6):\n",
"self.assertEqual(prepare_string_argume... | [
"def test_prepare_string_argument_sh(self):...\n",
"expected_results = ('\"normal_string\"', '\"string with spaces\"',\n '\"string with quotes\\\\\"a\"', '\"string with s-quotes\\'b\"', '\"bsn \\n A\"',\n '\"unrecognized \\\\q escape\"')\n",
"for string, result in zip(self.test_strings, expected_results):... | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"For",
"Expr'"
] |
[
"def FUNC_8(self, VAR_6):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_11 = OrderedDict(((self.objects_key, VAR_6[1]), ('total_entries', self.\n session.query(func.count(self.resource_type.id)).scalar()), (\n 'filtered_entries', VAR_6[0])))\n",
"VAR_14 = self._add_meta_props(VAR_11)\n",
"return VAR_14\n"
] | [
"def wrap_list_response(self, data):...\n",
"\"\"\"docstring\"\"\"\n",
"response = OrderedDict(((self.objects_key, data[1]), ('total_entries', self\n .session.query(func.count(self.resource_type.id)).scalar()), (\n 'filtered_entries', data[0])))\n",
"full_response = self._add_meta_props(response)\n",
... | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_0(self, VAR_2, VAR_3, VAR_4, VAR_5, VAR_6, VAR_7):...\n",
"VAR_34 = {}\n",
"for VAR_48 in self.browse(VAR_2, VAR_3, VAR_4, VAR_7=context):\n",
"VAR_34[VAR_48.id] = (VAR_48.parent_id and VAR_48.parent_id.name + '/' or ''\n ) + VAR_48.name\n",
"return VAR_34\n"
] | [
"def _complete_name(self, cr, uid, ids, name, args, context):...\n",
"res = {}\n",
"for m in self.browse(cr, uid, ids, context=context):\n",
"res[m.id] = (m.parent_id and m.parent_id.name + '/' or '') + m.name\n",
"return res\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"For",
"Assign'",
"Return'"
] |
[
"def FUNC_1(self, VAR_6, VAR_7=True, VAR_8=1):...\n",
"if not self.sshpool:\n",
"VAR_13 = self.configuration.san_password\n",
"VAR_12 = None\n",
"VAR_14 = self.configuration.san_private_key\n",
"VAR_17 = VAR_8\n",
"VAR_0.error(_('Error running SSH command: %s') % VAR_6)\n",
"VAR_15 = self.configuratio... | [
"def _run_ssh(self, command, check_exit_code=True, attempts=1):...\n",
"if not self.sshpool:\n",
"password = self.configuration.san_password\n",
"last_exception = None\n",
"privatekey = self.configuration.san_private_key\n",
"total_attempts = attempts\n",
"LOG.error(_('Error running SSH command: %s') % ... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Condition",
"Assign'",
"AugAssign'",
"Assign'",
"Return'",
"Expr'",
"Assign'",
"Expr'"
] |
[
"def __init__(self, VAR_7, VAR_54=None, VAR_55=None, *VAR_15, **VAR_16):...\n",
"self.min = VAR_54\n",
"self.max = VAR_55\n",
"CLASS_0.__init__(self, VAR_7, *VAR_15, **kw)\n"
] | [
"def __init__(self, param, min=None, max=None, *a, **kw):...\n",
"self.min = min\n",
"self.max = max\n",
"Validator.__init__(self, param, *a, **kw)\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_13(VAR_35, VAR_27):...\n",
"\"\"\"docstring\"\"\"\n",
"return 'string'.format(FUNC_12(VAR_35).replace('\\n', '<br>')).replace('\\n', '')\n"
] | [
"def _output_html(complete_output, categories):...\n",
"\"\"\"docstring\"\"\"\n",
"return \"\"\"<html>\n <head>\n <title>Automatically generated keywords by bibclassify</title>\n </head>\n <body>\n {0}\n </body>\n </html>\"\"\".format(\n _output_text(complete_output).replace('\\n', '<b... | [
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Return'"
] |
[
"def FUNC_6(self):...\n",
"if not (VAR_101.user_is_loggedin and VAR_101.site.is_moderator(VAR_101.user\n",
"abort(403, 'forbidden')\n"
] | [
"def run(self):...\n",
"if not (c.user_is_loggedin and c.site.is_moderator(c.user) or c.user_is_admin):\n",
"abort(403, 'forbidden')\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Expr'"
] |
[
"def FUNC_7(VAR_4, VAR_6):...\n",
"VAR_11 = f\"\"\"\n SELECT COUNT(p.id) AS total\n FROM {VAR_0} AS p INNER JOIN {VAR_1} AS pt ON\n p.product_type_id=pt.id\n WHERE p.rating >= {VAR_4['rating']}\n AND p.name LIKE '%{VAR_6}%'\n \"\"\"\n",
"VAR_12 =... | [
"def get_total_searched_products(filters, search):...\n",
"sql_query = f\"\"\"\n SELECT COUNT(p.id) AS total\n FROM {PRODUCTS_TABLE} AS p INNER JOIN {PRODUCTS_TYPES_TABLE} AS pt ON\n p.product_type_id=pt.id\n WHERE p.rating >= {filters['rating']}\n AND p.name... | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Return'"
] |
[
"def FUNC_42(self):...\n",
"self.compilation_ko(\"\"\"\nelsif true {\n\n}\n\"\"\")\n"
] | [
"def test_misplaced_elsif(self):...\n",
"self.compilation_ko(\"\"\"\nelsif true {\n\n}\n\"\"\")\n"
] | [
0,
0
] | [
"FunctionDef'",
"Expr'"
] |
[
"def FUNC_7(self, VAR_5, VAR_6, VAR_20, VAR_13=None, VAR_21=None):...\n",
"\"\"\"docstring\"\"\"\n",
"return super(CLASS_0, self)._check_recursion(VAR_5, VAR_6, VAR_20, VAR_21=\n parent or 'parent_id', VAR_13=context)\n"
] | [
"def _check_recursion(self, cr, uid, ids, context=None, parent=None):...\n",
"\"\"\"docstring\"\"\"\n",
"return super(budget_item, self)._check_recursion(cr, uid, ids, parent=\n parent or 'parent_id', context=context)\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Return'"
] |
[
"def FUNC_4(self, VAR_8):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_7 = \"SELECT ID FROM {0} WHERE path='{1}';\".format(VAR_2, VAR_8)\n",
"VAR_4 = self._run_command(VAR_7)\n",
"VAR_16 = VAR_4[0][0]\n",
"VAR_16 = None\n",
"return VAR_16\n"
] | [
"def getIDbyPath(self, pth):...\n",
"\"\"\"docstring\"\"\"\n",
"command = \"SELECT ID FROM {0} WHERE path='{1}';\".format(TABLE_NAME, pth)\n",
"data = self._run_command(command)\n",
"result = data[0][0]\n",
"result = None\n",
"return result\n"
] | [
0,
0,
4,
4,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_32(VAR_16, VAR_28):...\n",
"VAR_53 = \"SELECT DISTINCT scene FROM ranks WHERE player='{}'\".format(VAR_28)\n",
"VAR_79 = VAR_16.exec(VAR_53)\n",
"VAR_79 = [VAR_93[0] for VAR_93 in VAR_79]\n",
"VAR_88 = {VAR_93: FUNC_31(VAR_16, VAR_93, VAR_28) for VAR_93 in VAR_79}\n",
"return VAR_88\n"
] | [
"def get_bracket_graph_data(db, tag):...\n",
"sql = \"SELECT DISTINCT scene FROM ranks WHERE player='{}'\".format(tag)\n",
"scenes = db.exec(sql)\n",
"scenes = [s[0] for s in scenes]\n",
"bracket_placings_by_scene = {s: get_bracket_placings_in_scene(db, s, tag) for\n s in scenes}\n",
"return bracket_pl... | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_2():...\n",
"return inject_init.rules() + python_test_runner.rules()\n"
] | [
"def rules():...\n",
"return inject_init.rules() + python_test_runner.rules()\n"
] | [
0,
7
] | [
"FunctionDef'",
"Return'"
] |
[
"def __init__(self, VAR_12, VAR_13):...\n",
"self.icinga_lines = []\n",
"self.indent = CONFIG['INDENT']\n",
"self.icinga_lines.extend(VAR_13.serialize())\n",
"self.write_section('host', VAR_12.host)\n",
"for service in VAR_12.services:\n",
"self.write_section('service', service)\n"
] | [
"def __init__(self, yaml_config, header):...\n",
"self.icinga_lines = []\n",
"self.indent = CONFIG['INDENT']\n",
"self.icinga_lines.extend(header.serialize())\n",
"self.write_section('host', yaml_config.host)\n",
"for service in yaml_config.services:\n",
"self.write_section('service', service)\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"For",
"Expr'"
] |
[
"def __init__(self, VAR_5):...\n",
"self.user = VAR_5\n",
"super().__init__(VAR_5)\n"
] | [
"def __init__(self, user):...\n",
"self.user = user\n",
"super().__init__(user)\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'"
] |
[
"def FUNC_21(self, VAR_38):...\n",
"\"\"\"docstring\"\"\"\n",
"return self._parse_conditions(self.NOT_CONDITION, VAR_38)\n"
] | [
"def _parse_not(self, data):...\n",
"\"\"\"docstring\"\"\"\n",
"return self._parse_conditions(self.NOT_CONDITION, data)\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Return'"
] |
[
"def FUNC_15(self):...\n",
"VAR_1 = self.client.get(self.get_info_url, {'info_type': 'env_properties',\n 'env_group_id': self.group_new.pk})\n",
"VAR_8 = EnvGroup.objects.get(pk=self.group_new.pk)\n",
"VAR_7 = json.loads(serializers.serialize('json', VAR_8.property.all(),\n fields=('name', 'value')))\n"... | [
"def test_get_env_properties_by_group(self):...\n",
"response = self.client.get(self.get_info_url, {'info_type':\n 'env_properties', 'env_group_id': self.group_new.pk})\n",
"group = EnvGroup.objects.get(pk=self.group_new.pk)\n",
"expected_json = json.loads(serializers.serialize('json', group.property.all\n... | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_10(self, VAR_2=None, VAR_3=None, VAR_4=None):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_12 = []\n",
"def FUNC_31(VAR_1, VAR_16):...\n",
"VAR_30 = VAR_1.xsrf_token_data == {'some': 'data'}\n",
"VAR_12.append((VAR_16, VAR_30))\n",
"@api.public...\n",
"self.response.write(self.generate_xsrf_token(... | [
"def make_xsrf_handling_app(self, xsrf_token_enforce_on=None,...\n",
"\"\"\"docstring\"\"\"\n",
"calls = []\n",
"def record(request_handler, method):...\n",
"is_valid = request_handler.xsrf_token_data == {'some': 'data'}\n",
"calls.append((method, is_valid))\n",
"@api.public...\n",
"self.response.writ... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"FunctionDef'",
"Assign'",
"Expr'",
"Condition",
"Expr'",
"Condition",
"Expr'",
"Condition",
"Expr'",
"Condition",
"Expr'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Return'"
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.