lines
listlengths
1
444
raw_lines
listlengths
1
444
label
listlengths
1
444
type
listlengths
1
444
[ "def FUNC_28(self, VAR_23):...\n", "return self.mtime > VAR_23\n" ]
[ "def is_newer(self, time):...\n", "return self.mtime > time\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_36(self):...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = {'SEVERITY': {'MEDIUM': 12}, 'CONFIDENCE': {'LOW': 7, 'MEDIUM': 5}}\n", "self.check_example('sql_statements.py', VAR_2)\n" ]
[ "def test_sql_statements(self):...\n", "\"\"\"docstring\"\"\"\n", "expect = {'SEVERITY': {'MEDIUM': 12}, 'CONFIDENCE': {'LOW': 7, 'MEDIUM': 5}}\n", "self.check_example('sql_statements.py', expect)\n" ]
[ 0, 0, 4, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'" ]
[ "def FUNC_2(VAR_9):...\n", "VAR_34 = VAR_9.cmd('list-panes', '-F #{pane_pid}')\n", "return [int(p) for p in VAR_34.stdout]\n" ]
[ "def get_window_pid(window):...\n", "r = window.cmd('list-panes', '-F #{pane_pid}')\n", "return [int(p) for p in r.stdout]\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Return'" ]
[ "def FUNC_1(self):...\n", "return 'submit user test %s for task %s (ID %d) %s' % (repr(self.filenames),\n self.task[1], self.task[0], self.url)\n" ]
[ "def describe(self):...\n", "return 'submit user test %s for task %s (ID %d) %s' % (repr(self.filenames),\n self.task[1], self.task[0], self.url)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_12(self):...\n", "VAR_26 = os.path.join(self.path, 'manifest.json')\n", "if not os.path.exists(VAR_26):\n", "return\n", "FUNC_1('MANIFEST')\n", "\"\"\"\n Check if there is no comma syntax issue\n \"\"\"\n", "VAR_26 = json.loads(data_file.read())\n", "FUNC_4(\n \"[YEP-2.1] Sy...
[ "def check_manifest(self):...\n", "manifest = os.path.join(self.path, 'manifest.json')\n", "if not os.path.exists(manifest):\n", "return\n", "print_header('MANIFEST')\n", "\"\"\"\n Check if there is no comma syntax issue\n \"\"\"\n", "manifest = json.loads(data_file.read())\n", "print_er...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Return'", "Expr'", "Expr'", "Assign'", "Expr'", "Assign'", "For", "Condition", "Expr'", "Expr'", "For", "Expr'", "For", "Condition", "Expr'", "For", "Condition", "Condition", "Expr'", "Expr'", "Condition", "Condition", "Expr'...
[ "def FUNC_11(self, VAR_19):...\n", "\"\"\"docstring\"\"\"\n", "if self.test_name is None:\n", "self.initFile('line')\n", "if not self.reached_corner[0]:\n", "self.reached_corner[0] = self.navloc.goToPosition(0, 0)\n", "if self.navloc.goToPosition(VAR_19, 0):\n", "if self.reached_corner[0]:\n", "self...
[ "def testLine(self, length):...\n", "\"\"\"docstring\"\"\"\n", "if self.test_name is None:\n", "self.initFile('line')\n", "if not self.reached_corner[0]:\n", "self.reached_corner[0] = self.navloc.goToPosition(0, 0)\n", "if self.navloc.goToPosition(length, 0):\n", "if self.reached_corner[0]:\n", "sel...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Expr'", "Condition", "Assign'", "Condition", "Condition", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_6(self, VAR_64):...\n", "VAR_90 = VAR_64.split(',')\n", "return Thing._by_fullname(VAR_90, data=True, return_dict=False)\n" ]
[ "def run(self, id_str):...\n", "tids = id_str.split(',')\n", "return Thing._by_fullname(tids, data=True, return_dict=False)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Return'" ]
[ "def FUNC_0(VAR_0):...\n", "if VAR_0.method == 'POST':\n", "VAR_2 = VAR_0.POST.get('answer', '').lower()\n", "return render(VAR_0, 'exercises/first_exercise.html')\n", "VAR_3 = 0\n", "if 'hello' in VAR_2:\n", "VAR_3 += 1\n", "if 'a+' in VAR_2:\n", "VAR_3 += 1\n", "return render(VAR_0, 'exercises/f...
[ "def first(request):...\n", "if request.method == 'POST':\n", "submission = request.POST.get('answer', '').lower()\n", "return render(request, 'exercises/first_exercise.html')\n", "points = 0\n", "if 'hello' in submission:\n", "points += 1\n", "if 'a+' in submission:\n", "points += 1\n", "return r...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Return'", "Assign'", "Condition", "AugAssign'", "Condition", "AugAssign'", "Return'" ]
[ "import logging\n", "import os\n", "from xml.sax import SAXParseException\n", "from twisted.cred import credentials\n", "from twisted.internet import defer\n", "from twisted.web import util\n", "from twisted.web.error import FlattenerError\n", "from twisted.web.http import UNAUTHORIZED, OK\n", "from...
[ "import logging\n", "import os\n", "from xml.sax import SAXParseException\n", "from twisted.cred import credentials\n", "from twisted.internet import defer\n", "from twisted.web import util\n", "from twisted.web.error import FlattenerError\n", "from twisted.web.http import UNAUTHORIZED, OK\n", "from...
[ 0, 0, 0, 0, 0, 0, 5, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "FunctionDef'", "Assign'", "Return'" ]
[ "def FUNC_6(self, VAR_38):...\n", "if not VAR_38:\n", "if self.emp_error is not None:\n", "if len(VAR_38) > self.length:\n", "VAR_101.errors.add(self.emp_error)\n", "VAR_101.errors.add(self.len_error)\n", "return VAR_38\n" ]
[ "def run(self, title):...\n", "if not title:\n", "if self.emp_error is not None:\n", "if len(title) > self.length:\n", "c.errors.add(self.emp_error)\n", "c.errors.add(self.len_error)\n", "return title\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Condition", "Condition", "Expr'", "Expr'", "Return'" ]
[ "def FUNC_0(self):...\n", "VAR_8 = lambda x: dict(request_token_params={'scope': ''}, base_url=\n 'https://foo.bar/', request_token_url=None, access_token_url=\n 'https://foo.bar/oauth/access_token', authorize_url=\n 'https://foo.bar/oauth/authorize', consumer_key=x, consumer_secret=\n 'testsecret')\n...
[ "def setUp(self):...\n", "params = lambda x: dict(request_token_params={'scope': ''}, base_url=\n 'https://foo.bar/', request_token_url=None, access_token_url=\n 'https://foo.bar/oauth/access_token', authorize_url=\n 'https://foo.bar/oauth/authorize', consumer_key=x, consumer_secret=\n 'testsecret')\n...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "ImportFrom'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_0(self, VAR_11):...\n", "" ]
[ "def add_tag_by_name(self, tagname):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "def FUNC_4(VAR_30):...\n", "VAR_39 = 'meal_planner.db'\n", "VAR_50 = 'ingredients_' + str(VAR_13)\n", "conn.execute('CREATE TABLE IF NOT EXISTS ' + VAR_50 + ' (ingredients text)')\n", "conn.execute('INSERT INTO ' + VAR_50 + ' VALUES (?);', (VAR_30,))\n" ]
[ "def save_ingredients(ingredients):...\n", "database_file = 'meal_planner.db'\n", "tableName = 'ingredients_' + str(weekNumber)\n", "conn.execute('CREATE TABLE IF NOT EXISTS ' + tableName + ' (ingredients text)')\n", "conn.execute('INSERT INTO ' + tableName + ' VALUES (?);', (ingredients,))\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_11(self):...\n", "self.client.login(username=self.tester.username, password='password')\n", "VAR_1 = self.client.post(self.update_url, {'content_type':\n 'testruns.testcaserun', 'object_pk': self.case_run_1.pk, 'field':\n 'case_run_status', 'value': str(TestCaseRunStatus.objects.get(name=\n '...
[ "def test_change_case_run_status(self):...\n", "self.client.login(username=self.tester.username, password='password')\n", "response = self.client.post(self.update_url, {'content_type':\n 'testruns.testcaserun', 'object_pk': self.case_run_1.pk, 'field':\n 'case_run_status', 'value': str(TestCaseRunStatus.o...
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_6(self):...\n", "\"\"\"docstring\"\"\"\n", "if not VAR_3:\n", "VAR_14 = utils.FolderSearch.getFilepathsInclSubfolders(PIC_FOLDER,\n allowed_extensions=['txt', 'png', 'jpg', 'jpeg'])\n", "VAR_35 = utils.DropboxFolderSearch.getFilepathsInclSubfoldersDropboxPublic(\n DROPBOX_FOLDER_LINK, DROPBO...
[ "def updateFileList(self):...\n", "\"\"\"docstring\"\"\"\n", "if not FROM_DROPBOX:\n", "files = utils.FolderSearch.getFilepathsInclSubfolders(PIC_FOLDER,\n allowed_extensions=['txt', 'png', 'jpg', 'jpeg'])\n", "files_and_mods = (utils.DropboxFolderSearch.\n getFilepathsInclSubfoldersDropboxPublic(DROP...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Expr'", "Expr'", "Assign'", "Expr'" ]
[ "def FUNC_1(VAR_4, VAR_5):...\n", "VAR_10 = (VAR_4['page'] - 1) * VAR_4['perPage']\n", "VAR_11 = f\"\"\"string{VAR_0} AS p INNER JOIN {VAR_1} AS pt ON\n p.product_type_id=pt.id\n WHERE p.rating >= {VAR_4['rating']}\n AND pt.name = '{VAR_5}'\n ORDER BY p.id LIMIT {VAR_...
[ "def get_department_products(filters, department):...\n", "offset = (filters['page'] - 1) * filters['perPage']\n", "sql_query = f\"\"\"\n SELECT p.id, p.ean, p.name, p.description, pt.name AS type, p.company, p.price, p.rating, p.weight, p.quantity, p.image_url\n FROM {PRODUCTS_TABLE} AS p...
[ 0, 0, 4, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Return'" ]
[ "import logging\n", "import pandas as pd\n", "from webapp import db, app\n", "from webapp.utils import generate_matched_table_name, table_exists\n", "from collections import OrderedDict\n", "from webapp.logger import logger\n", "import numpy as np\n", "def FUNC_0(VAR_0, VAR_1, VAR_2, VAR_3):...\n", ...
[ "import logging\n", "import pandas as pd\n", "from webapp import db, app\n", "from webapp.utils import generate_matched_table_name, table_exists\n", "from collections import OrderedDict\n", "from webapp.logger import logger\n", "import numpy as np\n", "def get_histogram_bar_chart_data(data, distributi...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Import'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "For", "Return'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_5(self, VAR_1, VAR_2, VAR_8=None, VAR_9='form', VAR_4=None, VAR_14...\n", "VAR_29 = super(CLASS_0, self).fields_view_get(VAR_1, VAR_2, VAR_8, VAR_9,\n VAR_4, VAR_14=toolbar, VAR_15=submenu)\n", "if VAR_4 is None:\n", "VAR_4 = {}\n", "if 'location' in VAR_4 and VAR_4['location']:\n", "VAR_51 =...
[ "def fields_view_get(self, cr, uid, view_id=None, view_type='form', context=...\n", "res = super(product_product, self).fields_view_get(cr, uid, view_id,\n view_type, context, toolbar=toolbar, submenu=submenu)\n", "if context is None:\n", "context = {}\n", "if 'location' in context and context['location'...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "For", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Return'", "Assign'", "Condition", "Condition", "Condition", "Condition", "Assign'", "Condition", "Condition", "Condition", "Assign'", "Assign'", "Condition", "Condition", "Assign'", "Condition", "Condition...
[ "from __future__ import print_function\n", "import os\n", "import sys\n", "import etcd\n", "from custodia.store.interface import CSStore, CSStoreError, CSStoreExists\n", "def FUNC_0(VAR_0):...\n", "print(VAR_0, file=sys.stderr)\n", "def __init__(self, VAR_1):...\n", "self.server = VAR_1.get('etcd_se...
[ "from __future__ import print_function\n", "import os\n", "import sys\n", "import etcd\n", "from custodia.store.interface import CSStore, CSStoreError, CSStoreExists\n", "def log_error(error):...\n", "print(error, file=sys.stderr)\n", "def __init__(self, config):...\n", "self.server = config.get('et...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0 ]
[ "ImportFrom'", "Import'", "Import'", "Import'", "ImportFrom'", "FunctionDef'", "Expr'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "FunctionDef'", "Expr'", "Assign'", "Expr'", "Return'" ]
[ "def FUNC_7(self):...\n", "VAR_24 = time()\n", "for i in range(self.options.iterations):\n", "if self.db.campaign['aux']:\n", "VAR_25 = time()\n", "VAR_41 = Thread(VAR_18=self.aux.command, args=('./' + self.db.campaign[\n 'aux_command'],))\n", "VAR_34 = Thread(VAR_18=self.dut.command, args=('./' + se...
[ "def time_application(self):...\n", "start = time()\n", "for i in range(self.options.iterations):\n", "if self.db.campaign['aux']:\n", "end = time()\n", "aux_process = Thread(target=self.aux.command, args=('./' + self.db.campaign\n ['aux_command'],))\n", "dut_process = Thread(target=self.dut.command,...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "For", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Condition", "Expr'", "Condition", "Expr'", "Expr'" ]
[ "def FUNC_16(**VAR_19):...\n", "VAR_35 = FUNC_8('users/add', VAR_16=kwargs)\n", "VAR_46 = re.search('/user/([0-9]+)$', VAR_35.url)\n", "if VAR_46:\n", "VAR_10 = int(VAR_46.group(1))\n", "VAR_19['user_id'] = VAR_10\n", "VAR_3[VAR_10] = VAR_19\n", "VAR_35 = FUNC_8('contest/' + VAR_19['contest_id'] + '/u...
[ "def add_user(**kwargs):...\n", "r = admin_req('users/add', args=kwargs)\n", "g = re.search('/user/([0-9]+)$', r.url)\n", "if g:\n", "user_id = int(g.group(1))\n", "kwargs['user_id'] = user_id\n", "created_users[user_id] = kwargs\n", "r = admin_req('contest/' + kwargs['contest_id'] + '/users/add', arg...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Return'" ]
[ "def FUNC_0(self):...\n", "" ]
[ "def run_check(self):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "def FUNC_5(self, VAR_9):...\n", "self.virtual_fields.append(VAR_9)\n", "cached_property.clear_caches(self)\n" ]
[ "def add_virtual_field(self, field):...\n", "self.virtual_fields.append(field)\n", "cached_property.clear_caches(self)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Expr'" ]
[ "@staticmethod...\n", "VAR_17 = ['sup', 'i', 'span']\n", "return getattr(VAR_5, 'name', None) == 'a' and getattr(VAR_5.parent, 'name',\n None) not in VAR_17 and not VAR_5.has_attr('style')\n" ]
[ "@staticmethod...\n", "tags = ['sup', 'i', 'span']\n", "return getattr(element, 'name', None) == 'a' and getattr(element.parent,\n 'name', None) not in tags and not element.has_attr('style')\n" ]
[ 0, 0, 0 ]
[ "Condition", "Assign'", "Return'" ]
[ "@rest_utils.ajax()...\n", "\"\"\"docstring\"\"\"\n", "VAR_18 = VAR_1.DATA.get('length', 100)\n", "VAR_19 = api.nova.server_console_output(VAR_1, VAR_3, tail_length=log_length)\n", "return {'lines': [x for x in VAR_19.split('\\n')]}\n" ]
[ "@rest_utils.ajax()...\n", "\"\"\"docstring\"\"\"\n", "log_length = request.DATA.get('length', 100)\n", "console_lines = api.nova.server_console_output(request, server_id,\n tail_length=log_length)\n", "return {'lines': [x for x in console_lines.split('\\n')]}\n" ]
[ 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Return'" ]
[ "def __getstate__(self):...\n", "return self.serialize()\n" ]
[ "def __getstate__(self):...\n", "return self.serialize()\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_0(self):...\n", "for VAR_7 in self.G.edges():\n", "self.calcLKVM(VAR_7)\n" ]
[ "def calcAllLKVM(self):...\n", "for edge in self.G.edges():\n", "self.calcLKVM(edge)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "For", "Expr'" ]
[ "def FUNC_1(self, VAR_0):...\n", "self.__init__(VAR_4=False)\n", "self.name = VAR_0.get('name')\n", "self.vars = VAR_0.get('vars', dict())\n", "self.address = VAR_0.get('address', '')\n", "self._uuid = VAR_0.get('uuid', None)\n", "self.implicit = VAR_0.get('implicit', False)\n", "VAR_8 = VAR_0.get('gr...
[ "def deserialize(self, data):...\n", "self.__init__(gen_uuid=False)\n", "self.name = data.get('name')\n", "self.vars = data.get('vars', dict())\n", "self.address = data.get('address', '')\n", "self._uuid = data.get('uuid', None)\n", "self.implicit = data.get('implicit', False)\n", "groups = data.get('...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Expr'", "Expr'" ]
[ "@commands.command()...\n", "\"\"\"docstring\"\"\"\n", "await self.simple_embed('string')\n" ]
[ "@commands.command()...\n", "\"\"\"docstring\"\"\"\n", "await self.simple_embed(\n \"\"\"If you have boot9strap and Luma3DS installed after following Plailect's guide, run Luma Updater to make sure it is on the latest Luma3DS normal version and then you can proceed to update your 3DS through system settings....
[ 0, 0, 0 ]
[ "Condition", "Docstring", "Expr'" ]
[ "def FUNC_4(VAR_7, VAR_8, VAR_9, VAR_10):...\n", "if VAR_7.does_intersect_rule(VAR_4(VAR_8['bookmarks'], VAR_8['attributes'],\n", "VAR_25 = [VAR_9]\n", "return\n", "for VAR_24, VAR_11 in VAR_8['levels']:\n", "VAR_27 = VAR_11['bookmarks'] if 'bookmarks' in VAR_11 else []\n", "VAR_28 = VAR_11['treeattribu...
[ "def _traverse(searcher, rule, ctx, client):...\n", "if searcher.does_intersect_rule(RuleTraversalContext(rule['bookmarks'],\n", "pathlist = [ctx]\n", "return\n", "for leveltype, levelfields in rule['levels']:\n", "levelbookmarks = levelfields['bookmarks'] if 'bookmarks' in levelfields else []\n", "leve...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 1, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Return'", "For", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "For", "Assign'", "Assign'", "Condition", "Expr'", "Assign'", "Condition", "...
[ "def __init__(self):...\n", "self.updated_input = set()\n", "self.updated_input_run = set()\n", "self.missing_output = set()\n", "self.incomplete_output = set()\n", "self.forced = False\n", "self.noio = False\n", "self.nooutput = False\n", "self.derived = True\n" ]
[ "def __init__(self):...\n", "self.updated_input = set()\n", "self.updated_input_run = set()\n", "self.missing_output = set()\n", "self.incomplete_output = set()\n", "self.forced = False\n", "self.noio = False\n", "self.nooutput = False\n", "self.derived = True\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'" ]
[ "def FUNC_3():...\n", "VAR_9 = FUNC_0()\n", "VAR_10 = VAR_9.cursor()\n", "VAR_10.execute(VAR_8)\n", "VAR_9.commit()\n", "VAR_2.info('successfully created also table for the first time')\n" ]
[ "def create_also_table():...\n", "db = db_connect()\n", "cursor = db.cursor()\n", "cursor.execute(ALSO_TABLE)\n", "db.commit()\n", "logger.info('successfully created also table for the first time')\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'" ]
[ "@property...\n", "\"\"\"docstring\"\"\"\n", "return self.config.get(self.section, 'userdb_mongo_database')\n" ]
[ "@property...\n", "\"\"\"docstring\"\"\"\n", "return self.config.get(self.section, 'userdb_mongo_database')\n" ]
[ 0, 0, 0 ]
[ "Condition", "Docstring", "Return'" ]
[ "def FUNC_0(self):...\n", "\"\"\"docstring\"\"\"\n", "\"\"\"string\"\"\"\n", "VAR_15 = int(request.args.get('page'))\n", "VAR_16 = db_interac.get_messages(VAR_15)\n", "return VAR_16, 200\n" ]
[ "def get(self):...\n", "\"\"\"docstring\"\"\"\n", "\"\"\"\n # Check the request comes from appropriate location.\n if not utils.validate_ip(request.remote_addr)\n return {}, 403\n \"\"\"\n", "page = int(request.args.get('page'))\n", "response = db_interac.get_messages(page)\n...
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'", "Assign'", "Assign'", "Return'" ]
[ "\"\"\"string\"\"\"\n", "from urllib import quote, unquote\n", "from urllib.parse import quote, unquote\n", "from circuits.six import text_type\n", "from circuits import handler, BaseComponent, Event\n", "from circuits.web.utils import parse_qs\n", "from circuits.web.events import response\n", "from c...
[ "\"\"\"Dispatcher\n\nThis module implements a basic URL to Channel dispatcher.\nThis is the default dispatcher used by circuits.web\n\"\"\"\n", "from urllib import quote, unquote\n", "from urllib.parse import quote, unquote\n", "from circuits.six import text_type\n", "from circuits import handler, BaseCompo...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Expr'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "FunctionDef'", "Expr'", "Assign'", "FunctionDef'", "FunctionDef'", "Return'" ]
[ "from django.db import models\n", "from .common_info import CommonInfo\n", "from django.core.exceptions import ValidationError\n", "from .weight_fraction_type import WeightFractionType\n", "from .extracted_chemical import ExtractedChemical\n", "from .script import Script\n", "def FUNC_0(VAR_0):...\n", ...
[ "from django.db import models\n", "from .common_info import CommonInfo\n", "from django.core.exceptions import ValidationError\n", "from .weight_fraction_type import WeightFractionType\n", "from .extracted_chemical import ExtractedChemical\n", "from .script import Script\n", "def validate_wf_analysis(va...
[ 0, 0, 0, 6, 6, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "FunctionDef'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Return'" ]
[ "def __init__(self, *VAR_2, **VAR_3):...\n", "super(CLASS_1, self).__init__(*VAR_2, **kwargs)\n", "self.OIDC_OP_AUTH_ENDPOINT = import_from_settings(\n 'OIDC_OP_AUTHORIZATION_ENDPOINT')\n", "self.OIDC_RP_CLIENT_ID = import_from_settings('OIDC_RP_CLIENT_ID')\n" ]
[ "def __init__(self, *args, **kwargs):...\n", "super(OIDCAuthenticationRequestView, self).__init__(*args, **kwargs)\n", "self.OIDC_OP_AUTH_ENDPOINT = import_from_settings(\n 'OIDC_OP_AUTHORIZATION_ENDPOINT')\n", "self.OIDC_RP_CLIENT_ID = import_from_settings('OIDC_RP_CLIENT_ID')\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Assign'" ]
[ "def FUNC_0(self):...\n", "self.db_conn.commit()\n" ]
[ "def commit(self):...\n", "self.db_conn.commit()\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_6(self, VAR_44=None):...\n", "if not VAR_101.user_is_loggedin:\n", "if VAR_44 is not None and not valid_password(VAR_101.user, VAR_44):\n", "VAR_101.errors.add(errors.WRONG_PASSWORD)\n" ]
[ "def run(self, password=None):...\n", "if not c.user_is_loggedin:\n", "if password is not None and not valid_password(c.user, password):\n", "c.errors.add(errors.WRONG_PASSWORD)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Condition", "Expr'" ]
[ "@functools.wraps(VAR_0)...\n", "\"\"\"docstring\"\"\"\n", "if hasattr(VAR_9, 'interaction'):\n", "VAR_28 = map(lambda pp: pp.principal.id, VAR_9.interaction.participations)\n", "VAR_28 = []\n", "if hasattr(self, 'logFile'):\n", "VAR_29 = '%s %s - %s \"%s\" %d %s \"%s\" \"%s\"\\n' % (VAR_9.getClientIP()...
[ "@functools.wraps(f)...\n", "\"\"\"docstring\"\"\"\n", "if hasattr(request, 'interaction'):\n", "principals = map(lambda pp: pp.principal.id, request.interaction.participations\n )\n", "principals = []\n", "if hasattr(self, 'logFile'):\n", "line = '%s %s - %s \"%s\" %d %s \"%s\" \"%s\"\\n' % (request...
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Expr'" ]
[ "@property...\n", "\"\"\"docstring\"\"\"\n", "VAR_17 = [VAR_28.mtime for VAR_28 in self.expanded_output if VAR_28.exists]\n", "if self.benchmark and self.benchmark.exists:\n", "VAR_17.append(self.benchmark.mtime)\n", "if VAR_17:\n", "return min(VAR_17)\n", "return None\n" ]
[ "@property...\n", "\"\"\"docstring\"\"\"\n", "existing = [f.mtime for f in self.expanded_output if f.exists]\n", "if self.benchmark and self.benchmark.exists:\n", "existing.append(self.benchmark.mtime)\n", "if existing:\n", "return min(existing)\n", "return None\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Condition", "Expr'", "Condition", "Return'", "Return'" ]
[ "from django.db import models\n", "from dashboard.models import CommonInfo\n", "from .raw_chem import RawChem\n", "VAR_0 = models.CharField('Raw CAS', max_length=100, null=True, blank=True)\n", "VAR_1 = models.CharField('Raw chemical name', max_length=500, null=True,\n blank=True)\n", "VAR_2 = models.B...
[ "from django.db import models\n", "from dashboard.models import CommonInfo\n", "from .raw_chem import RawChem\n", "raw_cas_old = models.CharField('Raw CAS', max_length=100, null=True, blank=True\n )\n", "raw_chem_name_old = models.CharField('Raw chemical name', max_length=500,\n null=True, blank=True)...
[ 0, 6, 0, 0, 0, 0, 0, 0 ]
[ "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Condition", "Return'" ]
[ "from libtmux import Server\n", "from yaml import load, dump\n", "from setupParser import Loader\n", "from DepTree import Node, dep_resolve, CircularReferenceException\n", "import logging\n", "import os\n", "import socket\n", "import argparse\n", "from psutil import Process\n", "from subprocess im...
[ "from libtmux import Server\n", "from yaml import load, dump\n", "from setupParser import Loader\n", "from DepTree import Node, dep_resolve, CircularReferenceException\n", "import logging\n", "import os\n", "import socket\n", "import argparse\n", "from psutil import Process\n", "from subprocess im...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Import'", "Import'", "Import'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Import'", "ImportFrom'", "Import'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Assi...
[ "def __init__(self, VAR_25=None, VAR_40=None, VAR_41=None, VAR_42=None,...\n", "VAR_54 = config.NsxLibConfig(VAR_40=username or NSX_USER, VAR_41=password or\n NSX_PASSWORD, VAR_42=retries or NSX_HTTP_RETRIES, VAR_43=insecure if \n insecure is not None else NSX_INSECURE, VAR_44=ca_file or NSX_CERT,\n VAR_...
[ "def __init__(self, session_response=None, username=None, password=None,...\n", "nsxlib_config = config.NsxLibConfig(username=username or NSX_USER, password\n =password or NSX_PASSWORD, retries=retries or NSX_HTTP_RETRIES,\n insecure=insecure if insecure is not None else NSX_INSECURE, ca_file=\n ca_file ...
[ 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Expr'", "Assign'" ]
[ "@staticmethod...\n", "return CLASS_0._TalkToHandlerAsync(VAR_9, VAR_7, 'POST', VAR_10)\n" ]
[ "@staticmethod...\n", "return BaseRequest._TalkToHandlerAsync(data, handler, 'POST', timeout)\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_1(self):...\n", "\"\"\"docstring\"\"\"\n", "while True:\n", "VAR_48 = self.readline()\n", "sys.exit(0)\n", "self.execute(VAR_48)\n" ]
[ "def run(self):...\n", "\"\"\"docstring\"\"\"\n", "while True:\n", "line = self.readline()\n", "sys.exit(0)\n", "self.execute(line)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_1(self):...\n", "\"\"\"docstring\"\"\"\n", "VAR_15 = ['hp3par_api_url', 'hp3par_username', 'hp3par_password', 'san_ip',\n 'san_login', 'san_password']\n", "self.common.check_flags(self.configuration, VAR_15)\n" ]
[ "def _check_flags(self):...\n", "\"\"\"docstring\"\"\"\n", "required_flags = ['hp3par_api_url', 'hp3par_username', 'hp3par_password',\n 'san_ip', 'san_login', 'san_password']\n", "self.common.check_flags(self.configuration, required_flags)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'" ]
[ "def FUNC_0(VAR_0, VAR_1, VAR_2):...\n", "VAR_1['AWS_ACCESS_KEY_ID'] = VAR_0.get_input('username', default='')\n", "VAR_1['AWS_SECRET_ACCESS_KEY'] = VAR_0.get_input('password', default='')\n", "if VAR_0.has_input('security_token'):\n", "VAR_1['AWS_SECURITY_TOKEN'] = VAR_0.get_input('security_token', default...
[ "def aws(cred, env, private_data_dir):...\n", "env['AWS_ACCESS_KEY_ID'] = cred.get_input('username', default='')\n", "env['AWS_SECRET_ACCESS_KEY'] = cred.get_input('password', default='')\n", "if cred.has_input('security_token'):\n", "env['AWS_SECURITY_TOKEN'] = cred.get_input('security_token', default='')\...
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Condition", "Assign'" ]
[ "def FUNC_1(self):...\n", "VAR_4 = self.get_option('cluster') or '%'\n", "VAR_5 = self.get_option('datacenter') or '%'\n", "self.dbcmd = '/usr/share/ovirt-engine/dbscripts/engine-psql.sh -c \"'\n", "self.dbcmd += 'string' % (VAR_4, VAR_5)\n", "self.log_debug('Query command for ovirt DB set to: %s' % self....
[ "def format_db_cmd(self):...\n", "cluster = self.get_option('cluster') or '%'\n", "datacenter = self.get_option('datacenter') or '%'\n", "self.dbcmd = '/usr/share/ovirt-engine/dbscripts/engine-psql.sh -c \"'\n", "self.dbcmd += (\n 'select host_name from vds_static where cluster_id in (select cluster_id f...
[ 0, 2, 2, 2, 2, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "AugAssign'", "Expr'" ]
[ "def FUNC_1():...\n", "\"\"\"docstring\"\"\"\n", "VAR_5 = {}\n", "for arg in request.args:\n", "VAR_8 = re.findall('page_(.*)', arg)\n", "return VAR_5\n", "if VAR_8:\n", "VAR_5[VAR_8[0]] = int(request.args.get(arg))\n" ]
[ "def get_page_args():...\n", "\"\"\"docstring\"\"\"\n", "pages = {}\n", "for arg in request.args:\n", "re_match = re.findall('page_(.*)', arg)\n", "return pages\n", "if re_match:\n", "pages[re_match[0]] = int(request.args.get(arg))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "For", "Assign'", "Return'", "Condition", "Assign'" ]
[ "def FUNC_7(self, VAR_13, VAR_11, VAR_7, VAR_14, VAR_10):...\n", "VAR_18 = VAR_11['rules']\n", "for VAR_37, VAR_9 in itertools.product(VAR_18, VAR_14):\n", "VAR_8 = VAR_10.get_rule(VAR_37)\n", "return None\n", "FUNC_4(VAR_7, VAR_8, VAR_9, VAR_10)\n" ]
[ "def get_directories(self, levelctx, levelfields, searcher, ctxlist, client):...\n", "rulenames = levelfields['rules']\n", "for rulename, ctx in itertools.product(rulenames, ctxlist):\n", "rule = client.get_rule(rulename)\n", "return None\n", "_traverse(searcher, rule, ctx, client)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "For", "Assign'", "Return'", "Expr'" ]
[ "\"\"\"string\"\"\"\n", "from Tkinter import *\n", "from SeleniumDriver import *\n", "import pyperclip\n", "VAR_0 = None\n", "VAR_1 = None\n", "VAR_2 = None\n", "VAR_3 = None\n", "VAR_4 = None\n", "def __init__(self, VAR_5, VAR_6):...\n", "self.master = VAR_5\n", "VAR_5.title('Jira Automation'...
[ "\"\"\"\nAuthor; Garrett Breeden\nVersion: 1.0\n 1. Integrated SeleniumDriver into GUI \n - Data should now pass from GUI -> Selenium\n 2. Updated input field classes\n 3. Small GUI Tweaks\n\nVersion: 1.2\n 1. Corrected issue with data being passed to Selenium from GUI\n\n\nTO ONLY BE USED BY L2 ...
[ 0, 5, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ "Expr'", "ImportFrom'", "ImportFrom'", "Import'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Expr'", "Expr'", "Assi...
[ "def FUNC_1(self, VAR_10, VAR_11, VAR_12):...\n", "fd.write(VAR_12)\n", "fd.flush()\n", "fd.seek(0)\n", "VAR_35 = os.path.basename(fd.name)\n", "if self.shouldLog:\n", "self.logger.debug('Sending file %s with mime type : %s', VAR_35, VAR_11)\n", "VAR_30 = self.session.post(self.uploadUrl, files={self....
[ "def uploadFile(self, suffix, mime, payload):...\n", "fd.write(payload)\n", "fd.flush()\n", "fd.seek(0)\n", "filename = os.path.basename(fd.name)\n", "if self.shouldLog:\n", "self.logger.debug('Sending file %s with mime type : %s', filename, mime)\n", "fu = self.session.post(self.uploadUrl, files={sel...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 7, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Expr'", "Expr'", "Assign'", "Condition", "Expr'", "Assign'", "AugAssign'", "Condition", "Condition", "Return'", "Expr'", "Condition", "Expr'" ]
[ "def FUNC_20(self, VAR_19, VAR_20):...\n", "\"\"\"docstring\"\"\"\n", "for clause in reversed(self.order):\n", "VAR_40 = VAR_20.has_wildcards() - VAR_19.has_wildcards()\n", "VAR_51 = clause.index(VAR_19.name)\n", "if VAR_40 != 0:\n", "VAR_52 = clause.index(VAR_20.name)\n", "return VAR_40\n", "return...
[ "def compare(self, rule1, rule2):...\n", "\"\"\"docstring\"\"\"\n", "for clause in reversed(self.order):\n", "wildcard_cmp = rule2.has_wildcards() - rule1.has_wildcards()\n", "i = clause.index(rule1.name)\n", "if wildcard_cmp != 0:\n", "j = clause.index(rule2.name)\n", "return wildcard_cmp\n", "retu...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "For", "Assign'", "Assign'", "Condition", "Assign'", "Return'", "Return'", "Assign'", "Condition", "Assign'", "Condition", "Return'", "Assign'" ]
[ "def FUNC_19(self, VAR_6, VAR_14):...\n", "self.common.extend_volume(VAR_6, VAR_14)\n" ]
[ "def extend_volume(self, volume, new_size):...\n", "self.common.extend_volume(volume, new_size)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_3(VAR_21, VAR_22, VAR_23=BaseInlineFormSet, VAR_24=forms.ModelForm,...\n", "VAR_32 = VAR_22.detail_fields()\n", "if VAR_4:\n", "VAR_32 = [in_field for in_field in VAR_32 if not in_field in VAR_4]\n", "return forms.inlineformset_factory(VAR_21=parent_model, VAR_22=model,\n VAR_27=formset_fields,...
[ "def make_formset(parent_model, model, formset=BaseInlineFormSet, form=forms...\n", "formset_fields = model.detail_fields()\n", "if exclude:\n", "formset_fields = [in_field for in_field in formset_fields if not in_field in\n exclude]\n", "return forms.inlineformset_factory(parent_model=parent_model, mode...
[ 0, 0, 0, 0, 0 ]
[ "For", "Assign'", "Condition", "Assign'", "Return'" ]
[ "def __call__(self, VAR_20):...\n", "return set(filter(lambda VAR_53: str(VAR_53.value).startswith(str(self.\n prefix)), VAR_20))\n" ]
[ "def __call__(self, urls):...\n", "return set(filter(lambda u: str(u.value).startswith(str(self.prefix)), urls))\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_2(self, VAR_1, VAR_7, VAR_8, VAR_9, VAR_4=None):...\n", "if VAR_4 is None:\n", "VAR_4 = {}\n", "VAR_29 = super(CLASS_0, self).view_header_get(VAR_1, VAR_7, VAR_8, VAR_9, VAR_4\n )\n", "if VAR_29:\n", "return VAR_29\n", "if VAR_4.get('active_id', False) and VAR_4.get('active_model'\n", "re...
[ "def view_header_get(self, cr, user, view_id, view_type, context=None):...\n", "if context is None:\n", "context = {}\n", "res = super(product_product, self).view_header_get(cr, user, view_id,\n view_type, context)\n", "if res:\n", "return res\n", "if context.get('active_id', False) and context.get('...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Assign'", "Condition", "Return'", "Condition", "Return'", "Return'" ]
[ "def FUNC_2(self):...\n", "self.render('report.html')\n" ]
[ "def get(self):...\n", "self.render('report.html')\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_58():...\n", "VAR_36.execute('rollback')\n" ]
[ "def test():...\n", "cnxn.execute('rollback')\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "@integration_synonym_api...\n", "FUNC_2(VAR_2)\n", "FUNC_3(VAR_2, 'AILEEN ENTERPRISES', VAR_7='1')\n", "FUNC_5(VAR_4, VAR_5, VAR_11='ALAN HARGREAVES CORPORATION', VAR_10=[{'name':\n '----ALAN HARGREAVES'}, {'name': '----ALAN'}])\n" ]
[ "@integration_synonym_api...\n", "clean_database(solr)\n", "seed_database_with(solr, 'AILEEN ENTERPRISES', id='1')\n", "verify_results(client, jwt, query='ALAN HARGREAVES CORPORATION', expected=[\n {'name': '----ALAN HARGREAVES'}, {'name': '----ALAN'}])\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_0(VAR_6):...\n", "@wraps(VAR_6)...\n", "if 'logged_in' in VAR_10:\n", "return VAR_6(*VAR_8, **kwargs)\n", "flash('Unauthorized, Please login', 'danger')\n", "return redirect(url_for('login'))\n" ]
[ "def is_logged_in(f):...\n", "@wraps(f)...\n", "if 'logged_in' in session:\n", "return f(*args, **kwargs)\n", "flash('Unauthorized, Please login', 'danger')\n", "return redirect(url_for('login'))\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Condition", "Return'", "Expr'", "Return'" ]
[ "import socket\n", "import threading\n", "import tornado\n", "import tornado.concurrent\n", "import tornado.httputil\n", "import tornado.httpserver\n", "import tornado.gen\n", "import tornado.ioloop\n", "import tornado.process\n", "import tornado.web\n", "from bzs import const\n", "from bzs im...
[ "import socket\n", "import threading\n", "import tornado\n", "import tornado.concurrent\n", "import tornado.httputil\n", "import tornado.httpserver\n", "import tornado.gen\n", "import tornado.ioloop\n", "import tornado.process\n", "import tornado.web\n", "from bzs import const\n", "from bzs im...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 5, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "FunctionDef'", "Assign'", "Assig...
[ "import random\n", "from flask import url_for, redirect, render_template, request\n", "from . import bp as app\n", "@app.route('/')...\n", "return render_template('home.html')\n" ]
[ "import random\n", "from flask import url_for, redirect, render_template, request\n", "from . import bp as app\n", "@app.route('/')...\n", "return render_template('home.html')\n" ]
[ 0, 0, 0, 0, 3 ]
[ "Import'", "ImportFrom'", "ImportFrom'", "Condition", "Return'" ]
[ "def FUNC_12(self, VAR_13, VAR_12):...\n", "self.send_rep(VAR_13, 0, wzrpc.status.error, VAR_12)\n" ]
[ "def send_error_rep(self, reqid, data):...\n", "self.send_rep(reqid, 0, wzrpc.status.error, data)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_7(self, VAR_9, VAR_11=None):...\n", "get_and_check_project(VAR_9, VAR_11)\n", "VAR_25 = self.queryset.filter(VAR_1=project_pk)\n", "VAR_25 = filters.OrderingFilter().filter_queryset(self.request, VAR_25, self)\n", "VAR_26 = CLASS_1(VAR_25, many=True)\n", "return Response(VAR_26.data)\n" ]
[ "def list(self, request, project_pk=None):...\n", "get_and_check_project(request, project_pk)\n", "tasks = self.queryset.filter(project=project_pk)\n", "tasks = filters.OrderingFilter().filter_queryset(self.request, tasks, self)\n", "serializer = TaskSerializer(tasks, many=True)\n", "return Response(seria...
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_6(self, VAR_33):...\n", "VAR_78 = []\n", "for v in self.splitter.split(VAR_33):\n", "VAR_21 = self.id_re.match(v)\n", "return VAR_78\n", "if VAR_21:\n", "VAR_108 = CLASS_5(None, False).run(VAR_21.group(1))\n", "if VAR_108:\n", "VAR_78.append(VAR_108)\n" ]
[ "def run(self, val):...\n", "res = []\n", "for v in self.splitter.split(val):\n", "link_id = self.id_re.match(v)\n", "return res\n", "if link_id:\n", "l = VLink(None, False).run(link_id.group(1))\n", "if l:\n", "res.append(l)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "For", "Assign'", "Return'", "Condition", "Assign'", "Condition", "Expr'" ]
[ "@property...\n", "return self.type == 'CO'\n" ]
[ "@property...\n", "return self.type == 'CO'\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_3(VAR_4):...\n", "VAR_41 = list(set(itertools.chain.from_iterable(VAR_4.values())))\n", "VAR_42 = {obj.id: obj for obj in education_group_year.search(id=flat_root_ids)}\n", "return {obj_id: sorted([VAR_42[VAR_60] for VAR_60 in parents], VAR_57=lambda\n obj: obj.acronym) for obj_id, parents in VAR...
[ "def _convert_parent_ids_to_instances(root_ids_by_object_id):...\n", "flat_root_ids = list(set(itertools.chain.from_iterable(\n root_ids_by_object_id.values())))\n", "map_instance_by_id = {obj.id: obj for obj in education_group_year.search(id\n =flat_root_ids)}\n", "return {obj_id: sorted([map_instance_...
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_2(self):...\n", "return \"\"\"Task: %s (ID %d)\nFile: %s\n\"\"\" % (self.task[1], self.task[0], repr\n (self.filenames)) + GenericRequest.specific_info(self)\n" ]
[ "def specific_info(self):...\n", "return \"\"\"Task: %s (ID %d)\nFile: %s\n\"\"\" % (self.task[1], self.task[0], repr\n (self.filenames)) + GenericRequest.specific_info(self)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "async def FUNC_8(VAR_6):...\n", "VAR_10 = {'Authorization': f'Bearer {VAR_1}'}\n", "VAR_8 = await VAR_6.get('/results', VAR_10=headers)\n", "VAR_9 = await VAR_8.json()\n", "assert VAR_8.status == 200\n" ]
[ "async def test_positive_get_all_results(test_cli):...\n", "headers = {'Authorization': f'Bearer {access_token}'}\n", "resp = await test_cli.get('/results', headers=headers)\n", "resp_json = await resp.json()\n", "assert resp.status == 200\n" ]
[ 0, 0, 0, 0, 0 ]
[ "AsyncFunctionDef'", "Assign'", "Assign'", "Assign'", "Assert'" ]
[ "@VAR_2.route('/update_comment/<pic_name>', methods=['POST', 'GET'])...\n", "if request.method == 'POST':\n", "VAR_4 = (\n \"UPDATE commentsTable\\t\\t\\t\\t\\tSET comment = '%s'\\t\\t\\t\\t\\tWHERE comment_id = '%s' \"\n % (request.form['new_comment'], request.form['edit_value']))\n", "return redirect...
[ "@app.route('/update_comment/<pic_name>', methods=['POST', 'GET'])...\n", "if request.method == 'POST':\n", "query = (\n \"UPDATE commentsTable\\t\\t\\t\\t\\tSET comment = '%s'\\t\\t\\t\\t\\tWHERE comment_id = '%s' \"\n % (request.form['new_comment'], request.form['edit_value']))\n", "return redirect(u...
[ 0, 0, 4, 4, 0, 0 ]
[ "Condition", "Condition", "Assign'", "Return'", "Expr'", "Expr'" ]
[ "def FUNC_17(self, VAR_24):...\n", "" ]
[ "def process_wz_msg(self, frames):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "def FUNC_18(self):...\n", "\"\"\"docstring\"\"\"\n", "VAR_47 = self.auth_method\n", "if not VAR_47:\n", "for VAR_47 in self.get_auth_methods(config.ensure_configured()):\n", "if VAR_47 not in VAR_4:\n", "if VAR_47 in VAR_4:\n", "return VAR_4[VAR_47]\n" ]
[ "def _get_users_api(self):...\n", "\"\"\"docstring\"\"\"\n", "method = self.auth_method\n", "if not method:\n", "for method in self.get_auth_methods(config.ensure_configured()):\n", "if method not in _METHOD_TO_USERS_API:\n", "if method in _METHOD_TO_USERS_API:\n", "return _METHOD_TO_USERS_API[method]...
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "For", "Condition", "Condition", "Return'" ]
[ "def FUNC_26(VAR_30, *VAR_12, **VAR_17):...\n", "if not GLSetting.memory_copy.anomaly_checks:\n", "return VAR_20(VAR_30, *VAR_12, **kw)\n", "if GLSetting.anomalies_counter[VAR_1] > alarm_level[VAR_1]:\n", "if VAR_1 == 'new_submission':\n", "return VAR_20(VAR_30, *VAR_12, **kw)\n", "log.debug('Blocked a ...
[ "def call_handler(cls, *args, **kw):...\n", "if not GLSetting.memory_copy.anomaly_checks:\n", "return method_handler(cls, *args, **kw)\n", "if GLSetting.anomalies_counter[element] > alarm_level[element]:\n", "if element == 'new_submission':\n", "return method_handler(cls, *args, **kw)\n", "log.debug('Bl...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Return'", "Condition", "Condition", "Return'", "Expr'", "Condition", "Expr'", "Condition", "Expr'", "Condition", "Expr'", "Expr'" ]
[ "def FUNC_22(self):...\n", "self.assertRaises(NotImplementedError, self.driver.local_path, '')\n" ]
[ "def test_local_path(self):...\n", "self.assertRaises(NotImplementedError, self.driver.local_path, '')\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "@api.public...\n", "FUNC_31(self, 'DELETE')\n" ]
[ "@api.public...\n", "record(self, 'DELETE')\n" ]
[ 0, 0 ]
[ "Condition", "Expr'" ]
[ "def FUNC_0(VAR_0):...\n", "\"\"\"docstring\"\"\"\n", "if not VAR_83.included_stack:\n", "return None\n", "return os.path.join(os.path.dirname(VAR_83.included_stack[-1]), VAR_0)\n" ]
[ "def srcdir(path):...\n", "\"\"\"docstring\"\"\"\n", "if not workflow.included_stack:\n", "return None\n", "return os.path.join(os.path.dirname(workflow.included_stack[-1]), path)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Return'", "Return'" ]
[ "import time\n", "import mox\n", "import paramiko\n", "from cinder import context\n", "from cinder import exception\n", "from cinder.openstack.common import log as logging\n", "from cinder.openstack.common import processutils\n", "from cinder import test\n", "from cinder.volume import configuration ...
[ "import time\n", "import mox\n", "import paramiko\n", "from cinder import context\n", "from cinder import exception\n", "from cinder.openstack.common import log as logging\n", "from cinder.openstack.common import processutils\n", "from cinder import test\n", "from cinder.volume import configuration ...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "Import'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "FunctionDef'", "Expr'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "...
[ "def FUNC_2(VAR_0, VAR_1):...\n", "VAR_5 = frappe.session.user\n", "VAR_6 = get_leave_allocation_records(VAR_0.to_date)\n", "VAR_7 = get_leave_allocation_records(VAR_0.from_date)\n", "VAR_8 = frappe.get_all('Employee', VAR_0={'status': 'Active', 'company':\n filters.company}, fields=['name', 'employee_na...
[ "def get_data(filters, leave_types):...\n", "user = frappe.session.user\n", "allocation_records_based_on_to_date = get_leave_allocation_records(filters.\n to_date)\n", "allocation_records_based_on_from_date = get_leave_allocation_records(filters\n .from_date)\n", "active_employees = frappe.get_all('Em...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 4, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Return'", "Condition", "Assign'", "For", "Assign'", "Expr'", "Assign'", "Assign'", "AugAssign'" ]
[ "def FUNC_5(self, VAR_8):...\n", "\"\"\"docstring\"\"\"\n", "VAR_14 = hmac.new(b'd1b964811afb40118a12068ff74a12f4', digestmod=hashlib.sha1)\n", "VAR_15 = self.login_data['grant_type']\n", "VAR_16 = self.login_data['client_id']\n", "VAR_17 = self.login_data['source']\n", "VAR_14.update(bytes(VAR_15 + VAR...
[ "def _get_signature(self, timestamp):...\n", "\"\"\"docstring\"\"\"\n", "ha = hmac.new(b'd1b964811afb40118a12068ff74a12f4', digestmod=hashlib.sha1)\n", "grant_type = self.login_data['grant_type']\n", "client_id = self.login_data['client_id']\n", "source = self.login_data['source']\n", "ha.update(bytes(g...
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Return'" ]
[ "def FUNC_1(self, VAR_6, VAR_7, VAR_5):...\n", "VAR_14 = []\n", "VAR_14.append('string'.format(VAR_6))\n", "VAR_14.append('string'.format(VAR_6, VAR_5, VAR_7))\n", "for query in VAR_14:\n", "self.cur.execute(query)\n", "self.conn.commit()\n" ]
[ "def update_roster(self, username, role, server_id):...\n", "sql = []\n", "sql.append(\n \"\"\"INSERT INTO users (username)\n VALUES ('{0}')\n ON DUPLICATE KEY UPDATE username = '{0}';\n \"\"\"\n .format(username))\n", "sql.append(\n \"\"...
[ 0, 4, 4, 4, 4, 4, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Expr'", "For", "Expr'", "Expr'" ]
[ "def FUNC_24(VAR_5):...\n", "\"\"\"docstring\"\"\"\n", "VAR_40 = [VAR_5['name']]\n", "VAR_27 = VAR_5['holders']\n", "for holder_no, holder in enumerate(VAR_27):\n", "VAR_43 = VAR_33[min(len(VAR_27) - (holder_no + 1), 1)]\n", "return VAR_40\n", "VAR_44 = FUNC_24(holder)\n", "for line_no, line in enum...
[ "def format_tree(tree):...\n", "\"\"\"docstring\"\"\"\n", "result = [tree['name']]\n", "holders = tree['holders']\n", "for holder_no, holder in enumerate(holders):\n", "spacer_style = spacers[min(len(holders) - (holder_no + 1), 1)]\n", "return result\n", "subtree_lines = format_tree(holder)\n", "for...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "For", "Assign'", "Return'", "Assign'", "For", "Expr'" ]
[ "@api.multi...\n", "return self.write({'state': 'active'})\n" ]
[ "@api.multi...\n", "return self.write({'state': 'active'})\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_9(self, VAR_10, VAR_6, VAR_11, VAR_12, VAR_13, VAR_14):...\n", "if VAR_12:\n", "VAR_36 = db.log_event('Information', 'Debugger', 'Command', VAR_10, success\n =False)\n", "VAR_6 = [bytes(output, encoding='utf-8') for output in VAR_6]\n", "VAR_27 = ''\n", "if VAR_11 is None:\n", "VAR_11 = VAR...
[ "def command(self, command, expected_output, error_message, log_event,...\n", "if log_event:\n", "event = db.log_event('Information', 'Debugger', 'Command', command, success\n =False)\n", "expected_output = [bytes(output, encoding='utf-8') for output in\n expected_output]\n", "return_buffer = ''\n", ...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "AugAssign'", "AugAssign'", "Condition", "Expr'", "Condition", "Expr'", "For", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "A...
[ "def FUNC_0(VAR_0):...\n", "VAR_6 = (\"\"\"INSERT INTO dialogs (name, created_at)\nVALUES ('%s', NOW())\"\"\" %\n VAR_0)\n", "return {'status': 1}\n" ]
[ "def db_addDialog(nameDialog):...\n", "sql = (\"\"\"INSERT INTO dialogs (name, created_at)\nVALUES ('%s', NOW())\"\"\" %\n nameDialog)\n", "return {'status': 1}\n" ]
[ 0, 4, 4 ]
[ "FunctionDef'", "Assign'", "Return'" ]
[ "@rest_utils.ajax(data_required=True)...\n", "VAR_30 = VAR_1.DATA.get('flavor_access', [])\n", "VAR_27 = not VAR_30\n", "VAR_33 = api.nova.flavor_get_extras(VAR_1, VAR_5, raw=True)\n", "api.nova.flavor_delete(VAR_1, VAR_5)\n", "VAR_31 = api.nova.flavor_create(VAR_1, VAR_6=request.DATA['name'], memory=\n ...
[ "@rest_utils.ajax(data_required=True)...\n", "flavor_access = request.DATA.get('flavor_access', [])\n", "is_public = not flavor_access\n", "extras_dict = api.nova.flavor_get_extras(request, flavor_id, raw=True)\n", "api.nova.flavor_delete(request, flavor_id)\n", "flavor = api.nova.flavor_create(request, n...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "For", "Expr'", "Condition", "Expr'" ]
[ "def FUNC_12(self):...\n", "\"\"\"docstring\"\"\"\n", "VAR_50 = frappe.db.get_table_columns(self.doctype)\n", "VAR_51 = []\n", "for fld in self.fields:\n", "for VAR_34 in optional_fields:\n", "for fld in VAR_51:\n", "if VAR_34 in fld and not VAR_34 in VAR_50:\n", "VAR_51 = []\n", "VAR_51.append(fl...
[ "def set_optional_columns(self):...\n", "\"\"\"docstring\"\"\"\n", "columns = frappe.db.get_table_columns(self.doctype)\n", "to_remove = []\n", "for fld in self.fields:\n", "for f in optional_fields:\n", "for fld in to_remove:\n", "if f in fld and not f in columns:\n", "to_remove = []\n", "to_remo...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "For", "For", "For", "Condition", "Assign'", "Expr'", "For", "Condition", "For", "Assign'", "For", "Condition", "Condition", "Expr'", "Expr'" ]
[ "def __init__(self, VAR_6, VAR_7=None):...\n", "super().__init__(VAR_6, VAR_7)\n", "self.type = 'ip'\n" ]
[ "def __init__(self, name, display_name=None):...\n", "super().__init__(name, display_name)\n", "self.type = 'ip'\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Assign'" ]
[ "def FUNC_19(VAR_16):...\n", "VAR_37 = os.path.join(VAR_16, 'README.md')\n", "if os.path.isfile(VAR_37):\n", "return 'LICENSE' in open(VAR_37).read()\n", "return False\n" ]
[ "def license_mentionned_in_readme(path):...\n", "readme_path = os.path.join(path, 'README.md')\n", "if os.path.isfile(readme_path):\n", "return 'LICENSE' in open(readme_path).read()\n", "return False\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Return'", "Return'" ]
[ "def FUNC_1():...\n", "VAR_3 = {'foo': 'bar'}\n", "VAR_1 = json.dumps(VAR_3)\n", "VAR_2 = load_source(VAR_1)\n", "assert VAR_2 == VAR_3\n" ]
[ "def test_json_string():...\n", "native = {'foo': 'bar'}\n", "source = json.dumps(native)\n", "result = load_source(source)\n", "assert result == native\n" ]
[ 0, 7, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assert'" ]
[ "\"\"\"\nsabnzbd.database - Database Support\n\"\"\"\n", "import sqlite3\n", "import os\n", "import pysqlite2.dbapi2 as sqlite3\n", "import time\n", "import zlib\n", "import logging\n", "import sys\n", "import threading\n", "import sabnzbd\n", "import sabnzbd.cfg\n", "from sabnzbd.constants im...
[ "\"\"\"\nsabnzbd.database - Database Support\n\"\"\"\n", "import sqlite3\n", "import os\n", "import pysqlite2.dbapi2 as sqlite3\n", "import time\n", "import zlib\n", "import logging\n", "import sys\n", "import threading\n", "import sabnzbd\n", "import sabnzbd.cfg\n", "from sabnzbd.constants im...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Expr'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "FunctionDef'", "Docstring", "Condition", "Assign'", "Assign'", "Co...
[ "@transaction.atomic...\n", "VAR_6 = RegisterStudentForm(VAR_0.POST)\n", "if not VAR_6.is_valid():\n", "return FUNC_0(VAR_0, VAR_4=form)\n", "VAR_9 = Course.objects.get(course_number=request.POST['course_number'])\n", "VAR_10 = Student.objects.get(andrew_id=request.POST['andrew_id'])\n", "VAR_9.students...
[ "@transaction.atomic...\n", "form = RegisterStudentForm(request.POST)\n", "if not form.is_valid():\n", "return make_view(request, register_student_form=form)\n", "course = Course.objects.get(course_number=request.POST['course_number'])\n", "student = Student.objects.get(andrew_id=request.POST['andrew_id']...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Condition", "Return'", "Assign'", "Assign'", "Expr'", "Expr'", "Return'" ]
[ "@integration_synonym_api...\n", "FUNC_2(VAR_2)\n", "FUNC_3(VAR_2, 'FEDS', VAR_7='1')\n", "FUNC_5(VAR_4, VAR_5, VAR_11='FADS', VAR_10=[{'name': '----FADS'}])\n" ]
[ "@integration_synonym_api...\n", "clean_database(solr)\n", "seed_database_with(solr, 'FEDS', id='1')\n", "verify_results(client, jwt, query='FADS', expected=[{'name': '----FADS'}])\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_2():...\n", "super(CLASS_6, self).__init__(slideParser.parse(VAR_10, slideLexer))\n" ]
[ "def innerFunc():...\n", "super(Emph, self).__init__(slideParser.parse(txt, slideLexer))\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_28(self, VAR_2, VAR_3, VAR_4, VAR_32, VAR_7={}):...\n", "if 'hours' in VAR_32 and not VAR_32['hours']:\n", "VAR_32['hours'] = 0.0\n", "if 'hours' in VAR_32:\n", "for work in self.browse(VAR_2, VAR_3, VAR_4, VAR_7):\n", "return super(CLASS_3, self).write(VAR_2, VAR_3, VAR_4, VAR_32, VAR_7)\n", ...
[ "def write(self, cr, uid, ids, vals, context={}):...\n", "if 'hours' in vals and not vals['hours']:\n", "vals['hours'] = 0.0\n", "if 'hours' in vals:\n", "for work in self.browse(cr, uid, ids, context):\n", "return super(project_work, self).write(cr, uid, ids, vals, context)\n", "cr.execute(\n 'updat...
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Condition", "For", "Return'", "Expr'" ]
[ "def FUNC_25(self, VAR_21):...\n", "\"\"\"docstring\"\"\"\n", "VAR_57 = self.getfile(VAR_21, VAR_23=False)\n", "if VAR_57 == False:\n", "if not VAR_57[VAR_6] & FUNC_30.S_IFLNK:\n", "return VAR_57[VAR_9]\n" ]
[ "def readlink(self, path):...\n", "\"\"\"docstring\"\"\"\n", "p = self.getfile(path, follow_symlinks=False)\n", "if p == False:\n", "if not p[A_MODE] & stat.S_IFLNK:\n", "return p[A_TARGET]\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Condition", "Return'" ]
[ "def FUNC_0(self, VAR_1):...\n", "VAR_5 = ('select comment from comments order by date desc where userid=%d' %\n VAR_1)\n", "VAR_6 = sql.queryDB(self.conn, VAR_5)\n", "return VAR_6\n" ]
[ "def getCommentsByUser(self, userid):...\n", "sqlText = (\n 'select comment from comments order by date desc where userid=%d' % userid)\n", "result = sql.queryDB(self.conn, sqlText)\n", "return result\n" ]
[ 0, 4, 4, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Return'" ]
[ "import sublime\n", "import sublime_plugin\n", "import os\n", "import re\n", "import datetime\n", "import Urtext.urtext as Urtext\n", "import pprint\n", "import sys\n", "sys.path.append(os.path.join(os.path.dirname(__file__)))\n", "from anytree import Node, RenderTree\n", "import anytree\n", "...
[ "import sublime\n", "import sublime_plugin\n", "import os\n", "import re\n", "import datetime\n", "import Urtext.urtext as Urtext\n", "import pprint\n", "import sys\n", "sys.path.append(os.path.join(os.path.dirname(__file__)))\n", "from anytree import Node, RenderTree\n", "import anytree\n", "...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "Expr'", "ImportFrom'", "Import'", "Import'", "FunctionDef'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_7(self, VAR_2, VAR_3, VAR_4, *VAR_5):...\n", "self.write_data({'type': 'channel_discovered', 'event': VAR_5[0]})\n" ]
[ "def on_channel_discovered(self, subject, changetype, objectID, *args):...\n", "self.write_data({'type': 'channel_discovered', 'event': args[0]})\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]