lines
listlengths
1
444
raw_lines
listlengths
1
444
label
listlengths
1
444
type
listlengths
1
444
[ "def FUNC_1(self):...\n", "VAR_0 = self.objects.extext.pk\n", "self.assertIsNone(self.objects.extext.qa_group)\n", "self.assertEqual(len(QAGroup.objects.all()), 0)\n", "VAR_0 = self.objects.extext.extraction_script.pk\n", "VAR_1 = self.client.get(f'/qa/extractionscript/{VAR_0}/')\n", "self.assertEqual(VAR_1.status_code, 200)\n", "VAR_2 = QAGroup.objects.get(extraction_script=self.objects.extext.\n extraction_script)\n", "VAR_3 = ExtractedText.objects.get(VAR_2=qa_group)\n", "self.assertIsNotNone(VAR_3.qa_group)\n", "VAR_1 = self.client.get(f'/qa/extractedtext/{VAR_3.pk}/')\n" ]
[ "def test_qa_group_creation(self):...\n", "pk = self.objects.extext.pk\n", "self.assertIsNone(self.objects.extext.qa_group)\n", "self.assertEqual(len(QAGroup.objects.all()), 0)\n", "pk = self.objects.extext.extraction_script.pk\n", "response = self.client.get(f'/qa/extractionscript/{pk}/')\n", "self.assertEqual(response.status_code, 200)\n", "qa_group = QAGroup.objects.get(extraction_script=self.objects.extext.\n extraction_script)\n", "ext = ExtractedText.objects.get(qa_group=qa_group)\n", "self.assertIsNotNone(ext.qa_group)\n", "response = self.client.get(f'/qa/extractedtext/{ext.pk}/')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Expr'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'", "Assign'" ]
[ "def FUNC_17(self, VAR_35=False, VAR_36=False, VAR_37=None, VAR_38=None):...\n", "self.clean()\n", "return super().save(VAR_35, VAR_36, VAR_37, VAR_38)\n" ]
[ "def save(self, force_insert=False, force_update=False, using=None,...\n", "self.clean()\n", "return super().save(force_insert, force_update, using, update_fields)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Return'" ]
[ "def FUNC_6(self, VAR_11='', VAR_6=None):...\n", "VAR_5 = self.session.query(self.obj)\n", "VAR_5 = self._get_base_query(VAR_5=query, VAR_6=filters)\n", "VAR_25 = VAR_5.all()\n", "VAR_27 = GroupByDateYear(VAR_11, 'Group by Year')\n", "return VAR_27.apply(VAR_25)\n" ]
[ "def query_year_group(self, group_by='', filters=None):...\n", "query = self.session.query(self.obj)\n", "query = self._get_base_query(query=query, filters=filters)\n", "query_result = query.all()\n", "group_year = GroupByDateYear(group_by, 'Group by Year')\n", "return group_year.apply(query_result)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_6(VAR_6):...\n", "\"\"\"docstring\"\"\"\n", "VAR_23 = open('keys/%s/key_rsa.pub' % VAR_6, 'rb').read()\n", "return VAR_23\n" ]
[ "def loadClientRSAKey(client):...\n", "\"\"\"docstring\"\"\"\n", "key = open('keys/%s/key_rsa.pub' % client, 'rb').read()\n", "return key\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Return'" ]
[ "def FUNC_20(self, VAR_25, VAR_23):...\n", "VAR_19 = threading.Event()\n", "def FUNC_28(VAR_21):...\n", "if not VAR_25(VAR_21):\n", "return VAR_21, False\n", "VAR_19.set()\n", "return VAR_21, True\n" ]
[ "def _get_message_handle(self, match, handlername):...\n", "event = threading.Event()\n", "def handler(msg):...\n", "if not match(msg):\n", "return msg, False\n", "event.set()\n", "return msg, True\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "FunctionDef'", "Condition", "Return'", "Expr'", "Return'" ]
[ "def FUNC_5(VAR_1, *VAR_2, **VAR_3):...\n", "def FUNC_6():...\n", "if not request.method in VAR_0:\n", "return True\n", "if not request.environ.get('AUTH_TYPE') == 'cookie':\n", "return True\n", "if config.get('skip_authentication'):\n", "return True\n", "if request.environ.get('HTTP_REFERER'):\n", "VAR_7 = urlparse(request.environ.get('HTTP_REFERER'))\n", "if request.method == 'GET' and FUNC_4():\n", "VAR_8 = VAR_7.hostname\n", "return True\n", "return False\n", "if VAR_7.port:\n", "VAR_8 += ':' + str(VAR_7.port)\n", "if VAR_8.endswith(request.environ['adhocracy.domain']):\n", "if request.method != 'GET':\n", "return True\n" ]
[ "def _decorate(f, *a, **kw):...\n", "def check():...\n", "if not request.method in methods:\n", "return True\n", "if not request.environ.get('AUTH_TYPE') == 'cookie':\n", "return True\n", "if config.get('skip_authentication'):\n", "return True\n", "if request.environ.get('HTTP_REFERER'):\n", "ref_url = urlparse(request.environ.get('HTTP_REFERER'))\n", "if request.method == 'GET' and has_token():\n", "ref_host = ref_url.hostname\n", "return True\n", "return False\n", "if ref_url.port:\n", "ref_host += ':' + str(ref_url.port)\n", "if ref_host.endswith(request.environ['adhocracy.domain']):\n", "if request.method != 'GET':\n", "return True\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 5, 0, 0, 0, 0, 0, 0, 5, 0 ]
[ "FunctionDef'", "FunctionDef'", "Condition", "Return'", "Condition", "Return'", "Condition", "Return'", "Condition", "Assign'", "Condition", "Assign'", "Return'", "Return'", "Condition", "AugAssign'", "Condition", "Condition", "Return'" ]
[ "def __ne__(self, VAR_1):...\n", "return not self.__eq__(VAR_1)\n" ]
[ "def __ne__(self, other):...\n", "return not self.__eq__(other)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_9(self, *VAR_15, **VAR_75):...\n", "" ]
[ "def newfn(self, *a, **env):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "\"\"\"string\"\"\"\n", "from lib.core.agent import agent\n", "from lib.core.data import conf\n", "from lib.core.data import kb\n", "from lib.core.data import logger\n", "from lib.core.data import queries\n", "from lib.core.session import setUnion\n", "from lib.request.connect import Connect as Request\n", "def FUNC_0(VAR_0, VAR_1):...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = {}\n", "for count in range(0, 50):\n", "if kb.dbms == 'Oracle' and VAR_0.endswith(' FROM DUAL'):\n", "return None\n", "VAR_0 = VAR_0[:-len(' FROM DUAL')]\n", "if count:\n", "VAR_0 += ', NULL'\n", "if kb.dbms == 'Oracle':\n", "VAR_0 += ' FROM DUAL'\n", "VAR_5 = agent.postfixQuery(VAR_0, VAR_1)\n", "VAR_6 = agent.payload(newValue=commentedQuery)\n", "VAR_7 = Request.queryPage(VAR_6)\n", "if not VAR_7 in VAR_2.keys():\n", "VAR_2[VAR_7] = 1, VAR_5\n", "VAR_2[VAR_7] = VAR_2[VAR_7][0] + 1, VAR_5\n", "if count:\n", "for element in VAR_2.values():\n", "if element[0] == 1:\n", "if kb.injPlace == 'GET':\n", "VAR_4 = '%s?%s' % (conf.url, VAR_6)\n", "if kb.injPlace == 'POST':\n", "return VAR_4\n", "VAR_4 = \"URL:\\t'%s'\" % conf.url\n", "if kb.injPlace == 'Cookie':\n", "VAR_4 += \"\\nPOST:\\t'%s'\\n\" % VAR_6\n", "VAR_4 = \"URL:\\t'%s'\" % conf.url\n", "if kb.injPlace == 'User-Agent':\n", "VAR_4 += \"\\nCookie:\\t'%s'\\n\" % VAR_6\n", "VAR_4 = \"URL:\\t\\t'%s'\" % conf.url\n", "VAR_4 += \"\"\"\nUser-Agent:\t'%s'\n\"\"\" % VAR_6\n" ]
[ "\"\"\"\n$Id$\n\nThis file is part of the sqlmap project, http://sqlmap.sourceforge.net.\n\nCopyright (c) 2006-2008 Bernardo Damele A. G. <bernardo.damele@gmail.com>\n and Daniele Bellucci <daniele.bellucci@gmail.com>\n\nsqlmap is free software; you can redistribute it and/or modify it under\nthe terms of the GNU General Public License as published by the Free\nSoftware Foundation version 2 of the License.\n\nsqlmap is distributed in the hope that it will be useful, but WITHOUT ANY\nWARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS\nFOR A PARTICULAR PURPOSE. See the GNU General Public License for more\ndetails.\n\nYou should have received a copy of the GNU General Public License along\nwith sqlmap; if not, write to the Free Software Foundation, Inc., 51\nFranklin St, Fifth Floor, Boston, MA 02110-1301 USA\n\"\"\"\n", "from lib.core.agent import agent\n", "from lib.core.data import conf\n", "from lib.core.data import kb\n", "from lib.core.data import logger\n", "from lib.core.data import queries\n", "from lib.core.session import setUnion\n", "from lib.request.connect import Connect as Request\n", "def __effectiveUnionTest(query, comment):...\n", "\"\"\"docstring\"\"\"\n", "resultDict = {}\n", "for count in range(0, 50):\n", "if kb.dbms == 'Oracle' and query.endswith(' FROM DUAL'):\n", "return None\n", "query = query[:-len(' FROM DUAL')]\n", "if count:\n", "query += ', NULL'\n", "if kb.dbms == 'Oracle':\n", "query += ' FROM DUAL'\n", "commentedQuery = agent.postfixQuery(query, comment)\n", "payload = agent.payload(newValue=commentedQuery)\n", "newResult = Request.queryPage(payload)\n", "if not newResult in resultDict.keys():\n", "resultDict[newResult] = 1, commentedQuery\n", "resultDict[newResult] = resultDict[newResult][0] + 1, commentedQuery\n", "if count:\n", "for element in resultDict.values():\n", "if element[0] == 1:\n", "if kb.injPlace == 'GET':\n", "value = '%s?%s' % (conf.url, payload)\n", "if kb.injPlace == 'POST':\n", "return value\n", "value = \"URL:\\t'%s'\" % conf.url\n", "if kb.injPlace == 'Cookie':\n", "value += \"\\nPOST:\\t'%s'\\n\" % payload\n", "value = \"URL:\\t'%s'\" % conf.url\n", "if kb.injPlace == 'User-Agent':\n", "value += \"\\nCookie:\\t'%s'\\n\" % payload\n", "value = \"URL:\\t\\t'%s'\" % conf.url\n", "value += \"\"\"\nUser-Agent:\t'%s'\n\"\"\" % payload\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2, 0, 0, 0, 0, 0, 2, 0, 0, 0, 0, 2, 2, 0, 2, 0, 0, 0, 0, 2, 0, 0, 2, 0, 2 ]
[ "Expr'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "FunctionDef'", "Docstring", "Assign'", "For", "Condition", "Return'", "Assign'", "Condition", "AugAssign'", "Condition", "AugAssign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "For", "Condition", "Condition", "Assign'", "Condition", "Return'", "Assign'", "Condition", "AugAssign'", "Assign'", "Condition", "AugAssign'", "Assign'", "AugAssign'" ]
[ "def FUNC_9(self, *VAR_15, **VAR_75):...\n", "" ]
[ "def newfn(self, *a, **env):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "@property...\n", "\"\"\"docstring\"\"\"\n", "return self._attributes.get('dimensions', {}).copy()\n" ]
[ "@property...\n", "\"\"\"docstring\"\"\"\n", "return self._attributes.get('dimensions', {}).copy()\n" ]
[ 0, 0, 0 ]
[ "Condition", "Docstring", "Return'" ]
[ "def FUNC_1(self, VAR_2):...\n", "" ]
[ "def get(self, key):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "def FUNC_4(self, VAR_1, VAR_2):...\n", "VAR_12 = self._formatCredentials(VAR_1, VAR_8='current')\n", "VAR_9 = '{credentials} rclone touch current:{path}/.keep'.format(VAR_12=\n credentials, VAR_2=path)\n", "VAR_28 = self._execute(VAR_9)\n", "return {'message': 'Success'}\n" ]
[ "def mkdir(self, data, path):...\n", "credentials = self._formatCredentials(data, name='current')\n", "command = '{credentials} rclone touch current:{path}/.keep'.format(credentials\n =credentials, path=path)\n", "result = self._execute(command)\n", "return {'message': 'Success'}\n" ]
[ 0, 0, 2, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_15(self):...\n", "self._test_health(FUNC_1, cluster.ClusterHealth.RED)\n" ]
[ "def test_red_health(self):...\n", "self._test_health(_validate_conn_down, cluster.ClusterHealth.RED)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "@handled_slot(bool)...\n", "print('Starting thin etalon scan...')\n", "self.matisse.thin_etalon_scan()\n" ]
[ "@handled_slot(bool)...\n", "print('Starting thin etalon scan...')\n", "self.matisse.thin_etalon_scan()\n" ]
[ 0, 0, 0 ]
[ "Condition", "Expr'", "Expr'" ]
[ "from typing import Any, Dict, Iterator, TypeVar\n", "from typing_extensions import Protocol\n", "VAR_0 = TypeVar('T')\n", "def __iter__(self) ->'Iterator[IRow]':...\n", "...\n", "def FUNC_0(self, VAR_1: bool) ->bool:...\n", "...\n", "def __getattr__(self, VAR_2: str) ->Any:...\n", "...\n", "def __iter__(self) ->Iterator[CLASS_2]:...\n", "...\n", "selected_fields: 'Dict[str, ISpy]'\n", "is_subquery: bool\n", "VAR_2: str\n", "def FUNC_1(self, VAR_3: str) ->Iterator[CLASS_2]:...\n", "...\n" ]
[ "from typing import Any, Dict, Iterator, TypeVar\n", "from typing_extensions import Protocol\n", "T = TypeVar('T')\n", "def __iter__(self) ->'Iterator[IRow]':...\n", "...\n", "def filter(self, expression: bool) ->bool:...\n", "...\n", "def __getattr__(self, name: str) ->Any:...\n", "...\n", "def __iter__(self) ->Iterator[IRow]:...\n", "...\n", "selected_fields: 'Dict[str, ISpy]'\n", "is_subquery: bool\n", "name: str\n", "def query(self, query_string: str) ->Iterator[IRow]:...\n", "...\n" ]
[ 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "ImportFrom'", "ImportFrom'", "Assign'", "FunctionDef'", "Expr'", "FunctionDef'", "Expr'", "FunctionDef'", "Expr'", "FunctionDef'", "Expr'", "AnnAssign'", "AnnAssign'", "AnnAssign'", "FunctionDef'", "Expr'" ]
[ "def FUNC_21(self, VAR_1):...\n", "VAR_16 = [VAR_25.fieldname for VAR_25 in self.meta.get_table_fields() if \n VAR_25.options == VAR_1]\n", "return VAR_16[0] if VAR_16 else None\n" ]
[ "def get_parentfield_of_doctype(self, doctype):...\n", "fieldname = [df.fieldname for df in self.meta.get_table_fields() if df.\n options == doctype]\n", "return fieldname[0] if fieldname else None\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Return'" ]
[ "def FUNC_4(self):...\n", "return self._doc['globals'].keys()\n" ]
[ "def get_global_names(self):...\n", "return self._doc['globals'].keys()\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "@property...\n", "\"\"\"docstring\"\"\"\n", "return self.config.get(self.section, 'ssl_adapter')\n" ]
[ "@property...\n", "\"\"\"docstring\"\"\"\n", "return self.config.get(self.section, 'ssl_adapter')\n" ]
[ 0, 0, 0 ]
[ "Condition", "Docstring", "Return'" ]
[ "def __iter__(self):...\n", "for item in self.sequence:\n", "yield item\n" ]
[ "def __iter__(self):...\n", "for item in self.sequence:\n", "yield item\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "For", "Expr'" ]
[ "@VAR_2.route('/data/jobtypes/')...\n", "return {'jobtypes': VAR_3.jobtype_query()}\n" ]
[ "@app.route('/data/jobtypes/')...\n", "return {'jobtypes': JOBSDATA.jobtype_query()}\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_8(self):...\n", "\"\"\"docstring\"\"\"\n", "self.assertEqual(FUNC_1(), 0)\n", "self.assertEqual(FUNC_3(), 0)\n" ]
[ "def test_create_database_tables(self):...\n", "\"\"\"docstring\"\"\"\n", "self.assertEqual(drop(), 0)\n", "self.assertEqual(create(), 0)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'", "Expr'" ]
[ "def FUNC_10(self):...\n", "self.run_test_case(self.scenario.upsert_load_balancer(VAR_5=False))\n" ]
[ "def test_b_upsert_load_balancer_public(self):...\n", "self.run_test_case(self.scenario.upsert_load_balancer(use_vpc=False))\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_46(self, VAR_16):...\n", "\"\"\"docstring\"\"\"\n", "VAR_56 = self.regex.search(VAR_16)\n", "if VAR_56 is not None:\n", "return VAR_56.group('value')\n" ]
[ "def is_periodic(self, value):...\n", "\"\"\"docstring\"\"\"\n", "m = self.regex.search(value)\n", "if m is not None:\n", "return m.group('value')\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Return'" ]
[ "def FUNC_48(self):...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = {'SEVERITY': {'LOW': 1}, 'CONFIDENCE': {'HIGH': 1}}\n", "self.check_example('assert.py', VAR_2)\n" ]
[ "def test_asserts(self):...\n", "\"\"\"docstring\"\"\"\n", "expect = {'SEVERITY': {'LOW': 1}, 'CONFIDENCE': {'HIGH': 1}}\n", "self.check_example('assert.py', expect)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'" ]
[ "def FUNC_8(self):...\n", "\"\"\"docstring\"\"\"\n", "VAR_4 = '/api/apps'\n", "VAR_5 = self.client.post(VAR_4)\n", "self.assertEqual(VAR_5.status_code, 201)\n", "VAR_6 = VAR_5.data['id']\n", "VAR_4 = '/api/apps/{app_id}/builds'.format(**locals())\n", "VAR_8 = {'image': 'autotest/example', 'sha': 'a' * 40, 'procfile': json.\n dumps({'web': 'node server.js', 'worker': 'node worker.js'})}\n", "VAR_5 = self.client.post(VAR_4, json.dumps(VAR_8), content_type=\n 'application/json')\n", "self.assertEqual(VAR_5.status_code, 201)\n", "VAR_4 = '/api/apps/{app_id}/scale'.format(**locals())\n", "VAR_8 = {'web': 4, 'worker': 2}\n", "VAR_5 = self.client.post(VAR_4, json.dumps(VAR_8), content_type=\n 'application/json')\n", "self.assertEqual(VAR_5.status_code, 204)\n", "VAR_4 = '/api/apps/{app_id}/containers'.format(**locals())\n", "VAR_5 = self.client.get(VAR_4)\n", "self.assertEqual(VAR_5.status_code, 200)\n", "self.assertEqual(len(VAR_5.data['results']), 6)\n", "VAR_10 = VAR_5.data['results'][0]['uuid']\n", "VAR_11 = Container.objects.get(VAR_10=uuid)\n", "self.assertEqual(VAR_11.short_name(), '{}.{}.{}'.format(VAR_11.app, VAR_11.\n type, VAR_11.num))\n", "self.assertEqual(str(VAR_11), '{}.{}.{}'.format(VAR_11.app, VAR_11.type,\n VAR_11.num))\n" ]
[ "def test_container_str(self):...\n", "\"\"\"docstring\"\"\"\n", "url = '/api/apps'\n", "response = self.client.post(url)\n", "self.assertEqual(response.status_code, 201)\n", "app_id = response.data['id']\n", "url = '/api/apps/{app_id}/builds'.format(**locals())\n", "body = {'image': 'autotest/example', 'sha': 'a' * 40, 'procfile': json.\n dumps({'web': 'node server.js', 'worker': 'node worker.js'})}\n", "response = self.client.post(url, json.dumps(body), content_type=\n 'application/json')\n", "self.assertEqual(response.status_code, 201)\n", "url = '/api/apps/{app_id}/scale'.format(**locals())\n", "body = {'web': 4, 'worker': 2}\n", "response = self.client.post(url, json.dumps(body), content_type=\n 'application/json')\n", "self.assertEqual(response.status_code, 204)\n", "url = '/api/apps/{app_id}/containers'.format(**locals())\n", "response = self.client.get(url)\n", "self.assertEqual(response.status_code, 200)\n", "self.assertEqual(len(response.data['results']), 6)\n", "uuid = response.data['results'][0]['uuid']\n", "container = Container.objects.get(uuid=uuid)\n", "self.assertEqual(container.short_name(), '{}.{}.{}'.format(container.app,\n container.type, container.num))\n", "self.assertEqual(str(container), '{}.{}.{}'.format(container.app, container\n .type, container.num))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_2(VAR_0):...\n", "\"\"\"docstring\"\"\"\n", "VAR_6 = VAR_38(VAR_0)\n", "VAR_6 = VAR_6[VAR_6.find(\"'\") + 1:VAR_6.rfind(\"'\")]\n", "VAR_6 = ':'.join(VAR_6.rsplit('.', 1))\n", "return VAR_6\n" ]
[ "def class2str(value):...\n", "\"\"\"docstring\"\"\"\n", "s = str(value)\n", "s = s[s.find(\"'\") + 1:s.rfind(\"'\")]\n", "s = ':'.join(s.rsplit('.', 1))\n", "return s\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "from django.shortcuts import render\n", "from django.http import HttpResponse, JsonResponse\n", "from django.views.decorators.csrf import csrf_exempt\n", "from rest_framework.renderers import JSONRenderer\n", "from rest_framework.parsers import JSONParser\n", "from rest_framework.response import Response\n", "from rest_framework import viewsets\n", "from rest_framework.decorators import list_route\n", "from flask import escape\n", "from .models import BoxDetails, RegisteredServices\n", "from .serializers import BoxDetailsSerializer, RegisteredServicesSerializer\n", "import common, sqlite3, subprocess, NetworkManager, os, crypt, pwd, getpass, spwd\n", "VAR_0 = NetworkManager.NetworkManager\n", "VAR_1 = [d for d in VAR_0.Devices if isinstance(d, NetworkManager.Wireless)]\n", "def FUNC_0():...\n", "\"\"\"docstring\"\"\"\n", "VAR_17 = f.read().split('\\n')\n", "VAR_18 = VAR_17[4].split('=')[1].strip('\"')\n", "return VAR_18\n" ]
[ "from django.shortcuts import render\n", "from django.http import HttpResponse, JsonResponse\n", "from django.views.decorators.csrf import csrf_exempt\n", "from rest_framework.renderers import JSONRenderer\n", "from rest_framework.parsers import JSONParser\n", "from rest_framework.response import Response\n", "from rest_framework import viewsets\n", "from rest_framework.decorators import list_route\n", "from flask import escape\n", "from .models import BoxDetails, RegisteredServices\n", "from .serializers import BoxDetailsSerializer, RegisteredServicesSerializer\n", "import common, sqlite3, subprocess, NetworkManager, os, crypt, pwd, getpass, spwd\n", "nm = NetworkManager.NetworkManager\n", "wlans = [d for d in nm.Devices if isinstance(d, NetworkManager.Wireless)]\n", "def get_osversion():...\n", "\"\"\"docstring\"\"\"\n", "osfilecontent = f.read().split('\\n')\n", "version = osfilecontent[4].split('=')[1].strip('\"')\n", "return version\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 7, 0, 0, 0, 0, 0, 0, 0 ]
[ "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Import'", "Assign'", "Assign'", "FunctionDef'", "Docstring", "Assign'", "Assign'", "Return'" ]
[ "@aiohttp_jinja2.template('index.html')...\n", "VAR_9 = natsorted(glob(os.path.join(VAR_3.STORAGE_DIR, '**/*.jpg'),\n recursive=True), VAR_8=lambda x: x.upper())\n", "return {'images': (CLASS_0(x.replace(VAR_3.STORAGE_DIR, '')) for x in VAR_9)}\n" ]
[ "@aiohttp_jinja2.template('index.html')...\n", "images = natsorted(glob(os.path.join(settings.STORAGE_DIR, '**/*.jpg'),\n recursive=True), key=lambda x: x.upper())\n", "return {'images': (Item(x.replace(settings.STORAGE_DIR, '')) for x in images)}\n" ]
[ 0, 0, 0 ]
[ "Condition", "Assign'", "Return'" ]
[ "def FUNC_13(self, VAR_17, VAR_18, VAR_19):...\n", "if 'memory_mapped' in self.targets[VAR_18] and self.targets[VAR_18][\n", "VAR_10 = 'md'\n", "VAR_28 = self.command('rd ' + VAR_17, [':'], 'Error getting register value')\n", "if 'bits' in self.targets[VAR_18]['registers'][VAR_17]:\n", "return VAR_28.split('\\r')[0].split(':')[1].split()[0]\n", "VAR_43 = self.targets[VAR_18]['registers'][VAR_17]['bits']\n", "VAR_37 = self.targets[VAR_18]['base'][VAR_19] + self.targets[VAR_18][\n 'registers'][VAR_17]['offset']\n", "if VAR_43 == 8:\n", "VAR_28 = self.command(VAR_10 + ' ' + hex(VAR_37) + ' 1', [':'],\n 'Error getting register value')\n", "VAR_10 += 'b'\n", "if VAR_43 == 16:\n", "VAR_10 += 'h'\n", "if VAR_43 == 64:\n", "VAR_10 += 'd'\n" ]
[ "def get_register_value(self, register, target, target_index):...\n", "if 'memory_mapped' in self.targets[target] and self.targets[target][\n", "command = 'md'\n", "buff = self.command('rd ' + register, [':'], 'Error getting register value')\n", "if 'bits' in self.targets[target]['registers'][register]:\n", "return buff.split('\\r')[0].split(':')[1].split()[0]\n", "bits = self.targets[target]['registers'][register]['bits']\n", "address = self.targets[target]['base'][target_index] + self.targets[target][\n 'registers'][register]['offset']\n", "if bits == 8:\n", "buff = self.command(command + ' ' + hex(address) + ' 1', [':'],\n 'Error getting register value')\n", "command += 'b'\n", "if bits == 16:\n", "command += 'h'\n", "if bits == 64:\n", "command += 'd'\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Assign'", "Condition", "Return'", "Assign'", "Assign'", "Condition", "Assign'", "AugAssign'", "Condition", "AugAssign'", "Condition", "AugAssign'" ]
[ "@app.task...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = get_object_or_404(Project, pk=project_id)\n", "VAR_3 = create_routing_test_suite(VAR_2=project)\n", "VAR_4 = TestCaseResults.objects.filter(VAR_2=project)\n", "if VAR_4.count() > 2:\n", "VAR_7 = VAR_2.testcaseresults_set.latest('updated')\n", "TestCaseResults.objects.create(VAR_2=project, results=testcases)\n", "print(str(e))\n", "VAR_5 = 'TestCases updated.'\n", "if VAR_7.results == VAR_3:\n", "if not VAR_1:\n", "VAR_7.updated = datetime.now()\n", "VAR_11 = VAR_2.testcaseresults_set.earliest('updated')\n", "VAR_5 = FUNC_1(VAR_2.id)\n", "return VAR_5\n", "VAR_7.save()\n", "VAR_11.results = VAR_3\n", "VAR_11.updated = datetime.now()\n", "VAR_11.save()\n" ]
[ "@app.task...\n", "\"\"\"docstring\"\"\"\n", "project = get_object_or_404(Project, pk=project_id)\n", "testcases = create_routing_test_suite(project=project)\n", "tc_results = TestCaseResults.objects.filter(project=project)\n", "if tc_results.count() > 2:\n", "tc_latest = project.testcaseresults_set.latest('updated')\n", "TestCaseResults.objects.create(project=project, results=testcases)\n", "print(str(e))\n", "msg = 'TestCases updated.'\n", "if tc_latest.results == testcases:\n", "if not call_from:\n", "tc_latest.updated = datetime.now()\n", "tc_earliest = project.testcaseresults_set.earliest('updated')\n", "msg = push_testcases_to_testrail_celery(project.id)\n", "return msg\n", "tc_latest.save()\n", "tc_earliest.results = testcases\n", "tc_earliest.updated = datetime.now()\n", "tc_earliest.save()\n" ]
[ 1, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Expr'", "Expr'", "Assign'", "Condition", "Condition", "Assign'", "Assign'", "Assign'", "Return'", "Expr'", "Assign'", "Assign'", "Expr'" ]
[ "@VAR_0.route('/level-4')...\n", "if int(current_user.progress) >= 4:\n", "return render_template('ui.html', VAR_6='4', VAR_7='index', level_progress=\n current_user.level4_progress, max_level_progress=3)\n", "return redirect(VAR_4)\n" ]
[ "@app.route('/level-4')...\n", "if int(current_user.progress) >= 4:\n", "return render_template('ui.html', level='4', page='index', level_progress=\n current_user.level4_progress, max_level_progress=3)\n", "return redirect(url_prefix)\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Condition", "Return'", "Return'" ]
[ "def FUNC_2(self, VAR_2, VAR_3, VAR_4=False):...\n", "VAR_6 = os.path.join(self.namespace, VAR_2)\n", "self.etcd.write(VAR_6, VAR_3, prevExist=replace)\n", "FUNC_0('Error storing key %s: [%r]' % (VAR_2, repr(err)))\n" ]
[ "def set(self, key, value, replace=False):...\n", "path = os.path.join(self.namespace, key)\n", "self.etcd.write(path, value, prevExist=replace)\n", "log_error('Error storing key %s: [%r]' % (key, repr(err)))\n" ]
[ 0, 1, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_2(VAR_1):...\n", "VAR_7 = datetime.now()\n", "VAR_8 = sqlite3.connect('database.db')\n", "VAR_9 = VAR_8.cursor()\n", "VAR_10 = FUNC_4()\n", "for VAR_0 in reversed(list(VAR_1)):\n", "VAR_13 = [[-1, ''], [-2, ''], [-3, ''], [-4, '']]\n", "VAR_8.commit()\n", "for topLevelComment in VAR_0.comments:\n", "VAR_8.close()\n", "VAR_13.sort(key=operator.itemgetter(0), reverse=True)\n", "if topLevelComment.author.name == VAR_0.author.name:\n", "if 'Previous win:' not in topLevelComment.body and 'for winning' not in topLevelComment.body and 'for tying' not in topLevelComment.body and '|' not in topLevelComment.body and topLevelComment is not None and topLevelComment.author is not None:\n", "for i in range(0, 3):\n", "VAR_15 = False\n", "VAR_16 = max([int(VAR_16.replace(',', '')) for VAR_16 in re.findall(\n '(?<!round )(?<!~~)(?<!\\\\w)\\\\d+\\\\,?\\\\d+', topLevelComment.body)])\n", "VAR_16 = -1\n", "if 0 <= VAR_16 <= 32395:\n", "while VAR_13[i][0] == VAR_13[i + 1][0]:\n", "VAR_14 = str(VAR_0.id), FUNC_1(VAR_0), str(VAR_13[0][1]), str(VAR_13[1][1]\n ), str(VAR_13[2][1]), FUNC_0(VAR_0)\n", "if '!trackthisseries' in topLevelComment.body.lower():\n", "VAR_13.append([int(VAR_16), topLevelComment.author.name])\n", "VAR_13[i][1] += '|' + VAR_13[i + 1][1]\n", "VAR_9.execute(\n 'INSERT OR REPLACE INTO ChallengeRankings VALUES (?, ?, ?, ?, ?, ?)',\n VAR_14)\n", "print('Found track request: ' + str(VAR_0.id))\n", "if '!stoptracking' in topLevelComment.body.lower():\n", "VAR_9.execute(\"INSERT OR REPLACE INTO SeriesTracking VALUES ('\" + FUNC_1(\n VAR_0) + \"', '\" + FUNC_0(VAR_0) + \"')\")\n", "print('Found stop tracking request: ' + str(VAR_0.id))\n", "for reply in topLevelComment.replies:\n", "VAR_9.execute(\"DELETE FROM SeriesTracking WHERE SeriesTitle = '\" + FUNC_1(\n VAR_0) + \"'\")\n", "if reply.author.name == VAR_10:\n", "if not VAR_15:\n", "for reply in topLevelComment.replies:\n", "VAR_15 = True\n", "FUNC_3(topLevelComment, True)\n", "if reply.author.name == VAR_10:\n", "if not VAR_15:\n", "VAR_15 = True\n", "FUNC_3(topLevelComment, False)\n" ]
[ "def addToDatabase(submissionList):...\n", "startTime = datetime.now()\n", "database = sqlite3.connect('database.db')\n", "cursor = database.cursor()\n", "botUsername = getBotUsername()\n", "for submission in reversed(list(submissionList)):\n", "scoresInChallenge = [[-1, ''], [-2, ''], [-3, ''], [-4, '']]\n", "database.commit()\n", "for topLevelComment in submission.comments:\n", "database.close()\n", "scoresInChallenge.sort(key=operator.itemgetter(0), reverse=True)\n", "if topLevelComment.author.name == submission.author.name:\n", "if 'Previous win:' not in topLevelComment.body and 'for winning' not in topLevelComment.body and 'for tying' not in topLevelComment.body and '|' not in topLevelComment.body and topLevelComment is not None and topLevelComment.author is not None:\n", "for i in range(0, 3):\n", "alreadyReplied = False\n", "number = max([int(number.replace(',', '')) for number in re.findall(\n '(?<!round )(?<!~~)(?<!\\\\w)\\\\d+\\\\,?\\\\d+', topLevelComment.body)])\n", "number = -1\n", "if 0 <= number <= 32395:\n", "while scoresInChallenge[i][0] == scoresInChallenge[i + 1][0]:\n", "record = str(submission.id), getTitle(submission), str(scoresInChallenge[0][1]\n ), str(scoresInChallenge[1][1]), str(scoresInChallenge[2][1]), getDate(\n submission)\n", "if '!trackthisseries' in topLevelComment.body.lower():\n", "scoresInChallenge.append([int(number), topLevelComment.author.name])\n", "scoresInChallenge[i][1] += '|' + scoresInChallenge[i + 1][1]\n", "cursor.execute(\n 'INSERT OR REPLACE INTO ChallengeRankings VALUES (?, ?, ?, ?, ?, ?)',\n record)\n", "print('Found track request: ' + str(submission.id))\n", "if '!stoptracking' in topLevelComment.body.lower():\n", "cursor.execute(\"INSERT OR REPLACE INTO SeriesTracking VALUES ('\" + getTitle\n (submission) + \"', '\" + getDate(submission) + \"')\")\n", "print('Found stop tracking request: ' + str(submission.id))\n", "for reply in topLevelComment.replies:\n", "cursor.execute(\"DELETE FROM SeriesTracking WHERE SeriesTitle = '\" +\n getTitle(submission) + \"'\")\n", "if reply.author.name == botUsername:\n", "if not alreadyReplied:\n", "for reply in topLevelComment.replies:\n", "alreadyReplied = True\n", "replyToTrackRequest(topLevelComment, True)\n", "if reply.author.name == botUsername:\n", "if not alreadyReplied:\n", "alreadyReplied = True\n", "replyToTrackRequest(topLevelComment, False)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 4, 0, 0, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Expr'", "For", "Expr'", "Expr'", "Condition", "For", "For", "Assign'", "Assign'", "Assign'", "Condition", "Condition", "Assign'", "Condition", "Expr'", "AugAssign'", "Expr'", "Expr'", "Condition", "Expr'", "Expr'", "For", "Expr'", "Condition", "Condition", "For", "Assign'", "Expr'", "Condition", "Condition", "Assign'", "Expr'" ]
[ "def FUNC_23(VAR_11):...\n", "return '%s> ' % self.configuration.eqlx_group_name\n" ]
[ "def _fake_recv(ignore_arg):...\n", "return '%s> ' % self.configuration.eqlx_group_name\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "import json\n", "import os\n", "import time\n", "import urllib\n", "import psycopg2\n", "from websocket import create_connection\n", "import config\n", "VAR_0 = create_connection(config.WEBSOCKET_URL)\n", "VAR_1 = psycopg2.connect(**config.POSTGRES)\n", "VAR_2 = VAR_1.cursor()\n", "VAR_3 = 'TRUNCATE holders'\n", "VAR_2.execute(VAR_3)\n", "VAR_3 = 'ALTER SEQUENCE holders_hid_seq RESTART WITH 1'\n", "VAR_2.execute(VAR_3)\n", "VAR_1.commit()\n", "VAR_0.send('{\"id\":1, \"method\":\"call\", \"params\":[0,\"get_account_count\",[]]}')\n", "VAR_4 = VAR_0.recv()\n", "VAR_5 = json.loads(VAR_4)\n", "VAR_6 = int(VAR_5['result'])\n", "for ac in range(0, VAR_6):\n", "VAR_0.send('{\"id\":1, \"method\":\"call\", \"params\":[0,\"get_objects\",[[\"1.2.' +\n str(ac) + '\"]]]}')\n", "VAR_1.close()\n", "VAR_4 = VAR_0.recv()\n", "VAR_5 = json.loads(VAR_4)\n", "VAR_9 = VAR_5['result'][0]['id']\n", "VAR_0.send(\n '{\"id\":1, \"method\":\"call\", \"params\":[0,\"get_account_balances\",[\"' +\n VAR_9 + '\", [\"1.3.0\"]]]}')\n", "VAR_10 = VAR_5['result'][0]['name']\n", "VAR_7 = VAR_0.recv()\n", "VAR_8 = json.loads(VAR_7)\n", "if VAR_8['result'][0]['amount'] == 0:\n", "VAR_11 = VAR_8['result'][0]['amount']\n", "VAR_0.send('{\"id\":1, \"method\":\"call\", \"params\":[0,\"get_objects\",[[\"' +\n VAR_5['result'][0]['statistics'] + '\"]]]}')\n", "VAR_4 = VAR_0.recv()\n", "VAR_12 = json.loads(VAR_4)\n", "VAR_14 = VAR_12['result'][0]['total_core_in_orders']\n", "VAR_14 = 0\n", "VAR_11 = int(VAR_11) + int(VAR_14)\n", "VAR_13 = VAR_5['result'][0]['options']['voting_account']\n", "VAR_3 = (\n \"INSERT INTO holders (account_id, account_name, amount, voting_as) VALUES('\"\n + VAR_9 + \"', '\" + VAR_10 + \"','\" + str(VAR_11) + \"', '\" + VAR_13 + \"')\")\n", "VAR_2.execute(VAR_3)\n", "VAR_1.commit()\n" ]
[ "import json\n", "import os\n", "import time\n", "import urllib\n", "import psycopg2\n", "from websocket import create_connection\n", "import config\n", "ws = create_connection(config.WEBSOCKET_URL)\n", "con = psycopg2.connect(**config.POSTGRES)\n", "cur = con.cursor()\n", "query = 'TRUNCATE holders'\n", "cur.execute(query)\n", "query = 'ALTER SEQUENCE holders_hid_seq RESTART WITH 1'\n", "cur.execute(query)\n", "con.commit()\n", "ws.send('{\"id\":1, \"method\":\"call\", \"params\":[0,\"get_account_count\",[]]}')\n", "result = ws.recv()\n", "j = json.loads(result)\n", "account_count = int(j['result'])\n", "for ac in range(0, account_count):\n", "ws.send('{\"id\":1, \"method\":\"call\", \"params\":[0,\"get_objects\",[[\"1.2.' + str\n (ac) + '\"]]]}')\n", "con.close()\n", "result = ws.recv()\n", "j = json.loads(result)\n", "account_id = j['result'][0]['id']\n", "ws.send('{\"id\":1, \"method\":\"call\", \"params\":[0,\"get_account_balances\",[\"' +\n account_id + '\", [\"1.3.0\"]]]}')\n", "account_name = j['result'][0]['name']\n", "result3 = ws.recv()\n", "jb = json.loads(result3)\n", "if jb['result'][0]['amount'] == 0:\n", "amount = jb['result'][0]['amount']\n", "ws.send('{\"id\":1, \"method\":\"call\", \"params\":[0,\"get_objects\",[[\"' + j[\n 'result'][0]['statistics'] + '\"]]]}')\n", "result = ws.recv()\n", "js = json.loads(result)\n", "total_core_in_orders = js['result'][0]['total_core_in_orders']\n", "total_core_in_orders = 0\n", "amount = int(amount) + int(total_core_in_orders)\n", "voting_account = j['result'][0]['options']['voting_account']\n", "query = (\n \"INSERT INTO holders (account_id, account_name, amount, voting_as) VALUES('\"\n + account_id + \"', '\" + account_name + \"','\" + str(amount) + \"', '\" +\n voting_account + \"')\")\n", "cur.execute(query)\n", "con.commit()\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 4, 0, 0 ]
[ "Import'", "Import'", "Import'", "Import'", "Import'", "ImportFrom'", "Import'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Assign'", "Assign'", "Assign'", "For", "Expr'", "Expr'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_28(self, *VAR_76, **VAR_52):...\n", "def FUNC_39(VAR_101):...\n", "VAR_101.resources = VAR_76, VAR_52\n", "return VAR_101\n" ]
[ "def resources(self, *args, **resources):...\n", "def decorate(ruleinfo):...\n", "ruleinfo.resources = args, resources\n", "return ruleinfo\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "FunctionDef'", "Assign'", "Return'" ]
[ "def FUNC_11(self, VAR_9, VAR_10=None):...\n", "if isinstance(VAR_9, str) or callable(VAR_9):\n", "self.params.append(VAR_9)\n", "VAR_48 = len(self.params)\n", "if VAR_10:\n", "for VAR_51 in VAR_9:\n", "self.params.add_name(VAR_10)\n", "self._set_params_item(VAR_51)\n", "if VAR_10:\n", "self.params.set_name(VAR_10, VAR_48, end=len(self.params))\n" ]
[ "def _set_params_item(self, item, name=None):...\n", "if isinstance(item, str) or callable(item):\n", "self.params.append(item)\n", "start = len(self.params)\n", "if name:\n", "for i in item:\n", "self.params.add_name(name)\n", "self._set_params_item(i)\n", "if name:\n", "self.params.set_name(name, start, end=len(self.params))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Expr'", "Assign'", "Condition", "For", "Expr'", "Expr'", "Condition", "Expr'" ]
[ "def FUNC_2(self, *VAR_2, **VAR_1):...\n", "VAR_0 = super(CLASS_0, self).get(*VAR_2, **kwargs)\n", "self._store_xsrf_token(VAR_0)\n", "return VAR_0\n" ]
[ "def get(self, *args, **kwargs):...\n", "response = super(Session, self).get(*args, **kwargs)\n", "self._store_xsrf_token(response)\n", "return response\n" ]
[ 0, 0, 5, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Return'" ]
[ "def FUNC_2(self):...\n", "\"\"\"docstring\"\"\"\n" ]
[ "def stop_app(self):...\n", "\"\"\"docstring\"\"\"\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Docstring" ]
[ "def FUNC_5():...\n", "VAR_34 = FUNC_3(VAR_18, VAR_19)\n", "return CLASS_16, VAR_34\n" ]
[ "def two():...\n", "FunctionalUseFormSet = make_formset(parent, child)\n", "return ExtractedTextForm, FunctionalUseFormSet\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Return'" ]
[ "def FUNC_5(VAR_7):...\n", "\"\"\"docstring\"\"\"\n", "return inspect.isfunction(VAR_7) or inspect.ismethod(VAR_7) or FUNC_4(VAR_7)\n" ]
[ "def is_function_or_method(obj):...\n", "\"\"\"docstring\"\"\"\n", "return inspect.isfunction(obj) or inspect.ismethod(obj) or is_cython(obj)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_23(VAR_31, VAR_32, VAR_3, VAR_33, VAR_34=apply_wildcards, VAR_35=None...\n", "for VAR_10, VAR_9 in VAR_32.allitems():\n", "VAR_48 = len(VAR_31)\n", "VAR_49 = True\n", "if callable(VAR_9):\n", "if not_iterable(VAR_9):\n", "VAR_9 = VAR_9(VAR_33)\n", "if not_iterable(VAR_9):\n", "VAR_9 = [VAR_9]\n", "for item_ in VAR_9:\n", "VAR_9 = [VAR_9]\n", "for item_ in VAR_9:\n", "VAR_49 = False\n", "VAR_53 = VAR_34(item_, VAR_3)\n", "if VAR_10:\n", "VAR_49 = False\n", "if not isinstance(item_, str):\n", "VAR_31.append(VAR_53)\n", "VAR_31.set_name(VAR_10, VAR_48, end=len(newitems) if is_iterable else None)\n", "VAR_53 = VAR_34(item_, VAR_3)\n", "if VAR_35 is not None:\n", "VAR_31.append(VAR_53)\n", "VAR_35[VAR_53] = item_\n", "if VAR_35 is not None:\n", "VAR_35[VAR_53] = item_\n" ]
[ "def _apply_wildcards(newitems, olditems, wildcards, wildcards_obj,...\n", "for name, item in olditems.allitems():\n", "start = len(newitems)\n", "is_iterable = True\n", "if callable(item):\n", "if not_iterable(item):\n", "item = item(wildcards_obj)\n", "if not_iterable(item):\n", "item = [item]\n", "for item_ in item:\n", "item = [item]\n", "for item_ in item:\n", "is_iterable = False\n", "concrete = concretize(item_, wildcards)\n", "if name:\n", "is_iterable = False\n", "if not isinstance(item_, str):\n", "newitems.append(concrete)\n", "newitems.set_name(name, start, end=len(newitems) if is_iterable else None)\n", "concrete = concretize(item_, wildcards)\n", "if ruleio is not None:\n", "newitems.append(concrete)\n", "ruleio[concrete] = item_\n", "if ruleio is not None:\n", "ruleio[concrete] = item_\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "For", "Assign'", "Assign'", "Condition", "Condition", "Assign'", "Condition", "Assign'", "For", "Assign'", "For", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Expr'", "Expr'", "Assign'", "Condition", "Expr'", "Assign'", "Condition", "Assign'" ]
[ "def __repr__(self):...\n", "return 'Distribution({!r}, minimum_version={!r}, maximum_version={!r} jdk={!r})'.format(\n self._bin_path, self._minimum_version, self._maximum_version, self._jdk)\n" ]
[ "def __repr__(self):...\n", "return 'Distribution({!r}, minimum_version={!r}, maximum_version={!r} jdk={!r})'.format(\n self._bin_path, self._minimum_version, self._maximum_version, self._jdk)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_1(self, VAR_21, VAR_22):...\n", "\"\"\"docstring\"\"\"\n", "VAR_55 = VAR_21.rstrip('/').split('/')\n", "if VAR_21[0] == '/':\n", "VAR_22 = []\n", "VAR_22 = [x for x in VAR_22.split('/') if len(x) and x is not None]\n", "while 1:\n", "if not len(VAR_55):\n", "VAR_66 = VAR_55.pop(0)\n", "return '/%s' % ('/'.join(VAR_22),)\n", "if VAR_66 == '..':\n", "if len(VAR_22):\n", "if VAR_66 in ('.', ''):\n", "VAR_22.pop()\n", "VAR_22.append(VAR_66)\n" ]
[ "def resolve_path(self, path, cwd):...\n", "\"\"\"docstring\"\"\"\n", "pieces = path.rstrip('/').split('/')\n", "if path[0] == '/':\n", "cwd = []\n", "cwd = [x for x in cwd.split('/') if len(x) and x is not None]\n", "while 1:\n", "if not len(pieces):\n", "piece = pieces.pop(0)\n", "return '/%s' % ('/'.join(cwd),)\n", "if piece == '..':\n", "if len(cwd):\n", "if piece in ('.', ''):\n", "cwd.pop()\n", "cwd.append(piece)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Condition", "Assign'", "Return'", "Condition", "Condition", "Condition", "Expr'", "Expr'" ]
[ "def FUNC_0(self, VAR_2, VAR_3, VAR_4, VAR_5):...\n", "\"\"\"docstring\"\"\"\n", "VAR_6 = self.pool.get('res.users').browse(VAR_2, VAR_3, VAR_3).company_id.id\n", "VAR_7 = []\n", "VAR_8 = self.pool.get('ir.model.data')\n", "VAR_9 = self.pool.get('account.bank.statement')\n", "VAR_10 = self.pool.get('account.journal')\n", "VAR_2.execute(\n 'select DISTINCT journal_id from pos_journal_users where user_id=%d order by journal_id'\n % VAR_3)\n", "VAR_11 = map(lambda x1: x1[0], VAR_2.fetchall())\n", "VAR_2.execute('string' % ','.join(map(lambda x: \"'\" + str(x) + \"'\", VAR_11)))\n", "VAR_12 = map(lambda x1: x1[0], VAR_2.fetchall())\n", "for journal in VAR_10.browse(VAR_2, VAR_3, VAR_12):\n", "VAR_4 = VAR_9.search(VAR_2, VAR_3, [('state', '!=', 'confirm'), ('user_id',\n '=', VAR_3), ('journal_id', '=', journal.id)])\n", "VAR_13 = self.pool.get('ir.model.data')\n", "if not VAR_4:\n", "VAR_14 = VAR_13._get_id(VAR_2, VAR_3, 'account', 'view_bank_statement_tree')\n", "VAR_7.append(VAR_4[0])\n", "VAR_15 = VAR_13._get_id(VAR_2, VAR_3, 'account', 'view_bank_statement_form2')\n", "if not journal.check_dtls:\n", "if VAR_14:\n", "VAR_9.button_confirm_cash(VAR_2, VAR_3, VAR_4, VAR_5)\n", "VAR_14 = VAR_13.browse(VAR_2, VAR_3, VAR_14, VAR_5=context).res_id\n", "if VAR_15:\n", "VAR_15 = VAR_13.browse(VAR_2, VAR_3, VAR_15, VAR_5=context).res_id\n", "return {'domain': \"[('id','in',\" + str(VAR_7) + ')]', 'name':\n 'Close Statements', 'view_type': 'form', 'view_mode': 'tree,form',\n 'res_model': 'account.bank.statement', 'views': [(VAR_14, 'tree'), (\n VAR_15, 'form')], 'type': 'ir.actions.act_window'}\n" ]
[ "def close_statement(self, cr, uid, ids, context):...\n", "\"\"\"docstring\"\"\"\n", "company_id = self.pool.get('res.users').browse(cr, uid, uid).company_id.id\n", "list_statement = []\n", "mod_obj = self.pool.get('ir.model.data')\n", "statement_obj = self.pool.get('account.bank.statement')\n", "journal_obj = self.pool.get('account.journal')\n", "cr.execute(\n 'select DISTINCT journal_id from pos_journal_users where user_id=%d order by journal_id'\n % uid)\n", "j_ids = map(lambda x1: x1[0], cr.fetchall())\n", "cr.execute(\n \"\"\" select id from account_journal\n where auto_cash='True' and type='cash'\n and id in (%s)\"\"\"\n % ','.join(map(lambda x: \"'\" + str(x) + \"'\", j_ids)))\n", "journal_ids = map(lambda x1: x1[0], cr.fetchall())\n", "for journal in journal_obj.browse(cr, uid, journal_ids):\n", "ids = statement_obj.search(cr, uid, [('state', '!=', 'confirm'), ('user_id',\n '=', uid), ('journal_id', '=', journal.id)])\n", "data_obj = self.pool.get('ir.model.data')\n", "if not ids:\n", "id2 = data_obj._get_id(cr, uid, 'account', 'view_bank_statement_tree')\n", "list_statement.append(ids[0])\n", "id3 = data_obj._get_id(cr, uid, 'account', 'view_bank_statement_form2')\n", "if not journal.check_dtls:\n", "if id2:\n", "statement_obj.button_confirm_cash(cr, uid, ids, context)\n", "id2 = data_obj.browse(cr, uid, id2, context=context).res_id\n", "if id3:\n", "id3 = data_obj.browse(cr, uid, id3, context=context).res_id\n", "return {'domain': \"[('id','in',\" + str(list_statement) + ')]', 'name':\n 'Close Statements', 'view_type': 'form', 'view_mode': 'tree,form',\n 'res_model': 'account.bank.statement', 'views': [(id2, 'tree'), (id3,\n 'form')], 'type': 'ir.actions.act_window'}\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 4, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "For", "Assign'", "Assign'", "Condition", "Assign'", "Expr'", "Assign'", "Condition", "Condition", "Expr'", "Assign'", "Condition", "Assign'", "Return'" ]
[ "@property...\n", "if self.is_ordered():\n", "return self.current_columns.index(self.column_id) + 1\n" ]
[ "@property...\n", "if self.is_ordered():\n", "return self.current_columns.index(self.column_id) + 1\n" ]
[ 0, 0, 0 ]
[ "Condition", "Condition", "Return'" ]
[ "def FUNC_26():...\n", "frappe.throw(_('Cannot use sub-query or function in fields'), frappe.DataError)\n" ]
[ "def _raise_exception():...\n", "frappe.throw(_('Cannot use sub-query or function in fields'), frappe.DataError)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def __repr__(self):...\n", "return self.get_name()\n" ]
[ "def __repr__(self):...\n", "return self.get_name()\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "@VAR_0.route('/star/<hip>/elements')...\n", "" ]
[ "@app.route('/star/<hip>/elements')...\n", "" ]
[ 0, 0 ]
[ "Condition", "Condition" ]
[ "def FUNC_9(self, VAR_2, VAR_3, VAR_4, *VAR_5):...\n", "self.write_data({'type': 'torrent_finished', 'event': {'infohash': hexlify(\n VAR_4), 'name': VAR_5[0]}})\n" ]
[ "def on_torrent_finished(self, subject, changetype, objectID, *args):...\n", "self.write_data({'type': 'torrent_finished', 'event': {'infohash': hexlify(\n objectID), 'name': args[0]}})\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "@classmethod...\n", "super(CLASS_7, VAR_0).setUpTestData()\n", "VAR_0.get_info_url = reverse('ajax-info')\n", "VAR_0.group_nitrate = EnvGroupFactory(name='nitrate')\n", "VAR_0.group_new = EnvGroupFactory(name='NewGroup')\n", "VAR_0.property_os = EnvPropertyFactory(name='os')\n", "VAR_0.property_python = EnvPropertyFactory(name='python')\n", "VAR_0.property_django = EnvPropertyFactory(name='django')\n", "EnvGroupPropertyMapFactory(VAR_8=cls.group_nitrate, property=cls.property_os)\n", "EnvGroupPropertyMapFactory(VAR_8=cls.group_nitrate, property=cls.\n property_python)\n", "EnvGroupPropertyMapFactory(VAR_8=cls.group_new, property=cls.property_django)\n" ]
[ "@classmethod...\n", "super(TestGetObjectInfo, cls).setUpTestData()\n", "cls.get_info_url = reverse('ajax-info')\n", "cls.group_nitrate = EnvGroupFactory(name='nitrate')\n", "cls.group_new = EnvGroupFactory(name='NewGroup')\n", "cls.property_os = EnvPropertyFactory(name='os')\n", "cls.property_python = EnvPropertyFactory(name='python')\n", "cls.property_django = EnvPropertyFactory(name='django')\n", "EnvGroupPropertyMapFactory(group=cls.group_nitrate, property=cls.property_os)\n", "EnvGroupPropertyMapFactory(group=cls.group_nitrate, property=cls.\n property_python)\n", "EnvGroupPropertyMapFactory(group=cls.group_new, property=cls.property_django)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_2(self):...\n", "VAR_7 = self.window.menuBar()\n", "VAR_8 = VAR_7.addMenu('Console')\n", "self.clear_log_area_action = VAR_8.addAction('Clear Log')\n", "self.open_idle_action = VAR_8.addAction('Open Python Shell...')\n", "self.restart_action = VAR_8.addAction('Restart')\n", "VAR_9 = VAR_7.addMenu('Set')\n", "self.set_wavelength_action = VAR_9.addAction('Wavelength')\n", "self.set_bifi_approx_wavelength_action = VAR_9.addAction(\n 'BiFi Approx. Wavelength')\n", "self.set_bifi_motor_pos_action = VAR_9.addAction('BiFi Motor Position')\n", "self.set_thin_eta_motor_pos_action = VAR_9.addAction(\n 'Thin Etalon Motor Position')\n", "VAR_10 = VAR_7.addMenu('Scan')\n", "self.bifi_scan_action = VAR_10.addAction('Birefringent Filter')\n", "self.thin_eta_scan_action = VAR_10.addAction('Thin Etalon')\n", "VAR_11 = VAR_7.addMenu('Lock')\n", "self.lock_all_action = VAR_11.addAction('Lock All')\n", "self.lock_all_action.setCheckable(True)\n", "self.lock_slow_piezo_action = VAR_11.addAction('Lock Slow Piezo')\n", "self.lock_slow_piezo_action.setCheckable(True)\n", "self.lock_thin_etalon_action = VAR_11.addAction('Lock Thin Etalon')\n", "self.lock_thin_etalon_action.setCheckable(True)\n", "self.lock_piezo_etalon_action = VAR_11.addAction('Lock Piezo Etalon')\n", "self.lock_piezo_etalon_action.setCheckable(True)\n", "self.lock_fast_piezo_action = VAR_11.addAction('Lock Fast Piezo')\n", "self.lock_fast_piezo_action.setCheckable(True)\n", "VAR_12 = VAR_7.addMenu('Tools')\n", "self.lock_actions = [self.lock_slow_piezo_action, self.\n lock_thin_etalon_action, self.lock_piezo_etalon_action, self.\n lock_fast_piezo_action]\n" ]
[ "def setup_menus(self):...\n", "menu_bar = self.window.menuBar()\n", "console_menu = menu_bar.addMenu('Console')\n", "self.clear_log_area_action = console_menu.addAction('Clear Log')\n", "self.open_idle_action = console_menu.addAction('Open Python Shell...')\n", "self.restart_action = console_menu.addAction('Restart')\n", "set_menu = menu_bar.addMenu('Set')\n", "self.set_wavelength_action = set_menu.addAction('Wavelength')\n", "self.set_bifi_approx_wavelength_action = set_menu.addAction(\n 'BiFi Approx. Wavelength')\n", "self.set_bifi_motor_pos_action = set_menu.addAction('BiFi Motor Position')\n", "self.set_thin_eta_motor_pos_action = set_menu.addAction(\n 'Thin Etalon Motor Position')\n", "scan_menu = menu_bar.addMenu('Scan')\n", "self.bifi_scan_action = scan_menu.addAction('Birefringent Filter')\n", "self.thin_eta_scan_action = scan_menu.addAction('Thin Etalon')\n", "lock_menu = menu_bar.addMenu('Lock')\n", "self.lock_all_action = lock_menu.addAction('Lock All')\n", "self.lock_all_action.setCheckable(True)\n", "self.lock_slow_piezo_action = lock_menu.addAction('Lock Slow Piezo')\n", "self.lock_slow_piezo_action.setCheckable(True)\n", "self.lock_thin_etalon_action = lock_menu.addAction('Lock Thin Etalon')\n", "self.lock_thin_etalon_action.setCheckable(True)\n", "self.lock_piezo_etalon_action = lock_menu.addAction('Lock Piezo Etalon')\n", "self.lock_piezo_etalon_action.setCheckable(True)\n", "self.lock_fast_piezo_action = lock_menu.addAction('Lock Fast Piezo')\n", "self.lock_fast_piezo_action.setCheckable(True)\n", "tools_menu = menu_bar.addMenu('Tools')\n", "self.lock_actions = [self.lock_slow_piezo_action, self.\n lock_thin_etalon_action, self.lock_piezo_etalon_action, self.\n lock_fast_piezo_action]\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Assign'" ]
[ "def FUNC_0(self):...\n", "self.mail_service = mock()\n", "self.services_factory = mock()\n", "self.services_factory.mode = UserAgentMode(is_single_user=True)\n", "self.services = mock()\n", "self.services.mail_service = self.mail_service\n", "self.services_factory._services_by_user = {'someuserid': self.mail_service}\n", "when(self.services_factory).services(ANY()).thenReturn(self.services)\n", "self.mail_service.account_email = self.MAIL_ADDRESS\n", "VAR_1 = RootResource(self.services_factory)\n", "VAR_1._html_template = (\n '<html><head><title>$account_email</title></head></html>')\n", "VAR_1._mode = VAR_1\n", "self.web = DummySite(VAR_1)\n" ]
[ "def setUp(self):...\n", "self.mail_service = mock()\n", "self.services_factory = mock()\n", "self.services_factory.mode = UserAgentMode(is_single_user=True)\n", "self.services = mock()\n", "self.services.mail_service = self.mail_service\n", "self.services_factory._services_by_user = {'someuserid': self.mail_service}\n", "when(self.services_factory).services(ANY()).thenReturn(self.services)\n", "self.mail_service.account_email = self.MAIL_ADDRESS\n", "root_resource = RootResource(self.services_factory)\n", "root_resource._html_template = (\n '<html><head><title>$account_email</title></head></html>')\n", "root_resource._mode = root_resource\n", "self.web = DummySite(root_resource)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'" ]
[ "def FUNC_9(VAR_1):...\n", "return {'name': str(VAR_1), 'link': VAR_1.get_absolute_url()}\n" ]
[ "def course_entry(instance):...\n", "return {'name': str(instance), 'link': instance.get_absolute_url()}\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "@commands.command()...\n", "\"\"\"docstring\"\"\"\n", "VAR_7 = discord.Embed(VAR_2='Dsp1', VAR_3=discord.Color.green())\n", "VAR_7.set_author(name='zoogie', url='https://github.com/zoogie', icon_url=\n 'https://gbatemp.net/data/avatars/l/357/357147.jpg?1426471484')\n", "VAR_7.description = \"Dump 3DS's DSP component to SD for homebrew audio.\"\n", "VAR_7.set_thumbnail(url=\n 'https://raw.githubusercontent.com/Cruel/DspDump/master/icon.png')\n", "VAR_7.url = 'https://github.com/zoogie/DSP1/releases'\n", "await self.bot.say('', VAR_7=embed)\n" ]
[ "@commands.command()...\n", "\"\"\"docstring\"\"\"\n", "embed = discord.Embed(title='Dsp1', color=discord.Color.green())\n", "embed.set_author(name='zoogie', url='https://github.com/zoogie', icon_url=\n 'https://gbatemp.net/data/avatars/l/357/357147.jpg?1426471484')\n", "embed.description = \"Dump 3DS's DSP component to SD for homebrew audio.\"\n", "embed.set_thumbnail(url=\n 'https://raw.githubusercontent.com/Cruel/DspDump/master/icon.png')\n", "embed.url = 'https://github.com/zoogie/DSP1/releases'\n", "await self.bot.say('', embed=embed)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Expr'", "Assign'", "Expr'", "Assign'", "Expr'" ]
[ "def FUNC_0(VAR_0, VAR_1):...\n", "\"\"\"docstring\"\"\"\n", "VAR_18 = sql.connect('./resources/users.db')\n", "VAR_19 = False\n", "VAR_20 = \"\"\"SELECT * FROM users\n WHERE username = ? AND password = ?;\n \"\"\"\n", "VAR_21 = VAR_18.cursor()\n", "VAR_22 = VAR_0, VAR_1\n", "VAR_21.execute(VAR_20, VAR_22)\n", "VAR_23 = VAR_21.fetchone()\n", "if VAR_23 != None:\n", "VAR_19 = True\n", "VAR_19 = False\n", "VAR_5 = VAR_23[0]\n", "VAR_5 = None\n", "VAR_2 = VAR_23[2]\n", "VAR_2 = None\n", "VAR_3 = VAR_23[3]\n", "VAR_3 = None\n", "VAR_18.close()\n", "return VAR_19, VAR_5, VAR_2, VAR_3\n" ]
[ "def authenticate(username, password):...\n", "\"\"\"docstring\"\"\"\n", "con = sql.connect('./resources/users.db')\n", "authenticated = False\n", "cmd = \"\"\"SELECT * FROM users\n WHERE username = ? AND password = ?;\n \"\"\"\n", "cur = con.cursor()\n", "info = username, password\n", "cur.execute(cmd, info)\n", "user_info = cur.fetchone()\n", "if user_info != None:\n", "authenticated = True\n", "authenticated = False\n", "user_id = user_info[0]\n", "user_id = None\n", "firstname = user_info[2]\n", "firstname = None\n", "lastname = user_info[3]\n", "lastname = None\n", "con.close()\n", "return authenticated, user_id, firstname, lastname\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Return'" ]
[ "@hook.command(autohelp=False)...\n", "return FUNC_1(VAR_8, 'kick', VAR_4, VAR_5, VAR_1, VAR_6, VAR_7)\n" ]
[ "@hook.command(autohelp=False)...\n", "return process_vote(inp, 'kick', chan, mask, db, notice, conn)\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_5(self, VAR_1, VAR_2, VAR_8=None, VAR_9='form', VAR_4=None, VAR_14...\n", "VAR_29 = super(CLASS_0, self).fields_view_get(VAR_1, VAR_2, VAR_8, VAR_9,\n VAR_4, VAR_14=toolbar, VAR_15=submenu)\n", "if VAR_4 is None:\n", "VAR_4 = {}\n", "if 'location' in VAR_4 and VAR_4['location']:\n", "VAR_51 = self.pool.get('stock.location').browse(VAR_1, VAR_2, VAR_4['location']\n )\n", "return VAR_29\n", "VAR_52 = VAR_29.get('fields', {})\n", "if VAR_52:\n", "if VAR_51.usage == 'supplier':\n", "if VAR_52.get('virtual_available'):\n", "if VAR_51.usage == 'internal':\n", "VAR_29['fields']['virtual_available']['string'] = _('Future Receptions')\n", "if VAR_52.get('qty_available'):\n", "if VAR_52.get('virtual_available'):\n", "if VAR_51.usage == 'customer':\n", "VAR_29['fields']['qty_available']['string'] = _('Received Qty')\n", "VAR_29['fields']['virtual_available']['string'] = _('Future Stock')\n", "if VAR_52.get('virtual_available'):\n", "if VAR_51.usage == 'inventory':\n", "VAR_29['fields']['virtual_available']['string'] = _('Future Deliveries')\n", "if VAR_52.get('qty_available'):\n", "if VAR_52.get('virtual_available'):\n", "if VAR_51.usage == 'procurement':\n", "VAR_29['fields']['qty_available']['string'] = _('Delivered Qty')\n", "VAR_29['fields']['virtual_available']['string'] = _('Future P&L')\n", "if VAR_52.get('qty_available'):\n", "if VAR_52.get('virtual_available'):\n", "if VAR_51.usage == 'production':\n", "VAR_29['fields']['qty_available']['string'] = _('P&L Qty')\n", "VAR_29['fields']['virtual_available']['string'] = _('Future Qty')\n", "if VAR_52.get('qty_available'):\n", "if VAR_52.get('virtual_available'):\n", "VAR_29['fields']['qty_available']['string'] = _('Unplanned Qty')\n", "VAR_29['fields']['virtual_available']['string'] = _('Future Productions')\n", "if VAR_52.get('qty_available'):\n", "VAR_29['fields']['qty_available']['string'] = _('Produced Qty')\n" ]
[ "def fields_view_get(self, cr, uid, view_id=None, view_type='form', context=...\n", "res = super(product_product, self).fields_view_get(cr, uid, view_id,\n view_type, context, toolbar=toolbar, submenu=submenu)\n", "if context is None:\n", "context = {}\n", "if 'location' in context and context['location']:\n", "location_info = self.pool.get('stock.location').browse(cr, uid, context[\n 'location'])\n", "return res\n", "fields = res.get('fields', {})\n", "if fields:\n", "if location_info.usage == 'supplier':\n", "if fields.get('virtual_available'):\n", "if location_info.usage == 'internal':\n", "res['fields']['virtual_available']['string'] = _('Future Receptions')\n", "if fields.get('qty_available'):\n", "if fields.get('virtual_available'):\n", "if location_info.usage == 'customer':\n", "res['fields']['qty_available']['string'] = _('Received Qty')\n", "res['fields']['virtual_available']['string'] = _('Future Stock')\n", "if fields.get('virtual_available'):\n", "if location_info.usage == 'inventory':\n", "res['fields']['virtual_available']['string'] = _('Future Deliveries')\n", "if fields.get('qty_available'):\n", "if fields.get('virtual_available'):\n", "if location_info.usage == 'procurement':\n", "res['fields']['qty_available']['string'] = _('Delivered Qty')\n", "res['fields']['virtual_available']['string'] = _('Future P&L')\n", "if fields.get('qty_available'):\n", "if fields.get('virtual_available'):\n", "if location_info.usage == 'production':\n", "res['fields']['qty_available']['string'] = _('P&L Qty')\n", "res['fields']['virtual_available']['string'] = _('Future Qty')\n", "if fields.get('qty_available'):\n", "if fields.get('virtual_available'):\n", "res['fields']['qty_available']['string'] = _('Unplanned Qty')\n", "res['fields']['virtual_available']['string'] = _('Future Productions')\n", "if fields.get('qty_available'):\n", "res['fields']['qty_available']['string'] = _('Produced Qty')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "For", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Return'", "Assign'", "Condition", "Condition", "Condition", "Condition", "Assign'", "Condition", "Condition", "Condition", "Assign'", "Assign'", "Condition", "Condition", "Assign'", "Condition", "Condition", "Condition", "Assign'", "Assign'", "Condition", "Condition", "Condition", "Assign'", "Assign'", "Condition", "Condition", "Assign'", "Assign'", "Condition", "Assign'" ]
[ "def FUNC_17(self):...\n", "return self.get('__islocal')\n" ]
[ "def is_new(self):...\n", "return self.get('__islocal')\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "@Json...\n", "if VAR_2._chk_error(errors.NO_TITLE):\n", "VAR_2._chk_error(errors.TITLE_TOO_LONG)\n", "VAR_2._chk_errors((errors.NO_LOCATION, errors.NO_DESCRIPTION, errors.\n INVALID_DATE, errors.NO_DATE))\n", "VAR_2._focus('title')\n", "if VAR_2.error:\n", "return\n", "VAR_0 = Meetup(author_id=c.user._id, VAR_4=title, VAR_8=description, VAR_9=\n location, VAR_10=latitude, VAR_11=longitude, VAR_12=timestamp / 1000,\n VAR_13=tzoffset)\n", "g.rendercache.invalidate_key_group(Meetup.group_cache_key())\n", "VAR_0._commit()\n", "VAR_17 = Link._submit(FUNC_1(VAR_0), FUNC_0(VAR_0), c.user, Subreddit.\n _by_name('discussion'), VAR_14, [])\n", "VAR_17.meetup = VAR_0._id36\n", "VAR_17._commit()\n", "VAR_0.assoc_link = VAR_17._id\n", "VAR_0._commit()\n", "if g.write_query_queue:\n", "queries.new_link(VAR_17)\n", "VAR_2._redirect(url_for(action='show', id=meetup._id36))\n" ]
[ "@Json...\n", "if res._chk_error(errors.NO_TITLE):\n", "res._chk_error(errors.TITLE_TOO_LONG)\n", "res._chk_errors((errors.NO_LOCATION, errors.NO_DESCRIPTION, errors.\n INVALID_DATE, errors.NO_DATE))\n", "res._focus('title')\n", "if res.error:\n", "return\n", "meetup = Meetup(author_id=c.user._id, title=title, description=description,\n location=location, latitude=latitude, longitude=longitude, timestamp=\n timestamp / 1000, tzoffset=tzoffset)\n", "g.rendercache.invalidate_key_group(Meetup.group_cache_key())\n", "meetup._commit()\n", "l = Link._submit(meetup_article_title(meetup), meetup_article_text(meetup),\n c.user, Subreddit._by_name('discussion'), ip, [])\n", "l.meetup = meetup._id36\n", "l._commit()\n", "meetup.assoc_link = l._id\n", "meetup._commit()\n", "if g.write_query_queue:\n", "queries.new_link(l)\n", "res._redirect(url_for(action='show', id=meetup._id36))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Condition", "Expr'", "Expr'", "Expr'", "Condition", "Return'", "Assign'", "Expr'", "Expr'", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'", "Condition", "Expr'", "Expr'" ]
[ "def FUNC_0(VAR_9):...\n", "for x in VAR_4:\n", "if x in [y.id for y in VAR_9.message.author.roles]:\n", "return True\n", "return False\n" ]
[ "def whitelist_check(ctx):...\n", "for x in role_whitelist:\n", "if x in [y.id for y in ctx.message.author.roles]:\n", "return True\n", "return False\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "For", "For", "Return'", "Return'" ]
[ "def FUNC_37(self, VAR_29):...\n", "VAR_12.append('authorization_error')\n", "super(CLASS_7, self).authorization_error(VAR_29)\n" ]
[ "def authorization_error(self, err):...\n", "calls.append('authorization_error')\n", "super(Handler, self).authorization_error(err)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Expr'" ]
[ "def FUNC_4(self, VAR_14):...\n", "VAR_25 = self.env['res.partner'].search([('ref', '=', VAR_14)])\n", "if not VAR_25:\n", "return VAR_25\n" ]
[ "def get_vendor_id(self, vendor_ref):...\n", "vendor_id = self.env['res.partner'].search([('ref', '=', vendor_ref)])\n", "if not vendor_id:\n", "return vendor_id\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Return'" ]
[ "def FUNC_7(self):...\n", "return self.data\n" ]
[ "def to_json(self):...\n", "return self.data\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "@user_passes_test(user_is_staff)...\n", "\"\"\"docstring\"\"\"\n", "VAR_9 = push_testcases_to_testrail_celery.delay(VAR_2)\n", "VAR_4 = context_testcases()\n", "VAR_4['testrail'] = VAR_9\n", "VAR_4['link_id'] = VAR_2\n", "return render(VAR_0, 'testcases/testcases.html', VAR_4)\n" ]
[ "@user_passes_test(user_is_staff)...\n", "\"\"\"docstring\"\"\"\n", "testrail_contents = push_testcases_to_testrail_celery.delay(project_id)\n", "context = context_testcases()\n", "context['testrail'] = testrail_contents\n", "context['link_id'] = project_id\n", "return render(request, 'testcases/testcases.html', context)\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_2():...\n", "VAR_16 = InputMail()\n", "VAR_16.fdoc = CLASS_0({})\n", "VAR_16._chash = '123'\n", "VAR_16.as_dict = lambda : None\n", "return VAR_16\n" ]
[ "def input_mail():...\n", "mail = InputMail()\n", "mail.fdoc = TestDoc({})\n", "mail._chash = '123'\n", "mail.as_dict = lambda : None\n", "return mail\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_25(self):...\n", "\"\"\"docstring\"\"\"\n", "return self._normalized_jdk_paths\n" ]
[ "def all_jdk_paths(self):...\n", "\"\"\"docstring\"\"\"\n", "return self._normalized_jdk_paths\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def __init__(self):...\n", "CLASS_0.__init__(self)\n" ]
[ "def __init__(self):...\n", "BaseLevel.__init__(self)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_24(self, VAR_14):...\n", "" ]
[ "def is_pk(self, col_name):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "import unittest\n", "import mock\n", "from oslo_serialization import jsonutils\n", "from requests import exceptions as requests_exceptions\n", "import six.moves.urllib.parse as urlparse\n", "from vmware_nsxlib.tests.unit.v3 import mocks\n", "from vmware_nsxlib.tests.unit.v3 import nsxlib_testcase\n", "from vmware_nsxlib.v3 import client\n", "from vmware_nsxlib.v3 import client_cert\n", "from vmware_nsxlib.v3 import cluster\n", "from vmware_nsxlib.v3 import exceptions as nsxlib_exc\n", "def FUNC_0(*VAR_0, **VAR_1):...\n", "return\n" ]
[ "import unittest\n", "import mock\n", "from oslo_serialization import jsonutils\n", "from requests import exceptions as requests_exceptions\n", "import six.moves.urllib.parse as urlparse\n", "from vmware_nsxlib.tests.unit.v3 import mocks\n", "from vmware_nsxlib.tests.unit.v3 import nsxlib_testcase\n", "from vmware_nsxlib.v3 import client\n", "from vmware_nsxlib.v3 import client_cert\n", "from vmware_nsxlib.v3 import cluster\n", "from vmware_nsxlib.v3 import exceptions as nsxlib_exc\n", "def _validate_conn_up(*args, **kwargs):...\n", "return\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "Import'", "ImportFrom'", "ImportFrom'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "FunctionDef'", "Return'" ]
[ "def FUNC_2(self):...\n", "for VAR_25 in self.args:\n", "if self.args[VAR_25]:\n", "if self['sos_opt_line']:\n", "self[VAR_25] = self.args[VAR_25]\n", "self['sos_opt_line'] = pipes.quote(self['sos_opt_line'])\n" ]
[ "def parse_config(self):...\n", "for k in self.args:\n", "if self.args[k]:\n", "if self['sos_opt_line']:\n", "self[k] = self.args[k]\n", "self['sos_opt_line'] = pipes.quote(self['sos_opt_line'])\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "For", "Condition", "Condition", "Assign'", "Assign'" ]
[ "from base import Field\n", "VAR_0 = 'FLOAT'\n", "def __init__(self, *VAR_1, **VAR_2):...\n", "super(CLASS_0, self).__init__(*VAR_1, **kwargs)\n", "def FUNC_0(self, VAR_3):...\n", "VAR_5 = '%s float' % str(VAR_3)\n", "return VAR_5\n" ]
[ "from base import Field\n", "TYPE = 'FLOAT'\n", "def __init__(self, *args, **kwargs):...\n", "super(FloatField, self).__init__(*args, **kwargs)\n", "def create_field(self, name):...\n", "field_string = '%s float' % str(name)\n", "return field_string\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "ImportFrom'", "Assign'", "FunctionDef'", "Expr'", "FunctionDef'", "Assign'", "Return'" ]
[ "def FUNC_7(self):...\n", "self.controller.remove_frame(VAR_9)\n", "VAR_8 = self.controller.create_frame(VAR_9)\n" ]
[ "def new_search(self):...\n", "self.controller.remove_frame(SearchPage)\n", "new_frame = self.controller.create_frame(SearchPage)\n" ]
[ 0, 4, 4 ]
[ "FunctionDef'", "Expr'", "Assign'" ]
[ "def FUNC_16(self):...\n", "\"\"\"docstring\"\"\"\n", "logging.info('Using default log_out() method')\n", "self.clear_session()\n", "return self.redirect_to_goodbye()\n" ]
[ "def log_out(self):...\n", "\"\"\"docstring\"\"\"\n", "logging.info('Using default log_out() method')\n", "self.clear_session()\n", "return self.redirect_to_goodbye()\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'", "Expr'", "Return'" ]
[ "\"\"\"Unittest to exercise the code in url_helper.py.\"\"\"\n", "import logging\n", "import os\n", "import stat\n", "import StringIO\n", "import sys\n", "import tempfile\n", "import time\n", "import unittest\n", "import urllib\n", "import urllib2\n", "VAR_0 = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))\n", "sys.path.insert(0, VAR_0)\n", "import test_env\n", "test_env.setup_test_env()\n", "from depot_tools import auto_stub\n", "from third_party.mox import mox\n", "import url_helper\n", "def FUNC_0(self):...\n", "self._mox = mox.Mox()\n", "self.mock(logging, 'error', lambda *_: None)\n", "self.mock(logging, 'exception', lambda *_: None)\n", "self.mock(logging, 'info', lambda *_: None)\n", "self.mock(logging, 'warning', lambda *_: None)\n", "self._mox.StubOutWithMock(time, 'sleep')\n", "self._mox.StubOutWithMock(urllib2, 'urlopen')\n", "def FUNC_1(self):...\n", "self._mox.UnsetStubs()\n", "def FUNC_2(self):...\n", "self._mox.ReplayAll()\n", "self.assertEqual(url_helper.UrlOpen('url', max_tries=-1), None)\n", "self._mox.VerifyAll()\n", "def FUNC_3(self):...\n", "self._mox.ReplayAll()\n", "self.assertEqual(url_helper.UrlOpen('url', wait_duration=-1), None)\n", "self._mox.VerifyAll()\n", "def FUNC_4(self):...\n", "VAR_1 = 'http://my.url.com'\n", "VAR_2 = 'True'\n", "url_helper.urllib2.urlopen(mox.StrContains(VAR_1), timeout=mox.IgnoreArg()\n ).AndReturn(StringIO.StringIO(VAR_2))\n", "self._mox.ReplayAll()\n", "self.assertEqual(url_helper.UrlOpen(VAR_1, method='GET'), VAR_2)\n", "self._mox.VerifyAll()\n", "def FUNC_5(self):...\n", "VAR_1 = 'http://my.url.com'\n", "VAR_2 = 'True'\n", "url_helper.urllib2.urlopen(VAR_1, mox.IgnoreArg(), timeout=mox.IgnoreArg()\n ).AndReturn(StringIO.StringIO(VAR_2))\n", "self._mox.ReplayAll()\n", "self.assertEqual(url_helper.UrlOpen(VAR_1, method='POST'), VAR_2)\n", "self._mox.VerifyAll()\n", "def FUNC_6(self):...\n", "VAR_1 = 'http://my.url.com'\n", "VAR_2 = 'True'\n", "url_helper.urllib2.urlopen(mox.IsA(urllib2.Request), timeout=mox.IgnoreArg()\n ).AndReturn(StringIO.StringIO(VAR_2))\n", "self._mox.ReplayAll()\n", "self.assertEqual(url_helper.UrlOpen(VAR_1, method='POSTFORM'), VAR_2)\n", "self._mox.VerifyAll()\n", "def FUNC_7(self):...\n", "url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndRaise(urllib2.URLError('url'))\n", "time.sleep(mox.IgnoreArg())\n", "VAR_2 = 'True'\n", "url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndReturn(StringIO.StringIO(VAR_2))\n", "self._mox.ReplayAll()\n", "self.assertEqual(url_helper.UrlOpen('url', max_tries=2), VAR_2)\n", "self._mox.VerifyAll()\n", "def FUNC_8(self):...\n", "url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndRaise(urllib2.URLError('url'))\n", "self._mox.ReplayAll()\n", "self.assertIsNone(url_helper.UrlOpen('url', max_tries=1))\n", "self._mox.VerifyAll()\n", "def FUNC_9(self):...\n", "url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndRaise(urllib2.HTTPError('url', 400, 'error message',\n None, None))\n", "self._mox.ReplayAll()\n", "self.assertIsNone(url_helper.UrlOpen('url', max_tries=10))\n", "self._mox.VerifyAll()\n", "def FUNC_10(self):...\n", "VAR_2 = 'response'\n", "url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndRaise(urllib2.HTTPError('url', 500, 'error message',\n None, None))\n", "time.sleep(mox.IgnoreArg())\n", "url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndReturn(StringIO.StringIO(VAR_2))\n", "self._mox.ReplayAll()\n", "self.assertEqual(VAR_2, url_helper.UrlOpen('url', max_tries=10))\n", "self._mox.VerifyAll()\n", "def FUNC_11(self):...\n", "VAR_3 = 5\n", "for i in range(VAR_3):\n", "VAR_11 = urllib.urlencode({url_helper.swarm_constants.COUNT_KEY: i})\n", "self._mox.ReplayAll()\n", "url_helper.urllib2.urlopen(mox.IgnoreArg(), VAR_11, timeout=mox.IgnoreArg()\n ).AndRaise(urllib2.URLError('url'))\n", "self.assertEqual(url_helper.UrlOpen('url', max_tries=attempts), None)\n", "if i != VAR_3 - 1:\n", "self._mox.VerifyAll()\n", "time.sleep(mox.IgnoreArg())\n", "def FUNC_12(self):...\n", "VAR_4 = {url_helper.swarm_constants.COUNT_KEY: 1}\n", "self._mox.ReplayAll()\n", "self.assertEqual(url_helper.UrlOpen('url', VAR_4=data), None)\n", "self._mox.VerifyAll()\n", "def FUNC_13(self):...\n", "VAR_4 = {'r': u'not ascii £ һ'}\n", "VAR_1 = 'http://my.url.com'\n", "VAR_2 = 'True'\n", "url_helper.urllib2.urlopen(mox.StrContains(VAR_1), mox.IgnoreArg(), timeout\n =mox.IgnoreArg()).AndReturn(StringIO.StringIO(VAR_2))\n", "self._mox.ReplayAll()\n", "self.assertEqual(url_helper.UrlOpen(VAR_1, VAR_4=data), VAR_2)\n", "self._mox.VerifyAll()\n", "def FUNC_14(self):...\n", "VAR_5 = None\n", "VAR_5 = tempfile.NamedTemporaryFile(delete=False)\n", "if VAR_5:\n", "def FUNC_15(self):...\n", "VAR_5.close()\n", "os.remove(VAR_5.name)\n", "VAR_13 = 'fake_local_file.fake'\n", "if os.path.exists(VAR_13):\n", "def FUNC_16(self):...\n", "self._mox.StubOutWithMock(url_helper, 'UrlOpen')\n", "self._mox.StubOutWithMock(url_helper, 'UrlOpen')\n", "os.remove(VAR_13)\n", "VAR_6 = None\n", "VAR_12 = 'data'\n", "url_helper.UrlOpen(mox.IgnoreArg(), method='GET').AndReturn(None)\n", "VAR_6 = tempfile.NamedTemporaryFile(delete=False)\n", "if VAR_6:\n", "def FUNC_17(self):...\n", "url_helper.UrlOpen(mox.IgnoreArg(), method='GET').AndReturn(VAR_12)\n", "self._mox.ReplayAll()\n", "VAR_6.close()\n", "os.remove(VAR_6.name)\n", "VAR_7 = [('x', 'y'), (1, 2)]\n", "self._mox.ReplayAll()\n", "self.assertFalse(url_helper.DownloadFile(VAR_13, 'http://www.fakeurl.com'))\n", "os.chmod(VAR_6.name, stat.S_IREAD)\n", "VAR_8 = [('key', 'filename', 'file data')]\n", "self.assertTrue(url_helper.DownloadFile(VAR_5.name, 'http://www.fakeurl.com'))\n", "self._mox.VerifyAll()\n", "self._mox.StubOutWithMock(url_helper, 'UrlOpen')\n", "VAR_9, VAR_10 = url_helper.EncodeMultipartFormData()\n", "self.assertEqual(VAR_12, f.read())\n", "url_helper.UrlOpen(mox.IgnoreArg(), method='GET').AndReturn('data')\n", "self.assertTrue(VAR_9.startswith('multipart/form-data; boundary='))\n", "self._mox.VerifyAll()\n", "self._mox.ReplayAll()\n", "self.assertEqual('', VAR_10)\n", "self.assertFalse(url_helper.DownloadFile(VAR_6.name, 'http://www.fakeurl.com'))\n", "VAR_9, VAR_10 = url_helper.EncodeMultipartFormData(VAR_7=fields)\n", "self._mox.VerifyAll()\n", "self.assertTrue(VAR_9.startswith('multipart/form-data; boundary='))\n", "self.assertTrue('name=\"x\"\\r\\n\\r\\ny' in VAR_10, VAR_10)\n", "self.assertTrue('name=\"1\"\\r\\n\\r\\n2' in VAR_10, VAR_10)\n", "VAR_9, VAR_10 = url_helper.EncodeMultipartFormData(VAR_8=files)\n", "self.assertTrue(VAR_9.startswith('multipart/form-data; boundary='))\n", "self.assertTrue('name=\"key\"; filename=\"filename\"' in VAR_10, VAR_10)\n", "self.assertTrue('file data' in VAR_10, VAR_10)\n", "VAR_9, VAR_10 = url_helper.EncodeMultipartFormData(VAR_7=fields, VAR_8=files)\n", "self.assertTrue(VAR_9.startswith('multipart/form-data; boundary='))\n", "self.assertTrue('name=\"x\"\\r\\n\\r\\ny' in VAR_10, VAR_10)\n", "self.assertTrue('name=\"1\"\\r\\n\\r\\n2' in VAR_10, VAR_10)\n", "if __name__ == '__main__':\n", "logging.disable(logging.FATAL)\n", "unittest.main()\n" ]
[ "\"\"\"Unittest to exercise the code in url_helper.py.\"\"\"\n", "import logging\n", "import os\n", "import stat\n", "import StringIO\n", "import sys\n", "import tempfile\n", "import time\n", "import unittest\n", "import urllib\n", "import urllib2\n", "ROOT_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))\n", "sys.path.insert(0, ROOT_DIR)\n", "import test_env\n", "test_env.setup_test_env()\n", "from depot_tools import auto_stub\n", "from third_party.mox import mox\n", "import url_helper\n", "def setUp(self):...\n", "self._mox = mox.Mox()\n", "self.mock(logging, 'error', lambda *_: None)\n", "self.mock(logging, 'exception', lambda *_: None)\n", "self.mock(logging, 'info', lambda *_: None)\n", "self.mock(logging, 'warning', lambda *_: None)\n", "self._mox.StubOutWithMock(time, 'sleep')\n", "self._mox.StubOutWithMock(urllib2, 'urlopen')\n", "def tearDown(self):...\n", "self._mox.UnsetStubs()\n", "def testUrlOpenInvalidTryCount(self):...\n", "self._mox.ReplayAll()\n", "self.assertEqual(url_helper.UrlOpen('url', max_tries=-1), None)\n", "self._mox.VerifyAll()\n", "def testUrlOpenInvalidWaitDuration(self):...\n", "self._mox.ReplayAll()\n", "self.assertEqual(url_helper.UrlOpen('url', wait_duration=-1), None)\n", "self._mox.VerifyAll()\n", "def testUrlOpenGETSuccess(self):...\n", "url = 'http://my.url.com'\n", "response = 'True'\n", "url_helper.urllib2.urlopen(mox.StrContains(url), timeout=mox.IgnoreArg()\n ).AndReturn(StringIO.StringIO(response))\n", "self._mox.ReplayAll()\n", "self.assertEqual(url_helper.UrlOpen(url, method='GET'), response)\n", "self._mox.VerifyAll()\n", "def testUrlOpenPOSTSuccess(self):...\n", "url = 'http://my.url.com'\n", "response = 'True'\n", "url_helper.urllib2.urlopen(url, mox.IgnoreArg(), timeout=mox.IgnoreArg()\n ).AndReturn(StringIO.StringIO(response))\n", "self._mox.ReplayAll()\n", "self.assertEqual(url_helper.UrlOpen(url, method='POST'), response)\n", "self._mox.VerifyAll()\n", "def testUrlOpenPOSTFORMSuccess(self):...\n", "url = 'http://my.url.com'\n", "response = 'True'\n", "url_helper.urllib2.urlopen(mox.IsA(urllib2.Request), timeout=mox.IgnoreArg()\n ).AndReturn(StringIO.StringIO(response))\n", "self._mox.ReplayAll()\n", "self.assertEqual(url_helper.UrlOpen(url, method='POSTFORM'), response)\n", "self._mox.VerifyAll()\n", "def testUrlOpenSuccessAfterFailure(self):...\n", "url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndRaise(urllib2.URLError('url'))\n", "time.sleep(mox.IgnoreArg())\n", "response = 'True'\n", "url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndReturn(StringIO.StringIO(response))\n", "self._mox.ReplayAll()\n", "self.assertEqual(url_helper.UrlOpen('url', max_tries=2), response)\n", "self._mox.VerifyAll()\n", "def testUrlOpenFailure(self):...\n", "url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndRaise(urllib2.URLError('url'))\n", "self._mox.ReplayAll()\n", "self.assertIsNone(url_helper.UrlOpen('url', max_tries=1))\n", "self._mox.VerifyAll()\n", "def testUrlOpenHTTPErrorNoRetry(self):...\n", "url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndRaise(urllib2.HTTPError('url', 400, 'error message',\n None, None))\n", "self._mox.ReplayAll()\n", "self.assertIsNone(url_helper.UrlOpen('url', max_tries=10))\n", "self._mox.VerifyAll()\n", "def testUrlOpenHTTPErrorWithRetry(self):...\n", "response = 'response'\n", "url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndRaise(urllib2.HTTPError('url', 500, 'error message',\n None, None))\n", "time.sleep(mox.IgnoreArg())\n", "url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndReturn(StringIO.StringIO(response))\n", "self._mox.ReplayAll()\n", "self.assertEqual(response, url_helper.UrlOpen('url', max_tries=10))\n", "self._mox.VerifyAll()\n", "def testEnsureCountKeyIncludedInOpen(self):...\n", "attempts = 5\n", "for i in range(attempts):\n", "encoded_data = urllib.urlencode({url_helper.swarm_constants.COUNT_KEY: i})\n", "self._mox.ReplayAll()\n", "url_helper.urllib2.urlopen(mox.IgnoreArg(), encoded_data, timeout=mox.\n IgnoreArg()).AndRaise(urllib2.URLError('url'))\n", "self.assertEqual(url_helper.UrlOpen('url', max_tries=attempts), None)\n", "if i != attempts - 1:\n", "self._mox.VerifyAll()\n", "time.sleep(mox.IgnoreArg())\n", "def testCountKeyInData(self):...\n", "data = {url_helper.swarm_constants.COUNT_KEY: 1}\n", "self._mox.ReplayAll()\n", "self.assertEqual(url_helper.UrlOpen('url', data=data), None)\n", "self._mox.VerifyAll()\n", "def testNonAcsiiData(self):...\n", "data = {'r': u'not ascii £ һ'}\n", "url = 'http://my.url.com'\n", "response = 'True'\n", "url_helper.urllib2.urlopen(mox.StrContains(url), mox.IgnoreArg(), timeout=\n mox.IgnoreArg()).AndReturn(StringIO.StringIO(response))\n", "self._mox.ReplayAll()\n", "self.assertEqual(url_helper.UrlOpen(url, data=data), response)\n", "self._mox.VerifyAll()\n", "def testDownloadFile(self):...\n", "local_file = None\n", "local_file = tempfile.NamedTemporaryFile(delete=False)\n", "if local_file:\n", "def testDownloadFileDownloadError(self):...\n", "local_file.close()\n", "os.remove(local_file.name)\n", "fake_file = 'fake_local_file.fake'\n", "if os.path.exists(fake_file):\n", "def testDownloadFileSavingErrors(self):...\n", "self._mox.StubOutWithMock(url_helper, 'UrlOpen')\n", "self._mox.StubOutWithMock(url_helper, 'UrlOpen')\n", "os.remove(fake_file)\n", "file_readonly = None\n", "file_data = 'data'\n", "url_helper.UrlOpen(mox.IgnoreArg(), method='GET').AndReturn(None)\n", "file_readonly = tempfile.NamedTemporaryFile(delete=False)\n", "if file_readonly:\n", "def testEncodeMultipartFormData(self):...\n", "url_helper.UrlOpen(mox.IgnoreArg(), method='GET').AndReturn(file_data)\n", "self._mox.ReplayAll()\n", "file_readonly.close()\n", "os.remove(file_readonly.name)\n", "fields = [('x', 'y'), (1, 2)]\n", "self._mox.ReplayAll()\n", "self.assertFalse(url_helper.DownloadFile(fake_file, 'http://www.fakeurl.com'))\n", "os.chmod(file_readonly.name, stat.S_IREAD)\n", "files = [('key', 'filename', 'file data')]\n", "self.assertTrue(url_helper.DownloadFile(local_file.name,\n 'http://www.fakeurl.com'))\n", "self._mox.VerifyAll()\n", "self._mox.StubOutWithMock(url_helper, 'UrlOpen')\n", "content_type, body = url_helper.EncodeMultipartFormData()\n", "self.assertEqual(file_data, f.read())\n", "url_helper.UrlOpen(mox.IgnoreArg(), method='GET').AndReturn('data')\n", "self.assertTrue(content_type.startswith('multipart/form-data; boundary='))\n", "self._mox.VerifyAll()\n", "self._mox.ReplayAll()\n", "self.assertEqual('', body)\n", "self.assertFalse(url_helper.DownloadFile(file_readonly.name,\n 'http://www.fakeurl.com'))\n", "content_type, body = url_helper.EncodeMultipartFormData(fields=fields)\n", "self._mox.VerifyAll()\n", "self.assertTrue(content_type.startswith('multipart/form-data; boundary='))\n", "self.assertTrue('name=\"x\"\\r\\n\\r\\ny' in body, body)\n", "self.assertTrue('name=\"1\"\\r\\n\\r\\n2' in body, body)\n", "content_type, body = url_helper.EncodeMultipartFormData(files=files)\n", "self.assertTrue(content_type.startswith('multipart/form-data; boundary='))\n", "self.assertTrue('name=\"key\"; filename=\"filename\"' in body, body)\n", "self.assertTrue('file data' in body, body)\n", "content_type, body = url_helper.EncodeMultipartFormData(fields=fields,\n files=files)\n", "self.assertTrue(content_type.startswith('multipart/form-data; boundary='))\n", "self.assertTrue('name=\"x\"\\r\\n\\r\\ny' in body, body)\n", "self.assertTrue('name=\"1\"\\r\\n\\r\\n2' in body, body)\n", "if __name__ == '__main__':\n", "logging.disable(logging.FATAL)\n", "unittest.main()\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 5, 0, 0, 0, 0, 0, 0, 5, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 5, 0, 0, 5, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 5, 0, 0, 0, 0, 0, 0, 0, 0, 5, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 5, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Expr'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "Import'", "Assign'", "Expr'", "Import'", "Expr'", "ImportFrom'", "ImportFrom'", "Import'", "FunctionDef'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "Expr'", "FunctionDef'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "Expr'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "Expr'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "Assign'", "For", "Assign'", "Expr'", "Expr'", "Expr'", "Condition", "Expr'", "Expr'", "FunctionDef'", "Assign'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "FunctionDef'", "Assign'", "Assign'", "Condition", "FunctionDef'", "Expr'", "Expr'", "Assign'", "Condition", "FunctionDef'", "Expr'", "Expr'", "Expr'", "Assign'", "Assign'", "Expr'", "Assign'", "Condition", "FunctionDef'", "Expr'", "Expr'", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'", "Condition", "Expr'", "Expr'" ]
[ "def FUNC_5(self, VAR_0):...\n", "\"\"\"docstring\"\"\"\n", "VAR_22 = 'SELECT package.name, evr.epoch, evr.version, evr.release, arch.name'\n", "VAR_22 += ' FROM pkg_errata'\n", "VAR_22 += ' JOIN package ON package.id = pkg_errata.pkg_id'\n", "VAR_22 += ' JOIN evr ON evr.id = package.evr_id'\n", "VAR_22 += ' JOIN arch ON arch.id = package.arch_id'\n", "VAR_22 += ' WHERE pkg_errata.errata_id = %s' % str(VAR_0)\n", "self.cursor.execute(VAR_22)\n", "VAR_23 = self.cursor.fetchall()\n", "VAR_9 = []\n", "for VAR_1, VAR_12, VAR_13, VAR_14, VAR_15 in VAR_23:\n", "VAR_9.append(self.build_package_name(VAR_1, VAR_12, VAR_13, VAR_14, VAR_15))\n", "return VAR_9\n" ]
[ "def get_package_list_for_erratum_id(self, id):...\n", "\"\"\"docstring\"\"\"\n", "pkg_query = (\n 'SELECT package.name, evr.epoch, evr.version, evr.release, arch.name')\n", "pkg_query += ' FROM pkg_errata'\n", "pkg_query += ' JOIN package ON package.id = pkg_errata.pkg_id'\n", "pkg_query += ' JOIN evr ON evr.id = package.evr_id'\n", "pkg_query += ' JOIN arch ON arch.id = package.arch_id'\n", "pkg_query += ' WHERE pkg_errata.errata_id = %s' % str(id)\n", "self.cursor.execute(pkg_query)\n", "result = self.cursor.fetchall()\n", "package_list = []\n", "for name, epoch, version, release, arch in result:\n", "package_list.append(self.build_package_name(name, epoch, version, release,\n arch))\n", "return package_list\n" ]
[ 0, 0, 4, 4, 4, 4, 4, 4, 4, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "AugAssign'", "AugAssign'", "AugAssign'", "AugAssign'", "AugAssign'", "Expr'", "Assign'", "Assign'", "For", "Expr'", "Return'" ]
[ "def FUNC_3(VAR_1):...\n", "\"\"\"docstring\"\"\"\n", "VAR_3 = VAR_2.lexmatch.groupdict()\n", "VAR_1.value = beamr.interpreters.Emph(VAR_3['EMPH_FLAG'], VAR_3['EMPH_TXT'])\n", "return VAR_1\n" ]
[ "def t_EMPH(t):...\n", "\"\"\"docstring\"\"\"\n", "gd = lexer.lexmatch.groupdict()\n", "t.value = beamr.interpreters.Emph(gd['EMPH_FLAG'], gd['EMPH_TXT'])\n", "return t\n" ]
[ 0, 0, 2, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_2(VAR_1):...\n", "\"\"\"docstring\"\"\"\n", "VAR_17 = Trading(config_file=PATH_TO_YAML, warnings=True, timeout=20)\n", "print(e)\n", "return VAR_19\n", "VAR_18 = {'ActiveList': {'Include': True, 'Pagination': {'EntriesPerPage': \n 100, 'PageNumber': VAR_1}, 'IncludeWatchCount': True}, 'DetailLevel':\n 'ReturnAll'}\n", "print(e.response.dict())\n", "VAR_17.execute('GetMyeBaySelling', VAR_18)\n", "VAR_19 = VAR_17.response.dict()\n" ]
[ "def get_myebay_selling_request(page):...\n", "\"\"\"docstring\"\"\"\n", "api_trading = Trading(config_file=PATH_TO_YAML, warnings=True, timeout=20)\n", "print(e)\n", "return products\n", "api_request = {'ActiveList': {'Include': True, 'Pagination': {\n 'EntriesPerPage': 100, 'PageNumber': page}, 'IncludeWatchCount': True},\n 'DetailLevel': 'ReturnAll'}\n", "print(e.response.dict())\n", "api_trading.execute('GetMyeBaySelling', api_request)\n", "products = api_trading.response.dict()\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'", "Return'", "Assign'", "Expr'", "Expr'", "Assign'" ]
[ "def FUNC_3(VAR_11):...\n", "VAR_19 = '\"' + VAR_11 + '\"'\n", "VAR_20 = conn.cursor()\n", "VAR_20.execute('DELETE FROM recipe WHERE name = ' + '\"' + VAR_11 + '\"')\n", "print(VAR_20.rowcount)\n", "if VAR_20.rowcount == 1:\n", "messagebox.showinfo('Success', 'Recipe Deleted.')\n", "if VAR_20.rowcount == 0:\n", "VAR_6.pack_forget()\n", "messagebox.showerror('Cannot Delete', 'Cannot delete recipe, please try again.'\n )\n", "conn.close()\n", "VAR_5.pack(expand=True, fill='both')\n" ]
[ "def actually_delete(recipeName):...\n", "queryString = '\"' + recipeName + '\"'\n", "cursor = conn.cursor()\n", "cursor.execute('DELETE FROM recipe WHERE name = ' + '\"' + recipeName + '\"')\n", "print(cursor.rowcount)\n", "if cursor.rowcount == 1:\n", "messagebox.showinfo('Success', 'Recipe Deleted.')\n", "if cursor.rowcount == 0:\n", "menuFrame.pack_forget()\n", "messagebox.showerror('Cannot Delete', 'Cannot delete recipe, please try again.'\n )\n", "conn.close()\n", "viewRecipeFrame.pack(expand=True, fill='both')\n" ]
[ 0, 4, 0, 4, 4, 0, 0, 0, 0, 4, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Expr'", "Condition", "Expr'", "Condition", "Expr'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_5(self):...\n", "VAR_6 = self.client.get('/api/apps')\n", "VAR_8 = VAR_6.data['results'][0]['id']\n", "self.assertTrue(self.client.login(VAR_1='autotest-2', VAR_2='password'))\n", "VAR_5 = '/api/apps/{}/perms'.format(VAR_8)\n", "VAR_7 = {'username': 'autotest-2'}\n", "VAR_6 = self.client.post(VAR_5, json.dumps(VAR_7), content_type=\n 'application/json')\n", "self.assertEqual(VAR_6.status_code, 403)\n" ]
[ "def test_create_errors(self):...\n", "response = self.client.get('/api/apps')\n", "app_id = response.data['results'][0]['id']\n", "self.assertTrue(self.client.login(username='autotest-2', password='password'))\n", "url = '/api/apps/{}/perms'.format(app_id)\n", "body = {'username': 'autotest-2'}\n", "response = self.client.post(url, json.dumps(body), content_type=\n 'application/json')\n", "self.assertEqual(response.status_code, 403)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_1(self):...\n", "cmd.commands = self.orig_commands\n" ]
[ "def tearDown(self):...\n", "cmd.commands = self.orig_commands\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Assign'" ]
[ "async def FUNC_9(VAR_6, VAR_7):...\n", "VAR_8 = main.create_app()\n", "VAR_8.on_startup.append(FUNC_11)\n", "VAR_15 = await VAR_6(VAR_8)\n", "VAR_16 = await VAR_15.get(VAR_7)\n", "VAR_14 = await VAR_16.text()\n", "return VAR_14\n" ]
[ "async def _get(test_client, path):...\n", "app = main.create_app()\n", "app.on_startup.append(_clean_tables)\n", "client = await test_client(app)\n", "resp = await client.get(path)\n", "text = await resp.text()\n", "return text\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "AsyncFunctionDef'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_0(self, VAR_3):...\n", "VAR_21 = {'cls': self.encoder} if self.encoder else {}\n", "return json.dumps(VAR_3, **options)\n" ]
[ "def dumps(self, obj):...\n", "options = {'cls': self.encoder} if self.encoder else {}\n", "return json.dumps(obj, **options)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Return'" ]
[ "def FUNC_1(VAR_1, VAR_2=None):...\n", "if not 'instance' in VAR_1:\n", "VAR_13 = VAR_1['instance']\n", "FUNC_0(VAR_1)\n", "if not 'content' in VAR_1:\n", "VAR_1['content'] = CachedContent(VAR_13)\n", "def FUNC_21(VAR_14, VAR_15):...\n", "if not VAR_15 in VAR_1:\n", "VAR_1[VAR_15] = CachedPoints(VAR_13, VAR_14, VAR_1['content'])\n", "return VAR_1[VAR_15]\n" ]
[ "def _prepare_context(context, student=None):...\n", "if not 'instance' in context:\n", "instance = context['instance']\n", "_prepare_now(context)\n", "if not 'content' in context:\n", "context['content'] = CachedContent(instance)\n", "def points(user, key):...\n", "if not key in context:\n", "context[key] = CachedPoints(instance, user, context['content'])\n", "return context[key]\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Expr'", "Condition", "Assign'", "FunctionDef'", "Condition", "Assign'", "Return'" ]
[ "def FUNC_60(self):...\n", "self._test_strtype('blob', bytearray(VAR_13), VAR_23(VAR_13))\n" ]
[ "def t(self):...\n", "self._test_strtype('blob', bytearray(value), len(value))\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_4(self):...\n", "\"\"\"docstring\"\"\"\n", "config.set(xsrf_token_key=None)\n", "VAR_1 = utils.XsrfTool()\n", "VAR_2 = VAR_1.generate_token(12345, 'test_action')\n", "self.assertTrue(VAR_1.verify_token(VAR_2, 12345, 'test_action'))\n" ]
[ "def test_good_with_no_prior_key(self):...\n", "\"\"\"docstring\"\"\"\n", "config.set(xsrf_token_key=None)\n", "tool = utils.XsrfTool()\n", "token = tool.generate_token(12345, 'test_action')\n", "self.assertTrue(tool.verify_token(token, 12345, 'test_action'))\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'", "Assign'", "Assign'", "Expr'" ]
[ "@FUNC_6...\n", "\"\"\"docstring\"\"\"\n", "VAR_7 = users.find_one(VAR_1)\n", "def FUNC_18(VAR_16):...\n", "VAR_40 = ''\n", "VAR_41 = 1\n", "for item in VAR_16:\n", "if not item[0]:\n", "return VAR_40\n", "VAR_40 += '{}. {}\\n'.format(VAR_41, item[0])\n", "VAR_41 += 1\n" ]
[ "@cache_most_popular_items...\n", "\"\"\"docstring\"\"\"\n", "user = users.find_one(message)\n", "def list_to_ordered_str_list(list_of_gadgets):...\n", "string_roaster = ''\n", "index = 1\n", "for item in list_of_gadgets:\n", "if not item[0]:\n", "return string_roaster\n", "string_roaster += '{}. {}\\n'.format(index, item[0])\n", "index += 1\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "FunctionDef'", "Assign'", "Assign'", "For", "Condition", "Return'", "AugAssign'", "AugAssign'" ]
[ "def FUNC_19(self, VAR_8):...\n", "\"\"\"docstring\"\"\"\n", "VAR_8 = path.join(path.dirname(VAR_8), METADATA_FILENAME)\n", "VAR_4 = self.getCaption(VAR_8)\n", "return VAR_4\n" ]
[ "def getCaptionPic(self, pth):...\n", "\"\"\"docstring\"\"\"\n", "pth = path.join(path.dirname(pth), METADATA_FILENAME)\n", "data = self.getCaption(pth)\n", "return data\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_23():...\n", "" ]
[ "def stale_revision():...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "def FUNC_14(VAR_24=sys.argv):...\n", "VAR_39 = FUNC_12(VAR_24)\n", "if VAR_39.print_templates:\n", "print(VAR_39.templates)\n", "if not VAR_39.metadata:\n", "return 0\n", "if 'OS_CONFIG_FILES' in os.environ:\n", "if VAR_39.templates is None:\n", "VAR_28.error(e)\n", "return 0\n", "VAR_39.metadata = os.environ['OS_CONFIG_FILES'].split(':')\n", "VAR_39.metadata = FUNC_13(VAR_39.os_config_files)\n", "if VAR_39.key:\n", "return 1\n", "if not VAR_39.metadata and VAR_39.os_config_files == VAR_2:\n", "FUNC_2(VAR_39.metadata, VAR_39.key, VAR_39.type, VAR_39.key_default, VAR_39\n .fallback_metadata)\n", "FUNC_1(VAR_39.metadata, VAR_39.templates, VAR_39.output, VAR_39.validate,\n VAR_39.subhash, VAR_39.fallback_metadata)\n", "VAR_28.warning('DEPRECATED: falling back to %s' % VAR_3)\n", "VAR_28.info('success')\n", "VAR_39.metadata = FUNC_13(VAR_3)\n" ]
[ "def main(argv=sys.argv):...\n", "opts = parse_opts(argv)\n", "if opts.print_templates:\n", "print(opts.templates)\n", "if not opts.metadata:\n", "return 0\n", "if 'OS_CONFIG_FILES' in os.environ:\n", "if opts.templates is None:\n", "logger.error(e)\n", "return 0\n", "opts.metadata = os.environ['OS_CONFIG_FILES'].split(':')\n", "opts.metadata = load_list_from_json(opts.os_config_files)\n", "if opts.key:\n", "return 1\n", "if not opts.metadata and opts.os_config_files == OS_CONFIG_FILES_PATH:\n", "print_key(opts.metadata, opts.key, opts.type, opts.key_default, opts.\n fallback_metadata)\n", "install_config(opts.metadata, opts.templates, opts.output, opts.validate,\n opts.subhash, opts.fallback_metadata)\n", "logger.warning('DEPRECATED: falling back to %s' % OS_CONFIG_FILES_PATH_OLD)\n", "logger.info('success')\n", "opts.metadata = load_list_from_json(OS_CONFIG_FILES_PATH_OLD)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Expr'", "Condition", "Return'", "Condition", "Condition", "Expr'", "Return'", "Assign'", "Assign'", "Condition", "Return'", "Condition", "Expr'", "Expr'", "Expr'", "Expr'", "Assign'" ]
[ "def FUNC_1(self, VAR_1, VAR_2, VAR_5, VAR_6, VAR_4={}):...\n", "\"\"\"docstring\"\"\"\n", "VAR_22 = self.pool.get('stock.location')\n", "VAR_23 = self.pool.get('account.move')\n", "VAR_24 = self.pool.get('account.move.line')\n", "VAR_25 = VAR_6.get('new_price', 0.0)\n", "VAR_19 = VAR_6.get('stock_output_account', False)\n", "VAR_18 = VAR_6.get('stock_input_account', False)\n", "VAR_20 = VAR_6.get('stock_journal', False)\n", "VAR_17 = self.browse(VAR_1, VAR_2, VAR_5)[0]\n", "VAR_21 = VAR_17.categ_id.property_stock_variation\n", "VAR_26 = VAR_21 and VAR_21.id or False\n", "if not VAR_26:\n", "VAR_27 = []\n", "VAR_28 = VAR_22.search(VAR_1, VAR_2, [('usage', '=', 'internal')])\n", "for rec_id in VAR_5:\n", "for location in VAR_22.browse(VAR_1, VAR_2, VAR_28):\n", "return VAR_27\n", "VAR_49 = VAR_4.copy()\n", "self.write(VAR_1, VAR_2, rec_id, {'standard_price': VAR_25})\n", "VAR_49.update({'location': location.id, 'compute_child': False})\n", "VAR_45 = self.browse(VAR_1, VAR_2, rec_id, VAR_4=c)\n", "VAR_53 = VAR_45.qty_available\n", "VAR_54 = VAR_45.standard_price - VAR_25\n", "if not VAR_54:\n", "if VAR_53:\n", "VAR_56 = location.company_id and location.company_id.id or False\n", "if not VAR_56:\n", "if not VAR_20:\n", "VAR_20 = (VAR_45.categ_id.property_stock_journal and VAR_45.categ_id.\n property_stock_journal.id or False)\n", "if not VAR_20:\n", "VAR_57 = VAR_23.create(VAR_1, VAR_2, {'journal_id': VAR_20, 'company_id':\n VAR_56})\n", "VAR_27.append(VAR_57)\n", "if VAR_54 > 0:\n", "if not VAR_18:\n", "if VAR_54 < 0:\n", "VAR_18 = VAR_45.product_tmpl_id.property_stock_account_input.id\n", "if not VAR_18:\n", "if not VAR_19:\n", "VAR_18 = VAR_45.categ_id.property_stock_account_input_categ.id\n", "if not VAR_18:\n", "VAR_19 = VAR_45.product_tmpl_id.property_stock_account_output.id\n", "if not VAR_19:\n", "VAR_58 = VAR_53 * VAR_54\n", "VAR_19 = VAR_45.categ_id.property_stock_account_output_categ.id\n", "if not VAR_19:\n", "VAR_24.create(VAR_1, VAR_2, {'name': VAR_45.name, 'account_id': VAR_18,\n 'debit': VAR_58, 'move_id': VAR_57})\n", "VAR_58 = VAR_53 * -VAR_54\n", "VAR_24.create(VAR_1, VAR_2, {'name': VAR_45.categ_id.name, 'account_id':\n VAR_26, 'credit': VAR_58, 'move_id': VAR_57})\n", "VAR_24.create(VAR_1, VAR_2, {'name': VAR_45.name, 'account_id': VAR_19,\n 'credit': VAR_58, 'move_id': VAR_57})\n", "VAR_24.create(VAR_1, VAR_2, {'name': VAR_45.categ_id.name, 'account_id':\n VAR_26, 'debit': VAR_58, 'move_id': VAR_57})\n" ]
[ "def do_change_standard_price(self, cr, uid, ids, datas, context={}):...\n", "\"\"\"docstring\"\"\"\n", "location_obj = self.pool.get('stock.location')\n", "move_obj = self.pool.get('account.move')\n", "move_line_obj = self.pool.get('account.move.line')\n", "new_price = datas.get('new_price', 0.0)\n", "stock_output_acc = datas.get('stock_output_account', False)\n", "stock_input_acc = datas.get('stock_input_account', False)\n", "journal_id = datas.get('stock_journal', False)\n", "product_obj = self.browse(cr, uid, ids)[0]\n", "account_variation = product_obj.categ_id.property_stock_variation\n", "account_variation_id = account_variation and account_variation.id or False\n", "if not account_variation_id:\n", "move_ids = []\n", "loc_ids = location_obj.search(cr, uid, [('usage', '=', 'internal')])\n", "for rec_id in ids:\n", "for location in location_obj.browse(cr, uid, loc_ids):\n", "return move_ids\n", "c = context.copy()\n", "self.write(cr, uid, rec_id, {'standard_price': new_price})\n", "c.update({'location': location.id, 'compute_child': False})\n", "product = self.browse(cr, uid, rec_id, context=c)\n", "qty = product.qty_available\n", "diff = product.standard_price - new_price\n", "if not diff:\n", "if qty:\n", "company_id = location.company_id and location.company_id.id or False\n", "if not company_id:\n", "if not journal_id:\n", "journal_id = (product.categ_id.property_stock_journal and product.categ_id.\n property_stock_journal.id or False)\n", "if not journal_id:\n", "move_id = move_obj.create(cr, uid, {'journal_id': journal_id, 'company_id':\n company_id})\n", "move_ids.append(move_id)\n", "if diff > 0:\n", "if not stock_input_acc:\n", "if diff < 0:\n", "stock_input_acc = product.product_tmpl_id.property_stock_account_input.id\n", "if not stock_input_acc:\n", "if not stock_output_acc:\n", "stock_input_acc = product.categ_id.property_stock_account_input_categ.id\n", "if not stock_input_acc:\n", "stock_output_acc = product.product_tmpl_id.property_stock_account_output.id\n", "if not stock_output_acc:\n", "amount_diff = qty * diff\n", "stock_output_acc = product.categ_id.property_stock_account_output_categ.id\n", "if not stock_output_acc:\n", "move_line_obj.create(cr, uid, {'name': product.name, 'account_id':\n stock_input_acc, 'debit': amount_diff, 'move_id': move_id})\n", "amount_diff = qty * -diff\n", "move_line_obj.create(cr, uid, {'name': product.categ_id.name, 'account_id':\n account_variation_id, 'credit': amount_diff, 'move_id': move_id})\n", "move_line_obj.create(cr, uid, {'name': product.name, 'account_id':\n stock_output_acc, 'credit': amount_diff, 'move_id': move_id})\n", "move_line_obj.create(cr, uid, {'name': product.categ_id.name, 'account_id':\n account_variation_id, 'debit': amount_diff, 'move_id': move_id})\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "For", "For", "Return'", "Assign'", "Expr'", "Expr'", "Assign'", "Assign'", "Assign'", "Condition", "Condition", "Assign'", "Condition", "Condition", "Assign'", "Condition", "Assign'", "Expr'", "Condition", "Condition", "Condition", "Assign'", "Condition", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Expr'", "Assign'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_7(self, VAR_13, VAR_11, VAR_7, VAR_14, VAR_10):...\n", "return []\n" ]
[ "def get_directories(self, levelctx, levelfields, searcher, ctxlist, client):...\n", "return []\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_3(self, VAR_5):...\n", "self.common = self._init_common()\n", "self._check_flags()\n", "self.common.do_setup(VAR_5)\n" ]
[ "def do_setup(self, context):...\n", "self.common = self._init_common()\n", "self._check_flags()\n", "self.common.do_setup(context)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_1(VAR_0):...\n", "\"\"\"docstring\"\"\"\n", "VAR_1 = psycopg2.connect('dbname=forum')\n", "VAR_2 = VAR_1.cursor()\n", "VAR_2.execute(\"insert into posts values ('%s')\" % VAR_0)\n", "VAR_1.commit()\n", "VAR_1.close()\n" ]
[ "def add_post(content):...\n", "\"\"\"docstring\"\"\"\n", "conn = psycopg2.connect('dbname=forum')\n", "cursor = conn.cursor()\n", "cursor.execute(\"insert into posts values ('%s')\" % content)\n", "conn.commit()\n", "conn.close()\n" ]
[ 0, 0, 0, 0, 4, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_9(self):...\n", "VAR_7 = BokChoyTestSuite('', VAR_3=True)\n", "VAR_1 = 'tests'\n", "self.assertEqual(VAR_7.cmd, self._expected_command(VAR_1=name, VAR_3=True))\n" ]
[ "def test_verify_xss(self):...\n", "suite = BokChoyTestSuite('', verify_xss=True)\n", "name = 'tests'\n", "self.assertEqual(suite.cmd, self._expected_command(name=name, verify_xss=True))\n" ]
[ 0, 0, 0, 3 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_1():...\n", "VAR_13 = argparse.ArgumentParser(description='Stress tester for CMS')\n", "VAR_13.add_argument('-c', '--contest-id', action='store', type=int,\n required=True, help='ID of the contest to test against')\n", "VAR_13.add_argument('-n', '--actor-num', action='store', type=int, help=\n 'the number of actors to spawn')\n", "VAR_13.add_argument('-s', '--sort-actors', action='store_true', help=\n 'sort usernames alphabetically before slicing them')\n", "VAR_13.add_argument('-u', '--base-url', action='store', type=utf8_decoder,\n help='base URL for placing HTTP requests')\n", "VAR_13.add_argument('-S', '--submissions-path', action='store', type=\n utf8_decoder, help='base path for submission to send')\n", "VAR_13.add_argument('-p', '--prepare-path', action='store', type=\n utf8_decoder, help='file to put contest info to')\n", "VAR_13.add_argument('-r', '--read-from', action='store', type=utf8_decoder,\n help='file to read contest info from')\n", "VAR_13.add_argument('-t', '--time-coeff', action='store', type=float,\n default=10.0, help='average wait between actions')\n", "VAR_13.add_argument('-o', '--only-submit', action='store_true', help=\n 'whether the actor only submits solutions')\n", "VAR_14 = VAR_13.parse_args()\n", "if VAR_14.prepare_path is not None:\n", "VAR_12, VAR_8 = FUNC_0(VAR_14.contest_id)\n", "assert VAR_14.time_coeff > 0.0\n", "VAR_27 = dict()\n", "assert not (VAR_14.only_submit and VAR_14.submissions_path == '')\n", "VAR_27['users'] = VAR_12\n", "VAR_12 = []\n", "VAR_27['tasks'] = VAR_8\n", "VAR_8 = []\n", "file_.write('%s' % VAR_27)\n", "if VAR_14.read_from is None:\n", "return\n", "VAR_12, VAR_8 = FUNC_0(VAR_14.contest_id)\n", "VAR_27 = ast.literal_eval(file_.read())\n", "if VAR_14.actor_num is not None:\n", "VAR_12 = VAR_27['users']\n", "VAR_28 = VAR_12.items()\n", "if VAR_14.base_url is not None:\n", "VAR_8 = VAR_27['tasks']\n", "if VAR_14.sort_actors:\n", "VAR_10 = VAR_14.base_url\n", "VAR_10 = 'http://%s:%d/' % (get_service_address(ServiceCoord(\n 'ContestWebServer', 0))[0], config.contest_listen_port[0])\n", "VAR_28.sort()\n", "random.shuffle(VAR_28)\n", "VAR_7 = VAR_1\n", "VAR_12 = dict(VAR_28[:VAR_14.actor_num])\n", "VAR_7['time_coeff'] = VAR_14.time_coeff\n", "VAR_15 = CLASS_3\n", "if VAR_14.only_submit:\n", "VAR_15 = CLASS_4\n", "VAR_16 = [VAR_15(VAR_5, data['password'], VAR_7, VAR_8, VAR_9=RequestLog(\n log_dir=os.path.join('./test_logs', username)), VAR_10=base_url, VAR_11\n =args.submissions_path) for VAR_5, data in VAR_12.iteritems()]\n", "for actor in VAR_16:\n", "actor.start()\n", "while True:\n", "print('Taking down actors', file=sys.stderr)\n", "VAR_17 = False\n", "time.sleep(1)\n", "for actor in VAR_16:\n", "while not VAR_17:\n", "actor.die = True\n", "for actor in VAR_16:\n", "print('Test finished', file=sys.stderr)\n", "actor.join()\n", "VAR_18 = CLASS_0()\n", "for actor in VAR_16:\n", "VAR_18.merge(actor.log)\n", "VAR_18.print_stats()\n" ]
[ "def main():...\n", "parser = argparse.ArgumentParser(description='Stress tester for CMS')\n", "parser.add_argument('-c', '--contest-id', action='store', type=int,\n required=True, help='ID of the contest to test against')\n", "parser.add_argument('-n', '--actor-num', action='store', type=int, help=\n 'the number of actors to spawn')\n", "parser.add_argument('-s', '--sort-actors', action='store_true', help=\n 'sort usernames alphabetically before slicing them')\n", "parser.add_argument('-u', '--base-url', action='store', type=utf8_decoder,\n help='base URL for placing HTTP requests')\n", "parser.add_argument('-S', '--submissions-path', action='store', type=\n utf8_decoder, help='base path for submission to send')\n", "parser.add_argument('-p', '--prepare-path', action='store', type=\n utf8_decoder, help='file to put contest info to')\n", "parser.add_argument('-r', '--read-from', action='store', type=utf8_decoder,\n help='file to read contest info from')\n", "parser.add_argument('-t', '--time-coeff', action='store', type=float,\n default=10.0, help='average wait between actions')\n", "parser.add_argument('-o', '--only-submit', action='store_true', help=\n 'whether the actor only submits solutions')\n", "args = parser.parse_args()\n", "if args.prepare_path is not None:\n", "users, tasks = harvest_contest_data(args.contest_id)\n", "assert args.time_coeff > 0.0\n", "contest_data = dict()\n", "assert not (args.only_submit and args.submissions_path == '')\n", "contest_data['users'] = users\n", "users = []\n", "contest_data['tasks'] = tasks\n", "tasks = []\n", "file_.write('%s' % contest_data)\n", "if args.read_from is None:\n", "return\n", "users, tasks = harvest_contest_data(args.contest_id)\n", "contest_data = ast.literal_eval(file_.read())\n", "if args.actor_num is not None:\n", "users = contest_data['users']\n", "user_items = users.items()\n", "if args.base_url is not None:\n", "tasks = contest_data['tasks']\n", "if args.sort_actors:\n", "base_url = args.base_url\n", "base_url = 'http://%s:%d/' % (get_service_address(ServiceCoord(\n 'ContestWebServer', 0))[0], config.contest_listen_port[0])\n", "user_items.sort()\n", "random.shuffle(user_items)\n", "metrics = DEFAULT_METRICS\n", "users = dict(user_items[:args.actor_num])\n", "metrics['time_coeff'] = args.time_coeff\n", "actor_class = RandomActor\n", "if args.only_submit:\n", "actor_class = SubmitActor\n", "actors = [actor_class(username, data['password'], metrics, tasks, log=\n RequestLog(log_dir=os.path.join('./test_logs', username)), base_url=\n base_url, submissions_path=args.submissions_path) for username, data in\n users.iteritems()]\n", "for actor in actors:\n", "actor.start()\n", "while True:\n", "print('Taking down actors', file=sys.stderr)\n", "finished = False\n", "time.sleep(1)\n", "for actor in actors:\n", "while not finished:\n", "actor.die = True\n", "for actor in actors:\n", "print('Test finished', file=sys.stderr)\n", "actor.join()\n", "great_log = RequestLog()\n", "for actor in actors:\n", "great_log.merge(actor.log)\n", "great_log.print_stats()\n" ]
[ 0, 0, 0, 0, 0, 5, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 5, 0, 0, 5, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Assign'", "Condition", "Assign'", "Assert'", "Assign'", "Assert'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Condition", "Return'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "For", "Expr'", "Condition", "Expr'", "Assign'", "Expr'", "For", "Condition", "Assign'", "For", "Expr'", "Expr'", "Assign'", "For", "Expr'", "Expr'" ]
[ "def FUNC_5(VAR_3, VAR_7, VAR_8=None):...\n", "\"\"\"docstring\"\"\"\n", "VAR_5 = {'currentProvider': None, 'providers': [], 'secondaryProviders': [],\n 'finishAuthUrl': None, 'errorMessage': None,\n 'registerFormSubmitButtonText': _('Create Account')}\n", "if third_party_auth.is_enabled():\n", "VAR_20 = enterprise_customer_for_request(VAR_3)\n", "return VAR_5\n", "if not VAR_20:\n", "for enabled in third_party_auth.provider.Registry.displayed_for_login(VAR_8\n", "VAR_36 = pipeline.get(VAR_3)\n", "VAR_44 = {'id': enabled.provider_id, 'name': enabled.name, 'iconClass': \n enabled.icon_class or None, 'iconImage': enabled.icon_image.url if\n enabled.icon_image else None, 'loginUrl': pipeline.get_login_url(\n enabled.provider_id, pipeline.AUTH_ENTRY_LOGIN, redirect_url=\n redirect_to), 'registerUrl': pipeline.get_login_url(enabled.provider_id,\n pipeline.AUTH_ENTRY_REGISTER, redirect_url=redirect_to)}\n", "if VAR_36 is not None:\n", "VAR_5['providers' if not enabled.secondary else 'secondaryProviders'].append(\n VAR_44)\n", "VAR_41 = third_party_auth.provider.Registry.get_from_pipeline(VAR_36)\n", "for msg in messages.get_messages(VAR_3):\n", "if VAR_41 is not None:\n", "if msg.extra_tags.split()[0] == 'social-auth':\n", "VAR_5['currentProvider'] = VAR_41.name\n", "VAR_5['errorMessage'] = _(unicode(msg))\n", "VAR_5['finishAuthUrl'] = pipeline.get_complete_url(VAR_41.backend_name)\n", "if VAR_41.skip_registration_form:\n", "if not VAR_20:\n", "VAR_5['autoSubmitRegForm'] = True\n", "VAR_5['autoRegisterWelcomeMessage'] = (\n 'Thank you for joining {}. Just a couple steps before you start learning!'\n .format(configuration_helpers.get_value('PLATFORM_NAME', settings.\n PLATFORM_NAME)))\n", "VAR_5['registerFormSubmitButtonText'] = _('Continue')\n" ]
[ "def _third_party_auth_context(request, redirect_to, tpa_hint=None):...\n", "\"\"\"docstring\"\"\"\n", "context = {'currentProvider': None, 'providers': [], 'secondaryProviders':\n [], 'finishAuthUrl': None, 'errorMessage': None,\n 'registerFormSubmitButtonText': _('Create Account')}\n", "if third_party_auth.is_enabled():\n", "enterprise_customer = enterprise_customer_for_request(request)\n", "return context\n", "if not enterprise_customer:\n", "for enabled in third_party_auth.provider.Registry.displayed_for_login(tpa_hint\n", "running_pipeline = pipeline.get(request)\n", "info = {'id': enabled.provider_id, 'name': enabled.name, 'iconClass': \n enabled.icon_class or None, 'iconImage': enabled.icon_image.url if\n enabled.icon_image else None, 'loginUrl': pipeline.get_login_url(\n enabled.provider_id, pipeline.AUTH_ENTRY_LOGIN, redirect_url=\n redirect_to), 'registerUrl': pipeline.get_login_url(enabled.provider_id,\n pipeline.AUTH_ENTRY_REGISTER, redirect_url=redirect_to)}\n", "if running_pipeline is not None:\n", "context['providers' if not enabled.secondary else 'secondaryProviders'].append(\n info)\n", "current_provider = third_party_auth.provider.Registry.get_from_pipeline(\n running_pipeline)\n", "for msg in messages.get_messages(request):\n", "if current_provider is not None:\n", "if msg.extra_tags.split()[0] == 'social-auth':\n", "context['currentProvider'] = current_provider.name\n", "context['errorMessage'] = _(unicode(msg))\n", "context['finishAuthUrl'] = pipeline.get_complete_url(current_provider.\n backend_name)\n", "if current_provider.skip_registration_form:\n", "if not enterprise_customer:\n", "context['autoSubmitRegForm'] = True\n", "context['autoRegisterWelcomeMessage'] = (\n 'Thank you for joining {}. Just a couple steps before you start learning!'\n .format(configuration_helpers.get_value('PLATFORM_NAME', settings.\n PLATFORM_NAME)))\n", "context['registerFormSubmitButtonText'] = _('Continue')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Assign'", "Return'", "Condition", "For", "Assign'", "Assign'", "Condition", "Expr'", "Assign'", "For", "Condition", "Condition", "Assign'", "Assign'", "Assign'", "For", "Condition", "Assign'", "Assign'", "Assign'" ]
[ "import os\n", "from posixpath import normpath\n", "import re\n", "import warnings\n", "from wsgiref.headers import Headers\n", "from wsgiref.util import FileWrapper\n", "from .media_types import MediaTypes\n", "from .scantree import scantree\n", "from .responders import StaticFile, MissingFileError, IsDirectoryError, Redirect\n", "from .string_utils import decode_if_byte_string, decode_path_info, ensure_leading_trailing_slash\n", "VAR_0 = 10 * 365 * 24 * 60 * 60\n", "VAR_1 = ('autorefresh', 'max_age', 'allow_all_origins', 'charset',\n 'mimetypes', 'add_headers_function', 'index_file', 'immutable_file_test')\n", "VAR_2 = False\n", "VAR_3 = 60\n", "VAR_4 = True\n", "VAR_5 = 'utf-8'\n", "VAR_6 = None\n", "VAR_7 = None\n", "VAR_8 = None\n", "def __init__(self, VAR_9, VAR_10=None, VAR_11=None, **VAR_12):...\n", "for attr in self.config_attrs:\n", "if VAR_12:\n", "VAR_34 = VAR_12.pop(attr)\n", "VAR_34 = decode_if_byte_string(VAR_34)\n", "self.media_types = MediaTypes(extra_types=self.mimetypes)\n", "setattr(self, attr, VAR_34)\n", "self.application = VAR_9\n", "self.files = {}\n", "self.directories = []\n", "if self.index_file is True:\n", "self.index_file = 'index.html'\n", "if not callable(self.immutable_file_test):\n", "VAR_26 = re.compile(self.immutable_file_test)\n", "if VAR_10 is not None:\n", "self.immutable_file_test = lambda VAR_17, VAR_16: bool(VAR_26.search(VAR_16))\n", "self.add_files(VAR_10, VAR_11)\n", "def __call__(self, VAR_13, VAR_14):...\n", "VAR_17 = decode_path_info(VAR_13.get('PATH_INFO', ''))\n", "if self.autorefresh:\n", "VAR_15 = self.find_file(VAR_17)\n", "VAR_15 = self.files.get(VAR_17)\n", "if VAR_15 is None:\n", "return self.application(VAR_13, VAR_14)\n", "return self.serve(VAR_15, VAR_13, VAR_14)\n" ]
[ "import os\n", "from posixpath import normpath\n", "import re\n", "import warnings\n", "from wsgiref.headers import Headers\n", "from wsgiref.util import FileWrapper\n", "from .media_types import MediaTypes\n", "from .scantree import scantree\n", "from .responders import StaticFile, MissingFileError, IsDirectoryError, Redirect\n", "from .string_utils import decode_if_byte_string, decode_path_info, ensure_leading_trailing_slash\n", "FOREVER = 10 * 365 * 24 * 60 * 60\n", "config_attrs = ('autorefresh', 'max_age', 'allow_all_origins', 'charset',\n 'mimetypes', 'add_headers_function', 'index_file', 'immutable_file_test')\n", "autorefresh = False\n", "max_age = 60\n", "allow_all_origins = True\n", "charset = 'utf-8'\n", "mimetypes = None\n", "add_headers_function = None\n", "index_file = None\n", "def __init__(self, application, root=None, prefix=None, **kwargs):...\n", "for attr in self.config_attrs:\n", "if kwargs:\n", "value = kwargs.pop(attr)\n", "value = decode_if_byte_string(value)\n", "self.media_types = MediaTypes(extra_types=self.mimetypes)\n", "setattr(self, attr, value)\n", "self.application = application\n", "self.files = {}\n", "self.directories = []\n", "if self.index_file is True:\n", "self.index_file = 'index.html'\n", "if not callable(self.immutable_file_test):\n", "regex = re.compile(self.immutable_file_test)\n", "if root is not None:\n", "self.immutable_file_test = lambda path, url: bool(regex.search(url))\n", "self.add_files(root, prefix)\n", "def __call__(self, environ, start_response):...\n", "path = decode_path_info(environ.get('PATH_INFO', ''))\n", "if self.autorefresh:\n", "static_file = self.find_file(path)\n", "static_file = self.files.get(path)\n", "if static_file is None:\n", "return self.application(environ, start_response)\n", "return self.serve(static_file, environ, start_response)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "ImportFrom'", "Import'", "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "For", "Condition", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Expr'", "FunctionDef'", "Assign'", "Condition", "Assign'", "Assign'", "Condition", "Return'", "Return'" ]
[ "def FUNC_3(VAR_16):...\n", "return VAR_16.first_name + ' ' + VAR_16.last_name\n" ]
[ "def get_private_display_name(this):...\n", "return this.first_name + ' ' + this.last_name\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]