lines
listlengths 1
444
| raw_lines
listlengths 1
444
| label
listlengths 1
444
| type
listlengths 1
444
|
|---|---|---|---|
[
"@VAR_1.route('/get_cuisines')...\n",
"VAR_18 = VAR_3.get_cuisines()\n",
"if VAR_18 == -1:\n",
"return None\n",
"VAR_0.info('GET get_cuisines query')\n",
"return VAR_18\n"
] |
[
"@app.route('/get_cuisines')...\n",
"query_res = database.get_cuisines()\n",
"if query_res == -1:\n",
"return None\n",
"logger.info('GET get_cuisines query')\n",
"return query_res\n"
] |
[
0,
0,
0,
0,
0,
0
] |
[
"Condition",
"Assign'",
"Condition",
"Return'",
"Expr'",
"Return'"
] |
[
"def FUNC_4(VAR_6, VAR_8, VAR_7, VAR_9):...\n",
"VAR_19 = ''\n",
"print('Linking : discordId : ' + str(VAR_6) + ', osuName : ' + VAR_8 +\n ', osuId : ' + str(VAR_7) + ' to Database.', end=' ')\n",
"VAR_16 = sqlite3.connect(VAR_5)\n",
"VAR_17 = VAR_16.cursor()\n",
"VAR_17.execute('SELECT * FROM users WHERE discordId = ' + str(VAR_6))\n",
"if len(VAR_17.fetchall()) == 0:\n",
"VAR_17.execute(\n \"\"\"\n\t\tINSERT INTO users (discordId, osuName, osuId, rank) \n\t\tVALUES (?, ?, ?, ?)\n\t\t\"\"\"\n , (VAR_6, VAR_8, VAR_7, VAR_9))\n",
"VAR_17.execute(\"UPDATE users SET osuName = '\" + VAR_8 + \"', osuId = \" + str\n (VAR_7) + \", rank = '\" + VAR_9 + \"' WHERE discordId = \" + str(VAR_6))\n",
"VAR_16.commit()\n",
"VAR_16.commit()\n",
"print('Added')\n",
"print('Updated')\n",
"VAR_19 = 'linked'\n",
"VAR_19 = 'updated'\n",
"VAR_16.close()\n",
"return VAR_19\n"
] |
[
"def link_user(discordId, osuName, osuId, rank):...\n",
"result = ''\n",
"print('Linking : discordId : ' + str(discordId) + ', osuName : ' + osuName +\n ', osuId : ' + str(osuId) + ' to Database.', end=' ')\n",
"conn = sqlite3.connect(databasePath)\n",
"cursor = conn.cursor()\n",
"cursor.execute('SELECT * FROM users WHERE discordId = ' + str(discordId))\n",
"if len(cursor.fetchall()) == 0:\n",
"cursor.execute(\n \"\"\"\n\t\tINSERT INTO users (discordId, osuName, osuId, rank) \n\t\tVALUES (?, ?, ?, ?)\n\t\t\"\"\"\n , (discordId, osuName, osuId, rank))\n",
"cursor.execute(\"UPDATE users SET osuName = '\" + osuName + \"', osuId = \" +\n str(osuId) + \", rank = '\" + rank + \"' WHERE discordId = \" + str(discordId))\n",
"conn.commit()\n",
"conn.commit()\n",
"print('Added')\n",
"print('Updated')\n",
"result = 'linked'\n",
"result = 'updated'\n",
"conn.close()\n",
"return result\n"
] |
[
0,
0,
0,
0,
0,
2,
0,
0,
2,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"Condition",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"Return'"
] |
[
"@api_post...\n",
"VAR_7 = VAR_3.get('submit_id', 0)\n",
"VAR_8 = VAR_3.get('password', None)\n",
"VAR_9 = VAR_3.get('astree', True)\n",
"VAR_6 = VAR_0.get_files(VAR_7=submit_id, VAR_8=password, VAR_9=astree)\n",
"return JsonResponse({'status': True, 'data': VAR_6, 'defaults': FUNC_0()},\n encoder=JsonSerialize)\n"
] |
[
"@api_post...\n",
"submit_id = body.get('submit_id', 0)\n",
"password = body.get('password', None)\n",
"astree = body.get('astree', True)\n",
"data = submit_manager.get_files(submit_id=submit_id, password=password,\n astree=astree)\n",
"return JsonResponse({'status': True, 'data': data, 'defaults': defaults()},\n encoder=JsonSerialize)\n"
] |
[
0,
0,
0,
0,
1,
1
] |
[
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_0(VAR_0=['POST', 'GET', 'PUT', 'DELETE']):...\n",
"\"\"\"docstring\"\"\"\n",
"def FUNC_5(VAR_1, *VAR_2, **VAR_3):...\n",
"def FUNC_6():...\n",
"if not request.method in VAR_0:\n",
"return True\n",
"if not request.environ.get('AUTH_TYPE') == 'cookie':\n",
"return True\n",
"if config.get('skip_authentication'):\n",
"return True\n",
"if request.environ.get('HTTP_REFERER'):\n",
"VAR_7 = urlparse(request.environ.get('HTTP_REFERER'))\n",
"if request.method == 'GET' and FUNC_4():\n",
"VAR_8 = VAR_7.hostname\n",
"return True\n",
"return False\n",
"if VAR_7.port:\n",
"VAR_8 += ':' + str(VAR_7.port)\n",
"if VAR_8.endswith(request.environ['adhocracy.domain']):\n",
"if request.method != 'GET':\n",
"return True\n"
] |
[
"def RequireInternalRequest(methods=['POST', 'GET', 'PUT', 'DELETE']):...\n",
"\"\"\"docstring\"\"\"\n",
"def _decorate(f, *a, **kw):...\n",
"def check():...\n",
"if not request.method in methods:\n",
"return True\n",
"if not request.environ.get('AUTH_TYPE') == 'cookie':\n",
"return True\n",
"if config.get('skip_authentication'):\n",
"return True\n",
"if request.environ.get('HTTP_REFERER'):\n",
"ref_url = urlparse(request.environ.get('HTTP_REFERER'))\n",
"if request.method == 'GET' and has_token():\n",
"ref_host = ref_url.hostname\n",
"return True\n",
"return False\n",
"if ref_url.port:\n",
"ref_host += ':' + str(ref_url.port)\n",
"if ref_host.endswith(request.environ['adhocracy.domain']):\n",
"if request.method != 'GET':\n",
"return True\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
5,
0,
0,
0,
0,
0,
0,
5,
0
] |
[
"FunctionDef'",
"Docstring",
"FunctionDef'",
"FunctionDef'",
"Condition",
"Return'",
"Condition",
"Return'",
"Condition",
"Return'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Return'",
"Return'",
"Condition",
"AugAssign'",
"Condition",
"Condition",
"Return'"
] |
[
"def FUNC_36(self, VAR_14):...\n",
"return self.__class__(self.get_related_model(VAR_14), self.session)\n"
] |
[
"def get_related_interface(self, col_name):...\n",
"return self.__class__(self.get_related_model(col_name), self.session)\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Return'"
] |
[
"@VAR_0.route('/queries', methods=['GET', 'POST'])...\n",
"if request.method == 'GET':\n",
"return render_template('queries-form.html', queries=get_queries(app, g))\n",
"VAR_5 = request.form['query-selector']\n",
"VAR_6 = get_queries(VAR_0, g)[VAR_5]\n",
"VAR_7, VAR_4 = execute_query(VAR_0, g, VAR_6)\n",
"return jsonify([('', VAR_7, VAR_4)])\n"
] |
[
"@app.route('/queries', methods=['GET', 'POST'])...\n",
"if request.method == 'GET':\n",
"return render_template('queries-form.html', queries=get_queries(app, g))\n",
"query_key = request.form['query-selector']\n",
"query = get_queries(app, g)[query_key]\n",
"schema, data = execute_query(app, g, query)\n",
"return jsonify([('', schema, data)])\n"
] |
[
0,
0,
4,
0,
4,
4,
0
] |
[
"Condition",
"Condition",
"Return'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_12(self, VAR_14):...\n",
"\"\"\"docstring\"\"\"\n",
"return VAR_14\n"
] |
[
"def write(self, value):...\n",
"\"\"\"docstring\"\"\"\n",
"return value\n"
] |
[
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Return'"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self.config.get(self.section, 'listen_addr')\n"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self.config.get(self.section, 'listen_addr')\n"
] |
[
0,
0,
0
] |
[
"Condition",
"Docstring",
"Return'"
] |
[
"from __future__ import unicode_literals\n",
"from django.db import models, migrations\n",
"VAR_0 = [('userprofile', '0002_auto_20150427_1717'), ('exercise',\n '0014_ltiexercise')]\n",
"VAR_1 = [migrations.CreateModel(name='LearningObjectDisplay', fields=[('id',\n models.AutoField(verbose_name='ID', serialize=False, auto_created=True,\n primary_key=True)), ('timestamp', models.DateTimeField(auto_now_add=\n True)), ('learning_object', models.ForeignKey(to=\n 'exercise.LearningObject')), ('profile', models.ForeignKey(to=\n 'userprofile.UserProfile'))], options={}, bases=(models.Model,)),\n migrations.AlterField(model_name='learningobject', name='status', field\n =models.CharField(choices=[('ready', 'Ready'), ('unlisted',\n 'Unlisted in table of contents'), ('enrollment', 'Enrollment questions'\n ), ('hidden', 'Hidden from non course staff'), ('maintenance',\n 'Maintenance')], max_length=32, default='ready'), preserve_default=True)]\n"
] |
[
"from __future__ import unicode_literals\n",
"from django.db import models, migrations\n",
"dependencies = [('userprofile', '0002_auto_20150427_1717'), ('exercise',\n '0014_ltiexercise')]\n",
"operations = [migrations.CreateModel(name='LearningObjectDisplay', fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False,\n auto_created=True, primary_key=True)), ('timestamp', models.\n DateTimeField(auto_now_add=True)), ('learning_object', models.\n ForeignKey(to='exercise.LearningObject')), ('profile', models.\n ForeignKey(to='userprofile.UserProfile'))], options={}, bases=(models.\n Model,)), migrations.AlterField(model_name='learningobject', name=\n 'status', field=models.CharField(choices=[('ready', 'Ready'), (\n 'unlisted', 'Unlisted in table of contents'), ('enrollment',\n 'Enrollment questions'), ('hidden', 'Hidden from non course staff'), (\n 'maintenance', 'Maintenance')], max_length=32, default='ready'),\n preserve_default=True)]\n"
] |
[
0,
0,
0,
4
] |
[
"ImportFrom'",
"ImportFrom'",
"Assign'",
"Assign'"
] |
[
"def FUNC_20(VAR_6):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_35 = 'disk', 'partition'\n",
"if not isinstance(VAR_6, (list, tuple)):\n",
"VAR_6 = [VAR_6]\n",
"VAR_6 = [block.sys_block_path(VAR_15) for VAR_15 in VAR_6]\n",
"for VAR_4 in [FUNC_16(p) for p in VAR_6]:\n",
"if any(holder_type not in VAR_35 and VAR_15 not in VAR_6 for holder_type,\n"
] |
[
"def assert_clear(base_paths):...\n",
"\"\"\"docstring\"\"\"\n",
"valid = 'disk', 'partition'\n",
"if not isinstance(base_paths, (list, tuple)):\n",
"base_paths = [base_paths]\n",
"base_paths = [block.sys_block_path(path) for path in base_paths]\n",
"for holders_tree in [gen_holders_tree(p) for p in base_paths]:\n",
"if any(holder_type not in valid and path not in base_paths for holder_type,\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"For",
"For"
] |
[
"def FUNC_0(self, VAR_3):...\n",
"return '{:,}'.format(VAR_3.num_cycles)\n"
] |
[
"def render_num_cycles(self, record):...\n",
"return '{:,}'.format(record.num_cycles)\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Return'"
] |
[
"def FUNC_7(self, VAR_17):...\n",
"self._WriteHeader('text/css')\n",
"self._WriteTemplate(VAR_17)\n"
] |
[
"def _DoCSS(self, template):...\n",
"self._WriteHeader('text/css')\n",
"self._WriteTemplate(template)\n"
] |
[
0,
0,
0
] |
[
"FunctionDef'",
"Expr'",
"Expr'"
] |
[
"@api.depends('state')...\n",
"for rec in self:\n",
"if rec.state.readonly_fields:\n",
"rec.crapo_readonly_fields = ',{},'.format(rec.state.readonly_fields)\n",
"rec.crapo_readonly_fields = ',0,'\n"
] |
[
"@api.depends('state')...\n",
"for rec in self:\n",
"if rec.state.readonly_fields:\n",
"rec.crapo_readonly_fields = ',{},'.format(rec.state.readonly_fields)\n",
"rec.crapo_readonly_fields = ',0,'\n"
] |
[
0,
0,
0,
0,
0
] |
[
"Condition",
"For",
"Condition",
"Assign'",
"Assign'"
] |
[
"def FUNC_36(VAR_11):...\n",
"for VAR_51, VAR_58 in VAR_11.items():\n",
"if isinstance(VAR_58, str) or not isinstance(VAR_58, Iterable):\n",
"VAR_58 = [VAR_58]\n",
"yield [(VAR_51, VAR_16) for VAR_16 in VAR_58]\n"
] |
[
"def flatten(wildcards):...\n",
"for wildcard, values in wildcards.items():\n",
"if isinstance(values, str) or not isinstance(values, Iterable):\n",
"values = [values]\n",
"yield [(wildcard, value) for value in values]\n"
] |
[
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"For",
"Condition",
"Assign'",
"Expr'"
] |
[
"def FUNC_43(self, VAR_21=None):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_32 = list()\n",
"VAR_21 = VAR_21 or self.get_columns_list()\n",
"for VAR_14 in VAR_21:\n",
"if not self.is_relation(VAR_14):\n",
"return VAR_32\n",
"if hasattr(self.obj, VAR_14):\n",
"if not hasattr(getattr(self.obj, VAR_14), '__call__') or hasattr(getattr(\n",
"VAR_32.append(VAR_14)\n",
"VAR_32.append(VAR_14)\n"
] |
[
"def get_order_columns_list(self, list_columns=None):...\n",
"\"\"\"docstring\"\"\"\n",
"ret_lst = list()\n",
"list_columns = list_columns or self.get_columns_list()\n",
"for col_name in list_columns:\n",
"if not self.is_relation(col_name):\n",
"return ret_lst\n",
"if hasattr(self.obj, col_name):\n",
"if not hasattr(getattr(self.obj, col_name), '__call__') or hasattr(getattr(\n",
"ret_lst.append(col_name)\n",
"ret_lst.append(col_name)\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"For",
"Condition",
"Return'",
"Condition",
"Condition",
"Expr'",
"Expr'"
] |
[
"@app.route('/achievements/<achievement_id>')...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_5 = request.args.get('language', 'en')\n",
"VAR_6 = request.args.get('region', 'US')\n",
"VAR_11 = db.connection.cursor(db.pymysql.cursors.DictCursor)\n",
"VAR_11.execute(VAR_0 + 'WHERE ach.id = %(achievement_id)s', {'language':\n VAR_5, 'region': VAR_6, 'achievement_id': VAR_1})\n",
"return VAR_11.fetchone()\n"
] |
[
"@app.route('/achievements/<achievement_id>')...\n",
"\"\"\"docstring\"\"\"\n",
"language = request.args.get('language', 'en')\n",
"region = request.args.get('region', 'US')\n",
"cursor = db.connection.cursor(db.pymysql.cursors.DictCursor)\n",
"cursor.execute(SELECT_ACHIEVEMENTS_QUERY +\n 'WHERE ach.id = %(achievement_id)s', {'language': language, 'region':\n region, 'achievement_id': achievement_id})\n",
"return cursor.fetchone()\n"
] |
[
0,
0,
0,
0,
0,
0,
0
] |
[
"Condition",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Return'"
] |
[
"def FUNC_39(VAR_101):...\n",
"VAR_101.norun = True\n",
"return VAR_101\n"
] |
[
"def decorate(ruleinfo):...\n",
"ruleinfo.norun = True\n",
"return ruleinfo\n"
] |
[
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Return'"
] |
[
"def FUNC_7(self):...\n",
"VAR_7 = BokChoyTestSuite('', default_store='invalid')\n",
"VAR_1 = 'tests'\n",
"self.assertEqual(VAR_7.cmd, self._expected_command(VAR_1=name, VAR_2='invalid')\n )\n"
] |
[
"def test_invalid_default_store(self):...\n",
"suite = BokChoyTestSuite('', default_store='invalid')\n",
"name = 'tests'\n",
"self.assertEqual(suite.cmd, self._expected_command(name=name, store='invalid'))\n"
] |
[
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"@defer.inlineCallbacks...\n",
"VAR_58 = []\n",
"yield self.mail_store.add_mailbox(VAR_31)\n",
"for VAR_63 in range(VAR_30):\n",
"VAR_82 = MailBuilder().with_status(VAR_32).with_tags(VAR_33).with_to(VAR_34\n ).with_cc(VAR_35).with_bcc(VAR_36)\n",
"defer.returnValue(VAR_58)\n",
"VAR_82.with_body(str(random.random()))\n",
"VAR_29 = VAR_82.build_input_mail()\n",
"VAR_68 = yield self.mail_store.add_mail(VAR_31, VAR_29.raw)\n",
"if VAR_33:\n",
"VAR_68.tags |= set(VAR_33)\n",
"if VAR_32:\n",
"for flag in VAR_32:\n",
"if VAR_33 or VAR_32:\n",
"VAR_68.flags.add(flag)\n",
"yield self.mail_store.update_mail(VAR_68)\n",
"VAR_58.append(VAR_68)\n"
] |
[
"@defer.inlineCallbacks...\n",
"mails = []\n",
"yield self.mail_store.add_mailbox(mailbox)\n",
"for _ in range(num):\n",
"builder = MailBuilder().with_status(flags).with_tags(tags).with_to(to).with_cc(\n cc).with_bcc(bcc)\n",
"defer.returnValue(mails)\n",
"builder.with_body(str(random.random()))\n",
"input_mail = builder.build_input_mail()\n",
"mail = yield self.mail_store.add_mail(mailbox, input_mail.raw)\n",
"if tags:\n",
"mail.tags |= set(tags)\n",
"if flags:\n",
"for flag in flags:\n",
"if tags or flags:\n",
"mail.flags.add(flag)\n",
"yield self.mail_store.update_mail(mail)\n",
"mails.append(mail)\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"Condition",
"Assign'",
"Expr'",
"For",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Condition",
"AugAssign'",
"Condition",
"For",
"Condition",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_4(self):...\n",
"self.client.login(username=self.tester.username, password='password')\n",
"VAR_1 = self.client.post(self.many_comments_url, {'run': [self.case_run_1.\n pk, self.case_run_2.pk]})\n",
"self.assertJSONEqual(str(VAR_1.content, encoding=settings.DEFAULT_CHARSET),\n {'rc': 1, 'response': 'Comments needed'})\n"
] |
[
"def test_refuse_if_missing_comment(self):...\n",
"self.client.login(username=self.tester.username, password='password')\n",
"response = self.client.post(self.many_comments_url, {'run': [self.\n case_run_1.pk, self.case_run_2.pk]})\n",
"self.assertJSONEqual(str(response.content, encoding=settings.\n DEFAULT_CHARSET), {'rc': 1, 'response': 'Comments needed'})\n"
] |
[
0,
0,
0,
0
] |
[
"FunctionDef'",
"Expr'",
"Assign'",
"Expr'"
] |
[
"def FUNC_2(VAR_0):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_18 = 'og' + VAR_0[2:]\n",
"print('FAILED TO RESTORE ORIGINAL')\n",
"db.engine.execute('DROP TABLE \"{0}\"'.format(VAR_0))\n",
"db.engine.execute('CREATE TABLE \"{0}\" AS SELECT * FROM \"{1}\"'.format(VAR_0,\n VAR_18))\n"
] |
[
"def restore_original(table_name):...\n",
"\"\"\"docstring\"\"\"\n",
"original = 'og' + table_name[2:]\n",
"print('FAILED TO RESTORE ORIGINAL')\n",
"db.engine.execute('DROP TABLE \"{0}\"'.format(table_name))\n",
"db.engine.execute('CREATE TABLE \"{0}\" AS SELECT * FROM \"{1}\"'.format(\n table_name, original))\n"
] |
[
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"@app.route('/experiment', methods=['GET'])...\n",
"VAR_1 = ProcessingForm(request.form)\n",
"return render_template('experiment/index.html', title='Try it Out!',\n sitekey=app.config['G_CAPTCHA_SITEKEY'], VAR_1=form, files=utils.\n SAMPLE_FILES)\n"
] |
[
"@app.route('/experiment', methods=['GET'])...\n",
"form = ProcessingForm(request.form)\n",
"return render_template('experiment/index.html', title='Try it Out!',\n sitekey=app.config['G_CAPTCHA_SITEKEY'], form=form, files=utils.\n SAMPLE_FILES)\n"
] |
[
0,
0,
2
] |
[
"Condition",
"Assign'",
"Return'"
] |
[
"def FUNC_1(self):...\n",
"VAR_16 = 0\n",
"VAR_17 = FUNC_0(CLASS_0.objects.filter(VAR_12=self.qa_group, VAR_8=False),\n self, 'next')\n",
"if VAR_17:\n",
"VAR_16 = VAR_17.pk\n",
"if VAR_17 == self:\n",
"VAR_16 = 0\n",
"return VAR_16\n"
] |
[
"def next_extracted_text_in_qa_group(self):...\n",
"nextid = 0\n",
"extextnext = get_next_or_prev(ExtractedText.objects.filter(qa_group=self.\n qa_group, qa_checked=False), self, 'next')\n",
"if extextnext:\n",
"nextid = extextnext.pk\n",
"if extextnext == self:\n",
"nextid = 0\n",
"return nextid\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Return'"
] |
[
"@mock.patch('requests.post', FUNC_0)...\n",
"VAR_4 = '/api/apps'\n",
"VAR_5 = self.client.post(VAR_4)\n",
"self.assertEqual(VAR_5.status_code, 201)\n",
"VAR_6 = VAR_5.data['id']\n",
"VAR_4 = '/api/apps/{app_id}/containers'.format(**locals())\n",
"VAR_5 = self.client.get(VAR_4)\n",
"self.assertEqual(VAR_5.status_code, 200)\n",
"self.assertEqual(len(VAR_5.data['results']), 0)\n",
"VAR_4 = '/api/apps/{app_id}/builds'.format(**locals())\n",
"VAR_8 = {'image': 'autotest/example', 'sha': 'a' * 40, 'procfile': json.\n dumps({'web': 'node server.js', 'worker': 'node worker.js'})}\n",
"VAR_5 = self.client.post(VAR_4, json.dumps(VAR_8), content_type=\n 'application/json')\n",
"self.assertEqual(VAR_5.status_code, 201)\n",
"VAR_4 = '/api/apps/{app_id}/scale'.format(**locals())\n",
"VAR_8 = {'web': 1}\n",
"VAR_5 = self.client.post(VAR_4, json.dumps(VAR_8), content_type=\n 'application/json')\n",
"self.assertEqual(VAR_5.status_code, 204)\n",
"VAR_4 = '/api/apps/{app_id}/containers'.format(**locals())\n",
"VAR_5 = self.client.get(VAR_4)\n",
"self.assertEqual(VAR_5.status_code, 200)\n",
"self.assertEqual(len(VAR_5.data['results']), 1)\n",
"self.assertEqual(VAR_5.data['results'][0]['release'], 'v2')\n",
"VAR_4 = '/api/apps/{app_id}/builds'.format(**locals())\n",
"VAR_8 = {'image': 'autotest/example'}\n",
"VAR_5 = self.client.post(VAR_4, json.dumps(VAR_8), content_type=\n 'application/json')\n",
"self.assertEqual(VAR_5.status_code, 201)\n",
"self.assertEqual(VAR_5.data['image'], VAR_8['image'])\n",
"VAR_4 = '/api/apps/{app_id}/containers'.format(**locals())\n",
"VAR_5 = self.client.get(VAR_4)\n",
"self.assertEqual(VAR_5.status_code, 200)\n",
"self.assertEqual(len(VAR_5.data['results']), 1)\n",
"self.assertEqual(VAR_5.data['results'][0]['release'], 'v3')\n",
"VAR_4 = '/api/apps/{app_id}/config'.format(**locals())\n",
"VAR_8 = {'values': json.dumps({'KEY': 'value'})}\n",
"VAR_5 = self.client.post(VAR_4, json.dumps(VAR_8), content_type=\n 'application/json')\n",
"self.assertEqual(VAR_5.status_code, 201)\n",
"VAR_4 = '/api/apps/{app_id}/containers'.format(**locals())\n",
"VAR_5 = self.client.get(VAR_4)\n",
"self.assertEqual(VAR_5.status_code, 200)\n",
"self.assertEqual(len(VAR_5.data['results']), 1)\n",
"self.assertEqual(VAR_5.data['results'][0]['release'], 'v4')\n"
] |
[
"@mock.patch('requests.post', mock_import_repository_task)...\n",
"url = '/api/apps'\n",
"response = self.client.post(url)\n",
"self.assertEqual(response.status_code, 201)\n",
"app_id = response.data['id']\n",
"url = '/api/apps/{app_id}/containers'.format(**locals())\n",
"response = self.client.get(url)\n",
"self.assertEqual(response.status_code, 200)\n",
"self.assertEqual(len(response.data['results']), 0)\n",
"url = '/api/apps/{app_id}/builds'.format(**locals())\n",
"body = {'image': 'autotest/example', 'sha': 'a' * 40, 'procfile': json.\n dumps({'web': 'node server.js', 'worker': 'node worker.js'})}\n",
"response = self.client.post(url, json.dumps(body), content_type=\n 'application/json')\n",
"self.assertEqual(response.status_code, 201)\n",
"url = '/api/apps/{app_id}/scale'.format(**locals())\n",
"body = {'web': 1}\n",
"response = self.client.post(url, json.dumps(body), content_type=\n 'application/json')\n",
"self.assertEqual(response.status_code, 204)\n",
"url = '/api/apps/{app_id}/containers'.format(**locals())\n",
"response = self.client.get(url)\n",
"self.assertEqual(response.status_code, 200)\n",
"self.assertEqual(len(response.data['results']), 1)\n",
"self.assertEqual(response.data['results'][0]['release'], 'v2')\n",
"url = '/api/apps/{app_id}/builds'.format(**locals())\n",
"body = {'image': 'autotest/example'}\n",
"response = self.client.post(url, json.dumps(body), content_type=\n 'application/json')\n",
"self.assertEqual(response.status_code, 201)\n",
"self.assertEqual(response.data['image'], body['image'])\n",
"url = '/api/apps/{app_id}/containers'.format(**locals())\n",
"response = self.client.get(url)\n",
"self.assertEqual(response.status_code, 200)\n",
"self.assertEqual(len(response.data['results']), 1)\n",
"self.assertEqual(response.data['results'][0]['release'], 'v3')\n",
"url = '/api/apps/{app_id}/config'.format(**locals())\n",
"body = {'values': json.dumps({'KEY': 'value'})}\n",
"response = self.client.post(url, json.dumps(body), content_type=\n 'application/json')\n",
"self.assertEqual(response.status_code, 201)\n",
"url = '/api/apps/{app_id}/containers'.format(**locals())\n",
"response = self.client.get(url)\n",
"self.assertEqual(response.status_code, 200)\n",
"self.assertEqual(len(response.data['results']), 1)\n",
"self.assertEqual(response.data['results'][0]['release'], 'v4')\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"Condition",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_21(self, VAR_14):...\n",
""
] |
[
"def is_relation_one_to_many(self, col_name):...\n",
""
] |
[
0,
0
] |
[
"FunctionDef'",
"Condition"
] |
[
"def FUNC_20(self, VAR_7):...\n",
"if '%(key)s' in VAR_7:\n",
"VAR_7 = VAR_7.replace('%(key)s', 'name')\n",
"return frappe.db.sql(VAR_7, as_dict=not self.as_list)\n"
] |
[
"def run_custom_query(self, query):...\n",
"if '%(key)s' in query:\n",
"query = query.replace('%(key)s', 'name')\n",
"return frappe.db.sql(query, as_dict=not self.as_list)\n"
] |
[
0,
0,
0,
0
] |
[
"FunctionDef'",
"Condition",
"Assign'",
"Return'"
] |
[
"def FUNC_6(VAR_10):...\n",
"self.assertEqual(VAR_10, VAR_11)\n"
] |
[
"def on_callback(response):...\n",
"self.assertEqual(response, magnet_link)\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Expr'"
] |
[
"def FUNC_3(VAR_14, VAR_15):...\n",
"return FUNC_6(VAR_14, [[VAR_16] for line in VAR_15 for VAR_16 in line], False)\n"
] |
[
"def vStrip(dims, files):...\n",
"return smartGrid(dims, [[file] for line in files for file in line], False)\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Return'"
] |
[
"def __init__(self, VAR_3=None):...\n",
"if VAR_3 is None:\n",
"VAR_3 = []\n",
"self.nodes = VAR_3\n"
] |
[
"def __init__(self, nodes=None):...\n",
"if nodes is None:\n",
"nodes = []\n",
"self.nodes = nodes\n"
] |
[
0,
0,
0,
0
] |
[
"FunctionDef'",
"Condition",
"Assign'",
"Assign'"
] |
[
"@utils.synchronized('3par', external=True)...\n",
"self.common.client_login()\n",
"self.common.create_snapshot(VAR_8)\n",
"self.common.client_logout()\n"
] |
[
"@utils.synchronized('3par', external=True)...\n",
"self.common.client_login()\n",
"self.common.create_snapshot(snapshot)\n",
"self.common.client_logout()\n"
] |
[
0,
0,
0,
0
] |
[
"Condition",
"Expr'",
"Expr'",
"Expr'"
] |
[
"\"\"\"Tests for the XSRF tool.\"\"\"\n",
"import datetime\n",
"import unittest\n",
"import config\n",
"import utils\n",
"\"\"\"Test cases for utils.XsrfTool.\"\"\"\n",
"VAR_0 = datetime.datetime(2010, 1, 31, 18, 0, 0)\n",
"def FUNC_0(self):...\n",
"utils.set_utcnow_for_test(CLASS_0.TEST_NOW)\n",
"def FUNC_1(self):...\n",
"\"\"\"docstring\"\"\"\n",
"config.set(xsrf_token_key='abcdef')\n",
"VAR_1 = utils.XsrfTool()\n",
"VAR_2 = VAR_1.generate_token(12345, 'test_action')\n",
"self.assertTrue(VAR_1.verify_token(VAR_2, 12345, 'test_action'))\n",
"def FUNC_2(self):...\n",
"\"\"\"docstring\"\"\"\n",
"config.set(xsrf_token_key='abcdef')\n",
"VAR_1 = utils.XsrfTool()\n",
"VAR_3 = utils.get_timestamp(CLASS_0.TEST_NOW)\n",
"self.assertFalse(VAR_1.verify_token('NotTheRightDigest/%f' % VAR_3, 12345,\n 'test_action'))\n",
"def FUNC_3(self):...\n",
"\"\"\"docstring\"\"\"\n",
"config.set(xsrf_token_key='abcdef')\n",
"VAR_1 = utils.XsrfTool()\n",
"VAR_2 = VAR_1.generate_token(12345, 'test_action')\n",
"utils.set_utcnow_for_test(CLASS_0.TEST_NOW + datetime.timedelta(hours=4,\n minutes=1))\n",
"self.assertFalse(VAR_1.verify_token(VAR_2, 12345, 'test_action'))\n",
"def FUNC_4(self):...\n",
"\"\"\"docstring\"\"\"\n",
"config.set(xsrf_token_key=None)\n",
"VAR_1 = utils.XsrfTool()\n",
"VAR_2 = VAR_1.generate_token(12345, 'test_action')\n",
"self.assertTrue(VAR_1.verify_token(VAR_2, 12345, 'test_action'))\n",
"def FUNC_5(self):...\n",
"\"\"\"docstring\"\"\"\n",
"config.set(xsrf_token_key=None)\n",
"VAR_1 = utils.XsrfTool()\n",
"VAR_3 = utils.get_timestamp(CLASS_0.TEST_NOW)\n",
"self.assertFalse(VAR_1.verify_token('NotTheRightDigest/%f' % VAR_3, 12345,\n 'test_action'))\n"
] |
[
"\"\"\"Tests for the XSRF tool.\"\"\"\n",
"import datetime\n",
"import unittest\n",
"import config\n",
"import utils\n",
"\"\"\"Test cases for utils.XsrfTool.\"\"\"\n",
"TEST_NOW = datetime.datetime(2010, 1, 31, 18, 0, 0)\n",
"def setUp(self):...\n",
"utils.set_utcnow_for_test(XsrfToolTests.TEST_NOW)\n",
"def test_gen_and_verify_good_token(self):...\n",
"\"\"\"docstring\"\"\"\n",
"config.set(xsrf_token_key='abcdef')\n",
"tool = utils.XsrfTool()\n",
"token = tool.generate_token(12345, 'test_action')\n",
"self.assertTrue(tool.verify_token(token, 12345, 'test_action'))\n",
"def test_rejects_invalid_token(self):...\n",
"\"\"\"docstring\"\"\"\n",
"config.set(xsrf_token_key='abcdef')\n",
"tool = utils.XsrfTool()\n",
"timestamp = utils.get_timestamp(XsrfToolTests.TEST_NOW)\n",
"self.assertFalse(tool.verify_token('NotTheRightDigest/%f' % timestamp, \n 12345, 'test_action'))\n",
"def test_rejects_expired_token(self):...\n",
"\"\"\"docstring\"\"\"\n",
"config.set(xsrf_token_key='abcdef')\n",
"tool = utils.XsrfTool()\n",
"token = tool.generate_token(12345, 'test_action')\n",
"utils.set_utcnow_for_test(XsrfToolTests.TEST_NOW + datetime.timedelta(hours\n =4, minutes=1))\n",
"self.assertFalse(tool.verify_token(token, 12345, 'test_action'))\n",
"def test_good_with_no_prior_key(self):...\n",
"\"\"\"docstring\"\"\"\n",
"config.set(xsrf_token_key=None)\n",
"tool = utils.XsrfTool()\n",
"token = tool.generate_token(12345, 'test_action')\n",
"self.assertTrue(tool.verify_token(token, 12345, 'test_action'))\n",
"def test_bad_with_no_prior_key(self):...\n",
"\"\"\"docstring\"\"\"\n",
"config.set(xsrf_token_key=None)\n",
"tool = utils.XsrfTool()\n",
"timestamp = utils.get_timestamp(XsrfToolTests.TEST_NOW)\n",
"self.assertFalse(tool.verify_token('NotTheRightDigest/%f' % timestamp, \n 12345, 'test_action'))\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"Expr'",
"Import'",
"Import'",
"Import'",
"Import'",
"Expr'",
"Assign'",
"FunctionDef'",
"Expr'",
"FunctionDef'",
"Docstring",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"FunctionDef'",
"Docstring",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"FunctionDef'",
"Docstring",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"FunctionDef'",
"Docstring",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"FunctionDef'",
"Docstring",
"Expr'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"from __future__ import absolute_import\n",
"from __future__ import print_function\n",
"from __future__ import unicode_literals\n",
"import re\n",
"from bs4 import BeautifulSoup\n",
"from cmstestsuite.web import GenericRequest, LoginRequest\n",
"def FUNC_0(self):...\n",
"if not LoginRequest.test_success(self):\n",
"return False\n",
"VAR_4 = re.compile('Failed to log in.')\n",
"if VAR_4.search(self.res_data) is not None:\n",
"return False\n",
"VAR_5 = re.compile(self.username)\n",
"if VAR_5.search(self.res_data) is None:\n",
"return False\n",
"return True\n"
] |
[
"from __future__ import absolute_import\n",
"from __future__ import print_function\n",
"from __future__ import unicode_literals\n",
"import re\n",
"from bs4 import BeautifulSoup\n",
"from cmstestsuite.web import GenericRequest, LoginRequest\n",
"def test_success(self):...\n",
"if not LoginRequest.test_success(self):\n",
"return False\n",
"fail_re = re.compile('Failed to log in.')\n",
"if fail_re.search(self.res_data) is not None:\n",
"return False\n",
"username_re = re.compile(self.username)\n",
"if username_re.search(self.res_data) is None:\n",
"return False\n",
"return True\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"FunctionDef'",
"Condition",
"Return'",
"Assign'",
"Condition",
"Return'",
"Assign'",
"Condition",
"Return'",
"Return'"
] |
[
"def FUNC_9():...\n",
"for filename in ['cert_db.sqlite3', 'rf_model.pkl', 'rf_features.pkl']:\n",
"os.rename('temp_' + filename, filename)\n",
"os.remove(filename)\n"
] |
[
"def tearDownClass():...\n",
"for filename in ['cert_db.sqlite3', 'rf_model.pkl', 'rf_features.pkl']:\n",
"os.rename('temp_' + filename, filename)\n",
"os.remove(filename)\n"
] |
[
0,
0,
0,
0
] |
[
"FunctionDef'",
"For",
"Expr'",
"Expr'"
] |
[
"def FUNC_19(self, VAR_9=True):...\n",
"\"\"\"docstring\"\"\"\n",
"self._kill_process_type(VAR_13.PROCESS_TYPE_LOG_MONITOR, VAR_9=check_alive)\n"
] |
[
"def kill_log_monitor(self, check_alive=True):...\n",
"\"\"\"docstring\"\"\"\n",
"self._kill_process_type(ray_constants.PROCESS_TYPE_LOG_MONITOR, check_alive\n =check_alive)\n"
] |
[
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Expr'"
] |
[
"def FUNC_14(self, VAR_21, VAR_11):...\n",
"if VAR_0(VAR_11):\n",
"VAR_37, VAR_41 = self.process_lhs(VAR_21, VAR_11)\n",
"VAR_42, VAR_43 = self.process_rhs(VAR_21, VAR_11)\n",
"assert len(VAR_43) == 1, _('A dictionary must be provided as argument')\n",
"VAR_4, *VAR_44 = VAR_43\n",
"return '%s %s %s::jsonb' % (VAR_37, self.lookup_operator, VAR_42), [json_encode\n (VAR_4)]\n"
] |
[
"def as_sql(self, compiler, connection):...\n",
"if is_postgresql(connection):\n",
"lhs, lhs_params = self.process_lhs(compiler, connection)\n",
"rhs, rhs_params = self.process_rhs(compiler, connection)\n",
"assert len(rhs_params) == 1, _('A dictionary must be provided as argument')\n",
"value, *junk = rhs_params\n",
"return '%s %s %s::jsonb' % (lhs, self.lookup_operator, rhs), [json_encode(\n value)]\n"
] |
[
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Condition",
"Assign'",
"Assign'",
"Assert'",
"Assign'",
"Return'"
] |
[
"def FUNC_5(self):...\n",
"return reverse(self.success_url, VAR_4={'step': 0, 'uuid': self.object.uuid})\n"
] |
[
"def get_success_url(self):...\n",
"return reverse(self.success_url, kwargs={'step': 0, 'uuid': self.object.uuid})\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Return'"
] |
[
"def FUNC_13(self):...\n",
"self.run_test_case(self.scenario.upsert_load_balancer())\n"
] |
[
"def test_a_upsert_load_balancer(self):...\n",
"self.run_test_case(self.scenario.upsert_load_balancer())\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Expr'"
] |
[
"def FUNC_4(VAR_4):...\n",
"VAR_9 = FUNC_0()\n",
"VAR_10 = VAR_9.cursor()\n",
"VAR_10.execute(\" SELECT karma FROM people WHERE name='{}' \".format(VAR_4))\n",
"VAR_2.error('Execution failed with error: {}'.format(e))\n",
"VAR_11 = VAR_10.fetchone()\n",
"if VAR_11 is None:\n",
"VAR_2.debug('No karma found for name {}'.format(VAR_4))\n",
"VAR_11 = VAR_11[0]\n",
"VAR_9.close()\n",
"VAR_2.debug('karma of {} found for name {}'.format(VAR_11, VAR_4))\n",
"return VAR_11\n",
"VAR_9.close()\n",
"return VAR_11\n"
] |
[
"def karma_ask(name):...\n",
"db = db_connect()\n",
"cursor = db.cursor()\n",
"cursor.execute(\" SELECT karma FROM people WHERE name='{}' \".format(name))\n",
"logger.error('Execution failed with error: {}'.format(e))\n",
"karma = cursor.fetchone()\n",
"if karma is None:\n",
"logger.debug('No karma found for name {}'.format(name))\n",
"karma = karma[0]\n",
"db.close()\n",
"logger.debug('karma of {} found for name {}'.format(karma, name))\n",
"return karma\n",
"db.close()\n",
"return karma\n"
] |
[
0,
0,
0,
4,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Condition",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Return'",
"Expr'",
"Return'"
] |
[
"@VAR_0.route('/quiz')...\n",
"VAR_13 = request.args.get('varga')\n",
"VAR_14 = []\n",
"con.close()\n",
"con.row_factory = sql.Row\n",
"VAR_15 = con.cursor()\n",
"VAR_15.execute('string' % VAR_13)\n",
"VAR_14 = VAR_15.fetchall()\n",
"VAR_18 = VAR_14[0]['artha']\n",
"VAR_15.execute(\n \"select pada from pada where varga = '%s' and artha = '%s' order by id\" %\n (VAR_13, VAR_18))\n",
"VAR_19 = VAR_15.fetchall()\n",
"return render_template('quiz.html', VAR_14=rows, VAR_19=paryaya, VAR_13=varga)\n"
] |
[
"@app.route('/quiz')...\n",
"varga = request.args.get('varga')\n",
"rows = []\n",
"con.close()\n",
"con.row_factory = sql.Row\n",
"cur = con.cursor()\n",
"cur.execute(\n \"select * from pada inner join mula on pada.sloka_line = mula.sloka_line where pada.varga = '%s' order by random() limit 1;\"\n % varga)\n",
"rows = cur.fetchall()\n",
"artha = rows[0]['artha']\n",
"cur.execute(\n \"select pada from pada where varga = '%s' and artha = '%s' order by id\" %\n (varga, artha))\n",
"paryaya = cur.fetchall()\n",
"return render_template('quiz.html', rows=rows, paryaya=paryaya, varga=varga)\n"
] |
[
0,
0,
0,
0,
0,
0,
4,
0,
0,
4,
0,
0
] |
[
"Condition",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Return'"
] |
[
"def FUNC_16(self):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_2 = {'SEVERITY': {'LOW': 3}, 'CONFIDENCE': {'HIGH': 3}}\n",
"self.check_example('imports-from.py', VAR_2)\n"
] |
[
"def test_imports_from(self):...\n",
"\"\"\"docstring\"\"\"\n",
"expect = {'SEVERITY': {'LOW': 3}, 'CONFIDENCE': {'HIGH': 3}}\n",
"self.check_example('imports-from.py', expect)\n"
] |
[
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'"
] |
[
"def FUNC_0(self, VAR_3, VAR_4, VAR_5):...\n",
"\"\"\"docstring\"\"\"\n",
"if VAR_4.is_course_staff:\n",
"return True\n",
"if not VAR_5.visible_to_students:\n",
"self.error_msg(_('The resource is not currently visible.'))\n",
"VAR_6 = VAR_3.user\n",
"return False\n",
"VAR_11 = VAR_5.view_content_to\n",
"VAR_12 = VAR_5.VIEW_ACCESS\n",
"if VAR_11 != VAR_12.PUBLIC:\n",
"if not VAR_6.is_authenticated():\n",
"return True\n",
"self.error_msg(_('This course is not open for public.'))\n",
"if VAR_4.get_access_mode() == ACCESS.ENROLL:\n",
"return False\n",
"return self.enrollment_audience_check(VAR_3, VAR_5, VAR_6)\n",
"if VAR_11 == VAR_12.ENROLLED:\n",
"if not VAR_5.is_student(VAR_6):\n",
"if VAR_11 == VAR_12.ENROLLMENT_AUDIENCE:\n",
"self.error_msg(_('Only enrolled students shall pass.'))\n",
"return self.enrollment_audience_check(VAR_3, VAR_5, VAR_6)\n",
"return False\n"
] |
[
"def is_object_visible(self, request, view, course):...\n",
"\"\"\"docstring\"\"\"\n",
"if view.is_course_staff:\n",
"return True\n",
"if not course.visible_to_students:\n",
"self.error_msg(_('The resource is not currently visible.'))\n",
"user = request.user\n",
"return False\n",
"show_for = course.view_content_to\n",
"VA = course.VIEW_ACCESS\n",
"if show_for != VA.PUBLIC:\n",
"if not user.is_authenticated():\n",
"return True\n",
"self.error_msg(_('This course is not open for public.'))\n",
"if view.get_access_mode() == ACCESS.ENROLL:\n",
"return False\n",
"return self.enrollment_audience_check(request, course, user)\n",
"if show_for == VA.ENROLLED:\n",
"if not course.is_student(user):\n",
"if show_for == VA.ENROLLMENT_AUDIENCE:\n",
"self.error_msg(_('Only enrolled students shall pass.'))\n",
"return self.enrollment_audience_check(request, course, user)\n",
"return False\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
4,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Condition",
"Return'",
"Condition",
"Expr'",
"Assign'",
"Return'",
"Assign'",
"Assign'",
"Condition",
"Condition",
"Return'",
"Expr'",
"Condition",
"Return'",
"Return'",
"Condition",
"Condition",
"Condition",
"Expr'",
"Return'",
"Return'"
] |
[
"def FUNC_6(self, VAR_18, VAR_13=None):...\n",
"\"\"\"docstring\"\"\"\n",
"if not VAR_18.isdigit():\n",
"return CLASS_0.INVALID_ID\n",
"VAR_19 = self.__con.cursor()\n",
"if VAR_13 == None:\n",
"VAR_13 = 'NULL'\n",
"VAR_23 = CLASS_0.__format_date(VAR_13)\n",
"return CLASS_0.SUCCESS if VAR_19.execute(\n 'UPDATE tasks SET due_date=%s WHERE taskid=%s' % (VAR_13, int(VAR_18))\n ) else CLASS_0.DOES_NOT_EXIST\n",
"if VAR_23 == CLASS_0.INVALID_DATE:\n",
"return VAR_23\n",
"VAR_13 = \"'%s'\" % VAR_23\n"
] |
[
"def update_date(self, taskid, date=None):...\n",
"\"\"\"docstring\"\"\"\n",
"if not taskid.isdigit():\n",
"return Database.INVALID_ID\n",
"cur = self.__con.cursor()\n",
"if date == None:\n",
"date = 'NULL'\n",
"valid_date = Database.__format_date(date)\n",
"return Database.SUCCESS if cur.execute(\n 'UPDATE tasks SET due_date=%s WHERE taskid=%s' % (date, int(taskid))\n ) else Database.DOES_NOT_EXIST\n",
"if valid_date == Database.INVALID_DATE:\n",
"return valid_date\n",
"date = \"'%s'\" % valid_date\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
2,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Condition",
"Return'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Return'",
"Condition",
"Return'",
"Assign'"
] |
[
"def FUNC_24(VAR_5):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_40 = [VAR_5['name']]\n",
"VAR_27 = VAR_5['holders']\n",
"for holder_no, holder in enumerate(VAR_27):\n",
"VAR_43 = VAR_33[min(len(VAR_27) - (holder_no + 1), 1)]\n",
"return VAR_40\n",
"VAR_44 = FUNC_24(holder)\n",
"for line_no, line in enumerate(VAR_44):\n",
"VAR_40.append(VAR_43[min(line_no, 1)] + line)\n"
] |
[
"def format_tree(tree):...\n",
"\"\"\"docstring\"\"\"\n",
"result = [tree['name']]\n",
"holders = tree['holders']\n",
"for holder_no, holder in enumerate(holders):\n",
"spacer_style = spacers[min(len(holders) - (holder_no + 1), 1)]\n",
"return result\n",
"subtree_lines = format_tree(holder)\n",
"for line_no, line in enumerate(subtree_lines):\n",
"result.append(spacer_style[min(line_no, 1)] + line)\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"For",
"Assign'",
"Return'",
"Assign'",
"For",
"Expr'"
] |
[
"def FUNC_6(self, VAR_14):...\n",
"if PY2:\n",
"return {VAR_40: k for k, VAR_40 in VAR_14.iteritems()}\n",
"return {VAR_40: k for k, VAR_40 in VAR_14.items()}\n"
] |
[
"def invert(self, orig):...\n",
"if PY2:\n",
"return {v: k for k, v in orig.iteritems()}\n",
"return {v: k for k, v in orig.items()}\n"
] |
[
0,
0,
0,
0
] |
[
"FunctionDef'",
"Condition",
"Return'",
"Return'"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return [os.path.join(self.bin_dir_path, self.executable)] + self.args\n"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return [os.path.join(self.bin_dir_path, self.executable)] + self.args\n"
] |
[
0,
0,
2
] |
[
"Condition",
"Docstring",
"Return'"
] |
[
"def FUNC_9(self):...\n",
"if not VAR_0:\n",
"self.redirect('/')\n",
"self.write('string'.format(self.xsrf_form_html()))\n"
] |
[
"def get(self):...\n",
"if not enable_authentication:\n",
"self.redirect('/')\n",
"self.write(\n '<html><head><title>Liked Saved Downloader</title><link rel=\"stylesheet\" type=\"text/css\" href=\"webInterfaceNoAuth/index.css\"></head><body><h1>Login Required</h1><form action=\"/login\" method=\"post\">Name: <input type=\"text\" name=\"name\"><br />Password: <input type=\"password\" name=\"password\">{}<br /><input type=\"submit\" value=\"Sign in\"></form></body></html>'\n .format(self.xsrf_form_html()))\n"
] |
[
0,
0,
0,
0
] |
[
"FunctionDef'",
"Condition",
"Expr'",
"Expr'"
] |
[
"def FUNC_8(self, VAR_9, VAR_10=None, VAR_11=None):...\n",
"get_and_check_project(VAR_9, VAR_11)\n",
"VAR_28 = self.queryset.get(VAR_10=pk, VAR_1=project_pk)\n",
"VAR_26 = CLASS_1(VAR_28)\n",
"return Response(VAR_26.data)\n"
] |
[
"def retrieve(self, request, pk=None, project_pk=None):...\n",
"get_and_check_project(request, project_pk)\n",
"task = self.queryset.get(pk=pk, project=project_pk)\n",
"serializer = TaskSerializer(task)\n",
"return Response(serializer.data)\n"
] |
[
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Expr'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_0(self, VAR_1, VAR_2, VAR_3=None, VAR_4=None, VAR_5=0):...\n",
"VAR_3 = VAR_3 or [NTFY_UPDATE, NTFY_INSERT, NTFY_DELETE]\n",
"\"\"\"string\"\"\"\n",
"assert isinstance(VAR_3, list)\n",
"assert VAR_2 in self.SUBJECTS, 'Subject %s not in SUBJECTS' % VAR_2\n",
"VAR_9 = VAR_1, VAR_2, VAR_3, VAR_4, VAR_5\n",
"self.observerLock.acquire()\n",
"self.observers.append(VAR_9)\n",
"self.observerLock.release()\n"
] |
[
"def add_observer(self, func, subject, changeTypes=None, id=None, cache=0):...\n",
"changeTypes = changeTypes or [NTFY_UPDATE, NTFY_INSERT, NTFY_DELETE]\n",
"\"\"\"\n Add observer function which will be called upon certain event\n Example:\n addObserver(NTFY_TORRENTS, [NTFY_INSERT,NTFY_DELETE]) -> get callbacks\n when peers are added or deleted\n addObserver(NTFY_TORRENTS, [NTFY_SEARCH_RESULT], 'a_search_id') -> get\n callbacks when peer-searchresults of of search\n with id=='a_search_id' come in\n \"\"\"\n",
"assert isinstance(changeTypes, list)\n",
"assert subject in self.SUBJECTS, 'Subject %s not in SUBJECTS' % subject\n",
"obs = func, subject, changeTypes, id, cache\n",
"self.observerLock.acquire()\n",
"self.observers.append(obs)\n",
"self.observerLock.release()\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Expr'",
"Assert'",
"Assert'",
"Assign'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_2(VAR_2: HttpRequest):...\n",
"\"\"\"docstring\"\"\"\n",
"u: Profile = get_current_user(VAR_2)\n",
"VAR_5 = '/admin?success'\n",
"if u.rights > 0:\n",
"VAR_5 = '/admin/reservations'\n",
"if VAR_2.GET.get('redirect'):\n",
"VAR_5 = VAR_2.GET['redirect']\n",
"if 'payload' not in VAR_2.GET:\n",
"return HttpResponseRedirect('/admin?error=No%20id%20provided')\n",
"VAR_4 = GroupReservation.objects.get(id=int(request.GET['payload']))\n",
"if VAR_4.createdByUser != u and u.rights < 2:\n",
"return HttpResponseRedirect('/admin?error=noyb')\n",
"VAR_4.submitted = True\n",
"VAR_4.save()\n",
"res: HttpResponseRedirect = HttpResponseRedirect(VAR_5)\n",
"return res\n"
] |
[
"def write_db_reservation_action(request: HttpRequest):...\n",
"\"\"\"docstring\"\"\"\n",
"u: Profile = get_current_user(request)\n",
"forward_url = '/admin?success'\n",
"if u.rights > 0:\n",
"forward_url = '/admin/reservations'\n",
"if request.GET.get('redirect'):\n",
"forward_url = request.GET['redirect']\n",
"if 'payload' not in request.GET:\n",
"return HttpResponseRedirect('/admin?error=No%20id%20provided')\n",
"current_reservation = GroupReservation.objects.get(id=int(request.GET[\n 'payload']))\n",
"if current_reservation.createdByUser != u and u.rights < 2:\n",
"return HttpResponseRedirect('/admin?error=noyb')\n",
"current_reservation.submitted = True\n",
"current_reservation.save()\n",
"res: HttpResponseRedirect = HttpResponseRedirect(forward_url)\n",
"return res\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"AnnAssign'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Return'",
"Assign'",
"Condition",
"Return'",
"Assign'",
"Expr'",
"AnnAssign'",
"Return'"
] |
[
"def __init__(self, *VAR_12, VAR_13=VAR_1, VAR_14=dict, **VAR_4):...\n",
"self._sep = VAR_13\n",
"self._data = {}\n",
"self._create_on_missing = FUNC_1(VAR_14)\n",
"self.update(*VAR_12, **kwargs)\n"
] |
[
"def __init__(self, *args, path_separator=DEFAULT_PATH_SEPARATOR,...\n",
"self._sep = path_separator\n",
"self._data = {}\n",
"self._create_on_missing = create_on_missing(path_factory)\n",
"self.update(*args, **kwargs)\n"
] |
[
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_2(self):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_4 = 'katotestlb' + self.test_id\n",
"self.__use_lb_name = VAR_4\n",
"VAR_2 = self.bindings\n",
"VAR_5 = VAR_2['TEST_AWS_REGION']\n",
"VAR_6 = [VAR_5 + 'a', VAR_5 + 'b']\n",
"VAR_7 = {'Listener': {'InstancePort': 7001, 'LoadBalancerPort': 80}}\n",
"VAR_8 = {'HealthyThreshold': 8, 'UnhealthyThreshold': 3, 'Interval': 123,\n 'Timeout': 12, 'Target': 'HTTP:%d/healthcheck' % VAR_7['Listener'][\n 'InstancePort']}\n",
"VAR_9 = self.agent.type_to_payload('upsertAmazonLoadBalancerDescription', {\n 'credentials': VAR_2['AWS_CREDENTIALS'], 'clusterName': VAR_2[\n 'TEST_APP'], 'name': VAR_4, 'availabilityZones': {VAR_5: VAR_6},\n 'listeners': [{'internalProtocol': 'HTTP', 'internalPort': VAR_7[\n 'Listener']['InstancePort'], 'externalProtocol': 'HTTP', 'externalPort':\n VAR_7['Listener']['LoadBalancerPort']}], 'healthCheck': VAR_8['Target'],\n 'healthTimeout': VAR_8['Timeout'], 'healthInterval': VAR_8['Interval'],\n 'healthyThreshold': VAR_8['HealthyThreshold'], 'unhealthyThreshold':\n VAR_8['UnhealthyThreshold']})\n",
"VAR_10 = aws.AwsContractBuilder(self.aws_observer)\n",
"VAR_10.new_clause_builder('Load Balancer Added', retryable_for_secs=30\n ).collect_resources(aws_module='elb', command='describe-load-balancers',\n args=['--load-balancer-names', self.__use_lb_name]).contains_pred_list([\n jc.PathContainsPredicate('LoadBalancerDescriptions/HealthCheck', VAR_8),\n jc.PathPredicate('LoadBalancerDescriptions/AvailabilityZones', jc.\n LIST_SIMILAR(VAR_6)), jc.PathElementsContainPredicate(\n 'LoadBalancerDescriptions/ListenerDescriptions', VAR_7)])\n",
"return st.OperationContract(self.new_post_operation(title=\n 'upsert_amazon_load_balancer', data=payload, path='ops'), contract=\n builder.build())\n"
] |
[
"def upsert_load_balancer(self):...\n",
"\"\"\"docstring\"\"\"\n",
"detail_raw_name = 'katotestlb' + self.test_id\n",
"self.__use_lb_name = detail_raw_name\n",
"bindings = self.bindings\n",
"region = bindings['TEST_AWS_REGION']\n",
"avail_zones = [region + 'a', region + 'b']\n",
"listener = {'Listener': {'InstancePort': 7001, 'LoadBalancerPort': 80}}\n",
"health_check = {'HealthyThreshold': 8, 'UnhealthyThreshold': 3, 'Interval':\n 123, 'Timeout': 12, 'Target': 'HTTP:%d/healthcheck' % listener[\n 'Listener']['InstancePort']}\n",
"payload = self.agent.type_to_payload('upsertAmazonLoadBalancerDescription',\n {'credentials': bindings['AWS_CREDENTIALS'], 'clusterName': bindings[\n 'TEST_APP'], 'name': detail_raw_name, 'availabilityZones': {region:\n avail_zones}, 'listeners': [{'internalProtocol': 'HTTP', 'internalPort':\n listener['Listener']['InstancePort'], 'externalProtocol': 'HTTP',\n 'externalPort': listener['Listener']['LoadBalancerPort']}],\n 'healthCheck': health_check['Target'], 'healthTimeout': health_check[\n 'Timeout'], 'healthInterval': health_check['Interval'],\n 'healthyThreshold': health_check['HealthyThreshold'],\n 'unhealthyThreshold': health_check['UnhealthyThreshold']})\n",
"builder = aws.AwsContractBuilder(self.aws_observer)\n",
"builder.new_clause_builder('Load Balancer Added', retryable_for_secs=30\n ).collect_resources(aws_module='elb', command='describe-load-balancers',\n args=['--load-balancer-names', self.__use_lb_name]).contains_pred_list([\n jc.PathContainsPredicate('LoadBalancerDescriptions/HealthCheck',\n health_check), jc.PathPredicate(\n 'LoadBalancerDescriptions/AvailabilityZones', jc.LIST_SIMILAR(\n avail_zones)), jc.PathElementsContainPredicate(\n 'LoadBalancerDescriptions/ListenerDescriptions', listener)])\n",
"return st.OperationContract(self.new_post_operation(title=\n 'upsert_amazon_load_balancer', data=payload, path='ops'), contract=\n builder.build())\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Return'"
] |
[
"def FUNC_0(self):...\n",
"VAR_0.debug('Retrieving connection from the pool')\n",
"VAR_11 = self._pool.get()\n",
"VAR_0.debug('Type checking connection')\n",
"if not isinstance(VAR_11, MySQLdb.connections.Connection):\n",
"return -1\n",
"VAR_0.info('Successful MySQL connection get request')\n",
"return VAR_11\n"
] |
[
"def get_connection(self):...\n",
"logger.debug('Retrieving connection from the pool')\n",
"db = self._pool.get()\n",
"logger.debug('Type checking connection')\n",
"if not isinstance(db, MySQLdb.connections.Connection):\n",
"return -1\n",
"logger.info('Successful MySQL connection get request')\n",
"return db\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Expr'",
"Assign'",
"Expr'",
"Condition",
"Return'",
"Expr'",
"Return'"
] |
[
"def FUNC_5(VAR_1='', VAR_2=''):...\n",
"VAR_5 = tournament.registerPlayer(VAR_1=player_name, VAR_2=country)\n",
"return VAR_5\n"
] |
[
"def dummy_player(player_name='', country=''):...\n",
"s = tournament.registerPlayer(player_name=player_name, country=country)\n",
"return s\n"
] |
[
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Return'"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self._node_ip_address\n"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self._node_ip_address\n"
] |
[
0,
0,
0
] |
[
"Condition",
"Docstring",
"Return'"
] |
[
"@property...\n",
"return self.step.project if self.step else None\n"
] |
[
"@property...\n",
"return self.step.project if self.step else None\n"
] |
[
0,
0
] |
[
"Condition",
"Return'"
] |
[
"@utils.synchronized('3par', external=True)...\n",
"self.common.client_login()\n",
"VAR_16 = self.common.get_volume_stats(VAR_4)\n",
"VAR_16['storage_protocol'] = 'FC'\n",
"VAR_17 = self.configuration.safe_get('volume_backend_name')\n",
"VAR_16['volume_backend_name'] = VAR_17 or self.__class__.__name__\n",
"self.common.client_logout()\n",
"return VAR_16\n"
] |
[
"@utils.synchronized('3par', external=True)...\n",
"self.common.client_login()\n",
"stats = self.common.get_volume_stats(refresh)\n",
"stats['storage_protocol'] = 'FC'\n",
"backend_name = self.configuration.safe_get('volume_backend_name')\n",
"stats['volume_backend_name'] = backend_name or self.__class__.__name__\n",
"self.common.client_logout()\n",
"return stats\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"Condition",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Return'"
] |
[
"def __init__(self, *VAR_1, **VAR_2):...\n",
"super(CLASS_0, self).__init__(*VAR_1, **kwargs)\n"
] |
[
"def __init__(self, *args, **kwargs):...\n",
"super(TimestampField, self).__init__(*args, **kwargs)\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Expr'"
] |
[
"def FUNC_24(self):...\n",
"VAR_13 = 1234\n",
"self.cursor.execute('create table t1(n int)')\n",
"self.cursor.execute('insert into t1 values (?)', VAR_13)\n",
"VAR_20 = self.cursor.execute('select n from t1').fetchone()[0]\n",
"self.assertEqual(VAR_20, VAR_13)\n"
] |
[
"def test_int(self):...\n",
"value = 1234\n",
"self.cursor.execute('create table t1(n int)')\n",
"self.cursor.execute('insert into t1 values (?)', value)\n",
"result = self.cursor.execute('select n from t1').fetchone()[0]\n",
"self.assertEqual(result, value)\n"
] |
[
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Expr'"
] |
[
"def FUNC_18(self):...\n",
"VAR_47 = Sanic(__name__)\n",
"VAR_48 = str(VAR_0 / 'datasette' / 'templates')\n",
"VAR_49 = []\n",
"if self.template_dir:\n",
"VAR_49.append(self.template_dir)\n",
"VAR_49.extend([plugin['templates_path'] for plugin in get_plugins(pm) if\n plugin['templates_path']])\n",
"VAR_49.append(VAR_48)\n",
"VAR_50 = ChoiceLoader([FileSystemLoader(VAR_49), PrefixLoader({'default':\n FileSystemLoader(VAR_48)}, delimiter=':')])\n",
"self.jinja_env = Environment(loader=template_loader, autoescape=True)\n",
"self.jinja_env.filters['escape_css_string'] = escape_css_string\n",
"self.jinja_env.filters['quote_plus'] = lambda u: urllib.parse.quote_plus(u)\n",
"self.jinja_env.filters['escape_sqlite'] = escape_sqlite\n",
"self.jinja_env.filters['to_css_class'] = to_css_class\n",
"pm.hook.prepare_jinja2_environment(env=self.jinja_env)\n",
"VAR_47.add_route(IndexView.as_view(self), '/<as_format:(\\\\.jsono?)?$>')\n",
"VAR_47.add_route(FUNC_0, '/favicon.ico')\n",
"VAR_47.static('/-/static/', str(VAR_0 / 'datasette' / 'static'))\n",
"for VAR_61, dirname in self.static_mounts:\n",
"VAR_47.static(VAR_61, dirname)\n",
"for plugin in get_plugins(pm):\n",
"if plugin['static_path']:\n",
"VAR_47.add_route(JsonDataView.as_view(self, 'inspect.json', self.inspect),\n '/-/inspect<as_format:(\\\\.json)?$>')\n",
"VAR_65 = '/-/static-plugins/{}/'.format(plugin['name'])\n",
"VAR_47.add_route(JsonDataView.as_view(self, 'metadata.json', lambda : self.\n _metadata), '/-/metadata<as_format:(\\\\.json)?$>')\n",
"VAR_47.static(VAR_65, plugin['static_path'])\n",
"VAR_47.add_route(JsonDataView.as_view(self, 'versions.json', self.versions),\n '/-/versions<as_format:(\\\\.json)?$>')\n",
"VAR_47.add_route(JsonDataView.as_view(self, 'plugins.json', self.plugins),\n '/-/plugins<as_format:(\\\\.json)?$>')\n",
"VAR_47.add_route(JsonDataView.as_view(self, 'config.json', lambda : self.\n _config), '/-/config<as_format:(\\\\.json)?$>')\n",
"VAR_47.add_route(DatabaseDownload.as_view(self),\n '/<db_name:[^/]+?><as_db:(\\\\.db)$>')\n",
"VAR_47.add_route(DatabaseView.as_view(self),\n '/<db_name:[^/]+?><as_format:(\\\\.jsono?|\\\\.csv)?$>')\n",
"VAR_47.add_route(TableView.as_view(self),\n '/<db_name:[^/]+>/<table_and_format:[^/]+?$>')\n",
"VAR_47.add_route(RowView.as_view(self),\n '/<db_name:[^/]+>/<table:[^/]+?>/<pk_path:[^/]+?><as_format:(\\\\.jsono?)?$>'\n )\n",
"self.register_custom_units()\n",
"@VAR_47.middleware('response')...\n",
"if VAR_51.status == 404 and VAR_6.path.endswith('/'):\n",
"VAR_61 = VAR_6.path.rstrip('/')\n",
"@VAR_47.exception(Exception)...\n",
"if VAR_6.query_string:\n",
"VAR_56 = None\n",
"VAR_61 = '{}?{}'.format(VAR_61, VAR_6.query_string)\n",
"return response.redirect(VAR_61)\n",
"VAR_57 = None\n",
"if isinstance(VAR_52, NotFound):\n",
"VAR_66 = 404\n",
"if isinstance(VAR_52, InvalidUsage):\n",
"VAR_64 = {}\n",
"VAR_66 = 405\n",
"if isinstance(VAR_52, DatasetteError):\n",
"VAR_67 = VAR_52.args[0]\n",
"VAR_64 = {}\n",
"VAR_66 = VAR_52.status\n",
"VAR_66 = 500\n",
"VAR_58 = ['500.html']\n",
"VAR_67 = VAR_52.args[0]\n",
"VAR_64 = VAR_52.error_dict\n",
"VAR_64 = {}\n",
"if VAR_66 != 500:\n",
"VAR_67 = VAR_52.message\n",
"VAR_67 = str(VAR_52)\n",
"VAR_58 = ['{}.html'.format(VAR_66)] + VAR_58\n",
"VAR_64.update({'ok': False, 'error': VAR_67, 'status': VAR_66, 'title': VAR_56}\n )\n",
"if VAR_52.messagge_is_html:\n",
"traceback.print_exc()\n",
"if VAR_6 is not None and VAR_6.path.split('?')[0].endswith('.json'):\n",
"VAR_67 = Markup(VAR_67)\n",
"VAR_56 = VAR_52.title\n",
"return response.json(VAR_64, VAR_66=status)\n",
"VAR_68 = self.jinja_env.select_template(VAR_58)\n",
"return response.html(VAR_68.render(VAR_64), VAR_66=status)\n"
] |
[
"def app(self):...\n",
"app = Sanic(__name__)\n",
"default_templates = str(app_root / 'datasette' / 'templates')\n",
"template_paths = []\n",
"if self.template_dir:\n",
"template_paths.append(self.template_dir)\n",
"template_paths.extend([plugin['templates_path'] for plugin in get_plugins(\n pm) if plugin['templates_path']])\n",
"template_paths.append(default_templates)\n",
"template_loader = ChoiceLoader([FileSystemLoader(template_paths),\n PrefixLoader({'default': FileSystemLoader(default_templates)},\n delimiter=':')])\n",
"self.jinja_env = Environment(loader=template_loader, autoescape=True)\n",
"self.jinja_env.filters['escape_css_string'] = escape_css_string\n",
"self.jinja_env.filters['quote_plus'] = lambda u: urllib.parse.quote_plus(u)\n",
"self.jinja_env.filters['escape_sqlite'] = escape_sqlite\n",
"self.jinja_env.filters['to_css_class'] = to_css_class\n",
"pm.hook.prepare_jinja2_environment(env=self.jinja_env)\n",
"app.add_route(IndexView.as_view(self), '/<as_format:(\\\\.jsono?)?$>')\n",
"app.add_route(favicon, '/favicon.ico')\n",
"app.static('/-/static/', str(app_root / 'datasette' / 'static'))\n",
"for path, dirname in self.static_mounts:\n",
"app.static(path, dirname)\n",
"for plugin in get_plugins(pm):\n",
"if plugin['static_path']:\n",
"app.add_route(JsonDataView.as_view(self, 'inspect.json', self.inspect),\n '/-/inspect<as_format:(\\\\.json)?$>')\n",
"modpath = '/-/static-plugins/{}/'.format(plugin['name'])\n",
"app.add_route(JsonDataView.as_view(self, 'metadata.json', lambda : self.\n _metadata), '/-/metadata<as_format:(\\\\.json)?$>')\n",
"app.static(modpath, plugin['static_path'])\n",
"app.add_route(JsonDataView.as_view(self, 'versions.json', self.versions),\n '/-/versions<as_format:(\\\\.json)?$>')\n",
"app.add_route(JsonDataView.as_view(self, 'plugins.json', self.plugins),\n '/-/plugins<as_format:(\\\\.json)?$>')\n",
"app.add_route(JsonDataView.as_view(self, 'config.json', lambda : self.\n _config), '/-/config<as_format:(\\\\.json)?$>')\n",
"app.add_route(DatabaseDownload.as_view(self),\n '/<db_name:[^/]+?><as_db:(\\\\.db)$>')\n",
"app.add_route(DatabaseView.as_view(self),\n '/<db_name:[^/]+?><as_format:(\\\\.jsono?|\\\\.csv)?$>')\n",
"app.add_route(TableView.as_view(self),\n '/<db_name:[^/]+>/<table_and_format:[^/]+?$>')\n",
"app.add_route(RowView.as_view(self),\n '/<db_name:[^/]+>/<table:[^/]+?>/<pk_path:[^/]+?><as_format:(\\\\.jsono?)?$>'\n )\n",
"self.register_custom_units()\n",
"@app.middleware('response')...\n",
"if original_response.status == 404 and request.path.endswith('/'):\n",
"path = request.path.rstrip('/')\n",
"@app.exception(Exception)...\n",
"if request.query_string:\n",
"title = None\n",
"path = '{}?{}'.format(path, request.query_string)\n",
"return response.redirect(path)\n",
"help = None\n",
"if isinstance(exception, NotFound):\n",
"status = 404\n",
"if isinstance(exception, InvalidUsage):\n",
"info = {}\n",
"status = 405\n",
"if isinstance(exception, DatasetteError):\n",
"message = exception.args[0]\n",
"info = {}\n",
"status = exception.status\n",
"status = 500\n",
"templates = ['500.html']\n",
"message = exception.args[0]\n",
"info = exception.error_dict\n",
"info = {}\n",
"if status != 500:\n",
"message = exception.message\n",
"message = str(exception)\n",
"templates = ['{}.html'.format(status)] + templates\n",
"info.update({'ok': False, 'error': message, 'status': status, 'title': title})\n",
"if exception.messagge_is_html:\n",
"traceback.print_exc()\n",
"if request is not None and request.path.split('?')[0].endswith('.json'):\n",
"message = Markup(message)\n",
"title = exception.title\n",
"return response.json(info, status=status)\n",
"template = self.jinja_env.select_template(templates)\n",
"return response.html(template.render(info), status=status)\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"For",
"Expr'",
"For",
"Condition",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Condition",
"Condition",
"Assign'",
"Condition",
"Condition",
"Assign'",
"Assign'",
"Return'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Condition",
"Expr'",
"Condition",
"Assign'",
"Assign'",
"Return'",
"Assign'",
"Return'"
] |
[
"def FUNC_0(VAR_0):...\n",
"VAR_1 = VAR_0.GET.get('url', None)\n",
"VAR_6 = FUNC_1(VAR_1)\n",
"while VAR_6 != VAR_1:\n",
"VAR_1 = VAR_6\n",
"return JsonResponse({'url': VAR_6})\n",
"VAR_6 = FUNC_1(VAR_1)\n"
] |
[
"def expand_url(request):...\n",
"url = request.GET.get('url', None)\n",
"exurl = expand(url)\n",
"while exurl != url:\n",
"url = exurl\n",
"return JsonResponse({'url': exurl})\n",
"exurl = expand(url)\n"
] |
[
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Return'",
"Assign'"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self.config.get(self.section, 'vccs_url')\n"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self.config.get(self.section, 'vccs_url')\n"
] |
[
0,
0,
0
] |
[
"Condition",
"Docstring",
"Return'"
] |
[
"def FUNC_4(self):...\n",
"for VAR_11 in range(self.size):\n",
"self.addNewNode(VAR_11)\n"
] |
[
"def genNodes(self):...\n",
"for i in range(self.size):\n",
"self.addNewNode(i)\n"
] |
[
0,
1,
0
] |
[
"FunctionDef'",
"For",
"Expr'"
] |
[
"def FUNC_3(self):...\n",
"self.telnet.write(bytes('shutdown\\n', encoding='utf-8'))\n",
"super().close()\n",
"self.openocd.wait()\n",
"db.log_event('Information', 'Debugger', 'Closed openocd')\n"
] |
[
"def close(self):...\n",
"self.telnet.write(bytes('shutdown\\n', encoding='utf-8'))\n",
"super().close()\n",
"self.openocd.wait()\n",
"db.log_event('Information', 'Debugger', 'Closed openocd')\n"
] |
[
0,
0,
0,
0,
2
] |
[
"FunctionDef'",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_12(self, VAR_18):...\n",
"\"\"\"docstring\"\"\"\n",
"if not VAR_18.is_dir:\n",
"return False\n",
"VAR_30, VAR_41, VAR_42, VAR_43, VAR_46, VAR_47 = self._sqlify_fsnode(VAR_18)\n",
"self.fs_db.execute('string' % (VAR_30, VAR_41, VAR_42, VAR_43, VAR_46, VAR_47))\n",
"return\n"
] |
[
"def _insert_in_db(self, item):...\n",
"\"\"\"docstring\"\"\"\n",
"if not item.is_dir:\n",
"return False\n",
"(n_uuid, n_file_name, n_owner, n_upload_time, n_sub_folders_str,\n n_sub_files_str) = self._sqlify_fsnode(item)\n",
"self.fs_db.execute(\n \"INSERT INTO file_system (uuid, file_name, owner, upload_time, sub_folders, sub_files) VALUES ('%s', '%s', '%s', %f, %s, %s);\"\n % (n_uuid, n_file_name, n_owner, n_upload_time, n_sub_folders_str,\n n_sub_files_str))\n",
"return\n"
] |
[
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Condition",
"Return'",
"Assign'",
"Expr'",
"Return'"
] |
[
"@rest_utils.ajax()...\n",
"api.nova.server_delete(VAR_1, VAR_3)\n"
] |
[
"@rest_utils.ajax()...\n",
"api.nova.server_delete(request, server_id)\n"
] |
[
0,
0
] |
[
"Condition",
"Expr'"
] |
[
"def FUNC_6(self, VAR_6, VAR_7):...\n",
"self.vars[VAR_6] = VAR_7\n"
] |
[
"def set_variable(self, key, value):...\n",
"self.vars[key] = value\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Assign'"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self.format_wildcards(self.rule.message) if self.rule.message else None\n"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self.format_wildcards(self.rule.message) if self.rule.message else None\n"
] |
[
0,
0,
0
] |
[
"Condition",
"Docstring",
"Return'"
] |
[
"\"\"\"Unittest to exercise the code in url_helper.py.\"\"\"\n",
"import logging\n",
"import os\n",
"import stat\n",
"import StringIO\n",
"import sys\n",
"import tempfile\n",
"import time\n",
"import unittest\n",
"import urllib\n",
"import urllib2\n",
"VAR_0 = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))\n",
"sys.path.insert(0, VAR_0)\n",
"import test_env\n",
"test_env.setup_test_env()\n",
"from depot_tools import auto_stub\n",
"from third_party.mox import mox\n",
"import url_helper\n",
"def FUNC_0(self):...\n",
"self._mox = mox.Mox()\n",
"self.mock(logging, 'error', lambda *_: None)\n",
"self.mock(logging, 'exception', lambda *_: None)\n",
"self.mock(logging, 'info', lambda *_: None)\n",
"self.mock(logging, 'warning', lambda *_: None)\n",
"self._mox.StubOutWithMock(time, 'sleep')\n",
"self._mox.StubOutWithMock(urllib2, 'urlopen')\n",
"def FUNC_1(self):...\n",
"self._mox.UnsetStubs()\n",
"def FUNC_2(self):...\n",
"self._mox.ReplayAll()\n",
"self.assertEqual(url_helper.UrlOpen('url', max_tries=-1), None)\n",
"self._mox.VerifyAll()\n",
"def FUNC_3(self):...\n",
"self._mox.ReplayAll()\n",
"self.assertEqual(url_helper.UrlOpen('url', wait_duration=-1), None)\n",
"self._mox.VerifyAll()\n",
"def FUNC_4(self):...\n",
"VAR_1 = 'http://my.url.com'\n",
"VAR_2 = 'True'\n",
"url_helper.urllib2.urlopen(mox.StrContains(VAR_1), timeout=mox.IgnoreArg()\n ).AndReturn(StringIO.StringIO(VAR_2))\n",
"self._mox.ReplayAll()\n",
"self.assertEqual(url_helper.UrlOpen(VAR_1, method='GET'), VAR_2)\n",
"self._mox.VerifyAll()\n",
"def FUNC_5(self):...\n",
"VAR_1 = 'http://my.url.com'\n",
"VAR_2 = 'True'\n",
"url_helper.urllib2.urlopen(VAR_1, mox.IgnoreArg(), timeout=mox.IgnoreArg()\n ).AndReturn(StringIO.StringIO(VAR_2))\n",
"self._mox.ReplayAll()\n",
"self.assertEqual(url_helper.UrlOpen(VAR_1, method='POST'), VAR_2)\n",
"self._mox.VerifyAll()\n",
"def FUNC_6(self):...\n",
"VAR_1 = 'http://my.url.com'\n",
"VAR_2 = 'True'\n",
"url_helper.urllib2.urlopen(mox.IsA(urllib2.Request), timeout=mox.IgnoreArg()\n ).AndReturn(StringIO.StringIO(VAR_2))\n",
"self._mox.ReplayAll()\n",
"self.assertEqual(url_helper.UrlOpen(VAR_1, method='POSTFORM'), VAR_2)\n",
"self._mox.VerifyAll()\n",
"def FUNC_7(self):...\n",
"url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndRaise(urllib2.URLError('url'))\n",
"time.sleep(mox.IgnoreArg())\n",
"VAR_2 = 'True'\n",
"url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndReturn(StringIO.StringIO(VAR_2))\n",
"self._mox.ReplayAll()\n",
"self.assertEqual(url_helper.UrlOpen('url', max_tries=2), VAR_2)\n",
"self._mox.VerifyAll()\n",
"def FUNC_8(self):...\n",
"url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndRaise(urllib2.URLError('url'))\n",
"self._mox.ReplayAll()\n",
"self.assertIsNone(url_helper.UrlOpen('url', max_tries=1))\n",
"self._mox.VerifyAll()\n",
"def FUNC_9(self):...\n",
"url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndRaise(urllib2.HTTPError('url', 400, 'error message',\n None, None))\n",
"self._mox.ReplayAll()\n",
"self.assertIsNone(url_helper.UrlOpen('url', max_tries=10))\n",
"self._mox.VerifyAll()\n",
"def FUNC_10(self):...\n",
"VAR_2 = 'response'\n",
"url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndRaise(urllib2.HTTPError('url', 500, 'error message',\n None, None))\n",
"time.sleep(mox.IgnoreArg())\n",
"url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndReturn(StringIO.StringIO(VAR_2))\n",
"self._mox.ReplayAll()\n",
"self.assertEqual(VAR_2, url_helper.UrlOpen('url', max_tries=10))\n",
"self._mox.VerifyAll()\n",
"def FUNC_11(self):...\n",
"VAR_3 = 5\n",
"for i in range(VAR_3):\n",
"VAR_11 = urllib.urlencode({url_helper.swarm_constants.COUNT_KEY: i})\n",
"self._mox.ReplayAll()\n",
"url_helper.urllib2.urlopen(mox.IgnoreArg(), VAR_11, timeout=mox.IgnoreArg()\n ).AndRaise(urllib2.URLError('url'))\n",
"self.assertEqual(url_helper.UrlOpen('url', max_tries=attempts), None)\n",
"if i != VAR_3 - 1:\n",
"self._mox.VerifyAll()\n",
"time.sleep(mox.IgnoreArg())\n",
"def FUNC_12(self):...\n",
"VAR_4 = {url_helper.swarm_constants.COUNT_KEY: 1}\n",
"self._mox.ReplayAll()\n",
"self.assertEqual(url_helper.UrlOpen('url', VAR_4=data), None)\n",
"self._mox.VerifyAll()\n",
"def FUNC_13(self):...\n",
"VAR_4 = {'r': u'not ascii £ һ'}\n",
"VAR_1 = 'http://my.url.com'\n",
"VAR_2 = 'True'\n",
"url_helper.urllib2.urlopen(mox.StrContains(VAR_1), mox.IgnoreArg(), timeout\n =mox.IgnoreArg()).AndReturn(StringIO.StringIO(VAR_2))\n",
"self._mox.ReplayAll()\n",
"self.assertEqual(url_helper.UrlOpen(VAR_1, VAR_4=data), VAR_2)\n",
"self._mox.VerifyAll()\n",
"def FUNC_14(self):...\n",
"VAR_5 = None\n",
"VAR_5 = tempfile.NamedTemporaryFile(delete=False)\n",
"if VAR_5:\n",
"def FUNC_15(self):...\n",
"VAR_5.close()\n",
"os.remove(VAR_5.name)\n",
"VAR_13 = 'fake_local_file.fake'\n",
"if os.path.exists(VAR_13):\n",
"def FUNC_16(self):...\n",
"self._mox.StubOutWithMock(url_helper, 'UrlOpen')\n",
"self._mox.StubOutWithMock(url_helper, 'UrlOpen')\n",
"os.remove(VAR_13)\n",
"VAR_6 = None\n",
"VAR_12 = 'data'\n",
"url_helper.UrlOpen(mox.IgnoreArg(), method='GET').AndReturn(None)\n",
"VAR_6 = tempfile.NamedTemporaryFile(delete=False)\n",
"if VAR_6:\n",
"def FUNC_17(self):...\n",
"url_helper.UrlOpen(mox.IgnoreArg(), method='GET').AndReturn(VAR_12)\n",
"self._mox.ReplayAll()\n",
"VAR_6.close()\n",
"os.remove(VAR_6.name)\n",
"VAR_7 = [('x', 'y'), (1, 2)]\n",
"self._mox.ReplayAll()\n",
"self.assertFalse(url_helper.DownloadFile(VAR_13, 'http://www.fakeurl.com'))\n",
"os.chmod(VAR_6.name, stat.S_IREAD)\n",
"VAR_8 = [('key', 'filename', 'file data')]\n",
"self.assertTrue(url_helper.DownloadFile(VAR_5.name, 'http://www.fakeurl.com'))\n",
"self._mox.VerifyAll()\n",
"self._mox.StubOutWithMock(url_helper, 'UrlOpen')\n",
"VAR_9, VAR_10 = url_helper.EncodeMultipartFormData()\n",
"self.assertEqual(VAR_12, f.read())\n",
"url_helper.UrlOpen(mox.IgnoreArg(), method='GET').AndReturn('data')\n",
"self.assertTrue(VAR_9.startswith('multipart/form-data; boundary='))\n",
"self._mox.VerifyAll()\n",
"self._mox.ReplayAll()\n",
"self.assertEqual('', VAR_10)\n",
"self.assertFalse(url_helper.DownloadFile(VAR_6.name, 'http://www.fakeurl.com'))\n",
"VAR_9, VAR_10 = url_helper.EncodeMultipartFormData(VAR_7=fields)\n",
"self._mox.VerifyAll()\n",
"self.assertTrue(VAR_9.startswith('multipart/form-data; boundary='))\n",
"self.assertTrue('name=\"x\"\\r\\n\\r\\ny' in VAR_10, VAR_10)\n",
"self.assertTrue('name=\"1\"\\r\\n\\r\\n2' in VAR_10, VAR_10)\n",
"VAR_9, VAR_10 = url_helper.EncodeMultipartFormData(VAR_8=files)\n",
"self.assertTrue(VAR_9.startswith('multipart/form-data; boundary='))\n",
"self.assertTrue('name=\"key\"; filename=\"filename\"' in VAR_10, VAR_10)\n",
"self.assertTrue('file data' in VAR_10, VAR_10)\n",
"VAR_9, VAR_10 = url_helper.EncodeMultipartFormData(VAR_7=fields, VAR_8=files)\n",
"self.assertTrue(VAR_9.startswith('multipart/form-data; boundary='))\n",
"self.assertTrue('name=\"x\"\\r\\n\\r\\ny' in VAR_10, VAR_10)\n",
"self.assertTrue('name=\"1\"\\r\\n\\r\\n2' in VAR_10, VAR_10)\n",
"if __name__ == '__main__':\n",
"logging.disable(logging.FATAL)\n",
"unittest.main()\n"
] |
[
"\"\"\"Unittest to exercise the code in url_helper.py.\"\"\"\n",
"import logging\n",
"import os\n",
"import stat\n",
"import StringIO\n",
"import sys\n",
"import tempfile\n",
"import time\n",
"import unittest\n",
"import urllib\n",
"import urllib2\n",
"ROOT_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))\n",
"sys.path.insert(0, ROOT_DIR)\n",
"import test_env\n",
"test_env.setup_test_env()\n",
"from depot_tools import auto_stub\n",
"from third_party.mox import mox\n",
"import url_helper\n",
"def setUp(self):...\n",
"self._mox = mox.Mox()\n",
"self.mock(logging, 'error', lambda *_: None)\n",
"self.mock(logging, 'exception', lambda *_: None)\n",
"self.mock(logging, 'info', lambda *_: None)\n",
"self.mock(logging, 'warning', lambda *_: None)\n",
"self._mox.StubOutWithMock(time, 'sleep')\n",
"self._mox.StubOutWithMock(urllib2, 'urlopen')\n",
"def tearDown(self):...\n",
"self._mox.UnsetStubs()\n",
"def testUrlOpenInvalidTryCount(self):...\n",
"self._mox.ReplayAll()\n",
"self.assertEqual(url_helper.UrlOpen('url', max_tries=-1), None)\n",
"self._mox.VerifyAll()\n",
"def testUrlOpenInvalidWaitDuration(self):...\n",
"self._mox.ReplayAll()\n",
"self.assertEqual(url_helper.UrlOpen('url', wait_duration=-1), None)\n",
"self._mox.VerifyAll()\n",
"def testUrlOpenGETSuccess(self):...\n",
"url = 'http://my.url.com'\n",
"response = 'True'\n",
"url_helper.urllib2.urlopen(mox.StrContains(url), timeout=mox.IgnoreArg()\n ).AndReturn(StringIO.StringIO(response))\n",
"self._mox.ReplayAll()\n",
"self.assertEqual(url_helper.UrlOpen(url, method='GET'), response)\n",
"self._mox.VerifyAll()\n",
"def testUrlOpenPOSTSuccess(self):...\n",
"url = 'http://my.url.com'\n",
"response = 'True'\n",
"url_helper.urllib2.urlopen(url, mox.IgnoreArg(), timeout=mox.IgnoreArg()\n ).AndReturn(StringIO.StringIO(response))\n",
"self._mox.ReplayAll()\n",
"self.assertEqual(url_helper.UrlOpen(url, method='POST'), response)\n",
"self._mox.VerifyAll()\n",
"def testUrlOpenPOSTFORMSuccess(self):...\n",
"url = 'http://my.url.com'\n",
"response = 'True'\n",
"url_helper.urllib2.urlopen(mox.IsA(urllib2.Request), timeout=mox.IgnoreArg()\n ).AndReturn(StringIO.StringIO(response))\n",
"self._mox.ReplayAll()\n",
"self.assertEqual(url_helper.UrlOpen(url, method='POSTFORM'), response)\n",
"self._mox.VerifyAll()\n",
"def testUrlOpenSuccessAfterFailure(self):...\n",
"url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndRaise(urllib2.URLError('url'))\n",
"time.sleep(mox.IgnoreArg())\n",
"response = 'True'\n",
"url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndReturn(StringIO.StringIO(response))\n",
"self._mox.ReplayAll()\n",
"self.assertEqual(url_helper.UrlOpen('url', max_tries=2), response)\n",
"self._mox.VerifyAll()\n",
"def testUrlOpenFailure(self):...\n",
"url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndRaise(urllib2.URLError('url'))\n",
"self._mox.ReplayAll()\n",
"self.assertIsNone(url_helper.UrlOpen('url', max_tries=1))\n",
"self._mox.VerifyAll()\n",
"def testUrlOpenHTTPErrorNoRetry(self):...\n",
"url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndRaise(urllib2.HTTPError('url', 400, 'error message',\n None, None))\n",
"self._mox.ReplayAll()\n",
"self.assertIsNone(url_helper.UrlOpen('url', max_tries=10))\n",
"self._mox.VerifyAll()\n",
"def testUrlOpenHTTPErrorWithRetry(self):...\n",
"response = 'response'\n",
"url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndRaise(urllib2.HTTPError('url', 500, 'error message',\n None, None))\n",
"time.sleep(mox.IgnoreArg())\n",
"url_helper.urllib2.urlopen(mox.IgnoreArg(), mox.IgnoreArg(), timeout=mox.\n IgnoreArg()).AndReturn(StringIO.StringIO(response))\n",
"self._mox.ReplayAll()\n",
"self.assertEqual(response, url_helper.UrlOpen('url', max_tries=10))\n",
"self._mox.VerifyAll()\n",
"def testEnsureCountKeyIncludedInOpen(self):...\n",
"attempts = 5\n",
"for i in range(attempts):\n",
"encoded_data = urllib.urlencode({url_helper.swarm_constants.COUNT_KEY: i})\n",
"self._mox.ReplayAll()\n",
"url_helper.urllib2.urlopen(mox.IgnoreArg(), encoded_data, timeout=mox.\n IgnoreArg()).AndRaise(urllib2.URLError('url'))\n",
"self.assertEqual(url_helper.UrlOpen('url', max_tries=attempts), None)\n",
"if i != attempts - 1:\n",
"self._mox.VerifyAll()\n",
"time.sleep(mox.IgnoreArg())\n",
"def testCountKeyInData(self):...\n",
"data = {url_helper.swarm_constants.COUNT_KEY: 1}\n",
"self._mox.ReplayAll()\n",
"self.assertEqual(url_helper.UrlOpen('url', data=data), None)\n",
"self._mox.VerifyAll()\n",
"def testNonAcsiiData(self):...\n",
"data = {'r': u'not ascii £ һ'}\n",
"url = 'http://my.url.com'\n",
"response = 'True'\n",
"url_helper.urllib2.urlopen(mox.StrContains(url), mox.IgnoreArg(), timeout=\n mox.IgnoreArg()).AndReturn(StringIO.StringIO(response))\n",
"self._mox.ReplayAll()\n",
"self.assertEqual(url_helper.UrlOpen(url, data=data), response)\n",
"self._mox.VerifyAll()\n",
"def testDownloadFile(self):...\n",
"local_file = None\n",
"local_file = tempfile.NamedTemporaryFile(delete=False)\n",
"if local_file:\n",
"def testDownloadFileDownloadError(self):...\n",
"local_file.close()\n",
"os.remove(local_file.name)\n",
"fake_file = 'fake_local_file.fake'\n",
"if os.path.exists(fake_file):\n",
"def testDownloadFileSavingErrors(self):...\n",
"self._mox.StubOutWithMock(url_helper, 'UrlOpen')\n",
"self._mox.StubOutWithMock(url_helper, 'UrlOpen')\n",
"os.remove(fake_file)\n",
"file_readonly = None\n",
"file_data = 'data'\n",
"url_helper.UrlOpen(mox.IgnoreArg(), method='GET').AndReturn(None)\n",
"file_readonly = tempfile.NamedTemporaryFile(delete=False)\n",
"if file_readonly:\n",
"def testEncodeMultipartFormData(self):...\n",
"url_helper.UrlOpen(mox.IgnoreArg(), method='GET').AndReturn(file_data)\n",
"self._mox.ReplayAll()\n",
"file_readonly.close()\n",
"os.remove(file_readonly.name)\n",
"fields = [('x', 'y'), (1, 2)]\n",
"self._mox.ReplayAll()\n",
"self.assertFalse(url_helper.DownloadFile(fake_file, 'http://www.fakeurl.com'))\n",
"os.chmod(file_readonly.name, stat.S_IREAD)\n",
"files = [('key', 'filename', 'file data')]\n",
"self.assertTrue(url_helper.DownloadFile(local_file.name,\n 'http://www.fakeurl.com'))\n",
"self._mox.VerifyAll()\n",
"self._mox.StubOutWithMock(url_helper, 'UrlOpen')\n",
"content_type, body = url_helper.EncodeMultipartFormData()\n",
"self.assertEqual(file_data, f.read())\n",
"url_helper.UrlOpen(mox.IgnoreArg(), method='GET').AndReturn('data')\n",
"self.assertTrue(content_type.startswith('multipart/form-data; boundary='))\n",
"self._mox.VerifyAll()\n",
"self._mox.ReplayAll()\n",
"self.assertEqual('', body)\n",
"self.assertFalse(url_helper.DownloadFile(file_readonly.name,\n 'http://www.fakeurl.com'))\n",
"content_type, body = url_helper.EncodeMultipartFormData(fields=fields)\n",
"self._mox.VerifyAll()\n",
"self.assertTrue(content_type.startswith('multipart/form-data; boundary='))\n",
"self.assertTrue('name=\"x\"\\r\\n\\r\\ny' in body, body)\n",
"self.assertTrue('name=\"1\"\\r\\n\\r\\n2' in body, body)\n",
"content_type, body = url_helper.EncodeMultipartFormData(files=files)\n",
"self.assertTrue(content_type.startswith('multipart/form-data; boundary='))\n",
"self.assertTrue('name=\"key\"; filename=\"filename\"' in body, body)\n",
"self.assertTrue('file data' in body, body)\n",
"content_type, body = url_helper.EncodeMultipartFormData(fields=fields,\n files=files)\n",
"self.assertTrue(content_type.startswith('multipart/form-data; boundary='))\n",
"self.assertTrue('name=\"x\"\\r\\n\\r\\ny' in body, body)\n",
"self.assertTrue('name=\"1\"\\r\\n\\r\\n2' in body, body)\n",
"if __name__ == '__main__':\n",
"logging.disable(logging.FATAL)\n",
"unittest.main()\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
5,
0,
0,
0,
0,
0,
0,
5,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
5,
0,
0,
5,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
5,
0,
0,
0,
0,
0,
0,
0,
0,
5,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
5,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"Expr'",
"Import'",
"Import'",
"Import'",
"Import'",
"Import'",
"Import'",
"Import'",
"Import'",
"Import'",
"Import'",
"Assign'",
"Expr'",
"Import'",
"Expr'",
"ImportFrom'",
"ImportFrom'",
"Import'",
"FunctionDef'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"FunctionDef'",
"Expr'",
"FunctionDef'",
"Expr'",
"Expr'",
"Expr'",
"FunctionDef'",
"Expr'",
"Expr'",
"Expr'",
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"FunctionDef'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"FunctionDef'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"FunctionDef'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"FunctionDef'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"FunctionDef'",
"Assign'",
"For",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Condition",
"Expr'",
"Expr'",
"FunctionDef'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"FunctionDef'",
"Assign'",
"Assign'",
"Condition",
"FunctionDef'",
"Expr'",
"Expr'",
"Assign'",
"Condition",
"FunctionDef'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Condition",
"FunctionDef'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Condition",
"Expr'",
"Expr'"
] |
[
"def FUNC_10(self):...\n",
"self.env_var_override.set('VERIFY_XSS', 'True')\n",
"VAR_7 = BokChoyTestSuite('')\n",
"VAR_1 = 'tests'\n",
"self.assertEqual(VAR_7.cmd, self._expected_command(VAR_1=name, VAR_3=True))\n"
] |
[
"def test_verify_xss_env_var(self):...\n",
"self.env_var_override.set('VERIFY_XSS', 'True')\n",
"suite = BokChoyTestSuite('')\n",
"name = 'tests'\n",
"self.assertEqual(suite.cmd, self._expected_command(name=name, verify_xss=True))\n"
] |
[
0,
3,
0,
0,
0
] |
[
"FunctionDef'",
"Expr'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_39(VAR_101):...\n",
"VAR_101.params = VAR_71, VAR_72\n",
"return VAR_101\n"
] |
[
"def decorate(ruleinfo):...\n",
"ruleinfo.params = params, kwparams\n",
"return ruleinfo\n"
] |
[
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Return'"
] |
[
"def FUNC_17(self, *VAR_26, **VAR_27):...\n",
"if VAR_27.get('body') is not None:\n",
"VAR_27['body'] = jsonutils.dumps(VAR_27['body'], sort_keys=True)\n",
"VAR_19 = super(CLASS_1, self)._rest_call(*VAR_26, **kwargs)\n",
"return VAR_19.json() if VAR_19.content else VAR_19\n"
] |
[
"def _rest_call(self, *args, **kwargs):...\n",
"if kwargs.get('body') is not None:\n",
"kwargs['body'] = jsonutils.dumps(kwargs['body'], sort_keys=True)\n",
"result = super(JSONRESTClient, self)._rest_call(*args, **kwargs)\n",
"return result.json() if result.content else result\n"
] |
[
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Condition",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_27(self, VAR_43, VAR_44):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_63 = self.getfile(VAR_43)\n",
"if VAR_63 == False:\n",
"VAR_64 = self.getfile(VAR_44)\n",
"if VAR_64 != False:\n",
"self.get_path(os.path.dirname(VAR_43)).remove(VAR_63)\n",
"VAR_63[VAR_1] = os.path.basename(VAR_44)\n",
"self.get_path(os.path.dirname(VAR_44)).append(VAR_63)\n",
"return\n"
] |
[
"def rename(self, oldpath, newpath):...\n",
"\"\"\"docstring\"\"\"\n",
"old = self.getfile(oldpath)\n",
"if old == False:\n",
"new = self.getfile(newpath)\n",
"if new != False:\n",
"self.get_path(os.path.dirname(oldpath)).remove(old)\n",
"old[A_NAME] = os.path.basename(newpath)\n",
"self.get_path(os.path.dirname(newpath)).append(old)\n",
"return\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Expr'",
"Assign'",
"Expr'",
"Return'"
] |
[
"def FUNC_21(VAR_21):...\n",
"\"\"\"docstring\"\"\"\n",
"return json.load(VAR_0)\n",
"VAR_0.seek(0)\n",
"import yaml\n",
"return yaml.load(VAR_0)\n"
] |
[
"def _load_configfile(configpath):...\n",
"\"\"\"docstring\"\"\"\n",
"return json.load(f)\n",
"f.seek(0)\n",
"import yaml\n",
"return yaml.load(f)\n"
] |
[
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Return'",
"Expr'",
"Import'",
"Return'"
] |
[
"def FUNC_14(self):...\n",
"VAR_1 = self.client.get(self.get_info_url, {'info_type': 'env_properties'})\n",
"VAR_7 = json.loads(serializers.serialize('json', EnvProperty.objects.all(),\n fields=('name', 'value')))\n",
"self.assertJSONEqual(str(VAR_1.content, encoding=settings.DEFAULT_CHARSET),\n VAR_7)\n"
] |
[
"def test_get_env_properties(self):...\n",
"response = self.client.get(self.get_info_url, {'info_type': 'env_properties'})\n",
"expected_json = json.loads(serializers.serialize('json', EnvProperty.\n objects.all(), fields=('name', 'value')))\n",
"self.assertJSONEqual(str(response.content, encoding=settings.\n DEFAULT_CHARSET), expected_json)\n"
] |
[
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_5(self, VAR_5):...\n",
"if VAR_5 in self.groups:\n",
"self.groups.remove(VAR_5)\n",
"for oldg in VAR_5.get_ancestors():\n",
"if oldg.name != 'all':\n",
"for childg in self.groups:\n",
"if oldg in childg.get_ancestors():\n"
] |
[
"def remove_group(self, group):...\n",
"if group in self.groups:\n",
"self.groups.remove(group)\n",
"for oldg in group.get_ancestors():\n",
"if oldg.name != 'all':\n",
"for childg in self.groups:\n",
"if oldg in childg.get_ancestors():\n"
] |
[
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Condition",
"Expr'",
"For",
"Condition",
"For",
"Condition"
] |
[
"def FUNC_0(self):...\n",
"if not GenericRequest.test_success(self):\n",
"return False\n",
"self.get_user_test_info()\n",
"return False\n",
"return True\n"
] |
[
"def test_success(self):...\n",
"if not GenericRequest.test_success(self):\n",
"return False\n",
"self.get_user_test_info()\n",
"return False\n",
"return True\n"
] |
[
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Condition",
"Return'",
"Expr'",
"Return'",
"Return'"
] |
[
"\"\"\"string\"\"\"\n",
"import logging\n",
"from datetime import timedelta\n",
"import voluptuous as vol\n",
"from homeassistant.components.sensor import PLATFORM_SCHEMA\n",
"from homeassistant.const import TEMP_CELSIUS, DEVICE_CLASS_HUMIDITY, DEVICE_CLASS_TEMPERATURE, STATE_UNKNOWN\n",
"from homeassistant.helpers.entity import Entity\n",
"from homeassistant.util import Throttle\n",
"import homeassistant.helpers.config_validation as cv\n",
"VAR_0 = logging.getLogger(__name__)\n",
"VAR_1 = 'modules'\n",
"VAR_2 = 'station'\n",
"VAR_3 = ['netatmo']\n",
"VAR_4 = timedelta(seconds=600)\n",
"VAR_5 = {'temperature': ['Temperature', TEMP_CELSIUS, None,\n DEVICE_CLASS_TEMPERATURE], 'co2': ['CO2', 'ppm', 'mdi:cloud', None],\n 'pressure': ['Pressure', 'mbar', 'mdi:gauge', None], 'noise': ['Noise',\n 'dB', 'mdi:volume-high', None], 'humidity': ['Humidity', '%', None,\n DEVICE_CLASS_HUMIDITY], 'rain': ['Rain', 'mm', 'mdi:weather-rainy',\n None], 'sum_rain_1': ['sum_rain_1', 'mm', 'mdi:weather-rainy', None],\n 'sum_rain_24': ['sum_rain_24', 'mm', 'mdi:weather-rainy', None],\n 'battery_vp': ['Battery', '', 'mdi:battery', None], 'battery_lvl': [\n 'Battery_lvl', '', 'mdi:battery', None], 'min_temp': ['Min Temp.',\n TEMP_CELSIUS, 'mdi:thermometer', None], 'max_temp': ['Max Temp.',\n TEMP_CELSIUS, 'mdi:thermometer', None], 'windangle': ['Angle', '',\n 'mdi:compass', None], 'windangle_value': ['Angle Value', 'º',\n 'mdi:compass', None], 'windstrength': ['Strength', 'km/h',\n 'mdi:weather-windy', None], 'gustangle': ['Gust Angle', '',\n 'mdi:compass', None], 'gustangle_value': ['Gust Angle Value', 'º',\n 'mdi:compass', None], 'guststrength': ['Gust Strength', 'km/h',\n 'mdi:weather-windy', None], 'rf_status': ['Radio', '', 'mdi:signal',\n None], 'rf_status_lvl': ['Radio_lvl', '', 'mdi:signal', None],\n 'wifi_status': ['Wifi', '', 'mdi:wifi', None], 'wifi_status_lvl': [\n 'Wifi_lvl', 'dBm', 'mdi:wifi', None]}\n",
"VAR_6 = vol.Schema({vol.Required(cv.string): vol.All(cv.ensure_list, [vol.\n In(VAR_5)])})\n",
"VAR_7 = VAR_7.extend({vol.Optional(VAR_2): cv.string, vol.Optional(VAR_1):\n VAR_6})\n",
"def FUNC_0(VAR_8, VAR_9, VAR_10, VAR_11=None):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_12 = VAR_8.components.netatmo\n",
"VAR_13 = CLASS_1(VAR_12.NETATMO_AUTH, VAR_9.get(VAR_2, None))\n",
"VAR_14 = []\n",
"import pyatmo\n",
"if VAR_1 in VAR_9:\n",
"return None\n",
"VAR_10(VAR_14, True)\n",
"for VAR_16, monitored_conditions in VAR_9[VAR_1].items():\n",
"for VAR_16 in VAR_13.get_module_names():\n",
"\"\"\"Implementation of a Netatmo sensor.\"\"\"\n",
"if VAR_16 not in VAR_13.get_module_names():\n",
"for variable in VAR_13.station_data.monitoredConditions(VAR_16):\n",
"def __init__(self, VAR_15, VAR_16, VAR_17):...\n",
"VAR_0.error('Module name: \"%s\" not found', VAR_16)\n",
"for variable in monitored_conditions:\n",
"if variable in VAR_5.keys():\n",
"\"\"\"docstring\"\"\"\n",
"VAR_14.append(CLASS_0(VAR_13, VAR_16, variable))\n",
"VAR_14.append(CLASS_0(VAR_13, VAR_16, variable))\n",
"VAR_0.warning('Ignoring unknown var %s for mod %s', variable, VAR_16)\n",
"self._name = 'Netatmo {} {}'.format(VAR_16, VAR_5[VAR_17][0])\n",
"self.netatmo_data = VAR_15\n",
"self.module_name = VAR_16\n",
"self.type = VAR_17\n",
"self._state = None\n",
"self._device_class = VAR_5[self.type][3]\n",
"self._icon = VAR_5[self.type][2]\n",
"self._unit_of_measurement = VAR_5[self.type][1]\n",
"VAR_20 = self.netatmo_data.station_data.moduleByName(module=module_name)['_id']\n",
"self.module_id = VAR_20[1]\n",
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self._name\n"
] |
[
"\"\"\"\nSupport for the NetAtmo Weather Service.\n\nFor more details about this platform, please refer to the documentation at\nhttps://home-assistant.io/components/sensor.netatmo/\n\"\"\"\n",
"import logging\n",
"from datetime import timedelta\n",
"import voluptuous as vol\n",
"from homeassistant.components.sensor import PLATFORM_SCHEMA\n",
"from homeassistant.const import TEMP_CELSIUS, DEVICE_CLASS_HUMIDITY, DEVICE_CLASS_TEMPERATURE, STATE_UNKNOWN\n",
"from homeassistant.helpers.entity import Entity\n",
"from homeassistant.util import Throttle\n",
"import homeassistant.helpers.config_validation as cv\n",
"_LOGGER = logging.getLogger(__name__)\n",
"CONF_MODULES = 'modules'\n",
"CONF_STATION = 'station'\n",
"DEPENDENCIES = ['netatmo']\n",
"MIN_TIME_BETWEEN_UPDATES = timedelta(seconds=600)\n",
"SENSOR_TYPES = {'temperature': ['Temperature', TEMP_CELSIUS, None,\n DEVICE_CLASS_TEMPERATURE], 'co2': ['CO2', 'ppm', 'mdi:cloud', None],\n 'pressure': ['Pressure', 'mbar', 'mdi:gauge', None], 'noise': ['Noise',\n 'dB', 'mdi:volume-high', None], 'humidity': ['Humidity', '%', None,\n DEVICE_CLASS_HUMIDITY], 'rain': ['Rain', 'mm', 'mdi:weather-rainy',\n None], 'sum_rain_1': ['sum_rain_1', 'mm', 'mdi:weather-rainy', None],\n 'sum_rain_24': ['sum_rain_24', 'mm', 'mdi:weather-rainy', None],\n 'battery_vp': ['Battery', '', 'mdi:battery', None], 'battery_lvl': [\n 'Battery_lvl', '', 'mdi:battery', None], 'min_temp': ['Min Temp.',\n TEMP_CELSIUS, 'mdi:thermometer', None], 'max_temp': ['Max Temp.',\n TEMP_CELSIUS, 'mdi:thermometer', None], 'windangle': ['Angle', '',\n 'mdi:compass', None], 'windangle_value': ['Angle Value', 'º',\n 'mdi:compass', None], 'windstrength': ['Strength', 'km/h',\n 'mdi:weather-windy', None], 'gustangle': ['Gust Angle', '',\n 'mdi:compass', None], 'gustangle_value': ['Gust Angle Value', 'º',\n 'mdi:compass', None], 'guststrength': ['Gust Strength', 'km/h',\n 'mdi:weather-windy', None], 'rf_status': ['Radio', '', 'mdi:signal',\n None], 'rf_status_lvl': ['Radio_lvl', '', 'mdi:signal', None],\n 'wifi_status': ['Wifi', '', 'mdi:wifi', None], 'wifi_status_lvl': [\n 'Wifi_lvl', 'dBm', 'mdi:wifi', None]}\n",
"MODULE_SCHEMA = vol.Schema({vol.Required(cv.string): vol.All(cv.ensure_list,\n [vol.In(SENSOR_TYPES)])})\n",
"PLATFORM_SCHEMA = PLATFORM_SCHEMA.extend({vol.Optional(CONF_STATION): cv.\n string, vol.Optional(CONF_MODULES): MODULE_SCHEMA})\n",
"def setup_platform(hass, config, add_devices, discovery_info=None):...\n",
"\"\"\"docstring\"\"\"\n",
"netatmo = hass.components.netatmo\n",
"data = NetAtmoData(netatmo.NETATMO_AUTH, config.get(CONF_STATION, None))\n",
"dev = []\n",
"import pyatmo\n",
"if CONF_MODULES in config:\n",
"return None\n",
"add_devices(dev, True)\n",
"for module_name, monitored_conditions in config[CONF_MODULES].items():\n",
"for module_name in data.get_module_names():\n",
"\"\"\"Implementation of a Netatmo sensor.\"\"\"\n",
"if module_name not in data.get_module_names():\n",
"for variable in data.station_data.monitoredConditions(module_name):\n",
"def __init__(self, netatmo_data, module_name, sensor_type):...\n",
"_LOGGER.error('Module name: \"%s\" not found', module_name)\n",
"for variable in monitored_conditions:\n",
"if variable in SENSOR_TYPES.keys():\n",
"\"\"\"docstring\"\"\"\n",
"dev.append(NetAtmoSensor(data, module_name, variable))\n",
"dev.append(NetAtmoSensor(data, module_name, variable))\n",
"_LOGGER.warning('Ignoring unknown var %s for mod %s', variable, module_name)\n",
"self._name = 'Netatmo {} {}'.format(module_name, SENSOR_TYPES[sensor_type][0])\n",
"self.netatmo_data = netatmo_data\n",
"self.module_name = module_name\n",
"self.type = sensor_type\n",
"self._state = None\n",
"self._device_class = SENSOR_TYPES[self.type][3]\n",
"self._icon = SENSOR_TYPES[self.type][2]\n",
"self._unit_of_measurement = SENSOR_TYPES[self.type][1]\n",
"module_id = self.netatmo_data.station_data.moduleByName(module=module_name)[\n '_id']\n",
"self.module_id = module_id[1]\n",
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self._name\n"
] |
[
0,
0,
7,
0,
0,
0,
0,
7,
0,
0,
0,
0,
0,
7,
7,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"Expr'",
"Import'",
"ImportFrom'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Import'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Import'",
"Condition",
"Return'",
"Expr'",
"For",
"For",
"Expr'",
"Condition",
"For",
"FunctionDef'",
"Expr'",
"For",
"Condition",
"Docstring",
"Expr'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Docstring",
"Return'"
] |
[
"def __setitem__(self, VAR_7, VAR_15):...\n",
"VAR_9, VAR_21 = FUNC_3(self, VAR_7, VAR_2=self._sep, VAR_8=self.\n _create_on_missing)\n",
"if VAR_9 is self:\n",
"self._data[VAR_21] = VAR_15\n",
"VAR_9[VAR_21] = VAR_15\n"
] |
[
"def __setitem__(self, path, value):...\n",
"node, key = traverse(self, path, sep=self._sep, on_missing=self.\n _create_on_missing)\n",
"if node is self:\n",
"self._data[key] = value\n",
"node[key] = value\n"
] |
[
0,
0,
1,
1,
0
] |
[
"FunctionDef'",
"Assign'",
"Condition",
"Assign'",
"Assign'"
] |
[
"def FUNC_0(self):...\n",
"self.objects = load_model_objects()\n",
"self.client.login(username='Karyn', password='specialP@55word')\n"
] |
[
"def setUp(self):...\n",
"self.objects = load_model_objects()\n",
"self.client.login(username='Karyn', password='specialP@55word')\n"
] |
[
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Expr'"
] |
[
"def FUNC_28(VAR_13):...\n",
"self.assertEqual('/request', VAR_13.path)\n",
"VAR_12.append('applicable')\n",
"return VAR_5\n"
] |
[
"def applicable(request):...\n",
"self.assertEqual('/request', request.path)\n",
"calls.append('applicable')\n",
"return ident\n"
] |
[
0,
0,
0,
0
] |
[
"FunctionDef'",
"Expr'",
"Expr'",
"Return'"
] |
[
"def __init__(self, VAR_7, *VAR_15, **VAR_16):...\n",
"CLASS_0.__init__(self, VAR_7, *VAR_15, **kw)\n"
] |
[
"def __init__(self, param, *a, **kw):...\n",
"Validator.__init__(self, param, *a, **kw)\n"
] |
[
0,
0
] |
[
"FunctionDef'",
"Expr'"
] |
[
"def FUNC_7(VAR_0, VAR_5):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_20 = FUNC_5(VAR_0, VAR_5.number())\n",
"VAR_13, VAR_19 = FUNC_4(VAR_0, VAR_5)\n",
"return VAR_20.seed_games_and_powers(omitting_players=sitters,\n players_doubling_up=two_gamers)\n"
] |
[
"def _seed_games_and_powers(tournament, the_round):...\n",
"\"\"\"docstring\"\"\"\n",
"seeder = _create_game_seeder(tournament, the_round.number())\n",
"sitters, two_gamers = _sitters_and_two_gamers(tournament, the_round)\n",
"return seeder.seed_games_and_powers(omitting_players=sitters,\n players_doubling_up=two_gamers)\n"
] |
[
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Return'"
] |
[
"import sqlite3\n",
"VAR_0 = sqlite3.connect('contacts.sqlite')\n",
"for row in VAR_0.execute('SELECT * FROM contacts'):\n",
"print(row)\n",
"VAR_0.close()\n"
] |
[
"import sqlite3\n",
"db = sqlite3.connect('contacts.sqlite')\n",
"for row in db.execute('SELECT * FROM contacts'):\n",
"print(row)\n",
"db.close()\n"
] |
[
0,
0,
4,
0,
0
] |
[
"Import'",
"Assign'",
"For",
"Expr'",
"Expr'"
] |
[
"def FUNC_1(self):...\n",
"VAR_1 = VAR_0.copy()\n",
"VAR_2 = self._get(VAR_1)\n",
"self.assertEqual(VAR_2.status_code, 403)\n",
"self.assertEqual(User.objects.count(), 1)\n"
] |
[
"def test_invalid(self):...\n",
"meta = DEF_SHIBD_META.copy()\n",
"response = self._get(meta)\n",
"self.assertEqual(response.status_code, 403)\n",
"self.assertEqual(User.objects.count(), 1)\n"
] |
[
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Expr'"
] |
[
"def FUNC_3(self, VAR_1):...\n",
"VAR_5 = dict()\n",
"VAR_10, VAR_11 = self.get_epoch_day(VAR_1)\n",
"VAR_5['interval'] = {'from': self.convert_local_ts_to_utc(VAR_10, self.\n local_timezone), 'to': self.convert_local_ts_to_utc(VAR_11, self.\n local_timezone)}\n",
"VAR_12 = 'string'\n",
"VAR_5['data'] = list()\n",
"for VAR_13 in self.c.execute(VAR_12 % (VAR_10, VAR_11)):\n",
"VAR_5['data'].append({'time': VAR_13[0], 'power': VAR_13[1]})\n",
"if self.get_datetime(VAR_1).date() == datetime.today().date():\n",
"VAR_12 = \"\"\"\n SELECT SUM(EToday) as EToday\n FROM Inverters;\n \"\"\"\n",
"VAR_12 = 'string' % (VAR_10, VAR_11)\n",
"self.c.execute(VAR_12)\n",
"VAR_13 = self.c.fetchone()\n",
"if VAR_13 and VAR_13[0]:\n",
"VAR_5['total'] = VAR_13[0]\n",
"VAR_5['total'] = 0\n",
"VAR_12 = 'string'\n",
"self.c.execute(VAR_12)\n",
"VAR_14, VAR_15 = self.c.fetchone()\n",
"if VAR_14:\n",
"VAR_5['hasPrevious'] = VAR_14 < VAR_10\n",
"VAR_5['hasPrevious'] = False\n",
"if VAR_15:\n",
"VAR_5['hasNext'] = VAR_15 > VAR_11\n",
"VAR_5['hasNext'] = False\n",
"return VAR_5\n"
] |
[
"def get_requested_day(self, date):...\n",
"data = dict()\n",
"day_start, day_end = self.get_epoch_day(date)\n",
"data['interval'] = {'from': self.convert_local_ts_to_utc(day_start, self.\n local_timezone), 'to': self.convert_local_ts_to_utc(day_end, self.\n local_timezone)}\n",
"query = \"\"\"\n SELECT TimeStamp, SUM(Power) AS Power \n FROM DayData \n WHERE TimeStamp BETWEEN %s AND %s \n GROUP BY TimeStamp;\n \"\"\"\n",
"data['data'] = list()\n",
"for row in self.c.execute(query % (day_start, day_end)):\n",
"data['data'].append({'time': row[0], 'power': row[1]})\n",
"if self.get_datetime(date).date() == datetime.today().date():\n",
"query = \"\"\"\n SELECT SUM(EToday) as EToday\n FROM Inverters;\n \"\"\"\n",
"query = (\n \"\"\"\n SELECT SUM(DayYield) AS Power \n FROM MonthData \n WHERE TimeStamp BETWEEN %s AND %s\n GROUP BY TimeStamp\n \"\"\"\n % (day_start, day_end))\n",
"self.c.execute(query)\n",
"row = self.c.fetchone()\n",
"if row and row[0]:\n",
"data['total'] = row[0]\n",
"data['total'] = 0\n",
"query = \"\"\"\n SELECT MIN(TimeStamp) as Min, MAX(TimeStamp) as Max \n FROM ( SELECT TimeStamp FROM DayData GROUP BY TimeStamp );\n \"\"\"\n",
"self.c.execute(query)\n",
"first_data, last_data = self.c.fetchone()\n",
"if first_data:\n",
"data['hasPrevious'] = first_data < day_start\n",
"data['hasPrevious'] = False\n",
"if last_data:\n",
"data['hasNext'] = last_data > day_end\n",
"data['hasNext'] = False\n",
"return data\n"
] |
[
0,
0,
0,
0,
4,
0,
4,
0,
0,
0,
4,
4,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"For",
"Expr'",
"Condition",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Return'"
] |
[
"@FUNC_0...\n",
"return VAR_4\n"
] |
[
"@handle_html...\n",
"return base\n"
] |
[
0,
0
] |
[
"Condition",
"Return'"
] |
[
"def FUNC_6(self, VAR_40):...\n",
"if VAR_40 and len(VAR_40) > 500:\n",
"VAR_101.errors.add(errors.DESC_TOO_LONG)\n",
"return unkeep_space(VAR_40 or '')\n"
] |
[
"def run(self, description):...\n",
"if description and len(description) > 500:\n",
"c.errors.add(errors.DESC_TOO_LONG)\n",
"return unkeep_space(description or '')\n"
] |
[
0,
0,
0,
0
] |
[
"FunctionDef'",
"Condition",
"Expr'",
"Return'"
] |
[
"def FUNC_7(self):...\n",
"VAR_4 = '/api/apps'\n",
"VAR_5 = self.client.post(VAR_4)\n",
"self.assertEqual(VAR_5.status_code, 201)\n",
"VAR_6 = VAR_5.data['id']\n",
"VAR_4 = '/api/apps/{app_id}/scale'.format(**locals())\n",
"VAR_8 = {'web': 'not_an_int'}\n",
"VAR_5 = self.client.post(VAR_4, json.dumps(VAR_8), content_type=\n 'application/json')\n",
"self.assertContains(VAR_5, 'Invalid scaling format', status_code=400)\n",
"VAR_8 = {'invalid': 1}\n",
"VAR_5 = self.client.post(VAR_4, json.dumps(VAR_8), content_type=\n 'application/json')\n",
"self.assertContains(VAR_5, 'Container type invalid', status_code=400)\n"
] |
[
"def test_container_errors(self):...\n",
"url = '/api/apps'\n",
"response = self.client.post(url)\n",
"self.assertEqual(response.status_code, 201)\n",
"app_id = response.data['id']\n",
"url = '/api/apps/{app_id}/scale'.format(**locals())\n",
"body = {'web': 'not_an_int'}\n",
"response = self.client.post(url, json.dumps(body), content_type=\n 'application/json')\n",
"self.assertContains(response, 'Invalid scaling format', status_code=400)\n",
"body = {'invalid': 1}\n",
"response = self.client.post(url, json.dumps(body), content_type=\n 'application/json')\n",
"self.assertContains(response, 'Container type invalid', status_code=400)\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def __init__(self, VAR_20, **VAR_1):...\n",
"super().__init__(**kwargs)\n",
"self._new_name = VAR_20\n"
] |
[
"def __init__(self, new_name, **kwargs):...\n",
"super().__init__(**kwargs)\n",
"self._new_name = new_name\n"
] |
[
0,
0,
0
] |
[
"FunctionDef'",
"Expr'",
"Assign'"
] |
[
"def __init__(self, VAR_10, VAR_13, VAR_14, VAR_11=None):...\n",
"CLASS_1.__init__(self, VAR_10, VAR_11)\n",
"self.username = VAR_13\n",
"self.password = VAR_14\n",
"self.url = '%slogin' % self.base_url\n",
"self.data = {'username': self.username, 'password': self.password, 'next': '/'}\n"
] |
[
"def __init__(self, browser, username, password, base_url=None):...\n",
"GenericRequest.__init__(self, browser, base_url)\n",
"self.username = username\n",
"self.password = password\n",
"self.url = '%slogin' % self.base_url\n",
"self.data = {'username': self.username, 'password': self.password, 'next': '/'}\n"
] |
[
0,
0,
0,
0,
5,
5
] |
[
"FunctionDef'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'"
] |
[
"@api.public...\n",
"FUNC_31(self, 'DELETE')\n"
] |
[
"@api.public...\n",
"record(self, 'DELETE')\n"
] |
[
0,
0
] |
[
"Condition",
"Expr'"
] |
[
"def FUNC_39(self):...\n",
"\"\"\"docstring\"\"\"\n",
"self.remove_sos_archive()\n",
"if self.hash_retrieved:\n",
"self.remove_file(self.sos_path + '.md5')\n",
"VAR_48 = self.host.set_cleanup_cmd()\n",
"if VAR_48:\n",
"self.run_command(VAR_48)\n"
] |
[
"def cleanup(self):...\n",
"\"\"\"docstring\"\"\"\n",
"self.remove_sos_archive()\n",
"if self.hash_retrieved:\n",
"self.remove_file(self.sos_path + '.md5')\n",
"cleanup = self.host.set_cleanup_cmd()\n",
"if cleanup:\n",
"self.run_command(cleanup)\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Expr'",
"Condition",
"Expr'",
"Assign'",
"Condition",
"Expr'"
] |
[
"import sys\n",
"import logging\n",
"import sqlalchemy as sa\n",
"from . import filters\n",
"from sqlalchemy.orm import joinedload\n",
"from sqlalchemy.exc import IntegrityError\n",
"from sqlalchemy import func\n",
"from sqlalchemy.orm.properties import SynonymProperty\n",
"from ..base import BaseInterface\n",
"from ..group import GroupByDateYear, GroupByDateMonth, GroupByCol\n",
"from ..mixins import FileColumn, ImageColumn\n",
"from ...filemanager import FileManager, ImageManager\n",
"from ..._compat import as_unicode\n",
"from ...const import LOGMSG_ERR_DBI_ADD_GENERIC, LOGMSG_ERR_DBI_EDIT_GENERIC, LOGMSG_ERR_DBI_DEL_GENERIC, LOGMSG_WAR_DBI_ADD_INTEGRITY, LOGMSG_WAR_DBI_EDIT_INTEGRITY, LOGMSG_WAR_DBI_DEL_INTEGRITY\n",
"VAR_0 = logging.getLogger(__name__)\n",
"def FUNC_0(VAR_1):...\n",
"for key in VAR_6.__all__:\n",
"if not hasattr(VAR_1, key):\n",
"\"\"\"\n SQLAModel\n Implements SQLA support methods for views\n \"\"\"\n",
"setattr(VAR_1, key, getattr(VAR_6, key))\n",
"VAR_3 = None\n",
"VAR_4 = VAR_6.SQLAFilterConverter\n",
"def __init__(self, VAR_1, VAR_3=None):...\n",
"FUNC_0(self)\n",
"self.list_columns = dict()\n",
"self.list_properties = dict()\n",
"self.session = VAR_3\n",
"for VAR_36 in sa.orm.class_mapper(VAR_1).iterate_properties:\n",
"if type(VAR_36) != SynonymProperty:\n",
"for VAR_14 in VAR_1.__mapper__.columns.keys():\n",
"self.list_properties[VAR_36.key] = VAR_36\n",
"if VAR_14 in self.list_properties:\n",
"super(CLASS_0, self).__init__(VAR_1)\n",
"self.list_columns[VAR_14] = VAR_1.__mapper__.columns[VAR_14]\n",
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self.obj.__name__\n"
] |
[
"import sys\n",
"import logging\n",
"import sqlalchemy as sa\n",
"from . import filters\n",
"from sqlalchemy.orm import joinedload\n",
"from sqlalchemy.exc import IntegrityError\n",
"from sqlalchemy import func\n",
"from sqlalchemy.orm.properties import SynonymProperty\n",
"from ..base import BaseInterface\n",
"from ..group import GroupByDateYear, GroupByDateMonth, GroupByCol\n",
"from ..mixins import FileColumn, ImageColumn\n",
"from ...filemanager import FileManager, ImageManager\n",
"from ..._compat import as_unicode\n",
"from ...const import LOGMSG_ERR_DBI_ADD_GENERIC, LOGMSG_ERR_DBI_EDIT_GENERIC, LOGMSG_ERR_DBI_DEL_GENERIC, LOGMSG_WAR_DBI_ADD_INTEGRITY, LOGMSG_WAR_DBI_EDIT_INTEGRITY, LOGMSG_WAR_DBI_DEL_INTEGRITY\n",
"log = logging.getLogger(__name__)\n",
"def _include_filters(obj):...\n",
"for key in filters.__all__:\n",
"if not hasattr(obj, key):\n",
"\"\"\"\n SQLAModel\n Implements SQLA support methods for views\n \"\"\"\n",
"setattr(obj, key, getattr(filters, key))\n",
"session = None\n",
"filter_converter_class = filters.SQLAFilterConverter\n",
"def __init__(self, obj, session=None):...\n",
"_include_filters(self)\n",
"self.list_columns = dict()\n",
"self.list_properties = dict()\n",
"self.session = session\n",
"for prop in sa.orm.class_mapper(obj).iterate_properties:\n",
"if type(prop) != SynonymProperty:\n",
"for col_name in obj.__mapper__.columns.keys():\n",
"self.list_properties[prop.key] = prop\n",
"if col_name in self.list_properties:\n",
"super(SQLAInterface, self).__init__(obj)\n",
"self.list_columns[col_name] = obj.__mapper__.columns[col_name]\n",
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self.obj.__name__\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"Import'",
"Import'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Assign'",
"FunctionDef'",
"For",
"Condition",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"FunctionDef'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"For",
"Condition",
"For",
"Assign'",
"Condition",
"Expr'",
"Assign'",
"Condition",
"Docstring",
"Return'"
] |
[
"@VAR_0.route('/about')...\n",
"return render_template('about.html')\n"
] |
[
"@app.route('/about')...\n",
"return render_template('about.html')\n"
] |
[
0,
0
] |
[
"Condition",
"Return'"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self._device_class\n"
] |
[
"@property...\n",
"\"\"\"docstring\"\"\"\n",
"return self._device_class\n"
] |
[
0,
0,
0
] |
[
"Condition",
"Docstring",
"Return'"
] |
[
"def FUNC_1():...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_2 = 'string'\n",
"VAR_3 = psycopg2.connect(database=DBNAME)\n",
"VAR_4 = VAR_3.cursor()\n",
"VAR_4.execute(VAR_2)\n",
"VAR_5 = VAR_4.fetchall()\n",
"VAR_3.close()\n",
"return VAR_5\n"
] |
[
"def get_top_authors():...\n",
"\"\"\"docstring\"\"\"\n",
"query = (\n 'select au.name, count(alv.article) as views from articles a inner join article_log_view alv on a.slug = alv.article inner join authors au on a.author = au.id group by au.name order by count(alv.article) desc;'\n )\n",
"db = psycopg2.connect(database=DBNAME)\n",
"c = db.cursor()\n",
"c.execute(query)\n",
"rows = c.fetchall()\n",
"db.close()\n",
"return rows\n"
] |
[
0,
0,
4,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Return'"
] |
[
"def FUNC_9(self):...\n",
"self._hosts_cache = None\n",
"for VAR_12 in self.parent_groups:\n",
"VAR_12.clear_hosts_cache()\n"
] |
[
"def clear_hosts_cache(self):...\n",
"self._hosts_cache = None\n",
"for g in self.parent_groups:\n",
"g.clear_hosts_cache()\n"
] |
[
0,
0,
1,
1
] |
[
"FunctionDef'",
"Assign'",
"For",
"Expr'"
] |
[
"def FUNC_1(self):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_15 = ['hp3par_api_url', 'hp3par_username', 'hp3par_password', 'san_ip',\n 'san_login', 'san_password']\n",
"self.common.check_flags(self.configuration, VAR_15)\n"
] |
[
"def _check_flags(self):...\n",
"\"\"\"docstring\"\"\"\n",
"required_flags = ['hp3par_api_url', 'hp3par_username', 'hp3par_password',\n 'san_ip', 'san_login', 'san_password']\n",
"self.common.check_flags(self.configuration, required_flags)\n"
] |
[
0,
0,
0,
0
] |
[
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'"
] |
[
"def FUNC_5(VAR_5):...\n",
""
] |
[
"def urlopen(url):...\n",
""
] |
[
0,
0
] |
[
"FunctionDef'",
"Condition"
] |
[
"def __init__(self, VAR_5, VAR_6):...\n",
"self.master = VAR_5\n",
"VAR_5.title('Jira Automation')\n",
"self.firefox = VAR_6\n",
"self.colorWhiteText = '#FFF'\n",
"self.colorGreyBackground = '#708090'\n",
"VAR_12 = Label(VAR_5, text=' ')\n",
"VAR_12.grid(column=0, row=0)\n",
"self.login_button = Button(VAR_5, text='JIRA Login', command=self.\n setCredentials)\n",
"self.login_button.grid(sticky='E', column=3, row=0)\n",
"self.summary_static_text = Label(VAR_5, text='Input Sync Summary:')\n",
"self.summary_static_text.grid(columnspan=4, sticky='W', row=1)\n",
"self.summary_field = Entry(VAR_9, width=61)\n",
"self.summary_field.configure(background=self.colorGreyBackground,\n foreground=self.colorWhiteText)\n",
"self.summary_field.grid(columnspan=4, column=0, row=2, sticky='W')\n",
"VAR_13 = Label(VAR_5, text=' ')\n",
"VAR_13.grid(column=0, row=3)\n",
"self.clarify_static_text = Label(VAR_5, text='Input Case Number:')\n",
"self.clarify_static_text.grid(columnspan=4, sticky='W', row=4)\n",
"self.clarify_field = Entry(VAR_9, width=61)\n",
"self.clarify_field.configure(background=self.colorGreyBackground,\n foreground=self.colorWhiteText)\n",
"self.clarify_field.grid(column=0, row=5, columnspan=4, sticky='W')\n",
"VAR_14 = Label(VAR_5, text=' ')\n",
"VAR_14.grid(column=0, row=6)\n",
"self.detailed_static_text = Label(VAR_5, text='Input Detailed Information')\n",
"self.detailed_static_text.grid(columnspan=4, sticky='W', row=7)\n",
"self.detailed_field = Text(VAR_9, height=35, width=79)\n",
"self.detailed_field.configure(background=self.colorGreyBackground,\n foreground=self.colorWhiteText)\n",
"self.detailed_field.grid(columnspan=4, row=8, column=0, sticky='W')\n",
"self.run_split_button = Button(VAR_5, text='Create JIRA', command=self.\n createCase)\n",
"self.run_split_button.grid(row=10, column=1, sticky='W')\n",
"self.close_button = Button(VAR_5, text='Close', command=master.quit)\n",
"self.close_button.grid(row=10, column=2, sticky='E')\n"
] |
[
"def __init__(self, master, selenium):...\n",
"self.master = master\n",
"master.title('Jira Automation')\n",
"self.firefox = selenium\n",
"self.colorWhiteText = '#FFF'\n",
"self.colorGreyBackground = '#708090'\n",
"empty_spacer_1 = Label(master, text=' ')\n",
"empty_spacer_1.grid(column=0, row=0)\n",
"self.login_button = Button(master, text='JIRA Login', command=self.\n setCredentials)\n",
"self.login_button.grid(sticky='E', column=3, row=0)\n",
"self.summary_static_text = Label(master, text='Input Sync Summary:')\n",
"self.summary_static_text.grid(columnspan=4, sticky='W', row=1)\n",
"self.summary_field = Entry(root, width=61)\n",
"self.summary_field.configure(background=self.colorGreyBackground,\n foreground=self.colorWhiteText)\n",
"self.summary_field.grid(columnspan=4, column=0, row=2, sticky='W')\n",
"empty_spacer_2 = Label(master, text=' ')\n",
"empty_spacer_2.grid(column=0, row=3)\n",
"self.clarify_static_text = Label(master, text='Input Case Number:')\n",
"self.clarify_static_text.grid(columnspan=4, sticky='W', row=4)\n",
"self.clarify_field = Entry(root, width=61)\n",
"self.clarify_field.configure(background=self.colorGreyBackground,\n foreground=self.colorWhiteText)\n",
"self.clarify_field.grid(column=0, row=5, columnspan=4, sticky='W')\n",
"empty_spacer_3 = Label(master, text=' ')\n",
"empty_spacer_3.grid(column=0, row=6)\n",
"self.detailed_static_text = Label(master, text='Input Detailed Information')\n",
"self.detailed_static_text.grid(columnspan=4, sticky='W', row=7)\n",
"self.detailed_field = Text(root, height=35, width=79)\n",
"self.detailed_field.configure(background=self.colorGreyBackground,\n foreground=self.colorWhiteText)\n",
"self.detailed_field.grid(columnspan=4, row=8, column=0, sticky='W')\n",
"self.run_split_button = Button(master, text='Create JIRA', command=self.\n createCase)\n",
"self.run_split_button.grid(row=10, column=1, sticky='W')\n",
"self.close_button = Button(master, text='Close', command=master.quit)\n",
"self.close_button.grid(row=10, column=2, sticky='E')\n"
] |
[
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
[
"FunctionDef'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Assign'",
"Expr'"
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.