diff --git "a/test.jsonl" "b/test.jsonl"
--- "a/test.jsonl"
+++ "b/test.jsonl"
@@ -117,14 +117,12 @@
{"code": " public function setHomePage()\n {\n if (!$this->request->is('post')) {\n throw new MethodNotAllowedException(__('This endpoint only aaccepts POST requests.'));\n }\n if (empty($this->request->data['path'])) {\n $this->request->data = array('path' => $this->request->data);\n }\n if (empty($this->request->data['path'])) {\n throw new InvalidArgumentException(__('No path POSTed.'));\n }\n $setting = array(\n 'UserSetting' => array(\n 'user_id' => $this->Auth->user('id'),\n 'setting' => 'homepage',\n 'value' => json_encode(array('path' => $this->request->data['path']))\n )\n );\n $result = $this->UserSetting->setSetting($this->Auth->user(), $setting);\n return $this->RestResponse->saveSuccessResponse('UserSettings', 'setHomePage', false, $this->response->type(), 'Homepage set to ' . $this->request->data['path']);\n }", "label_name": "Base", "label": 1.0}
{"code": " public function checkAuthorisation($id, $user, $write)\n {\n // fetch the bare template\n $template = $this->find('first', array(\n 'conditions' => array('id' => $id),\n 'recursive' => -1,\n ));\n\n // if not found return false\n if (empty($template)) {\n return false;\n }\n\n //if the user is a site admin, return the template withoug question\n if ($user['Role']['perm_site_admin']) {\n return $template;\n }\n\n if ($write) {\n // if write access is requested, check if template belongs to user's org and whether the user is authorised to edit templates\n if ($user['Organisation']['name'] == $template['Template']['org'] && $user['Role']['perm_template']) {\n return $template;\n }\n return false;\n } else {\n\n // if read access is requested, check if the template belongs to the user's org or alternatively whether the template is shareable\n if ($user['Organisation']['name'] == $template['Template']['org'] || $template['Template']['share']) {\n return $template;\n }\n return false;\n }\n }", "label_name": "Base", "label": 1.0}
{"code": " public function isBlacklisted($ip, $username)\n {\n // first remove old expired rows\n $this->clean();\n // count\n $params = array('conditions' => array(\n 'Bruteforce.ip' => $ip,\n 'Bruteforce.username' => $username),);\n $count = $this->find('count', $params);\n if ($count >= Configure::read('SecureAuth.amount')) {\n return true;\n } else {\n return false;\n }\n }", "label_name": "Base", "label": 1.0}
-{"code": " public function isBlacklisted($ip, $username)\n {\n // first remove old expired rows\n $this->clean();\n // count\n $params = array('conditions' => array(\n 'Bruteforce.ip' => $ip,\n 'Bruteforce.username' => $username),);\n $count = $this->find('count', $params);\n if ($count >= Configure::read('SecureAuth.amount')) {\n return true;\n } else {\n return false;\n }\n }", "label_name": NaN, "label": NaN}
{"code": "\tpublic function Archive() {\n\n\t\t// Update \"main\" ticket\n\t\t$upd_stmt = Database::prepare('\n\t\t\tUPDATE `' . TABLE_PANEL_TICKETS . '` SET `archived` = \"1\" WHERE `id` = :tid'\n\t\t);\n\t\tDatabase::pexecute($upd_stmt, array('tid' => $this->tid));\n\n\t\t// Update \"answers\" to ticket\n\t\t$upd_stmt = Database::prepare('\n\t\t\tUPDATE `' . TABLE_PANEL_TICKETS . '` SET `archived` = \"1\" WHERE `answerto` = :tid'\n\t\t);\n\t\tDatabase::pexecute($upd_stmt, array('tid' => $this->tid));\n\t\treturn true;\n\t}", "label_name": "Class", "label": 2.0}
{"code": " public function browse() {\n $params = func_get_args();\n\n $this->path = join('/', $params);\n // make sure there's a / at the end\n if (substr($this->path, -1, 1) != '/')\n $this->path .= '/';\n\n //security\n // we dont allow back link\n if (strpos($this->path, '..') !== false) {\n /*\n if (Plugin::isEnabled('statistics_api')) {\n $user = null;\n if (AuthUser::isLoggedIn())\n $user = AuthUser::getUserName();\n $ip = isset($_SERVER['HTTP_X_FORWARDED_FOR']) ? $_SERVER['HTTP_X_FORWARDED_FOR'] : ($_SERVER['REMOTE_ADDR']);\n $event = array('event_type' => 'hack_attempt', // simple event type identifier\n 'description' => __('A possible hack attempt was detected.'), // translatable description\n 'ipaddress' => $ip,\n 'username' => $user);\n Observer::notify('stats_file_manager_hack_attempt', $event);\n }\n */\n }\n $this->path = str_replace('..', '', $this->path);\n\n // clean up nicely\n $this->path = str_replace('//', '', $this->path);\n\n // we dont allow leading slashes\n $this->path = preg_replace('/^\\//', '', $this->path);\n\n $this->fullpath = FILES_DIR . '/' . $this->path;\n\n // clean up nicely\n $this->fullpath = preg_replace('/\\/\\//', '/', $this->fullpath);\n\n $this->display('file_manager/views/index', array(\n 'dir' => $this->path,\n //'files' => $this->_getListFiles()\n 'files' => $this->_listFiles()\n ));\n }", "label_name": "Class", "label": 2.0}
{"code": " public function delete() {\n if (!AuthUser::hasPermission('file_manager_delete')) {\n Flash::set('error', __('You do not have sufficient permissions to delete a file or directory.'));\n redirect(get_url('plugin/file_manager/browse/'));\n }\n\n $paths = func_get_args();\n\n $file = urldecode(join('/', $paths));\n \n // CSRF checks\n if (isset($_GET['csrf_token'])) {\n $csrf_token = $_GET['csrf_token'];\n if (!SecureToken::validateToken($csrf_token, BASE_URL.'plugin/file_manager/delete/'.$file)) {\n Flash::set('error', __('Invalid CSRF token found!'));\n redirect(get_url('plugin/file_manager/browse/'));\n }\n }\n else {\n Flash::set('error', __('No CSRF token found!'));\n redirect(get_url('plugin/file_manager/browse/'));\n }\n\n $file = FILES_DIR . '/' . str_replace('..', '', $file);\n $filename = array_pop($paths);\n $paths = join('/', $paths);\n\n if (is_file($file)) {\n if (!unlink($file))\n Flash::set('error', __('Permission denied!'));\n }\n else {\n if (!$this->_rrmdir($file))\n Flash::set('error', __('Permission denied!'));\n }\n\n redirect(get_url('plugin/file_manager/browse/' . $paths));\n }", "label_name": "Class", "label": 2.0}
{"code": " $object->size = convert_size($cur->getSize());\n $object->mtime = date('D, j M, Y', $cur->getMTime());\n list($object->perms, $object->chmod) = $this->_getPermissions($cur->getPerms());\n \n // Find the file type\n $object->type = $this->_getFileType($cur);\n\n // make the link depending on if it's a file or a dir\n if ($cur->isDir()) {\n $object->link = 'path . $object->name) . '\">' . $object->name . '';\n } else {\n $object->link = 'path . $object->name . (endsWith($object->name, URL_SUFFIX) ? '?has_url_suffix=1' : '')) . '\">' . $object->name . '';\n }\n\n $files[$object->name] = $object;\n }", "label_name": "Class", "label": 2.0}
{"code": "\t\t\t$user = db_fetch_cell_prepared('SELECT username FROM user_auth WHERE id = ?', array($check['user']), 'username');\n\t\t\tform_alternate_row('line' . $check['id']);\n\t\t\t$name = get_data_source_title($check['datasource']);\n\t\t\t$title = $name;\n\t\t\tif (strlen($name) > 50) {\n\t\t\t\t$name = substr($name, 0, 50);\n\t\t\t}\n\t\t\tform_selectable_cell('' . $name . '', $check['id']);\n\t\t\tform_selectable_cell($user, $check['id']);\n\t\t\tform_selectable_cell(date('F j, Y, G:i', $check['started']), $check['id']);\n\t\t\tform_selectable_cell($check['datasource'], $check['id']);\n\t\t\tform_selectable_cell(debug_icon(($check['done'] ? (strlen($issue_line) ? 'off' : 'on' ) : '')), $check['id'], '', 'text-align: center;');\n\t\t\tform_selectable_cell(debug_icon($info['rrd_writable']), $check['id'], '', 'text-align: center;');\n\t\t\tform_selectable_cell(debug_icon($info['rrd_exists']), $check['id'], '', 'text-align: center;');\n\t\t\tform_selectable_cell(debug_icon($info['active']), $check['id'], '', 'text-align: center;');\n\t\t\tform_selectable_cell(debug_icon($info['rrd_match']), $check['id'], '', 'text-align: center;');\n\t\t\tform_selectable_cell(debug_icon($info['valid_data']), $check['id'], '', 'text-align: center;');\n\t\t\tform_selectable_cell(debug_icon(($info['rra_timestamp2'] != '' ? 1 : '')), $check['id'], '', 'text-align: center;');\n\t\t\tform_selectable_cell('' . html_escape(strlen(trim($issue_line)) ? $issue_line : '') . '', $check['id']);\n\t\t\tform_checkbox_cell($check['id'], $check['id']);\n\t\t\tform_end_row();\n\t\t}\n\t}else{", "label_name": "Base", "label": 1.0}
{"code": "\t\t\tform_selectable_cell(filter_value($vdef['name'], get_request_var('filter'), 'vdef.php?action=edit&id=' . $vdef['id']), $vdef['id']);\n form_selectable_cell($disabled ? __('No'):__('Yes'), $vdef['id'], '', 'text-align:right');\n form_selectable_cell(number_format_i18n($vdef['graphs'], '-1'), $vdef['id'], '', 'text-align:right');\n form_selectable_cell(number_format_i18n($vdef['templates'], '-1'), $vdef['id'], '', 'text-align:right');\n form_checkbox_cell($vdef['name'], $vdef['id'], $disabled);\n form_end_row();\n }", "label_name": "Base", "label": 1.0}
-{"code": "function form_confirm_buttons($action_url, $cancel_url) {\n\tglobal $config;\n\t?>\n\t
\n\t\t| \n\t\t\t\")' value=''>\n\t\t\t\")' value=''>\n\t\t | \n\t
\n $length) {\n\t\t\tfor($i = 0; $i < $forgiveness; $i++) {\n\t\t\t\tif (substr($string, $length-$i, 1) == \" \") {\n\t\t\t\t\t$new_string .= substr($string, 0, $length-$i) . \"
\";\n\n\t\t\t\t\tbreak;\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t$string = substr($string, $length-$i);\n\t\t} else {\n\t\t\t$new_string .= $string;\n\t\t\t$done = true;\n\t\t}\n\n\t\t$j++;\n\t\tif ($j > 4) break;\n\t}\n\n\treturn $new_string;\n}", "label_name": "Base", "label": 1.0}
{"code": "\t\t} elseif (strtolower($icon) == 'desc') {\n\t\t\t$icon = 'fa fa-sort-desc';\n\t\t} else {\n\t\t\t$icon = 'fa fa-unsorted';\n\t\t}\n\n\t\tif (($db_column == '') || (substr_count($db_column, 'nosort'))) {\n\t\t\tprint '' . $display_text . \" | \\n\";", "label_name": "Base", "label": 1.0}
{"code": " $fileName = ltrim(dirname($fileName) . '/' . basename($fileName, '.js'), '/.');\n if (empty($fileName)) {\n continue;\n }\n $templateBase = $inflector->filter(array(\n 'module' => $moduleName,\n 'controller' => $controllerName,\n 'file' => $fileName)\n );\n\n $templateExtend = $inflector->filter(array(\n 'module' => $moduleName,\n 'controller' => $this->Request()->getControllerName(),\n 'file' => $fileName)\n );\n if ($this->View()->templateExists($templateBase)) {\n $template .= '{include file=\"' . $templateBase. '\"}' . \"\\n\";\n }\n if ($this->View()->templateExists($templateExtend)) {\n $template .= '{include file=\"' . $templateExtend. '\"}' . \"\\n\";\n }\n }", "label_name": "Class", "label": 2.0}
@@ -335,14 +333,12 @@
{"code": " public function take($value)\n {\n return $this->limit($value);\n }", "label_name": "Base", "label": 1.0}
{"code": " public function from($dirName)\n {\n $this->from = $dirName;\n\n return $this;\n }", "label_name": "Base", "label": 1.0}
{"code": " public function getModel()\n {\n return $this->model;\n }", "label_name": "Base", "label": 1.0}
-{"code": " elseif ($permission == $rolePermission && $rolePermissions[$permission] == 1) {\n $matched = true;\n break;\n }", "label_name": NaN, "label": NaN}
{"code": " public function clearLoginAttempts()\n {\n // If our login attempts is already at zero\n // we do not need to do anything. Additionally,\n // if we are suspended, we are not going to do\n // anything either as clearing login attempts\n // makes us unsuspended. We need to manually\n // call unsuspend() in order to unsuspend.\n if ($this->getLoginAttempts() == 0 or $this->is_suspended) {\n return;\n }\n\n $this->attempts = 0;\n $this->last_attempt_at = null;\n $this->is_suspended = false;\n $this->suspended_at = null;\n $this->save();\n }", "label_name": NaN, "label": NaN}
{"code": " foreach ($mergedPermissions as $mergedPermission => $value) {\n // Strip the '*' off the beginning of the permission.\n $checkPermission = substr($permission, 1);\n\n // We will make sure that the merged permission does not\n // exactly match our permission, but ends with it.\n if ($checkPermission != $mergedPermission && ends_with($mergedPermission, $checkPermission) && $value === 1) {\n $matched = true;\n break;\n }\n }", "label_name": NaN, "label": NaN}
{"code": "function sqlrequest($database,$sql,$id=false){\n\n\t// Get the global value\n\tglobal $database_host;\n\tglobal $database_username;\n\tglobal $database_password;\n\n\t$connexion = mysqli_connect($database_host, $database_username, $database_password, $database);\n\tif (!$connexion) {\n\t\techo \"\";\n\t\techo \"- Alert EyesOfNetwork - Message EON-database connect
\";\n\t\techo \"- Could not connect to database : $database ($database_host)
\";\n\t\techo \"
\";\n\t\texit(1);\n\t}\n\n\tif ( $database == \"eonweb\" ) {\n\t\t// Force UTF-8\n\t\tmysqli_query($connexion, \"SET NAMES 'utf8'\");\n\t}\n\t$result=mysqli_query($connexion, \"$sql\");\n\n\tif($id==true)\n\t\t$result=mysqli_insert_id($connexion);\n\t\t\n\tmysqli_close($connexion);\n\treturn $result;\n}", "label_name": "Base", "label": 1.0}
{"code": "function list_process($bp,$display,$bdd){\n\t$sql = \"select name from bp where is_define = 1 and name!='\".$bp.\"' and priority = '\" . $display . \"'\";\n\t$req = $bdd->query($sql);\n\t$process = $req->fetchall();\n\n echo json_encode($process);\n}", "label_name": "Base", "label": 1.0}
{"code": "function edit($selected_events, $queue)\n{\n\tglobal $database_ged;\n\n\t// get all needed infos into variables\n\t$value_parts = explode(\":\", $selected_events);\n\t$id = $value_parts[0];\n\t$ged_type = $value_parts[1];\n\n\t$sql = \"SELECT comments FROM \".$ged_type.\"_queue_\".$queue.\" WHERE id = $id\";\n\t$result = sqlrequest($database_ged, $sql);\n\t$event = mysqli_fetch_assoc($result);\n\n\t$event[\"comments\"] = str_replace(\"\\'\", \"'\", $event[\"comments\"]);\n\t$event[\"comments\"] = str_replace(\"\\#\", \"#'\", $event[\"comments\"]);\n\n\techo \"\n\t\";\n}", "label_name": "Base", "label": 1.0}
{"code": " protected function generateVerifyCode()\n {\n if ($this->minLength > $this->maxLength) {\n $this->maxLength = $this->minLength;\n }\n if ($this->minLength < 3) {\n $this->minLength = 3;\n }\n if ($this->maxLength > 20) {\n $this->maxLength = 20;\n }\n $length = mt_rand($this->minLength, $this->maxLength);\n\n $letters = 'bcdfghjklmnpqrstvwxyz';\n $vowels = 'aeiou';\n $code = '';\n for ($i = 0; $i < $length; ++$i) {\n if ($i % 2 && mt_rand(0, 10) > 2 || !($i % 2) && mt_rand(0, 10) > 9) {\n $code .= $vowels[mt_rand(0, 4)];\n } else {\n $code .= $letters[mt_rand(0, 20)];\n }\n }\n\n return $code;\n }", "label_name": "Class", "label": 2.0}
-{"code": " public function testSandboxAllowMethodToString()\n {\n $twig = $this->getEnvironment(true, [], self::$templates, [], [], ['FooObject' => '__toString']);\n FooObject::reset();\n $this->assertEquals('foo', $twig->load('1_basic5')->render(self::$params), 'Sandbox allow some methods');\n $this->assertEquals(1, FooObject::$called['__toString'], 'Sandbox only calls method once');\n }", "label_name": NaN, "label": NaN}
{"code": "\tpublic function\tqueryLimit($cols, $table, $where=\"1=1\", $order=\"\", $offset=0, $max=100)\r\n\t{\t\t\r\n\t\t$this->lastError = '';\r\n\t\t$this->lastResult = '';\t\r\n\t\t$fetch = PDO::FETCH_ASSOC;\r\n\r\n\t\ttry\r\n\t\t{\r\n\t\t\t$sql = ' SELECT SQL_CALC_FOUND_ROWS '.$cols.'\r\n\t\t\t\t\t FROM '.$table.'\r\n\t\t\t\t\t WHERE '.$where.'\r\n \t\t\t\t ORDER BY '.$order.' \r\n\t\t\t\t\t LIMIT '.$max.'\r\n\t\t\t\t\t OFFSET '.$offset;\r\n\r\n\t\t\t$statement = $this->db->query($sql);\r\n $this->queries_count++;\r\n\t\t\t$statement->setFetchMode($fetch);\r\n\t\t\t$this->lastResult = $statement->fetchAll();\r\n\t\t\t$statement->closeCursor();\r\n\t\t\tunset($statement);\r\n\t\t}\r\n\t\tcatch(PDOException $e)\r\n\t\t{\r\n\t\t\t$this->lastError = $e->getMessage();\r\n\t\t}\r\n\t\tcatch(Exception $e)\r\n\t\t{\r\n\t\t\treturn false;\r\n\t\t}\t\t\r\n\t\t\r\n\t\treturn empty($this->lastError);\t\t\t\r\n\t}\r", "label_name": "Base", "label": 1.0}
{"code": " public static function remove_spam()\r\n {\r\n global $DB;\r\n global $website;\r\n\r\n $count = $DB->query_single(\r\n\t 'count(*) as total',\r\n\t 'nv_comments',\r\n\t 'website = '.protect($website->id).' AND status = 3'\r\n );\r\n\r\n $ok = $DB->execute('\r\n\t\t\tDELETE FROM nv_comments\r\n WHERE website = '.protect($website->id).'\r\n AND status = 3\r\n ');\r\n\r\n if($ok)\r\n return $count;\r\n }\r", "label_name": "Base", "label": 1.0}
{"code": " public static function pending_count()\r\n {\r\n global $DB;\r\n global $website;\r\n\r\n $pending_comments = $DB->query_single(\r\n 'COUNT(*)',\r\n 'nv_comments',\r\n ' website = '.protect($website->id).' AND\r\n status = -1'\r\n );\r\n\r\n return $pending_comments;\r\n }\r", "label_name": "Base", "label": 1.0}
@@ -652,8 +648,6 @@
{"code": " var fillTable = function(chart_data) {\n\n var tableRows = [],\n rowData, rowId;\n for(i=0; i' + rowData.label + '(id:'+ rowId +') | ');\n tableRows.push('' + rowData.data.filesizeformat() + ' | ');\n }\n \n $('#drivespaceTable tbody').html(tableRows.join(\"\"));\n $('#drivespaceTable').show();\n };", "label_name": "Base", "label": 1.0}
{"code": " var create_tag_html = function(text, description, owner, id, parent_id,\n is_tagset) {\n var cls = is_tagset ? \"alltags-tagset\" :\n (parent_id ? \"alltags-childtag\" : \"alltags-tag\");\n var html = \"\" + encode_html(text) + \"
\";\n return html;\n };", "label_name": "Base", "label": 1.0}
{"code": "String.prototype.escapeHTML = function(){\n /*\n HTML Escape Before Inserting Untrusted Data into HTML Element Content\n https://www.owasp.org/index.php/XSS_%28Cross_Site_Scripting%29_Prevention\n _Cheat_Sheet#RULE_.231_-_HTML_Escape_Before_Inserting_Untrusted_Data_into\n _HTML_Element_Content\n */\n var s = this;\n if (!s) {\n return \"\";\n }\n s = s + \"\";\n return s.replace(/[\\&\"<>\\\\]/g, function(s) {\n switch(s) {\n case \"&\": return \"&\";\n case \"\\\\\": return \"\\";\n case '\"': return '\\\"';\n case \"<\": return \"<\";\n case \">\": return \">\";\n default: return s;\n }\n });\n};", "label_name": "Base", "label": 1.0}
-{"code": "\t\t\t\t\t\tpattern: RegExp(/((?:^|[^\\\\])(?:\\\\\\\\)*)\\$\\{(?:)*?\\}/.source.replace(//g, function () { return FTL_EXPR; })),", "label_name": NaN, "label": NaN}
-{"code": "!function(e){e.languages.sass=e.languages.extend(\"css\",{comment:{pattern:/^([ \\t]*)\\/[\\/*].*(?:(?:\\r?\\n|\\r)\\1[ \\t]+.+)*/m,lookbehind:!0}}),e.languages.insertBefore(\"sass\",\"atrule\",{\"atrule-line\":{pattern:/^(?:[ \\t]*)[@+=].+/m,inside:{atrule:/(?:@[\\w-]+|[+=])/m}}}),delete e.languages.sass.atrule;var t=/\\$[-\\w]+|#\\{\\$[-\\w]+\\}/,a=[/[+*\\/%]|[=!]=|<=?|>=?|\\b(?:and|or|not)\\b/,{pattern:/(\\s+)-(?=\\s)/,lookbehind:!0}];e.languages.insertBefore(\"sass\",\"property\",{\"variable-line\":{pattern:/^[ \\t]*\\$.+/m,inside:{punctuation:/:/,variable:t,operator:a}},\"property-line\":{pattern:/^[ \\t]*(?:[^:\\s]+ *:.*|:[^:\\s]+.*)/m,inside:{property:[/[^:\\s]+(?=\\s*:)/,{pattern:/(:)[^:\\s]+/,lookbehind:!0}],punctuation:/:/,variable:t,operator:a,important:e.languages.sass.important}}}),delete e.languages.sass.property,delete e.languages.sass.important,e.languages.insertBefore(\"sass\",\"punctuation\",{selector:{pattern:/([ \\t]*)\\S(?:,?[^,\\r\\n]+)*(?:,(?:\\r?\\n|\\r)\\1[ \\t]+\\S(?:,?[^,\\r\\n]+)*)*/,lookbehind:!0}})}(Prism);", "label_name": NaN, "label": NaN}
{"code": " select: function (context, coverage, statMode) {\n coverage = parseFloat(coverage)\n var usage = browserslist.usage.global\n if (statMode) {\n if (statMode.match(/^\\s+in\\s+my\\s+stats$/)) {\n if (!context.customUsage) {\n throw new BrowserslistError(\n 'Custom usage statistics was not provided'\n )\n }\n usage = context.customUsage\n } else {\n var match = statMode.match(/\\s+in\\s+((alt-)?\\w\\w)/)\n var place = match[1]\n if (place.length === 2) {\n place = place.toUpperCase()\n } else {\n place = place.toLowerCase()\n }\n env.loadCountry(browserslist.usage, place, browserslist.data)\n usage = browserslist.usage[place]\n }\n }\n var versions = Object.keys(usage).sort(function (a, b) {\n return usage[b] - usage[a]\n })\n var coveraged = 0\n var result = []\n var version\n for (var i = 0; i <= versions.length; i++) {\n version = versions[i]\n if (usage[version] === 0) break\n coveraged += usage[version]\n result.push(version)\n if (coveraged >= coverage) break\n }\n return result\n }", "label_name": NaN, "label": NaN}
{"code": " select: function (context, version) {\n var versionToUse = normalizeElectron(version)\n var chrome = e2c[versionToUse]\n if (!chrome) {\n throw new BrowserslistError(\n 'Unknown version ' + version + ' of electron')\n }\n return ['chrome ' + chrome]\n }", "label_name": NaN, "label": NaN}
{"code": "exports.get = function(obj, path) {\n\n\tvar cachekey = 'G=' + path;\n\n\tif (F.temporary.other[cachekey])\n\t\treturn F.temporary.other[cachekey](obj);\n\n\tvar arr = parsepath(path);\n\tvar builder = [];\n\n\tfor (var i = 0, length = arr.length - 1; i < length; i++)\n\t\tbuilder.push('if(!w' + (!arr[i] || arr[i][0] === '[' ? '' : '.') + arr[i] + ')return');\n\n\tvar v = arr[arr.length - 1];\n\tvar fn = (new Function('w', builder.join(';') + ';return w' + (v[0] === '[' ? '' : '.') + v));\n\tF.temporary.other[cachekey] = fn;\n\treturn fn(obj);\n};", "label_name": "Base", "label": 1.0}
@@ -1001,7 +995,6 @@
{"code": "iakerb_gss_export_sec_context(OM_uint32 *minor_status,\n gss_ctx_id_t *context_handle,\n gss_buffer_t interprocess_token)\n{\n OM_uint32 maj;\n iakerb_ctx_id_t ctx = (iakerb_ctx_id_t)context_handle;\n\n /* We don't currently support exporting partially established contexts. */\n if (!ctx->established)\n return GSS_S_UNAVAILABLE;\n\n maj = krb5_gss_export_sec_context(minor_status, &ctx->gssc,\n interprocess_token);\n if (ctx->gssc == GSS_C_NO_CONTEXT) {\n iakerb_release_context(ctx);\n *context_handle = GSS_C_NO_CONTEXT;\n }\n return maj;\n}", "label_name": "Class", "label": 2.0}
{"code": "getprivs_ret * get_privs_2_svc(krb5_ui_4 *arg, struct svc_req *rqstp)\n{\n static getprivs_ret ret;\n gss_buffer_desc client_name, service_name;\n OM_uint32 minor_stat;\n kadm5_server_handle_t handle;\n const char *errmsg = NULL;\n\n xdr_free(xdr_getprivs_ret, &ret);\n\n if ((ret.code = new_server_handle(*arg, rqstp, &handle)))\n goto exit_func;\n\n if ((ret.code = check_handle((void *)handle)))\n goto exit_func;\n\n ret.api_version = handle->api_version;\n\n if (setup_gss_names(rqstp, &client_name, &service_name) < 0) {\n ret.code = KADM5_FAILURE;\n goto exit_func;\n }\n\n ret.code = kadm5_get_privs((void *)handle, &ret.privs);\n if( ret.code != 0 )\n errmsg = krb5_get_error_message(handle->context, ret.code);\n\n log_done(\"kadm5_get_privs\", client_name.value, errmsg,\n &client_name, &service_name, rqstp);\n\n if (errmsg != NULL)\n krb5_free_error_message(handle->context, errmsg);\n\n gss_release_buffer(&minor_stat, &client_name);\n gss_release_buffer(&minor_stat, &service_name);\nexit_func:\n free_server_handle(handle);\n return &ret;\n}", "label_name": "Base", "label": 1.0}
{"code": "decode_sequence(const uint8_t *asn1, size_t len, const struct seq_info *seq,\n void *val)\n{\n krb5_error_code ret;\n const uint8_t *contents;\n size_t i, j, clen;\n taginfo t;\n\n assert(seq->n_fields > 0);\n for (i = 0; i < seq->n_fields; i++) {\n if (len == 0)\n break;\n ret = get_tag(asn1, len, &t, &contents, &clen, &asn1, &len);\n if (ret)\n goto error;\n /*\n * Find the applicable sequence field. This logic is a little\n * oversimplified; we could match an element to an optional extensible\n * choice or optional stored-DER type when we ought to match a\n * subsequent non-optional field. But it's unwise and (hopefully) very\n * rare for ASN.1 modules to require such precision.\n */\n for (; i < seq->n_fields; i++) {\n if (check_atype_tag(seq->fields[i], &t))\n break;\n ret = omit_atype(seq->fields[i], val);\n if (ret)\n goto error;\n }\n /* We currently model all sequences as extensible. We should consider\n * changing this before making the encoder visible to plugins. */\n if (i == seq->n_fields)\n break;\n ret = decode_atype(&t, contents, clen, seq->fields[i], val);\n if (ret)\n goto error;\n }\n /* Initialize any fields in the C object which were not accounted for in\n * the sequence. Error out if any of them aren't optional. */\n for (; i < seq->n_fields; i++) {\n ret = omit_atype(seq->fields[i], val);\n if (ret)\n goto error;\n }\n return 0;\n\nerror:\n /* Free what we've decoded so far. Free pointers in a second pass in\n * case multiple fields refer to the same pointer. */\n for (j = 0; j < i; j++)\n free_atype(seq->fields[j], val);\n for (j = 0; j < i; j++)\n free_atype_ptr(seq->fields[j], val);\n return ret;\n}", "label_name": "Class", "label": 2.0}
-{"code": "static void frag_kfree_skb(struct netns_frags *nf, struct sk_buff *skb)\n{\n\tatomic_sub(skb->truesize, &nf->mem);\n\tkfree_skb(skb);\n}", "label_name": NaN, "label": NaN}
{"code": "asmlinkage void __sched schedule(void)\n{\n\tstruct task_struct *prev, *next;\n\tunsigned long *switch_count;\n\tstruct rq *rq;\n\tint cpu;\n\nneed_resched:\n\tpreempt_disable();\n\tcpu = smp_processor_id();\n\trq = cpu_rq(cpu);\n\trcu_note_context_switch(cpu);\n\tprev = rq->curr;\n\n\trelease_kernel_lock(prev);\nneed_resched_nonpreemptible:\n\n\tschedule_debug(prev);\n\n\tif (sched_feat(HRTICK))\n\t\thrtick_clear(rq);\n\n\traw_spin_lock_irq(&rq->lock);\n\tclear_tsk_need_resched(prev);\n\n\tswitch_count = &prev->nivcsw;\n\tif (prev->state && !(preempt_count() & PREEMPT_ACTIVE)) {\n\t\tif (unlikely(signal_pending_state(prev->state, prev))) {\n\t\t\tprev->state = TASK_RUNNING;\n\t\t} else {\n\t\t\t/*\n\t\t\t * If a worker is going to sleep, notify and\n\t\t\t * ask workqueue whether it wants to wake up a\n\t\t\t * task to maintain concurrency. If so, wake\n\t\t\t * up the task.\n\t\t\t */\n\t\t\tif (prev->flags & PF_WQ_WORKER) {\n\t\t\t\tstruct task_struct *to_wakeup;\n\n\t\t\t\tto_wakeup = wq_worker_sleeping(prev, cpu);\n\t\t\t\tif (to_wakeup)\n\t\t\t\t\ttry_to_wake_up_local(to_wakeup);\n\t\t\t}\n\t\t\tdeactivate_task(rq, prev, DEQUEUE_SLEEP);\n\t\t}\n\t\tswitch_count = &prev->nvcsw;\n\t}\n\n\tpre_schedule(rq, prev);\n\n\tif (unlikely(!rq->nr_running))\n\t\tidle_balance(cpu, rq);\n\n\tput_prev_task(rq, prev);\n\tnext = pick_next_task(rq);\n\n\tif (likely(prev != next)) {\n\t\tsched_info_switch(prev, next);\n\t\tperf_event_task_sched_out(prev, next);\n\n\t\trq->nr_switches++;\n\t\trq->curr = next;\n\t\t++*switch_count;\n\n\t\tcontext_switch(rq, prev, next); /* unlocks the rq */\n\t\t/*\n\t\t * The context switch have flipped the stack from under us\n\t\t * and restored the local variables which were saved when\n\t\t * this task called schedule() in the past. prev == current\n\t\t * is still correct, but it can be moved to another cpu/rq.\n\t\t */\n\t\tcpu = smp_processor_id();\n\t\trq = cpu_rq(cpu);\n\t} else\n\t\traw_spin_unlock_irq(&rq->lock);\n\n\tpost_schedule(rq);\n\n\tif (unlikely(reacquire_kernel_lock(prev)))\n\t\tgoto need_resched_nonpreemptible;\n\n\tpreempt_enable_no_resched();\n\tif (need_resched())\n\t\tgoto need_resched;\n}", "label_name": "Base", "label": 1.0}
{"code": "int ip_build_and_send_pkt(struct sk_buff *skb, struct sock *sk,\n\t\t\t __be32 saddr, __be32 daddr, struct ip_options *opt)\n{\n\tstruct inet_sock *inet = inet_sk(sk);\n\tstruct rtable *rt = skb_rtable(skb);\n\tstruct iphdr *iph;\n\n\t/* Build the IP header. */\n\tskb_push(skb, sizeof(struct iphdr) + (opt ? opt->optlen : 0));\n\tskb_reset_network_header(skb);\n\tiph = ip_hdr(skb);\n\tiph->version = 4;\n\tiph->ihl = 5;\n\tiph->tos = inet->tos;\n\tif (ip_dont_fragment(sk, &rt->dst))\n\t\tiph->frag_off = htons(IP_DF);\n\telse\n\t\tiph->frag_off = 0;\n\tiph->ttl = ip_select_ttl(inet, &rt->dst);\n\tiph->daddr = rt->rt_dst;\n\tiph->saddr = rt->rt_src;\n\tiph->protocol = sk->sk_protocol;\n\tip_select_ident(iph, &rt->dst, sk);\n\n\tif (opt && opt->optlen) {\n\t\tiph->ihl += opt->optlen>>2;\n\t\tip_options_build(skb, opt, daddr, rt, 0);\n\t}\n\n\tskb->priority = sk->sk_priority;\n\tskb->mark = sk->sk_mark;\n\n\t/* Send it out. */\n\treturn ip_local_out(skb);\n}", "label_name": "Class", "label": 2.0}
{"code": "static struct ip_options *tcp_v4_save_options(struct sock *sk,\n\t\t\t\t\t struct sk_buff *skb)\n{\n\tstruct ip_options *opt = &(IPCB(skb)->opt);\n\tstruct ip_options *dopt = NULL;\n\n\tif (opt && opt->optlen) {\n\t\tint opt_size = optlength(opt);\n\t\tdopt = kmalloc(opt_size, GFP_ATOMIC);\n\t\tif (dopt) {\n\t\t\tif (ip_options_echo(dopt, skb)) {\n\t\t\t\tkfree(dopt);\n\t\t\t\tdopt = NULL;\n\t\t\t}\n\t\t}\n\t}\n\treturn dopt;\n}", "label_name": "Class", "label": 2.0}
@@ -1065,15 +1058,11 @@
{"code": "static inline bool key_is_instantiated(const struct key *key)\n{\n\treturn test_bit(KEY_FLAG_INSTANTIATED, &key->flags) &&\n\t\t!test_bit(KEY_FLAG_NEGATIVE, &key->flags);\n}", "label_name": "Class", "label": 2.0}
{"code": "static int shash_no_setkey(struct crypto_shash *tfm, const u8 *key,\n\t\t\t unsigned int keylen)\n{\n\treturn -ENOSYS;\n}", "label_name": "Base", "label": 1.0}
{"code": "static int encrypt(struct blkcipher_desc *desc,\n\t\t struct scatterlist *dst, struct scatterlist *src,\n\t\t unsigned int nbytes)\n{\n\tstruct blkcipher_walk walk;\n\tstruct crypto_blkcipher *tfm = desc->tfm;\n\tstruct salsa20_ctx *ctx = crypto_blkcipher_ctx(tfm);\n\tint err;\n\n\tblkcipher_walk_init(&walk, dst, src, nbytes);\n\terr = blkcipher_walk_virt_block(desc, &walk, 64);\n\n\tsalsa20_ivsetup(ctx, walk.iv);\n\n\tif (likely(walk.nbytes == nbytes))\n\t{\n\t\tsalsa20_encrypt_bytes(ctx, walk.src.virt.addr,\n\t\t\t\t walk.dst.virt.addr, nbytes);\n\t\treturn blkcipher_walk_done(desc, &walk, 0);\n\t}\n\n\twhile (walk.nbytes >= 64) {\n\t\tsalsa20_encrypt_bytes(ctx, walk.src.virt.addr,\n\t\t\t\t walk.dst.virt.addr,\n\t\t\t\t walk.nbytes - (walk.nbytes % 64));\n\t\terr = blkcipher_walk_done(desc, &walk, walk.nbytes % 64);\n\t}\n\n\tif (walk.nbytes) {\n\t\tsalsa20_encrypt_bytes(ctx, walk.src.virt.addr,\n\t\t\t\t walk.dst.virt.addr, walk.nbytes);\n\t\terr = blkcipher_walk_done(desc, &walk, 0);\n\t}\n\n\treturn err;\n}", "label_name": "Class", "label": 2.0}
-{"code": "static void sas_eh_defer_cmd(struct scsi_cmnd *cmd)\n{\n\tstruct domain_device *dev = cmd_to_domain_dev(cmd);\n\tstruct sas_ha_struct *ha = dev->port->ha;\n\tstruct sas_task *task = TO_SAS_TASK(cmd);\n\n\tif (!dev_is_sata(dev)) {\n\t\tsas_eh_finish_cmd(cmd);\n\t\treturn;\n\t}\n\n\t/* report the timeout to libata */\n\tsas_end_task(cmd, task);\n\tlist_move_tail(&cmd->eh_entry, &ha->eh_ata_q);\n}", "label_name": NaN, "label": NaN}
{"code": "static void vgacon_scrollback_reset(int vc_num, size_t reset_size)\n{\n\tstruct vgacon_scrollback_info *scrollback = &vgacon_scrollbacks[vc_num];\n\n\tif (scrollback->data && reset_size > 0)\n\t\tmemset(scrollback->data, 0, reset_size);\n\n\tscrollback->cnt = 0;\n\tscrollback->tail = 0;\n\tscrollback->cur = 0;\n}", "label_name": "Base", "label": 1.0}
{"code": "static void vgacon_flush_scrollback(struct vc_data *c)\n{\n\tsize_t size = CONFIG_VGACON_SOFT_SCROLLBACK_SIZE * 1024;\n\n\tvgacon_scrollback_reset(c->vc_num, size);\n}", "label_name": "Base", "label": 1.0}
{"code": "static void clear_evtchn_to_irq_row(unsigned row)\n{\n\tunsigned col;\n\n\tfor (col = 0; col < EVTCHN_PER_ROW; col++)\n\t\tevtchn_to_irq[row][col] = -1;\n}", "label_name": "Class", "label": 2.0}
{"code": "static void clear_evtchn_to_irq_row(unsigned row)\n{\n\tunsigned col;\n\n\tfor (col = 0; col < EVTCHN_PER_ROW; col++)\n\t\tevtchn_to_irq[row][col] = -1;\n}", "label_name": "Variant", "label": 0.0}
-{"code": "static void xen_irq_lateeoi_locked(struct irq_info *info)\n{\n\tevtchn_port_t evtchn;\n\n\tevtchn = info->evtchn;\n\tif (!VALID_EVTCHN(evtchn))\n\t\treturn;\n\n\tunmask_evtchn(evtchn);\n}", "label_name": NaN, "label": NaN}
-{"code": "static int evtchn_fifo_percpu_deinit(unsigned int cpu)\n{\n\t__evtchn_fifo_handle_events(cpu, true);\n\treturn 0;\n}", "label_name": NaN, "label": NaN}
{"code": "static int list_devices(struct file *filp, struct dm_ioctl *param, size_t param_size)\n{\n\tunsigned int i;\n\tstruct hash_cell *hc;\n\tsize_t len, needed = 0;\n\tstruct gendisk *disk;\n\tstruct dm_name_list *orig_nl, *nl, *old_nl = NULL;\n\tuint32_t *event_nr;\n\n\tdown_write(&_hash_lock);\n\n\t/*\n\t * Loop through all the devices working out how much\n\t * space we need.\n\t */\n\tfor (i = 0; i < NUM_BUCKETS; i++) {\n\t\tlist_for_each_entry (hc, _name_buckets + i, name_list) {\n\t\t\tneeded += align_val(offsetof(struct dm_name_list, name) + strlen(hc->name) + 1);\n\t\t\tneeded += align_val(sizeof(uint32_t));\n\t\t}\n\t}\n\n\t/*\n\t * Grab our output buffer.\n\t */\n\tnl = orig_nl = get_result_buffer(param, param_size, &len);\n\tif (len < needed) {\n\t\tparam->flags |= DM_BUFFER_FULL_FLAG;\n\t\tgoto out;\n\t}\n\tparam->data_size = param->data_start + needed;\n\n\tnl->dev = 0;\t/* Flags no data */\n\n\t/*\n\t * Now loop through filling out the names.\n\t */\n\tfor (i = 0; i < NUM_BUCKETS; i++) {\n\t\tlist_for_each_entry (hc, _name_buckets + i, name_list) {\n\t\t\tif (old_nl)\n\t\t\t\told_nl->next = (uint32_t) ((void *) nl -\n\t\t\t\t\t\t\t (void *) old_nl);\n\t\t\tdisk = dm_disk(hc->md);\n\t\t\tnl->dev = huge_encode_dev(disk_devt(disk));\n\t\t\tnl->next = 0;\n\t\t\tstrcpy(nl->name, hc->name);\n\n\t\t\told_nl = nl;\n\t\t\tevent_nr = align_ptr(nl->name + strlen(hc->name) + 1);\n\t\t\t*event_nr = dm_get_event_nr(hc->md);\n\t\t\tnl = align_ptr(event_nr + 1);\n\t\t}\n\t}\n\t/*\n\t * If mismatch happens, security may be compromised due to buffer\n\t * overflow, so it's better to crash.\n\t */\n\tBUG_ON((char *)nl - (char *)orig_nl != needed);\n\n out:\n\tup_write(&_hash_lock);\n\treturn 0;\n}", "label_name": "Base", "label": 1.0}
-{"code": "struct nfs_client *nfs4_alloc_client(const struct nfs_client_initdata *cl_init)\n{\n\tint err;\n\tstruct nfs_client *clp = nfs_alloc_client(cl_init);\n\tif (IS_ERR(clp))\n\t\treturn clp;\n\n\terr = nfs_get_cb_ident_idr(clp, cl_init->minorversion);\n\tif (err)\n\t\tgoto error;\n\n\tif (cl_init->minorversion > NFS4_MAX_MINOR_VERSION) {\n\t\terr = -EINVAL;\n\t\tgoto error;\n\t}\n\n\tspin_lock_init(&clp->cl_lock);\n\tINIT_DELAYED_WORK(&clp->cl_renewd, nfs4_renew_state);\n\tINIT_LIST_HEAD(&clp->cl_ds_clients);\n\trpc_init_wait_queue(&clp->cl_rpcwaitq, \"NFS client\");\n\tclp->cl_state = 1 << NFS4CLNT_LEASE_EXPIRED;\n\tclp->cl_mvops = nfs_v4_minor_ops[cl_init->minorversion];\n\tclp->cl_mig_gen = 1;\n#if IS_ENABLED(CONFIG_NFS_V4_1)\n\tinit_waitqueue_head(&clp->cl_lock_waitq);\n#endif\n\tINIT_LIST_HEAD(&clp->pending_cb_stateids);\n\treturn clp;\n\nerror:\n\tnfs_free_client(clp);\n\treturn ERR_PTR(err);\n}", "label_name": NaN, "label": NaN}
{"code": "static int vt_kdsetmode(struct vc_data *vc, unsigned long mode)\n{\n\tswitch (mode) {\n\tcase KD_GRAPHICS:\n\t\tbreak;\n\tcase KD_TEXT0:\n\tcase KD_TEXT1:\n\t\tmode = KD_TEXT;\n\t\tfallthrough;\n\tcase KD_TEXT:\n\t\tbreak;\n\tdefault:\n\t\treturn -EINVAL;\n\t}\n\n\t/* FIXME: this needs the console lock extending */\n\tif (vc->vc_mode == mode)\n\t\treturn 0;\n\n\tvc->vc_mode = mode;\n\tif (vc->vc_num != fg_console)\n\t\treturn 0;\n\n\t/* explicitly blank/unblank the screen if switching modes */\n\tconsole_lock();\n\tif (mode == KD_TEXT)\n\t\tdo_unblank_screen(1);\n\telse\n\t\tdo_blank_screen(1);\n\tconsole_unlock();\n\n\treturn 0;\n}", "label_name": "Class", "label": 2.0}
{"code": "static bool tipc_crypto_key_rcv(struct tipc_crypto *rx, struct tipc_msg *hdr)\n{\n\tstruct tipc_crypto *tx = tipc_net(rx->net)->crypto_tx;\n\tstruct tipc_aead_key *skey = NULL;\n\tu16 key_gen = msg_key_gen(hdr);\n\tu16 size = msg_data_sz(hdr);\n\tu8 *data = msg_data(hdr);\n\n\tspin_lock(&rx->lock);\n\tif (unlikely(rx->skey || (key_gen == rx->key_gen && rx->key.keys))) {\n\t\tpr_err(\"%s: key existed <%p>, gen %d vs %d\\n\", rx->name,\n\t\t rx->skey, key_gen, rx->key_gen);\n\t\tgoto exit;\n\t}\n\n\t/* Allocate memory for the key */\n\tskey = kmalloc(size, GFP_ATOMIC);\n\tif (unlikely(!skey)) {\n\t\tpr_err(\"%s: unable to allocate memory for skey\\n\", rx->name);\n\t\tgoto exit;\n\t}\n\n\t/* Copy key from msg data */\n\tskey->keylen = ntohl(*((__be32 *)(data + TIPC_AEAD_ALG_NAME)));\n\tmemcpy(skey->alg_name, data, TIPC_AEAD_ALG_NAME);\n\tmemcpy(skey->key, data + TIPC_AEAD_ALG_NAME + sizeof(__be32),\n\t skey->keylen);\n\n\t/* Sanity check */\n\tif (unlikely(size != tipc_aead_key_size(skey))) {\n\t\tkfree(skey);\n\t\tskey = NULL;\n\t\tgoto exit;\n\t}\n\n\trx->key_gen = key_gen;\n\trx->skey_mode = msg_key_mode(hdr);\n\trx->skey = skey;\n\trx->nokey = 0;\n\tmb(); /* for nokey flag */\n\nexit:\n\tspin_unlock(&rx->lock);\n\n\t/* Schedule the key attaching on this crypto */\n\tif (likely(skey && queue_delayed_work(tx->wq, &rx->work, 0)))\n\t\treturn true;\n\n\treturn false;\n}", "label_name": "Class", "label": 2.0}
{"code": "void jbd2_journal_lock_updates(journal_t *journal)\n{\n\tDEFINE_WAIT(wait);\n\n\tjbd2_might_wait_for_commit(journal);\n\n\twrite_lock(&journal->j_state_lock);\n\t++journal->j_barrier_count;\n\n\t/* Wait until there are no reserved handles */\n\tif (atomic_read(&journal->j_reserved_credits)) {\n\t\twrite_unlock(&journal->j_state_lock);\n\t\twait_event(journal->j_wait_reserved,\n\t\t\t atomic_read(&journal->j_reserved_credits) == 0);\n\t\twrite_lock(&journal->j_state_lock);\n\t}\n\n\t/* Wait until there are no running t_updates */\n\tjbd2_journal_wait_updates(journal);\n\n\twrite_unlock(&journal->j_state_lock);\n\n\t/*\n\t * We have now established a barrier against other normal updates, but\n\t * we also need to barrier against other jbd2_journal_lock_updates() calls\n\t * to make sure that we serialise special journal-locked operations\n\t * too.\n\t */\n\tmutex_lock(&journal->j_barrier);\n}", "label_name": "Variant", "label": 0.0}
@@ -1081,8 +1070,6 @@
{"code": "void rose_start_t2timer(struct sock *sk)\n{\n\tstruct rose_sock *rose = rose_sk(sk);\n\n\tdel_timer(&rose->timer);\n\n\trose->timer.function = rose_timer_expiry;\n\trose->timer.expires = jiffies + rose->t2;\n\n\tadd_timer(&rose->timer);\n}", "label_name": "Variant", "label": 0.0}
{"code": "static int sgi_clock_set(clockid_t clockid, struct timespec *tp)\n{\n\n\tu64 nsec;\n\tu64 rem;\n\n\tnsec = rtc_time() * sgi_clock_period;\n\n\tsgi_clock_offset.tv_sec = tp->tv_sec - div_long_long_rem(nsec, NSEC_PER_SEC, &rem);\n\n\tif (rem <= tp->tv_nsec)\n\t\tsgi_clock_offset.tv_nsec = tp->tv_sec - rem;\n\telse {\n\t\tsgi_clock_offset.tv_nsec = tp->tv_sec + NSEC_PER_SEC - rem;\n\t\tsgi_clock_offset.tv_sec--;\n\t}\n\treturn 0;\n}", "label_name": NaN, "label": NaN}
{"code": "static inline long div_ll_X_l_rem(long long divs, long div, long *rem)\n{\n\tlong dum2;\n\tasm(\"divl %2\":\"=a\"(dum2), \"=d\"(*rem)\n\t : \"rm\"(div), \"A\"(divs));\n\n\treturn dum2;\n\n}", "label_name": NaN, "label": NaN}
-{"code": "static void nfs4_open_confirm_release(void *calldata)\n{\n\tstruct nfs4_opendata *data = calldata;\n\tstruct nfs4_state *state = NULL;\n\n\t/* If this request hasn't been cancelled, do nothing */\n\tif (data->cancelled == 0)\n\t\tgoto out_free;\n\t/* In case of error, no cleanup! */\n\tif (!data->rpc_done)\n\t\tgoto out_free;\n\tstate = nfs4_opendata_to_nfs4_state(data);\n\tif (!IS_ERR(state))\n\t\tnfs4_close_state(&data->path, state, data->o_arg.open_flags);\nout_free:\n\tnfs4_opendata_put(data);\n}", "label_name": NaN, "label": NaN}
-{"code": "static void nfs4_open_prepare(struct rpc_task *task, void *calldata)\n{\n\tstruct nfs4_opendata *data = calldata;\n\tstruct nfs4_state_owner *sp = data->owner;\n\n\tif (nfs_wait_on_sequence(data->o_arg.seqid, task) != 0)\n\t\treturn;\n\t/*\n\t * Check if we still need to send an OPEN call, or if we can use\n\t * a delegation instead.\n\t */\n\tif (data->state != NULL) {\n\t\tstruct nfs_delegation *delegation;\n\n\t\tif (can_open_cached(data->state, data->o_arg.open_flags & (FMODE_READ|FMODE_WRITE|O_EXCL)))\n\t\t\tgoto out_no_action;\n\t\trcu_read_lock();\n\t\tdelegation = rcu_dereference(NFS_I(data->state->inode)->delegation);\n\t\tif (delegation != NULL &&\n\t\t test_bit(NFS_DELEGATION_NEED_RECLAIM, &delegation->flags) == 0) {\n\t\t\trcu_read_unlock();\n\t\t\tgoto out_no_action;\n\t\t}\n\t\trcu_read_unlock();\n\t}\n\t/* Update sequence id. */\n\tdata->o_arg.id = sp->so_owner_id.id;\n\tdata->o_arg.clientid = sp->so_client->cl_clientid;\n\tif (data->o_arg.claim == NFS4_OPEN_CLAIM_PREVIOUS) {\n\t\ttask->tk_msg.rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_OPEN_NOATTR];\n\t\tnfs_copy_fh(&data->o_res.fh, data->o_arg.fh);\n\t}\n\tdata->timestamp = jiffies;\n\trpc_call_start(task);\n\treturn;\nout_no_action:\n\ttask->tk_action = NULL;\n\n}", "label_name": NaN, "label": NaN}
{"code": "static int __init sit_init(void)\n{\n\tint err;\n\n\tprintk(KERN_INFO \"IPv6 over IPv4 tunneling driver\\n\");\n\n\tif (xfrm4_tunnel_register(&sit_handler, AF_INET6) < 0) {\n\t\tprintk(KERN_INFO \"sit init: Can't add protocol\\n\");\n\t\treturn -EAGAIN;\n\t}\n\n\terr = register_pernet_device(&sit_net_ops);\n\tif (err < 0)\n\t\txfrm4_tunnel_deregister(&sit_handler, AF_INET6);\n\n\treturn err;\n}", "label_name": "Class", "label": 2.0}
{"code": "static int __init xfrm6_tunnel_spi_init(void)\n{\n\txfrm6_tunnel_spi_kmem = kmem_cache_create(\"xfrm6_tunnel_spi\",\n\t\t\t\t\t\t sizeof(struct xfrm6_tunnel_spi),\n\t\t\t\t\t\t 0, SLAB_HWCACHE_ALIGN,\n\t\t\t\t\t\t NULL);\n\tif (!xfrm6_tunnel_spi_kmem)\n\t\treturn -ENOMEM;\n\treturn 0;\n}", "label_name": "Class", "label": 2.0}
{"code": "static int iscsi_add_notunderstood_response(\n\tchar *key,\n\tchar *value,\n\tstruct iscsi_param_list *param_list)\n{\n\tstruct iscsi_extra_response *extra_response;\n\n\tif (strlen(value) > VALUE_MAXLEN) {\n\t\tpr_err(\"Value for notunderstood key \\\"%s\\\" exceeds %d,\"\n\t\t\t\" protocol error.\\n\", key, VALUE_MAXLEN);\n\t\treturn -1;\n\t}\n\n\textra_response = kzalloc(sizeof(struct iscsi_extra_response), GFP_KERNEL);\n\tif (!extra_response) {\n\t\tpr_err(\"Unable to allocate memory for\"\n\t\t\t\" struct iscsi_extra_response.\\n\");\n\t\treturn -1;\n\t}\n\tINIT_LIST_HEAD(&extra_response->er_list);\n\n\tstrncpy(extra_response->key, key, strlen(key) + 1);\n\tstrncpy(extra_response->value, NOTUNDERSTOOD,\n\t\t\tstrlen(NOTUNDERSTOOD) + 1);\n\n\tlist_add_tail(&extra_response->er_list,\n\t\t\t¶m_list->extra_response_list);\n\treturn 0;\n}", "label_name": "Class", "label": 2.0}
@@ -1124,7 +1111,6 @@
{"code": "static inline int crypto_rng_seedsize(struct crypto_rng *tfm)\n{\n\treturn tfm->seedsize;\n}", "label_name": "Base", "label": 1.0}
{"code": "static void oz_usb_handle_ep_data(struct oz_usb_ctx *usb_ctx,\n\tstruct oz_usb_hdr *usb_hdr, int len)\n{\n\tstruct oz_data *data_hdr = (struct oz_data *)usb_hdr;\n\n\tswitch (data_hdr->format) {\n\tcase OZ_DATA_F_MULTIPLE_FIXED: {\n\t\t\tstruct oz_multiple_fixed *body =\n\t\t\t\t(struct oz_multiple_fixed *)data_hdr;\n\t\t\tu8 *data = body->data;\n\t\t\tint n;\n\t\t\tif (!body->unit_size)\n\t\t\t\tbreak;\n\t\t\tn = (len - sizeof(struct oz_multiple_fixed)+1)\n\t\t\t\t/ body->unit_size;\n\t\t\twhile (n--) {\n\t\t\t\toz_hcd_data_ind(usb_ctx->hport, body->endpoint,\n\t\t\t\t\tdata, body->unit_size);\n\t\t\t\tdata += body->unit_size;\n\t\t\t}\n\t\t}\n\t\tbreak;\n\tcase OZ_DATA_F_ISOC_FIXED: {\n\t\t\tstruct oz_isoc_fixed *body =\n\t\t\t\t(struct oz_isoc_fixed *)data_hdr;\n\t\t\tint data_len = len-sizeof(struct oz_isoc_fixed)+1;\n\t\t\tint unit_size = body->unit_size;\n\t\t\tu8 *data = body->data;\n\t\t\tint count;\n\t\t\tint i;\n\n\t\t\tif (!unit_size)\n\t\t\t\tbreak;\n\t\t\tcount = data_len/unit_size;\n\t\t\tfor (i = 0; i < count; i++) {\n\t\t\t\toz_hcd_data_ind(usb_ctx->hport,\n\t\t\t\t\tbody->endpoint, data, unit_size);\n\t\t\t\tdata += unit_size;\n\t\t\t}\n\t\t}\n\t\tbreak;\n\t}\n\n}", "label_name": "Class", "label": 2.0}
{"code": "static int rfcomm_sock_bind(struct socket *sock, struct sockaddr *addr, int addr_len)\n{\n\tstruct sockaddr_rc *sa = (struct sockaddr_rc *) addr;\n\tstruct sock *sk = sock->sk;\n\tint chan = sa->rc_channel;\n\tint err = 0;\n\n\tBT_DBG(\"sk %p %pMR\", sk, &sa->rc_bdaddr);\n\n\tif (!addr || addr->sa_family != AF_BLUETOOTH)\n\t\treturn -EINVAL;\n\n\tlock_sock(sk);\n\n\tif (sk->sk_state != BT_OPEN) {\n\t\terr = -EBADFD;\n\t\tgoto done;\n\t}\n\n\tif (sk->sk_type != SOCK_STREAM) {\n\t\terr = -EINVAL;\n\t\tgoto done;\n\t}\n\n\twrite_lock(&rfcomm_sk_list.lock);\n\n\tif (chan && __rfcomm_get_listen_sock_by_addr(chan, &sa->rc_bdaddr)) {\n\t\terr = -EADDRINUSE;\n\t} else {\n\t\t/* Save source address */\n\t\tbacpy(&rfcomm_pi(sk)->src, &sa->rc_bdaddr);\n\t\trfcomm_pi(sk)->channel = chan;\n\t\tsk->sk_state = BT_BOUND;\n\t}\n\n\twrite_unlock(&rfcomm_sk_list.lock);\n\ndone:\n\trelease_sock(sk);\n\treturn err;\n}", "label_name": "Base", "label": 1.0}
-{"code": "static bool __init is_skylake_era(void)\n{\n\tif (boot_cpu_data.x86_vendor == X86_VENDOR_INTEL &&\n\t boot_cpu_data.x86 == 6) {\n\t\tswitch (boot_cpu_data.x86_model) {\n\t\tcase INTEL_FAM6_SKYLAKE_MOBILE:\n\t\tcase INTEL_FAM6_SKYLAKE_DESKTOP:\n\t\tcase INTEL_FAM6_SKYLAKE_X:\n\t\tcase INTEL_FAM6_KABYLAKE_MOBILE:\n\t\tcase INTEL_FAM6_KABYLAKE_DESKTOP:\n\t\t\treturn true;\n\t\t}\n\t}\n\treturn false;\n}", "label_name": NaN, "label": NaN}
{"code": "void vmacache_flush_all(struct mm_struct *mm)\n{\n\tstruct task_struct *g, *p;\n\n\tcount_vm_vmacache_event(VMACACHE_FULL_FLUSHES);\n\n\t/*\n\t * Single threaded tasks need not iterate the entire\n\t * list of process. We can avoid the flushing as well\n\t * since the mm's seqnum was increased and don't have\n\t * to worry about other threads' seqnum. Current's\n\t * flush will occur upon the next lookup.\n\t */\n\tif (atomic_read(&mm->mm_users) == 1)\n\t\treturn;\n\n\trcu_read_lock();\n\tfor_each_process_thread(g, p) {\n\t\t/*\n\t\t * Only flush the vmacache pointers as the\n\t\t * mm seqnum is already set and curr's will\n\t\t * be set upon invalidation when the next\n\t\t * lookup is done.\n\t\t */\n\t\tif (mm == p->mm)\n\t\t\tvmacache_flush(p);\n\t}\n\trcu_read_unlock();\n}", "label_name": "Variant", "label": 0.0}
{"code": "static void update_blocked_averages(int cpu)\n{\n\tstruct rq *rq = cpu_rq(cpu);\n\tstruct cfs_rq *cfs_rq, *pos;\n\tconst struct sched_class *curr_class;\n\tstruct rq_flags rf;\n\tbool done = true;\n\n\trq_lock_irqsave(rq, &rf);\n\tupdate_rq_clock(rq);\n\n\t/*\n\t * Iterates the task_group tree in a bottom up fashion, see\n\t * list_add_leaf_cfs_rq() for details.\n\t */\n\tfor_each_leaf_cfs_rq_safe(rq, cfs_rq, pos) {\n\t\tstruct sched_entity *se;\n\n\t\t/* throttled entities do not contribute to load */\n\t\tif (throttled_hierarchy(cfs_rq))\n\t\t\tcontinue;\n\n\t\tif (update_cfs_rq_load_avg(cfs_rq_clock_task(cfs_rq), cfs_rq))\n\t\t\tupdate_tg_load_avg(cfs_rq, 0);\n\n\t\t/* Propagate pending load changes to the parent, if any: */\n\t\tse = cfs_rq->tg->se[cpu];\n\t\tif (se && !skip_blocked_update(se))\n\t\t\tupdate_load_avg(cfs_rq_of(se), se, 0);\n\n\t\t/*\n\t\t * There can be a lot of idle CPU cgroups. Don't let fully\n\t\t * decayed cfs_rqs linger on the list.\n\t\t */\n\t\tif (cfs_rq_is_decayed(cfs_rq))\n\t\t\tlist_del_leaf_cfs_rq(cfs_rq);\n\n\t\t/* Don't need periodic decay once load/util_avg are null */\n\t\tif (cfs_rq_has_blocked(cfs_rq))\n\t\t\tdone = false;\n\t}\n\n\tcurr_class = rq->curr->sched_class;\n\tupdate_rt_rq_load_avg(rq_clock_task(rq), rq, curr_class == &rt_sched_class);\n\tupdate_dl_rq_load_avg(rq_clock_task(rq), rq, curr_class == &dl_sched_class);\n\tupdate_irq_load_avg(rq, 0);\n\t/* Don't need periodic decay once load/util_avg are null */\n\tif (others_have_blocked(rq))\n\t\tdone = false;\n\n#ifdef CONFIG_NO_HZ_COMMON\n\trq->last_blocked_load_update_tick = jiffies;\n\tif (done)\n\t\trq->has_blocked_load = 0;\n#endif\n\trq_unlock_irqrestore(rq, &rf);\n}", "label_name": "Base", "label": 1.0}
{"code": "__be32 ipv6_proxy_select_ident(struct net *net, struct sk_buff *skb)\n{\n\tstatic u32 ip6_proxy_idents_hashrnd __read_mostly;\n\tstruct in6_addr buf[2];\n\tstruct in6_addr *addrs;\n\tu32 id;\n\n\taddrs = skb_header_pointer(skb,\n\t\t\t\t skb_network_offset(skb) +\n\t\t\t\t offsetof(struct ipv6hdr, saddr),\n\t\t\t\t sizeof(buf), buf);\n\tif (!addrs)\n\t\treturn 0;\n\n\tnet_get_random_once(&ip6_proxy_idents_hashrnd,\n\t\t\t sizeof(ip6_proxy_idents_hashrnd));\n\n\tid = __ipv6_select_ident(net, ip6_proxy_idents_hashrnd,\n\t\t\t\t &addrs[1], &addrs[0]);\n\treturn htonl(id);\n}", "label_name": "Class", "label": 2.0}
@@ -1192,7 +1178,6 @@
{"code": "static u32 read_32(cdk_stream_t s)\n{\n\tbyte buf[4];\n\tsize_t nread;\n\n\tassert(s != NULL);\n\n\tstream_read(s, buf, 4, &nread);\n\tif (nread != 4)\n\t\treturn (u32) - 1;\n\treturn buf[0] << 24 | buf[1] << 16 | buf[2] << 8 | buf[3];\n}", "label_name": "Base", "label": 1.0}
{"code": "static u16 read_16(cdk_stream_t s)\n{\n\tbyte buf[2];\n\tsize_t nread;\n\n\tassert(s != NULL);\n\n\tstream_read(s, buf, 2, &nread);\n\tif (nread != 2)\n\t\treturn (u16) - 1;\n\treturn buf[0] << 8 | buf[1];\n}", "label_name": "Base", "label": 1.0}
{"code": "static void print_value(int output, int num, const char *devname,\n\t\t\tconst char *value, const char *name, size_t valsz)\n{\n\tif (output & OUTPUT_VALUE_ONLY) {\n\t\tfputs(value, stdout);\n\t\tfputc('\\n', stdout);\n\n\t} else if (output & OUTPUT_UDEV_LIST) {\n\t\tprint_udev_format(name, value);\n\n\t} else if (output & OUTPUT_EXPORT_LIST) {\n\t\tif (num == 1 && devname)\n\t\t\tprintf(\"DEVNAME=%s\\n\", devname);\n\t\tfputs(name, stdout);\n\t\tfputs(\"=\", stdout);\n\t\tsafe_print(value, valsz, NULL);\n\t\tfputs(\"\\n\", stdout);\n\n\t} else {\n\t\tif (num == 1 && devname)\n\t\t\tprintf(\"%s:\", devname);\n\t\tfputs(\" \", stdout);\n\t\tfputs(name, stdout);\n\t\tfputs(\"=\\\"\", stdout);\n\t\tsafe_print(value, valsz, \"\\\"\");\n\t\tfputs(\"\\\"\", stdout);\n\t}\n}", "label_name": "Class", "label": 2.0}
-{"code": "int xmkstemp(char **tmpname, char *dir)\n{\n\tchar *localtmp;\n\tchar *tmpenv;\n\tmode_t old_mode;\n\tint fd, rc;\n\n\t/* Some use cases must be capable of being moved atomically\n\t * with rename(2), which is the reason why dir is here. */\n\tif (dir != NULL)\n\t\ttmpenv = dir;\n\telse\n\t\ttmpenv = getenv(\"TMPDIR\");\n\n\tif (tmpenv)\n\t\trc = asprintf(&localtmp, \"%s/%s.XXXXXX\", tmpenv,\n\t\t\t program_invocation_short_name);\n\telse\n\t\trc = asprintf(&localtmp, \"%s/%s.XXXXXX\", _PATH_TMP,\n\t\t\t program_invocation_short_name);\n\n\tif (rc < 0)\n\t\treturn -1;\n\n\told_mode = umask(077);\n\tfd = mkostemp(localtmp, O_RDWR|O_CREAT|O_EXCL|O_CLOEXEC);\n\tumask(old_mode);\n\tif (fd == -1) {\n\t\tfree(localtmp);\n\t\tlocaltmp = NULL;\n\t}\n\t*tmpname = localtmp;\n\treturn fd;\n}", "label_name": NaN, "label": NaN}
{"code": "wb_id(netdissect_options *ndo,\n const struct pkt_id *id, u_int len)\n{\n\tint i;\n\tconst char *cp;\n\tconst struct id_off *io;\n\tchar c;\n\tint nid;\n\n\tND_PRINT((ndo, \" wb-id:\"));\n\tif (len < sizeof(*id) || !ND_TTEST(*id))\n\t\treturn (-1);\n\tlen -= sizeof(*id);\n\n\tND_PRINT((ndo, \" %u/%s:%u (max %u/%s:%u) \",\n\t EXTRACT_32BITS(&id->pi_ps.slot),\n\t ipaddr_string(ndo, &id->pi_ps.page.p_sid),\n\t EXTRACT_32BITS(&id->pi_ps.page.p_uid),\n\t EXTRACT_32BITS(&id->pi_mslot),\n\t ipaddr_string(ndo, &id->pi_mpage.p_sid),\n\t EXTRACT_32BITS(&id->pi_mpage.p_uid)));\n\n\tnid = EXTRACT_16BITS(&id->pi_ps.nid);\n\tlen -= sizeof(*io) * nid;\n\tio = (struct id_off *)(id + 1);\n\tcp = (char *)(io + nid);\n\tif (!ND_TTEST2(cp, len)) {\n\t\tND_PRINT((ndo, \"\\\"\"));\n\t\tfn_print(ndo, (u_char *)cp, (u_char *)cp + len);\n\t\tND_PRINT((ndo, \"\\\"\"));\n\t}\n\n\tc = '<';\n\tfor (i = 0; i < nid && ND_TTEST(*io); ++io, ++i) {\n\t\tND_PRINT((ndo, \"%c%s:%u\",\n\t\t c, ipaddr_string(ndo, &io->id), EXTRACT_32BITS(&io->off)));\n\t\tc = ',';\n\t}\n\tif (i >= nid) {\n\t\tND_PRINT((ndo, \">\"));\n\t\treturn (0);\n\t}\n\treturn (-1);\n}", "label_name": "Class", "label": 2.0}
{"code": "parserep(netdissect_options *ndo,\n register const struct sunrpc_msg *rp, register u_int length)\n{\n\tregister const uint32_t *dp;\n\tu_int len;\n\tenum sunrpc_accept_stat astat;\n\n\t/*\n\t * Portability note:\n\t * Here we find the address of the ar_verf credentials.\n\t * Originally, this calculation was\n\t *\tdp = (uint32_t *)&rp->rm_reply.rp_acpt.ar_verf\n\t * On the wire, the rp_acpt field starts immediately after\n\t * the (32 bit) rp_stat field. However, rp_acpt (which is a\n\t * \"struct accepted_reply\") contains a \"struct opaque_auth\",\n\t * whose internal representation contains a pointer, so on a\n\t * 64-bit machine the compiler inserts 32 bits of padding\n\t * before rp->rm_reply.rp_acpt.ar_verf. So, we cannot use\n\t * the internal representation to parse the on-the-wire\n\t * representation. Instead, we skip past the rp_stat field,\n\t * which is an \"enum\" and so occupies one 32-bit word.\n\t */\n\tdp = ((const uint32_t *)&rp->rm_reply) + 1;\n\tND_TCHECK(dp[1]);\n\tlen = EXTRACT_32BITS(&dp[1]);\n\tif (len >= length)\n\t\treturn (NULL);\n\t/*\n\t * skip past the ar_verf credentials.\n\t */\n\tdp += (len + (2*sizeof(uint32_t) + 3)) / sizeof(uint32_t);\n\tND_TCHECK2(dp[0], 0);\n\n\t/*\n\t * now we can check the ar_stat field\n\t */\n\tastat = (enum sunrpc_accept_stat) EXTRACT_32BITS(dp);\n\tif (astat != SUNRPC_SUCCESS) {\n\t\tND_PRINT((ndo, \" %s\", tok2str(sunrpc_str, \"ar_stat %d\", astat)));\n\t\tnfserr = 1;\t\t/* suppress trunc string */\n\t\treturn (NULL);\n\t}\n\t/* successful return */\n\tND_TCHECK2(*dp, sizeof(astat));\n\treturn ((const uint32_t *) (sizeof(astat) + ((const char *)dp)));\ntrunc:\n\treturn (0);\n}", "label_name": "Base", "label": 1.0}
{"code": "lookup_bytestring(netdissect_options *ndo, register const u_char *bs,\n\t\t const unsigned int nlen)\n{\n\tstruct enamemem *tp;\n\tregister u_int i, j, k;\n\n\tif (nlen >= 6) {\n\t\tk = (bs[0] << 8) | bs[1];\n\t\tj = (bs[2] << 8) | bs[3];\n\t\ti = (bs[4] << 8) | bs[5];\n\t} else if (nlen >= 4) {\n\t\tk = (bs[0] << 8) | bs[1];\n\t\tj = (bs[2] << 8) | bs[3];\n\t\ti = 0;\n\t} else\n\t\ti = j = k = 0;\n\n\ttp = &bytestringtable[(i ^ j) & (HASHNAMESIZE-1)];\n\twhile (tp->e_nxt)\n\t\tif (tp->e_addr0 == i &&\n\t\t tp->e_addr1 == j &&\n\t\t tp->e_addr2 == k &&\n\t\t memcmp((const char *)bs, (const char *)(tp->e_bs), nlen) == 0)\n\t\t\treturn tp;\n\t\telse\n\t\t\ttp = tp->e_nxt;\n\n\ttp->e_addr0 = i;\n\ttp->e_addr1 = j;\n\ttp->e_addr2 = k;\n\n\ttp->e_bs = (u_char *) calloc(1, nlen + 1);\n\tif (tp->e_bs == NULL)\n\t\t(*ndo->ndo_error)(ndo, \"lookup_bytestring: calloc\");\n\n\tmemcpy(tp->e_bs, bs, nlen);\n\ttp->e_nxt = (struct enamemem *)calloc(1, sizeof(*tp));\n\tif (tp->e_nxt == NULL)\n\t\t(*ndo->ndo_error)(ndo, \"lookup_bytestring: calloc\");\n\n\treturn tp;\n}", "label_name": "Base", "label": 1.0}
@@ -1247,7 +1232,6 @@
{"code": "const char *cJSON_GetErrorPtr( void )\n{\n\treturn ep;\n}", "label_name": "Base", "label": 1.0}
{"code": "cJSON *cJSON_CreateInt( int64_t num )\n{\n\tcJSON *item = cJSON_New_Item();\n\tif ( item ) {\n\t\titem->type = cJSON_Number;\n\t\titem->valuefloat = num;\n\t\titem->valueint = num;\n\t}\n\treturn item;\n}", "label_name": "Base", "label": 1.0}
{"code": "void cJSON_AddItemReferenceToArray( cJSON *array, cJSON *item )\n{\n\tcJSON_AddItemToArray( array, create_reference( item ) );\n}", "label_name": "Base", "label": 1.0}
-{"code": "local unsigned long crc32_big(crc, buf, len)\n unsigned long crc;\n const unsigned char FAR *buf;\n unsigned len;\n{\n register z_crc_t c;\n register const z_crc_t FAR *buf4;\n\n c = ZSWAP32((z_crc_t)crc);\n c = ~c;\n while (len && ((ptrdiff_t)buf & 3)) {\n c = crc_table[4][(c >> 24) ^ *buf++] ^ (c << 8);\n len--;\n }\n\n buf4 = (const z_crc_t FAR *)(const void FAR *)buf;\n buf4--;\n while (len >= 32) {\n DOBIG32;\n len -= 32;\n }\n while (len >= 4) {\n DOBIG4;\n len -= 4;\n }\n buf4++;\n buf = (const unsigned char FAR *)buf4;\n\n if (len) do {\n c = crc_table[4][(c >> 24) ^ *buf++] ^ (c << 8);\n } while (--len);\n c = ~c;\n return (unsigned long)(ZSWAP32(c));\n}", "label_name": NaN, "label": NaN}
{"code": "int ZEXPORT deflateCopy (dest, source)\n z_streamp dest;\n z_streamp source;\n{\n#ifdef MAXSEG_64K\n return Z_STREAM_ERROR;\n#else\n deflate_state *ds;\n deflate_state *ss;\n ushf *overlay;\n\n\n if (deflateStateCheck(source) || dest == Z_NULL) {\n return Z_STREAM_ERROR;\n }\n\n ss = source->state;\n\n zmemcpy((voidpf)dest, (voidpf)source, sizeof(z_stream));\n\n ds = (deflate_state *) ZALLOC(dest, 1, sizeof(deflate_state));\n if (ds == Z_NULL) return Z_MEM_ERROR;\n dest->state = (struct internal_state FAR *) ds;\n zmemcpy((voidpf)ds, (voidpf)ss, sizeof(deflate_state));\n ds->strm = dest;\n\n ds->window = (Bytef *) ZALLOC(dest, ds->w_size, 2*sizeof(Byte));\n ds->prev = (Posf *) ZALLOC(dest, ds->w_size, sizeof(Pos));\n ds->head = (Posf *) ZALLOC(dest, ds->hash_size, sizeof(Pos));\n overlay = (ushf *) ZALLOC(dest, ds->lit_bufsize, sizeof(ush)+2);\n ds->pending_buf = (uchf *) overlay;\n\n if (ds->window == Z_NULL || ds->prev == Z_NULL || ds->head == Z_NULL ||\n ds->pending_buf == Z_NULL) {\n deflateEnd (dest);\n return Z_MEM_ERROR;\n }\n /* following zmemcpy do not work for 16-bit MSDOS */\n zmemcpy(ds->window, ss->window, ds->w_size * 2 * sizeof(Byte));\n zmemcpy((voidpf)ds->prev, (voidpf)ss->prev, ds->w_size * sizeof(Pos));\n zmemcpy((voidpf)ds->head, (voidpf)ss->head, ds->hash_size * sizeof(Pos));\n zmemcpy(ds->pending_buf, ss->pending_buf, (uInt)ds->pending_buf_size);\n\n ds->pending_out = ds->pending_buf + (ss->pending_out - ss->pending_buf);\n ds->d_buf = overlay + ds->lit_bufsize/sizeof(ush);\n ds->l_buf = ds->pending_buf + (1+sizeof(ush))*ds->lit_bufsize;\n\n ds->l_desc.dyn_tree = ds->dyn_ltree;\n ds->d_desc.dyn_tree = ds->dyn_dtree;\n ds->bl_desc.dyn_tree = ds->bl_tree;\n\n return Z_OK;\n#endif /* MAXSEG_64K */\n}", "label_name": "Base", "label": 1.0}
{"code": "static size_t consume_init_expr (ut8 *buf, ut8 *max, ut8 eoc, void *out, ut32 *offset) {\n\tut32 i = 0;\n\twhile (buf + i < max && buf[i] != eoc) {\n\t\t// TODO: calc the expresion with the bytcode (ESIL?)\n\t\ti += 1;\n\t}\n\tif (buf[i] != eoc) {\n\t\treturn 0;\n\t}\n\tif (offset) {\n\t\t*offset += i + 1;\n\t}\n\treturn i + 1;\n}", "label_name": "Base", "label": 1.0}
{"code": "static void process_constructors (RBinFile *bf, RList *ret, int bits) {\n\tRList *secs = sections (bf);\n\tRListIter *iter;\n\tRBinSection *sec;\n\tint i, type;\n\tr_list_foreach (secs, iter, sec) {\n\t\ttype = -1;\n\t\tif (!strcmp (sec->name, \".fini_array\")) {\n\t\t\ttype = R_BIN_ENTRY_TYPE_FINI;\n\t\t} else if (!strcmp (sec->name, \".init_array\")) {\n\t\t\ttype = R_BIN_ENTRY_TYPE_INIT;\n\t\t} else if (!strcmp (sec->name, \".preinit_array\")) {\n\t\t\ttype = R_BIN_ENTRY_TYPE_PREINIT;\n\t\t}\n\t\tif (type != -1) {\n\t\t\tut8 *buf = calloc (sec->size, 1);\n\t\t\tif (!buf) {\n\t\t\t\tcontinue;\n\t\t\t}\n\t\t\t(void)r_buf_read_at (bf->buf, sec->paddr, buf, sec->size);\n\t\t\tif (bits == 32) {\n\t\t\t\tfor (i = 0; i < sec->size; i += 4) {\n\t\t\t\t\tut32 addr32 = r_read_le32 (buf + i);\n\t\t\t\t\tif (addr32) {\n\t\t\t\t\t\tRBinAddr *ba = newEntry (sec->paddr + i, (ut64)addr32, type, bits);\n\t\t\t\t\t\tr_list_append (ret, ba);\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tfor (i = 0; i < sec->size; i += 8) {\n\t\t\t\t\tut64 addr64 = r_read_le64 (buf + i);\n\t\t\t\t\tif (addr64) {\n\t\t\t\t\t\tRBinAddr *ba = newEntry (sec->paddr + i, addr64, type, bits);\n\t\t\t\t\t\tr_list_append (ret, ba);\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t\tfree (buf);\n\t\t}\n\t}\n\tr_list_free (secs);\n}", "label_name": "Base", "label": 1.0}
@@ -1278,13 +1262,11 @@
{"code": "fiber_switch(mrb_state *mrb, mrb_value self, mrb_int len, const mrb_value *a, mrb_bool resume, mrb_bool vmexec)\n{\n struct mrb_context *c = fiber_check(mrb, self);\n struct mrb_context *old_c = mrb->c;\n enum mrb_fiber_state status;\n mrb_value value;\n\n fiber_check_cfunc(mrb, c);\n status = c->status;\n switch (status) {\n case MRB_FIBER_TRANSFERRED:\n if (resume) {\n mrb_raise(mrb, E_FIBER_ERROR, \"resuming transferred fiber\");\n }\n break;\n case MRB_FIBER_RUNNING:\n case MRB_FIBER_RESUMED:\n mrb_raise(mrb, E_FIBER_ERROR, \"double resume\");\n break;\n case MRB_FIBER_TERMINATED:\n mrb_raise(mrb, E_FIBER_ERROR, \"resuming dead fiber\");\n break;\n default:\n break;\n }\n old_c->status = resume ? MRB_FIBER_RESUMED : MRB_FIBER_TRANSFERRED;\n c->prev = resume ? mrb->c : (c->prev ? c->prev : mrb->root_c);\n fiber_switch_context(mrb, c);\n if (status == MRB_FIBER_CREATED) {\n mrb_value *b, *e;\n\n if (!c->ci->proc) {\n mrb_raise(mrb, E_FIBER_ERROR, \"double resume (current)\");\n }\n mrb_stack_extend(mrb, len+2); /* for receiver and (optional) block */\n b = c->stbase+1;\n e = b + len;\n while (bci--; /* pop dummy callinfo */\n }\n c->cibase->n = len;\n value = c->stbase[0] = MRB_PROC_ENV(c->cibase->proc)->stack[0];\n }\n else {\n value = fiber_result(mrb, a, len);\n if (vmexec) {\n c->ci[1].stack[0] = value;\n }\n }\n\n if (vmexec) {\n c->vmexec = TRUE;\n value = mrb_vm_exec(mrb, c->ci->proc, c->ci->pc);\n mrb->c = old_c;\n }\n else {\n MARK_CONTEXT_MODIFY(c);\n }\n return value;\n}", "label_name": "Base", "label": 1.0}
{"code": "mrb_obj_clone(mrb_state *mrb, mrb_value self)\n{\n struct RObject *p;\n mrb_value clone;\n\n if (mrb_immediate_p(self)) {\n mrb_raisef(mrb, E_TYPE_ERROR, \"can't clone %S\", self);\n }\n if (mrb_type(self) == MRB_TT_SCLASS) {\n mrb_raise(mrb, E_TYPE_ERROR, \"can't clone singleton class\");\n }\n p = (struct RObject*)mrb_obj_alloc(mrb, mrb_type(self), mrb_obj_class(mrb, self));\n p->c = mrb_singleton_class_clone(mrb, self);\n mrb_field_write_barrier(mrb, (struct RBasic*)p, (struct RBasic*)p->c);\n clone = mrb_obj_value(p);\n init_copy(mrb, clone, self);\n p->flags = mrb_obj_ptr(self)->flags;\n\n return clone;\n}", "label_name": "Base", "label": 1.0}
{"code": "read_pbm_integer(j_compress_ptr cinfo, FILE *infile, unsigned int maxval)\n/* Read an unsigned decimal integer from the PPM file */\n/* Swallows one trailing character after the integer */\n/* Note that on a 16-bit-int machine, only values up to 64k can be read. */\n/* This should not be a problem in practice. */\n{\n register int ch;\n register unsigned int val;\n\n /* Skip any leading whitespace */\n do {\n ch = pbm_getc(infile);\n if (ch == EOF)\n ERREXIT(cinfo, JERR_INPUT_EOF);\n } while (ch == ' ' || ch == '\\t' || ch == '\\n' || ch == '\\r');\n\n if (ch < '0' || ch > '9')\n ERREXIT(cinfo, JERR_PPM_NONNUMERIC);\n\n val = ch - '0';\n while ((ch = pbm_getc(infile)) >= '0' && ch <= '9') {\n val *= 10;\n val += ch - '0';\n }\n\n if (val > maxval)\n ERREXIT(cinfo, JERR_PPM_TOOLARGE);\n\n return val;\n}", "label_name": "Base", "label": 1.0}
-{"code": "gimp_write_and_read_file (Gimp *gimp,\n gboolean with_unusual_stuff,\n gboolean compat_paths,\n gboolean use_gimp_2_8_features)\n{\n GimpImage *image;\n GimpImage *loaded_image;\n GimpPlugInProcedure *proc;\n gchar *filename;\n GFile *file;\n\n /* Create the image */\n image = gimp_create_mainimage (gimp,\n with_unusual_stuff,\n compat_paths,\n use_gimp_2_8_features);\n\n /* Assert valid state */\n gimp_assert_mainimage (image,\n with_unusual_stuff,\n compat_paths,\n use_gimp_2_8_features);\n\n /* Write to file */\n filename = g_build_filename (g_get_tmp_dir (), \"gimp-test.xcf\", NULL);\n file = g_file_new_for_path (filename);\n g_free (filename);\n\n proc = gimp_plug_in_manager_file_procedure_find (image->gimp->plug_in_manager,\n GIMP_FILE_PROCEDURE_GROUP_SAVE,\n file,\n NULL /*error*/);\n file_save (gimp,\n image,\n NULL /*progress*/,\n file,\n proc,\n GIMP_RUN_NONINTERACTIVE,\n FALSE /*change_saved_state*/,\n FALSE /*export_backward*/,\n FALSE /*export_forward*/,\n NULL /*error*/);\n\n /* Load from file */\n loaded_image = gimp_test_load_image (image->gimp, file);\n\n /* Assert on the loaded file. If success, it means that there is no\n * significant information loss when we wrote the image to a file\n * and loaded it again\n */\n gimp_assert_mainimage (loaded_image,\n with_unusual_stuff,\n compat_paths,\n use_gimp_2_8_features);\n\n g_file_delete (file, NULL, NULL);\n g_object_unref (file);\n}", "label_name": NaN, "label": NaN}
{"code": "GF_AV1Config *gf_isom_av1_config_get(GF_ISOFile *the_file, u32 trackNumber, u32 DescriptionIndex)\n{\n\tGF_TrackBox *trak;\n\tGF_MPEGVisualSampleEntryBox *entry;\n\tif (gf_isom_get_reference_count(the_file, trackNumber, GF_ISOM_REF_TBAS)) {\n\t\tu32 ref_track;\n\t\tGF_Err e = gf_isom_get_reference(the_file, trackNumber, GF_ISOM_REF_TBAS, 1, &ref_track);\n\t\tif (e == GF_OK) {\n\t\t\ttrackNumber = ref_track;\n\t\t}\n\t}\n\ttrak = gf_isom_get_track_from_file(the_file, trackNumber);\n\tif (!trak || !trak->Media || !DescriptionIndex) return NULL;\n\tentry = (GF_MPEGVisualSampleEntryBox*)gf_list_get(trak->Media->information->sampleTable->SampleDescription->child_boxes, DescriptionIndex - 1);\n\tif (!entry || !entry->av1_config) return NULL;\n\treturn AV1_DuplicateConfig(entry->av1_config->config);\n}", "label_name": "Base", "label": 1.0}
{"code": "}\n\nvoid print_udta(GF_ISOFile *file, u32 track_number, Bool has_itags)\n{\n\tu32 i, count;\n\n\tcount = gf_isom_get_udta_count(file, track_number);\n\tif (!count) return;\n\n\tif (has_itags) {\n\t\tfor (i=0; iSampleSize->sampleCount) {\n\t\tstbl->SampleSize->sampleSize = size;\n\t\tstbl->SampleSize->sampleCount += nb_pack;\n\t\treturn GF_OK;\n\t}\n\tif (stbl->SampleSize->sampleSize && (stbl->SampleSize->sampleSize==size)) {\n\t\tstbl->SampleSize->sampleCount += nb_pack;\n\t\treturn GF_OK;\n\t}\n\tif (!stbl->SampleSize->sizes || (stbl->SampleSize->sampleCount+nb_pack > stbl->SampleSize->alloc_size)) {\n\t\tBool init_table = (stbl->SampleSize->sizes==NULL) ? 1 : 0;\n\t\tALLOC_INC(stbl->SampleSize->alloc_size);\n\t\tif (stbl->SampleSize->sampleCount+nb_pack > stbl->SampleSize->alloc_size)\n\t\t\tstbl->SampleSize->alloc_size = stbl->SampleSize->sampleCount+nb_pack;\n\n\t\tstbl->SampleSize->sizes = (u32 *)gf_realloc(stbl->SampleSize->sizes, sizeof(u32)*stbl->SampleSize->alloc_size);\n\t\tif (!stbl->SampleSize->sizes) return GF_OUT_OF_MEM;\n\t\tmemset(&stbl->SampleSize->sizes[stbl->SampleSize->sampleCount], 0, sizeof(u32) * (stbl->SampleSize->alloc_size - stbl->SampleSize->sampleCount) );\n\n\t\tif (init_table) {\n\t\t\tfor (i=0; iSampleSize->sampleCount; i++)\n\t\t\t\tstbl->SampleSize->sizes[i] = stbl->SampleSize->sampleSize;\n\t\t}\n\t}\n\tstbl->SampleSize->sampleSize = 0;\n\tfor (i=0; iSampleSize->sizes[stbl->SampleSize->sampleCount+i] = size;\n\t}\n\tstbl->SampleSize->sampleCount += nb_pack;\n\tif (size > stbl->SampleSize->max_size)\n\t\tstbl->SampleSize->max_size = size;\n\tstbl->SampleSize->total_size += size;\n\tstbl->SampleSize->total_samples += nb_pack;\n\treturn GF_OK;\n}", "label_name": "Base", "label": 1.0}
{"code": "static s32 svc_parse_slice(GF_BitStream *bs, AVCState *avc, AVCSliceInfo *si)\n{\n\ts32 pps_id;\n\n\t/*s->current_picture.reference= h->nal_ref_idc != 0;*/\n\tgf_bs_read_ue_log(bs, \"first_mb_in_slice\");\n\tsi->slice_type = gf_bs_read_ue_log(bs, \"slice_type\");\n\tif (si->slice_type > 9) return -1;\n\n\tpps_id = gf_bs_read_ue_log(bs, \"pps_id\");\n\tif (pps_id > 255)\n\t\treturn -1;\n\tsi->pps = &avc->pps[pps_id];\n\tsi->pps->id = pps_id;\n\tif (!si->pps->slice_group_count)\n\t\treturn -2;\n\tsi->sps = &avc->sps[si->pps->sps_id + GF_SVC_SSPS_ID_SHIFT];\n\tif (!si->sps->log2_max_frame_num)\n\t\treturn -2;\n\n\tsi->frame_num = gf_bs_read_int_log(bs, si->sps->log2_max_frame_num, \"frame_num\");\n\n\tsi->field_pic_flag = 0;\n\tif (si->sps->frame_mbs_only_flag) {\n\t\t/*s->picture_structure= PICT_FRAME;*/\n\t}\n\telse {\n\t\tsi->field_pic_flag = gf_bs_read_int_log(bs, 1, \"field_pic_flag\");\n\t\tif (si->field_pic_flag) si->bottom_field_flag = gf_bs_read_int_log(bs, 1, \"bottom_field_flag\");\n\t}\n\tif (si->nal_unit_type == GF_AVC_NALU_IDR_SLICE || si->NalHeader.idr_pic_flag)\n\t\tsi->idr_pic_id = gf_bs_read_ue_log(bs, \"idr_pic_id\");\n\n\tif (si->sps->poc_type == 0) {\n\t\tsi->poc_lsb = gf_bs_read_int_log(bs, si->sps->log2_max_poc_lsb, \"poc_lsb\");\n\t\tif (si->pps->pic_order_present && !si->field_pic_flag) {\n\t\t\tsi->delta_poc_bottom = gf_bs_read_se_log(bs, \"delta_poc_bottom\");\n\t\t}\n\t}\n\telse if ((si->sps->poc_type == 1) && !si->sps->delta_pic_order_always_zero_flag) {\n\t\tsi->delta_poc[0] = gf_bs_read_se_log(bs, \"delta_poc0\");\n\t\tif ((si->pps->pic_order_present == 1) && !si->field_pic_flag)\n\t\t\tsi->delta_poc[1] = gf_bs_read_se_log(bs, \"delta_poc1\");\n\t}\n\tif (si->pps->redundant_pic_cnt_present) {\n\t\tsi->redundant_pic_cnt = gf_bs_read_ue_log(bs, \"redundant_pic_cnt\");\n\t}\n\treturn 0;\n}", "label_name": "Base", "label": 1.0}
{"code": "u32 GetHintFormat(GF_TrackBox *trak)\n{\n\tGF_HintMediaHeaderBox *hmhd = (GF_HintMediaHeaderBox *)trak->Media->information->InfoHeader;\n\tif (hmhd->type != GF_ISOM_BOX_TYPE_HMHD)\n\t\treturn 0;\n\t\t\n\tif (!hmhd || !hmhd->subType) {\n\t\tGF_Box *a = (GF_Box *)gf_list_get(trak->Media->information->sampleTable->SampleDescription->child_boxes, 0);\n\t\tif (!hmhd) return a ? a->type : 0;\n\t\tif (a) hmhd->subType = a->type;\n\t\treturn hmhd->subType;\n\t}\n\treturn hmhd->subType;\n}", "label_name": "Base", "label": 1.0}
-{"code": "static void gf_dump_vrml_simple_field(GF_SceneDumper *sdump, GF_FieldInfo field, GF_Node *parent)\n{\n\tu32 i, sf_type;\n\tGF_ChildNodeItem *list;\n\tvoid *slot_ptr;\n\n\tswitch (field.fieldType) {\n\tcase GF_SG_VRML_SFNODE:\n\t\tassert ( *(GF_Node **)field.far_ptr);\n\t\tgf_dump_vrml_node(sdump, *(GF_Node **)field.far_ptr, 0, NULL);\n\t\treturn;\n\tcase GF_SG_VRML_MFNODE:\n\t\tlist = * ((GF_ChildNodeItem **) field.far_ptr);\n\t\tassert( list );\n\t\tsdump->indent++;\n\t\twhile (list) {\n\t\t\tgf_dump_vrml_node(sdump, list->node, 1, NULL);\n\t\t\tlist = list->next;\n\t\t}\n\t\tsdump->indent--;\n\t\treturn;\n\tcase GF_SG_VRML_SFCOMMANDBUFFER:\n\t\treturn;\n\t}\n\tif (gf_sg_vrml_is_sf_field(field.fieldType)) {\n\t\tif (sdump->XMLDump) StartAttribute(sdump, \"value\");\n\t\tgf_dump_vrml_sffield(sdump, field.fieldType, field.far_ptr, 0, parent);\n\t\tif (sdump->XMLDump) EndAttribute(sdump);\n\t} else {\n\t\tGenMFField *mffield;\n\t\tmffield = (GenMFField *) field.far_ptr;\n\t\tsf_type = gf_sg_vrml_get_sf_type(field.fieldType);\n\t\tif (!sdump->XMLDump) {\n\t\t\tgf_fprintf(sdump->trace, \"[\");\n\t\t} else if (sf_type==GF_SG_VRML_SFSTRING) {\n\t\t\tgf_fprintf(sdump->trace, \" value=\\'\");\n\t\t} else {\n\t\t\tStartAttribute(sdump, \"value\");\n\t\t}\n\t\tfor (i=0; icount; i++) {\n\t\t\tif (i) gf_fprintf(sdump->trace, \" \");\n\t\t\tgf_sg_vrml_mf_get_item(field.far_ptr, field.fieldType, &slot_ptr, i);\n\t\t\t/*this is to cope with single MFString which shall appear as SF in XMT*/\n\t\t\tgf_dump_vrml_sffield(sdump, sf_type, slot_ptr, 1, parent);\n\t\t}\n\t\tif (!sdump->XMLDump) {\n\t\t\tgf_fprintf(sdump->trace, \"]\");\n\t\t} else if (sf_type==GF_SG_VRML_SFSTRING) {\n\t\t\tgf_fprintf(sdump->trace, \"\\'\");\n\t\t} else {\n\t\t\tEndAttribute(sdump);\n\t\t}\n\t}\n}", "label_name": NaN, "label": NaN}
{"code": "char *M_fs_path_join_parts(const M_list_str_t *path, M_fs_system_t sys_type)\n{\n\tM_list_str_t *parts;\n\tconst char *part;\n\tchar *out;\n\tsize_t len;\n\tsize_t i;\n\tsize_t count;\n\n\tif (path == NULL) {\n\t\treturn NULL;\n\t}\n\tlen = M_list_str_len(path);\n\tif (len == 0) {\n\t\treturn NULL;\n\t}\n\n\tsys_type = M_fs_path_get_system_type(sys_type);\n\n\t/* Remove any empty parts (except for the first part which denotes an abs path on Unix\n \t * or a UNC path on Windows). */\n\tparts = M_list_str_duplicate(path);\n\tfor (i=len-1; i>0; i--) {\n\t\tpart = M_list_str_at(parts, i);\n\t\tif (part == NULL || *part == '\\0') {\n\t\t\tM_list_str_remove_at(parts, i);\n\t\t}\n\t}\n\n\tlen = M_list_str_len(parts);\n\n\t/* Join puts the sep between items. If there are no items then the sep\n\t * won't be written. */\n\tpart = M_list_str_at(parts, 0);\n\tif (len == 1 && (part == NULL || *part == '\\0')) {\n\t\tM_list_str_destroy(parts);\n\t\tif (sys_type == M_FS_SYSTEM_WINDOWS) {\n\t\t\treturn M_strdup(\"\\\\\\\\\");\n\t\t}\n\t\treturn M_strdup(\"/\");\n\t}\n\n\t/* Handle windows abs path because they need two separators. */\n\tif (sys_type == M_FS_SYSTEM_WINDOWS && len > 0) {\n\t\tpart = M_list_str_at(parts, 0);\n\t\t/* If we have 1 item we need to add two empties so we get the second separator. */\n\t\tcount = (len == 1) ? 2 : 1;\n\t\t/* If we're dealing with a unc path add the second sep so we get two separators for the UNC base. */\n\t\tif (part != NULL && *part == '\\0') {\n\t\t\tfor (i=0; idata->uid) < 0)\n return -1;\n\n s = imap_next_word (s);\n }\n else if (ascii_strncasecmp (\"INTERNALDATE\", s, 12) == 0)\n {\n s += 12;\n SKIPWS (s);\n if (*s != '\\\"')\n {\n dprint (1, (debugfile, \"msg_parse_fetch(): bogus INTERNALDATE entry: %s\\n\", s));\n return -1;\n }\n s++;\n ptmp = tmp;\n while (*s && *s != '\\\"')\n *ptmp++ = *s++;\n if (*s != '\\\"')\n return -1;\n s++; /* skip past the trailing \" */\n *ptmp = 0;\n h->received = imap_parse_date (tmp);\n }\n else if (ascii_strncasecmp (\"RFC822.SIZE\", s, 11) == 0)\n {\n s += 11;\n SKIPWS (s);\n ptmp = tmp;\n while (isdigit ((unsigned char) *s))\n *ptmp++ = *s++;\n *ptmp = 0;\n if (mutt_atol (tmp, &h->content_length) < 0)\n return -1;\n }\n else if (!ascii_strncasecmp (\"BODY\", s, 4) ||\n !ascii_strncasecmp (\"RFC822.HEADER\", s, 13))\n {\n /* handle above, in msg_fetch_header */\n return -2;\n }\n else if (*s == ')')\n s++; /* end of request */\n else if (*s)\n {\n /* got something i don't understand */\n imap_error (\"msg_parse_fetch\", s);\n return -1;\n }\n }\n\n return 0;\n}", "label_name": "Base", "label": 1.0}
{"code": "static int pop_sync_mailbox(struct Context *ctx, int *index_hint)\n{\n int i, j, ret = 0;\n char buf[LONG_STRING];\n struct PopData *pop_data = (struct PopData *) ctx->data;\n struct Progress progress;\n#ifdef USE_HCACHE\n header_cache_t *hc = NULL;\n#endif\n\n pop_data->check_time = 0;\n\n while (true)\n {\n if (pop_reconnect(ctx) < 0)\n return -1;\n\n mutt_progress_init(&progress, _(\"Marking messages deleted...\"),\n MUTT_PROGRESS_MSG, WriteInc, ctx->deleted);\n\n#ifdef USE_HCACHE\n hc = pop_hcache_open(pop_data, ctx->path);\n#endif\n\n for (i = 0, j = 0, ret = 0; ret == 0 && i < ctx->msgcount; i++)\n {\n if (ctx->hdrs[i]->deleted && ctx->hdrs[i]->refno != -1)\n {\n j++;\n if (!ctx->quiet)\n mutt_progress_update(&progress, j, -1);\n snprintf(buf, sizeof(buf), \"DELE %d\\r\\n\", ctx->hdrs[i]->refno);\n ret = pop_query(pop_data, buf, sizeof(buf));\n if (ret == 0)\n {\n mutt_bcache_del(pop_data->bcache, ctx->hdrs[i]->data);\n#ifdef USE_HCACHE\n mutt_hcache_delete(hc, ctx->hdrs[i]->data, strlen(ctx->hdrs[i]->data));\n#endif\n }\n }\n\n#ifdef USE_HCACHE\n if (ctx->hdrs[i]->changed)\n {\n mutt_hcache_store(hc, ctx->hdrs[i]->data, strlen(ctx->hdrs[i]->data),\n ctx->hdrs[i], 0);\n }\n#endif\n }\n\n#ifdef USE_HCACHE\n mutt_hcache_close(hc);\n#endif\n\n if (ret == 0)\n {\n mutt_str_strfcpy(buf, \"QUIT\\r\\n\", sizeof(buf));\n ret = pop_query(pop_data, buf, sizeof(buf));\n }\n\n if (ret == 0)\n {\n pop_data->clear_cache = true;\n pop_clear_cache(pop_data);\n pop_data->status = POP_DISCONNECTED;\n return 0;\n }\n\n if (ret == -2)\n {\n mutt_error(\"%s\", pop_data->err_msg);\n return -1;\n }\n }\n}", "label_name": "Base", "label": 1.0}
@@ -1323,7 +1305,6 @@
{"code": "apr_byte_t oidc_cache_get(request_rec *r, const char *section, const char *key,\n\t\tchar **value) {\n\n\toidc_cfg *cfg = ap_get_module_config(r->server->module_config,\n\t\t\t&auth_openidc_module);\n\tint encrypted = oidc_cfg_cache_encrypt(r);\n\tapr_byte_t rc = TRUE;\n\tchar *msg = NULL;\n\n\toidc_debug(r, \"enter: %s (section=%s, decrypt=%d, type=%s)\", key, section,\n\t\t\tencrypted, cfg->cache->name);\n\n\t/* see if encryption is turned on */\n\tif (encrypted == 1)\n\t\tkey = oidc_cache_get_hashed_key(r, cfg->crypto_passphrase, key);\n\n\t/* get the value from the cache */\n\tconst char *cache_value = NULL;\n\tif (cfg->cache->get(r, section, key, &cache_value) == FALSE) {\n\t\trc = FALSE;\n\t\tgoto out;\n\t}\n\n\t/* see if it is any good */\n\tif (cache_value == NULL)\n\t\tgoto out;\n\n\t/* see if encryption is turned on */\n\tif (encrypted == 0) {\n\t\t*value = apr_pstrdup(r->pool, cache_value);\n\t\tgoto out;\n\t}\n\n\trc = (oidc_cache_crypto_decrypt(r, cache_value,\n\t\t\toidc_cache_hash_passphrase(r, cfg->crypto_passphrase),\n\t\t\t(unsigned char **) value) > 0);\n\nout:\n\t/* log the result */\n\tmsg = apr_psprintf(r->pool, \"from %s cache backend for %skey %s\",\n\t\t\tcfg->cache->name, encrypted ? \"encrypted \" : \"\", key);\n\tif (rc == TRUE)\n\t\tif (*value != NULL)\n\t\t\toidc_debug(r, \"cache hit: return %d bytes %s\",\n\t\t\t\t\t*value ? (int )strlen(*value) : 0, msg);\n\t\telse\n\t\t\toidc_debug(r, \"cache miss %s\", msg);\n\telse\n\t\toidc_warn(r, \"error retrieving value %s\", msg);\n\n\treturn rc;\n}", "label_name": "Class", "label": 2.0}
{"code": "static unsigned char *oidc_cache_hash_passphrase(request_rec *r,\n\t\tconst char *passphrase) {\n\n\tunsigned char *key = NULL;\n\tunsigned int key_len = 0;\n\toidc_jose_error_t err;\n\n\tif (oidc_jose_hash_bytes(r->pool, OIDC_JOSE_ALG_SHA256,\n\t\t\t(const unsigned char *) passphrase, strlen(passphrase), &key,\n\t\t\t&key_len, &err) == FALSE) {\n\t\toidc_error(r, \"oidc_jose_hash_bytes returned an error: %s\", err.text);\n\t\treturn NULL;\n\t}\n\n\treturn key;\n}", "label_name": "Class", "label": 2.0}
{"code": "static void utee_param_to_param(struct tee_ta_param *p, struct utee_params *up)\n{\n\tsize_t n;\n\tuint32_t types = up->types;\n\n\tp->types = types;\n\tfor (n = 0; n < TEE_NUM_PARAMS; n++) {\n\t\tuintptr_t a = up->vals[n * 2];\n\t\tsize_t b = up->vals[n * 2 + 1];\n\n\t\tswitch (TEE_PARAM_TYPE_GET(types, n)) {\n\t\tcase TEE_PARAM_TYPE_MEMREF_INPUT:\n\t\tcase TEE_PARAM_TYPE_MEMREF_OUTPUT:\n\t\tcase TEE_PARAM_TYPE_MEMREF_INOUT:\n\t\t\tp->u[n].mem.mobj = &mobj_virt;\n\t\t\tp->u[n].mem.offs = a;\n\t\t\tp->u[n].mem.size = b;\n\t\t\tbreak;\n\t\tcase TEE_PARAM_TYPE_VALUE_INPUT:\n\t\tcase TEE_PARAM_TYPE_VALUE_INOUT:\n\t\t\tp->u[n].val.a = a;\n\t\t\tp->u[n].val.b = b;\n\t\t\tbreak;\n\t\tdefault:\n\t\t\tmemset(&p->u[n], 0, sizeof(p->u[n]));\n\t\t\tbreak;\n\t\t}\n\t}\n}", "label_name": "Class", "label": 2.0}
-{"code": "int fit_image_verify(const void *fit, int image_noffset)\n{\n\tconst void\t*data;\n\tsize_t\t\tsize;\n\tint\t\tnoffset = 0;\n\tchar\t\t*err_msg = \"\";\n\n\t/* Get image data and data length */\n\tif (fit_image_get_data_and_size(fit, image_noffset, &data, &size)) {\n\t\terr_msg = \"Can't get image data/size\";\n\t\tprintf(\"error!\\n%s for '%s' hash node in '%s' image node\\n\",\n\t\t err_msg, fit_get_name(fit, noffset, NULL),\n\t\t fit_get_name(fit, image_noffset, NULL));\n\t\treturn 0;\n\t}\n\n\treturn fit_image_verify_with_data(fit, image_noffset, data, size);\n}", "label_name": NaN, "label": NaN}
{"code": "static int do_i2c_crc(struct cmd_tbl *cmdtp, int flag, int argc,\n\t\t char *const argv[])\n{\n\tuint\tchip;\n\tulong\taddr;\n\tint\talen;\n\tint\tcount;\n\tuchar\tbyte;\n\tulong\tcrc;\n\tulong\terr;\n\tint ret = 0;\n#if CONFIG_IS_ENABLED(DM_I2C)\n\tstruct udevice *dev;\n#endif\n\n\tif (argc < 4)\n\t\treturn CMD_RET_USAGE;\n\n\t/*\n\t * Chip is always specified.\n\t */\n\tchip = hextoul(argv[1], NULL);\n\n\t/*\n\t * Address is always specified.\n\t */\n\taddr = hextoul(argv[2], NULL);\n\talen = get_alen(argv[2], DEFAULT_ADDR_LEN);\n\tif (alen > 3)\n\t\treturn CMD_RET_USAGE;\n\n#if CONFIG_IS_ENABLED(DM_I2C)\n\tret = i2c_get_cur_bus_chip(chip, &dev);\n\tif (!ret && alen != -1)\n\t\tret = i2c_set_chip_offset_len(dev, alen);\n\tif (ret)\n\t\treturn i2c_report_err(ret, I2C_ERR_READ);\n#endif\n\t/*\n\t * Count is always specified\n\t */\n\tcount = hextoul(argv[3], NULL);\n\n\tprintf (\"CRC32 for %08lx ... %08lx ==> \", addr, addr + count - 1);\n\t/*\n\t * CRC a byte at a time. This is going to be slooow, but hey, the\n\t * memories are small and slow too so hopefully nobody notices.\n\t */\n\tcrc = 0;\n\terr = 0;\n\twhile (count-- > 0) {\n#if CONFIG_IS_ENABLED(DM_I2C)\n\t\tret = dm_i2c_read(dev, addr, &byte, 1);\n#else\n\t\tret = i2c_read(chip, addr, alen, &byte, 1);\n#endif\n\t\tif (ret)\n\t\t\terr++;\n\t\tcrc = crc32(crc, &byte, 1);\n\t\taddr++;\n\t}\n\tif (err > 0)\n\t\ti2c_report_err(ret, I2C_ERR_READ);\n\telse\n\t\tprintf (\"%08lx\\n\", crc);\n\n\treturn 0;\n}", "label_name": "Base", "label": 1.0}
{"code": "static int do_i2c_read(struct cmd_tbl *cmdtp, int flag, int argc,\n\t\t char *const argv[])\n{\n\tuint\tchip;\n\tuint\tdevaddr, length;\n\tint alen;\n\tu_char *memaddr;\n\tint ret;\n#if CONFIG_IS_ENABLED(DM_I2C)\n\tstruct udevice *dev;\n#endif\n\n\tif (argc != 5)\n\t\treturn CMD_RET_USAGE;\n\n\t/*\n\t * I2C chip address\n\t */\n\tchip = hextoul(argv[1], NULL);\n\n\t/*\n\t * I2C data address within the chip. This can be 1 or\n\t * 2 bytes long. Some day it might be 3 bytes long :-).\n\t */\n\tdevaddr = hextoul(argv[2], NULL);\n\talen = get_alen(argv[2], DEFAULT_ADDR_LEN);\n\tif (alen > 3)\n\t\treturn CMD_RET_USAGE;\n\n\t/*\n\t * Length is the number of objects, not number of bytes.\n\t */\n\tlength = hextoul(argv[3], NULL);\n\n\t/*\n\t * memaddr is the address where to store things in memory\n\t */\n\tmemaddr = (u_char *)hextoul(argv[4], NULL);\n\n#if CONFIG_IS_ENABLED(DM_I2C)\n\tret = i2c_get_cur_bus_chip(chip, &dev);\n\tif (!ret && alen != -1)\n\t\tret = i2c_set_chip_offset_len(dev, alen);\n\tif (!ret)\n\t\tret = dm_i2c_read(dev, devaddr, memaddr, length);\n#else\n\tret = i2c_read(chip, devaddr, alen, memaddr, length);\n#endif\n\tif (ret)\n\t\treturn i2c_report_err(ret, I2C_ERR_READ);\n\n\treturn 0;\n}", "label_name": "Base", "label": 1.0}
{"code": "static void Sp_search(js_State *J)\n{\n\tjs_Regexp *re;\n\tconst char *text;\n\tResub m;\n\n\ttext = checkstring(J, 0);\n\n\tif (js_isregexp(J, 1))\n\t\tjs_copy(J, 1);\n\telse if (js_isundefined(J, 1))\n\t\tjs_newregexp(J, \"\", 0);\n\telse\n\t\tjs_newregexp(J, js_tostring(J, 1), 0);\n\n\tre = js_toregexp(J, -1);\n\n\tif (!js_regexec(re->prog, text, &m, 0))\n\t\tjs_pushnumber(J, js_utfptrtoidx(text, m.sub[0].sp));\n\telse\n\t\tjs_pushnumber(J, -1);\n}", "label_name": "Class", "label": 2.0}
@@ -1397,13 +1378,11 @@
{"code": "static void uc_invalidate_tb(struct uc_struct *uc, uint64_t start_addr, size_t len) \n{\n tb_page_addr_t start, end;\n\n // GVA to GPA (GPA -> HVA via page_find, HVA->HPA via host mmu)\n start = get_page_addr_code(uc->cpu->env_ptr, start_addr) & (target_ulong)(-1);\n \n // For 32bit target.\n end = (start + len) & (target_ulong)(-1);\n\n // We get a wrap?\n if (start > end) {\n return;\n }\n\n tb_invalidate_phys_range(uc, start, end);\n}", "label_name": "Pillar", "label": 3.0}
{"code": "static ram_addr_t find_ram_offset(struct uc_struct *uc, ram_addr_t size)\n{\n RAMBlock *block, *next_block;\n ram_addr_t offset = RAM_ADDR_MAX, mingap = RAM_ADDR_MAX;\n\n assert(size != 0); /* it would hand out same offset multiple times */\n\n if (QLIST_EMPTY(&uc->ram_list.blocks)) {\n return 0;\n }\n\n RAMBLOCK_FOREACH(block) {\n ram_addr_t candidate, next = RAM_ADDR_MAX;\n\n /* Align blocks to start on a 'long' in the bitmap\n * which makes the bitmap sync'ing take the fast path.\n */\n candidate = block->offset + block->max_length;\n candidate = ROUND_UP(candidate, BITS_PER_LONG << TARGET_PAGE_BITS);\n\n /* Search for the closest following block\n * and find the gap.\n */\n RAMBLOCK_FOREACH(next_block) {\n if (next_block->offset >= candidate) {\n next = MIN(next, next_block->offset);\n }\n }\n\n /* If it fits remember our place and remember the size\n * of gap, but keep going so that we might find a smaller\n * gap to fill so avoiding fragmentation.\n */\n if (next - candidate >= size && next - candidate < mingap) {\n offset = candidate;\n mingap = next - candidate;\n }\n }\n\n if (offset == RAM_ADDR_MAX) {\n fprintf(stderr, \"Failed to find gap of requested size: %\" PRIu64 \"\\n\",\n (uint64_t)size);\n abort();\n }\n\n return offset;\n}", "label_name": "Base", "label": 1.0}
{"code": "void qemu_ram_free(struct uc_struct *uc, RAMBlock *block)\n{\n if (!block) {\n return;\n }\n\n //if (block->host) {\n // ram_block_notify_remove(block->host, block->max_length);\n //}\n\n QLIST_REMOVE(block, next);\n uc->ram_list.mru_block = NULL;\n /* Write list before version */\n //smp_wmb();\n // call_rcu(block, reclaim_ramblock, rcu);\n reclaim_ramblock(uc, block);\n}", "label_name": "Base", "label": 1.0}
-{"code": "static int match_func(struct libmnt_fs *fs,\n\t\t void *data __attribute__ ((__unused__)))\n{\n\tint rc = flags & FL_INVERT ? 1 : 0;\n\tconst char *m;\n\tvoid *md;\n\n\tm = get_match(COL_FSTYPE);\n\tif (m && !mnt_fs_match_fstype(fs, m))\n\t\treturn rc;\n\n\tm = get_match(COL_OPTIONS);\n\tif (m && !mnt_fs_match_options(fs, m))\n\t\treturn rc;\n\n\tmd = get_match_data(COL_MAJMIN);\n\tif (md && mnt_fs_get_devno(fs) != *((dev_t *) md))\n\t\treturn rc;\n\n\tm = get_match(COL_TARGET);\n\tif (m && !mnt_fs_match_target(fs, m, cache))\n\t\treturn rc;\n\n\tm = get_match(COL_SOURCE);\n\tif (m && !mnt_fs_match_source(fs, m, cache))\n\t\treturn rc;\n\n\tif ((flags & FL_DF) && !(flags & FL_ALL)) {\n\t\tconst char *type = mnt_fs_get_fstype(fs);\n\n\t\tif (type && strstr(type, \"tmpfs\"))\t/* tmpfs is wanted */\n\t\t\treturn !rc;\n\n\t\tif (mnt_fs_is_pseudofs(fs))\n\t\t\treturn rc;\n\t}\n\n\tif ((flags & FL_REAL) && mnt_fs_is_pseudofs(fs))\n\t return rc;\n\n\tif ((flags & FL_PSEUDO) && !mnt_fs_is_pseudofs(fs))\n\t return rc;\n\n\tif ((flags & FL_SHADOWED)) {\n\t\tstruct libmnt_table *tb = NULL;\n\n\t\tmnt_fs_get_table(fs, &tb);\n\t\tif (tb && mnt_table_over_fs(tb, fs, NULL) != 0)\n\t\t\treturn rc;\n\t}\n\n\tif ((flags & FL_DELETED) && !mnt_fs_is_deleted(fs))\n\t\treturn rc;\n\n\treturn !rc;\n}", "label_name": NaN, "label": NaN}
{"code": "static inline struct htx_blk *htx_add_header(struct htx *htx, const struct ist name,\n\t\t\t\t\t const struct ist value)\n{\n\tstruct htx_blk *blk;\n\n\t/* FIXME: check name.len (< 256B) and value.len (< 1MB) */\n\tblk = htx_add_blk(htx, HTX_BLK_HDR, name.len + value.len);\n\tif (!blk)\n\t\treturn NULL;\n\n\tblk->info += (value.len << 8) + name.len;\n\tist2bin_lc(htx_get_blk_ptr(htx, blk), name);\n\tmemcpy(htx_get_blk_ptr(htx, blk) + name.len, value.ptr, value.len);\n\treturn blk;\n}", "label_name": "Base", "label": 1.0}
{"code": "void ZydisFormatterBufferInit(ZydisFormatterBuffer* buffer, char* user_buffer,\n ZyanUSize length)\n{\n ZYAN_ASSERT(buffer);\n ZYAN_ASSERT(user_buffer);\n ZYAN_ASSERT(length);\n\n buffer->is_token_list = ZYAN_FALSE;\n buffer->string.flags = ZYAN_STRING_HAS_FIXED_CAPACITY;\n buffer->string.vector.allocator = ZYAN_NULL;\n buffer->string.vector.element_size = sizeof(char);\n buffer->string.vector.size = 1;\n buffer->string.vector.capacity = length;\n buffer->string.vector.data = user_buffer;\n *user_buffer = '\\0';\n}", "label_name": "Variant", "label": 0.0}
{"code": "void ZydisFormatterBufferInitTokenized(ZydisFormatterBuffer* buffer,\n ZydisFormatterToken** first_token, void* user_buffer, ZyanUSize length)\n{\n ZYAN_ASSERT(buffer);\n ZYAN_ASSERT(first_token);\n ZYAN_ASSERT(user_buffer);\n ZYAN_ASSERT(length);\n\n *first_token = user_buffer;\n (*first_token)->type = ZYDIS_TOKEN_INVALID;\n (*first_token)->next = 0;\n\n user_buffer = (ZyanU8*)user_buffer + sizeof(ZydisFormatterToken);\n length -= sizeof(ZydisFormatterToken);\n\n buffer->is_token_list = ZYAN_TRUE;\n buffer->capacity = length;\n buffer->string.flags = ZYAN_STRING_HAS_FIXED_CAPACITY;\n buffer->string.vector.allocator = ZYAN_NULL;\n buffer->string.vector.element_size = sizeof(char);\n buffer->string.vector.size = 1;\n buffer->string.vector.capacity = length;\n buffer->string.vector.data = user_buffer;\n *(char*)user_buffer = '\\0';\n}", "label_name": "Variant", "label": 0.0}
{"code": "destroyUserInformationLists(DUL_USERINFO * userInfo)\n{\n PRV_SCUSCPROLE\n * role;\n\n role = (PRV_SCUSCPROLE*)LST_Dequeue(&userInfo->SCUSCPRoleList);\n while (role != NULL) {\n free(role);\n role = (PRV_SCUSCPROLE*)LST_Dequeue(&userInfo->SCUSCPRoleList);\n }\n LST_Destroy(&userInfo->SCUSCPRoleList);\n\n /* extended negotiation */\n delete userInfo->extNegList; userInfo->extNegList = NULL;\n\n /* user identity negotiation */\n delete userInfo->usrIdent; userInfo->usrIdent = NULL;\n}", "label_name": "Variant", "label": 0.0}
{"code": "SWTPM_NVRAM_CheckHeader(unsigned char *data, uint32_t length,\n uint32_t *dataoffset, uint16_t *hdrflags,\n uint8_t *hdrversion, bool quiet)\n{\n blobheader *bh = (blobheader *)data;\n\n if (length < sizeof(bh)) {\n if (!quiet)\n logprintf(STDERR_FILENO,\n \"not enough bytes for header: %u\\n\", length);\n return TPM_BAD_PARAMETER;\n }\n\n if (ntohl(bh->totlen) != length) {\n if (!quiet)\n logprintf(STDERR_FILENO,\n \"broken header: bh->totlen %u != %u\\n\",\n htonl(bh->totlen), length);\n return TPM_BAD_PARAMETER;\n }\n\n if (bh->min_version > BLOB_HEADER_VERSION) {\n if (!quiet)\n logprintf(STDERR_FILENO,\n \"Minimum required version for the blob is %d, we \"\n \"only support version %d\\n\", bh->min_version,\n BLOB_HEADER_VERSION);\n return TPM_BAD_VERSION;\n }\n\n *hdrversion = bh->version;\n *dataoffset = ntohs(bh->hdrsize);\n *hdrflags = ntohs(bh->flags);\n\n return TPM_SUCCESS;\n}", "label_name": "Base", "label": 1.0}
-{"code": "static void create_layer_surface(struct swaylock_surface *surface) {\n\tstruct swaylock_state *state = surface->state;\n\n\tsurface->image = select_image(state, surface);\n\n\tsurface->surface = wl_compositor_create_surface(state->compositor);\n\tassert(surface->surface);\n\n\tsurface->child = wl_compositor_create_surface(state->compositor);\n\tassert(surface->child);\n\tsurface->subsurface = wl_subcompositor_get_subsurface(state->subcompositor, surface->child, surface->surface);\n\tassert(surface->subsurface);\n\twl_subsurface_set_sync(surface->subsurface);\n\n\tsurface->layer_surface = zwlr_layer_shell_v1_get_layer_surface(\n\t\t\tstate->layer_shell, surface->surface, surface->output,\n\t\t\tZWLR_LAYER_SHELL_V1_LAYER_OVERLAY, \"lockscreen\");\n\tassert(surface->layer_surface);\n\n\tzwlr_layer_surface_v1_set_size(surface->layer_surface, 0, 0);\n\tzwlr_layer_surface_v1_set_anchor(surface->layer_surface,\n\t\t\tZWLR_LAYER_SURFACE_V1_ANCHOR_TOP |\n\t\t\tZWLR_LAYER_SURFACE_V1_ANCHOR_RIGHT |\n\t\t\tZWLR_LAYER_SURFACE_V1_ANCHOR_BOTTOM |\n\t\t\tZWLR_LAYER_SURFACE_V1_ANCHOR_LEFT);\n\tzwlr_layer_surface_v1_set_exclusive_zone(surface->layer_surface, -1);\n\tzwlr_layer_surface_v1_set_keyboard_interactivity(\n\t\t\tsurface->layer_surface, true);\n\tzwlr_layer_surface_v1_add_listener(surface->layer_surface,\n\t\t\t&layer_surface_listener, surface);\n\n\tif (surface_is_opaque(surface) &&\n\t\t\tsurface->state->args.mode != BACKGROUND_MODE_CENTER &&\n\t\t\tsurface->state->args.mode != BACKGROUND_MODE_FIT) {\n\t\tstruct wl_region *region =\n\t\t\twl_compositor_create_region(surface->state->compositor);\n\t\twl_region_add(region, 0, 0, INT32_MAX, INT32_MAX);\n\t\twl_surface_set_opaque_region(surface->surface, region);\n\t\twl_region_destroy(region);\n\t}\n\n\twl_surface_commit(surface->surface);\n}", "label_name": NaN, "label": NaN}
{"code": " def get_ipcache_entry(self, client):\n \"\"\"Build a cache of dns results.\"\"\"\n if client in self.ipcache:\n if self.ipcache[client]:\n return self.ipcache[client]\n else:\n raise socket.gaierror\n else:\n # need to add entry\n try:\n ipaddr = socket.gethostbyname(client)\n self.ipcache[client] = (ipaddr, client)\n return (ipaddr, client)\n except socket.gaierror:\n cmd = \"getent hosts %s\" % client\n ipaddr = Popen(cmd, shell=True, \\\n stdout=PIPE).stdout.read().strip().split()\n if ipaddr:\n self.ipcache[client] = (ipaddr, client)\n return (ipaddr, client)\n self.ipcache[client] = False\n self.logger.error(\"Failed to find IP address for %s\" % client)\n raise socket.gaierror", "label_name": "Class", "label": 2.0}
{"code": " def verify_cert_against_key(self, filename, key_filename):\n \"\"\"\n check that a certificate validates against its private key.\n \"\"\"\n cert = self.data + filename\n key = self.data + key_filename\n cmd = \"openssl x509 -noout -modulus -in %s | openssl md5\" % cert\n cert_md5 = Popen(cmd, shell=True, stdout=PIPE, stderr=STDOUT).stdout.read()\n cmd = \"openssl rsa -noout -modulus -in %s | openssl md5\" % key\n key_md5 = Popen(cmd, shell=True, stdout=PIPE, stderr=STDOUT).stdout.read()\n if cert_md5 == key_md5:\n return True\n return False", "label_name": "Class", "label": 2.0}
{"code": " def get_revision(self):\n \"\"\"Read svn revision information for the Bcfg2 repository.\"\"\"\n try:\n data = Popen((\"env LC_ALL=C svn info %s\" %\n (self.datastore)), shell=True,\n stdout=PIPE).communicate()[0].split('\\n')\n return [line.split(': ')[1] for line in data \\\n if line[:9] == 'Revision:'][-1]\n except IndexError:\n logger.error(\"Failed to read svn info; disabling svn support\")\n logger.error('''Ran command \"svn info %s\"''' % (self.datastore))\n logger.error(\"Got output: %s\" % data)\n raise Bcfg2.Server.Plugin.PluginInitError", "label_name": "Class", "label": 2.0}
@@ -1446,7 +1425,6 @@
{"code": " def __getattr__(_self, attr):\n if attr == \"nameResolver\":\n return nameResolver\n else:\n return getattr(real_reactor, attr)", "label_name": "Base", "label": 1.0}
{"code": " def prepare(self, reactor, clock, hs):\n # build a replication server\n server_factory = ReplicationStreamProtocolFactory(hs)\n self.streamer = hs.get_replication_streamer()\n self.server = server_factory.buildProtocol(None)\n\n # Make a new HomeServer object for the worker\n self.reactor.lookups[\"testserv\"] = \"1.2.3.4\"\n self.worker_hs = self.setup_test_homeserver(\n http_client=None,\n homeserver_to_use=GenericWorkerServer,\n config=self._get_worker_hs_config(),\n reactor=self.reactor,\n )\n\n # Since we use sqlite in memory databases we need to make sure the\n # databases objects are the same.\n self.worker_hs.get_datastore().db_pool = hs.get_datastore().db_pool\n\n self.test_handler = self._build_replication_data_handler()\n self.worker_hs._replication_data_handler = self.test_handler\n\n repl_handler = ReplicationCommandHandler(self.worker_hs)\n self.client = ClientReplicationStreamProtocol(\n self.worker_hs, \"client\", \"test\", clock, repl_handler,\n )\n\n self._client_transport = None\n self._server_transport = None", "label_name": "Base", "label": 1.0}
{"code": " def make_homeserver(self, reactor, clock):\n config = self.default_config()\n config[\"redaction_retention_period\"] = \"30d\"\n return self.setup_test_homeserver(\n resource_for_federation=Mock(), http_client=None, config=config\n )", "label_name": "Base", "label": 1.0}
-{"code": " def makepasv(self):\n if self.af == socket.AF_INET:\n host, port = parse227(self.sendcmd('PASV'))\n else:\n host, port = parse229(self.sendcmd('EPSV'), self.sock.getpeername())\n return host, port", "label_name": NaN, "label": NaN}
{"code": "def preprocess_input_exprs_arg_string(input_exprs_str):\n \"\"\"Parses input arg into dictionary that maps input key to python expression.\n\n Parses input string in the format of 'input_key=' into a\n dictionary that maps each input_key to its python expression.\n\n Args:\n input_exprs_str: A string that specifies python expression for input keys.\n Each input is separated by semicolon. For each input key:\n 'input_key='\n\n Returns:\n A dictionary that maps input keys to their values.\n\n Raises:\n RuntimeError: An error when the given input string is in a bad format.\n \"\"\"\n input_dict = {}\n\n for input_raw in filter(bool, input_exprs_str.split(';')):\n if '=' not in input_exprs_str:\n raise RuntimeError('--input_exprs \"%s\" format is incorrect. Please follow'\n '\"=\"' % input_exprs_str)\n input_key, expr = input_raw.split('=', 1)\n # ast.literal_eval does not work with numpy expressions\n input_dict[input_key] = eval(expr) # pylint: disable=eval-used\n return input_dict", "label_name": "Base", "label": 1.0}
{"code": " def testPartialIndexInsert(self):\n with ops.Graph().as_default() as G:\n with ops.device('/cpu:0'):\n x = array_ops.placeholder(dtypes.float32)\n f = array_ops.placeholder(dtypes.float32)\n v = array_ops.placeholder(dtypes.float32)\n pi = array_ops.placeholder(dtypes.int64)\n gi = array_ops.placeholder(dtypes.int64)\n with ops.device(test.gpu_device_name()):\n stager = data_flow_ops.MapStagingArea(\n [dtypes.float32, dtypes.float32, dtypes.float32])\n stage_xf = stager.put(pi, [x, f], [0, 2])\n stage_v = stager.put(pi, [v], [1])\n key, ret = stager.get(gi)\n size = stager.size()\n isize = stager.incomplete_size()\n\n G.finalize()\n\n with self.session(graph=G) as sess:\n # 0 complete and incomplete entries\n self.assertTrue(sess.run([size, isize]) == [0, 0])\n # Stage key 0, x and f tuple entries\n sess.run(stage_xf, feed_dict={pi: 0, x: 1, f: 2})\n self.assertTrue(sess.run([size, isize]) == [0, 1])\n # Stage key 1, x and f tuple entries\n sess.run(stage_xf, feed_dict={pi: 1, x: 1, f: 2})\n self.assertTrue(sess.run([size, isize]) == [0, 2])\n\n # Now complete key 0 with tuple entry v\n sess.run(stage_v, feed_dict={pi: 0, v: 1})\n # 1 complete and 1 incomplete entry\n self.assertTrue(sess.run([size, isize]) == [1, 1])\n # We can now obtain tuple associated with key 0\n self.assertTrue(sess.run([key, ret], feed_dict={gi: 0}) == [0, [1, 1, 2]])\n\n # 0 complete and 1 incomplete entry\n self.assertTrue(sess.run([size, isize]) == [0, 1])\n # Now complete key 1 with tuple entry v\n sess.run(stage_v, feed_dict={pi: 1, v: 3})\n # We can now obtain tuple associated with key 1\n self.assertTrue(sess.run([key, ret], feed_dict={gi: 1}) == [1, [1, 3, 2]])", "label_name": "Base", "label": 1.0}
{"code": " def testMultiple(self):\n with ops.Graph().as_default() as G:\n with ops.device('/cpu:0'):\n x = array_ops.placeholder(dtypes.float32)\n pi = array_ops.placeholder(dtypes.int64)\n gi = array_ops.placeholder(dtypes.int64)\n v = 2. * (array_ops.zeros([128, 128]) + x)\n with ops.device(test.gpu_device_name()):\n stager = data_flow_ops.MapStagingArea([dtypes.float32, dtypes.float32])\n stage = stager.put(pi, [x, v], [0, 1])\n k, (z, y) = stager.get(gi)\n y = math_ops.reduce_max(z * math_ops.matmul(y, y))\n\n G.finalize()\n\n with self.session(graph=G) as sess:\n sess.run(stage, feed_dict={x: -1, pi: 0})\n for i in range(10):\n _, yval = sess.run([stage, y], feed_dict={x: i, pi: i + 1, gi: i})\n self.assertAllClose(\n 4 * (i - 1) * (i - 1) * (i - 1) * 128, yval, rtol=1e-4)", "label_name": "Base", "label": 1.0}
@@ -1480,7 +1458,6 @@
{"code": "def test_digest_object_stale():\n credentials = (\"joe\", \"password\")\n host = None\n request_uri = \"/digest/stale/\"\n headers = {}\n response = httplib2.Response({})\n response[\"www-authenticate\"] = (\n 'Digest realm=\"myrealm\", nonce=\"bd669f\", '\n 'algorithm=MD5, qop=\"auth\", stale=true'\n )\n response.status = 401\n content = b\"\"\n d = httplib2.DigestAuthentication(\n credentials, host, request_uri, headers, response, content, None\n )\n # Returns true to force a retry\n assert d.response(response, content)", "label_name": "Class", "label": 2.0}
{"code": "def authorized():\n resp = google_remote_app().authorized_response()\n access_token = resp[\"access_token\"]\n\n if access_token is None:\n logger.warning(\"Access token missing in call back request.\")\n flash(\"Validation error. Please retry.\")\n return redirect(url_for(\"redash.login\"))\n\n profile = get_user_profile(access_token)\n if profile is None:\n flash(\"Validation error. Please retry.\")\n return redirect(url_for(\"redash.login\"))\n\n if \"org_slug\" in session:\n org = models.Organization.get_by_slug(session.pop(\"org_slug\"))\n else:\n org = current_org\n\n if not verify_profile(org, profile):\n logger.warning(\n \"User tried to login with unauthorized domain name: %s (org: %s)\",\n profile[\"email\"],\n org,\n )\n flash(\"Your Google Apps account ({}) isn't allowed.\".format(profile[\"email\"]))\n return redirect(url_for(\"redash.login\", org_slug=org.slug))\n\n picture_url = \"%s?sz=40\" % profile[\"picture\"]\n user = create_and_login_user(org, profile[\"name\"], profile[\"email\"], picture_url)\n if user is None:\n return logout_and_redirect_to_index()\n\n unsafe_next_path = request.args.get(\"state\") or url_for(\n \"redash.index\", org_slug=org.slug\n )\n next_path = get_next_path(unsafe_next_path)\n\n return redirect(next_path)", "label_name": "Base", "label": 1.0}
{"code": " def test_get_response_httperror_exception(self, mock_get):\n mock_response = mock.Mock()\n mock_response.status_code = 500\n mock_response.text = \"Server Error\"\n http_error = requests.HTTPError()\n mock_response.raise_for_status.side_effect = http_error\n mock_get.return_value = mock_response\n\n url = \"https://example.com/\"\n query_runner = BaseHTTPQueryRunner({})\n response, error = query_runner.get_response(url)\n mock_get.assert_called_once_with(\"get\", url, auth=None)\n self.assertIsNotNone(error)\n self.assertIn(\"Failed to execute query\", error)", "label_name": "Base", "label": 1.0}
-{"code": " def gen_response_objects(self):\n for o in self.event_json['Event']['Object']:\n self.response += object_to_entity(o)", "label_name": NaN, "label": NaN}
{"code": "def env() -> Environment:\n from openapi_python_client import utils\n\n TEMPLATE_FILTERS = {\"snakecase\": utils.snake_case, \"spinalcase\": utils.spinal_case}\n env = Environment(loader=PackageLoader(\"openapi_python_client\"), trim_blocks=True, lstrip_blocks=True)\n env.filters.update(TEMPLATE_FILTERS)\n return env", "label_name": "Base", "label": 1.0}
{"code": " def __init__(self, *, openapi: GeneratorData) -> None:\n self.openapi: GeneratorData = openapi\n self.env: Environment = Environment(loader=PackageLoader(__package__), trim_blocks=True, lstrip_blocks=True)\n\n self.project_name: str = self.project_name_override or f\"{utils.kebab_case(openapi.title).lower()}-client\"\n self.project_dir: Path = Path.cwd() / self.project_name\n\n self.package_name: str = self.package_name_override or self.project_name.replace(\"-\", \"_\")\n self.package_dir: Path = self.project_dir / self.package_name\n self.package_description: str = f\"A client library for accessing {self.openapi.title}\"\n self.version: str = openapi.version\n\n self.env.filters.update(self.TEMPLATE_FILTERS)", "label_name": "Base", "label": 1.0}
{"code": " def __post_init__(self) -> None:\n super().__post_init__()\n if self.default is not None:\n self.default = f'\"{self.default}\"'", "label_name": "Base", "label": 1.0}
@@ -1516,7 +1493,6 @@
{"code": "def adv_search_serie(q, include_series_inputs, exclude_series_inputs):\n for serie in include_series_inputs:\n q = q.filter(db.Books.series.any(db.Series.id == serie))\n for serie in exclude_series_inputs:\n q = q.filter(not_(db.Books.series.any(db.Series.id == serie)))\n return q", "label_name": "Base", "label": 1.0}
{"code": "def adv_search_shelf(q, include_shelf_inputs, exclude_shelf_inputs):\n q = q.outerjoin(ub.BookShelf, db.Books.id == ub.BookShelf.book_id)\\\n .filter(or_(ub.BookShelf.shelf == None, ub.BookShelf.shelf.notin_(exclude_shelf_inputs)))\n if len(include_shelf_inputs) > 0:\n q = q.filter(ub.BookShelf.shelf.in_(include_shelf_inputs))\n return q", "label_name": "Base", "label": 1.0}
{"code": " def strip_illegal_bytes_parser(xml):\n return parse(BytesIO(re_xml_illegal_bytes.sub(b'', xml)))", "label_name": "Base", "label": 1.0}
-{"code": "def setup_passwordless_with_other_nodes(init_node):\n \"\"\"\n Setup passwordless with other cluster nodes\n\n Should fetch the node list from init node, then swap the key\n \"\"\"\n # Fetch cluster nodes list\n cmd = \"ssh -o StrictHostKeyChecking=no root@{} crm_node -l\".format(init_node)\n rc, out, err = utils.get_stdout_stderr(cmd)\n if rc != 0:\n error(\"Can't fetch cluster nodes list from {}: {}\".format(init_node, err))\n cluster_nodes_list = []\n for line in out.splitlines():\n _, node, stat = line.split()\n if stat == \"member\":\n cluster_nodes_list.append(node)\n\n # Filter out init node from cluster_nodes_list\n cmd = \"ssh -o StrictHostKeyChecking=no root@{} hostname\".format(init_node)\n rc, out, err = utils.get_stdout_stderr(cmd)\n if rc != 0:\n error(\"Can't fetch hostname of {}: {}\".format(init_node, err))\n if out in cluster_nodes_list:\n cluster_nodes_list.remove(out)\n\n # Swap ssh public key between join node and other cluster nodes\n for node in cluster_nodes_list:\n swap_public_ssh_key(node)", "label_name": NaN, "label": NaN}
{"code": "def useruid(s, login):\n \"\"\"Connect to a LDAP and check the uid matching the given field data\"\"\"\n uid = False\n c = Connection(s, config.LDAPACC, \n password=config.LDAPPASS, auto_bind=True)\n\n if c.result[\"description\"] != \"success\":\n app.logger.error(\"Error connecting to the LDAP with the service account\")\n return False\n\n # Look for the user entry.\n if not c.search(config.LDAPBASE,\n \"(\" + config.LDAPFIELD + \"=\" + login + \")\") :\n app.logger.error(\"Error: Connection to the LDAP with service account failed\")\n else:\n if len(c.entries) >= 1 :\n if len(c.entries) > 1 :\n app.logger.error(\"Error: multiple entries with this login. \"+ \\\n \"Trying first entry...\")\n uid = c.entries[0].entry_dn\n else:\n app.logger.error(\"Error: Login not found\")\n c.unbind()\n \n return uid", "label_name": "Class", "label": 2.0}
{"code": "def test_post_broken_body():\n response = client.post(\"/items/\", data={\"name\": \"Foo\", \"price\": 50.5})\n assert response.status_code == 422, response.text\n assert response.json() == {\n \"detail\": [\n {\n \"ctx\": {\n \"colno\": 1,\n \"doc\": \"name=Foo&price=50.5\",\n \"lineno\": 1,\n \"msg\": \"Expecting value\",\n \"pos\": 0,\n },\n \"loc\": [\"body\", 0],\n \"msg\": \"Expecting value: line 1 column 1 (char 0)\",\n \"type\": \"value_error.jsondecode\",\n }\n ]\n }\n with patch(\"json.loads\", side_effect=Exception):\n response = client.post(\"/items/\", json={\"test\": \"test2\"})\n assert response.status_code == 400, response.text\n assert response.json() == {\"detail\": \"There was an error parsing the body\"}", "label_name": "Compound", "label": 4.0}
{"code": "def parse_jwt_token(request: func.HttpRequest) -> Result[UserInfo]:\n \"\"\"Obtains the Access Token from the Authorization Header\"\"\"\n token_str = get_auth_token(request)\n if token_str is None:\n return Error(\n code=ErrorCode.INVALID_REQUEST,\n errors=[\"unable to find authorization token\"],\n )\n\n # This token has already been verified by the azure authentication layer\n token = jwt.decode(token_str, options={\"verify_signature\": False})\n\n application_id = UUID(token[\"appid\"]) if \"appid\" in token else None\n object_id = UUID(token[\"oid\"]) if \"oid\" in token else None\n upn = token.get(\"upn\")\n return UserInfo(application_id=application_id, object_id=object_id, upn=upn)", "label_name": "Class", "label": 2.0}
@@ -1548,8 +1524,6 @@
{"code": "def get_object_src_http(dataset, rel_path):\n path = _get_obj_abosolute_path(dataset, rel_path)\n response = send_file(path,\n cache_timeout=datetime.timedelta(\n days=365).total_seconds(),\n add_etags=True,\n conditional=True)\n return response", "label_name": "Base", "label": 1.0}
{"code": " public SAXParser newSAXParser() throws ParserConfigurationException {\n final SAXParser ret;\n try {\n ret = new XercesJAXPSAXParser(this, features, validating, handleXInclude);\n } catch (final SAXException se) {\n // Translate to ParserConfigurationException\n throw new OXFException(se); // so we see a decent stack trace!\n// throw new ParserConfigurationException(se.getMessage());\n }\n return ret;\n }", "label_name": NaN, "label": NaN}
{"code": "\tpublic Template getTemplate(\n\t\tString templateId, String templateContent, String errorTemplateId,\n\t\tString errorTemplateContent, TemplateContextType templateContextType) {\n\n\t\tif (templateContextType.equals(TemplateContextType.EMPTY)) {\n\t\t\treturn new FreeMarkerTemplate(\n\t\t\t\t\ttemplateId, templateContent, errorTemplateId,\n\t\t\t\t\terrorTemplateContent, null, _configuration,\n\t\t\t\t\t_templateContextHelper, _stringTemplateLoader);\n\t\t}\n\t\telse if (templateContextType.equals(TemplateContextType.RESTRICTED)) {\n\t\t\treturn new RestrictedTemplate(\n\t\t\t\tnew FreeMarkerTemplate(\n\t\t\t\t\ttemplateId, templateContent, errorTemplateId,\n\t\t\t\t\terrorTemplateContent, _restrictedHelperUtilities,\n\t\t\t\t\t_configuration, _templateContextHelper,\n\t\t\t\t\t_stringTemplateLoader),\n\t\t\t\t_templateContextHelper.getRestrictedVariables());\n\t\t}\n\t\telse if (templateContextType.equals(TemplateContextType.STANDARD)) {\n\t\t\treturn new FreeMarkerTemplate(\n\t\t\t\ttemplateId, templateContent, errorTemplateId,\n\t\t\t\terrorTemplateContent, _standardHelperUtilities, _configuration,\n\t\t\t\t_templateContextHelper, _stringTemplateLoader);\n\t\t}\n\n\t\treturn null;\n\t}", "label_name": NaN, "label": NaN}
-{"code": " public void run() {\n try {\n mcs.joinGroup(MULTICAST);\n ready.signal();\n\n while(true) {\n byte[] buf = new byte[2048];\n DatagramPacket p = new DatagramPacket(buf,buf.length);\n mcs.receive(p);\n\n SocketAddress sender = p.getSocketAddress();\n\n // prepare a response\n TcpSlaveAgentListener tal = jenkins.getTcpSlaveAgentListener();\n\n StringBuilder rsp = new StringBuilder(\"\");\n tag(rsp,\"version\", Jenkins.VERSION);\n tag(rsp,\"url\", jenkins.getRootUrl());\n tag(rsp,\"server-id\", Util.getDigestOf(jenkins.getSecretKey()));\n tag(rsp,\"slave-port\",tal==null?null:tal.getPort());\n\n for (UDPBroadcastFragment f : UDPBroadcastFragment.all())\n f.buildFragment(rsp,sender);\n\n rsp.append(\"\");\n\n byte[] response = rsp.toString().getBytes(\"UTF-8\");\n mcs.send(new DatagramPacket(response,response.length,sender));\n }\n } catch (ClosedByInterruptException e) {\n // shut down\n } catch (BindException e) {\n // if we failed to listen to UDP, just silently abandon it, as a stack trace\n // makes people unnecessarily concerned, for a feature that currently does no good.\n LOGGER.log(Level.WARNING, \"Failed to listen to UDP port \"+PORT,e);\n } catch (IOException e) {\n if (shutdown) return; // forcibly closed\n LOGGER.log(Level.WARNING, \"UDP handling problem\",e);\n }\n }", "label_name": NaN, "label": NaN}
-{"code": " protected String getSecretKey() {\n return Jenkins.getInstance().getSecretKey();\n }", "label_name": NaN, "label": NaN}
{"code": " private SecretKey getKey() {\n try {\n if (secret==null) {\n synchronized (this) {\n if (secret==null) {\n byte[] payload = load();\n if (payload==null) {\n payload = ConfidentialStore.get().randomBytes(256);\n store(payload);\n }\n // Due to the stupid US export restriction JDK only ships 128bit version.\n secret = new SecretKeySpec(payload,0,128/8, ALGORITHM);\n }\n }\n }\n return secret;\n } catch (IOException e) {\n throw new Error(\"Failed to load the key: \"+getId(),e);\n }\n }", "label_name": "Class", "label": 2.0}
{"code": " private boolean exractAndLoad(ArrayList errors, String version, String customPath, String resourcePath) {\n URL resource = classLoader.getResource(resourcePath);\n if( resource !=null ) {\n \n String libName = name + \"-\" + getBitModel();\n if( version !=null) {\n libName += \"-\" + version;\n }\n \n if( customPath!=null ) {\n // Try to extract it to the custom path...\n File target = file(customPath, map(libName));\n if( extract(errors, resource, target) ) {\n if( load(errors, target) ) {\n return true;\n }\n }\n }\n \n // Fall back to extracting to the tmp dir\n customPath = System.getProperty(\"java.io.tmpdir\");\n File target = file(customPath, map(libName));\n if( extract(errors, resource, target) ) {\n if( load(errors, target) ) {\n return true;\n }\n }\n }\n return false;\n }", "label_name": "Base", "label": 1.0}
{"code": " public void onSubmit()\n {\n final String sessionCsrfToken = getCsrfSessionToken();\n final String postedCsrfToken = this.csrfTokenField.getInput();\n if (StringUtils.equals(sessionCsrfToken, postedCsrfToken) == false) {\n log.error(\"Cross site request forgery alert. csrf token doesn't match! session csrf token=\"\n + sessionCsrfToken\n + \", posted csrf token=\"\n + postedCsrfToken);\n throw new InternalErrorException(\"errorpage.csrfError\");\n }\n }", "label_name": "Compound", "label": 4.0}
@@ -1672,7 +1646,6 @@
{"code": " private void ondata(String data) {\n try {\n this.decoder.add(data);\n } catch (DecodingException e) {\n this.onerror(e);\n }\n }", "label_name": "Base", "label": 1.0}
{"code": " public AjaxResult updateAvatar(@RequestParam(\"avatarfile\") MultipartFile file)\r\n {\r\n SysUser currentUser = getSysUser();\r\n try\r\n {\r\n if (!file.isEmpty())\r\n {\r\n String avatar = FileUploadUtils.upload(RuoYiConfig.getAvatarPath(), file);\r\n currentUser.setAvatar(avatar);\r\n if (userService.updateUserInfo(currentUser) > 0)\r\n {\r\n setSysUser(userService.selectUserById(currentUser.getUserId()));\r\n return success();\r\n }\r\n }\r\n return error();\r\n }\r\n catch (Exception e)\r\n {\r\n log.error(\"\u4fee\u6539\u5934\u50cf\u5931\u8d25\uff01\", e);\r\n return error(e.getMessage());\r\n }\r\n }\r", "label_name": "Base", "label": 1.0}
{"code": "armv6pmu_handle_irq(int irq_num,\n\t\t void *dev)\n{\n\tunsigned long pmcr = armv6_pmcr_read();\n\tstruct perf_sample_data data;\n\tstruct cpu_hw_events *cpuc;\n\tstruct pt_regs *regs;\n\tint idx;\n\n\tif (!armv6_pmcr_has_overflowed(pmcr))\n\t\treturn IRQ_NONE;\n\n\tregs = get_irq_regs();\n\n\t/*\n\t * The interrupts are cleared by writing the overflow flags back to\n\t * the control register. All of the other bits don't have any effect\n\t * if they are rewritten, so write the whole value back.\n\t */\n\tarmv6_pmcr_write(pmcr);\n\n\tperf_sample_data_init(&data, 0);\n\n\tcpuc = &__get_cpu_var(cpu_hw_events);\n\tfor (idx = 0; idx <= armpmu->num_events; ++idx) {\n\t\tstruct perf_event *event = cpuc->events[idx];\n\t\tstruct hw_perf_event *hwc;\n\n\t\tif (!test_bit(idx, cpuc->active_mask))\n\t\t\tcontinue;\n\n\t\t/*\n\t\t * We have a single interrupt for all counters. Check that\n\t\t * each counter has overflowed before we process it.\n\t\t */\n\t\tif (!armv6_pmcr_counter_has_overflowed(pmcr, idx))\n\t\t\tcontinue;\n\n\t\thwc = &event->hw;\n\t\tarmpmu_event_update(event, hwc, idx, 1);\n\t\tdata.period = event->hw.last_period;\n\t\tif (!armpmu_event_set_period(event, hwc, idx))\n\t\t\tcontinue;\n\n\t\tif (perf_event_overflow(event, 0, &data, regs))\n\t\t\tarmpmu->disable(hwc, idx);\n\t}\n\n\t/*\n\t * Handle the pending perf events.\n\t *\n\t * Note: this call *must* be run with interrupts disabled. For\n\t * platforms that can have the PMU interrupts raised as an NMI, this\n\t * will not work.\n\t */\n\tirq_work_run();\n\n\treturn IRQ_HANDLED;\n}", "label_name": "Class", "label": 2.0}
-{"code": "static int em_ret_near_imm(struct x86_emulate_ctxt *ctxt)\n{\n\tint rc;\n\n\tctxt->dst.type = OP_REG;\n\tctxt->dst.addr.reg = &ctxt->_eip;\n\tctxt->dst.bytes = ctxt->op_bytes;\n\trc = emulate_pop(ctxt, &ctxt->dst.val, ctxt->op_bytes);\n\tif (rc != X86EMUL_CONTINUE)\n\t\treturn rc;\n\trsp_increment(ctxt, ctxt->src.val);\n\treturn X86EMUL_CONTINUE;\n}", "label_name": NaN, "label": NaN}
{"code": "compat_mpt_command(struct file *filp, unsigned int cmd,\n\t\t\tunsigned long arg)\n{\n\tstruct mpt_ioctl_command32 karg32;\n\tstruct mpt_ioctl_command32 __user *uarg = (struct mpt_ioctl_command32 __user *) arg;\n\tstruct mpt_ioctl_command karg;\n\tMPT_ADAPTER *iocp = NULL;\n\tint iocnum, iocnumX;\n\tint nonblock = (filp->f_flags & O_NONBLOCK);\n\tint ret;\n\n\tif (copy_from_user(&karg32, (char __user *)arg, sizeof(karg32)))\n\t\treturn -EFAULT;\n\n\t/* Verify intended MPT adapter */\n\tiocnumX = karg32.hdr.iocnum & 0xFF;\n\tif (((iocnum = mpt_verify_adapter(iocnumX, &iocp)) < 0) ||\n\t (iocp == NULL)) {\n\t\tprintk(KERN_DEBUG MYNAM \"::compat_mpt_command @%d - ioc%d not found!\\n\",\n\t\t\t__LINE__, iocnumX);\n\t\treturn -ENODEV;\n\t}\n\n\tif ((ret = mptctl_syscall_down(iocp, nonblock)) != 0)\n\t\treturn ret;\n\n\tdctlprintk(iocp, printk(MYIOC_s_DEBUG_FMT \"compat_mpt_command() called\\n\",\n\t iocp->name));\n\t/* Copy data to karg */\n\tkarg.hdr.iocnum = karg32.hdr.iocnum;\n\tkarg.hdr.port = karg32.hdr.port;\n\tkarg.timeout = karg32.timeout;\n\tkarg.maxReplyBytes = karg32.maxReplyBytes;\n\n\tkarg.dataInSize = karg32.dataInSize;\n\tkarg.dataOutSize = karg32.dataOutSize;\n\tkarg.maxSenseBytes = karg32.maxSenseBytes;\n\tkarg.dataSgeOffset = karg32.dataSgeOffset;\n\n\tkarg.replyFrameBufPtr = (char __user *)(unsigned long)karg32.replyFrameBufPtr;\n\tkarg.dataInBufPtr = (char __user *)(unsigned long)karg32.dataInBufPtr;\n\tkarg.dataOutBufPtr = (char __user *)(unsigned long)karg32.dataOutBufPtr;\n\tkarg.senseDataPtr = (char __user *)(unsigned long)karg32.senseDataPtr;\n\n\t/* Pass new structure to do_mpt_command\n\t */\n\tret = mptctl_do_mpt_command (karg, &uarg->MF);\n\n\tmutex_unlock(&iocp->ioctl_cmds.mutex);\n\n\treturn ret;\n}", "label_name": "Class", "label": 2.0}
{"code": "YCPBoolean IniAgent::Write(const YCPPath &path, const YCPValue& value, const YCPValue& arg)\n{\n if (!parser.isStarted())\n {\n\ty2warning(\"Can't execute Write before being mounted.\");\n\treturn YCPBoolean (false);\n }\n // no need to update if modified, we are changing value\n\n bool ok = false; // is the _path_ ok?\n // return value\n YCPBoolean b (true);\n\n if (0 == path->length ())\n {\n\tif (value->isString() && value->asString()->value() == \"force\")\n\t parser.inifile.setDirty();\n\telse if (value->isString () && value->asString()->value() == \"clean\")\n\t parser.inifile.clean ();\n\tif (0 != parser.write ())\n\t b = false;\n\tok = true;\n }\n else\n {\n\tif (( parser.repeatNames () && value->isList ()) ||\n\t (!parser.repeatNames () && (value->isString () || value->isInteger())) ||\n\t path->component_str(0) == \"all\"\n\t )\n\t {\n\t\tok = true;\n\t\tif (parser.inifile.Write (path, value, parser.HaveRewrites ()))\n\t\t b = false;\n\t }\n else if (value->isVoid ())\n\t {\n\t\tint wb = -1;\n\t\tstring del_sec = \"\";\n\t\tok = true;\n\t\tif (2 == path->length ())\n\t\t{\n\t\t string pc = path->component_str(0);\n\t\t if (\"s\" == pc || \"section\" == pc)\n\t\t {\t// request to delete section. Find the file name\n\t\t\tdel_sec = path->component_str (1);\n\t\t\twb = parser.inifile.getSubSectionRewriteBy (del_sec.c_str());\n\t\t }\n\t\t}\n\t\tif (parser.inifile.Delete (path))\n\t\t b = false;\n\t\telse if (del_sec != \"\")\n\t\t{\n\t\t parser.deleted_sections.insert (parser.getFileName (del_sec, wb));\n\t\t}\n\t }\n\telse\n\t{\n\t ycp2error (\"Wrong value for path %s: %s\", path->toString ().c_str (), value->toString ().c_str ());\n\t b = false;\n\t}\n }\n if (!ok)\n {\n \tycp2error ( \"Wrong path '%s' in Write().\", path->toString().c_str () );\n }\n\n return b;\n}", "label_name": "Class", "label": 2.0}
{"code": " IniSection (const IniSection &s) :\n\tIniBase (s),\n\t ip (s.ip),\n\t end_comment (s.end_comment), rewrite_by (s.rewrite_by),\n\t container (s.container)\n\t{ reindex (); }", "label_name": "Class", "label": 2.0}
@@ -1681,8 +1654,6 @@
{"code": "void AllocateDataSet(cmsIT8* it8)\n{\n TABLE* t = GetTable(it8);\n\n if (t -> Data) return; // Already allocated\n\n t-> nSamples = atoi(cmsIT8GetProperty(it8, \"NUMBER_OF_FIELDS\"));\n t-> nPatches = atoi(cmsIT8GetProperty(it8, \"NUMBER_OF_SETS\"));\n\n t-> Data = (char**)AllocChunk (it8, ((cmsUInt32Number) t->nSamples + 1) * ((cmsUInt32Number) t->nPatches + 1) *sizeof (char*));\n if (t->Data == NULL) {\n\n SynError(it8, \"AllocateDataSet: Unable to allocate data array\");\n }\n\n}", "label_name": "Base", "label": 1.0}
{"code": "void AllocateDataSet(cmsIT8* it8)\n{\n TABLE* t = GetTable(it8);\n\n if (t -> Data) return; // Already allocated\n\n t-> nSamples = atoi(cmsIT8GetProperty(it8, \"NUMBER_OF_FIELDS\"));\n t-> nPatches = atoi(cmsIT8GetProperty(it8, \"NUMBER_OF_SETS\"));\n\n t-> Data = (char**)AllocChunk (it8, ((cmsUInt32Number) t->nSamples + 1) * ((cmsUInt32Number) t->nPatches + 1) *sizeof (char*));\n if (t->Data == NULL) {\n\n SynError(it8, \"AllocateDataSet: Unable to allocate data array\");\n }\n\n}", "label_name": "Base", "label": 1.0}
{"code": "BGD_DECLARE(void *) gdImageWBMPPtr(gdImagePtr im, int *size, int fg)\n{\n\tvoid *rv;\n\tgdIOCtx *out = gdNewDynamicCtx(2048, NULL);\n\tif (out == NULL) return NULL;\n\tgdImageWBMPCtx(im, fg, out);\n\trv = gdDPExtractData(out, size);\n\tout->gd_free(out);\n\treturn rv;\n}", "label_name": "Variant", "label": 0.0}
-{"code": "Variant HHVM_FUNCTION(apc_store,\n const Variant& key_or_array,\n const Variant& var /* = null */,\n int64_t ttl /* = 0 */) {\n if (!apcExtension::Enable) return Variant(false);\n\n if (key_or_array.isArray()) {\n Array valuesArr = key_or_array.toArray();\n\n for (ArrayIter iter(valuesArr); iter; ++iter) {\n Variant key = iter.first();\n if (!key.isString()) {\n throw_invalid_argument(\"apc key: (not a string)\");\n return Variant(false);\n }\n Variant v = iter.second();\n apc_store().set(key.toString(), v, ttl);\n }\n\n return Variant(ArrayData::Create());\n }\n\n if (!key_or_array.isString()) {\n throw_invalid_argument(\"apc key: (not a string)\");\n return Variant(false);\n }\n String strKey = key_or_array.toString();\n apc_store().set(strKey, var, ttl);\n return Variant(true);\n}", "label_name": NaN, "label": NaN}
-{"code": "Variant HHVM_FUNCTION(apc_add,\n const Variant& key_or_array,\n const Variant& var /* = null */,\n int64_t ttl /* = 0 */) {\n if (!apcExtension::Enable) return false;\n\n if (key_or_array.isArray()) {\n Array valuesArr = key_or_array.toArray();\n\n // errors stores all keys corresponding to entries that could not be cached\n ArrayInit errors(valuesArr.size(), ArrayInit::Map{});\n\n for (ArrayIter iter(valuesArr); iter; ++iter) {\n Variant key = iter.first();\n if (!key.isString()) {\n throw_invalid_argument(\"apc key: (not a string)\");\n return false;\n }\n Variant v = iter.second();\n if (!apc_store().add(key.toString(), v, ttl)) {\n errors.add(key, -1);\n }\n }\n return errors.toVariant();\n }\n\n if (!key_or_array.isString()) {\n throw_invalid_argument(\"apc key: (not a string)\");\n return false;\n }\n String strKey = key_or_array.toString();\n return apc_store().add(strKey, var, ttl);\n}", "label_name": NaN, "label": NaN}
{"code": "void PCRECache::dump(const std::string& filename) {\n std::ofstream out(filename.c_str());\n switch (m_kind) {\n case CacheKind::Static:\n for (auto& it : *m_staticCache) {\n out << it.first->data() << \"\\n\";\n }\n break;\n case CacheKind::Lru:\n case CacheKind::Scalable:\n {\n std::vector keys;\n if (m_kind == CacheKind::Lru) {\n m_lruCache->snapshotKeys(keys);\n } else {\n m_scalableCache->snapshotKeys(keys);\n }\n for (auto& key: keys) {\n out << key.c_str() << \"\\n\";\n }\n }\n break;\n }\n out.close();\n}", "label_name": "Base", "label": 1.0}
{"code": "String StringUtil::Crypt(const String& input, const char *salt /* = \"\" */) {\n if (salt && salt[0] == '\\0') {\n raise_notice(\"crypt(): No salt parameter was specified.\"\n \" You must use a randomly generated salt and a strong\"\n \" hash function to produce a secure hash.\");\n }\n return String(string_crypt(input.c_str(), salt), AttachString);\n}", "label_name": "Base", "label": 1.0}
{"code": "inline int StringData::size() const { return m_len; }", "label_name": "Base", "label": 1.0}
@@ -1726,7 +1697,6 @@
{"code": "void RemoteFsDevice::serviceAdded(const QString &name)\n{\n if (name==details.serviceName && constSambaAvahiProtocol==details.url.scheme()) {\n sub=tr(\"Available\");\n updateStatus();\n }\n}", "label_name": "Class", "label": 2.0}
{"code": "static inline bool isMountable(const RemoteFsDevice::Details &d)\n{\n return RemoteFsDevice::constSshfsProtocol==d.url.scheme() ||\n RemoteFsDevice::constSambaProtocol==d.url.scheme() || RemoteFsDevice::constSambaAvahiProtocol==d.url.scheme();\n}", "label_name": "Class", "label": 2.0}
{"code": "void RemoteFsDevice::load()\n{\n if (RemoteFsDevice::constSambaAvahiProtocol==details.url.scheme()) {\n // Start Avahi listener...\n Avahi::self();\n QUrlQuery q(details.url);\n if (q.hasQueryItem(constServiceNameQuery)) {\n details.serviceName=q.queryItemValue(constServiceNameQuery);\n }\n\n if (!details.serviceName.isEmpty()) {\n AvahiService *srv=Avahi::self()->getService(details.serviceName);\n if (!srv || srv->getHost().isEmpty()) {\n sub=tr(\"Not Available\");\n } else {\n sub=tr(\"Available\");\n }\n }\n connect(Avahi::self(), SIGNAL(serviceAdded(QString)), SLOT(serviceAdded(QString)));\n connect(Avahi::self(), SIGNAL(serviceRemoved(QString)), SLOT(serviceRemoved(QString)));\n }\n\n if (isConnected()) {\n setAudioFolder();\n readOpts(settingsFileName(), opts, true);\n rescan(false); // Read from cache if we have it!\n }\n}", "label_name": "Class", "label": 2.0}
-{"code": "void Rectangle(double x,double y,double w,double h) {\n outpos += sprintf(outpos,\"\\n %12.3f %12.3f %12.3f %12.3f re\",x,y,w,h);\n}", "label_name": NaN, "label": NaN}
{"code": "void ResourceHandle::FromProto(const ResourceHandleProto& proto) {\n set_device(proto.device());\n set_container(proto.container());\n set_name(proto.name());\n set_hash_code(proto.hash_code());\n set_maybe_type_name(proto.maybe_type_name());\n std::vector dtypes_and_shapes;\n for (const auto& dtype_and_shape : proto.dtypes_and_shapes()) {\n DataType dtype = dtype_and_shape.dtype();\n PartialTensorShape shape(dtype_and_shape.shape());\n dtypes_and_shapes.push_back(DtypeAndPartialTensorShape{dtype, shape});\n }\n dtypes_and_shapes_ = std::move(dtypes_and_shapes);\n}", "label_name": "Base", "label": 1.0}
{"code": "StatusOr SpecializeType(const AttrSlice& attrs,\n const OpDef& op_def) {\n FullTypeDef ft;\n ft.set_type_id(TFT_PRODUCT);\n\n for (int i = 0; i < op_def.output_arg_size(); i++) {\n auto* t = ft.add_args();\n\n *t = op_def.output_arg(i).experimental_full_type();\n\n // Resolve dependent types. The convention for op registrations is to use\n // attributes as type variables.\n // See https://www.tensorflow.org/guide/create_op#type_polymorphism.\n // Once the op signature can be defined entirely in FullType, this\n // convention can be deprecated.\n //\n // Note: While this code performs some basic verifications, it generally\n // assumes consistent op defs and attributes. If more complete\n // verifications are needed, they should be done by separately, and in a\n // way that can be reused for type inference.\n for (int j = 0; j < t->args_size(); j++) {\n auto* arg = t->mutable_args(i);\n if (arg->type_id() == TFT_VAR) {\n const auto* attr = attrs.Find(arg->s());\n DCHECK(attr != nullptr);\n if (attr->value_case() == AttrValue::kList) {\n const auto& attr_list = attr->list();\n arg->set_type_id(TFT_PRODUCT);\n for (int i = 0; i < attr_list.type_size(); i++) {\n map_dtype_to_tensor(attr_list.type(i), arg->add_args());\n }\n\n } else if (attr->value_case() == AttrValue::kType) {\n map_dtype_to_tensor(attr->type(), arg);\n\n } else {\n return Status(error::UNIMPLEMENTED,\n absl::StrCat(\"unknown attribute type\",\n attrs.DebugString(), \" key=\", arg->s()));\n }\n\n arg->clear_s();\n }\n }\n }\n\n return ft;\n}", "label_name": "Base", "label": 1.0}
{"code": "Status OpLevelCostEstimator::PredictFusedBatchNorm(\n const OpContext& op_context, NodeCosts* node_costs) const {\n bool found_unknown_shapes = false;\n const auto& op_info = op_context.op_info;\n // x: op_info.inputs(0)\n // scale: op_info.inputs(1)\n // offset: op_info.inputs(2)\n // mean: op_info.inputs(3) --> only for inference\n // variance: op_info.inputs(4) --> only for inference\n ConvolutionDimensions dims = OpDimensionsFromInputs(\n op_info.inputs(0).shape(), op_info, &found_unknown_shapes);\n const bool is_training = IsTraining(op_info);\n\n int64_t ops = 0;\n const auto rsqrt_cost = Eigen::internal::functor_traits<\n Eigen::internal::scalar_rsqrt_op>::Cost;\n if (is_training) {\n ops = dims.iz * (dims.batch * dims.ix * dims.iy * 4 + 6 + rsqrt_cost);\n } else {\n ops = dims.batch * dims.ix * dims.iy * dims.iz * 2;\n }\n node_costs->num_compute_ops = ops;\n\n const int64_t size_nhwc =\n CalculateTensorSize(op_info.inputs(0), &found_unknown_shapes);\n const int64_t size_c =\n CalculateTensorSize(op_info.inputs(1), &found_unknown_shapes);\n if (is_training) {\n node_costs->num_input_bytes_accessed = {size_nhwc, size_c, size_c};\n node_costs->num_output_bytes_accessed = {size_nhwc, size_c, size_c, size_c,\n size_c};\n // FusedBatchNorm in training mode internally re-reads the input tensor:\n // one for mean/variance, and the 2nd internal read forthe actual scaling.\n // Assume small intermediate data such as mean / variance (size_c) can be\n // cached on-chip.\n node_costs->internal_read_bytes = size_nhwc;\n } else {\n node_costs->num_input_bytes_accessed = {size_nhwc, size_c, size_c, size_c,\n size_c};\n node_costs->num_output_bytes_accessed = {size_nhwc};\n }\n node_costs->max_memory = node_costs->num_total_output_bytes();\n\n if (found_unknown_shapes) {\n node_costs->inaccurate = true;\n node_costs->num_nodes_with_unknown_shapes = 1;\n }\n return Status::OK();\n}", "label_name": "Base", "label": 1.0}
@@ -1818,7 +1788,6 @@
{"code": " def self.generate_key_pair(key_base_name, recipient, real_name)\n public_key_file_name = \"#{key_base_name}.pub\"\n private_key_file_name = \"#{key_base_name}.sec\"\n script = generate_key_script(public_key_file_name, private_key_file_name, recipient, real_name)\n script_file = Tempfile.new('gpg-script')\n begin\n script_file.write(script)\n script_file.close\n result = system(\"gpg --batch --gen-key #{script_file.path}\")\n raise RuntimeError.new('gpg failed') unless result\n ensure\n script_file.close\n script_file.unlink\n end\n end", "label_name": "Base", "label": 1.0}
{"code": " it \"should find a user by first name or last name\" do\n @cur_user.stub(:pref).and_return(:activity_user => 'Billy')\n controller.instance_variable_set(:@current_user, @cur_user)\n User.should_receive(:where).with(\"upper(first_name) LIKE upper('%Billy%') OR upper(last_name) LIKE upper('%Billy%')\").and_return([@user])\n controller.send(:activity_user).should == 1\n end", "label_name": "Base", "label": 1.0}
{"code": " def tmp_dir\n ENV['TMPDIR'] || ENV['TEMP'] || '/tmp'\n end", "label_name": NaN, "label": NaN}
-{"code": " def cat(path, identifier=nil)\n p = CGI.escape(scm_iconv(@path_encoding, 'UTF-8', path))\n hg 'rhcat', '-r', CGI.escape(hgrev(identifier)), hgtarget(p) do |io|\n io.binmode\n io.read\n end\n rescue HgCommandAborted\n nil # means not found\n end", "label_name": NaN, "label": NaN}
{"code": " def schedule_all\n Log.add_info(request, params.inspect)\n\n date_s = params[:date]\n if date_s.nil? or date_s.empty?\n @date = Date.today\n else\n @date = Date.parse(date_s)\n end\n\n if @login_user.nil? or params[:display].nil? or params[:display] == 'all'\n params[:display] = 'all'\n con = EquipmentHelper.get_scope_condition_for(@login_user)\n else\n display_type = params[:display].split('_').first\n display_id = params[:display].split('_').last\n\n case display_type\n when 'group'\n if @login_user.get_groups_a(true).include?(display_id)\n con = ApplicationHelper.get_sql_like([:groups], \"|#{display_id}|\")\n end\n when 'team'\n if @login_user.get_teams_a.include?(display_id)\n con = ApplicationHelper.get_sql_like([:teams], \"|#{display_id}|\")\n end\n end\n end", "label_name": "Base", "label": 1.0}
{"code": " def save_page\n Log.add_info(request, params.inspect)\n\n # Next page\n pave_val = params[:page].to_i + 1\n @page = sprintf('%02d', pave_val)\n\n page_num = Dir.glob(File.join(Research.page_dir, \"_q[0-9][0-9].html.erb\")).length\n\n unless params[:research].nil?\n params[:research].each do |key, value|\n if value.instance_of?(Array)\n value.compact!\n value.delete('')\n if value.empty?\n params[:research][key] = nil\n else\n params[:research][key] = value.join(\"\\n\") + \"\\n\"\n end\n end\n end\n end\n\n if params[:research_id].nil? or params[:research_id].empty?\n @research = Research.new(params.require(:research).permit(Research::PERMIT_BASE))\n @research.status = Research::U_STATUS_IN_ACTON\n @research.update_attribute(:user_id, @login_user.id)\n else\n @research = Research.find(params[:research_id])\n @research.update_attributes(params.require(:research).permit(Research::PERMIT_BASE))\n end\n\n if pave_val <= page_num\n\n render(:action => 'edit_page')\n\n else\n\n tmpl_folder, tmpl_q_folder = TemplatesHelper.get_tmpl_subfolder(TemplatesHelper::TMPL_RESEARCH)\n\n if tmpl_q_folder.nil?\n ary = TemplatesHelper.setup_tmpl_folder\n tmpl_q_folder = ary[4]\n end\n\n items = Folder.get_items_admin(tmpl_q_folder.id, 'xorder ASC')\n\n @q_caps_h = {}\n unless items.nil?\n items.each do |item|\n\n desc = item.description\n next if desc.nil? or desc.empty?\n\n hash = Research.select_q_caps(desc)\n hash.each do |key, val|\n @q_caps_h[key] = val\n end\n end\n end\n\n render(:action => 'confirm')\n end\n\n rescue => evar\n Log.add_error(request, evar)\n @page = '01'\n render(:action => 'edit_page')\n end", "label_name": "Base", "label": 1.0}
{"code": " def show_tree\n if params[:action] == 'show_tree'\n Log.add_info(request, params.inspect)\n end\n\n con = []\n con << \"(user_id=#{@login_user.id})\"\n\n account_xtype = params[:mail_account_xtype]\n\n unless account_xtype.blank?\n SqlHelper.validate_token([account_xtype])\n con << \"(xtype='#{account_xtype}')\"\n end\n @mail_accounts = MailAccount.find_all(con.join(' and '))\n\n mail_account_ids = []\n @mail_accounts.each do |mail_account|\n\n mail_account_ids << mail_account.id\n\n if MailFolder.where(\"mail_account_id=#{mail_account.id}\").count <= 0\n @login_user.create_default_mail_folders(mail_account.id)\n end\n\n Email.destroy_by_user(@login_user.id, \"status='#{Email::STATUS_TEMPORARY}'\")\n end\n\n @folder_tree = MailFolder.get_tree_for(@login_user, mail_account_ids)\n end", "label_name": "Base", "label": 1.0}
@@ -1890,10 +1859,7 @@
{"code": " it \"should be created correctly\" do\n invite = Fabricate(:invite, email: 'walter.white@email.com')\n user = InviteRedeemer.create_user_from_invite(invite: invite, email: invite.email, username: 'walter', name: 'Walter White')\n expect(user.username).to eq('walter')\n expect(user.name).to eq('Walter White')\n expect(user.email).to eq('walter.white@email.com')\n expect(user.approved).to eq(true)\n expect(user.active).to eq(false)\n end", "label_name": "Class", "label": 2.0}
{"code": " def email_login\n raise Discourse::NotFound if !SiteSetting.enable_local_logins_via_email\n second_factor_token = params[:second_factor_token]\n second_factor_method = params[:second_factor_method].to_i\n token = params[:token]\n valid_token = !!EmailToken.valid_token_format?(token)\n user = EmailToken.confirmable(token)&.user\n\n if valid_token && user&.totp_enabled?\n if !second_factor_token.present?\n @second_factor_required = true\n @backup_codes_enabled = true if user&.backup_codes_enabled?\n return render layout: 'no_ember'\n elsif !user.authenticate_second_factor(second_factor_token, second_factor_method)\n RateLimiter.new(nil, \"second-factor-min-#{request.remote_ip}\", 3, 1.minute).performed!\n @error = I18n.t('login.invalid_second_factor_code')\n return render layout: 'no_ember'\n end\n end\n\n if user = EmailToken.confirm(token)\n if login_not_approved_for?(user)\n @error = login_not_approved[:error]\n elsif payload = login_error_check(user)\n @error = payload[:error]\n else\n log_on_user(user)\n return redirect_to path(\"/\")\n end\n else\n @error = I18n.t('email_login.invalid_token')\n end\n\n render layout: 'no_ember'\n end", "label_name": "Class", "label": 2.0}
{"code": " it 'does not log in with incorrect backup code' do\n post \"/session/email-login/#{email_token.token}\", params: {\n second_factor_token: \"0000\",\n second_factor_method: UserSecondFactor.methods[:backup_codes]\n }\n\n expect(response.status).to eq(200)\n expect(CGI.unescapeHTML(response.body)).to include(I18n.t(\n \"login.invalid_second_factor_code\"\n ))\n end", "label_name": "Class", "label": 2.0}
-{"code": " it 'returns the right response' do\n get \"/session/email-login/adasdad\"\n\n expect(response.status).to eq(200)\n\n expect(CGI.unescapeHTML(response.body)).to match(\n I18n.t('email_login.invalid_token')\n )\n end", "label_name": NaN, "label": NaN}
{"code": " it 'logs in correctly' do\n post \"/session/email-login/#{email_token.token}\", params: {\n second_factor_token: ROTP::TOTP.new(user_second_factor.data).now,\n second_factor_method: UserSecondFactor.methods[:totp]\n }\n\n expect(response).to redirect_to(\"/\")\n end", "label_name": "Class", "label": 2.0}
-{"code": " it 'returns success' do\n get \"/session/email-login/#{email_token.token}\"\n\n expect(response).to redirect_to(\"/\")\n end", "label_name": NaN, "label": NaN}
-{"code": " it 'does not log in with incorrect two factor' do\n post \"/session/email-login/#{email_token.token}\", params: {\n second_factor_token: \"0000\",\n second_factor_method: UserSecondFactor.methods[:totp]\n }\n\n expect(response.status).to eq(200)\n\n expect(CGI.unescapeHTML(response.body)).to include(I18n.t(\n \"login.invalid_second_factor_code\"\n ))\n end", "label_name": NaN, "label": NaN}
{"code": " it 'does not log in with incorrect two factor' do\n post \"/session/email-login/#{email_token.token}\", params: {\n second_factor_token: \"0000\",\n second_factor_method: UserSecondFactor.methods[:totp]\n }\n\n expect(response.status).to eq(200)\n\n expect(CGI.unescapeHTML(response.body)).to include(I18n.t(\n \"login.invalid_second_factor_code\"\n ))\n end", "label_name": "Class", "label": 2.0}
{"code": " it \"fails when user is suspended\" do\n user.update!(\n suspended_till: 2.days.from_now,\n suspended_at: Time.zone.now\n )\n\n get \"/session/email-login/#{email_token.token}\"\n\n expect(response.status).to eq(200)\n\n expect(CGI.unescapeHTML(response.body)).to include(I18n.t(\"login.suspended\",\n date: I18n.l(user.suspended_till, format: :date_only)\n ))\n end", "label_name": "Class", "label": 2.0}
{"code": " def show\n respond_to do |format|\n format.html\n format.json { render json: @application }\n end\n end", "label_name": "Class", "label": 2.0}
@@ -1913,8 +1879,6 @@
{"code": "func (cs chunkState) String() string {\n\tstateString := \"\"\n\tswitch cs {\n\tcase readChunkHeader:\n\t\tstateString = \"readChunkHeader\"\n\tcase readChunkTrailer:\n\t\tstateString = \"readChunkTrailer\"\n\tcase readChunk:\n\t\tstateString = \"readChunk\"\n\tcase verifyChunk:\n\t\tstateString = \"verifyChunk\"\n\tcase eofChunk:\n\t\tstateString = \"eofChunk\"\n\n\t}\n\treturn stateString\n}", "label_name": "Base", "label": 1.0}
{"code": "func (rarFormat) Read(input io.Reader, destination string) error {\n\trr, err := rardecode.NewReader(input, \"\")\n\tif err != nil {\n\t\treturn fmt.Errorf(\"read: failed to create reader: %v\", err)\n\t}\n\n\tfor {\n\t\theader, err := rr.Next()\n\t\tif err == io.EOF {\n\t\t\tbreak\n\t\t} else if err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif header.IsDir {\n\t\t\terr = mkdir(filepath.Join(destination, header.Name))\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\t// if files come before their containing folders, then we must\n\t\t// create their folders before writing the file\n\t\terr = mkdir(filepath.Dir(filepath.Join(destination, header.Name)))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\terr = writeNewFile(filepath.Join(destination, header.Name), rr, header.Mode())\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "label_name": "Base", "label": 1.0}
{"code": "func Test_isRepositoryGitPath(t *testing.T) {\n\ttests := []struct {\n\t\tpath string\n\t\twantVal bool\n\t}{\n\t\t{path: filepath.Join(\".\", \".git\"), wantVal: true},\n\t\t{path: filepath.Join(\".\", \".git\", \"\"), wantVal: true},\n\t\t{path: filepath.Join(\".\", \".git\", \"hooks\", \"pre-commit\"), wantVal: true},\n\t\t{path: filepath.Join(\".git\", \"hooks\"), wantVal: true},\n\t\t{path: filepath.Join(\"dir\", \".git\"), wantVal: true},\n\n\t\t{path: filepath.Join(\".\", \".git.\"), wantVal: true},\n\t\t{path: filepath.Join(\".\", \".git.\", \"\"), wantVal: true},\n\t\t{path: filepath.Join(\".\", \".git.\", \"hooks\", \"pre-commit\"), wantVal: true},\n\t\t{path: filepath.Join(\".git.\", \"hooks\"), wantVal: true},\n\t\t{path: filepath.Join(\"dir\", \".git.\"), wantVal: true},\n\n\t\t{path: filepath.Join(\".gitignore\"), wantVal: false},\n\t\t{path: filepath.Join(\"dir\", \".gitkeep\"), wantVal: false},\n\t}\n\tfor _, test := range tests {\n\t\tt.Run(\"\", func(t *testing.T) {\n\t\t\tassert.Equal(t, test.wantVal, isRepositoryGitPath(test.path))\n\t\t})\n\t}\n}", "label_name": "Base", "label": 1.0}
-{"code": "func NewHandle(pamh unsafe.Pointer) (*Handle, error) {\n\tvar err error\n\th := &Handle{\n\t\thandle: (*C.pam_handle_t)(pamh),\n\t\tstatus: C.PAM_SUCCESS,\n\t}\n\n\tvar pamUsername *C.char\n\th.status = C.pam_get_user(h.handle, &pamUsername, nil)\n\tif err = h.err(); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif h.PamUser, err = user.Lookup(C.GoString(pamUsername)); err != nil {\n\t\treturn nil, err\n\t}\n\tif h.OrigUser, err = util.EffectiveUser(); err != nil {\n\t\treturn nil, err\n\t}\n\treturn h, nil\n}", "label_name": NaN, "label": NaN}
-{"code": "func (h *Handle) StartAsPamUser() error {\n\tif _, err := security.UserKeyringID(h.PamUser, true); err != nil {\n\t\tlog.Printf(\"Setting up keyrings in PAM: %v\", err)\n\t}\n\treturn security.SetProcessPrivileges(h.PamUser)\n}", "label_name": NaN, "label": NaN}
{"code": "func (p *BinaryProtocol) ReadBinary() ([]byte, error) {\n\tsize, e := p.ReadI32()\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\tif size < 0 {\n\t\treturn nil, invalidDataLength\n\t}\n\tif uint64(size) > p.trans.RemainingBytes() {\n\t\treturn nil, invalidDataLength\n\t}\n\n\tisize := int(size)\n\tbuf := make([]byte, isize)\n\t_, err := io.ReadFull(p.trans, buf)\n\treturn buf, NewProtocolException(err)\n}", "label_name": "Base", "label": 1.0}
{"code": "func TestSkipUnknownTypeBinaryProtocol(t *testing.T) {\n\tvar m MyTestStruct\n\td := NewDeserializer()\n\tf := NewBinaryProtocolFactoryDefault()\n\td.Protocol = f.GetProtocol(d.Transport)\n\t// skip over a map with invalid key/value type and 1.7B entries\n\tdata := []byte(\"\\n\\x10\\rO\\t6\\x03\\n\\n\\n\\x10\\r\\n\\tslice\\x00\")\n\tstart := time.Now()\n\terr := d.Read(&m, data)\n\tif err == nil {\n\t\tt.Fatalf(\"Parsed invalid message correctly\")\n\t} else if !strings.Contains(err.Error(), \"unknown type\") {\n\t\tt.Fatalf(\"Failed for reason besides unknown type\")\n\t}\n\n\tif time.Now().Sub(start).Seconds() > 5 {\n\t\tt.Fatalf(\"It should not take seconds to parse a small message\")\n\t}\n}", "label_name": "Base", "label": 1.0}
{"code": "func (voteSet *VoteSet) MakeCommit() *Commit {\n\tif voteSet.signedMsgType != tmproto.PrecommitType {\n\t\tpanic(\"Cannot MakeCommit() unless VoteSet.Type is PrecommitType\")\n\t}\n\tvoteSet.mtx.Lock()\n\tdefer voteSet.mtx.Unlock()\n\n\t// Make sure we have a 2/3 majority\n\tif voteSet.maj23 == nil {\n\t\tpanic(\"Cannot MakeCommit() unless a blockhash has +2/3\")\n\t}\n\n\t// For every validator, get the precommit\n\tcommitSigs := make([]CommitSig, len(voteSet.votes))\n\tfor i, v := range voteSet.votes {\n\t\tcommitSigs[i] = v.CommitSig()\n\t}\n\n\treturn NewCommit(voteSet.GetHeight(), voteSet.GetRound(), *voteSet.maj23, commitSigs)\n}", "label_name": "Base", "label": 1.0}
@@ -1956,8 +1920,6 @@
{"code": "func (AppModuleBasic) ConsensusVersion() uint64 {\n\treturn 1\n}", "label_name": "Class", "label": 2.0}
{"code": "func TestBuilder_BuildBootstrapLayeredRuntime(t *testing.T) {\n\tb := New(\"localhost:1111\", \"localhost:2222\", filemgr.NewManager(), nil)\n\tstaticCfg, err := b.BuildBootstrapLayeredRuntime()\n\tassert.NoError(t, err)\n\ttestutil.AssertProtoJSONEqual(t, `\n\t\t{ \"layers\": [{\n\t\t\t\"name\": \"static_layer_0\",\n\t\t\t\"staticLayer\": {\n\t\t\t\t\"overload\": {\n\t\t\t\t\t\"global_downstream_max_connections\": 50000\n\t\t\t\t}\n\t\t\t}\n\t\t}] }\n\t`, staticCfg)\n}", "label_name": "Class", "label": 2.0}
{"code": "func main() {\n\thostPtr := flag.String(\"host\", \"localhost:9159\", \"the hostname that the server should listen on.\")\n\ttokenPtr := flag.String(\"token\", \"\", \"the Proxy Access Token used to restrict access to the server.\")\n\tallowedOriginsPtr := flag.String(\"allowed-origins\", \"*\", \"a comma separated list of allowed origins.\")\n\tbannedOutputsPtr := flag.String(\"banned-outputs\", \"\", \"a comma separated list of banned outputs.\")\n\tflag.Parse()\n\n\tfinished := make(chan bool)\n\tlibproxy.Initialize(*tokenPtr, *hostPtr, *allowedOriginsPtr, *bannedOutputsPtr, onProxyStateChangeServer, false, finished)\n\n\t<-finished\n}", "label_name": "Base", "label": 1.0}
-{"code": "func CreatePayloadFromItems(items []types.MessageItem, title string, colors [types.MessageLevelCount]uint, omitted int) (WebhookPayload, error) {\n\n\tmetaCount := 1\n\tif omitted < 1 && len(title) < 1 {\n\t\tmetaCount = 0\n\t}\n\titemCount := util.Min(9, len(items))\n\n\tembeds := make([]embedItem, metaCount, itemCount+metaCount)\n\n\tfor _, item := range items {\n\n\t\tcolor := uint(0)\n\t\tif item.Level >= types.Unknown && int(item.Level) < len(colors) {\n\t\t\tcolor = colors[item.Level]\n\t\t}\n\n\t\tei := embedItem{\n\t\t\tContent: item.Text,\n\t\t\tColor: color,\n\t\t}\n\n\t\tif item.Level != types.Unknown {\n\t\t\tei.Footer = &embedFooter{\n\t\t\t\tText: item.Level.String(),\n\t\t\t}\n\t\t}\n\n\t\tif !item.Timestamp.IsZero() {\n\t\t\tei.Timestamp = item.Timestamp.UTC().Format(time.RFC3339)\n\t\t}\n\n\t\tembeds = append(embeds, ei)\n\t}\n\n\tembeds[0].Title = title\n\tif omitted > 0 {\n\t\tembeds[0].Footer = &embedFooter{\n\t\t\tText: fmt.Sprintf(\"... (%v character(s) where omitted)\", omitted),\n\t\t}\n\t}\n\n\treturn WebhookPayload{\n\t\tEmbeds: embeds,\n\t}, nil\n}", "label_name": NaN, "label": NaN}
-{"code": "func PartitionMessage(input string, limits t.MessageLimit, distance int) (items []t.MessageItem, omitted int) {\n\trunes := []rune(input)\n\tchunkOffset := 0\n\tmaxTotal := Min(len(runes), limits.TotalChunkSize)\n\tmaxCount := limits.ChunkCount - 1\n\n\tfor i := 0; i < maxCount; i++ {\n\t\t// If no suitable split point is found, use the chunkSize\n\t\tchunkEnd := chunkOffset + limits.ChunkSize\n\t\t// ... and start next chunk directly after this one\n\t\tnextChunkStart := chunkEnd\n\t\tif chunkEnd > maxTotal {\n\t\t\t// The chunk is smaller than the limit, no need to search\n\t\t\tchunkEnd = maxTotal\n\t\t\tnextChunkStart = maxTotal\n\t\t} else {\n\t\t\tfor r := 0; r < distance; r++ {\n\t\t\t\trp := chunkEnd - r\n\t\t\t\tif runes[rp] == '\\n' || runes[rp] == ' ' {\n\t\t\t\t\t// Suitable split point found\n\t\t\t\t\tchunkEnd = rp\n\t\t\t\t\t// Since the split is on a whitespace, skip it in the next chunk\n\t\t\t\t\tnextChunkStart = chunkEnd + 1\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\titems = append(items, t.MessageItem{\n\t\t\tText: string(runes[chunkOffset:chunkEnd]),\n\t\t})\n\n\t\tchunkOffset = nextChunkStart\n\t\tif chunkOffset >= maxTotal {\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn items, len(runes) - chunkOffset\n}", "label_name": NaN, "label": NaN}
{"code": "func cmdList(args *docopt.Args, client *tuf.Client) error {\n\tif _, err := client.Update(); err != nil && !tuf.IsLatestSnapshot(err) {\n\t\treturn err\n\t}\n\ttargets, err := client.Targets()\n\tif err != nil {\n\t\treturn err\n\t}\n\tw := tabwriter.NewWriter(os.Stdout, 1, 2, 2, ' ', 0)\n\tdefer w.Flush()\n\tfmt.Fprintln(w, \"PATH\\tSIZE\")\n\tfor path, meta := range targets {\n\t\tfmt.Fprintf(w, \"%s\\t%s\\n\", path, humanize.Bytes(uint64(meta.Length)))\n\t}\n\treturn nil\n}", "label_name": "Base", "label": 1.0}
{"code": " function buildBinary(isDownloaded) \n {\n var buildString = \"node-gyp configure build --IBM_DB_HOME=\\\"$IBM_DB_HOME\\\"\";\n if(isDownloaded) {\n buildString = buildString + \" --IS_DOWNLOADED=true\";\n } else {\n buildString = buildString + \" --IS_DOWNLOADED=false\";\n }\n if( platform == 'win32') \n {\n buildString = buildString + \" --IBM_DB_HOME_WIN=%IBM_DB_HOME%\";\n }\n var childProcess = exec(buildString, function (error, stdout, stderr) {\n console.log(stdout);\n if (error !== null) {\n console.log(error);\n process.exit(1);\n }\n\n if(platform == 'darwin' && arch == 'x64') \n {\n // Run the install_name_tool\n var nameToolCommand = \"install_name_tool -change libdb2.dylib $IBM_DB_HOME/lib/libdb2.dylib ./build/Release/odbc_bindings.node\"\n var nameToolCmdProcess = exec(nameToolCommand , \n function (error1, stdout1, stderr1) {\n if (error1 !== null) {\n console.log('Error setting up the lib path to ' +\n 'odbc_bindings.node file.Error trace:\\n'+error1);\n process.exit(1);\n }\n });\n }\n removeUsedPackages();\n });\n } //buildBinary", "label_name": NaN, "label": NaN}
{"code": " componentDidMount () {\n this.store.gotoUrl(this.props.params.url);\n return this.store.generateToken();\n }", "label_name": "Class", "label": 2.0}
@@ -1965,15 +1927,12 @@
{"code": " let getNftParentId = async (tokenIdHex: string) => {\n let txnhex = (await asyncSlpValidator.getRawTransactions([tokenIdHex]))[0];\n let tx = Primatives.Transaction.parseFromBuffer(Buffer.from(txnhex, 'hex'));\n let nftBurnTxnHex = (await asyncSlpValidator.getRawTransactions([tx.inputs[0].previousTxHash]))[0];\n let nftBurnTxn = Primatives.Transaction.parseFromBuffer(Buffer.from(nftBurnTxnHex, 'hex'));\n let slp = new Slp(this.BITBOX);\n let nftBurnSlp = slp.parseSlpOutputScript(Buffer.from(nftBurnTxn.outputs[0].scriptPubKey));\n if (nftBurnSlp.transactionType === SlpTransactionType.GENESIS) {\n return tx.inputs[0].previousTxHash;\n }\n else {\n return nftBurnSlp.tokenIdHex;\n }\n }", "label_name": "Class", "label": 2.0}
{"code": " 'X-Parse-Session-Token': user1.getSessionToken(),\n },\n },\n });\n\n let foundGraphQLClassReadPreference = false;\n let foundUserClassReadPreference = false;\n databaseAdapter.database.serverConfig.cursor.calls\n .all()\n .forEach((call) => {\n if (\n call.args[0].ns.collection.indexOf('GraphQLClass') >= 0\n ) {\n foundGraphQLClassReadPreference = true;\n expect(call.args[0].options.readPreference.mode).toBe(\n ReadPreference.PRIMARY\n );\n } else if (\n call.args[0].ns.collection.indexOf('_User') >= 0\n ) {\n foundUserClassReadPreference = true;\n expect(call.args[0].options.readPreference.mode).toBe(\n ReadPreference.PRIMARY\n );\n }\n });\n\n expect(foundGraphQLClassReadPreference).toBe(true);\n expect(foundUserClassReadPreference).toBe(true);\n } catch (e) {\n handleError(e);\n }\n });", "label_name": "Class", "label": 2.0}
{"code": " ): Promise => {\n event.preventDefault();\n\n if (SingleSignOn.isSingleSignOnLoginWindow(frameName)) {\n return new SingleSignOn(main, event, url, options).init();\n }\n\n this.logger.log('Opening an external window from a webview.');\n return shell.openExternal(url);\n };", "label_name": "Class", "label": 2.0}
-{"code": "const mapPolls = () => {\n const poll = Polls.findOne({});\n if (!poll) {\n return { pollExists: false };\n }\n\n const { answers } = poll;\n let stackOptions = false;\n\n answers.map((obj) => {\n if (stackOptions) return obj;\n if (obj.key.length > MAX_CHAR_LENGTH) {\n stackOptions = true;\n }\n return obj;\n });\n\n const amIRequester = poll.requester !== 'userId';\n\n return {\n poll: {\n answers: poll.answers,\n pollId: poll.id,\n stackOptions,\n },\n pollExists: true,\n amIRequester,\n handleVote(pollId, answerId) {\n makeCall('publishVote', pollId, answerId.id);\n },\n };\n};", "label_name": NaN, "label": NaN}
{"code": " function addSockListeners() {\n if (!accept && !reject) {\n sock.once('connect', onconnect);\n sock.on('data', ondata);\n sock.once('error', onerror);\n sock.once('close', onclose);\n } else {\n var chan;\n sock.once('connect', function() {\n chan = accept();\n var isDone = false;\n function onDone() {\n if (isDone)\n return;\n sock.destroy();\n isDone = true;\n }\n chan.once('end', onDone)\n .once('close', onDone)\n .on('data', function(data) {\n sock.write(data);\n });\n sock.on('data', function(data) {\n chan.write(data);\n });\n });\n sock.once('close', function() {\n if (!chan)\n reject();\n });\n }\n }", "label_name": "Base", "label": 1.0}
-{"code": "function methodByPath(target, path) {\n path = pathToArray(path);\n\n const values = breadcrumbs(target, path);\n\n if (values.length < path.length) {\n return noop;\n }\n\n if (typeof values[values.length - 1] !== 'function') {\n return noop;\n }\n\n if (values.length > 1) {\n return values[values.length - 1].bind(values[values.length - 2]);\n }\n else {\n return values[0].bind(target);\n }\n}", "label_name": NaN, "label": NaN}
{"code": "\t\t([ key, value ]) => ({ [key]: exec.bind(null, value) }),\n\t),", "label_name": "Base", "label": 1.0}
{"code": "module.exports = async function(tag) {\n\tif (!tag || ![ 'string', 'number' ].includes(typeof tag)) {\n\t\tthrow new TypeError(`string was expected, instead got ${tag}`);\n\t}\n\n\tconst { message, author, email } = this;\n\n\tawait Promise.all([\n\t\texec(`git config user.name \"${await author}\"`),\n\t\texec(`git config user.email \"${await email}\"`),\n\t]);\n\tawait exec(`git tag -a ${JSON.stringify(tag)} -m \"${await message}\"`);\n\tawait exec(`git push origin ${JSON.stringify(`refs/tags/${tag}`)}`);\n};", "label_name": "Base", "label": 1.0}
{"code": " GraphQLPlayground.init(root, ${JSON.stringify(\n extendedOptions,\n null,\n 2,\n )})\n })", "label_name": "Base", "label": 1.0}
{"code": "const getCdnMarkup = ({ version, cdnUrl = '//cdn.jsdelivr.net/npm', faviconUrl }) => `\n {\r\n const getQueryString = queryString.stringify(this.getOptions(options));\r\n fetch(`${this.apiUrl}${endpoint}?${getQueryString}`)\r\n .then(response => EtsyClient._response(response, resolve, reject))\r\n .catch(reject);\r\n });\r\n }\r", "label_name": "Class", "label": 2.0}
-{"code": "\tconst pdf = await mdToPdf({ path: resolve(__dirname, 'mathjax', 'math.md') });\n\n\tt.is(pdf.filename, '');\n\tt.truthy(pdf.content);\n\n\tconst doc = await getDocument({ data: pdf.content }).promise;\n\tconst page = await doc.getPage(1);\n\tconst text = (await page.getTextContent()).items.map(({ str }) => str).join('');\n\n\tt.true(text.startsWith('Formulas with MathJax'));\n\tt.true(text.includes('a\u22600'));\n});", "label_name": NaN, "label": NaN}
{"code": "declare function getSetCookie(cookieName: string, value?: string, options?: any): string;\n\n\n// backw compat, later, do once per file instead (don't want a global 'r').\n//\n// ReactDOMFactories looks like:\n//\n// var ReactDOMFactories = {\n// a: createDOMFactory('a'),\n// abbr: ...\n//\n// function createDOMFactory(type) {\n// var factory = React.createElement.bind(null, type);\n// factory.type = type; // makes: `.type === Foo` work\n// return factory;\n// };\n//\n// and React.createElement(type: keyof ReactHTML, props, ...children) returns:\n// DetailedReactHTMLElement\n//\nconst r: { [elmName: string]: (props?: any, ...children) => RElm } = ReactDOMFactories;", "label_name": "Base", "label": 1.0}
{"code": " switchToInvites: async () => {\n await this.waitAndClick('.e_InvTabB');\n await this.invitedUsersList.waitUntilLoaded();\n },", "label_name": "Base", "label": 1.0}
{"code": "export function getExtensionPath(): string {\n\treturn extensionPath;\n}", "label_name": "Class", "label": 2.0}