ngram
listlengths
0
82k
[ "constant \"a\" as a float of 32 bits and #", "the constants \"a\" and \"b\", as \"total\", # i.e., total", "the Operative System Library as operative_system import os as operative_system", "Import the Operative System Library as operative_system import os as", "the constant \"b\" tensorflow.print(\"b = \",...
[ "generic test utils.\"\"\" from unittest.mock import patch import pytest @pytest.fixture(autouse=True)", "pytest @pytest.fixture(autouse=True) def patch_requests(): \"\"\"Stub out services that makes requests.\"\"\"", "services that makes requests.\"\"\" patch_client = patch(\"homeassistant.components.meteo_fra...
[ "'7077')}\") \\ if config is not None: for key, value", "from pyspark.sql import SparkSession from .log import logger @contextmanager def", "pre_spark = pre_spark.config(key, value) spark = pre_spark.getOrCreate() logger.info(\"Created Spark session\")", "pre_spark.config(key, value) spark = pre_spark.getOrCr...
[ "# get all Data for one owl # fill all", "'#FF69B4', '#CD5C5C', '#4B0082', '#FFFFF0', '#F0E68C', '#E6E6FA', '#FFF0F5', '#7CFC00', '#FFFACD', '#ADD8E6',", "= dist return months months = fillMonths(monthTable, months) X =", "month == \"09\": curOwl[8] = i[3] if month == \"10\":", "if month == \"10\": curOwl[9...
[ "allowed_domains = ['yensaophuyen.vn'] start_urls = ['http://yensaophuyen.vn/'] tracking_url = '' sitemap_urls", "\"//div[@class='nav_center']/span/a/span\", 'description' : \"//div[@class='intro']\", 'images' : \"//div[@class='ui-corner-all']/div/div/a/@href\", 'canonical' : \"\",", "\"//div[@class='intro']\",...
[ "list)): values = [values] return tf.train.Feature(int64_list=tf.train.Int64List(value=values)) def floats_feature(value): return tf.train.Feature(float_list=tf.train.FloatList(value=value))", "2.0 (the \"License\"); # you may not use this file", "image_to_tfexample(image_data, image_format, height, width, clas...
[ "- f == 0 # # The generator is: 2.", "DB0933D7 1E8C94E0 4A25619D CEE3D226 1AD2EE6B F12FFA06 D98A0864 D8760273 3EC86A64 521F2B18", "A637ED6B 0BFF5CB6 F406B7ED EE386BFB 5A899FA5 AE9F2411 7C4B1FE6 49286651 ECE45B3D C2007CB8", "# # This group is assigned id 14. # #", "} f = 2**6144 - 2**6080 - 1 + 2**64", "# ...
[ "= \"1.2.0\" setup( name=PROJECT_NAME, version=VERSION, packages=find_packages(), include_package_data=True, install_requires=[\"opencv-python\", \"pyyaml\", \"imutils\"],", "VERSION = \"1.2.0\" setup( name=PROJECT_NAME, version=VERSION, packages=find_packages(), include_package_data=True, install_requires=[\"ope...
[ "(open('yadashcomp/version.py').read()) setup( name='yadashcomp', version=__version__, author='pingf', packages=['yadashcomp'], include_package_data=True, license='MIT', description='yet another", "author='pingf', packages=['yadashcomp'], include_package_data=True, license='MIT', description='yet another dash com...
[ "= '%s/%s'%(dirName[len(path)+1:], fname) print(finame) res=s3.meta.client.upload_file(dirName + '/' + fname, bucket,", "import os camera='1' bucket='mikey.com-security' path='/mnt/cameraimages/images' s3 = boto3.resource('s3') for dirName,", "+ fname, bucket, finame) print res os.unlink(dirName + '/' +", "fo...
[ "in curr_dist_compl: if key in dist_compl: for mag_key in curr_dist_compl[key]:", "return upa_to_name def get_statistics(ids, GOLD, upa_name=None): ''' get statistics from", "upa_to_name = {list(query_results.keys())[0]:\"\"} currdir = os.path.dirname(__file__) gold_path = os.path.join(currdir,'data','GOLD-meta...
[ "2.88, 3.12, 3.36, 3.6, 3.84]) == 80 assert benchmark(gps, 14,", "2.1, 2.31, 2.52, 3.25]) == 219 assert benchmark(gps, 20, [0.0,", "assert benchmark(gps, 14, [0.0, 0.01, 0.36, 0.6, 0.84, 1.05, 1.26,", "import gps def test_gps(benchmark): assert benchmark(gps, 20, [0.0, 0.23, 0.46,", "3.6, 3.84]) == 80 asser...
[ "| :class:`thelper.gui.annotators.Annotator` \"\"\" if \"annotator\" not in config or not", "on the type contained in the config dictionary. The tool", "GUI elements. \"\"\" import logging import thelper.utils logger = logging.getLogger(__name__)", "annotation tool based on the type contained in the config", ...
[]
[ "['ip'] RESULT_KEY = 'origin' class GetUserAgentAction(HttpbinAction): METHOD = 'GET' URL_COMPONENTS", "'origin' class GetUserAgentAction(HttpbinAction): METHOD = 'GET' URL_COMPONENTS = ['user-agent'] RESULT_KEY", "URL_COMPONENTS = ['ip'] RESULT_KEY = 'origin' class GetUserAgentAction(HttpbinAction): METHOD =",...
[ "index, entry in enumerate({\"Tam\", \"Tom\", \"Tim\"}): print(f\"{index}: {entry}\") for index,", "for index, entry in enumerate({\"Tam\", \"Tom\", \"Tim\"}): print(f\"{index}: {entry}\") for", "\"Tim\", \"Tom\"}): print(f\"{index}: {entry}\") for index, entry in enumerate({\"Tam\", \"Tom\",", "entry in enum...
[ "---------- <NAME>., & <NAME>. (2018). Expected policy gradients. AAAI. \"\"\"", "rllib.algorithms.eac import ExpectedActorCritic from .actor_critic_agent import ActorCriticAgent class ExpectedActorCriticAgent(ActorCriticAgent): \"\"\"Implementation", "TODO: build compatible function approximation. References -...
[ "activation='relu6') x = self.net_blocks.bottleneck_v3(x, 80 , 3, e=240, s=2, alpha=alpha,", "x = self.net_blocks.bottleneck_v3(x, 80 , 3, e=240, s=2, alpha=alpha, squeeze=False,", "n_filters = self.net_blocks.make_divisible(c[0] * alpha, 8) x = self.net_blocks.conv_block(I, n_filters,", "3, e=240, s=2, alpha...
[ "\"magenta\": Back.MAGENTA, \"light magenta\": Back.LIGHTMAGENTA_EX, \"yellow\": Back.YELLOW, \"light yellow\": Back.LIGHTYELLOW_EX,", "Back.CYAN, \"light cyan\": Back.LIGHTCYAN_EX, \"white\": Back.WHITE} # Check the shade", "background colour to use if prompt_back_col in back_cols: prompt_back_col =", "the t...
[ "\" + values_by_keys(morse_dict, code.strip()) decoded += str(values_by_keys(morse_dict, code)) # the", "list, add a space to the # decoded message and", "letters # can be stored in the keys and code", "\"\"\" # The function retrieves the morse code from a", "code is not a morse symbol, so the values_by_key...
[ "intergalactic medium. Returns ------- Tuple containing the current time, redshift,", "self.pf['include_cgm']: data_cgm = self.parcel_cgm.grid.data.copy() # Evolve in time! while z", "Flip to descending order (in redshift) z_inits = self.inits['z'][-1::-1] Tk_inits", "in self.parcel_igm.grid.data.keys(): if k...
[ "1.11.15 on 2018-08-08 15:00 from __future__ import unicode_literals from django.db", "# Generated by Django 1.11.15 on 2018-08-08 15:00 from __future__", "models class Migration(migrations.Migration): dependencies = [ ('web', '0041_auto_20180808_1450'), ] operations", "] operations = [ migrations.AlterField(...
[ "self._collate_tags(tags=list(list(data.values())[0].keys())) # e.g. [error, overlap, ...] group_names = list(tag_groups.keys()) #", "Union[pd.DataFrame, Dict[str, pd.DataFrame]]], save_path: str, ) -> None: \"\"\"Make plots", "the unified case, full dataset is loaded into memory. \"\"\"", "* num_columns + co...
[]
[ "\".repo-config.yaml\"): if os.path.exists(path): with open(path) as file: config = yaml.safe_load(file.read())", "json={\"query\": query}, headers=headers ).json() if \"errors\" in response: err =", "apply settings to repos you forked as well \"\"\" config", "import Any import requests import yaml if os.path...
[ "+= 1 except: pdb.set_trace() if featureHitCountDict[featureToBeHit] == 1: featuresHitAtLeastOnce +=", "that is keyed by locus tag essentialityDict = {} locusTags", "== 'genbank': sequence = ImportGenBankSequence(genomeFile) elif format == 'fasta': sequence", "# # ---------------------------------------------...
[ "uncomment from onegram import save, unsave def test_follow(logged, user, cassette):", "test_follow(logged, user, cassette): if logged: response = follow(user) assert response", "'ok', 'post_id': post['id']} response = unsave(post) assert response == {'status':", "follow(user) assert response == {'result': 'f...
[ "# Note that you need to specify a delimiter and", "delimiter and an encapsulator # for your account (either in", "= cim_api.create_profile( card_number=u\"4111111111111111\", expiration_date=u\"2008-07\", customer_id=u\"test_account\") # Store the profile id", "of our users. tree = cim_api.create_profile( ca...
[ "e) = (-1, -1, -1) if triple: (c, d, e)", "triple: (c, d, e) = triple return a * b,", "while front_index != end_index: a = sorted_report[front_index] b = sorted_report[end_index]", "+ b == target_num: return a, b elif a +", "Tuple[int, int]: report_data = parse_input(input_data_filepath) report_data.sort() ...
[ "def prep_notebook(spark, aws_cluster_size: int = None, aws_asg_name: str = None,", "env.aws_asg_name region_name = env.aws_region_name cluster_size = env.aws_cluster_size adjust_ec2_asg(asg_name, region_name, cluster_size)", "if inst['LifecycleState'] == 'InService']) def adjust_ec2_asg(asg_name: str, region_n...
[ "self.following_categories.union(cleaned_data.get('adding_categories')) return cleaned_data class Meta(TopicOrganizeForm.Meta): fields = ('following_categories', 'adding_categories', )", "self.fields['following_categories'].queryset = self.following_categories class Meta: model = User fields =", "model = User f...
[ "<reponame>boomletsgo/falcon-rest from contextlib import contextmanager from sqlalchemy.orm import sessionmaker @contextmanager", "a context, and close it afterwards\"\"\" db_session = sessionmaker(bind=db_engine, **kwargs)()", "context, and close it afterwards\"\"\" db_session = sessionmaker(bind=db_engine, **...
[ "self.register_event_type('on_move_x') self.register_event_type('on_move_y') self.register_event_type('on_stop') self.register_event_type('on_stop_x') self.register_event_type('on_stop_y') self.bind(on_update=self.move) def _update_velocity(self): \"\"\"Change velocity", "default :type speed_limit: float \"\"\" s...
[ "model.\"\"\" secret_key: str = \"\" mongo_url: str = \"\" testing:", "= False @lru_cache(typed=False) def get_settings() -> Settings: \"\"\"Initialize settings.\"\"\" return", "lru_cache from pydantic import BaseSettings class Settings(BaseSettings): \"\"\"Settings model.\"\"\" secret_key:", "= \"\" testing:...
[ "as nm import time # mp.set_policy(policy.OnlyNumPyPolicy()) def test_autograd(): @convert_args def", "Wh, b, dnext_h: minpy_rnn_step_forward(x, h, Wx, Wh, b) * nm(dnext_h)", "test_sum(): x_np = np.array([[1, 2], [3, 4], [5, 6]]) x_grad", "@convert_args def minpy_rnn_step_forward(x, prev_h, Wx, Wh, b): next_h...
[ "l2 = list(filter(lambda el: el.x % 2 == 0, l1))", "i*i), range(-5, 6))) l2 = list(filter(lambda el: el.x % 2", "Point __author__ = 'pzqa' l1 = list(map(lambda i: Point(i, i*i),", "point import Point __author__ = 'pzqa' l1 = list(map(lambda i:", "i: Point(i, i*i), range(-5, 6))) l2 = list(filter(lambda el: ...
[ "else: l_r.append(j) return l_r x= [[1,\"a\",[\"cat\"],2],[[[3]],\"dog\"],4,5] print(flatten(x)) #SORU 2_________________________________________________________ def", "rev(l): for j in l: if type(j) == list: j.reverse()", "l: if type(j) == list: j.reverse() l.reverse() return l x=[[1,", "flatten(l): for j in...
[ "serialize=False, verbose_name='ID')), ('textArea', models.CharField(default='', max_length=300, null=True)), ('username', models.CharField(default='username', max_length=100)), ('name',", "models.IntegerField(default=0)), ('invitations_sent_accepted', models.IntegerField(default=0)), ('invitations_sent_rejected'...
[ "import context menu: delete, reset, select existing, as new, sk", "and all tracks doesn't screw up the ordering) + status:", "after enter is pressed (a problem with the groupname) -", "if there are any tracks that arent selected or marked", "time units - cuesheet has frames, everywhere else are seconds", ...
[ "f: for line in f: word = line.replace('\\n','') if word", "+ 1 with open(\"../output/analysis output/entity.txt\", \"r\", encoding=\"utf-8\") as f: for", "as f: for line in f: word = line.replace('\\n','') if", "for line in f: word = line.replace('\\n','') if word not", "word not in entity: entity[word] = ...
[ "Just checking # Please, pay attention, that this class imported", "value return res class SyntaxVectorizer(ABC): def setup_rules(self): pass def text_structures_initializer(self):", "os import spacy_udpipe from .utils import load_pickled_file from .settings import", "= {} if len(tag_representation.split('|')...
[ "regularization.forward(param) l += loss.forward(input, target) dout = loss.backward() for layer", "input): layers = self.layers for layer in layers: if isinstance(layer,", "self.loss = loss self.lr = lr self.regularization = regularization def", "Network(object): def __init__(self): super(Network, self).__in...
[ "input(\"Create TTL Indexes? [y/N] \") if create_ttl_indexes == 'y' or", "\")) db['download_cache'].create_index([(\"update_time\", 1)], expireAfterSeconds=download_cache_expires_in) illust_detail_cache_expires_in = int(input(\"Illust detail cache expires", "db['illust_detail_cache'].drop_index([(\"update_time\...
[ "version='1.2', description='Disk mounting shortcut for use with dmenu.', author='<NAME>', url='https://github.com/renatoliveira/monta',", "url='https://github.com/renatoliveira/monta', include_package_data=True, package_data={ '': [ 'monta', 'license.txt' ] }, scripts=[", "[ 'monta', 'license.txt' ] }, scripts...
[ "tf.keras.layers.Dense(units=1, input_shape=[1]) model = tf.keras.Sequential([fc0]) # model = tf.keras.Sequential([tf.keras.layers.Dense(units=1, input_shape=[1])])", "import logging logger = tf.get_logger() logger.setLevel(logging.ERROR) # training data celsius_q", "input_shape=[1]) model = tf.keras.Sequential...
[ "the database await self._init_database() async def _init_database(self): logger.info(\"Checking database...\") #", "feed. locations: [List of locations ids] Media with those locations", "open(\"comments/video_comments.txt\") as f: comments = f.readlines() self.video_comments_list = [x.strip() for", "1 params...
[ "file: template_data = file.read() soup = BeautifulSoup(template_data, 'html.parser') for tagname,", "= uuid.uuid1().hex[:15] path = os.path.join(tempfile.gettempdir(), self.tempuuid) if not os.path.isdir(path): os.mkdir(path)", "time.sleep(1) yield self.temp_uuid, self.temp_path subprocess.run( ('python3' if o...
[ "tags tags = [] for structure in file: if structure.info['config_type']", "structure.info['config_type'] not in tags: tags.append(structure.info['config_type']) # extract unique tags and", "tags: for structure in file: if structure.info['config_type'] == key and", "key in tags: for structure in file: if struc...
[ "global img_cp if event == cv2.EVENT_LBUTTONDOWN: cv2.circle(img_cp, (x, y), 5,", "cv2.resize(img, dsize=(1000, 1000)) img = cv2.resize(img, (0, 0), fx=0.75, fy=0.75,", "dsize=(1000, 1000)) img = cv2.resize(img, (0, 0), fx=0.75, fy=0.75, interpolation=cv2.INTER_NEAREST)", "= [] img = None img_cp = None def dr...
[ "import get_files, tei_reader from pprint import pprint class TestText(unittest.TestCase): def", "= corpora.text with open(expected, encoding='utf-8') as f: diffs = list(diff", "for diff in diffs if diff[0:2] != ' ']), 0,", "= list(diff for diff in differ.compare( [line.strip() for line in", "diff in differ...
[ "as rsautil import scryptos.crypto.attack.knapsackutil as knapsackutil import scryptos.crypto.attack.prngutil as prngutil", "import scryptos.crypto.attack.rsautil as rsautil import scryptos.crypto.attack.knapsackutil as knapsackutil import scryptos.crypto.attack.prngutil", "scryptos.crypto.attack.rsautil as rsa...
[ "import p1.m2 import p1.m3 import p1.m4 p1.m4.mm_main() import p1.pp1.a1 import", "import p1.m1 import p1.m2 import p1.m3 import p1.m4 p1.m4.mm_main() import", "p1.m2 import p1.m3 import p1.m4 p1.m4.mm_main() import p1.pp1.a1 import p1.pp1.a2", "p1 import * import p1.m1 import p1.m2 import p1.m3 import", "p...
[ "pass @abstractmethod def parse_data(self) -> pd.DataFrame: pass class MLS(DataBase): ext", "from abc import ABC, abstractmethod class DataBase(ABC): @abstractmethod def make_tidy(self):", "df_test, df_dev], ignore_index=True).assign(base=self.basename) class CommonVoice(DataBase): ext = \".mp3\" basename =", ...
[ "= ImageFont.truetype(font_file, font_size) text_width, text_height = width, 0 except KeyboardInterrupt:", "FONT font = ImageFont.truetype(font_file, font_size) text_width, text_height = width, 0", "unicornhathd.rotation(270) unicornhathd.brightness(0.8) width, height = unicornhathd.get_shape() text_x = width t...
[ "(By.XPATH, '/html/body/div/div[3]/input') def click_on_btn_3(self): btn_1 = self.driver.find_element(*self.RADIO_BTN_1) btn_2 = self.driver.find_element(*self.RADIO_BTN_2)", "import By class RadioButtonPage(BasePage): RADIO_BTN_1 = (By.ID, 'radio-button-1') RADIO_BTN_2 =", "'radio-button-1') RADIO_BTN_2 = (By....
[ "'Delivery Note', 'Stock Entry']: frappe.db.sql(\"\"\" update `tabStock Ledger Entry` sle,", "and sle.project is null and parent_doc.project is not null and", "sle.voucher_type = %s and sle.project is null and parent_doc.project is", "import frappe def execute(): for doctype in ['Sales Invoice', 'Delivery", ...
[ "p2 = p*2 p3 = p*3 plt.plot(p,p2,color = 'b',ls =", "= '-',linewidth = 3) plt.title(\"Two line in same plot\") plt.xlabel(\"X-Axis\")", "<reponame>abhayanigam/Learn_Python_Programming<filename>matplotlib/two_line_in_same_plot.py<gh_stars>1-10 import numpy as np from matplotlib import pyplot as", "numpy as np ...
[ "a 64-bit processor but in 32-bits mode, maybe try: return", "# assume we have 'uname' mach = os.popen('uname -m', 'r').read().strip()", "{'i386': 'i386', 'i486': 'i386', 'i586': 'i386', 'i686': 'i386', 'i86pc': 'i386',", "platform.machine() except ImportError: pass if not mach: platform = sys.platform.lower(...
[ "class Colour_Cyan(Colour): def __init__(self): super() self.code = 96 class Colour_White(Colour):", "= 33 class Colour_DarkBlue(Colour): def __init__(self): super() self.code = 34", "self.code = 36 class Colour_Grey(Colour): def __init__(self): super() self.code =", "def __init__(self): super() self.code = 4...
[ "action=\"store\", default=\"ftpdel.conf\", help=\"ftpdel setting file\" ) @pytest.fixture def cli_conf(request): return", "options.\"\"\" #RefURL: https://docs.python.org/3/library/argparse.html#the-add-argument-method parser.addoption( \"--change-conf\", action=\"store\", default=\"ftpdel.conf\", help=\"ftpdel ...
[ "'value': ', '.join(extra_found) })) # Now validate the actual data", "Check range min_errors = None max_errors = None if colspec.get('min',", "'column': series.name, 'sheet': sheet })) # Check range min_errors =", "wb['data'] errors.extend(validate_sheet_generic(df, 'data', spec)) # Next check columns in 'da...
[ "def IouCal(Box1, Box2): inner_x1 = torch.max(Box1[0], Box2[0]) inner_y1 = torch.max(Box1[1],", "target location by :math: `sum{P(y_i) * y_i}`, P(y_i) denotes the", "def forward(self, x): \"\"\"Forward feature from the regression head to", "your new dataset or related settings. \"\"\" def __init__(self, reg_m...
[ "\"x\", #1 \":\", #2 \"\\\\neg\", #3 \"P(x)\" ) for size,pos,formula", "= Text(\"What's your Openion, MOHI?\", height = 0.8) source.set_color(GREEN) source1.set_color(RED)", "first\",\" test\", font =\"Arial\" , font_size = 44, text_color =", "self.wait(3) class deff(Scene): def construct(self): text = TexTex...
[ "\" }} {{range $index,$value := $b}} {{- if or (hasPrefix", "n st += \"\"\" }} {{if or (eq $a \"stickers\")", "\"gif\") (eq $a \"gifss\") }} {{deleteTrigger 0 }} {{if eq", "$a \"gifss\") }} {{deleteTrigger 0 }} {{if eq (len .Args)", "+ name + \"\\\" \" st += n st +=", "{{$r := str $r}} {{$r}} {{ deleteRes...
[ "= cv2.imread(infile) #conver to grayscale gray_image = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY) #blur", "nogui=False): # The first argument is the image image =", "cv2.COLOR_BGR2GRAY) #blur it blurred_image = cv2.GaussianBlur(image, (7,7), 0) if nogui:", "argument is the image image = cv2.imread(infile) #conv...
[ "of best practicing with multiple state variables.\"\"\" __author__ = \"<NAME>\"", "under the License. \"\"\" __license__ = \"Apache 2.0\" def fibonacci_wrong_way(n):", "1 l = [] for i in range(n): l.append(x) t", "0, 1, [] for i in range(n): l.append(x) x, y", "range(n): l.append(x) x, y = y, x + y print(l...
[ "\"w:gz\") for fil in files: logger.info (\"Packing %s into statistics", "= {} json_data['output_files']= [] if (\"createStructure\" in args.config[\"arguments\"][\"operations\"]) or (\"create3DfromNucleaR\"", "/home/MuG/MuG_Chromatin_equ_structure/src_test; bash run.sh %s %s %s\" % (tmp_dir +\"/nucleR_to_3D_nu...
[ "``True`` a ``.png`` figure is saved to hardrive filename_out (str):", "numpy as np # Suppress matplotlib warnings np.warnings.filterwarnings('ignore') import xarray", "#RGB/256 cdict = {'red': ((0.0, c1[0], c1[0]), (1.0, c2[0], c2[0])),", "plt.figure(figsize=(19,9)) # ax = fig.add_axes([0.05,0.05,0.80,1]) # ...
[ "import init from .root import root # add sub-command functions", ".init_cfg import init from .root import root # add sub-command", "<gh_stars>0 from .init_cfg import init from .root import root #", "init from .root import root # add sub-command functions here", "from .root import root # add sub-command fun...
[ "attribute ``object.my_attribute`` is accessed for the first time. Example: Use", "of Python attributes. In the example below, the ``expensive_computation()`` is", "first time. Example: Use the ``@lazy`` decorator for a function", "function that returns the result of the computation. Access it", "evaluation...
[ "allows to repair individuals after crossover if necessary. \"\"\" def", "abstractmethod class Repair: \"\"\" This class is allows to repair", "\"\"\" def do(self, problem, pop, **kwargs): return self._do(problem, pop, **kwargs)", "do(self, problem, pop, **kwargs): return self._do(problem, pop, **kwargs) @abs...
[ "DAO Object for BlockSite table \"\"\" from dbs.dao.Oracle.BlockSite.Insert import Insert", "BlockSite table \"\"\" from dbs.dao.Oracle.BlockSite.Insert import Insert as OraBlockSiteInsert class", "python \"\"\" DAO Object for BlockSite table \"\"\" from dbs.dao.Oracle.BlockSite.Insert", "#!/usr/bin/env pytho...
[ "& (df.ses == name[3]) & (df.task == name[4]) & (df.roi", "df = pd.read_csv(df_fn) if not os.path.exists(qc_fn) : qc = get_qc_metrics()", "[] fn_list += glob(\"raclopride/out_rcl/groupLevelQC/coreg_roc/test_group_qc_auc.csv\") #fn_list += glob(\"scott/out_fdg/groupLevelQC/coreg_roc/test_group_qc_auc.csv\") #fn_...
[ "'Fri', # d0e1f1 in iso-8859-7 '\\u03a3\\u03b1\\u03b2': 'Sat', # d3e1e2 in", "# cce1fa in iso-8859-7 '\\u039c\\u03b1\\u03b9': 'May', # cce1e9 in iso-8859-7", "a string according to a Greek 8-bit date format.''' m", "%(day)s %(month)s %(year)s %(hour)s:%(minute)s:%(second)s %(zonediff)s' % \\ {'wday': wday, 'd...
[]
[ "), migrations.CreateModel( name='UserProject', fields=[ ('uuid', models.UUIDField(default=uuid.uuid4, editable=False, primary_key=True, serialize=False)), ('name',", "settings from django.db import migrations, models import django.db.models.deletion import uuid", "'Orthomosaic'), (core.models.ArtifactType['SHA...
[ "-data Classifier -vec allvecs.vec -bg negatives.txt -numNeg 1000 -numPos 3000", "60 -h 60 -vec output${filename}.vec -maxzangle 0.5 -maxyangle 0.3 -maxxangle", "= cv2.CascadeClassifier('cascade.xml') img = cv2.imread('orange.jpg') gray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY) oranges", "being collected # $ fo...
[ "} DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': os.path.join(ARMORY_CONFIG['ARMORY_BASE_PATH'],", "], } DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME':", "= abolute_home + \"/tools/armory_custom/reports\" custom_webapps = abolute_home + \"/tools/armory_we...
[ "self.lastScannedBlock: self.lastScannedBlock += 1 self.checkBlock(self.lastScannedBlock) cursor = self.dbCon.cursor() cursor.execute('UPDATE heights", "self.lastScannedBlock = cursor.execute('SELECT height FROM heights WHERE chain = \"TN\"').fetchall()[0][0]", "def __init__(self, config): self.config = config ...
[ "instances have started. \"\"\" RUNNING = \"RUNNING\" \"\"\" The computer", "may no longer be changed and is considered to be", "being stopped. Some instances may already have stopped, however the", "aggregated view of the statuses of compute machine instances provisioned", "have transitioned to RUNNING or ...
[]
[ "def test_validate_config(self): for i in range(2, 5): with self.subTest(i=i): config", "def test_load_config(self): self.assertIsNone(load_config('tests/config/invalid_config_1.yaml')) self.assertDictEqual( validate_config(load_config('tests/config/valid_config_5.yaml')), { c.LEVELS: [{ c.BUY_PRICE: 95,", "c.O...
[ "code, Name, Continent,population,HeadofState from country\" while True: Comparison = input(\"Enter", "elif choice == \"2\": print(\"Cities by Population\") print(\"--------------------\") while True:", "in car: print (p) except : print (\"******Error Occurred while", "exist\") print(\"-----------------------...
[ "\") # ^^This asks your class subject; assigns it to", "a filename A = input('Do you want this to be", "keep record of grades. Made by <NAME>. 0.1-PUBLIC # NOTE!", "Gradebook! v 0.1 # # YOUR LIGHT WEIGHT SCHOOL RECORD", "(including dates)! print \"\"\"############################################ # Welcome t...
[ "in line: is_one_to_one = False is_ortho_plus_paralog = False break if", "f: is_one_to_one = True is_ortho_plus_paralog = True for group in", "[] one_to_one_notallspecies = [] ortho_plus_paralog_notallspecies = [] group_single = {}", "# # for key in group_paralog: # print(\"\\n{} paralog\\n\".format(key)) #",...
[ "unicode_literals from django.conf.urls import url, include from . import views,", "<gh_stars>0 from __future__ import absolute_import, unicode_literals from django.conf.urls import url,", "import url, include from . import views, api_urls urlpatterns =", "import absolute_import, unicode_literals from django....
[ "on 2020-04-01 13:24 from django.db import migrations class Migration(migrations.Migration): dependencies", "Django 3.0.4 on 2020-04-01 13:24 from django.db import migrations class", "migrations class Migration(migrations.Migration): dependencies = [ ('reading', '0007_auto_20200331_2133'), ] operations", "13:...
[ "api call failed' % (e, e)) if not self._ctx.pretty: print(json.dumps(records,", "print(json.dumps(records, indent=4)) return records_by_type = {} types = {} for", ") parser_zones.add_argument( \"-z\", \"--zone-name\", help=\"optional zone name\", ) parser_dns =", "for rec in result: records.append(rec) if pa...
[ "= True global csvfile csvfile = csv.writer(local_csvfile, delimiter=';') if new", "= ('pep', 'run') # retrieve input params try: protocol =", "insert page into database if 'error' not in performance_metrics: #", "timestamp performance['cacheWarming'] = cache_warming performance['error'] = error values = file...
[ "Technology, HeiGIT gGmbH, Heidelberg, Germany. /*************************************************************************** * * * This", "and toolbar icons inside the QGIS GUI.\"\"\" QgsApplication.processingRegistry().addProvider(self.provider) self.dialog.initGui() def", "Implementation.\"\"\" # noinspection...
[ "<filename>dashmat/custom/reviews/main.py from dashmat.core_modules.base import Module class Reviews(Module): @classmethod def dependencies(kls):", "dashmat.core_modules.base import Module class Reviews(Module): @classmethod def dependencies(kls): yield \"dashmat.core_modules.components.main:Components\"", "fro...
[ "import Simulation import rospy if __name__ == \"__main__\" : rospy.init_node('learn_to_manipulate')", "import rospy if __name__ == \"__main__\" : rospy.init_node('learn_to_manipulate') sim =", "rospy if __name__ == \"__main__\" : rospy.init_node('learn_to_manipulate') sim = Simulation.load_simulation('/home/ma...
[ "python # encoding: utf-8 \"\"\" @author: zhanghe @software: PyCharm @file:", "structure_key_item = 'enum_item' structure_key_items = 'enum_items' structure_key_item_cn = '枚举类型' structure_key_items_cn", "@file: enum_items.py @time: 2018-08-23 15:55 \"\"\" from __future__ import unicode_literals", "#!/usr/bin/...
[]
[ "-*- coding: utf-8 -*- \"\"\" Text used by GOTO to", "used by GOTO to do UX. \"\"\" from .text import", "coding: utf-8 -*- \"\"\" Text used by GOTO to do", "by GOTO to do UX. \"\"\" from .text import GotoError,", "# -*- coding: utf-8 -*- \"\"\" Text used by GOTO", "utf-8 -*- \"\"\" Text used by GOTO to do...
[]
[ "import setup VERSION = \"1.0.4\" NAMESPACE = \"newstore\" NAME =", "= os.path.join(os.path.dirname(__file__), file_name) with open(path, \"rt\") as fp: file_data =", "NAME = \"{}.json_encoder\".format(NAMESPACE) def local_text_file(file_name): path = os.path.join(os.path.dirname(__file__), file_name) with", ...
[ "self.get_children(depth=-1, include_self=True, gen=True): part.draw_create() assert \"Contain\" in getBaseClassNames(parent) or parent", "exists as an old order then it's removed. Returns key", "import getBaseClassNames, SigInfo, dict_insert, wrapper_transfer def set_parent_hook(self, parent, _draw=True): \"\"...
[ "from os_aio_pod.cmdline import execute def main(): command_packages = [\"os_aio_pod.commands\"] execute(command_packages=command_packages)", "<filename>src/os_aio_pod/main.py from os_aio_pod.cmdline import execute def main(): command_packages = [\"os_aio_pod.commands\"]" ]
[ "in_topo.json os.system('python gen_router_json.py') worker.put_file(\"routernew.json\", \"/tmp/routernew.json\") print \"***** Experiment Setup Start", "my_cmd in data[\"host_cmnds\"] : print \"Execute Command on Host ...\",", "<filename>MaxiNet/WorkerServer/tst_driver.py<gh_stars>1-10 #!/usr/bin/python2 # # T...