ngram
listlengths
0
82k
[ "# все транзакты Terminate(sgm, label=\"to_term\", deltaTerminate=0) # ЗАПУСК ---------------------- m.start(terminationCount=MAX_TIME,", "MAX_TIME = 20 # list_all_transact = [] # MAX_TIME =", "t[k]) for k in t.keys() if k in [TIME_CREATED, TERMINATED_TIME]])", "self.assertNotIn(F_1, t[FACILITY])) # Terminate...
[ "compute_gradients(self, samples): \"\"\"Returns critic, actor gradients.\"\"\" return self.model.compute_gradients(samples) def apply_gradients(self,", "samples): \"\"\"Returns critic, actor gradients.\"\"\" return self.model.compute_gradients(samples) def apply_gradients(self, grads):", "gradients.\"\"\" retu...
[ "e.g. hosts or switches. And the connections between them. '''", "List[dict] ''' return self._vertexes.hosts + self._vertexes.switches @property def switches(self): '''", "''' Draw a dot graph of the network graph. :params", "contains the entities in the network e.g. hosts or switches.", "graphargs: Argumen...
[ "for idx, (blobs, color, title) in enumerate(sequence): ax[idx].set_title(title) ax[idx].imshow(image) for", "axes.ravel() for idx, (blobs, color, title) in enumerate(sequence): ax[idx].set_title(title) ax[idx].imshow(image)", "'Determinant of Hessian'] sequence = zip(blobs_list, colors, titles) fig, axes", "...
[ "value): if value in listofnames: value_index = listofnames.index(value) return(listofnames, value_index)", "value in listofnames: value_index = listofnames.index(value) return(listofnames, value_index) else: return(-1)", "if value in listofnames: value_index = listofnames.index(value) return(listofnames, value...
[ "1 def move(self): new_x = self.xcor() + SPEED*self.x_bounce new_y =", "self.move_speed = 0.1 self.y_bounce = 1 self.x_bounce = 1 def", "= 10 class Ball(Turtle): def __init__(self): super().__init__() self.penup() self.color(\"white\") self.shape(\"circle\")", "def move(self): new_x = self.xcor() + SPEED*self...
[ "self.term class TreeInnerNode(TreeNode): def __init__(self, pred, left, right): self.pred =", "pred, left, right): self.pred = pred self.left = left self.right", "getExpr(self): return self.term class TreeInnerNode(TreeNode): def __init__(self, pred, left, right):", "class TreeNode: pass class TreeLeaf(TreeN...
[ "base ring as well. zeta = self._zeta if zeta_order is", "\\frac{B_{k,\\varepsilon}}{k!} t^k. ALGORITHM: The ``'recurrence'`` algorithm computes generalized Bernoulli numbers", "fixed:: sage: G = DirichletGroup(1) sage: chi = G.one() sage:", "sum associated to `\\chi` is .. MATH:: g_a(\\chi) = \\sum_{r", "t...
[ "write(self, file): \"\"\" Write the contents of this :class:`File` object", "Base class for all line based text files. When reading", "name if isinstance(file, str): with open(file, \"r\") as f: while", "file (or file-like object). Parameters ---------- file_name : file-like object", "isinstance(file, str)...
[ "if res.status_code == 200: logger.info(\"%r sent, FCM id: %r\", pnt,", "prepared push notification translation to be sent :type pnt: ~cms.models.push_notifications.push_notification_translation.PushNotificationTranslation", ").slug # Testumgebung - prevent sending PNs to actual users", "\"\"\" status = True ...
[ "decription: CREATE ASC INDEX # # Dependencies: # CREATE DATABASE", "id: functional.index.create.03 # title: CREATE ASC INDEX # decription: CREATE", "# resources: None substitutions_1 = [] init_script_1 = \"\"\"CREATE TABLE", "from firebird.qa import db_factory, isql_act, Action # version: 1.0 #", "substitu...
[ "information about the pare we are refering, index, items per", "contains information about the pare we are refering, index, items", "are refering, index, items per page, etc. \"\"\" page_index =", "we are refering, index, items per page, etc. \"\"\" page_index", "Page object file \"\"\" class Page(): \"\"\...
[ "# # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law", "get_names(): return list(__factory.keys()) def init_dataset(name, *args, **kwargs): if name not", "# # Licensed under the Apache License, Version 2.0 (the", "compliance with the License. # You may obtain a copy", "an \"...
[ "Decrease in Profits: {minmonth} (${minpl})\\\")\\n\" ] }, { \"cell_type\": \"code\",", "\"print(f\\\"Total Months: {n_months}\\\")\\n\", \"print(f\\\"Total: ${round(pl_total,2)}\\\")\\n\", \"print(f\\\"Average Change: ${round(avg_pl_change,2)}\\\")\\n\", \"print(f\\\"Greatest Increase", "for max and min PL cha...
[ "@dll_import('OleAut32') def VariantInit( pvarg : POINTER(VARIANT) ) -> None: ...", "..wintypes import VARIANT, dll_import @dll_import('OleAut32') def VariantInit( pvarg : POINTER(VARIANT)", "from ctypes import POINTER, Structure from ..wintypes import VARIANT, dll_import", "VARIANT, dll_import @dll_import('O...
[ ".models import EventTypeDescriptor from .models import ExternalConfigurationDescriptor from .models import", "from .models import EventTypeDescriptor from .models import ExternalConfigurationDescriptor from .models", "will be lost if the code is regenerated. # --------------------------------------------------...
[ "def __init__(self): global __logger if self.__logger: raise RuntimeError(\"Logger instance already", "\"INFO\") def log_warning(self, warning): self.__log(warning, \"WARNING\") def log_error(self, error): self.__log(error,", "instance already exists\") @staticmethod def get_logger(): global __logger if not", ...
[ "give unique orientations when the # Burgers transformation is applied", "cubic_syms = Quat.symEqv(\"cubic\") # subset of cubic symmetries that give", "unq_hex_syms = [ hex_syms[0], hex_syms[5], hex_syms[4], hex_syms[2], hex_syms[10], hex_syms[11] ]", "the # Burgers transformation is applied unq_hex_syms = [ ...
[]
[ "required\" })), 400 if not VALIDATOR.username_exists(username): return jsonify({ \"status\": 404,", "auth_token = user.generate_auth_token(username) return make_response(jsonify({ \"status\": 200, \"message\": 'Logged in", "not email or not email.split(): return make_response(jsonify({ \"status\": 400, \"messa...
[ "(UnitsContainer) from pint.converters import (ScaleConverter, OffsetConverter) from pint.definitions import (Definition,", ")) self.assertEqual(x.symbol, 'k') self.assertEqual(x.converter.to_reference(1000), 1) self.assertEqual(x.converter.from_reference(.001), 1) def test_baseunit_definition(self): x", "impor...
[ "# Check if there is a valid signature for the", "A PARTICULAR PURPOSE AND # NONINFRINGEMENT. IN NO EVENT SHALL", "CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN # ACTION", "this software and associated documentation files # (the \"Software\"), to", "# KSK-2010: dns.rrset.from_text('.', 15202, 'IN', 'DNSK...
[ "= Enum(\"D3D11_RTV_DIMENSION\", [ \"D3D11_RTV_DIMENSION_UNKNOWN\", \"D3D11_RTV_DIMENSION_BUFFER\", \"D3D11_RTV_DIMENSION_TEXTURE1D\", \"D3D11_RTV_DIMENSION_TEXTURE1DARRAY\", \"D3D11_RTV_DIMENSION_TEXTURE2D\", \"D3D11_RTV_DIMENSION_TEXTURE2DARRAY\", \"D3D11_RTV_DIMENSION_TEXTURE2DMS\",", "Struct(\"D3D11_TEX2DMS_D...
[ "op == 'nop': index += 1 if op == 'acc':", "result(input_): # Part 1 part_one = boot(input_)[1] # Part 2", "all_seqs.append(seq) if value[:3] == 'jmp': seq = deepcopy(list_) seq[idx] =", "== 'nop': seq = deepcopy(list_) seq[idx] = 'jmp' + value[3:]", "from copy import deepcopy def boot(seq): index = 0 playe...
[ "lr = tf.train.exponential_decay(0.001, global_step, 10000, 0.8, staircase=True) #lr = tf.constant(0.001,", "samples=input_data.get_images_labels(tfRecorf_dir,num_classes,66, crop_size=[224,224], batch_size=4) batch_queue = slim.prefetch_queue.prefetch_queue(samples, capacity=128 ) tra_batch =", "@author: LiHon...
[ "Python :: 3.8', 'Topic :: System', 'Topic :: System ::", ":: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language", "long_description=codecs.open('README.rst', 'r', encoding='utf-8', errors='ignore').read(), keywords='file libmagic magic infer numbers magicnumbers", "dependencies.', long_desc...
[ "templates from netmiko import ConnectHandler import json user = 'ntc'", "dicts print (type(sh_ip_int_br[0])) for each_dict in sh_ip_int_br: print \"\\n\" for", "if type(value) is list: print key + \" is \"", "value: print list_entry if type(value) is str: print key +", "'config_register': '0x2102', 'hardwa...
[ "JWT passed to your App Engine app by Identity-Aware Proxy.", "a public key from the list published by Identity-Aware Proxy,", "found'.format(key_id)) return key # Used to cache the Identity-Aware Proxy", "Cloud project. Returns: (user_id, user_email, error_str). \"\"\" expected_audience = '/projects/{}/apps/...
[ "self.txt ) def compute(self): try: self.txt = str(eval(self.txt.replace(\"x\",\"*\"))) self.aff.set( self.txt", "self.compute() elif key in [\"Delete\",\"Backspace\"]: self.clean() #-/-/-/-/-/-/-/-/-/-/-/-/-/-/-/-/-/-/-/-/-/-/-/-/-/-/-/-/-/-/-/ def press(self,val): self.txt", "# and execute it in a pywebview i...
[ "return (x-3)*(x+2)*x*0.2 g = Graph(800,600,8,6, 'example1.svg') g.bg() g.grid() g.axes() g.graph(func)", "g = Graph(800,600,8,6, 'example1.svg') g.bg() g.grid() g.axes() g.graph(func) g.save() g.display()", "from giraphics.graphing.graph import Graph def func(x): return (x-3)*(x+2)*x*0.2 g =", "<filename>res...
[ "for i in range(0,df.shape[0]): for j in range(0,df['num'][i]): poly=json.loads(df['polygon'][i]) GeoFunc.normData(poly,scale[index])", "name=[\"ga\",\"albano\",\"blaz1\",\"blaz2\",\"dighe1\",\"dighe2\",\"fu\",\"han\",\"jakobs1\",\"jakobs2\",\"mao\",\"marques\",\"shapes\",\"shirts\",\"swim\",\"trousers\"] print(\...
[ "import collect_hierarchical_module_name, collect_hierarchical_parameter_name, get_batch_n, to_value, \\ # safe_lookup, len_batch from", "\\ # safe_lookup, len_batch from .export import as_image_ui8, as_rgb_image, export_image,", "# safe_lookup, len_batch from .export import as_image_ui8, as_rgb_image, export_i...
[ "from tqdm import tqdm from pathos.multiprocessing import ProcessingPool as Pool", "with Pool(opt_threads) as p: pool_results = list(tqdm(p.imap(pool_worker, pool_items), total=len(fp_items), desc=desc))", "range(20): im = cv.blur(im, (35,35)) return result # end pool", "default=(None, None), help='Slice list...
[ "<filename>src/learndash/api_resources/user.py import learndash from learndash.api_resources.abstract import ListableAPIResource from learndash.api_resources.abstract import", "User(RetrievableAPIResource[UserDict], ListableAPIResource[UserDict]): api_path = learndash.path_users def course_progress(self, id=None)...
[ "subclasses.\"\"\" self.app_info = app_info self.resolver_kwds = kwds def _get_config_option(self, key,", "getattr(self.app_info, key) else: return default @abstractmethod def resolve(self, enabled_container_types, tool_info,", "argument is a :class:`galaxy.tool_util.deps.containers.ToolInfo` description of the...
[ "file_names: df: pd.DataFrame = pd.read_csv(f'{SRC_DIR}/{file_name}').set_index('EyeTrackerTimestamp').sort_index()[ ['GazePointX (ADCSpx)', 'GazePointY (ADCSpx)', 'PupilLeft',", "python3 import glob import os import pandas as pd import", "df['d'] = (df['dl'] + df['dr']) / 2 # start with", "to ms df['t'] = (d...
[ "and ('learning_rate' in data.columns): sortedData=data.sort_values(by=\"epochs\", axis=0, ascending=True) epoch=np.array(sortedData['epochs']) trainLoss=np.array(sortedData['trainLoss']) valLoss=np.array(sortedData['valLoss'])", "DIRECTORIES containing non-hidden files ending in FILENAME def getDataDirectories(D...
[ "char in dic: stack.append(char) elif stack and dic.get(stack[-1])!=char: return False", "in dic: stack.append(char) elif stack and dic.get(stack[-1])!=char: return False else:", "elif stack and dic.get(stack[-1])!=char: return False else: stack.pop() continue return", "[] dic = {'{':'}','[':']','(':')'} for ...
[ "None: info = _ConcatInfo() self._cache[path] = info with open(path, 'r')", "yaml.load(fp) info.files = config.get('files', []) info.delim = config.get('delim', \"\\n\") info.timestamp", "getDependencies(self, path): info = self._load(path) return info.files def getOutputFilenames(self, filename):", "import l...
[ "CellPressed(Event): def __init__(self, position): self.position = position def get_position(self): return", "import Event class CellPressed(Event): def __init__(self, position): self.position = position", "from src.events import Event class CellPressed(Event): def __init__(self, position): self.position", "E...
[ "discriminator value allowed for non-b jets minBDiscBJets = cms.double(1.0), maxBDiscLightJets", "useBTagging = cms.bool(False), ## choose algorithm for b-tagging bTagAlgorithm =", "as cms # # module to make the MaxSumPtWMass jet", "jets and ## maximum b discriminator value allowed for non-b", "b-tagging tw...
[ "#-*- coding:utf-8 -*- __all__ = [\"args\", \"colors\", \"libcolors\", \"routine\"] __version__", "-*- __all__ = [\"args\", \"colors\", \"libcolors\", \"routine\"] __version__ = \"0.96\"", "<reponame>runapp/xortool<filename>xortool/__init__.py #!/usr/bin/env python #-*- coding:utf-8 -*- __all__ = [\"args\", \"c...
[ "# TODO : ButtonRessourcePack.style.create_surface(size) class _RessourcePack: def __init__(self): self.font =", "class ScenesRessourcePack(RessourcePack): def __init__(self, background_color=(170, 170, 170), ): assert is_color(background_color)", "Object from baopig.pybao.issomething import * class RessourcePa...
[ "int(infilename.split(\".\")[0].split(\"_\")[-1]) cmatch = 0 counter = 0 # loop all", "extent change from colormesh plt.title(title) if doText: plt.text(0.01, 0.98, \"#stations:", "year station_names = np.array([]) fixed_station = np.array([]) latitudes = np.array([])", "= \"MARS - SYNOP - {}\".format(year) e...
[ "Authorized Connect Apps belonging to an account. :var Page: The", "used to avoid listing duplicated resources if new ones are", "get(self, FriendlyName=None, Page=None, PageSize=None, AfterSid=None): \"\"\" Fetch the Applications belonging", "= http.Request('GET', self.get_url(), params) return request, pars...
[ "feature extraction dimensions. Default is 64.\") args_opt = parser.parse_args() #", "64.\") args_opt = parser.parse_args() # Runtime context.set_context(mode=context.GRAPH_MODE, device_target='Ascend', device_id=0) #", "value t_sum = np.sum(e_x, axis=1, keepdims=True) # returns sum of", "2.0 (the \"License\"...
[ "= 0 elif pykeops.config.torch_found: import torch if all([type(var) in [torch.Tensor,", "= OrderedDict([('host',0), ('device',1)]) possible_options_list = ['auto', 'CPU', 'GPU', 'GPU_1D', 'GPU_1D_device',", "OrderedDict import pykeops import pykeops.config ######################################################...
[ "= wavfile.read(wav_path) x = x.astype(np.float64) f0, timeaxis = pyworld.dio(x, fs,", "y = P.adjust_frame_lengths(x, y, pad=True, divisible_by=2) # Save np.save(src_path, x)", "= FileSourceDataset(MGCSource(DATA_ROOT, [target_speaker], max_files=max_files)) skip_feature_extraction = exists(join(dst_dir, \"X\")...
[ "json.loads(el.deck_gl_json_chart.json) self.assertEqual(actual[\"layers\"][0][\"@@type\"], \"ScatterplotLayer\") self.assertEqual( actual[\"layers\"][0][\"data\"], [ {\"lat\": 1, \"lon\": 10},", "# # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law", "testutil import streamlit a...
[ "'pipeline_options'): self.project_id = ( sink.pipeline_options.view_as(GoogleCloudOptions).project) assert self.project_id is not None", "self.project_id is not None self.dataset_id = self.sink.table_reference.datasetId self.table_id = self.sink.table_reference.tableId", "values are not JSON compliant # This c...
[ "version of Qt libraries when there is installed another application", "there is installed another application (e.g. QtCreator) if is_win: from", "import extend_system_path extend_system_path([os.path.join(x, 'PyQt5') for x in getsitepackages()]) extend_system_path([os.path.join(os.path.dirname(get_module_file_...
[ "== 666 def test_compare_nodes(): old_node = vy_ast.parse_to_ast(\"foo = 42\") new_node", "new_node = vy_ast.Int.from_node(old_node, value=666) assert isinstance(new_node, vy_ast.Int) def test_source(): old_node", "new_node = vy_ast.Int.from_node(old_node, value=666) assert old_node.value == 42 assert new_node....
[ "cd opencv/build && \\ cmake -D CMAKE_BUILD_TYPE=RELEASE \\ -D CMAKE_INSTALL_PREFIX=/usr/local", "WITH_GTK=ON \\ -D WITH_LIBV4L=ON \\ -D BUILD_TESTS=OFF \\ -D BUILD_PERF_TESTS=OFF", "libopencore-amrwb-dev \\ libtheora-dev \\ libvorbis-dev \\ libxvidcore-dev \\ x264 \\", "\\ cmake -D CMAKE_BUILD_TYPE=RELEASE \...
[ "iterations = 1000000000 while n<iterations: for s in range(0,len(program)): arguments", "16 stage = map(chr, range(ord('a'),ord('a')+stage_length)) def spin(amount): \"\"\"To save time,", "history.index(stage) complete_cycles = (iterations - n) / loop_length n +=", "elif action_list[s] == 'x': swap((int(argu...
[ "as np from numpy.testing import assert_equal, assert_array_equal from nose.tools import", "= felzenszwalb(img, sigma=0) # we expect 4 segments: assert_equal(len(np.unique(seg)), 4)", "i in range(4): hist = np.histogram(img[seg == i], bins=[0, 0.1,", "for i in range(4): hist = np.histogram(img[seg == i], bins...
[ "from masonite.app import App from masonite.middleware import CsrfMiddleware from masonite.testsuite.TestSuite", "generate_wsgi import pytest from masonite.exceptions import InvalidCSRFToken class TestCSRFMiddleware: def", "[] with pytest.raises(InvalidCSRFToken): self.middleware.before() def test_incoming_toke...
[ "as f: faq_page = f.read() return {'content': faq_page} @view_config(route_name='conventions', renderer='conventions.mako')", "with open(conventions_file, 'r') as file: conventions_page = file.read().replace('\\n', '') return", "faq_file = os.path.join(dir_path, 'static/faq_with_indexes.html') with open(faq_fil...
[ "list ordered by \"id\" in ascending order \"\"\" RESTApiTestCase.process_test(self, 'computers',", "def test_computers_mixed3(self): \"\"\" url parameters: id, transport_type, orderby \"\"\" node_pk", "extras_filter with pagination ############# def test_node_extras_filter_pagination(self): \"\"\" Check that n...
[ "return False ts = datetime.now().strftime(\"%Y-%m-%d_%H-%M-%S\") filename = f\"{ts}_process-metrics-report_{name}_{duration}_{interval}.csv\" with open(f\"{filename}\",", "print(\"interval parameter is greater than duration parameter\") return ExitStatus.INTERVAL_GT_DURATION print(\"-----------------------------...
[ "normal_batch def test_loader(self): with tf.variable_scope('test_loader'): dataset = tf.data.Dataset.from_tensor_slices(self.test_x).repeat() normal_batch =", "label, *_ = line.split('\\t') self.anno[filename] = label def init_train(self): train_x,", "self.train_normal, name='normal_batch') random_crop_batch =...
[ "= True break if not ok: stderr.write(\"No markdown file found", "os.listdir(\".\"): if file.endswith(\".md\"): ok = True break if not ok:", "+ \"/src\"): os.symlink(script_path + \"/src\", tmp_dir + \"/src\") copy_tree(\".\", tmp_dir)", "args = parser.parse_args() # Check directory ok = False for", "!= \"\...
[ "visited, is_possible) visited[cur] = False if __name__ == '__main__': input", "dfs(node, 0, visited, is_possible) if is_possible[0]: return 1 return 0", "if visited[cur]: return if depth == target_depth: is_possible[0] = True", "[False] for node in range(N): visited = [False for _", "target_depth = 4 N, M ...
[ "to share' 'private data on BigchainDB.'), url='https://github.com/vrde/bst/', author='<NAME>', author_email='<EMAIL>', license='AGPLv3',", "3.5', 'Operating System :: MacOS :: MacOS X', 'Operating System", "Developers', 'Topic :: Database', 'Topic :: Database :: Database Engines/Servers',", "Linux', ], packa...
[ "<gh_stars>0 from django.contrib import admin from db.models.job_resources import JobResources admin.site.register(JobResources)" ]
[ "list_output_dict = [] # output directory path outdir = \"../results/run1/\"", "weight_column_idx] new_X_train = np.delete(X_train, weight_column_idx, axis=1) new_questions = column_names new_questions.remove('weight')", "= { 'criterion': ['gini', 'entropy'], 'max_depth': [2, 3, 4, 5,", "o_models_file.write(s...
[ "self.account.id) self.assertIn(hero.gender, (game_relations.GENDER.MALE, game_relations.GENDER.FEMALE)) self.assertEqual(hero.preferences.energy_regeneration_type, hero.race.energy_regeneration) self.assertEqual(hero.habit_honor.raw_value, 0) self.assertEqual(hero.habit_peacefulness.raw_value, 0)", "is_fast=Fals...
[]
[ "information from __future__ import absolute_import from __future__ import division from", "The code is released under the MIT Licence. # See", "y_inlet = self.case.boundary_data(\"inlet\", sort=\"y\")[0][:, 1] inlet_edge_length = tbl.edge_lengths(self.case, \"inlet\") self.d", "See LICENCE.txt and the Legal ...
[ "__init__(self, frequency=10): SteppableBasePy.__init__(self, frequency) def step(self, mcs): if mcs in", "def __init__(self, frequency=10): SteppableBasePy.__init__(self, frequency) def step(self, mcs): if mcs", "step(self, mcs): if mcs in [3, 5, 19,20, 23, 29,", "def step(self, mcs): if mcs in [3, 5, 19,20,...
[ "f(local_opt): name = (kwargs and kwargs.pop(\"name\")) or local_opt.__name__ gpu_optimizer.register(name, local_opt,", "\"matrix_ops_db2\" abstract_batch_norm_db = LocalGroupDB() abstract_batch_norm_db2 = LocalGroupDB(local_opt=GraphToGPULocalOptGroup) abstract_batch_norm_db2.__name__ = \"abstract_batch_norm_db2...
[ "urlquote log = logging.getLogger(__name__) class Node(JenkinsBase): \"\"\" Class to hold", "of the node :param jenkins_obj: ref to the jenkins obj", "are taking this node offline \"\"\" if not self._data['offline']: self.toggle_temporarily_offline(message)", "self.jenkins def __str__(self): return self.name ...
[ "= self.createTimeRange(0, 240) self._ddhhmmTime = self.getCurrentTime( argDict, \"%d%H%M\", shiftToLocal=0, stripLeading=0)", "the public domain, furnished \"as is\", without technical # support,", "self.getCurrentTime( argDict, \"%l%M %p %Z %a %b %e %Y\", stripLeading=1)", "*|\\n\\n\" ## # bullets = bullets...
[ "= splited_aux_lig[1] pose = area[0] buried_total = \"{:.4f}\".format(area[1]) #line =", "as normalized_buried_area FROM database a JOIN buried_area_total_sort b ON b.ligand", "Popen(command,shell=True, stdout=PIPE, stderr=PIPE) stdout, stderr = process.communicate() sasa_complex = get_value_from_xvg_sasa(f_tem...
[ "'smtp_server', request=request) if len(rq['smtp_server']) > 128: error_flag = True error_msg['smtp_server']", "#todo 仮でこのエラーは名前に入れている error_msg['mail_disp_name'] += get_message('MOSJA27215', request.user.get_lang_mode()) + '\\n' logger.user_log('LOSM07005', rq['smtp_server'],", "for key_value in defs.SMTP_PROT...
[ "-> Dog: return self.dogs.remove() def dequeCat(self) -> Cat: return self.cats.remove()", "= self.tail.next_node def remove(self) -> Any: if self.head is None:", "None def __str__(self) -> str: current = self.head string =", "__str__(self) -> str: return f\"{self.name}\" class Node: def __init__(self, data:",...
[ "Any datum kwargs that should go to all children. \"\"\"", "cam __all__ = ['DetectorBase', 'AreaDetector', 'AdscDetector', 'Andor3Detector', 'AndorDetector', 'BrukerDetector', 'DexelaDetector',", "for the hardware-specific classes that follow. Note that Plugin also", "_html_docs = ['EVTDoc.html'] cam = C(cam....
[ "(m): ')) area(f'A área do seu terreno {l}X{c} é de", "(m): ')) c = float(input('Comprimento (m): ')) area(f'A área do", "def area(msg):#declaracao da funcao com o parametro msg print(msg)#aqui msg", "funcao com o parametro msg print(msg)#aqui msg e a area", "20) l = float(input('Largura (m): ')) c = float(...
[ "html template.\"\"\" pyfile_path = os.path.dirname(os.path.abspath(__file__)) path = join_path(pyfile_path, 'templates', filename)", "Tuple with endpoints where access must not be checked. \"\"\"", "response = authorizer.validate() return response def authorize_endpoint(function): @wraps(function) def authoriz...
[ "name='_id', field=models.AutoField(editable=False, primary_key=True, serialize=False), ), migrations.AlterField( model_name='orderedproduct', name='_id', field=models.AutoField(editable=False, primary_key=True,", "name='dateTimeCreated', ), migrations.AlterField( model_name='order', name='_id', field=models.Auto...
[ "('pages', models.PositiveIntegerField(blank=True, default=0, null=True)), ('ddc', models.CharField(blank=True, default='', max_length=1024)), ('llcc', models.CharField(blank=True,", "('description', models.TextField(blank=True, default='')), ('publisher', models.CharField(blank=True, default='', max_length=512))...
[ "<filename>compliance_suite/exceptions/user_config_exception.py<gh_stars>1-10 # -*- coding: utf-8 -*- \"\"\"Module compliance_suite.exceptions.user_config_exception.py This module", "module contains class definition for user config file exceptions. \"\"\"", "# -*- coding: utf-8 -*- \"\"\"Module compliance_suite...
[ "self._ncols @property def nrows (self): \"\"\"The number of rows in", "RiskMap.load(filename) start = Point(0, 0) end = Point(rmap.ncols - 1,", "def ncols (self): \"\"\"The number of columns in this `RiskMap`.\"\"\"", "pos.x in range(0, self.ncols) def search (rmap, start, end): \"\"\"Searches", "position ...
[ "modify it under the terms of the MIT License; see", "it and/or # modify it under the terms of the", "software; you can redistribute it and/or # modify it under", "# modify it under the terms of the MIT License;", "free software; you can redistribute it and/or # modify it", "terms of the MIT License; see ...
[ "64 encoded value of the primary read-write key. :vartype primary_master_key:", "str :ivar secondary_master_key: Base 64 encoded value of the secondary", "key. :vartype secondary_readonly_master_key: str :ivar primary_master_key: Base 64 encoded value", "key. :vartype secondary_master_key: str \"\"\" _validat...
[ "link, \"displayed_link\": displayed_link, \"snippet\": snippet, \"author\": author, \"author_link\": author_link, \"date_published\":", "a.yKioRe:nth-child(1)::attr(href)\").get() more_editions_link = book_result.css(\".R1n8Q a.yKioRe:nth-child(2)::attr(href)\").get() books_results.append({ \"title\": title, \"l...
[ "analysis on different games. \"\"\" from hol import ( cards,", "<reponame>AustinTSchaffer/DailyProgrammer<filename>Python/Higher-Or-Lower/hol/__init__.py r\"\"\" Contains classes and methods that can be used", "be used when simulating the game Higher-or-Lower and performing statistical", "constants, ) from h...
[ "Box( (x, y, self.column_1 + self.column_2, self.text_height)) self.w.box.text = TextBox(", "\"update\", callback=self.update_font_callback, sizeStyle=self.size_style) # x slider x = self.padding_x y", "= self.w.x_slider.get() yValue = self.w.y_slider.get() x = self._moveX - xValue", "def slide_callback(self,...
[ "(str(p[0]), str(p[1])) for p in _params]).encode(\"utf-8\") sign = sha1(sign).hexdigest() sign_type", "return value if isinstance(value, six.text_type): return value.encode(encoding) return six.binary_type(value) def", "make_error_page(url): with io.open( os.path.join(os.path.dirname(__file__), 'contrib/error....
[ "from tensorflow.python.compiler.tensorrt import trt_convert as trt # pylint: enable=g-import-not-at-top #", "from tensorflow.python.ops.ragged import ragged_operators as _ragged_operators from tensorflow.python.ops.random_ops import *", "pylint: enable=redefined-builtin from tensorflow.python.eager import wrap...
[ "disable=global-statement global START_TIME db = tango.Database() elapsed = time.time() -", "num_devices)) print('>> Time taken to start devices: {:.4f} s ({:.4f}", "print('- Register devices: {:.4f} s ({:.4f} s/device)' .format(elapsed, elapsed /", "exported_devices = list(db.get_device_exported('test/*')) p...
[ "\"Sample text for document-2\", 'meta': {\"source\": \"wiki2\"}}, {\"text\": \"Sample text", "assert output[\"answers\"][0][\"answer\"].startswith(\"Using tests\") if isinstance(document_store, ElasticsearchDocumentStore): output = pipeline.run(query=\"How to", "prediction is not None assert prediction[\"query...
[ "polish_iterations: break if check_exist(asm_cns): return asm_cns else: print(\"polishing failed for", "in read_list: output.write(read + \"\\n\") else: # TODO: think about", "variant reads sv_reads = sv_reads_dir + \"/contig\" + str(k) sv_reads_rename", "None if check_exist(consensus): consensus_rename = os....
[ "lib._parse_constant(composite, valid=FAMILIES, valid_modifiers=VIAS) assert parsed == expected def test_parse_constant_fails(): \"\"\"", "pylint: disable=pointless-statement ses.create(\"session1\") assert ses.session_pointer is not None ses.destroy() with", "GMTTempFile() as outfile: lib.call_module(\"info\",...
[ "use these features, so I don't think its worth including", "We don't use these features, so I don't think its", "can always change it later. class ExceptionInfo: @property def value(self)", "# This class actually has more functions than are specified", "\"\"\"Type stubs for _pytest._code.\"\"\" # This clas...
[ "prime number as not prime return primes # returns the", "by a particular prime number prime_factors.append(prime) # add the prime", "0: # checks if the number is divisible by a", "prime numbers while n != 1: # keeps diving the", "prime number else: break # if the number is not", "# loops over all multipl...
[ "def _extended_gcd(self, a: int, b: int) -> tuple[int, int, int]:", "self._simple_new(_empty_range) first = self._range[::-1] if self.step < 0 else self._range", "array is saved in ``_cache``. \"\"\" return np.arange(self.start, self.stop, self.step,", "(attr, formatted_value) \"\"\" attrs = self._get_data_as...
[ "forward(self, x): out = self.l1(x) out = self.relu(out) out =", "self.relu(out) out = self.l4(out) # no activation and no softmax", "out = self.relu(out) out = self.l4(out) # no activation and", "out = self.l2(out) out = self.relu(out) out = self.l3(out) out", "nn.Linear(hidden_size, hidden_size) self.l3 =...
[ "file named ``config.json`` placed in the ``utils`` directory. References ----------", "name. The name of the ``log_file`` is a combination of", "os import logging from jwql.logging.logging_functions import configure_logging from jwql.logging.logging_functions import", "<NAME>, 2013 (WFC3 QL Version) Use --- ...
[ "typing import Optional session: Optional[ClientSession] = None __all__ = (session,)", "from typing import Optional session: Optional[ClientSession] = None __all__ =", "from aiohttp import ClientSession from typing import Optional session: Optional[ClientSession]", "aiohttp import ClientSession from typing im...
[ "_md5.hexdigest() def sha1(path): _sha1 = hashlib.sha1() fh = open(path, 'rb')", "= open(path, 'rb') if start > 0: fh.seek(start) if stop", "os.path.getsize(path) pos = start while pos < stop: size =", "= start while pos < stop: size = min(CHUNK, stop", "break _sha256.update(chunk) fh.close() return _sha256...