text
stringlengths
67
7.88k
<|fim_prefix|>def <|fim_suffix|>(self): # Test example from page 4 of # https://www.haystack.mit.edu/tech/vlbi/mark5/docs/230.3.pdf stream_hex = '0000 002D 0330 0000' + 'FFFF FFFF' + '4053 2143 3805 5' self.crc_hex = '284' self.crc12 = CRC(0x180f) self.crcstack12 = CRCStack(0x180f) self.stre...
<|fim_prefix|>def <|fim_suffix|>(tokenizer, train_batch_size, eval_batch_size): # Load dataset train_dataset, test_dataset = load_dataset("imdb", split=["train", "test"]) # Preprocess train dataset train_dataset = train_dataset.map( lambda e: tokenizer(e["text"], truncation=True, padding="max_le...
<|fim_prefix|>def <|fim_suffix|>(self): return self.METHOD_NAME<|fim_middle|>auto_decompress<|file_separator|>
<|fim_prefix|>async def <|fim_suffix|>(data, objectService): """We can upload a small amount of literal data""" name = f"taskcluster/test/client-py/{taskcluster.slugid.v4()}" await upload.uploadFromBuf( projectId="taskcluster", name=name, contentType="text/plain", contentLeng...
<|fim_prefix|>def <|fim_suffix|>(self, name): return self.has_callback(name) and self.num_callbacks(name) > 0<|fim_middle|>will_callback<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(y: Float64Array, x: Float64Array) -> Float64Array: """ Projection of y on x from y Parameters ---------- y : ndarray Array to project (nobs by nseries) x : ndarray Array to project onto (nobs by nvar) Returns ------- ndarray Pr...
<|fim_prefix|>def <|fim_suffix|>(current_actor_context): """ Report if there's configuration for a type we don't recognize. """ current_actor_context.feed(IfCfg( filename="/NM/ifcfg-pigeon0", properties=(IfCfgProperty(name="TYPE", value="AvianCarrier"),) )) current_actor_context....
<|fim_prefix|>def <|fim_suffix|>(self): """test label formatter for histogram for None""" formatted_label = self.histogram._format_bin_labels(None) assert formatted_label == "null and up"<|fim_middle|>test_histogram_label_formatter_none<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self): return all( self.datasets[key].METHOD_NAME for key in self.datasets )<|fim_middle|>supports_fetch_outside_dataloader<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>() -> None: """ Main function """ # Dir path if len(sys.argv) != 2: sys.stderr.write(f'Usage: {sys.argv[0]} <setup_dir>\n') exit(2) dirpath = sys.argv[1] # Dir non existence if os.path.exists(dirpath): sys.stderr.write(f'Directory: ...
<|fim_prefix|>def <|fim_suffix|>(scene): img = ImageMobject( np.uint8([[63, 0, 0, 0], [0, 127, 0, 0], [0, 0, 191, 0], [0, 0, 0, 255]]), ) img.height = 2 img1 = img.copy() img2 = img.copy() img3 = img.copy() img4 = img.copy() img5 = img.copy() img1.set_resampling_algorithm(RES...
<|fim_prefix|>def <|fim_suffix|>(devName): devName = resolveDevName(devName) return os.path.exists(os.path.join("/sys/devices/virtual/block/", devName))<|fim_middle|>is_virtual_device<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(elec_txt_dataframe): """Parse keys from EIA series_id string.""" input_ = elec_txt_dataframe.iloc[[2], :] expected = pd.DataFrame( { "series_code": ["RECEIPTS_BTU"], "fuel_agg": ["NG"], "geo_agg": ["US"], "sector_agg": ...
<|fim_prefix|>def <|fim_suffix|>(self, context, data_dict, fields_types, query_dict): '''Modify queries made on datastore_delete The overall design is that every IDatastore extension will receive the ``query_dict`` with the modifications made by previous extensions, then it can add/remove stuff into it ...
<|fim_prefix|>f <|fim_suffix|>(self):<|fim_middle|>test_counter<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(plugins: list[dict[str, str]]) -> list[dict[str, str]]: plugins_dict: dict[str, dict[str, str]] = {} for plugin in plugins: # Plugins from later indexes override others plugin["name"] = plugin["name"].lower() plugins_dict[plugin["name"]] = plugin retu...
<|fim_prefix|>def <|fim_suffix|>(self): self.file.METHOD_NAME()<|fim_middle|>close<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(train_data, train_labels, predict_data, nClasses): # Create an algorithm object and call compute train_algo = d4p.bf_knn_classification_training(nClasses=nClasses, fptype="float") train_result = train_algo.METHOD_NAME(train_data, train_labels) # Create an algorithm objec...
<|fim_prefix|>def <|fim_suffix|>(self): request = Mock(method="BADWOLF") view = Mock() obj = Mock() perm_obj = KolibriAuthPermissions() self.assertFalse(perm_obj.has_object_permission(request, view, obj))<|fim_middle|>test_bad_request_method<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self, *args, **kwargs): public_doc_records = public_doc_query() for doc in public_doc_records: pdf_name = create_name(doc) # We don't want to delete the original so we are not moving and we can't rename and copy at the same time try: temp_name...
<|fim_prefix|>def <|fim_suffix|>(bucket, file_name): """ Exposes helper method without breaking existing bindings/dependencies """ return storage_service_key_source_function(bucket, file_name)<|fim_middle|>storage_service_key<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(argv, log=False): """Call nvcc with arguments assembled from argv. Args: argv: A list of strings, possibly the argv passed to main(). log: True if logging is requested. Returns: The return value of calling os.system('nvcc ' + args) """ src_files = [f for f in argv ...
<|fim_prefix|>def <|fim_suffix|>(fmt, *args): s = fmt.format(*args) return (s, len(s))<|fim_middle|>wiki_format_raw<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(list_name, msgid): return "{}/arch/msg/{}/{}".format(settings.MAILING_LIST_ARCHIVE_URL, list_name, hash_list_message_id(list_name, msgid))<|fim_middle|>construct_message_url<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self, model): data = model.db.get_unit_operation_parameters("co2_addition") model.fs.unit.load_parameters_from_database() assert model.fs.unit.energy_electric_flow_vol_inlet.fixed assert ( model.fs.unit.energy_electric_flow_vol_inlet.value == data["energy...
<|fim_prefix|>def <|fim_suffix|>(self, train_conf, vars_conf): new_opt_confs = [] for param_group in self.param_groups: assert ( param_group["contiguous_params"] != True ), "contiguous_params cannot be used in graph" optimizer_conf = train_conf.optimizer_conf.add() lr...
<|fim_prefix|>def <|fim_suffix|>(self): """Return the precached index of the object. :rtype: int """ # Get the index of the object in its precache table METHOD_NAME = string_tables[self.precache_table][self._path] # Is the object precached? if METHOD_NAME != INVALID_STRING_INDEX: # R...
<|fim_prefix|>def <|fim_suffix|>(self, method, device): """Test pershot save density matrix instruction""" backend = self.backend(method=method, device=device) # Stabilizer test circuit circ = QuantumCircuit(1) circ.x(0) circ.reset(0) circ.h(0) circ.sdg(0) # Target statevector ta...
<|fim_prefix|>def <|fim_suffix|>(self, path, mode): path = path.decode(self.encoding) _evil_name(path) return errno_call(self._context, os.METHOD_NAME, path, mode)<|fim_middle|>chmod<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self) -> str: """ Resource Name. """ return pulumi.get(self, "name")<|fim_middle|>name<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self): return bool(self.__flags & DEF_BOUND)<|fim_middle|>is_local<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(cfg, in_channels): return RetinaNetModule(cfg, in_channels)<|fim_middle|>build_retinanet<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(pip_version=None): # type: (Optional[PipVersionValue]) -> bool pip_ver = pip_version or PipVersion.DEFAULT return pip_ver.version < Version("23.2")<|fim_middle|>supports_legacy_resolver<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self,nb,data=None): tab_num = nb.get_current_page() tab = nb.get_nth_page(tab_num) alloc = tab.get_allocation() x, y, w, h = (alloc.x, alloc.y, alloc.width, alloc.height) pixbuf = self.svg.get_pixbuf_sub(f'#layer{tab_num}').scale_simple(w-10, h-10, GdkPixbuf.InterpTy...
<|fim_prefix|>def <|fim_suffix|>(d: datetime) -> datetime: # There are two types of datetime objects in Python: naive and aware # Assume any dbt snapshot timestamp that is naive is meant to represent UTC if d is None: return d elif is_aware(d): return d else: return d.replace...
<|fim_prefix|>def <|fim_suffix|>(cql, namespaces, fes_version='1.0'): """transforms Common Query Language (CQL) query into OGC fes1 syntax""" filters = [] tmp_list = [] logical_op = None LOGGER.debug('CQL: %s', cql) if fes_version.startswith('1.0'): element_or = 'ogc:Or' element_...
<|fim_prefix|>def <|fim_suffix|>(redis, key: str) -> Any: """ Gets a JSON serialized value from the cache. """ cached_value = redis.get(key) if cached_value: logger.debug(f"Redis Cache - hit {key}") try: deserialized = json.loads(cached_value) return deseriali...
<|fim_prefix|>def <|fim_suffix|>(self) -> None: self.store.METHOD_NAME() try: flask.current_app.config.METHOD_NAME() except RuntimeError: pass<|fim_middle|>clear<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self, description, uuids=None, report=None): """Check that the delta description is correct.""" report = report if report is not None else self.report uuids = sorted(uuids or [REPORT_ID, NOTIFICATION_DESTINATION_ID]) self.assertEqual({"uuids": uuids, "email": self.email,...
<|fim_prefix|>def <|fim_suffix|>(mocker, cobbler_api): # Arrange mock_builtins_open = mocker.patch("builtins.open", mocker.mock_open()) mock_system = System(cobbler_api) mock_system.name = "test_manager_regen_ethers_system" mock_system.interfaces = {"default": NetworkInterface(cobbler_api)} mock...
<|fim_prefix|>async def <|fim_suffix|>(self, key, user=None, is_iter=False): """ Process a dict lookup message """ logging.debug("Looking up {} for {}".format(key, user)) orig_key = key # Priv and shared keys are handled slighlty differently key_type, key = key.decode("utf8").split("/", 1) t...
<|fim_prefix|>def <|fim_suffix|>(self): """Check that all foreign keys are created correctly""" # filter questions on survey questions = list(StudentQuestionBase.objects.filter(survey=self.survey)) self.assertTrue(self.grading_q.pk in [q.pk for q in questions]) self.assertTrue(self.slider_q.pk in [q...
<|fim_prefix|>def <|fim_suffix|>(self): self.args = [ "command_dummy", "--host", TESTSERVER_URL + "/service?request=GetCapabilities", ] with open(SERVICE_EXCEPTION_FILE, "rb") as fp: capabilities_doc = fp.read() with mock_httpd( TESTSERVER_ADDRESS, ...
<|fim_prefix|>def <|fim_suffix|>(self): if LXML_PRESENT: self.assertEqual(registry.lookup('html'), LXMLTreeBuilder) self.assertEqual(registry.lookup('xml'), LXMLTreeBuilderForXML) else: self.assertEqual(registry.lookup('xml'), None) if HTML5LIB_PRESENT: self.assertEqu...
<|fim_prefix|>def <|fim_suffix|>(n_fp, order): """ Prepare DOF permutation vector for each possible facet orientation. """ from sfepy.base.base import dict_to_array if n_fp == 2: mtx = make_line_matrix(order) ori_map = ori_line_to_iter fo = order - 1 elif n_fp == 3: ...
<|fim_prefix|>def <|fim_suffix|>(self) -> List[NamedUser]: ...<|fim_middle|>assignees<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(): model = SemanticSegmentation(2) model.eval() model.serve()<|fim_middle|>test_serve<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self): """Returns the string representation of the model""" return pprint.pformat(self.to_dict())<|fim_middle|>to_str<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self, max_norm, aggregate_norm_fn=None): """Clips gradient norm.""" return utils.clip_grad_norm_(self.params, max_norm, aggregate_norm_fn)<|fim_middle|>clip_grad_norm<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(agent_args, technologies, stock): from muse.agents.agent import Agent from muse.agents.factories import create_agent agent_args["share"] = "agent_share_zero" agent = create_agent( agent_type="Retrofit", technologies=technologies, capacity=stock.ca...
<|fim_prefix|>def <|fim_suffix|>(self, handler: ErrorHandler[object] | None) -> None: ...<|fim_middle|>set_error_handler<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self): pass<|fim_middle|>pre_operations<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>( repo, settings, fs=None, prefix: Optional[Tuple[str, ...]] = None, hash_name: Optional[str] = None, **kwargs, ): from dvc.fs import get_cloud_fs if not settings: return None cls, config, fs_path = get_cloud_fs(repo.config, **settings) fs ...
<|fim_prefix|>def <|fim_suffix|>(self): pass<|fim_middle|>pre_operations<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self, skip_run=False, executable=None): # start with the run command cmd = [] if skip_run else self.build_run_cmd(executable=executable) # add arguments and insert dummary key value separators which are replaced with "=" later for key, value in self.args: cmd.ext...
<|fim_prefix|>def <|fim_suffix|>(self): raise ValueError("this dispatcher is not writable")<|fim_middle|>handle_write_event<|file_separator|>
<|fim_prefix|>f <|fim_suffix|>(self):<|fim_middle|>get_zone<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self): """Exporter EntryPoint to call.""" return self.get_record_value('entry_point')<|fim_middle|>entry_point<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(scope): if scope == "list": return [None] elif scope in ["create", "import:backup"]: return [ { "owner": {"id": random.randrange(400, 500)}, "assignee": {"id": random.randrange(500, 600)}, "organization"...
<|fim_prefix|>def <|fim_suffix|>(input_event, relation, text_encoder, max_e1, max_r, force): abort = False e1_tokens, rel_tokens, _ = data.conceptnet_data.do_example(text_encoder, input_event, relation, None) if len(e1_tokens) > max_e1: if force: XMB = torch.zeros(1, len(e1_tokens) + max...
<|fim_prefix|>def <|fim_suffix|>(self): return self.maxsize > 0 and len(self.queue) == self.maxsize<|fim_middle|>full<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self, item, identity_field=None): identity = {} from_ = item if isinstance(item, dict) and 'data' in item: from_ = item['data']['message'][identity_field] identity['username'] = from_.get('username', None) identity['email'] = None identity['name'] = from_...
<|fim_prefix|>def <|fim_suffix|>(self): return self.client.format_url( "/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Cache/redisEnterprise/{clusterName}/databases/{databaseName}", **self.url_parameters )<|fim_middle|>url<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(test_file): with Image.open(test_file) as im: assert im.tell() == 0 # prior to first image raises an error, both blatant and borderline with pytest.raises(EOFError): im.seek(-1) with pytest.raises(EOFError): im.seek(-523) ...
<|fim_prefix|>def <|fim_suffix|>(): """See base_runner.""" return True<|fim_middle|>has_heartbeat<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(configuration_policy_group_name: Optional[pulumi.Input[str]] = None, resource_group_name: Optional[pulumi.Input[str]] = None, vpn_server_configuration_name: Optional[pulumi.Input[str]] = None, ...
<|fim_prefix|>def <|fim_suffix|>(wn, tn): if tn is None: # special case. i.e. the last tile. #dont know why it is a special case, if the ALL-1 is received with a tile #then it should always be one, if the tile is consired received by the method #below, then it should not be false ...
<|fim_prefix|>def <|fim_suffix|>( self, painter: QtGui.QPainter, option: QtWidgets.QStyleOptionGraphicsItem, # pylint: disable=unused-argument widget: Optional[QtWidgets.QWidget] = ..., ): # pylint: disable=unused-argument self._paint_boundary(painter)<|fim_middle|>paint<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self): """Configure the DUT (Router in our case) which we use for testing the EMU functionality prior to the test""" sys.stdout.flush() if not CTRexScenario.router_cfg['no_dut_config']: sys.stdout.write('Configuring DUT... ') start_time = time.time() ...
<|fim_prefix|>def <|fim_suffix|>(self, peer_id: ID) -> PublicKey: """ :param peer_id: peer ID to get public key for :return: public key of the peer :raise PeerStoreError: if peer ID not found """<|fim_middle|>pubkey<|file_separator|>
<|fim_prefix|>f <|fim_suffix|>(self, flag):<|fim_middle|>set_wires<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self, get_ipython, clear_output): """ Context manager that monkeypatches get_ipython and clear_output """ original_clear_output = widget_output.clear_output original_get_ipython = widget_output.get_ipython widget_output.get_ipython = get_ipython widget_output.clear_o...
<|fim_prefix|>def <|fim_suffix|>(): with pytest.raises(ImportError): from ddtrace.constants import INVALID_CONSTANT # noqa<|fim_middle|>test_invalid<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(lnst_config): for preset_name in PRESETS: preset = lnst_config.get_option("colours", preset_name) if preset == None: continue fg, bg, bf = preset extended_re = "^extended\([0-9]+\)$" if fg == "default": fg = None ...
<|fim_prefix|>def <|fim_suffix|>(self) -> LoggingMode: """ Get the logger's mode. :return: The logger mode. """ return self._mode<|fim_middle|>mode<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(data): """ Calculates the CRC32C checksum of the provided data. Args: data: the bytes over which the checksum should be calculated. Returns: An int representing the CRC32C checksum of the provided bytes. """ return _crc32c(six.ensure_binary(data))...
<|fim_prefix|>def <|fim_suffix|>(self, model): pass # Not needed<|fim_middle|>add_type<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self): size = 100 matrix = torch.randn(size, size, dtype=torch.float64) matrix = matrix.matmul(matrix.mT) matrix.div_(matrix.norm()) matrix.add_(torch.eye(matrix.size(-1), dtype=torch.float64).mul_(1e-1)) # set up vector rhs rhs = torch.randn(size, dtype=torc...
<|fim_prefix|>def <|fim_suffix|>(): if Promise.unhandled_exceptions: for exctype, value, tb in Promise.unhandled_exceptions: if value: raise value.with_traceback(tb) # traceback.print_exception(exctype, value, tb)<|fim_middle|>rereaise_unhandled<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(src_lines, var_name, lines): add_comments_header(lines) lines.append("const char *{0} =".format(var_name)) for src_line in src_lines: lines.append('"' + cpp_escape(src_line) + "\\n\"") lines[len(lines) - 1] += ';' add_comments_footer(lines)<|fim_middle|>add_cpp_string_literal<|fi...
<|fim_prefix|>def <|fim_suffix|>(self): return os.path.join(self.save_path, 'train')<|fim_middle|>train_path<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(): mc = MailChimp(mc_api=app.config["MAILCHIMP_KEY"]) try: email = request.form.get("email") try: data = mc.lists.members.create_or_update( list_id=app.config["MAILCHIMP_LIST"], subscriber_hash=get_subscriber_hash(email...
<|fim_prefix|>def <|fim_suffix|>(): vip = utils.config_get('vip') vip_iface = utils.config_get('vip_iface') vip_cidr = utils.config_get('vip_cidr') corosync_bindiface = utils.config_get('ha-bindiface') corosync_mcastport = utils.config_get('ha-mcastport') if None in [vip, vip_cidr, vip_iface]: ...
<|fim_prefix|>def <|fim_suffix|>( query: str, offset: int, length: int, expected_result: Any, expected_num_rows_total: int ) -> None: # simulate index file index_file_location = "index.duckdb" con = duckdb.connect(index_file_location) con.execute("INSTALL 'httpfs';") con.execute("LOAD 'httpfs';"...
<|fim_prefix|>async def <|fim_suffix|>( db: AsyncSession = Depends(get_async_db), form_data: OAuth2PasswordRequestForm = Depends(),<|fim_middle|>login_for_access_token<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(cls): super().METHOD_NAME() cls.sth_related = SomethingRelated.objects.create(name='Rel1') cls.pol_1 = PolymorphicModelTest.objects.create( name='Pol1', sth_related=cls.sth_related ) cls.pol_2 = PolymorphicModelTest.objects.create( name='Pol2'...
<|fim_prefix|>def <|fim_suffix|>(self): return []<|fim_middle|>get_keywords_to_be_excluded<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self): factory = RequestFactory() bad_referers = ( "http://otherdomain/bar/", "http://otherdomain/admin/forms/form/", ) for referer in bad_referers: with self.subTest(referer=referer): request = factory.get( "/api/v1/fo...
<|fim_prefix|> <|fim_suffix|>(self):<|fim_middle|>run<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self): 'od.keys() -> list of keys in od' return list(self)<|fim_middle|>keys<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(minion_id, package_name, state_tree): module_contents = """ def get_test_package_name(): return "{}" """.format( package_name ) top_file_contents = """ base: {}: - install-package """.format( minion_id ) insta...
<|fim_prefix|>def <|fim_suffix|>(vineyard_client): df = pd.DataFrame({'a': [1, 2, 3, 4], 'b': [5, 6, 7, 8]}) object_id = vineyard_client.put(df) pd.testing.assert_frame_equal(df, vineyard_client.get(object_id))<|fim_middle|>test_pandas_dataframe<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self) -> str: """ Resource type. """ return pulumi.get(self, "type")<|fim_middle|>type<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self): """ Test rectangular transfer of self.d_array1 to self.d_array2 """ # Reference o1 = self.offset1 o2 = self.offset2 T = self.transfer_shape logger.info("""Testing D->D rectangular copy with (N1_y, N1_x) = %s, (N2_y, N2_x) = %s: ...
<|fim_prefix|>def <|fim_suffix|>(self, view_size): if self.selection_valid(): before_selection = view_size // 2 self.view.start = max(0, self.selection - before_selection) self.view.end = self.view.start self.max_view_size = view_size self._expand_view() else: sel...
<|fim_prefix|>def <|fim_suffix|>(event): """ Control-P in vi edit mode on readline is history next, unlike default prompt toolkit. If completer is open this still select previous completion. """ event.current_buffer.auto_up()<|fim_middle|>previous_history_or_previous_completion<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>(self, X, queue=None): y = super()._predict(X, _backend.linear_model.regression, queue) return y<|fim_middle|>predict<|file_separator|>
<|fim_prefix|>def <|fim_suffix|>( record_property, get_host_key, setup_splunk, setup_sc4s, event ): host = get_host_key dt = datetime.datetime.now(datetime.timezone.utc) iso, _, _, _, _, _, epoch = time_operations(dt) # Tune time functions iso = dt.isoformat()[0:23] epoch = epoch[:-3] m...
<|fim_prefix|>def <|fim_suffix|>(self) -> None: warnings.filterwarnings("default", category=DeprecationWarning) warnings.filterwarnings("ignore", category=DeprecationWarning, module="pyscf") warnings.filterwarnings(action="ignore", category=DeprecationWarning, module=".*drivers*") warnings.filterwarning...
<|fim_prefix|>def <|fim_suffix|>(self, channel, on): pass<|fim_middle|>channel_on_off<|file_separator|>